Home
last modified time | relevance | path

Searched refs:nodemask (Results 1 – 25 of 31) sorted by relevance

12

/linux/mm/
H A Dshow_mem.c119 static bool show_mem_node_skip(unsigned int flags, int nid, nodemask_t *nodemask) in show_mem_node_skip() argument
129 if (!nodemask) in show_mem_node_skip()
130 nodemask = &cpuset_current_mems_allowed; in show_mem_node_skip()
132 return !node_isset(nid, *nodemask); in show_mem_node_skip()
180 static void show_free_areas(unsigned int filter, nodemask_t *nodemask, int max_zone_idx) in show_free_areas() argument
190 if (show_mem_node_skip(filter, zone_to_nid(zone), nodemask)) in show_free_areas()
227 if (show_mem_node_skip(filter, pgdat->node_id, nodemask)) in show_free_areas()
290 if (show_mem_node_skip(filter, zone_to_nid(zone), nodemask)) in show_free_areas()
361 if (show_mem_node_skip(filter, zone_to_nid(zone), nodemask)) in show_free_areas()
391 if (show_mem_node_skip(filter, nid, nodemask)) in show_free_areas()
402 __show_mem(unsigned int filter,nodemask_t * nodemask,int max_zone_idx) __show_mem() argument
[all...]
H A Dhugetlb_cma.h8 int nid, nodemask_t *nodemask);
21 gfp_t gfp_mask, int nid, nodemask_t *nodemask) in hugetlb_cma_alloc_folio()
22 hugetlb_cma_alloc_folio(int order,gfp_t gfp_mask,int nid,nodemask_t * nodemask) hugetlb_cma_alloc_folio() argument
H A Dmempolicy.c1442 nodemask_t *nodemask; in alloc_migration_target_by_mpol() local
1447 nodemask = policy_nodemask(gfp, pol, ilx, &nid); in alloc_migration_target_by_mpol()
1448 return alloc_hugetlb_folio_nodemask(h, nid, nodemask, gfp, in alloc_migration_target_by_mpol()
2200 nodemask_t nodemask; in weighted_interleave_nid() local
2207 nr_nodes = read_once_policy_nodemask(pol, &nodemask); in weighted_interleave_nid()
2219 for_each_node_mask(nid, nodemask) in weighted_interleave_nid()
2224 nid = first_node(nodemask); in weighted_interleave_nid()
2231 nid = next_node_in(nid, nodemask); in weighted_interleave_nid()
2244 nodemask_t nodemask; in interleave_nid() local
2249 nnodes = read_once_policy_nodemask(pol, &nodemask); in interleave_nid()
[all …]
H A Doom_kill.c95 const nodemask_t *mask = oc->nodemask; in oom_cpuset_eligible()
282 * This is not a __GFP_THISNODE allocation, so a truncated nodemask in in constrained_alloc()
286 if (oc->nodemask && in constrained_alloc()
287 !nodes_subset(node_states[N_MEMORY], *oc->nodemask)) { in constrained_alloc()
289 for_each_node_mask(nid, *oc->nodemask) in constrained_alloc()
296 highest_zoneidx, oc->nodemask) in constrained_alloc()
317 /* p may not have freeable memory in nodemask */ in oom_evaluate_task()
390 /* p may not have freeable memory in nodemask */ in dump_task()
450 pr_info("oom-kill:constraint=%s,nodemask=%*pbl", in dump_oom_victim()
452 nodemask_pr_args(oc->nodemask)); in dump_oom_victim()
[all...]
H A Dpage_alloc.c3508 ac->nodemask) { in unreserve_highatomic_pageblock()
3827 ac->nodemask) { in get_page_from_freelist()
4005 static void warn_alloc_show_mem(gfp_t gfp_mask, nodemask_t *nodemask) in warn_alloc_show_mem() argument
4021 __show_mem(filter, nodemask, gfp_zone(gfp_mask)); in warn_alloc_show_mem()
4025 void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...) in warn_alloc() argument
4041 nodemask_pr_args(nodemask)); in warn_alloc()
4047 warn_alloc_show_mem(gfp_mask, nodemask); in warn_alloc()
4075 .nodemask = ac->nodemask, in __alloc_pages_may_oom()
4317 ac->highest_zoneidx, ac->nodemask) { in should_compact_retry()
4425 ac->nodemask); in __perform_reclaim()
[all …]
H A Dhugetlb_cma.c28 int nid, nodemask_t *nodemask)
41 for_each_node_mask(node, *nodemask) { in hugetlb_cma_alloc_folio()
30 hugetlb_cma_alloc_folio(int order,gfp_t gfp_mask,int nid,nodemask_t * nodemask) hugetlb_cma_alloc_folio() argument
H A Dhugetlb.c1380 nodemask_t *nodemask; in dequeue_hugetlb_folio_vma() local
1391 nid = huge_node(vma, address, gfp_mask, &mpol, &nodemask); in dequeue_hugetlb_folio_vma()
1395 nid, nodemask); in dequeue_hugetlb_folio_vma()
1398 nodemask = NULL; in dequeue_hugetlb_folio_vma()
1403 nid, nodemask); in dequeue_hugetlb_folio_vma()
1414 int nid, nodemask_t *nodemask) in alloc_gigantic_frozen_folio() argument
1418 folio = hugetlb_cma_alloc_frozen_folio(order, gfp_mask, nid, nodemask); in alloc_gigantic_frozen_folio()
1426 nid, nodemask); in alloc_gigantic_frozen_folio()
1431 nodemask_t *nodemask) in alloc_gigantic_frozen_folio() argument
2215 nodemask_t *nodemask; in alloc_buddy_hugetlb_folio_with_mpol() local
[all …]
H A Dnuma_memblks.c23 static void __init numa_nodemask_from_meminfo(nodemask_t *nodemask, in numa_nodemask_from_meminfo() argument
31 node_set(mi->blk[i].nid, *nodemask); in numa_nodemask_from_meminfo()
H A Dvmscan.c82 nodemask_t *nodemask; member
6243 sc->reclaim_idx, sc->nodemask) { in shrink_zones()
6373 sc->nodemask) { in do_try_to_free_pages()
6487 nodemask_t *nodemask) in throttle_direct_reclaim() argument
6525 gfp_zone(gfp_mask), nodemask) { in throttle_direct_reclaim()
6567 gfp_t gfp_mask, nodemask_t *nodemask) in try_to_free_pages() argument
6575 .nodemask = nodemask, in try_to_free_pages()
6595 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages()
H A Dcompaction.c2443 ac->highest_zoneidx, ac->nodemask) { in compaction_zonelist_suitable()
2829 ac->highest_zoneidx, ac->nodemask) { in try_to_compact_pages()
/linux/tools/testing/selftests/kvm/x86/
H A Dxapic_ipi_test.c254 unsigned long nodemask = 0; in do_migrations() local
255 unsigned long nodemasks[sizeof(nodemask) * 8]; in do_migrations()
270 kvm_get_mempolicy(NULL, &nodemask, sizeof(nodemask) * 8, in do_migrations()
275 sizeof(nodemask) * 8, nodemask); in do_migrations()
281 for (i = 0, bit = 1; i < sizeof(nodemask) * 8; i++, bit <<= 1) { in do_migrations()
282 if (nodemask & bit) { in do_migrations()
283 nodemasks[nodes] = nodemask & bit; in do_migrations()
/linux/tools/testing/selftests/kvm/
H A Dguest_memfd_test.c82 unsigned long nodemask = 0; in test_mbind() local
83 unsigned long maxnode = BITS_PER_TYPE(nodemask); in test_mbind()
95 kvm_get_mempolicy(&policy, &nodemask, maxnode, mem, MPOL_F_ADDR); in test_mbind()
96 TEST_ASSERT(policy == MPOL_INTERLEAVE && nodemask == nodemask_0, in test_mbind()
98 MPOL_INTERLEAVE, nodemask_0, policy, nodemask); in test_mbind()
102 kvm_get_mempolicy(&policy, &nodemask, maxnode, mem + page_size * 2, MPOL_F_ADDR); in test_mbind()
103 TEST_ASSERT(policy == MPOL_BIND && nodemask == nodemask_0, in test_mbind()
105 MPOL_BIND, nodemask_0, policy, nodemask); in test_mbind()
109 kvm_get_mempolicy(&policy, &nodemask, maxnode, mem, MPOL_F_ADDR); in test_mbind()
110 TEST_ASSERT(policy == MPOL_DEFAULT && !nodemask, in test_mbind()
[all …]
/linux/include/linux/
H A Dcpuset.h86 int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask);
166 static inline void set_mems_allowed(nodemask_t nodemask) in set_mems_allowed() argument
173 current->mems_allowed = nodemask; in set_mems_allowed()
227 static inline int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask) in cpuset_nodemask_valid_mems_allowed() argument
288 static inline void set_mems_allowed(nodemask_t nodemask) in set_mems_allowed() argument
H A Dgfp.h230 nodemask_t *nodemask);
234 nodemask_t *nodemask);
238 nodemask_t *nodemask, int nr_pages,
457 gfp_t gfp_mask, int nid, nodemask_t *nodemask);
462 int nid, nodemask_t *nodemask);
H A Dmempolicy.h144 struct mempolicy **mpol, nodemask_t **nodemask);
260 struct mempolicy **mpol, nodemask_t **nodemask) in huge_node() argument
263 *nodemask = NULL; in huge_node()
H A Doom.h33 nodemask_t *nodemask; member
H A Dmmzone.h1800 #define for_each_zone_zonelist_nodemask(zone, z, zlist, highidx, nodemask) \ argument
1801 for (z = first_zones_zonelist(zlist, highidx, nodemask), zone = zonelist_zone(z); \
1803 z = next_zones_zonelist(++z, highidx, nodemask), \
1806 #define for_next_zone_zonelist_nodemask(zone, z, highidx, nodemask) \ argument
1809 z = next_zones_zonelist(++z, highidx, nodemask), \
/linux/tools/testing/selftests/futex/functional/
H A Dfutex_numa_mpol.c183 unsigned long nodemask; in TEST() local
186 nodemask = 1 << i; in TEST()
187 ret = mbind(futex_ptr, mem_size, MPOL_BIND, &nodemask, in TEST()
188 sizeof(nodemask) * 8, 0); in TEST()
/linux/Documentation/admin-guide/mm/
H A Dnuma_memory_policy.rst215 is always preferred by passing an empty nodemask with this
216 mode. If an empty nodemask is passed, the policy cannot use
248 satisfied from the nodemask specified in the policy. If there is
249 a memory pressure on all nodes in the nodemask, the allocation
265 This flag specifies that the nodemask passed by
270 change in the set of allowed nodes, the preferred nodemask (Preferred
271 Many), preferred node (Preferred) or nodemask (Bind, Interleave) is
284 3 is allowed from the user's nodemask, the "interleave" only
285 occurs over that node. If no nodes from the user's nodemask are
290 MPOL_PREFERRED policies that were created with an empty nodemask
[all …]
/linux/Documentation/translations/zh_CN/core-api/
H A Dprintk-formats.rst532 位图及其衍生物,如cpumask和nodemask
540 对于打印位图(bitmap)及其派生的cpumask和nodemask,%*pb输出以字段宽度为位数的位图,
544 nodemask_pr_args()来方便打印cpumask和nodemask
/linux/tools/testing/selftests/kvm/include/
H A Dnumaif.h30 const unsigned long *, nodemask, unsigned long, maxnode,
/linux/tools/workqueue/
H A Dwq_dump.py53 from drgn.helpers.linux.nodemask import for_each_node
/linux/Documentation/core-api/
H A Dprintk-formats.rst587 bitmap and its derivatives such as cpumask and nodemask
595 For printing bitmap and its derivatives such as cpumask and nodemask,
601 printing cpumask and nodemask.
/linux/kernel/irq/
H A Dmanage.c623 const struct cpumask *nodemask = cpumask_of_node(node); in irq_setup_affinity() local
626 if (cpumask_intersects(&mask, nodemask)) in irq_setup_affinity()
627 cpumask_and(&mask, &mask, nodemask); in irq_setup_affinity()
/linux/drivers/tty/
H A Dsysrq.c390 .nodemask = NULL, in moom_callback()

12