| /linux/mm/ |
| H A D | show_mem.c | 119 static bool show_mem_node_skip(unsigned int flags, int nid, nodemask_t *nodemask) in show_mem_node_skip() argument 129 if (!nodemask) in show_mem_node_skip() 130 nodemask = &cpuset_current_mems_allowed; in show_mem_node_skip() 132 return !node_isset(nid, *nodemask); in show_mem_node_skip() 180 static void show_free_areas(unsigned int filter, nodemask_t *nodemask, int max_zone_idx) in show_free_areas() argument 190 if (show_mem_node_skip(filter, zone_to_nid(zone), nodemask)) in show_free_areas() 227 if (show_mem_node_skip(filter, pgdat->node_id, nodemask)) in show_free_areas() 290 if (show_mem_node_skip(filter, zone_to_nid(zone), nodemask)) in show_free_areas() 361 if (show_mem_node_skip(filter, zone_to_nid(zone), nodemask)) in show_free_areas() 391 if (show_mem_node_skip(filter, nid, nodemask)) in show_free_areas() 402 __show_mem(unsigned int filter,nodemask_t * nodemask,int max_zone_idx) __show_mem() argument [all...] |
| H A D | hugetlb_cma.h | 8 int nid, nodemask_t *nodemask); 21 gfp_t gfp_mask, int nid, nodemask_t *nodemask) in hugetlb_cma_alloc_folio() 22 hugetlb_cma_alloc_folio(int order,gfp_t gfp_mask,int nid,nodemask_t * nodemask) hugetlb_cma_alloc_folio() argument
|
| H A D | mempolicy.c | 1442 nodemask_t *nodemask; in alloc_migration_target_by_mpol() local 1447 nodemask = policy_nodemask(gfp, pol, ilx, &nid); in alloc_migration_target_by_mpol() 1448 return alloc_hugetlb_folio_nodemask(h, nid, nodemask, gfp, in alloc_migration_target_by_mpol() 2200 nodemask_t nodemask; in weighted_interleave_nid() local 2207 nr_nodes = read_once_policy_nodemask(pol, &nodemask); in weighted_interleave_nid() 2219 for_each_node_mask(nid, nodemask) in weighted_interleave_nid() 2224 nid = first_node(nodemask); in weighted_interleave_nid() 2231 nid = next_node_in(nid, nodemask); in weighted_interleave_nid() 2244 nodemask_t nodemask; in interleave_nid() local 2249 nnodes = read_once_policy_nodemask(pol, &nodemask); in interleave_nid() [all …]
|
| H A D | oom_kill.c | 95 const nodemask_t *mask = oc->nodemask; in oom_cpuset_eligible() 282 * This is not a __GFP_THISNODE allocation, so a truncated nodemask in in constrained_alloc() 286 if (oc->nodemask && in constrained_alloc() 287 !nodes_subset(node_states[N_MEMORY], *oc->nodemask)) { in constrained_alloc() 289 for_each_node_mask(nid, *oc->nodemask) in constrained_alloc() 296 highest_zoneidx, oc->nodemask) in constrained_alloc() 317 /* p may not have freeable memory in nodemask */ in oom_evaluate_task() 390 /* p may not have freeable memory in nodemask */ in dump_task() 450 pr_info("oom-kill:constraint=%s,nodemask=%*pbl", in dump_oom_victim() 452 nodemask_pr_args(oc->nodemask)); in dump_oom_victim() [all...] |
| H A D | page_alloc.c | 3508 ac->nodemask) { in unreserve_highatomic_pageblock() 3827 ac->nodemask) { in get_page_from_freelist() 4005 static void warn_alloc_show_mem(gfp_t gfp_mask, nodemask_t *nodemask) in warn_alloc_show_mem() argument 4021 __show_mem(filter, nodemask, gfp_zone(gfp_mask)); in warn_alloc_show_mem() 4025 void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...) in warn_alloc() argument 4041 nodemask_pr_args(nodemask)); in warn_alloc() 4047 warn_alloc_show_mem(gfp_mask, nodemask); in warn_alloc() 4075 .nodemask = ac->nodemask, in __alloc_pages_may_oom() 4317 ac->highest_zoneidx, ac->nodemask) { in should_compact_retry() 4425 ac->nodemask); in __perform_reclaim() [all …]
|
| H A D | hugetlb_cma.c | 28 int nid, nodemask_t *nodemask) 41 for_each_node_mask(node, *nodemask) { in hugetlb_cma_alloc_folio() 30 hugetlb_cma_alloc_folio(int order,gfp_t gfp_mask,int nid,nodemask_t * nodemask) hugetlb_cma_alloc_folio() argument
|
| H A D | hugetlb.c | 1380 nodemask_t *nodemask; in dequeue_hugetlb_folio_vma() local 1391 nid = huge_node(vma, address, gfp_mask, &mpol, &nodemask); in dequeue_hugetlb_folio_vma() 1395 nid, nodemask); in dequeue_hugetlb_folio_vma() 1398 nodemask = NULL; in dequeue_hugetlb_folio_vma() 1403 nid, nodemask); in dequeue_hugetlb_folio_vma() 1414 int nid, nodemask_t *nodemask) in alloc_gigantic_frozen_folio() argument 1418 folio = hugetlb_cma_alloc_frozen_folio(order, gfp_mask, nid, nodemask); in alloc_gigantic_frozen_folio() 1426 nid, nodemask); in alloc_gigantic_frozen_folio() 1431 nodemask_t *nodemask) in alloc_gigantic_frozen_folio() argument 2215 nodemask_t *nodemask; in alloc_buddy_hugetlb_folio_with_mpol() local [all …]
|
| H A D | numa_memblks.c | 23 static void __init numa_nodemask_from_meminfo(nodemask_t *nodemask, in numa_nodemask_from_meminfo() argument 31 node_set(mi->blk[i].nid, *nodemask); in numa_nodemask_from_meminfo()
|
| H A D | vmscan.c | 82 nodemask_t *nodemask; member 6243 sc->reclaim_idx, sc->nodemask) { in shrink_zones() 6373 sc->nodemask) { in do_try_to_free_pages() 6487 nodemask_t *nodemask) in throttle_direct_reclaim() argument 6525 gfp_zone(gfp_mask), nodemask) { in throttle_direct_reclaim() 6567 gfp_t gfp_mask, nodemask_t *nodemask) in try_to_free_pages() argument 6575 .nodemask = nodemask, in try_to_free_pages() 6595 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages()
|
| H A D | compaction.c | 2443 ac->highest_zoneidx, ac->nodemask) { in compaction_zonelist_suitable() 2829 ac->highest_zoneidx, ac->nodemask) { in try_to_compact_pages()
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | xapic_ipi_test.c | 254 unsigned long nodemask = 0; in do_migrations() local 255 unsigned long nodemasks[sizeof(nodemask) * 8]; in do_migrations() 270 kvm_get_mempolicy(NULL, &nodemask, sizeof(nodemask) * 8, in do_migrations() 275 sizeof(nodemask) * 8, nodemask); in do_migrations() 281 for (i = 0, bit = 1; i < sizeof(nodemask) * 8; i++, bit <<= 1) { in do_migrations() 282 if (nodemask & bit) { in do_migrations() 283 nodemasks[nodes] = nodemask & bit; in do_migrations()
|
| /linux/tools/testing/selftests/kvm/ |
| H A D | guest_memfd_test.c | 82 unsigned long nodemask = 0; in test_mbind() local 83 unsigned long maxnode = BITS_PER_TYPE(nodemask); in test_mbind() 95 kvm_get_mempolicy(&policy, &nodemask, maxnode, mem, MPOL_F_ADDR); in test_mbind() 96 TEST_ASSERT(policy == MPOL_INTERLEAVE && nodemask == nodemask_0, in test_mbind() 98 MPOL_INTERLEAVE, nodemask_0, policy, nodemask); in test_mbind() 102 kvm_get_mempolicy(&policy, &nodemask, maxnode, mem + page_size * 2, MPOL_F_ADDR); in test_mbind() 103 TEST_ASSERT(policy == MPOL_BIND && nodemask == nodemask_0, in test_mbind() 105 MPOL_BIND, nodemask_0, policy, nodemask); in test_mbind() 109 kvm_get_mempolicy(&policy, &nodemask, maxnode, mem, MPOL_F_ADDR); in test_mbind() 110 TEST_ASSERT(policy == MPOL_DEFAULT && !nodemask, in test_mbind() [all …]
|
| /linux/include/linux/ |
| H A D | cpuset.h | 86 int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask); 166 static inline void set_mems_allowed(nodemask_t nodemask) in set_mems_allowed() argument 173 current->mems_allowed = nodemask; in set_mems_allowed() 227 static inline int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask) in cpuset_nodemask_valid_mems_allowed() argument 288 static inline void set_mems_allowed(nodemask_t nodemask) in set_mems_allowed() argument
|
| H A D | gfp.h | 230 nodemask_t *nodemask); 234 nodemask_t *nodemask); 238 nodemask_t *nodemask, int nr_pages, 457 gfp_t gfp_mask, int nid, nodemask_t *nodemask); 462 int nid, nodemask_t *nodemask);
|
| H A D | mempolicy.h | 144 struct mempolicy **mpol, nodemask_t **nodemask); 260 struct mempolicy **mpol, nodemask_t **nodemask) in huge_node() argument 263 *nodemask = NULL; in huge_node()
|
| H A D | oom.h | 33 nodemask_t *nodemask; member
|
| H A D | mmzone.h | 1800 #define for_each_zone_zonelist_nodemask(zone, z, zlist, highidx, nodemask) \ argument 1801 for (z = first_zones_zonelist(zlist, highidx, nodemask), zone = zonelist_zone(z); \ 1803 z = next_zones_zonelist(++z, highidx, nodemask), \ 1806 #define for_next_zone_zonelist_nodemask(zone, z, highidx, nodemask) \ argument 1809 z = next_zones_zonelist(++z, highidx, nodemask), \
|
| /linux/tools/testing/selftests/futex/functional/ |
| H A D | futex_numa_mpol.c | 183 unsigned long nodemask; in TEST() local 186 nodemask = 1 << i; in TEST() 187 ret = mbind(futex_ptr, mem_size, MPOL_BIND, &nodemask, in TEST() 188 sizeof(nodemask) * 8, 0); in TEST()
|
| /linux/Documentation/admin-guide/mm/ |
| H A D | numa_memory_policy.rst | 215 is always preferred by passing an empty nodemask with this 216 mode. If an empty nodemask is passed, the policy cannot use 248 satisfied from the nodemask specified in the policy. If there is 249 a memory pressure on all nodes in the nodemask, the allocation 265 This flag specifies that the nodemask passed by 270 change in the set of allowed nodes, the preferred nodemask (Preferred 271 Many), preferred node (Preferred) or nodemask (Bind, Interleave) is 284 3 is allowed from the user's nodemask, the "interleave" only 285 occurs over that node. If no nodes from the user's nodemask are 290 MPOL_PREFERRED policies that were created with an empty nodemask [all …]
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | printk-formats.rst | 532 位图及其衍生物,如cpumask和nodemask 540 对于打印位图(bitmap)及其派生的cpumask和nodemask,%*pb输出以字段宽度为位数的位图, 544 nodemask_pr_args()来方便打印cpumask和nodemask。
|
| /linux/tools/testing/selftests/kvm/include/ |
| H A D | numaif.h | 30 const unsigned long *, nodemask, unsigned long, maxnode,
|
| /linux/tools/workqueue/ |
| H A D | wq_dump.py | 53 from drgn.helpers.linux.nodemask import for_each_node
|
| /linux/Documentation/core-api/ |
| H A D | printk-formats.rst | 587 bitmap and its derivatives such as cpumask and nodemask 595 For printing bitmap and its derivatives such as cpumask and nodemask, 601 printing cpumask and nodemask.
|
| /linux/kernel/irq/ |
| H A D | manage.c | 623 const struct cpumask *nodemask = cpumask_of_node(node); in irq_setup_affinity() local 626 if (cpumask_intersects(&mask, nodemask)) in irq_setup_affinity() 627 cpumask_and(&mask, &mask, nodemask); in irq_setup_affinity()
|
| /linux/drivers/tty/ |
| H A D | sysrq.c | 390 .nodemask = NULL, in moom_callback()
|