| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | xapic_ipi_test.c | 254 unsigned long nodemask = 0; in do_migrations() local 255 unsigned long nodemasks[sizeof(nodemask) * 8]; in do_migrations() 270 kvm_get_mempolicy(NULL, &nodemask, sizeof(nodemask) * 8, in do_migrations() 275 sizeof(nodemask) * 8, nodemask); in do_migrations() 281 for (i = 0, bit = 1; i < sizeof(nodemask) * 8; i++, bit <<= 1) { in do_migrations() 282 if (nodemask & bit) { in do_migrations() 283 nodemasks[nodes] = nodemask & bit; in do_migrations()
|
| /linux/include/linux/ |
| H A D | cpuset.h | 86 int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask); 166 static inline void set_mems_allowed(nodemask_t nodemask) in set_mems_allowed() argument 173 current->mems_allowed = nodemask; in set_mems_allowed() 227 static inline int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask) in cpuset_nodemask_valid_mems_allowed() argument 288 static inline void set_mems_allowed(nodemask_t nodemask) in set_mems_allowed() argument
|
| H A D | gfp.h | 230 nodemask_t *nodemask); 234 nodemask_t *nodemask); 238 nodemask_t *nodemask, int nr_pages, 457 gfp_t gfp_mask, int nid, nodemask_t *nodemask); 462 int nid, nodemask_t *nodemask);
|
| H A D | mempolicy.h | 145 struct mempolicy **mpol, nodemask_t **nodemask); 261 struct mempolicy **mpol, nodemask_t **nodemask) in huge_node() argument 264 *nodemask = NULL; in huge_node()
|
| H A D | oom.h | 33 nodemask_t *nodemask; member
|
| H A D | mm.h | 4011 extern void __show_mem(unsigned int flags, nodemask_t *nodemask, int max_zone_idx); 4021 void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...);
|
| /linux/tools/testing/selftests/kvm/ |
| H A D | guest_memfd_test.c | 82 unsigned long nodemask = 0; in test_mbind() local 83 unsigned long maxnode = BITS_PER_TYPE(nodemask); in test_mbind() 95 kvm_get_mempolicy(&policy, &nodemask, maxnode, mem, MPOL_F_ADDR); in test_mbind() 96 TEST_ASSERT(policy == MPOL_INTERLEAVE && nodemask == nodemask_0, in test_mbind() 98 MPOL_INTERLEAVE, nodemask_0, policy, nodemask); in test_mbind() 102 kvm_get_mempolicy(&policy, &nodemask, maxnode, mem + page_size * 2, MPOL_F_ADDR); in test_mbind() 103 TEST_ASSERT(policy == MPOL_BIND && nodemask == nodemask_0, in test_mbind() 105 MPOL_BIND, nodemask_0, policy, nodemask); in test_mbind() 109 kvm_get_mempolicy(&policy, &nodemask, maxnode, mem, MPOL_F_ADDR); in test_mbind() 110 TEST_ASSERT(policy == MPOL_DEFAULT && !nodemask, in test_mbind() [all …]
|
| /linux/tools/testing/selftests/futex/functional/ |
| H A D | futex_numa_mpol.c | 185 unsigned long nodemask; in TEST() 188 nodemask = 1 << i; in TEST() 189 ret = mbind(futex_ptr, mem_size, MPOL_BIND, &nodemask, in TEST() 190 sizeof(nodemask) * 8, 0); in TEST() 183 unsigned long nodemask; TEST() local
|
| /linux/mm/ |
| H A D | mempolicy.c | 1448 nodemask_t *nodemask; in alloc_migration_target_by_mpol() local 1453 nodemask = policy_nodemask(gfp, pol, ilx, &nid); in alloc_migration_target_by_mpol() 1454 return alloc_hugetlb_folio_nodemask(h, nid, nodemask, gfp, in alloc_migration_target_by_mpol() 2206 nodemask_t nodemask; in weighted_interleave_nid() local 2213 nr_nodes = read_once_policy_nodemask(pol, &nodemask); in weighted_interleave_nid() 2225 for_each_node_mask(nid, nodemask) in weighted_interleave_nid() 2230 nid = first_node(nodemask); in weighted_interleave_nid() 2237 nid = next_node_in(nid, nodemask); in weighted_interleave_nid() 2250 nodemask_t nodemask; in interleave_nid() local 2255 nnodes = read_once_policy_nodemask(pol, &nodemask); in interleave_nid() [all …]
|
| H A D | page_alloc.c | 3492 ac->nodemask) { in unreserve_highatomic_pageblock() 3811 ac->nodemask) { in get_page_from_freelist() 3982 static void warn_alloc_show_mem(gfp_t gfp_mask, nodemask_t *nodemask) in warn_alloc_show_mem() argument 3998 __show_mem(filter, nodemask, gfp_zone(gfp_mask)); in warn_alloc_show_mem() 4002 void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...) in warn_alloc() argument 4018 nodemask_pr_args(nodemask)); in warn_alloc() 4024 warn_alloc_show_mem(gfp_mask, nodemask); in warn_alloc() 4052 .nodemask = ac->nodemask, in __alloc_pages_may_oom() 4294 ac->highest_zoneidx, ac->nodemask) { in should_compact_retry() 4402 ac->nodemask); in __perform_reclaim() [all …]
|
| H A D | hugetlb_cma.c | 28 int nid, nodemask_t *nodemask) in hugetlb_cma_alloc_frozen_folio() argument 41 for_each_node_mask(node, *nodemask) { in hugetlb_cma_alloc_frozen_folio()
|
| H A D | hugetlb.c | 1344 nodemask_t *nodemask; in dequeue_hugetlb_folio_vma() local 1355 nid = huge_node(vma, address, gfp_mask, &mpol, &nodemask); in dequeue_hugetlb_folio_vma() 1359 nid, nodemask); in dequeue_hugetlb_folio_vma() 1362 nodemask = NULL; in dequeue_hugetlb_folio_vma() 1367 nid, nodemask); in dequeue_hugetlb_folio_vma() 1378 int nid, nodemask_t *nodemask) in alloc_gigantic_frozen_folio() argument 1382 folio = hugetlb_cma_alloc_frozen_folio(order, gfp_mask, nid, nodemask); in alloc_gigantic_frozen_folio() 1390 nid, nodemask); in alloc_gigantic_frozen_folio() 1395 nodemask_t *nodemask) in alloc_gigantic_frozen_folio() argument 2179 nodemask_t *nodemask; in alloc_buddy_hugetlb_folio_with_mpol() local [all …]
|
| H A D | numa_memblks.c | 23 static void __init numa_nodemask_from_meminfo(nodemask_t *nodemask, in numa_nodemask_from_meminfo() argument 31 node_set(mi->blk[i].nid, *nodemask); in numa_nodemask_from_meminfo()
|
| H A D | memcontrol.c | 1890 .nodemask = NULL, in mem_cgroup_out_of_memory()
|
| /linux/tools/testing/selftests/kvm/include/ |
| H A D | numaif.h | 30 const unsigned long *, nodemask, unsigned long, maxnode,
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | printk-formats.rst | 532 位图及其衍生物,如cpumask和nodemask 540 对于打印位图(bitmap)及其派生的cpumask和nodemask,%*pb输出以字段宽度为位数的位图, 544 nodemask_pr_args()来方便打印cpumask和nodemask。
|
| /linux/tools/workqueue/ |
| H A D | wq_dump.py | 53 from drgn.helpers.linux.nodemask import for_each_node
|
| /linux/Documentation/core-api/ |
| H A D | printk-formats.rst | 587 bitmap and its derivatives such as cpumask and nodemask 595 For printing bitmap and its derivatives such as cpumask and nodemask, 601 printing cpumask and nodemask.
|
| /linux/kernel/irq/ |
| H A D | manage.c | 623 const struct cpumask *nodemask = cpumask_of_node(node); in irq_setup_affinity() local 626 if (cpumask_intersects(&mask, nodemask)) in irq_setup_affinity() 627 cpumask_and(&mask, &mask, nodemask); in irq_setup_affinity()
|
| /linux/drivers/tty/ |
| H A D | sysrq.c | 390 .nodemask = NULL, in moom_callback()
|
| /linux/kernel/cgroup/ |
| H A D | cpuset.c | 4152 int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask) in cpuset_nodemask_valid_mems_allowed() argument 4154 return nodes_intersects(*nodemask, current->mems_allowed); in cpuset_nodemask_valid_mems_allowed()
|
| /linux/kernel/sched/ |
| H A D | core.c | 3508 const struct cpumask *nodemask = NULL; in select_fallback_rq() local 3518 nodemask = cpumask_of_node(nid); in select_fallback_rq() 3521 for_each_cpu(dest_cpu, nodemask) { in select_fallback_rq()
|
| /linux/ |
| H A D | MAINTAINERS | 4504 F: include/linux/nodemask.h
|