Home
last modified time | relevance | path

Searched refs:nodemask (Results 1 – 24 of 24) sorted by relevance

/linux/mm/
H A Dshow_mem.c125 static bool show_mem_node_skip(unsigned int flags, int nid, nodemask_t *nodemask) in show_mem_node_skip() argument
135 if (!nodemask) in show_mem_node_skip()
136 nodemask = &cpuset_current_mems_allowed; in show_mem_node_skip()
138 return !node_isset(nid, *nodemask); in show_mem_node_skip()
186 static void show_free_areas(unsigned int filter, nodemask_t *nodemask, int max_zone_idx) in show_free_areas() argument
196 if (show_mem_node_skip(filter, zone_to_nid(zone), nodemask)) in show_free_areas()
233 if (show_mem_node_skip(filter, pgdat->node_id, nodemask)) in show_free_areas()
297 if (show_mem_node_skip(filter, zone_to_nid(zone), nodemask)) in show_free_areas()
360 if (show_mem_node_skip(filter, zone_to_nid(zone), nodemask)) in show_free_areas()
390 if (show_mem_node_skip(filter, nid, nodemask)) in show_free_areas()
400 __show_mem(unsigned int filter,nodemask_t * nodemask,int max_zone_idx) __show_mem() argument
[all...]
H A Dmempolicy.c89 #include <linux/nodemask.h>
121 #define MPOL_MF_INVERT (MPOL_MF_INTERNAL << 1) /* Invert check for nodemask */
255 * mpol_set_nodemask is called after mpol_new() to set up the nodemask, if
313 * MPOL_F_RELATIVE_NODES if the nodemask is empty (local allocation). in mpol_new()
314 * All other modes require a valid pointer to a non-empty nodemask. in mpol_new()
386 * OOM/allocation failure due to parallel nodemask modification.
411 * Rebind each vma in mm to new nodemask.
887 * Return nodemask for policy for get_mempolicy() query
1224 nodemask_t *nodemask; in alloc_migration_target_by_mpol()
1229 nodemask in alloc_migration_target_by_mpol()
1222 nodemask_t *nodemask; alloc_migration_target_by_mpol() local
1980 nodemask_t nodemask; weighted_interleave_nid() local
2024 nodemask_t nodemask; interleave_nid() local
2046 nodemask_t *nodemask = NULL; policy_nodemask() local
2102 huge_node(struct vm_area_struct * vma,unsigned long addr,gfp_t gfp_flags,struct mempolicy ** mpol,nodemask_t ** nodemask) huge_node() argument
2189 alloc_pages_preferred_many(gfp_t gfp,unsigned int order,int nid,nodemask_t * nodemask) alloc_pages_preferred_many() argument
2222 nodemask_t *nodemask; alloc_pages_mpol_noprof() local
2534 nodemask_t *nodemask; alloc_pages_bulk_array_mempolicy_noprof() local
[all...]
H A Doom_kill.c95 const nodemask_t *mask = oc->nodemask; in oom_cpuset_eligible()
282 * This is not a __GFP_THISNODE allocation, so a truncated nodemask in in constrained_alloc()
286 if (oc->nodemask && in constrained_alloc()
287 !nodes_subset(node_states[N_MEMORY], *oc->nodemask)) { in constrained_alloc()
289 for_each_node_mask(nid, *oc->nodemask) in constrained_alloc()
296 highest_zoneidx, oc->nodemask) in constrained_alloc()
317 /* p may not have freeable memory in nodemask */ in oom_evaluate_task()
390 /* p may not have freeable memory in nodemask */ in dump_task()
445 pr_info("oom-kill:constraint=%s,nodemask=%*pbl", in dump_oom_victim()
447 nodemask_pr_args(oc->nodemask)); in dump_oom_victim()
[all...]
H A Dpage_alloc.c2077 ac->nodemask) { in unreserve_highatomic_pageblock()
3335 ac->nodemask) { in get_page_from_freelist()
3491 static void warn_alloc_show_mem(gfp_t gfp_mask, nodemask_t *nodemask) in warn_alloc_show_mem() argument
3507 __show_mem(filter, nodemask, gfp_zone(gfp_mask)); in warn_alloc_show_mem()
3510 void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...) in warn_alloc() argument
3526 nodemask_pr_args(nodemask)); in warn_alloc()
3532 warn_alloc_show_mem(gfp_mask, nodemask); in warn_alloc()
3561 .nodemask = ac->nodemask, in __alloc_pages_may_oom()
3803 ac->highest_zoneidx, ac->nodemask) { in should_compact_retry()
3911 ac->nodemask); in __perform_reclaim()
[all …]
H A Dhugetlb.c13 #include <linux/nodemask.h>
1403 nodemask_t *nodemask; in dequeue_hugetlb_folio_vma()
1419 nid = huge_node(vma, address, gfp_mask, &mpol, &nodemask); in dequeue_hugetlb_folio_vma()
1423 nid, nodemask); in dequeue_hugetlb_folio_vma()
1426 nodemask = NULL; in dequeue_hugetlb_folio_vma()
1431 nid, nodemask); in dequeue_hugetlb_folio_vma()
1519 int nid, nodemask_t *nodemask) in __destroy_compound_gigantic_folio()
1537 for_each_node_mask(node, *nodemask) { in __destroy_compound_gigantic_folio()
1549 folio = folio_alloc_gigantic(order, gfp_mask, nid, nodemask); in destroy_compound_gigantic_folio()
1568 int nid, nodemask_t *nodemask)
1402 nodemask_t *nodemask; dequeue_hugetlb_folio_vma() local
1570 alloc_gigantic_folio(struct hstate * h,gfp_t gfp_mask,int nid,nodemask_t * nodemask) alloc_gigantic_folio() argument
1608 alloc_gigantic_folio(struct hstate * h,gfp_t gfp_mask,int nid,nodemask_t * nodemask) alloc_gigantic_folio() argument
1616 alloc_gigantic_folio(struct hstate * h,gfp_t gfp_mask,int nid,nodemask_t * nodemask) alloc_gigantic_folio() argument
2548 nodemask_t *nodemask; alloc_buddy_hugetlb_folio_with_mpol() local
6623 nodemask_t *nodemask; alloc_hugetlb_folio_vma() local
[all...]
H A Dvmscan.c81 nodemask_t *nodemask; member
6138 sc->reclaim_idx, sc->nodemask) { in shrink_zones()
6275 sc->nodemask) { in do_try_to_free_pages()
6393 nodemask_t *nodemask) in throttle_direct_reclaim() argument
6431 gfp_zone(gfp_mask), nodemask) { in throttle_direct_reclaim()
6473 gfp_t gfp_mask, nodemask_t *nodemask) in try_to_free_pages() argument
6481 .nodemask = nodemask, in try_to_free_pages()
6501 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages()
H A Dcompaction.c2463 ac->highest_zoneidx, ac->nodemask) { in compaction_suitable()
2823 ac->highest_zoneidx, ac->nodemask) { in compact_zone_order()
H A Dinternal.h459 * nodemask, migratetype and highest_zoneidx are initialized only once in
469 nodemask_t *nodemask;
452 nodemask_t *nodemask; global() member
H A Dmemcontrol.c1582 .nodemask = NULL, in mem_cgroup_out_of_memory()
/linux/Documentation/translations/zh_CN/core-api/
H A Dmemory-hotplug.rst73 - status_change_nid_normal是当nodemask的N_NORMAL_MEMORY被设置/清除时设置节
74 点id,如果是-1,则nodemask状态不改变。
76 - status_change_nid是当nodemask的N_MEMORY被(将)设置/清除时设置的节点id。这
78 存。如果这个值为-1,那么nodemask的状态就不会改变。
H A Dprintk-formats.rst533 位图及其衍生物,如cpumask和nodemask
541 对于打印位图(bitmap)及其派生的cpumask和nodemask,%*pb输出以字段宽度为位数的位图,
545 nodemask_pr_args()来方便打印cpumask和nodemask
/linux/tools/testing/selftests/kvm/x86_64/
H A Dxapic_ipi_test.c253 unsigned long nodemask = 0; in do_migrations() local
254 unsigned long nodemasks[sizeof(nodemask) * 8]; in do_migrations()
269 r = get_mempolicy(NULL, &nodemask, sizeof(nodemask) * 8, in do_migrations()
275 sizeof(nodemask) * 8, nodemask); in do_migrations()
281 for (i = 0, bit = 1; i < sizeof(nodemask) * 8; i++, bit <<= 1) { in do_migrations()
282 if (nodemask & bit) { in do_migrations()
283 nodemasks[nodes] = nodemask & bit; in do_migrations()
/linux/Documentation/admin-guide/mm/
H A Dnuma_memory_policy.rst215 is always preferred by passing an empty nodemask with this
216 mode. If an empty nodemask is passed, the policy cannot use
248 satisfied from the nodemask specified in the policy. If there is
249 a memory pressure on all nodes in the nodemask, the allocation
265 This flag specifies that the nodemask passed by
270 change in the set of allowed nodes, the preferred nodemask (Preferred
271 Many), preferred node (Preferred) or nodemask (Bind, Interleave) is
284 3 is allowed from the user's nodemask, the "interleave" only
285 occurs over that node. If no nodes from the user's nodemask are
290 MPOL_PREFERRED policies that were created with an empty nodemask
[all …]
/linux/Documentation/core-api/
H A Dmemory-hotplug.rst65 - status_change_nid_normal is set node id when N_NORMAL_MEMORY of nodemask
66 is (will be) set/clear, if this is -1, then nodemask status is not changed.
67 - status_change_nid is set node id when N_MEMORY of nodemask is (will be)
69 node loses all memory. If this is -1, then nodemask status is not changed.
/linux/include/linux/
H A Doom.h34 nodemask_t *nodemask; member
H A Dmmzone.h1705 #define for_each_zone_zonelist_nodemask(zone, z, zlist, highidx, nodemask) \ argument
1706 for (z = first_zones_zonelist(zlist, highidx, nodemask), zone = zonelist_zone(z); \
1708 z = next_zones_zonelist(++z, highidx, nodemask), \
1711 #define for_next_zone_zonelist_nodemask(zone, z, highidx, nodemask) \ argument
1714 z = next_zones_zonelist(++z, highidx, nodemask), \
H A Dmm.h3244 extern void __show_mem(unsigned int flags, nodemask_t *nodemask, int max_zone_idx);
3254 void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...);
/linux/kernel/
H A Dsmp.c751 const struct cpumask *nodemask;
760 nodemask = cpumask_of_node(cpu_to_node(cpu)); in smp_call_function_many_cond()
761 for (cpu = cpumask_first_and(nodemask, mask); cpu < nr_cpu_ids; in smp_call_function_many_cond()
762 cpu = cpumask_next_and(cpu, nodemask, mask)) { in smp_call_function_many_cond()
723 const struct cpumask *nodemask; smp_call_function_any() local
/linux/tools/workqueue/
H A Dwq_dump.py53 from drgn.helpers.linux.nodemask import for_each_node
/linux/kernel/irq/
H A Dmanage.c625 const struct cpumask *nodemask = cpumask_of_node(node); in irq_setup_affinity()
627 /* make sure at least one of the cpus in nodemask is online */ in irq_setup_affinity()
628 if (cpumask_intersects(&mask, nodemask)) in irq_setup_affinity() local
629 cpumask_and(&mask, &mask, nodemask); in irq_setup_affinity()
/linux/Documentation/mm/
H A Dphysical_memory.rst188 ``include/linux/nodemask.h``.
/linux/kernel/cgroup/
H A Dcpuset.c4079 int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask) in cpuset_nodemask_valid_mems_allowed() argument
4081 return nodes_intersects(*nodemask, current->mems_allowed); in cpuset_nodemask_valid_mems_allowed()
/linux/kernel/sched/
H A Dcore.c3449 const struct cpumask *nodemask = NULL; in select_fallback_rq() local
3459 nodemask = cpumask_of_node(nid); in select_fallback_rq()
3462 for_each_cpu(dest_cpu, nodemask) { in select_fallback_rq()
/linux/
H A DMAINTAINERS3926 F: include/linux/nodemask.h