| /linux/include/linux/ |
| H A D | oom.h | 12 struct zonelist; 30 struct zonelist *zonelist; member
|
| H A D | mmzone.h | 1329 struct zonelist { struct 1394 struct zonelist node_zonelists[MAX_ZONELISTS]; 1767 static inline struct zoneref *first_zones_zonelist(struct zonelist *zonelist, in first_zones_zonelist() argument 1771 return next_zones_zonelist(zonelist->_zonerefs, in first_zones_zonelist() 1814 struct zonelist *zonelist; in movable_only_nodes() local 1827 zonelist = &NODE_DATA(nid)->node_zonelists[ZONELIST_FALLBACK]; in movable_only_nodes() 1828 z = first_zones_zonelist(zonelist, ZONE_NORMAL, nodes); in movable_only_nodes()
|
| H A D | swap.h | 374 extern unsigned long try_to_free_pages(struct zonelist *zonelist, int order,
|
| H A D | gfp.h | 213 static inline struct zonelist *node_zonelist(int nid, gfp_t flags) in node_zonelist()
|
| /linux/Documentation/translations/zh_CN/mm/ |
| H A D | numa.rst | 57 中的一个或多个]构建了一个有序的“区列表”。zonelist指定了当一个选定的区/节点不能满足分配请求 63 代表了相对稀缺的资源。Linux选择了一个默认的Node ordered zonelist。这意味着在使用按NUMA距
|
| /linux/Documentation/mm/ |
| H A D | numa.rst | 74 an ordered "zonelist". A zonelist specifies the zones/nodes to visit when a 84 a default Node ordered zonelist. This means it tries to fallback to other zones 89 Linux will attempt to allocate from the first node in the appropriate zonelist 92 nodes' zones in the selected zonelist looking for the first zone in the list 120 zonelist--will not be the node itself. Rather, it will be the node that the
|
| /linux/mm/ |
| H A D | page_alloc.c | 3452 struct zonelist *zonelist = ac->zonelist; in unreserve_highatomic_pageblock() local 3460 for_each_zone_zonelist_nodemask(zone, z, zonelist, ac->highest_zoneidx, in unreserve_highatomic_pageblock() 4027 .zonelist = ac->zonelist, in __alloc_pages_may_oom() 4269 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_compact_retry() 4377 progress = try_to_free_pages(ac->zonelist, order, gfp_mask, in __perform_reclaim() 4437 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, highest_zoneidx, in wake_all_kswapds() 4581 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_reclaim_retry() 4716 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 4727 struct zoneref *z = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 4825 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() [all …]
|
| H A D | vmscan.c | 6221 static void shrink_zones(struct zonelist *zonelist, struct scan_control *sc) in shrink_zones() argument 6242 for_each_zone_zonelist_nodemask(zone, z, zonelist, in shrink_zones() 6344 static unsigned long do_try_to_free_pages(struct zonelist *zonelist, in do_try_to_free_pages() argument 6362 shrink_zones(zonelist, sc); in do_try_to_free_pages() 6379 for_each_zone_zonelist_nodemask(zone, z, zonelist, sc->reclaim_idx, in do_try_to_free_pages() 6493 static bool throttle_direct_reclaim(gfp_t gfp_mask, struct zonelist *zonelist, in throttle_direct_reclaim() argument 6531 for_each_zone_zonelist_nodemask(zone, z, zonelist, in throttle_direct_reclaim() 6573 unsigned long try_to_free_pages(struct zonelist *zonelist, int order, in try_to_free_pages() argument 6602 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages() 6608 nr_reclaimed = do_try_to_free_pages(zonelist, &sc); in try_to_free_pages() [all …]
|
| H A D | oom_kill.c | 271 if (!oc->zonelist) in constrained_alloc() 295 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc()
|
| H A D | mm_init.c | 71 struct zonelist *zonelist; in mminit_verify_zonelist() local 79 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist() 90 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
|
| H A D | internal.h | 589 struct zonelist *zonelist; member
|
| H A D | mempolicy.c | 2170 struct zonelist *zonelist; in mempolicy_slab_node() local 2172 zonelist = &NODE_DATA(node)->node_zonelists[ZONELIST_FALLBACK]; in mempolicy_slab_node() 2173 z = first_zones_zonelist(zonelist, highest_zoneidx, in mempolicy_slab_node()
|
| H A D | compaction.c | 2442 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in compaction_zonelist_suitable() 2828 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in try_to_compact_pages()
|
| H A D | hugetlb.c | 1340 struct zonelist *zonelist; in dequeue_hugetlb_folio_nodemask() local 1349 zonelist = node_zonelist(nid, gfp_mask); in dequeue_hugetlb_folio_nodemask() 1353 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_hugetlb_folio_nodemask()
|
| H A D | slub.c | 3561 struct zonelist *zonelist; in get_any_partial() local 3592 zonelist = node_zonelist(mempolicy_slab_node(), pc->flags); in get_any_partial() 3593 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in get_any_partial()
|
| H A D | memcontrol.c | 1664 .zonelist = NULL, in mem_cgroup_out_of_memory()
|
| /linux/Documentation/admin-guide/sysctl/ |
| H A D | vm.rst | 718 In non-NUMA case, a zonelist for GFP_KERNEL is ordered as following. 724 Assume 2 node NUMA and below is zonelist of Node(0)'s GFP_KERNEL::
|
| /linux/Documentation/admin-guide/mm/ |
| H A D | numa_memory_policy.rst | 234 node zonelist.
|
| /linux/Documentation/admin-guide/ |
| H A D | kernel-parameters.txt | 4669 numa_zonelist_order= [KNL, BOOT] Select zonelist order for NUMA.
|