| /linux/include/linux/ |
| H A D | oom.h | 12 struct zonelist; 30 struct zonelist *zonelist; member
|
| H A D | mmzone.h | 1329 struct zonelist { struct 1394 struct zonelist node_zonelists[MAX_ZONELISTS]; 1781 static inline struct zoneref *first_zones_zonelist(struct zonelist *zonelist, in first_zones_zonelist() argument 1785 return next_zones_zonelist(zonelist->_zonerefs, in first_zones_zonelist() 1828 struct zonelist *zonelist; in movable_only_nodes() local 1841 zonelist = &NODE_DATA(nid)->node_zonelists[ZONELIST_FALLBACK]; in movable_only_nodes() 1842 z = first_zones_zonelist(zonelist, ZONE_NORMAL, nodes); in movable_only_nodes()
|
| H A D | swap.h | 372 extern unsigned long try_to_free_pages(struct zonelist *zonelist, int order,
|
| H A D | gfp.h | 217 static inline struct zonelist *node_zonelist(int nid, gfp_t flags) in node_zonelist()
|
| /linux/Documentation/translations/zh_CN/mm/ |
| H A D | numa.rst | 57 中的一个或多个]构建了一个有序的“区列表”。zonelist指定了当一个选定的区/节点不能满足分配请求 63 代表了相对稀缺的资源。Linux选择了一个默认的Node ordered zonelist。这意味着在使用按NUMA距
|
| /linux/Documentation/mm/ |
| H A D | numa.rst | 74 an ordered "zonelist". A zonelist specifies the zones/nodes to visit when a 84 a default Node ordered zonelist. This means it tries to fallback to other zones 89 Linux will attempt to allocate from the first node in the appropriate zonelist 92 nodes' zones in the selected zonelist looking for the first zone in the list 120 zonelist--will not be the node itself. Rather, it will be the node that the
|
| /linux/mm/ |
| H A D | page_alloc.c | 3499 struct zonelist *zonelist = ac->zonelist; in unreserve_highatomic_pageblock() local 3507 for_each_zone_zonelist_nodemask(zone, z, zonelist, ac->highest_zoneidx, in unreserve_highatomic_pageblock() 4074 .zonelist = ac->zonelist, in __alloc_pages_may_oom() 4316 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_compact_retry() 4424 progress = try_to_free_pages(ac->zonelist, order, gfp_mask, in __perform_reclaim() 4484 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, highest_zoneidx, in wake_all_kswapds() 4628 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_reclaim_retry() 4778 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 4789 struct zoneref *z = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 4821 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() [all …]
|
| H A D | vmscan.c | 6221 static void shrink_zones(struct zonelist *zonelist, struct scan_control *sc) in shrink_zones() argument 6242 for_each_zone_zonelist_nodemask(zone, z, zonelist, in shrink_zones() 6344 static unsigned long do_try_to_free_pages(struct zonelist *zonelist, in do_try_to_free_pages() argument 6362 shrink_zones(zonelist, sc); in do_try_to_free_pages() 6372 for_each_zone_zonelist_nodemask(zone, z, zonelist, sc->reclaim_idx, in do_try_to_free_pages() 6486 static bool throttle_direct_reclaim(gfp_t gfp_mask, struct zonelist *zonelist, in throttle_direct_reclaim() argument 6524 for_each_zone_zonelist_nodemask(zone, z, zonelist, in throttle_direct_reclaim() 6566 unsigned long try_to_free_pages(struct zonelist *zonelist, int order, in try_to_free_pages() argument 6595 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages() 6601 nr_reclaimed = do_try_to_free_pages(zonelist, &sc); in try_to_free_pages() [all …]
|
| H A D | oom_kill.c | 271 if (!oc->zonelist) in constrained_alloc() 295 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc()
|
| H A D | mm_init.c | 71 struct zonelist *zonelist; in mminit_verify_zonelist() local 79 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist() 84 /* Print information about the zonelist */ in mminit_verify_zonelist() 85 printk(KERN_DEBUG "mminit::zonelist %s %d:%s = ", in mminit_verify_zonelist() 89 /* Iterate the zonelist */ in mminit_verify_zonelist() 90 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
|
| H A D | internal.h | 658 struct zonelist *zonelist; member
|
| H A D | mempolicy.c | 2168 struct zonelist *zonelist; in mempolicy_slab_node() local 2170 zonelist = &NODE_DATA(node)->node_zonelists[ZONELIST_FALLBACK]; in mempolicy_slab_node() 2171 z = first_zones_zonelist(zonelist, highest_zoneidx, in mempolicy_slab_node()
|
| H A D | slub.c | 3822 struct zonelist *zonelist; in get_from_any_partial() local 3861 zonelist = node_zonelist(mempolicy_slab_node(), pc->flags); in get_from_any_partial() 3862 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in get_from_any_partial() 7055 struct zonelist *zonelist; in __refill_objects_any() local 7069 zonelist = node_zonelist(mempolicy_slab_node(), gfp); in __refill_objects_any() 7070 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in __refill_objects_any()
|
| H A D | compaction.c | 2442 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in compaction_zonelist_suitable() 2828 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in try_to_compact_pages()
|
| H A D | hugetlb.c | 1332 struct zonelist *zonelist; in dequeue_hugetlb_folio_nodemask() local 1341 zonelist = node_zonelist(nid, gfp_mask); in dequeue_hugetlb_folio_nodemask() 1345 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_hugetlb_folio_nodemask()
|
| H A D | memcontrol.c | 1677 .zonelist = NULL, in mem_cgroup_out_of_memory()
|
| /linux/drivers/tty/ |
| H A D | sysrq.c | 389 .zonelist = node_zonelist(first_memory_node, gfp_mask), in moom_callback()
|
| /linux/Documentation/admin-guide/sysctl/ |
| H A D | vm.rst | 744 In non-NUMA case, a zonelist for GFP_KERNEL is ordered as following. 750 Assume 2 node NUMA and below is zonelist of Node(0)'s GFP_KERNEL::
|
| /linux/Documentation/admin-guide/mm/ |
| H A D | numa_memory_policy.rst | 234 node zonelist.
|
| /linux/Documentation/admin-guide/ |
| H A D | kernel-parameters.txt | 4785 numa_zonelist_order= [KNL, BOOT] Select zonelist order for NUMA.
|