| /linux/mm/ |
| H A D | mmzone.c | 34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone() 39 zone = pgdat->node_zones; in next_zone()
|
| H A D | shuffle.c | 157 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in __shuffle_free_memory()
|
| H A D | mm_init.c | 80 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist() 677 struct zone *zone = &pgdat->node_zones[zid]; in __init_page_from_nid() 976 struct zone *zone = node->node_zones + j; in memmap_init() 1298 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) { in reset_memoryless_node_totalpages() 1344 struct zone *zone = pgdat->node_zones + i; in calculate_node_totalpages() 1585 struct zone *zone = pgdat->node_zones + z; in free_area_init_core_hotplug() 1602 struct zone *zone = pgdat->node_zones + j; in free_area_init_core() 1754 struct zone *zone = &pgdat->node_zones[zone_type]; in check_for_memory() 2146 zone = pgdat->node_zones + pgdat->nr_zones - 1; in deferred_init_memmap()
|
| H A D | memory_hotplug.c | 524 for (zone = pgdat->node_zones; in update_pgdat_span() 525 zone < pgdat->node_zones + MAX_NR_ZONES; zone++) { in update_pgdat_span() 861 zone = pgdat->node_zones + i; in auto_movable_can_online_movable() 905 struct zone *zone = &pgdat->node_zones[zid]; in default_kernel_zone_for_pfn() 911 return &pgdat->node_zones[ZONE_NORMAL]; in default_kernel_zone_for_pfn() 1020 return &NODE_DATA(nid)->node_zones[ZONE_MOVABLE]; in auto_movable_zone_for_pfn() 1030 struct zone *movable_zone = &NODE_DATA(nid)->node_zones[ZONE_MOVABLE]; in default_zone_for_pfn() 1057 return &NODE_DATA(nid)->node_zones[ZONE_MOVABLE]; in zone_for_pfn_range()
|
| H A D | vmstat.c | 329 zone = &pgdat->node_zones[i]; in set_pgdat_percpu_threshold() 982 struct zone *zones = NODE_DATA(node)->node_zones; in drain_zonestat() 996 struct zone *zones = NODE_DATA(node)->node_zones; in sum_zone_node_page_state() 1530 struct zone *node_zones = pgdat->node_zones; in frag_next() 1533 for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { in frag_stop() 1742 struct zone *compare = &pgdat->node_zones[zid]; 1545 struct zone *node_zones = pgdat->node_zones; walk_zones_in_node() local
|
| H A D | show_mem.c | 98 struct zone *zone = &pgdat->node_zones[zone_type]; in si_meminfo_node() 166 if (zone_managed_pages(pgdat->node_zones + zone_idx)) in node_has_managed_zones()
|
| H A D | memremap.c | 228 zone = &NODE_DATA(nid)->node_zones[ZONE_DEVICE]; in pagemap_range() 242 memmap_init_zone_device(&NODE_DATA(nid)->node_zones[ZONE_DEVICE], in pagemap_range()
|
| H A D | compaction.c | 407 struct zone *zone = &pgdat->node_zones[zoneid]; in reset_isolation_suitable() 2205 zone = &pgdat->node_zones[zoneid]; in fragmentation_score_node() 2907 zone = &pgdat->node_zones[zoneid]; in compact_node() 3039 zone = &pgdat->node_zones[zoneid]; in kcompactd_node_suitable() 3081 zone = &pgdat->node_zones[zoneid]; in kcompactd_do_work()
|
| H A D | migrate.c | 2620 struct zone *zone = pgdat->node_zones + z; in migrate_balanced_pgdat() 2691 if (managed_zone(pgdat->node_zones + z)) in migrate_misplaced_folio_prepare() 2702 wakeup_kswapd(pgdat->node_zones + z, 0, in migrate_misplaced_folio_prepare()
|
| H A D | page-writeback.c | 250 struct zone *zone = pgdat->node_zones + z; 286 z = &NODE_DATA(node)->node_zones[i]; in highmem_dirtyable_memory()
|
| H A D | memblock.c | 2320 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_managed_pages()
|
| H A D | page_alloc.c | 5559 zone = pgdat->node_zones + zone_type; in build_zonerefs_node() 6349 struct zone *zone = pgdat->node_zones + i; in calculate_totalreserve_pages() 6409 struct zone *zone = &pgdat->node_zones[i]; in setup_per_zone_lowmem_reserve() 6415 struct zone *upper_zone = &pgdat->node_zones[j]; in setup_per_zone_lowmem_reserve() 7602 if (managed_zone(&pgdat->node_zones[zone])) in has_managed_zone()
|
| H A D | vmscan.c | 284 for ((idx) = 0, (zone) = (pgdat)->node_zones; \ 4857 struct zone *zone = lruvec_pgdat(lruvec)->node_zones + i; in should_abort_scan() 6741 zone = pgdat->node_zones + i; in pgdat_watermark_boosted() 7008 zone = pgdat->node_zones + i; in balance_pgdat() 7141 zone = pgdat->node_zones + i; in balance_pgdat()
|
| /linux/tools/testing/vma/linux/ |
| H A D | mmzone.h | 34 struct zone node_zones[MAX_NR_ZONES]; member
|
| /linux/tools/testing/memblock/linux/ |
| H A D | mmzone.h | 34 struct zone node_zones[MAX_NR_ZONES]; member
|
| /linux/tools/perf/util/bpf_skel/vmlinux/ |
| H A D | vmlinux.h | 211 struct zone node_zones[6]; /* value for all possible config */ member
|
| /linux/include/linux/ |
| H A D | mmzone.h | 1387 struct zone node_zones[MAX_NR_ZONES]; member 1590 #define zone_idx(zone) ((zone) - (zone)->zone_pgdat->node_zones) 1709 for (zone = (first_online_pgdat())->node_zones; \ 1714 for (zone = (first_online_pgdat())->node_zones; \
|
| H A D | mm_inline.h | 48 __mod_zone_page_state(&pgdat->node_zones[zid], in __update_lru_size()
|
| H A D | mm.h | 2212 return &NODE_DATA(page_to_nid(page))->node_zones[page_zonenum(page)]; in page_zone() 2227 return &folio_pgdat(folio)->node_zones[folio_zonenum(folio)]; in folio_zone()
|
| /linux/drivers/acpi/ |
| H A D | acpi_mrrm.c | 43 struct zone *zone = NODE_DATA(nid)->node_zones + z; in get_node_num()
|
| /linux/kernel/ |
| H A D | vmcore_info.c | 202 VMCOREINFO_OFFSET(pglist_data, node_zones); in crash_save_vmcoreinfo_init()
|
| /linux/tools/perf/util/bpf_skel/ |
| H A D | lock_contention.bpf.c | 869 zone_off = offsetof(struct pglist_data, node_zones); in collect_zone_lock()
|
| /linux/drivers/base/ |
| H A D | memory.c | 733 zone = pgdat->node_zones + i; in early_node_zone_for_memory_block()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_acpi.c | 886 zone_managed_pages(&pgdat->node_zones[zone_type]); in amdgpu_acpi_get_numa_size()
|
| /linux/Documentation/admin-guide/kdump/ |
| H A D | vmcoreinfo.rst | 150 (pglist_data, node_zones|nr_zones|node_mem_map|node_start_pfn|node_spanned_pages|node_id)
|