Home
last modified time | relevance | path

Searched refs:pgdat (Results 1 – 25 of 39) sorted by relevance

12

/linux/mm/
H A Dmm_init.c68 pg_data_t *pgdat = NODE_DATA(nid); in mminit_verify_zonelist() local
79 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist()
80 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist()
671 pg_data_t *pgdat; in __init_page_from_nid() local
674 pgdat = NODE_DATA(nid); in __init_page_from_nid()
677 struct zone *zone = &pgdat->node_zones[zid]; in __init_page_from_nid()
690 static inline void pgdat_set_deferred_range(pg_data_t *pgdat) in pgdat_set_deferred_range() argument
692 pgdat->first_deferred_pfn = ULONG_MAX; in pgdat_set_deferred_range()
753 static inline void pgdat_set_deferred_range(pg_data_t *pgdat) {} in pgdat_set_deferred_range() argument
1114 struct pglist_data *pgdat = zone->zone_pgdat; in memmap_init_zone_device() local
[all …]
H A Dshow_mem.c95 pg_data_t *pgdat = NODE_DATA(nid); in si_meminfo_node() local
98 struct zone *zone = &pgdat->node_zones[zone_type]; in si_meminfo_node()
107 val->sharedram = node_page_state(pgdat, NR_SHMEM); in si_meminfo_node()
162 static bool node_has_managed_zones(pg_data_t *pgdat, int max_zone_idx) in node_has_managed_zones() argument
166 if (zone_managed_pages(pgdat->node_zones + zone_idx)) in node_has_managed_zones()
185 pg_data_t *pgdat; in show_free_areas() local
226 for_each_online_pgdat(pgdat) { in show_free_areas()
227 if (show_mem_node_skip(filter, pgdat->node_id, nodemask)) in show_free_areas()
229 if (!node_has_managed_zones(pgdat, max_zone_idx)) in show_free_areas()
258 pgdat in show_free_areas()
[all...]
H A Dvmscan.c283 #define for_each_managed_zone_pgdat(zone, pgdat, idx, highidx) \ argument
284 for ((idx) = 0, (zone) = (pgdat)->node_zones; \
346 struct pglist_data *pgdat = NODE_DATA(nid); in can_demote() local
349 if (!pgdat || !numa_demotion_enabled) in can_demote()
354 node_get_allowed_targets(pgdat, &allowed_mask); in can_demote()
502 static bool skip_throttle_noprogress(pg_data_t *pgdat) in skip_throttle_noprogress() argument
511 if (kswapd_test_hopeless(pgdat)) in skip_throttle_noprogress()
519 for_each_managed_zone_pgdat(zone, pgdat, i, MAX_NR_ZONES - 1) { in skip_throttle_noprogress()
530 void reclaim_throttle(pg_data_t *pgdat, enum vmscan_throttle_state reason) in reclaim_throttle() argument
532 wait_queue_head_t *wqh = &pgdat->reclaim_wait[reason]; in reclaim_throttle()
[all …]
H A Dvmstat.c277 struct pglist_data *pgdat; in refresh_zone_stat_thresholds() local
282 /* Zero current pgdat thresholds */ in refresh_zone_stat_thresholds()
283 for_each_online_pgdat(pgdat) { in refresh_zone_stat_thresholds()
285 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold = 0; in refresh_zone_stat_thresholds()
290 struct pglist_data *pgdat = zone->zone_pgdat; in refresh_zone_stat_thresholds() local
302 pgdat_threshold = per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold; in refresh_zone_stat_thresholds()
303 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold in refresh_zone_stat_thresholds()
320 void set_pgdat_percpu_threshold(pg_data_t *pgdat, in set_pgdat_percpu_threshold() argument
328 for (i = 0; i < pgdat->nr_zones; i++) { in set_pgdat_percpu_threshold()
329 zone = &pgdat in set_pgdat_percpu_threshold()
376 __mod_node_page_state(struct pglist_data * pgdat,enum node_stat_item item,long delta) __mod_node_page_state() argument
456 __inc_node_state(struct pglist_data * pgdat,enum node_stat_item item) __inc_node_state() argument
512 __dec_node_state(struct pglist_data * pgdat,enum node_stat_item item) __dec_node_state() argument
618 mod_node_state(struct pglist_data * pgdat,enum node_stat_item item,int delta,int overstep_mode) mod_node_state() argument
668 mod_node_page_state(struct pglist_data * pgdat,enum node_stat_item item,long delta) mod_node_page_state() argument
675 inc_node_state(struct pglist_data * pgdat,enum node_stat_item item) inc_node_state() argument
728 inc_node_state(struct pglist_data * pgdat,enum node_stat_item item) inc_node_state() argument
738 mod_node_page_state(struct pglist_data * pgdat,enum node_stat_item item,long delta) mod_node_page_state() argument
752 struct pglist_data *pgdat; inc_node_page_state() local
816 struct pglist_data *pgdat; refresh_cpu_vm_stats() local
906 struct pglist_data *pgdat; cpu_vm_stats_fold() local
1024 node_page_state_pages(struct pglist_data * pgdat,enum node_stat_item item) node_page_state_pages() argument
1035 node_page_state(struct pglist_data * pgdat,enum node_stat_item item) node_page_state() argument
1513 pg_data_t *pgdat; frag_start() local
1526 pg_data_t *pgdat = (pg_data_t *)arg; frag_next() local
1540 walk_zones_in_node(struct seq_file * m,pg_data_t * pgdat,bool assert_populated,bool nolock,void (* print)(struct seq_file * m,pg_data_t *,struct zone *)) walk_zones_in_node() argument
1562 frag_show_print(struct seq_file * m,pg_data_t * pgdat,struct zone * zone) frag_show_print() argument
1582 pg_data_t *pgdat = (pg_data_t *)arg; frag_show() local
1588 pagetypeinfo_showfree_print(struct seq_file * m,pg_data_t * pgdat,struct zone * zone) pagetypeinfo_showfree_print() argument
1633 pg_data_t *pgdat = (pg_data_t *)arg; pagetypeinfo_showfree() local
1645 pagetypeinfo_showblockcount_print(struct seq_file * m,pg_data_t * pgdat,struct zone * zone) pagetypeinfo_showblockcount_print() argument
1680 pg_data_t *pgdat = (pg_data_t *)arg; pagetypeinfo_showblockcount() local
1696 pagetypeinfo_showmixedcount(struct seq_file * m,pg_data_t * pgdat) pagetypeinfo_showmixedcount() argument
1722 pg_data_t *pgdat = (pg_data_t *)arg; pagetypeinfo_show() local
1752 is_zone_first_populated(pg_data_t * pgdat,struct zone * zone) is_zone_first_populated() argument
1766 zoneinfo_show_print(struct seq_file * m,pg_data_t * pgdat,struct zone * zone) zoneinfo_show_print() argument
1873 pg_data_t *pgdat = (pg_data_t *)arg; zoneinfo_show() local
2331 unusable_show_print(struct seq_file * m,pg_data_t * pgdat,struct zone * zone) unusable_show_print() argument
2360 pg_data_t *pgdat = (pg_data_t *)arg; unusable_show() local
2381 extfrag_show_print(struct seq_file * m,pg_data_t * pgdat,struct zone * zone) extfrag_show_print() argument
2406 pg_data_t *pgdat = (pg_data_t *)arg; extfrag_show() local
[all...]
H A Dcompaction.c402 void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument
407 struct zone *zone = &pgdat->node_zones[zoneid]; in reset_isolation_suitable()
763 pg_data_t *pgdat = cc->zone->zone_pgdat; in too_many_isolated() local
768 inactive = node_page_state(pgdat, NR_INACTIVE_FILE) + in too_many_isolated()
769 node_page_state(pgdat, NR_INACTIVE_ANON); in too_many_isolated()
770 active = node_page_state(pgdat, NR_ACTIVE_FILE) + in too_many_isolated()
771 node_page_state(pgdat, NR_ACTIVE_ANON); in too_many_isolated()
772 isolated = node_page_state(pgdat, NR_ISOLATED_FILE) + in too_many_isolated()
773 node_page_state(pgdat, NR_ISOLATED_ANON); in too_many_isolated()
788 wake_throttle_isolated(pgdat); in too_many_isolated()
[all …]
H A Dmmzone.c18 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat) in next_online_pgdat() argument
20 int nid = next_online_node(pgdat->node_id); in next_online_pgdat()
32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() local
34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone()
37 pgdat = next_online_pgdat(pgdat); in next_zone()
38 if (pgdat) in next_zone()
39 zone = pgdat->node_zones; in next_zone()
H A Dworkingset.c199 static void *pack_shadow(int memcgid, pg_data_t *pgdat, unsigned long eviction, in pack_shadow() argument
204 eviction = (eviction << NODES_SHIFT) | pgdat->node_id; in pack_shadow()
210 static void unpack_shadow(void *shadow, int *memcgidp, pg_data_t **pgdat, in unpack_shadow() argument
225 *pgdat = NODE_DATA(nid); in unpack_shadow()
245 struct pglist_data *pgdat = folio_pgdat(folio); in lru_gen_eviction() local
249 lruvec = mem_cgroup_lruvec(memcg, pgdat); in lru_gen_eviction()
257 return pack_shadow(mem_cgroup_private_id(memcg), pgdat, token, workingset); in lru_gen_eviction()
270 struct pglist_data *pgdat; in lru_gen_test_recent() local
272 unpack_shadow(shadow, &memcg_id, &pgdat, token, workingset); in lru_gen_test_recent()
275 *lruvec = mem_cgroup_lruvec(memcg, pgdat); in lru_gen_test_recent()
383 struct pglist_data *pgdat = folio_pgdat(folio); workingset_eviction() local
427 struct pglist_data *pgdat; workingset_test_recent() local
537 struct pglist_data *pgdat; workingset_refault() local
[all...]
H A Dmemory-tiers.c262 pg_data_t *pgdat; in __node_get_memory_tier() local
264 pgdat = NODE_DATA(node); in __node_get_memory_tier()
265 if (!pgdat) in __node_get_memory_tier()
272 return rcu_dereference_check(pgdat->memtier, in __node_get_memory_tier()
280 pg_data_t *pgdat; in node_is_toptier() local
283 pgdat = NODE_DATA(node); in node_is_toptier()
284 if (!pgdat) in node_is_toptier()
288 memtier = rcu_dereference(pgdat->memtier); in node_is_toptier()
302 void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets() argument
312 memtier = rcu_dereference(pgdat->memtier); in node_get_allowed_targets()
[all …]
H A Dshuffle.h11 extern void __shuffle_free_memory(pg_data_t *pgdat);
13 static inline void __meminit shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument
17 __shuffle_free_memory(pgdat); in shuffle_free_memory()
40 static inline void shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument
H A Dmemory_hotplug.c519 static void update_pgdat_span(struct pglist_data *pgdat) in update_pgdat_span() argument
524 for (zone = pgdat->node_zones; in update_pgdat_span()
525 zone < pgdat->node_zones + MAX_NR_ZONES; zone++) { in update_pgdat_span()
543 pgdat->node_start_pfn = node_start_pfn; in update_pgdat_span()
544 pgdat->node_spanned_pages = node_end_pfn - node_start_pfn; in update_pgdat_span()
552 struct pglist_data *pgdat = zone->zone_pgdat; in remove_pfn_range_from_zone() local
577 update_pgdat_span(pgdat); in remove_pfn_range_from_zone()
714 static void __meminit resize_pgdat_range(struct pglist_data *pgdat, unsigned long start_pfn, in resize_pgdat_range() argument
717 unsigned long old_end_pfn = pgdat_end_pfn(pgdat); in resize_pgdat_range()
719 if (!pgdat in resize_pgdat_range()
753 struct pglist_data *pgdat = zone->zone_pgdat; move_pfn_range_to_zone() local
859 pg_data_t *pgdat = NODE_DATA(nid); auto_movable_can_online_movable() local
901 struct pglist_data *pgdat = NODE_DATA(nid); default_kernel_zone_for_pfn() local
1255 struct pglist_data *pgdat; hotadd_init_pgdat() local
1290 pg_data_t *pgdat; __try_online_node() local
1899 struct pglist_data *pgdat = zone->zone_pgdat; offline_pages() local
[all...]
H A Dshuffle.c153 void __meminit __shuffle_free_memory(pg_data_t *pgdat) in __shuffle_free_memory() argument
157 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in __shuffle_free_memory()
H A Dnuma.c37 pg_data_t *pgdat; in alloc_offline_node_data() local
38 node_data[nid] = memblock_alloc_or_panic(sizeof(*pgdat), SMP_CACHE_BYTES); in alloc_offline_node_data()
H A Dpage_alloc.c2944 struct pglist_data *pgdat = zone->zone_pgdat; in free_frozen_page_commit() local
2954 if (kswapd_test_hopeless(pgdat) && in free_frozen_page_commit()
2955 next_memory_node(pgdat->node_id) < MAX_NUMNODES) in free_frozen_page_commit()
2956 kswapd_clear_hopeless(pgdat, KSWAPD_CLEAR_HOPELESS_PCP); in free_frozen_page_commit()
5551 static int build_zonerefs_node(pg_data_t *pgdat, struct zoneref *zonerefs) in build_zonerefs_node() argument
5559 zone = pgdat->node_zones + zone_type; in build_zonerefs_node()
5669 static void build_zonelists_in_node_order(pg_data_t *pgdat, int *node_order, in build_zonelists_in_node_order() argument
5675 zonerefs = pgdat->node_zonelists[ZONELIST_FALLBACK]._zonerefs; in build_zonelists_in_node_order()
5692 static void build_thisnode_zonelists(pg_data_t *pgdat) in build_thisnode_zonelists() argument
5697 zonerefs = pgdat->node_zonelists[ZONELIST_NOFALLBACK]._zonerefs; in build_thisnode_zonelists()
[all …]
H A Dmemcontrol.c794 pg_data_t *pgdat = folio_pgdat(folio); in lruvec_stat_mod_folio() local
802 mod_node_page_state(pgdat, idx, val); in lruvec_stat_mod_folio()
806 lruvec = mem_cgroup_lruvec(memcg, pgdat); in lruvec_stat_mod_folio()
814 pg_data_t *pgdat = page_pgdat(virt_to_page(p)); in mod_lruvec_kmem_state() local
828 mod_node_page_state(pgdat, idx, val); in mod_lruvec_kmem_state()
830 lruvec = mem_cgroup_lruvec(memcg, pgdat); in mod_lruvec_kmem_state()
1042 int nid = reclaim->pgdat->node_id; in mem_cgroup_iter()
2574 struct pglist_data *pgdat, in account_slab_nmi_safe() argument
2580 lruvec = mem_cgroup_lruvec(memcg, pgdat); in account_slab_nmi_safe()
2583 struct mem_cgroup_per_node *pn = memcg->nodeinfo[pgdat->node_id]; in account_slab_nmi_safe()
[all …]
H A Dinternal.h479 void __acct_reclaim_writeback(pg_data_t *pgdat, struct folio *folio,
483 pg_data_t *pgdat = folio_pgdat(folio); in acct_reclaim_writeback() local
484 int nr_throttled = atomic_read(&pgdat->nr_writeback_throttled); in acct_reclaim_writeback()
487 __acct_reclaim_writeback(pgdat, folio, nr_throttled); in acct_reclaim_writeback()
490 static inline void wake_throttle_isolated(pg_data_t *pgdat) in wake_throttle_isolated() argument
494 wqh = &pgdat->reclaim_wait[VMSCAN_THROTTLE_ISOLATED]; in wake_throttle_isolated()
617 extern void reclaim_throttle(pg_data_t *pgdat, enum vmscan_throttle_state reason);
619 struct mem_cgroup *memcg, pg_data_t *pgdat);
1292 static inline int node_reclaim(struct pglist_data *pgdat, gfp_t mask, in node_reclaim() argument
/linux/tools/testing/vma/linux/
H A Dmmzone.h9 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat);
11 #define for_each_online_pgdat(pgdat) \ argument
12 for (pgdat = first_online_pgdat(); \
13 pgdat; \
14 pgdat = next_online_pgdat(pgdat))
/linux/tools/testing/memblock/linux/
H A Dmmzone.h9 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat);
11 #define for_each_online_pgdat(pgdat) \ argument
12 for (pgdat = first_online_pgdat(); \
13 pgdat; \
14 pgdat = next_online_pgdat(pgdat))
/linux/drivers/base/
H A Dnode.c446 struct pglist_data *pgdat = NODE_DATA(nid); in node_read_meminfo() local
452 sreclaimable = node_page_state_pages(pgdat, NR_SLAB_RECLAIMABLE_B); in node_read_meminfo()
453 sunreclaimable = node_page_state_pages(pgdat, NR_SLAB_UNRECLAIMABLE_B); in node_read_meminfo()
455 swapcached = node_page_state_pages(pgdat, NR_SWAPCACHE); in node_read_meminfo()
474 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON) + in node_read_meminfo()
475 node_page_state(pgdat, NR_ACTIVE_FILE)), in node_read_meminfo()
476 nid, K(node_page_state(pgdat, NR_INACTIVE_ANON) + in node_read_meminfo()
477 node_page_state(pgdat, NR_INACTIVE_FILE)), in node_read_meminfo()
478 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON)), in node_read_meminfo()
479 nid, K(node_page_state(pgdat, NR_INACTIVE_ANON)), in node_read_meminfo()
[all …]
/linux/include/linux/
H A Dvmstat.h169 static inline void node_page_state_add(long x, struct pglist_data *pgdat, in node_page_state_add() argument
172 atomic_long_add(x, &pgdat->vm_stat[item]); in node_page_state_add()
258 extern unsigned long node_page_state(struct pglist_data *pgdat,
260 extern unsigned long node_page_state_pages(struct pglist_data *pgdat,
302 void set_pgdat_percpu_threshold(pg_data_t *pgdat,
317 static inline void __mod_node_page_state(struct pglist_data *pgdat,
331 node_page_state_add(delta, pgdat, item); in __mod_node_page_state()
340 static inline void __inc_node_state(struct pglist_data *pgdat, enum node_stat_item item)
342 atomic_long_inc(&pgdat->vm_stat[item]); in __inc_node_state()
352 static inline void __dec_node_state(struct pglist_data *pgdat, enu in __dec_zone_state()
318 __mod_node_page_state(struct pglist_data * pgdat,enum node_stat_item item,int delta) __mod_node_page_state() argument
341 __inc_node_state(struct pglist_data * pgdat,enum node_stat_item item) __inc_node_state() argument
353 __dec_node_state(struct pglist_data * pgdat,enum node_stat_item item) __dec_node_state() argument
401 set_pgdat_percpu_threshold(pgdat,callback) global() argument
[all...]
H A Dmemcontrol.h60 pg_data_t *pgdat; member
708 struct pglist_data *pgdat) in mem_cgroup_lruvec() argument
714 lruvec = &pgdat->__lruvec; in mem_cgroup_lruvec()
721 mz = memcg->nodeinfo[pgdat->node_id]; in mem_cgroup_lruvec()
729 if (unlikely(lruvec->pgdat != pgdat)) in mem_cgroup_lruvec()
730 lruvec->pgdat = pgdat; in mem_cgroup_lruvec()
1171 struct pglist_data *pgdat) in mem_cgroup_lruvec() argument
1173 return &pgdat->__lruvec; in mem_cgroup_lruvec()
1178 struct pglist_data *pgdat = folio_pgdat(folio); in folio_lruvec() local
1179 return &pgdat->__lruvec; in folio_lruvec()
[all …]
H A Dcompaction.h96 extern void reset_isolation_suitable(pg_data_t *pgdat);
108 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx);
111 static inline void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument
129 static inline void wakeup_kcompactd(pg_data_t *pgdat, in wakeup_kcompactd() argument
H A Dmmzone.h617 void lru_gen_init_pgdat(struct pglist_data *pgdat);
630 static inline void lru_gen_init_pgdat(struct pglist_data *pgdat) in lru_gen_init_pgdat() argument
695 struct pglist_data *pgdat; member
1529 static inline unsigned long pgdat_end_pfn(pg_data_t *pgdat) in pgdat_end_pfn() argument
1531 return pgdat->node_start_pfn + pgdat->node_spanned_pages; in pgdat_end_pfn()
1536 void build_all_zonelists(pg_data_t *pgdat);
1553 void kswapd_try_clear_hopeless(struct pglist_data *pgdat,
1555 void kswapd_clear_hopeless(pg_data_t *pgdat, enum kswapd_clear_hopeless_reason reason);
1556 bool kswapd_test_hopeless(pg_data_t *pgdat);
1575 return lruvec->pgdat; in lruvec_pgdat()
[all …]
H A Dmemory-tiers.h57 void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets);
65 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets() argument
109 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets() argument
/linux/mm/damon/
H A Dops-common.c301 struct list_head *migrate_folios, struct pglist_data *pgdat, in __damon_migrate_folio_list() argument
316 if (pgdat->node_id == target_nid || target_nid == NUMA_NO_NODE) in __damon_migrate_folio_list()
331 struct pglist_data *pgdat, in damon_migrate_folio_list() argument
361 &migrate_folios, pgdat, target_nid); in damon_migrate_folio_list()
/linux/tools/perf/util/bpf_skel/
H A Dlock_contention.bpf.c893 struct pglist_data *pgdat = NULL; in collect_zone_lock() local
896 err = bpf_core_read(&pgdat, sizeof(pgdat), &node_data[i]); in collect_zone_lock()
897 if (err < 0 || pgdat == NULL) in collect_zone_lock()
900 nr_zones = BPF_CORE_READ(pgdat, nr_zones); in collect_zone_lock()
907 zone_addr = (__u64)(void *)pgdat + (sizeof_zone * k) + zone_off; in collect_zone_lock()

12