| /linux/include/linux/ |
| H A D | compaction.h | 96 extern void reset_isolation_suitable(pg_data_t *pgdat); 108 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx); 111 static inline void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() 129 static inline void wakeup_kcompactd(pg_data_t *pgdat, in wakeup_kcompactd()
|
| H A D | memory-tiers.h | 57 void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets); 65 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets() 109 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets()
|
| H A D | mmzone.h | 1521 } pg_data_t; typedef 1529 static inline unsigned long pgdat_end_pfn(pg_data_t *pgdat) in pgdat_end_pfn() 1536 void build_all_zonelists(pg_data_t *pgdat); 1555 void kswapd_clear_hopeless(pg_data_t *pgdat, enum kswapd_clear_hopeless_reason reason); 1556 bool kswapd_test_hopeless(pg_data_t *pgdat);
|
| H A D | memcontrol.h | 60 pg_data_t *pgdat; 1876 unsigned long memcg1_soft_limit_reclaim(pg_data_t *pgdat, int order, 1904 unsigned long memcg1_soft_limit_reclaim(pg_data_t *pgdat, int order, in memcg1_soft_limit_reclaim()
|
| H A D | swap.h | 390 pg_data_t *pgdat,
|
| H A D | vmstat.h | 302 void set_pgdat_percpu_threshold(pg_data_t *pgdat,
|
| /linux/mm/ |
| H A D | shuffle.h | 11 extern void __shuffle_free_memory(pg_data_t *pgdat); 13 static inline void __meminit shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() 40 static inline void shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory()
|
| H A D | numa.c | 14 const size_t nd_size = roundup(sizeof(pg_data_t), SMP_CACHE_BYTES); in alloc_node_data() 32 memset(NODE_DATA(nid), 0, sizeof(pg_data_t)); in alloc_node_data() 37 pg_data_t *pgdat; in alloc_offline_node_data()
|
| H A D | vmstat.c | 320 void set_pgdat_percpu_threshold(pg_data_t *pgdat, in set_pgdat_percpu_threshold() 1498 pg_data_t *pgdat; 1511 pg_data_t *pgdat = (pg_data_t *)arg; in frag_start() 1525 static void walk_zones_in_node(struct seq_file *m, pg_data_t *pgdat, in frag_next() 1527 void (*print)(struct seq_file *m, pg_data_t *, struct zone *)) in frag_next() 1547 static void frag_show_print(struct seq_file *m, pg_data_t *pgdat, in walk_zones_in_node() 1567 pg_data_t *pgdat = (pg_data_t *)arg; in frag_show_print() 1573 pg_data_t *pgda in frag_show_print() 1540 walk_zones_in_node(struct seq_file * m,pg_data_t * pgdat,bool assert_populated,bool nolock,void (* print)(struct seq_file * m,pg_data_t *,struct zone *)) walk_zones_in_node() argument [all...] |
| H A D | compaction.c | 402 void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() 763 pg_data_t *pgdat = cc->zone->zone_pgdat; in too_many_isolated() 840 pg_data_t *pgdat = cc->zone->zone_pgdat; in isolate_migratepages_block() 2152 static bool kswapd_is_running(pg_data_t *pgdat) in kswapd_is_running() 2197 static unsigned int fragmentation_score_node(pg_data_t *pgdat) in fragmentation_score_node() 2223 static bool should_proactive_compact_node(pg_data_t *pgdat) in should_proactive_compact_node() 2262 pg_data_t *pgdat; in __compact_finished() 2893 static int compact_node(pg_data_t *pgdat, bool proactive) in compact_node() 2957 pg_data_t *pgdat = NODE_DATA(nid); in compaction_proactiveness_sysctl_handler() 3023 static inline bool kcompactd_work_requested(pg_data_t *pgdat) in kcompactd_work_requested() [all …]
|
| H A D | memory-tiers.c | 262 pg_data_t *pgdat; in __node_get_memory_tier() 280 pg_data_t *pgdat; in node_is_toptier() 302 void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets() 547 pg_data_t *pgdat = NODE_DATA(node); in set_node_memory_tier() 583 pg_data_t *pgdat; in clear_node_memory_tier()
|
| H A D | show_mem.c | 95 pg_data_t *pgdat = NODE_DATA(nid); in si_meminfo_node() 162 static bool node_has_managed_zones(pg_data_t *pgdat, int max_zone_idx) in node_has_managed_zones() 185 pg_data_t *pgdat; in show_free_areas()
|
| H A D | vmscan.c | 502 static bool skip_throttle_noprogress(pg_data_t *pgdat) in skip_throttle_noprogress() 530 void reclaim_throttle(pg_data_t *pgdat, enum vmscan_throttle_state reason) in reclaim_throttle() 604 void __acct_reclaim_writeback(pg_data_t *pgdat, struct folio *folio, in __acct_reclaim_writeback() 2317 static void prepare_scan_control(pg_data_t *pgdat, struct scan_control *sc) in prepare_scan_control() 5960 static void shrink_node_memcgs(pg_data_t *pgdat, struct scan_control *sc) in shrink_node_memcgs() 6039 static void shrink_node(pg_data_t *pgdat, struct scan_control *sc) in shrink_node() 6183 static void consider_reclaim_throttle(pg_data_t *pgdat, struct scan_control *sc) in consider_reclaim_throttle() 6228 pg_data_t *last_pgdat = NULL; in shrink_zones() 6229 pg_data_t *first_pgdat = NULL; in shrink_zones() 6313 static void snapshot_refaults(struct mem_cgroup *target_memcg, pg_data_t *pgdat) in snapshot_refaults() [all …]
|
| H A D | mm_init.c | 68 pg_data_t *pgdat = NODE_DATA(nid); in mminit_verify_zonelist() 671 pg_data_t *pgdat; 690 static inline void pgdat_set_deferred_range(pg_data_t *pgdat) in __init_page_from_nid() 753 static inline void pgdat_set_deferred_range(pg_data_t *pgdat) {} in __init_deferred_page() 1716 pg_data_t *pgdat = NODE_DATA(nid); 1720 /* pg_data_t should be reset to zero when it's allocated */ in free_area_init_node() 1749 static void __init check_for_memory(pg_data_t *pgdat) in free_area_init_node() 1809 * free_area_init - Initialise all pg_data_t and zone data in set_high_memory() 1897 pg_data_t *pgdat; in free_area_init() 2110 pg_data_t *pgda in deferred_page_init_max_threads() [all...] |
| H A D | mmzone.c | 32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone()
|
| H A D | shuffle.c | 153 void __meminit __shuffle_free_memory(pg_data_t *pgdat) in __shuffle_free_memory()
|
| H A D | internal.h | 479 void __acct_reclaim_writeback(pg_data_t *pgdat, struct folio *folio, 483 pg_data_t *pgdat = folio_pgdat(folio); in acct_reclaim_writeback() 490 static inline void wake_throttle_isolated(pg_data_t *pgdat) in wake_throttle_isolated() 617 extern void reclaim_throttle(pg_data_t *pgdat, enum vmscan_throttle_state reason); 619 struct mem_cgroup *memcg, pg_data_t *pgdat);
|
| H A D | workingset.c | 199 static void *pack_shadow(int memcgid, pg_data_t *pgdat, unsigned long eviction, in pack_shadow() 210 static void unpack_shadow(void *shadow, int *memcgidp, pg_data_t **pgdat, in unpack_shadow()
|
| H A D | page_alloc.c | 4475 pg_data_t *last_pgdat = NULL; in wake_all_kswapds() 5551 static int build_zonerefs_node(pg_data_t *pgdat, struct zoneref *zonerefs) in build_zonerefs_node() 5669 static void build_zonelists_in_node_order(pg_data_t *pgdat, int *node_order, in build_zonelists_in_node_order() 5680 pg_data_t *node = NODE_DATA(node_order[i]); in build_zonelists_in_node_order() 5692 static void build_thisnode_zonelists(pg_data_t *pgdat) in build_thisnode_zonelists() 5704 static void build_zonelists(pg_data_t *pgdat) in build_zonelists() 5760 static void build_zonelists(pg_data_t *pgdat) in build_zonelists() 5801 pg_data_t *self = data; in __build_all_zonelists() 5833 pg_data_t *pgdat = NODE_DATA(nid); in __build_all_zonelists() 5889 void __ref build_all_zonelists(pg_data_t *pgdat) in build_all_zonelists() [all …]
|
| H A D | memory_hotplug.c | 859 pg_data_t *pgdat = NODE_DATA(nid); in auto_movable_can_online_movable() 1253 static pg_data_t *hotadd_init_pgdat(int nid) in hotadd_init_pgdat() 1290 pg_data_t *pgdat; in __try_online_node()
|
| H A D | memcontrol-v1.c | 273 pg_data_t *pgdat, in mem_cgroup_soft_reclaim() 322 unsigned long memcg1_soft_limit_reclaim(pg_data_t *pgdat, int order, in memcg1_soft_limit_reclaim() 1947 pg_data_t *pgdat; in memcg1_stat_format()
|
| H A D | slab.h | 165 static inline pg_data_t *slab_pgdat(const struct slab *slab)
|
| /linux/tools/testing/vma/linux/ |
| H A D | mmzone.h | 36 } pg_data_t; typedef
|
| /linux/tools/testing/memblock/linux/ |
| H A D | mmzone.h | 36 } pg_data_t; typedef
|
| /linux/mm/kmsan/ |
| H A D | init.c | 77 const size_t nd_size = sizeof(pg_data_t); in kmsan_init_shadow()
|