Home
last modified time | relevance | path

Searched refs:pgdat (Results 1 – 25 of 48) sorted by relevance

12

/linux/mm/
H A Dshow_mem.c95 pg_data_t *pgdat = NODE_DATA(nid); in si_meminfo_node() local
98 managed_pages += zone_managed_pages(&pgdat->node_zones[zone_type]); in si_meminfo_node()
100 val->sharedram = node_page_state(pgdat, NR_SHMEM); in si_meminfo_node()
104 struct zone *zone = &pgdat->node_zones[zone_type]; in si_meminfo_node()
168 static bool node_has_managed_zones(pg_data_t *pgdat, int max_zone_idx) in node_has_managed_zones() argument
172 if (zone_managed_pages(pgdat->node_zones + zone_idx)) in node_has_managed_zones()
191 pg_data_t *pgdat; in show_free_areas() local
232 for_each_online_pgdat(pgdat) { in show_free_areas()
233 if (show_mem_node_skip(filter, pgdat->node_id, nodemask)) in show_free_areas()
235 if (!node_has_managed_zones(pgdat, max_zone_id in show_free_areas()
[all...]
H A Dmm_init.c50 pg_data_t *pgdat = NODE_DATA(nid); in mminit_verify_zonelist()
62 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist()
63 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist()
654 static inline void pgdat_set_deferred_range(pg_data_t *pgdat) in pgdat_set_deferred_range() argument
656 pgdat->first_deferred_pfn = ULONG_MAX; in pgdat_set_deferred_range()
709 pg_data_t *pgdat; in init_reserved_page() local
715 pgdat = NODE_DATA(nid); in init_reserved_page()
718 struct zone *zone = &pgdat->node_zones[zid]; in init_reserved_page()
726 static inline void pgdat_set_deferred_range(pg_data_t *pgdat) {} in pgdat_set_deferred_range() argument
1071 struct pglist_data *pgdat in memmap_init_zone_device() local
49 pg_data_t *pgdat = NODE_DATA(nid); mminit_verify_zonelist() local
1251 reset_memoryless_node_totalpages(struct pglist_data * pgdat) reset_memoryless_node_totalpages() argument
1269 calculate_node_totalpages(struct pglist_data * pgdat,unsigned long node_start_pfn,unsigned long node_end_pfn) calculate_node_totalpages() argument
1333 pgdat_init_split_queue(struct pglist_data * pgdat) pgdat_init_split_queue() argument
1342 pgdat_init_split_queue(struct pglist_data * pgdat) pgdat_init_split_queue() argument
1346 pgdat_init_kcompactd(struct pglist_data * pgdat) pgdat_init_kcompactd() argument
1351 pgdat_init_kcompactd(struct pglist_data * pgdat) pgdat_init_kcompactd() argument
1354 pgdat_init_internals(struct pglist_data * pgdat) pgdat_init_internals() argument
1403 struct pglist_data *pgdat = zone->zone_pgdat; init_currently_empty_zone() local
1503 free_area_init_core_hotplug(struct pglist_data * pgdat) free_area_init_core_hotplug() argument
1555 free_area_init_core(struct pglist_data * pgdat) free_area_init_core() argument
1636 alloc_node_mem_map(struct pglist_data * pgdat) alloc_node_mem_map() argument
1673 alloc_node_mem_map(struct pglist_data * pgdat) alloc_node_mem_map() argument
1706 pg_data_t *pgdat = NODE_DATA(nid); free_area_init_node() local
1739 check_for_memory(pg_data_t * pgdat) check_for_memory() argument
1869 pg_data_t *pgdat; free_area_init() local
2178 pg_data_t *pgdat = data; deferred_init_memmap() local
2271 pg_data_t *pgdat = zone->zone_pgdat; deferred_grow_zone() local
[all...]
H A Dvmstat.c274 struct pglist_data *pgdat; in refresh_zone_stat_thresholds() local
280 for_each_online_pgdat(pgdat) { in refresh_zone_stat_thresholds()
282 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold = 0; in refresh_zone_stat_thresholds()
287 struct pglist_data *pgdat = zone->zone_pgdat; in refresh_zone_stat_thresholds() local
299 pgdat_threshold = per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold; in refresh_zone_stat_thresholds()
300 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold in refresh_zone_stat_thresholds()
317 void set_pgdat_percpu_threshold(pg_data_t *pgdat, in set_pgdat_percpu_threshold() argument
325 for (i = 0; i < pgdat->nr_zones; i++) { in set_pgdat_percpu_threshold()
326 zone = &pgdat->node_zones[i]; in set_pgdat_percpu_threshold()
373 void __mod_node_page_state(struct pglist_data *pgdat, enum node_stat_item item, in __mod_node_page_state() argument
[all …]
H A Dvmscan.c466 static bool skip_throttle_noprogress(pg_data_t *pgdat) in skip_throttle_noprogress() argument
475 if (pgdat->kswapd_failures >= MAX_RECLAIM_RETRIES) in skip_throttle_noprogress()
484 struct zone *zone = pgdat->node_zones + i; in skip_throttle_noprogress()
499 void reclaim_throttle(pg_data_t *pgdat, enum vmscan_throttle_state reason) in reclaim_throttle() argument
501 wait_queue_head_t *wqh = &pgdat->reclaim_wait[reason]; in reclaim_throttle()
530 if (atomic_inc_return(&pgdat->nr_writeback_throttled) == 1) { in reclaim_throttle()
531 WRITE_ONCE(pgdat->nr_reclaim_start, in reclaim_throttle()
532 node_page_state(pgdat, NR_THROTTLED_WRITTEN)); in reclaim_throttle()
539 if (skip_throttle_noprogress(pgdat)) { in reclaim_throttle()
561 atomic_dec(&pgdat in reclaim_throttle()
573 __acct_reclaim_writeback(pg_data_t * pgdat,struct folio * folio,int nr_throttled) __acct_reclaim_writeback() argument
955 demote_folio_list(struct list_head * demote_folios,struct pglist_data * pgdat) demote_folio_list() argument
1012 shrink_folio_list(struct list_head * folio_list,struct pglist_data * pgdat,struct scan_control * sc,struct reclaim_stat * stat,bool ignore_references) shrink_folio_list() argument
1754 too_many_isolated(struct pglist_data * pgdat,int file,struct scan_control * sc) too_many_isolated() argument
1891 struct pglist_data *pgdat = lruvec_pgdat(lruvec); shrink_inactive_list() local
2012 struct pglist_data *pgdat = lruvec_pgdat(lruvec); shrink_active_list() local
2094 reclaim_folio_list(struct list_head * folio_list,struct pglist_data * pgdat,bool ignore_references) reclaim_folio_list() argument
2220 prepare_scan_control(pg_data_t * pgdat,struct scan_control * sc) prepare_scan_control() argument
2337 struct pglist_data *pgdat = lruvec_pgdat(lruvec); get_scan_count() local
2535 can_age_anon_pages(struct pglist_data * pgdat,struct scan_control * sc) can_age_anon_pages() argument
2591 struct pglist_data *pgdat = NODE_DATA(nid); get_lruvec() local
2612 struct pglist_data *pgdat = lruvec_pgdat(lruvec); get_swappiness() local
2762 struct pglist_data *pgdat = lruvec_pgdat(walk->lruvec); get_next_mm() local
3300 get_pfn_folio(unsigned long pfn,struct mem_cgroup * memcg,struct pglist_data * pgdat,bool can_swap) get_pfn_folio() argument
3341 struct pglist_data *pgdat = lruvec_pgdat(walk->lruvec); walk_pte_range() local
3409 struct pglist_data *pgdat = lruvec_pgdat(walk->lruvec); walk_pmd_range_locked() local
3516 struct pglist_data *pgdat = lruvec_pgdat(walk->lruvec); walk_pmd_range() local
3650 set_mm_walk(struct pglist_data * pgdat,bool force_alloc) set_mm_walk() argument
3938 lru_gen_age_node(struct pglist_data * pgdat,struct scan_control * sc) lru_gen_age_node() argument
4001 struct pglist_data *pgdat = folio_pgdat(folio); lru_gen_look_around() local
4109 struct pglist_data *pgdat = lruvec_pgdat(lruvec); lru_gen_rotate_memcg() local
4158 struct pglist_data *pgdat = NODE_DATA(nid); lru_gen_online_memcg() local
4193 struct pglist_data *pgdat = NODE_DATA(nid); lru_gen_release_memcg() local
4521 struct pglist_data *pgdat = lruvec_pgdat(lruvec); evict_folios() local
4757 struct pglist_data *pgdat = lruvec_pgdat(lruvec); shrink_one() local
4793 shrink_many(struct pglist_data * pgdat,struct scan_control * sc) shrink_many() argument
4884 set_initial_priority(struct pglist_data * pgdat,struct scan_control * sc) set_initial_priority() argument
4906 lru_gen_shrink_node(struct pglist_data * pgdat,struct scan_control * sc) lru_gen_shrink_node() argument
5533 lru_gen_init_pgdat(struct pglist_data * pgdat) lru_gen_init_pgdat() argument
5624 lru_gen_age_node(struct pglist_data * pgdat,struct scan_control * sc) lru_gen_age_node() argument
5634 lru_gen_shrink_node(struct pglist_data * pgdat,struct scan_control * sc) lru_gen_shrink_node() argument
5777 should_continue_reclaim(struct pglist_data * pgdat,unsigned long nr_reclaimed,struct scan_control * sc) should_continue_reclaim() argument
5829 shrink_node_memcgs(pg_data_t * pgdat,struct scan_control * sc) shrink_node_memcgs() argument
5887 shrink_node(pg_data_t * pgdat,struct scan_control * sc) shrink_node() argument
6035 consider_reclaim_throttle(pg_data_t * pgdat,struct scan_control * sc) consider_reclaim_throttle() argument
6165 snapshot_refaults(struct mem_cgroup * target_memcg,pg_data_t * pgdat) snapshot_refaults() argument
6285 allow_direct_reclaim(pg_data_t * pgdat) allow_direct_reclaim() argument
6339 pg_data_t *pgdat = NULL; throttle_direct_reclaim() local
6462 mem_cgroup_shrink_node(struct mem_cgroup * memcg,gfp_t gfp_mask,bool noswap,pg_data_t * pgdat,unsigned long * nr_scanned) mem_cgroup_shrink_node() argument
6539 kswapd_age_node(struct pglist_data * pgdat,struct scan_control * sc) kswapd_age_node() argument
6565 pgdat_watermark_boosted(pg_data_t * pgdat,int highest_zoneidx) pgdat_watermark_boosted() argument
6593 pgdat_balanced(pg_data_t * pgdat,int order,int highest_zoneidx) pgdat_balanced() argument
6629 clear_pgdat_congested(pg_data_t * pgdat) clear_pgdat_congested() argument
6645 prepare_kswapd_sleep(pg_data_t * pgdat,int order,int highest_zoneidx) prepare_kswapd_sleep() argument
6684 kswapd_shrink_node(pg_data_t * pgdat,struct scan_control * sc) kswapd_shrink_node() argument
6721 update_reclaim_active(pg_data_t * pgdat,int highest_zoneidx,bool active) update_reclaim_active() argument
6740 set_reclaim_active(pg_data_t * pgdat,int highest_zoneidx) set_reclaim_active() argument
6746 clear_reclaim_active(pg_data_t * pgdat,int highest_zoneidx) clear_reclaim_active() argument
6764 balance_pgdat(pg_data_t * pgdat,int order,int highest_zoneidx) balance_pgdat() argument
6992 kswapd_highest_zoneidx(pg_data_t * pgdat,enum zone_type prev_highest_zoneidx) kswapd_highest_zoneidx() argument
7000 kswapd_try_to_sleep(pg_data_t * pgdat,int alloc_order,int reclaim_order,unsigned int highest_zoneidx) kswapd_try_to_sleep() argument
7101 pg_data_t *pgdat = (pg_data_t *)p; kswapd() local
7185 pg_data_t *pgdat; wakeup_kswapd() local
7271 pg_data_t *pgdat = NODE_DATA(nid); kswapd_run() local
7293 pg_data_t *pgdat = NODE_DATA(nid); kswapd_stop() local
7345 node_unmapped_file_pages(struct pglist_data * pgdat) node_unmapped_file_pages() argument
7360 node_pagecache_reclaimable(struct pglist_data * pgdat) node_pagecache_reclaimable() argument
7390 __node_reclaim(struct pglist_data * pgdat,gfp_t gfp_mask,unsigned int order) __node_reclaim() argument
7443 node_reclaim(struct pglist_data * pgdat,gfp_t gfp_mask,unsigned int order) node_reclaim() argument
[all...]
H A Dcompaction.c453 void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument
458 struct zone *zone = &pgdat->node_zones[zoneid]; in reset_isolation_suitable()
817 pg_data_t *pgdat = cc->zone->zone_pgdat; in too_many_isolated() local
822 inactive = node_page_state(pgdat, NR_INACTIVE_FILE) + in too_many_isolated()
823 node_page_state(pgdat, NR_INACTIVE_ANON); in too_many_isolated()
824 active = node_page_state(pgdat, NR_ACTIVE_FILE) + in too_many_isolated()
825 node_page_state(pgdat, NR_ACTIVE_ANON); in too_many_isolated()
826 isolated = node_page_state(pgdat, NR_ISOLATED_FILE) + in too_many_isolated()
827 node_page_state(pgdat, NR_ISOLATED_ANON); in too_many_isolated()
842 wake_throttle_isolated(pgdat); in too_many_isolated()
894 pg_data_t *pgdat = cc->zone->zone_pgdat; isolate_migratepages_block() local
2203 kswapd_is_running(pg_data_t * pgdat) kswapd_is_running() argument
2248 fragmentation_score_node(pg_data_t * pgdat) fragmentation_score_node() argument
2278 should_proactive_compact_node(pg_data_t * pgdat) should_proactive_compact_node() argument
2317 pg_data_t *pgdat; __compact_finished() local
2900 compact_node(pg_data_t * pgdat,bool proactive) compact_node() argument
2964 pg_data_t *pgdat = NODE_DATA(nid); compaction_proactiveness_sysctl_handler() local
3030 kcompactd_work_requested(pg_data_t * pgdat) kcompactd_work_requested() argument
3036 kcompactd_node_suitable(pg_data_t * pgdat) kcompactd_node_suitable() argument
3059 kcompactd_do_work(pg_data_t * pgdat) kcompactd_do_work() argument
3137 wakeup_kcompactd(pg_data_t * pgdat,int order,int highest_zoneidx) wakeup_kcompactd() argument
3169 pg_data_t *pgdat = (pg_data_t *)p; kcompactd() local
3244 pg_data_t *pgdat = NODE_DATA(nid); kcompactd_run() local
3281 pg_data_t *pgdat = NODE_DATA(nid); kcompactd_cpu_online() local
[all...]
H A Dmmzone.c18 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat) in next_online_pgdat() argument
20 int nid = next_online_node(pgdat->node_id); in next_online_pgdat()
32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() local
34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone()
37 pgdat = next_online_pgdat(pgdat); in next_zone()
38 if (pgdat) in next_zone()
39 zone = pgdat->node_zones; in next_zone()
H A Dworkingset.c199 static void *pack_shadow(int memcgid, pg_data_t *pgdat, unsigned long eviction, in pack_shadow() argument
204 eviction = (eviction << NODES_SHIFT) | pgdat->node_id; in pack_shadow()
210 static void unpack_shadow(void *shadow, int *memcgidp, pg_data_t **pgdat, in unpack_shadow() argument
225 *pgdat = NODE_DATA(nid); in unpack_shadow()
244 struct pglist_data *pgdat = folio_pgdat(folio); in lru_gen_eviction() local
248 lruvec = mem_cgroup_lruvec(memcg, pgdat); in lru_gen_eviction()
256 return pack_shadow(mem_cgroup_id(memcg), pgdat, token, refs); in lru_gen_eviction()
269 struct pglist_data *pgdat; in lru_gen_test_recent() local
271 unpack_shadow(shadow, &memcg_id, &pgdat, token, workingset); in lru_gen_test_recent()
274 *lruvec = mem_cgroup_lruvec(memcg, pgdat); in lru_gen_test_recent()
384 struct pglist_data *pgdat = folio_pgdat(folio); workingset_eviction() local
426 struct pglist_data *pgdat; workingset_test_recent() local
533 struct pglist_data *pgdat; workingset_refault() local
[all...]
H A Dshuffle.h11 extern void __shuffle_free_memory(pg_data_t *pgdat);
13 static inline void __meminit shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument
17 __shuffle_free_memory(pgdat); in shuffle_free_memory()
40 static inline void shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument
H A Dmemory-tiers.c242 pg_data_t *pgdat; in __node_get_memory_tier()
244 pgdat = NODE_DATA(node); in __node_get_memory_tier()
245 if (!pgdat) in __node_get_memory_tier()
252 return rcu_dereference_check(pgdat->memtier, in node_is_toptier()
260 pg_data_t *pgdat; in node_is_toptier()
263 pgdat = NODE_DATA(node); in node_is_toptier()
264 if (!pgdat) in node_is_toptier()
268 memtier = rcu_dereference(pgdat->memtier); in node_is_toptier()
282 void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets()
292 memtier = rcu_dereference(pgdat in node_get_allowed_targets()
235 pg_data_t *pgdat; __node_get_memory_tier() local
253 pg_data_t *pgdat; node_is_toptier() local
275 node_get_allowed_targets(pg_data_t * pgdat,nodemask_t * targets) node_get_allowed_targets() argument
509 pg_data_t *pgdat = NODE_DATA(node); set_node_memory_tier() local
536 pg_data_t *pgdat; clear_node_memory_tier() local
[all...]
H A Dmemory_hotplug.c499 static void update_pgdat_span(struct pglist_data *pgdat) in update_pgdat_span() argument
504 for (zone = pgdat->node_zones; in update_pgdat_span()
505 zone < pgdat->node_zones + MAX_NR_ZONES; zone++) { in update_pgdat_span()
523 pgdat->node_start_pfn = node_start_pfn; in update_pgdat_span()
524 pgdat->node_spanned_pages = node_end_pfn - node_start_pfn; in update_pgdat_span()
532 struct pglist_data *pgdat = zone->zone_pgdat; in remove_pfn_range_from_zone() local
557 update_pgdat_span(pgdat); in remove_pfn_range_from_zone()
716 static void __meminit resize_pgdat_range(struct pglist_data *pgdat, unsigned long start_pfn, in resize_pgdat_range() argument
719 unsigned long old_end_pfn = pgdat_end_pfn(pgdat); in resize_pgdat_range()
721 if (!pgdat in resize_pgdat_range()
754 struct pglist_data *pgdat = zone->zone_pgdat; move_pfn_range_to_zone() local
849 pg_data_t *pgdat = NODE_DATA(nid); auto_movable_can_online_movable() local
900 struct pglist_data *pgdat = NODE_DATA(nid); default_kernel_zone_for_pfn() local
1239 struct pglist_data *pgdat; hotadd_init_pgdat() local
1274 pg_data_t *pgdat; __try_online_node() local
1887 struct pglist_data *pgdat = zone->zone_pgdat; node_states_check_changes_offline() local
[all...]
H A Dbootmem_info.c102 void __init register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument
105 int node = pgdat->node_id; in register_page_bootmem_info_node()
109 page = virt_to_page(pgdat); in register_page_bootmem_info_node()
114 pfn = pgdat->node_start_pfn; in register_page_bootmem_info_node()
115 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node()
H A Dshuffle.c153 void __meminit __shuffle_free_memory(pg_data_t *pgdat) in __shuffle_free_memory() argument
157 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in __shuffle_free_memory()
H A Dsparse.c321 static inline phys_addr_t pgdat_to_phys(struct pglist_data *pgdat) in pgdat_to_phys() argument
324 VM_BUG_ON(pgdat != &contig_page_data); in pgdat_to_phys()
327 return __pa(pgdat); in pgdat_to_phys()
332 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, in sparse_early_usemaps_alloc_pgdat_section() argument
342 * a pgdat can prevent a section being removed. If section A in sparse_early_usemaps_alloc_pgdat_section()
343 * contains a pgdat and section B contains the usemap, both in sparse_early_usemaps_alloc_pgdat_section()
345 * from the same section as the pgdat where possible to avoid in sparse_early_usemaps_alloc_pgdat_section()
348 goal = pgdat_to_phys(pgdat) & (PAGE_SECTION_MASK << PAGE_SHIFT); in sparse_early_usemaps_alloc_pgdat_section()
366 struct pglist_data *pgdat = NODE_DATA(nid); in check_usemap_section_nr() local
376 pgdat_snr = pfn_to_section_nr(pgdat_to_phys(pgdat) >> PAGE_SHIF in check_usemap_section_nr()
404 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data * pgdat,unsigned long size) sparse_early_usemaps_alloc_pgdat_section() argument
[all...]
H A Dpage_owner.c414 pg_data_t *pgdat, struct zone *zone) in __folio_copy_owner()
490 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in pagetypeinfo_showmixedcount_print()
760 static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) in lseek_page_owner()
827 pgdat->node_id, zone->name, count); in init_pages_in_zone()
830 static void init_zones_in_node(pg_data_t *pgdat) in init_pages_in_zone()
833 struct zone *node_zones = pgdat->node_zones; in init_pages_in_zone()
839 init_pages_in_zone(pgdat, zone); in init_zones_in_node()
845 pg_data_t *pgdat; in init_zones_in_node()
847 for_each_online_pgdat(pgdat)
848 init_zones_in_node(pgdat); in init_early_allocated_pages()
419 pagetypeinfo_showmixedcount_print(struct seq_file * m,pg_data_t * pgdat,struct zone * zone) pagetypeinfo_showmixedcount_print() argument
765 init_pages_in_zone(pg_data_t * pgdat,struct zone * zone) init_pages_in_zone() argument
835 init_zones_in_node(pg_data_t * pgdat) init_zones_in_node() argument
850 pg_data_t *pgdat; init_early_allocated_pages() local
[all...]
/linux/include/linux/
H A Dmemory_hotplug.h30 extern void arch_refresh_nodedata(int nid, pg_data_t *pgdat);
44 memblock_alloc(sizeof(*pgdat), SMP_CACHE_BYTES); \
48 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata() argument
50 node_data[nid] = pgdat; in arch_refresh_nodedata()
61 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata() argument
230 static inline void pgdat_kswapd_lock(pg_data_t *pgdat) in pgdat_kswapd_lock() argument
232 mutex_lock(&pgdat->kswapd_lock); in pgdat_kswapd_lock()
235 static inline void pgdat_kswapd_unlock(pg_data_t *pgdat) in pgdat_kswapd_unlock() argument
237 mutex_unlock(&pgdat->kswapd_lock); in pgdat_kswapd_unlock()
240 static inline void pgdat_kswapd_lock_init(pg_data_t *pgdat) in pgdat_kswapd_lock_init() argument
[all …]
H A Dcompaction.h91 extern void reset_isolation_suitable(pg_data_t *pgdat);
103 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx);
106 static inline void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument
123 static inline void wakeup_kcompactd(pg_data_t *pgdat, in wakeup_kcompactd() argument
H A Dvmstat.h172 static inline void node_page_state_add(long x, struct pglist_data *pgdat, in node_page_state_add() argument
175 atomic_long_add(x, &pgdat->vm_stat[item]); in node_page_state_add()
261 extern unsigned long node_page_state(struct pglist_data *pgdat,
263 extern unsigned long node_page_state_pages(struct pglist_data *pgdat,
311 void set_pgdat_percpu_threshold(pg_data_t *pgdat,
325 static inline void __mod_node_page_state(struct pglist_data *pgdat, in __mod_node_page_state() argument
339 node_page_state_add(delta, pgdat, item); in __mod_node_page_state()
348 static inline void __inc_node_state(struct pglist_data *pgdat, enum node_stat_item item) in __inc_node_state() argument
350 atomic_long_inc(&pgdat->vm_stat[item]); in __inc_node_state()
360 static inline void __dec_node_state(struct pglist_data *pgdat, enu argument
408 set_pgdat_percpu_threshold(pgdat,callback) global() argument
[all...]
H A Dmemcontrol.h59 pg_data_t *pgdat; member
731 * @pgdat: pglist_data
734 * @pgdat combination. This can be the node lruvec, if the memory
738 struct pglist_data *pgdat) in mem_cgroup_lruvec() argument
744 lruvec = &pgdat->__lruvec; in mem_cgroup_lruvec()
751 mz = memcg->nodeinfo[pgdat->node_id]; in mem_cgroup_lruvec()
756 * we have to be prepared to initialize lruvec->pgdat here; in mem_cgroup_lruvec()
759 if (unlikely(lruvec->pgdat != pgdat)) in mem_cgroup_lruvec()
760 lruvec->pgdat in mem_cgroup_lruvec()
1315 mem_cgroup_lruvec(struct mem_cgroup * memcg,struct pglist_data * pgdat) mem_cgroup_lruvec() argument
1322 struct pglist_data *pgdat = folio_pgdat(folio); folio_lruvec() local
1378 struct pglist_data *pgdat = folio_pgdat(folio); folio_lruvec_lock() local
1386 struct pglist_data *pgdat = folio_pgdat(folio); folio_lruvec_lock_irq() local
1395 struct pglist_data *pgdat = folio_pgdat(folio); folio_lruvec_lock_irqsave() local
1627 mem_cgroup_soft_limit_reclaim(pg_data_t * pgdat,int order,gfp_t gfp_mask,unsigned long * total_scanned) mem_cgroup_soft_limit_reclaim() argument
[all...]
H A Dmemory-tiers.h56 void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets); in next_demotion_node()
64 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets()
107 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) in node_get_allowed_targets()
61 node_get_allowed_targets(pg_data_t * pgdat,nodemask_t * targets) node_get_allowed_targets() argument
104 node_get_allowed_targets(pg_data_t * pgdat,nodemask_t * targets) node_get_allowed_targets() argument
H A Dnode.h131 struct pglist_data *pgdat = NODE_DATA(nid); in register_one_node() local
132 unsigned long start_pfn = pgdat->node_start_pfn; in register_one_node()
133 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; in register_one_node()
H A Dmmzone.h558 void lru_gen_init_pgdat(struct pglist_data *pgdat);
571 static inline void lru_gen_init_pgdat(struct pglist_data *pgdat) in lru_gen_init_pgdat() argument
635 struct pglist_data *pgdat; member
1426 static inline unsigned long pgdat_end_pfn(pg_data_t *pgdat) in pgdat_end_pfn() argument
1428 return pgdat->node_start_pfn + pgdat->node_spanned_pages; in pgdat_end_pfn()
1433 void build_all_zonelists(pg_data_t *pgdat);
1461 return lruvec->pgdat; in lruvec_pgdat()
1575 extern struct pglist_data *next_online_pgdat(struct pglist_data *pgdat);
1582 #define for_each_online_pgdat(pgdat) \ argument
1583 for (pgdat = first_online_pgdat(); \
[all …]
H A Dbootmem_info.h21 void __init register_page_bootmem_info_node(struct pglist_data *pgdat);
48 static inline void register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument
H A Dpage_ext.h57 extern void pgdat_page_ext_init(struct pglist_data *pgdat);
104 static inline void pgdat_page_ext_init(struct pglist_data *pgdat)
105 pgdat_page_ext_init(struct pglist_data * pgdat) pgdat_page_ext_init() argument
/linux/tools/testing/memblock/linux/
H A Dmmzone.h8 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat);
10 #define for_each_online_pgdat(pgdat) \ argument
11 for (pgdat = first_online_pgdat(); \
12 pgdat; \
13 pgdat = next_online_pgdat(pgdat))
/linux/drivers/base/
H A Dnode.c375 struct pglist_data *pgdat = NODE_DATA(nid); in node_read_meminfo() local
381 sreclaimable = node_page_state_pages(pgdat, NR_SLAB_RECLAIMABLE_B); in node_read_meminfo()
382 sunreclaimable = node_page_state_pages(pgdat, NR_SLAB_UNRECLAIMABLE_B); in node_read_meminfo()
384 swapcached = node_page_state_pages(pgdat, NR_SWAPCACHE); in node_read_meminfo()
403 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON) + in node_read_meminfo()
404 node_page_state(pgdat, NR_ACTIVE_FILE)), in node_read_meminfo()
405 nid, K(node_page_state(pgdat, NR_INACTIVE_ANON) + in node_read_meminfo()
406 node_page_state(pgdat, NR_INACTIVE_FILE)), in node_read_meminfo()
407 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON)), in node_read_meminfo()
408 nid, K(node_page_state(pgdat, NR_INACTIVE_ANON)), in node_read_meminfo()
[all …]

12