Home
last modified time | relevance | path

Searched refs:zone_managed_pages (Results 1 – 5 of 5) sorted by relevance

/linux/mm/
H A Dshow_mem.c99 managed_pages += zone_managed_pages(zone); in si_meminfo_node()
101 managed_highpages += zone_managed_pages(zone); in si_meminfo_node()
166 if (zone_managed_pages(pgdat->node_zones + zone_idx)) in node_has_managed_zones()
343 K(zone_managed_pages(zone)), in show_free_areas()
413 reserved += zone->present_pages - zone_managed_pages(zone); in __show_mem()
H A Dpage_alloc.c2173 if ((pageblock_nr_pages * 4) > zone_managed_pages(zone)) in boost_watermark()
3409 if ((zone_managed_pages(zone) / 100) < pageblock_nr_pages) in reserve_highatomic_pageblock()
3411 max_managed = ALIGN((zone_managed_pages(zone) / 100), pageblock_nr_pages); in reserve_highatomic_pageblock()
5478 unsigned long size = zone_managed_pages(zone); in nr_free_zone_pages()
5895 batch = min(zone_managed_pages(zone) >> 12, SZ_256K / PAGE_SIZE); in zone_batchsize()
5953 total_pages = zone_managed_pages(zone) / high_fraction; in zone_highsize()
6312 unsigned long managed_pages = zone_managed_pages(zone); in calculate_totalreserve_pages()
6372 bool clear = !ratio || !zone_managed_pages(zone); in setup_per_zone_lowmem_reserve()
6378 managed_pages += zone_managed_pages(upper_zone); in setup_per_zone_lowmem_reserve()
6404 lowmem_pages += zone_managed_pages(zone); in __setup_per_zone_wmarks()
[all …]
H A Dhighmem.c134 pages += zone_managed_pages(zone); in __totalhigh_pages()
H A Dvmstat.c260 mem = zone_managed_pages(zone) >> (27 - PAGE_SHIFT); in calculate_normal_threshold()
1801 zone_managed_pages(zone), in zoneinfo_show_print()
/linux/include/linux/
H A Dmmzone.h1103 static inline unsigned long zone_managed_pages(const struct zone *zone) in zone_managed_pages() function
1599 return zone_managed_pages(zone); in managed_zone()