| /linux/mm/ |
| H A D | mm_init.c | 317 unsigned long start_pfn, end_pfn; in early_calculate_totalpages() local 320 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in early_calculate_totalpages() 321 unsigned long pages = end_pfn - start_pfn; in early_calculate_totalpages() 479 unsigned long start_pfn, end_pfn; in find_zone_movable_pfns_for_nodes() local 497 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in find_zone_movable_pfns_for_nodes() 501 if (start_pfn >= end_pfn) in find_zone_movable_pfns_for_nodes() 507 kernel_pages = min(end_pfn, usable_startpfn) in find_zone_movable_pfns_for_nodes() 516 if (end_pfn <= usable_startpfn) { in find_zone_movable_pfns_for_nodes() 524 zone_movable_pfn[nid] = end_pfn; in find_zone_movable_pfns_for_nodes() 535 size_pages = end_pfn - start_pfn; in find_zone_movable_pfns_for_nodes() [all …]
|
| H A D | page_idle.c | 124 unsigned long pfn, end_pfn; in page_idle_bitmap_read() local 134 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_read() 135 if (end_pfn > max_pfn) in page_idle_bitmap_read() 136 end_pfn = max_pfn; in page_idle_bitmap_read() 138 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_read() 169 unsigned long pfn, end_pfn; in page_idle_bitmap_write() local 179 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_write() 180 if (end_pfn > max_pfn) in page_idle_bitmap_write() 181 end_pfn = max_pfn; in page_idle_bitmap_write() 183 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_write()
|
| H A D | memory_hotplug.c | 392 const unsigned long end_pfn = pfn + nr_pages; in __add_pages() local 419 for (; pfn < end_pfn; pfn += cur_nr_pages) { in __add_pages() 421 cur_nr_pages = min(end_pfn - pfn, in __add_pages() 433 /* find the smallest valid pfn in the range [start_pfn, end_pfn) */ 436 unsigned long end_pfn) in find_smallest_section_pfn() argument 438 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { in find_smallest_section_pfn() 454 /* find the biggest valid pfn in the range [start_pfn, end_pfn). */ 457 unsigned long end_pfn) in find_biggest_section_pfn() argument 462 pfn = end_pfn - 1; in find_biggest_section_pfn() 480 unsigned long end_pfn) in shrink_zone_span() argument 526 unsigned long end_pfn = zone_end_pfn(zone); update_pgdat_span() local 551 const unsigned long end_pfn = start_pfn + nr_pages; remove_pfn_range_from_zone() local 596 const unsigned long end_pfn = pfn + nr_pages; __remove_pages() local 660 const unsigned long end_pfn = start_pfn + nr_pages; online_pages_range() local 969 unsigned long online_pages = 0, max_pages, end_pfn; auto_movable_zone_for_pfn() local 1093 unsigned long end_pfn = pfn + nr_pages; mhp_init_memmap_on_memory() local 1123 unsigned long end_pfn = pfn + nr_pages; mhp_deinit_memmap_on_memory() local 1785 do_migrate_range(unsigned long start_pfn,unsigned long end_pfn) do_migrate_range() argument 1898 const unsigned long end_pfn = start_pfn + nr_pages; offline_pages() local [all...] |
| H A D | compaction.c | 558 unsigned long end_pfn, in isolate_freepages_block() argument 577 for (; blockpfn < end_pfn; blockpfn += stride, page += stride) { in isolate_freepages_block() 602 (blockpfn + (1UL << order) <= end_pfn)) { in isolate_freepages_block() 657 if (unlikely(blockpfn > end_pfn)) in isolate_freepages_block() 658 blockpfn = end_pfn; in isolate_freepages_block() 671 if (strict && blockpfn < end_pfn) in isolate_freepages_block() 696 unsigned long start_pfn, unsigned long end_pfn) in isolate_freepages_range() argument 710 for (; pfn < end_pfn; pfn += isolated, in isolate_freepages_range() 726 block_end_pfn = min(block_end_pfn, end_pfn); in isolate_freepages_range() 750 if (pfn < end_pfn) { in isolate_freepages_range() [all …]
|
| H A D | shuffle.c | 84 unsigned long end_pfn = zone_end_pfn(z); in __shuffle_zone() local 90 for (i = start_pfn; i < end_pfn; i += order_pages) { in __shuffle_zone()
|
| /linux/include/trace/events/ |
| H A D | page_isolation.h | 14 unsigned long end_pfn, 17 TP_ARGS(start_pfn, end_pfn, fin_pfn), 21 __field(unsigned long, end_pfn) 27 __entry->end_pfn = end_pfn; 32 __entry->start_pfn, __entry->end_pfn, __entry->fin_pfn, 33 __entry->end_pfn <= __entry->fin_pfn ? "success" : "fail")
|
| H A D | compaction.h | 18 unsigned long end_pfn, 22 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken), 26 __field(unsigned long, end_pfn) 33 __entry->end_pfn = end_pfn; 40 __entry->end_pfn, 49 unsigned long end_pfn, 53 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 60 unsigned long end_pfn, 64 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 71 unsigned long end_pfn, [all …]
|
| /linux/arch/sh/mm/ |
| H A D | numa.c | 25 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local 31 end_pfn = PFN_DOWN(end); in setup_bootmem_node() 38 __add_active_range(nid, start_pfn, end_pfn); in setup_bootmem_node() 49 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node()
|
| /linux/drivers/base/ |
| H A D | arch_numa.c | 195 static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) in setup_node_data() argument 197 if (start_pfn >= end_pfn) in setup_node_data() 204 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_node_data() 217 unsigned long start_pfn, end_pfn; in numa_register_nodes() local 219 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in numa_register_nodes() 220 setup_node_data(nid, start_pfn, end_pfn); in numa_register_nodes()
|
| /linux/arch/x86/platform/efi/ |
| H A D | efi_32.c | 38 u64 start_pfn, end_pfn, end; in efi_map_region() local 45 end_pfn = PFN_UP(end); in efi_map_region() 47 if (pfn_range_is_mapped(start_pfn, end_pfn)) { in efi_map_region()
|
| /linux/arch/sh/kernel/ |
| H A D | swsusp.c | 22 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave() local 24 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
|
| H A D | setup.c | 194 unsigned long end_pfn) in __add_active_range() argument 202 end = end_pfn << PAGE_SHIFT; in __add_active_range() 211 start_pfn, end_pfn); in __add_active_range() 235 memblock_set_node(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn), in __add_active_range()
|
| /linux/security/integrity/ima/ |
| H A D | ima_kexec.c | 310 unsigned long start_pfn, end_pfn; in ima_validate_range() local 317 end_pfn = PHYS_PFN(end_phys); in ima_validate_range() 320 if (!pfn_range_is_mapped(start_pfn, end_pfn)) in ima_validate_range() 322 if (!page_is_ram(start_pfn) || !page_is_ram(end_pfn)) in ima_validate_range()
|
| /linux/drivers/hv/ |
| H A D | hv_balloon.c | 429 unsigned long end_pfn; member 439 unsigned long end_pfn; member 593 if (pfn >= gap->start_pfn && pfn < gap->end_pfn) in has_pfn_is_backed() 615 (pfn < has->end_pfn) && in hv_page_offline_check() 773 (pfn + (1UL << order) > has->end_pfn)) in hv_online_page() 796 if (start_pfn < has->start_pfn || start_pfn >= has->end_pfn) in pfn_covered() 812 gap->end_pfn = start_pfn; in pfn_covered() 822 if ((start_pfn + pfn_cnt) > has->end_pfn) { in pfn_covered() 824 residual = (start_pfn + pfn_cnt - has->end_pfn); in pfn_covered() 825 has->end_pfn += ALIGN(residual, ha_pages_in_chunk); in pfn_covered() [all …]
|
| /linux/kernel/liveupdate/ |
| H A D | kexec_handover.c | 165 unsigned long end_pfn) in __kho_unpreserve() argument 169 while (pfn < end_pfn) { in __kho_unpreserve() 170 order = min(count_trailing_zeros(pfn), ilog2(end_pfn - pfn)); in __kho_unpreserve() 307 const unsigned long end_pfn = start_pfn + nr_pages; in kho_restore_pages() local 310 while (pfn < end_pfn) { in kho_restore_pages() 312 min(count_trailing_zeros(pfn), ilog2(end_pfn - pfn)); in kho_restore_pages() 861 const unsigned long end_pfn = start_pfn + nr_pages; in kho_preserve_pages() local 871 while (pfn < end_pfn) { in kho_preserve_pages() 873 min(count_trailing_zeros(pfn), ilog2(end_pfn - pfn)); in kho_preserve_pages() 905 const unsigned long end_pfn = start_pfn + nr_pages; in kho_unpreserve_pages() local [all …]
|
| /linux/arch/x86/virt/vmx/tdx/ |
| H A D | tdx.c | 193 unsigned long end_pfn, int nid) in add_tdx_memblock() argument 203 tmb->end_pfn = end_pfn; in add_tdx_memblock() 231 unsigned long start_pfn, end_pfn; in build_tdx_memlist() local 234 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in build_tdx_memlist() 242 if (start_pfn >= end_pfn) in build_tdx_memlist() 250 ret = add_tdx_memblock(tmb_list, start_pfn, end_pfn, nid); in build_tdx_memlist() 392 end = TDMR_ALIGN_UP(PFN_PHYS(tmb->end_pfn)); in fill_out_tdmrs() 491 if (tmb->end_pfn > PHYS_PFN(tdmr->base)) in tdmr_get_nid() 748 end = PFN_PHYS(tmb->end_pfn); in tdmr_populate_rsvd_holes() 1347 static bool is_tdx_memory(unsigned long start_pfn, unsigned long end_pfn) in is_tdx_memory() argument [all …]
|
| /linux/arch/loongarch/kernel/ |
| H A D | numa.c | 145 unsigned long start_pfn, end_pfn; in node_mem_init() local 152 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 154 node, start_pfn, end_pfn); in node_mem_init()
|
| /linux/arch/powerpc/mm/ |
| H A D | mem.c | 123 unsigned long end_pfn = PFN_UP(start + size); in update_end_of_memory_vars() local 125 if (end_pfn > max_pfn) { in update_end_of_memory_vars() 126 max_pfn = end_pfn; in update_end_of_memory_vars() 127 max_low_pfn = end_pfn; in update_end_of_memory_vars()
|
| /linux/include/asm-generic/ |
| H A D | memory_model.h | 35 #define for_each_valid_pfn(pfn, start_pfn, end_pfn) \ argument 37 (pfn) < min_t(unsigned long, (end_pfn), \
|
| /linux/kernel/power/ |
| H A D | snapshot.c | 397 unsigned long end_pfn; /* Zone end page frame + 1 */ member 556 zone->end_pfn = end; in create_zone_bm_rtree() 757 if (pfn >= zone->start_pfn && pfn < zone->end_pfn) in memory_bm_find_bit() 764 if (pfn >= curr->start_pfn && pfn < curr->end_pfn) { in memory_bm_find_bit() 935 pages = bm->cur.zone->end_pfn - bm->cur.zone->start_pfn; in memory_bm_next_pfn() 958 unsigned long end_pfn; member 997 void __init register_nosave_region(unsigned long start_pfn, unsigned long end_pfn) in register_nosave_region() argument 1001 if (start_pfn >= end_pfn) in register_nosave_region() 1008 if (region->end_pfn == start_pfn) { in register_nosave_region() 1009 region->end_pfn in register_nosave_region() [all...] |
| /linux/arch/x86/mm/ |
| H A D | numa.c | 135 unsigned long start_pfn, end_pfn; in numa_register_nodes() local 141 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in numa_register_nodes() 142 if (start_pfn >= end_pfn) in numa_register_nodes()
|
| /linux/include/linux/ |
| H A D | node.h | 123 unsigned long end_pfn); 127 unsigned long end_pfn) in register_memory_blocks_under_node_hotplug() argument
|
| H A D | mmzone.h | 2108 void online_mem_sections(unsigned long start_pfn, unsigned long end_pfn); 2109 void offline_mem_sections(unsigned long start_pfn, unsigned long end_pfn); 2213 static inline unsigned long first_valid_pfn(unsigned long pfn, unsigned long end_pfn) in first_valid_pfn() argument 2219 while (nr <= __highest_present_section_nr && pfn < end_pfn) { in first_valid_pfn() 2234 return end_pfn; in first_valid_pfn() 2237 static inline unsigned long next_valid_pfn(unsigned long pfn, unsigned long end_pfn) in next_valid_pfn() argument 2241 if (pfn >= end_pfn) in next_valid_pfn() 2242 return end_pfn; in next_valid_pfn() 2254 return first_valid_pfn(pfn, end_pfn); in next_valid_pfn()
|
| /linux/arch/x86/include/asm/ |
| H A D | mtrr.h | 59 extern int mtrr_trim_uncached_memory(unsigned long end_pfn); 101 static inline int mtrr_trim_uncached_memory(unsigned long end_pfn) in mtrr_trim_uncached_memory() argument
|
| /linux/fs/proc/ |
| H A D | kcore.c | 253 unsigned long end_pfn; in kcore_ram_list() local 257 end_pfn = 0; in kcore_ram_list() 261 if (end_pfn < node_end) in kcore_ram_list() 262 end_pfn = node_end; in kcore_ram_list() 265 ret = walk_system_ram_range(0, end_pfn, list, kclist_add_private); in kcore_ram_list()
|