Lines Matching refs:end_pfn
300 unsigned long start_pfn, end_pfn;
303 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) {
304 unsigned long pages = end_pfn - start_pfn;
462 unsigned long start_pfn, end_pfn;
480 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) {
484 if (start_pfn >= end_pfn)
490 kernel_pages = min(end_pfn, usable_startpfn)
499 if (end_pfn <= usable_startpfn) {
507 zone_movable_pfn[nid] = end_pfn;
515 * start_pfn->end_pfn. Calculate size_pages as the
518 size_pages = end_pfn - start_pfn;
549 unsigned long start_pfn, end_pfn;
554 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);
555 if (zone_movable_pfn[nid] >= end_pfn)
602 unsigned long start_pfn, end_pfn;
608 nid = memblock_search_pfn_nid(pfn, &start_pfn, &end_pfn);
611 state->last_end = end_pfn;
672 defer_init(int nid, unsigned long pfn, unsigned long end_pfn)
680 if (end_pfn < pgdat_end_pfn(NODE_DATA(nid)))
690 if (prev_end_pfn != end_pfn) {
691 prev_end_pfn = end_pfn;
737 static inline bool defer_init(int nid, unsigned long pfn, unsigned long end_pfn)
757 unsigned long end_pfn = PFN_UP(end);
759 for (; start_pfn < end_pfn; start_pfn++) {
856 unsigned long pfn, end_pfn = start_pfn + size;
859 if (highest_memmap_pfn < end_pfn - 1)
860 highest_memmap_pfn = end_pfn - 1;
876 end_pfn = altmap->base_pfn + vmem_altmap_offset(altmap);
880 for (pfn = start_pfn; pfn < end_pfn; ) {
920 unsigned long end_pfn,
928 end_pfn = clamp(end_pfn, zone_start_pfn, zone_end_pfn);
930 if (start_pfn >= end_pfn)
933 memmap_init_range(end_pfn - start_pfn, nid, zone_id, start_pfn,
939 *hole_pfn = end_pfn;
944 unsigned long start_pfn, end_pfn;
948 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) {
957 memmap_init_zone_range(zone, start_pfn, end_pfn,
973 end_pfn = round_up(end_pfn, PAGES_PER_SECTION);
974 if (hole_pfn < end_pfn)
976 init_unavailable_range(hole_pfn, end_pfn, zone_id, nid);
1051 unsigned long pfn, end_pfn = head_pfn + nr_pages;
1055 for (pfn = head_pfn + 1; pfn < end_pfn; pfn++) {
1077 unsigned long pfn, end_pfn = start_pfn + nr_pages;
1095 nr_pages = end_pfn - start_pfn;
1098 for (pfn = start_pfn; pfn < end_pfn; pfn += pfns_per_compound) {
1160 unsigned long start_pfn, end_pfn;
1163 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) {
1165 end_pfn = clamp(end_pfn, range_start_pfn, range_end_pfn);
1166 nr_absent -= end_pfn - start_pfn;
1174 * @end_pfn: The end PFN to stop searching for holes
1179 unsigned long end_pfn)
1181 return __absent_pages_in_range(MAX_NUMNODES, start_pfn, end_pfn);
1204 unsigned long start_pfn, end_pfn;
1210 end_pfn = clamp(memblock_region_memory_end_pfn(r),
1215 nr_absent += end_pfn - start_pfn;
1219 nr_absent += end_pfn - start_pfn;
1278 unsigned long start_pfn, end_pfn;
1287 end_pfn = PFN_DOWN(end_addr);
1289 if (start_pfn < end_pfn) {
1290 nr_all_pages += end_pfn - start_pfn;
1293 end_pfn = clamp(end_pfn, 0, high_zone_low);
1295 nr_kernel_pages += end_pfn - start_pfn;
1649 * @end_pfn: Passed by reference. On return, it will have the node end_pfn.
1656 unsigned long *start_pfn, unsigned long *end_pfn)
1662 *end_pfn = 0;
1666 *end_pfn = max(*end_pfn, this_end_pfn);
1677 unsigned long end_pfn = 0;
1682 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);
1688 if (start_pfn != end_pfn) {
1691 end_pfn ? ((u64)end_pfn << PAGE_SHIFT) - 1 : 0);
1693 calculate_node_totalpages(pgdat, start_pfn, end_pfn);
1761 unsigned long start_pfn, end_pfn;
1783 end_pfn = max(max_zone_pfn[zone], start_pfn);
1785 arch_zone_highest_possible_pfn[zone] = end_pfn;
1787 start_pfn = end_pfn;
1825 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) {
1828 ((u64)end_pfn << PAGE_SHIFT) - 1);
1829 subsection_map_init(start_pfn, end_pfn - start_pfn);
1963 unsigned long pfn, unsigned long end_pfn)
1966 unsigned long nr_pages = end_pfn - pfn;
1970 for (; pfn < end_pfn; pfn++, page++)
2022 unsigned long *end_pfn)
2025 unsigned long spfn = *start_pfn, epfn = *end_pfn;
2030 for_each_free_mem_pfn_range_in_zone_from(j, zone, start_pfn, end_pfn) {
2036 t = min(mo_pfn, *end_pfn);
2039 if (mo_pfn < *end_pfn) {
2065 deferred_init_memmap_chunk(unsigned long start_pfn, unsigned long end_pfn,
2078 while (spfn < end_pfn) {