Lines Matching refs:new_rgn

1246 	struct memblock_region *new_rgn = &memblock.reserved.regions[0];  in alloc_nid_top_down_numa_simple_check()  local
1267 ASSERT_EQ(new_rgn->size, size); in alloc_nid_top_down_numa_simple_check()
1268 ASSERT_EQ(new_rgn->base, region_end(req_node) - size); in alloc_nid_top_down_numa_simple_check()
1269 ASSERT_LE(req_node->base, new_rgn->base); in alloc_nid_top_down_numa_simple_check()
1298 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_top_down_numa_small_node_check() local
1319 ASSERT_EQ(new_rgn->size, size); in alloc_nid_top_down_numa_small_node_check()
1320 ASSERT_EQ(new_rgn->base, region_end(exp_node) - size); in alloc_nid_top_down_numa_small_node_check()
1321 ASSERT_LE(exp_node->base, new_rgn->base); in alloc_nid_top_down_numa_small_node_check()
1351 struct memblock_region *new_rgn = &memblock.reserved.regions[1]; in alloc_nid_top_down_numa_node_reserved_check() local
1373 ASSERT_EQ(new_rgn->size, size); in alloc_nid_top_down_numa_node_reserved_check()
1374 ASSERT_EQ(new_rgn->base, region_end(exp_node) - size); in alloc_nid_top_down_numa_node_reserved_check()
1375 ASSERT_LE(exp_node->base, new_rgn->base); in alloc_nid_top_down_numa_node_reserved_check()
1403 struct memblock_region *new_rgn = &memblock.reserved.regions[1]; in alloc_nid_top_down_numa_part_reserved_check() local
1428 ASSERT_EQ(new_rgn->size, size); in alloc_nid_top_down_numa_part_reserved_check()
1429 ASSERT_EQ(new_rgn->base, region_end(req_node) - size); in alloc_nid_top_down_numa_part_reserved_check()
1430 ASSERT_LE(req_node->base, new_rgn->base); in alloc_nid_top_down_numa_part_reserved_check()
1462 struct memblock_region *new_rgn = &memblock.reserved.regions[1]; in alloc_nid_top_down_numa_part_reserved_fallback_check() local
1489 ASSERT_EQ(new_rgn->size, size); in alloc_nid_top_down_numa_part_reserved_fallback_check()
1490 ASSERT_EQ(new_rgn->base, region_end(exp_node) - size); in alloc_nid_top_down_numa_part_reserved_fallback_check()
1491 ASSERT_LE(exp_node->base, new_rgn->base); in alloc_nid_top_down_numa_part_reserved_fallback_check()
1524 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_top_down_numa_split_range_low_check() local
1545 ASSERT_EQ(new_rgn->size, size); in alloc_nid_top_down_numa_split_range_low_check()
1546 ASSERT_EQ(new_rgn->base, req_node_end - size); in alloc_nid_top_down_numa_split_range_low_check()
1547 ASSERT_LE(req_node->base, new_rgn->base); in alloc_nid_top_down_numa_split_range_low_check()
1581 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_top_down_numa_split_range_high_check() local
1602 ASSERT_EQ(new_rgn->size, size); in alloc_nid_top_down_numa_split_range_high_check()
1603 ASSERT_EQ(new_rgn->base, exp_node_end - size); in alloc_nid_top_down_numa_split_range_high_check()
1604 ASSERT_LE(exp_node->base, new_rgn->base); in alloc_nid_top_down_numa_split_range_high_check()
1637 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_top_down_numa_no_overlap_split_check() local
1658 ASSERT_EQ(new_rgn->size, size); in alloc_nid_top_down_numa_no_overlap_split_check()
1659 ASSERT_EQ(new_rgn->base, region_end(req_node) - size); in alloc_nid_top_down_numa_no_overlap_split_check()
1660 ASSERT_LE(req_node->base, new_rgn->base); in alloc_nid_top_down_numa_no_overlap_split_check()
1694 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_top_down_numa_no_overlap_low_check() local
1714 ASSERT_EQ(new_rgn->size, size); in alloc_nid_top_down_numa_no_overlap_low_check()
1715 ASSERT_EQ(new_rgn->base, max_addr - size); in alloc_nid_top_down_numa_no_overlap_low_check()
1716 ASSERT_LE(max_node->base, new_rgn->base); in alloc_nid_top_down_numa_no_overlap_low_check()
1750 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_top_down_numa_no_overlap_high_check() local
1770 ASSERT_EQ(new_rgn->size, size); in alloc_nid_top_down_numa_no_overlap_high_check()
1771 ASSERT_EQ(new_rgn->base, max_addr - size); in alloc_nid_top_down_numa_no_overlap_high_check()
1772 ASSERT_LE(max_node->base, new_rgn->base); in alloc_nid_top_down_numa_no_overlap_high_check()
1790 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_bottom_up_numa_simple_check() local
1811 ASSERT_EQ(new_rgn->size, size); in alloc_nid_bottom_up_numa_simple_check()
1812 ASSERT_EQ(new_rgn->base, req_node->base); in alloc_nid_bottom_up_numa_simple_check()
1813 ASSERT_LE(region_end(new_rgn), region_end(req_node)); in alloc_nid_bottom_up_numa_simple_check()
1842 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_bottom_up_numa_small_node_check() local
1863 ASSERT_EQ(new_rgn->size, size); in alloc_nid_bottom_up_numa_small_node_check()
1864 ASSERT_EQ(new_rgn->base, exp_node->base); in alloc_nid_bottom_up_numa_small_node_check()
1865 ASSERT_LE(region_end(new_rgn), region_end(exp_node)); in alloc_nid_bottom_up_numa_small_node_check()
1896 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_bottom_up_numa_node_reserved_check() local
1918 ASSERT_EQ(new_rgn->size, size); in alloc_nid_bottom_up_numa_node_reserved_check()
1919 ASSERT_EQ(new_rgn->base, exp_node->base); in alloc_nid_bottom_up_numa_node_reserved_check()
1920 ASSERT_LE(region_end(new_rgn), region_end(exp_node)); in alloc_nid_bottom_up_numa_node_reserved_check()
1948 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_bottom_up_numa_part_reserved_check() local
1975 ASSERT_EQ(new_rgn->size, total_size); in alloc_nid_bottom_up_numa_part_reserved_check()
1976 ASSERT_EQ(new_rgn->base, req_node->base); in alloc_nid_bottom_up_numa_part_reserved_check()
1977 ASSERT_LE(region_end(new_rgn), region_end(req_node)); in alloc_nid_bottom_up_numa_part_reserved_check()
2009 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_bottom_up_numa_part_reserved_fallback_check() local
2036 ASSERT_EQ(new_rgn->size, size); in alloc_nid_bottom_up_numa_part_reserved_fallback_check()
2037 ASSERT_EQ(new_rgn->base, exp_node->base); in alloc_nid_bottom_up_numa_part_reserved_fallback_check()
2038 ASSERT_LE(region_end(new_rgn), region_end(exp_node)); in alloc_nid_bottom_up_numa_part_reserved_fallback_check()
2071 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_bottom_up_numa_split_range_low_check() local
2092 ASSERT_EQ(new_rgn->size, size); in alloc_nid_bottom_up_numa_split_range_low_check()
2093 ASSERT_EQ(new_rgn->base, req_node->base); in alloc_nid_bottom_up_numa_split_range_low_check()
2094 ASSERT_LE(region_end(new_rgn), req_node_end); in alloc_nid_bottom_up_numa_split_range_low_check()
2128 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_bottom_up_numa_split_range_high_check() local
2150 ASSERT_EQ(new_rgn->size, size); in alloc_nid_bottom_up_numa_split_range_high_check()
2151 ASSERT_EQ(new_rgn->base, exp_node->base); in alloc_nid_bottom_up_numa_split_range_high_check()
2152 ASSERT_LE(region_end(new_rgn), exp_node_end); in alloc_nid_bottom_up_numa_split_range_high_check()
2185 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_bottom_up_numa_no_overlap_split_check() local
2206 ASSERT_EQ(new_rgn->size, size); in alloc_nid_bottom_up_numa_no_overlap_split_check()
2207 ASSERT_EQ(new_rgn->base, req_node->base); in alloc_nid_bottom_up_numa_no_overlap_split_check()
2208 ASSERT_LE(region_end(new_rgn), region_end(req_node)); in alloc_nid_bottom_up_numa_no_overlap_split_check()
2242 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_bottom_up_numa_no_overlap_low_check() local
2262 ASSERT_EQ(new_rgn->size, size); in alloc_nid_bottom_up_numa_no_overlap_low_check()
2263 ASSERT_EQ(new_rgn->base, min_addr); in alloc_nid_bottom_up_numa_no_overlap_low_check()
2264 ASSERT_LE(region_end(new_rgn), region_end(min_node)); in alloc_nid_bottom_up_numa_no_overlap_low_check()
2298 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_bottom_up_numa_no_overlap_high_check() local
2318 ASSERT_EQ(new_rgn->size, size); in alloc_nid_bottom_up_numa_no_overlap_high_check()
2319 ASSERT_EQ(new_rgn->base, min_addr); in alloc_nid_bottom_up_numa_no_overlap_high_check()
2320 ASSERT_LE(region_end(new_rgn), region_end(min_node)); in alloc_nid_bottom_up_numa_no_overlap_high_check()
2392 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; in alloc_nid_numa_reserved_full_merge_generic_check() local
2424 ASSERT_EQ(new_rgn->size, total_size); in alloc_nid_numa_reserved_full_merge_generic_check()
2425 ASSERT_EQ(new_rgn->base, r2.base); in alloc_nid_numa_reserved_full_merge_generic_check()
2427 ASSERT_LE(new_rgn->base, req_node->base); in alloc_nid_numa_reserved_full_merge_generic_check()
2428 ASSERT_LE(region_end(req_node), region_end(new_rgn)); in alloc_nid_numa_reserved_full_merge_generic_check()