| /linux/drivers/gpu/drm/i915/selftests/ |
| H A D | i915_gem_gtt.c | 245 u64 hole_start, u64 hole_end, in lowlevel_hole() argument 259 for (size = 12; (hole_end - hole_start) >> size; size++) { in lowlevel_hole() 266 hole_size = (hole_end - hole_start) >> aligned_size; in lowlevel_hole() 272 __func__, hole_start, hole_end, size, hole_size); in lowlevel_hole() 288 GEM_BUG_ON(hole_start + count * BIT_ULL(aligned_size) > hole_end); in lowlevel_hole() 312 u64 addr = hole_start + order[n] * BIT_ULL(aligned_size); in lowlevel_hole() 320 hole_end = hole_start; /* quit */ in lowlevel_hole() 371 u64 addr = hole_start + order[n] * BIT_ULL(aligned_size); in lowlevel_hole() 410 u64 hole_start, u64 hole_end, in fill_hole() argument 413 const u64 hole_size = hole_end - hole_start; in fill_hole() [all …]
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_ggtt.c | 976 u64 hole_start, hole_end, hole_size; in xe_ggtt_largest_hole() local 980 drm_mm_for_each_hole(entry, mm, hole_start, hole_end) { in xe_ggtt_largest_hole() 981 hole_start = max(hole_start, ggtt->start); in xe_ggtt_largest_hole() 982 hole_start = ALIGN(hole_start, alignment); in xe_ggtt_largest_hole() 984 if (hole_start >= hole_end) in xe_ggtt_largest_hole() 986 hole_size = hole_end - hole_start; in xe_ggtt_largest_hole() 1153 u64 hole_start, hole_end, hole_size; in xe_ggtt_print_holes() local 1158 drm_mm_for_each_hole(entry, mm, hole_start, hole_end) { in xe_ggtt_print_holes() 1159 hole_start = max(hole_start, ggtt->start); in xe_ggtt_print_holes() 1160 hole_start = ALIGN(hole_start, alignment); in xe_ggtt_print_holes() [all …]
|
| /linux/drivers/gpu/drm/tests/ |
| H A D | drm_mm_test.c | 42 u64 hole_start, __always_unused hole_end; in assert_no_holes() local 46 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) in assert_no_holes() 67 u64 hole_start, hole_end; in assert_one_hole() local 75 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) { in assert_one_hole() 76 if (start != hole_start || end != hole_end) { in assert_one_hole() 80 hole_start, hole_end, start, end); in assert_one_hole()
|
| /linux/fs/btrfs/tests/ |
| H A D | raid-stripe-tree-tests.c | 50 u64 hole_start = logical1 + SZ_256K; in test_punch_hole_3extents() local 133 ret = btrfs_delete_raid_extent(trans, hole_start, hole_len); in test_punch_hole_3extents() 136 hole_start, hole_start + hole_len); in test_punch_hole_3extents() 367 u64 hole_start = logical1 + SZ_32K; in test_punch_hole() local 369 u64 logical2 = hole_start + hole_len; in test_punch_hole() 427 ret = btrfs_delete_raid_extent(trans, hole_start, hole_len); in test_punch_hole() 430 hole_start, hole_start + hole_len); in test_punch_hole() 479 ret = btrfs_get_raid_extent_offset(fs_info, hole_start, &hole_len, in test_punch_hole() 484 hole_start, hole_start + SZ_64K); in test_punch_hole()
|
| /linux/kernel/ |
| H A D | kexec_core.c | 421 unsigned long hole_start, hole_end, size; in kimage_alloc_crash_control_pages() local 426 hole_start = ALIGN(image->control_page, size); in kimage_alloc_crash_control_pages() 427 hole_end = hole_start + size - 1; in kimage_alloc_crash_control_pages() 441 if ((hole_end >= mstart) && (hole_start <= mend)) { in kimage_alloc_crash_control_pages() 443 hole_start = ALIGN(mend, size); in kimage_alloc_crash_control_pages() 444 hole_end = hole_start + size - 1; in kimage_alloc_crash_control_pages() 450 pages = pfn_to_page(hole_start >> PAGE_SHIFT); in kimage_alloc_crash_control_pages()
|
| /linux/fs/hugetlbfs/ |
| H A D | inode.c | 685 loff_t hole_start, hole_end; in hugetlbfs_punch_hole() local 690 hole_start = round_up(offset, hpage_size); in hugetlbfs_punch_hole() 704 if (offset < hole_start) in hugetlbfs_punch_hole() 706 offset, min(offset + len, hole_start)); in hugetlbfs_punch_hole() 709 if (hole_end > hole_start) { in hugetlbfs_punch_hole() 712 hole_start >> PAGE_SHIFT, in hugetlbfs_punch_hole() 717 if ((offset + len) > hole_end && (offset + len) > hole_start) in hugetlbfs_punch_hole() 724 if (hole_end > hole_start) in hugetlbfs_punch_hole() 725 remove_inode_hugepages(inode, hole_start, hole_end); in hugetlbfs_punch_hole()
|
| /linux/fs/ext4/ |
| H A D | extents.c | 4170 ext4_lblk_t hole_start, len; in ext4_ext_determine_insert_hole() local 4173 hole_start = lblk; in ext4_ext_determine_insert_hole() 4174 len = ext4_ext_find_hole(inode, path, &hole_start); in ext4_ext_determine_insert_hole() 4176 ext4_es_find_extent_range(inode, &ext4_es_is_delayed, hole_start, in ext4_ext_determine_insert_hole() 4177 hole_start + len - 1, &es); in ext4_ext_determine_insert_hole() 4190 len -= lblk - hole_start; in ext4_ext_determine_insert_hole() 4191 hole_start = lblk; in ext4_ext_determine_insert_hole() 4211 len = min(es.es_lblk - hole_start, len); in ext4_ext_determine_insert_hole() 4216 ext_debug(inode, " -> %u:%u\n", hole_start, len); in ext4_ext_determine_insert_hole() 4217 ext4_es_cache_extent(inode, hole_start, len, ~0, EXTENT_STATUS_HOLE); in ext4_ext_determine_insert_hole() [all …]
|
| /linux/fs/btrfs/ |
| H A D | volumes.c | 1683 u64 *hole_start, u64 *hole_size, in dev_extent_hole_check_zoned() argument 1691 ASSERT(IS_ALIGNED(*hole_start, zone_size), in dev_extent_hole_check_zoned() 1692 "hole_start=%llu zone_size=%llu", *hole_start, zone_size); in dev_extent_hole_check_zoned() 1695 pos = btrfs_find_allocatable_zones(device, *hole_start, in dev_extent_hole_check_zoned() 1696 *hole_start + *hole_size, in dev_extent_hole_check_zoned() 1698 if (pos != *hole_start) { in dev_extent_hole_check_zoned() 1699 *hole_size = *hole_start + *hole_size - pos; in dev_extent_hole_check_zoned() 1700 *hole_start = pos; in dev_extent_hole_check_zoned() 1714 *hole_start += *hole_size; in dev_extent_hole_check_zoned() 1719 *hole_start += zone_size; in dev_extent_hole_check_zoned() [all …]
|
| H A D | zoned.c | 1052 u64 btrfs_find_allocatable_zones(struct btrfs_device *device, u64 hole_start, in btrfs_find_allocatable_zones() argument 1058 u64 pos = hole_start; in btrfs_find_allocatable_zones() 1063 ASSERT(IS_ALIGNED(hole_start, zinfo->zone_size), in btrfs_find_allocatable_zones() 1064 "hole_start=%llu zinfo->zone_size=%llu", hole_start, zinfo->zone_size); in btrfs_find_allocatable_zones()
|
| H A D | relocation.c | 5921 u64 hole_start, u64 hole_length) in remove_range_from_remap_tree() argument 5932 hole_end = hole_start + hole_length; in remove_range_from_remap_tree() 5983 if (hole_start > remap_start) { in remove_range_from_remap_tree() 5985 hole_start - remap_start, new_addr); in remove_range_from_remap_tree() 6009 max_t(u64, hole_start, remap_start); in remove_range_from_remap_tree() 6021 hole_start - remap_start + new_addr, in remove_range_from_remap_tree()
|
| H A D | inode.c | 5304 u64 hole_start = ALIGN(oldsize, fs_info->sectorsize); in btrfs_cont_expand() local 5320 if (size <= hole_start) in btrfs_cont_expand() 5323 btrfs_lock_and_flush_ordered_range(inode, hole_start, block_end - 1, in btrfs_cont_expand() 5325 cur_offset = hole_start; in btrfs_cont_expand() 5381 btrfs_unlock_extent(io_tree, hole_start, block_end - 1, &cached_state); in btrfs_cont_expand()
|