| /linux/drivers/gpu/drm/i915/selftests/ |
| H A D | i915_gem_gtt.c | 245 u64 hole_start, u64 hole_end, in lowlevel_hole() argument 259 for (size = 12; (hole_end - hole_start) >> size; size++) { in lowlevel_hole() 266 hole_size = (hole_end - hole_start) >> aligned_size; in lowlevel_hole() 272 __func__, hole_start, hole_end, size, hole_size); in lowlevel_hole() 288 GEM_BUG_ON(hole_start + count * BIT_ULL(aligned_size) > hole_end); in lowlevel_hole() 312 u64 addr = hole_start + order[n] * BIT_ULL(aligned_size); in lowlevel_hole() 320 hole_end = hole_start; /* quit */ in lowlevel_hole() 371 u64 addr = hole_start + order[n] * BIT_ULL(aligned_size); in lowlevel_hole() 410 u64 hole_start, u64 hole_end, in fill_hole() argument 413 const u64 hole_size = hole_end - hole_start; in fill_hole() [all …]
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_ggtt.c | 921 u64 hole_start, hole_end, hole_size; in xe_ggtt_largest_hole() local 925 drm_mm_for_each_hole(entry, mm, hole_start, hole_end) { in xe_ggtt_largest_hole() 926 hole_start = max(hole_start, ggtt->start); in xe_ggtt_largest_hole() 927 hole_start = ALIGN(hole_start, alignment); in xe_ggtt_largest_hole() 929 if (hole_start >= hole_end) in xe_ggtt_largest_hole() 931 hole_size = hole_end - hole_start; in xe_ggtt_largest_hole() 1095 u64 hole_start, hole_end, hole_size; in xe_ggtt_print_holes() local 1100 drm_mm_for_each_hole(entry, mm, hole_start, hole_end) { in xe_ggtt_print_holes() 1101 hole_start = max(hole_start, ggtt->start); in xe_ggtt_print_holes() 1102 hole_start = ALIGN(hole_start, alignment); in xe_ggtt_print_holes() [all …]
|
| /linux/fs/btrfs/tests/ |
| H A D | raid-stripe-tree-tests.c | 50 u64 hole_start = logical1 + SZ_256K; in test_punch_hole_3extents() local 133 ret = btrfs_delete_raid_extent(trans, hole_start, hole_len); in test_punch_hole_3extents() 136 hole_start, hole_start + hole_len); in test_punch_hole_3extents() 367 u64 hole_start = logical1 + SZ_32K; in test_punch_hole() local 369 u64 logical2 = hole_start + hole_len; in test_punch_hole() 427 ret = btrfs_delete_raid_extent(trans, hole_start, hole_len); in test_punch_hole() 430 hole_start, hole_start + hole_len); in test_punch_hole() 479 ret = btrfs_get_raid_extent_offset(fs_info, hole_start, &hole_len, in test_punch_hole() 484 hole_start, hole_start + SZ_64K); in test_punch_hole()
|
| /linux/fs/ext4/ |
| H A D | extents.c | 4195 ext4_lblk_t hole_start, len; in ext4_ext_determine_insert_hole() local 4198 hole_start = lblk; in ext4_ext_determine_insert_hole() 4199 len = ext4_ext_find_hole(inode, path, &hole_start); in ext4_ext_determine_insert_hole() 4201 ext4_es_find_extent_range(inode, &ext4_es_is_delayed, hole_start, in ext4_ext_determine_insert_hole() 4202 hole_start + len - 1, &es); in ext4_ext_determine_insert_hole() 4215 len -= lblk - hole_start; in ext4_ext_determine_insert_hole() 4216 hole_start = lblk; in ext4_ext_determine_insert_hole() 4236 len = min(es.es_lblk - hole_start, len); in ext4_ext_determine_insert_hole() 4241 ext_debug(inode, " -> %u:%u\n", hole_start, len); in ext4_ext_determine_insert_hole() 4242 ext4_es_cache_extent(inode, hole_start, len, ~0, EXTENT_STATUS_HOLE); in ext4_ext_determine_insert_hole() [all …]
|
| /linux/fs/btrfs/ |
| H A D | volumes.c | 1684 u64 *hole_start, u64 *hole_size, in dev_extent_hole_check_zoned() argument 1692 ASSERT(IS_ALIGNED(*hole_start, zone_size), in dev_extent_hole_check_zoned() 1693 "hole_start=%llu zone_size=%llu", *hole_start, zone_size); in dev_extent_hole_check_zoned() 1696 pos = btrfs_find_allocatable_zones(device, *hole_start, in dev_extent_hole_check_zoned() 1697 *hole_start + *hole_size, in dev_extent_hole_check_zoned() 1699 if (pos != *hole_start) { in dev_extent_hole_check_zoned() 1700 *hole_size = *hole_start + *hole_size - pos; in dev_extent_hole_check_zoned() 1701 *hole_start = pos; in dev_extent_hole_check_zoned() 1715 *hole_start += *hole_size; in dev_extent_hole_check_zoned() 1720 *hole_start += zone_size; in dev_extent_hole_check_zoned() [all …]
|
| H A D | zoned.c | 1052 u64 btrfs_find_allocatable_zones(struct btrfs_device *device, u64 hole_start, in btrfs_find_allocatable_zones() argument 1058 u64 pos = hole_start; in btrfs_find_allocatable_zones() 1063 ASSERT(IS_ALIGNED(hole_start, zinfo->zone_size), in btrfs_find_allocatable_zones() 1064 "hole_start=%llu zinfo->zone_size=%llu", hole_start, zinfo->zone_size); in btrfs_find_allocatable_zones()
|
| H A D | relocation.c | 5931 u64 hole_start, u64 hole_length) in remove_range_from_remap_tree() argument 5942 hole_end = hole_start + hole_length; in remove_range_from_remap_tree() 5993 if (hole_start > remap_start) { in remove_range_from_remap_tree() 5995 hole_start - remap_start, new_addr); in remove_range_from_remap_tree() 6019 max_t(u64, hole_start, remap_start); in remove_range_from_remap_tree() 6031 hole_start - remap_start + new_addr, in remove_range_from_remap_tree()
|
| H A D | inode.c | 5304 u64 hole_start = ALIGN(oldsize, fs_info->sectorsize); in btrfs_cont_expand() local 5320 if (size <= hole_start) in btrfs_cont_expand() 5323 btrfs_lock_and_flush_ordered_range(inode, hole_start, block_end - 1, in btrfs_cont_expand() 5325 cur_offset = hole_start; in btrfs_cont_expand() 5381 btrfs_unlock_extent(io_tree, hole_start, block_end - 1, &cached_state); in btrfs_cont_expand()
|