Searched refs:hole_start (Results 1 – 6 of 6) sorted by relevance
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_ggtt.c | 895 u64 hole_start, hole_end, hole_size; in xe_ggtt_largest_hole() local 900 drm_mm_for_each_hole(entry, mm, hole_start, hole_end) { in xe_ggtt_largest_hole() 901 hole_start = max(hole_start, hole_min_start); in xe_ggtt_largest_hole() 902 hole_start = ALIGN(hole_start, alignment); in xe_ggtt_largest_hole() 904 if (hole_start >= hole_end) in xe_ggtt_largest_hole() 906 hole_size = hole_end - hole_start; in xe_ggtt_largest_hole() 1074 u64 hole_start, hole_end, hole_size; in xe_ggtt_print_holes() local 1080 drm_mm_for_each_hole(entry, mm, hole_start, hole_end) { in xe_ggtt_print_holes() 1081 hole_start = max(hole_start, hole_min_start); in xe_ggtt_print_holes() 1082 hole_start = ALIGN(hole_start, alignment); in xe_ggtt_print_holes() [all …]
|
| /linux/fs/btrfs/tests/ |
| H A D | raid-stripe-tree-tests.c | 50 u64 hole_start = logical1 + SZ_256K; in test_punch_hole_3extents() local 133 ret = btrfs_delete_raid_extent(trans, hole_start, hole_len); in test_punch_hole_3extents() 136 hole_start, hole_start + hole_len); in test_punch_hole_3extents() 367 u64 hole_start = logical1 + SZ_32K; in test_punch_hole() local 369 u64 logical2 = hole_start + hole_len; in test_punch_hole() 427 ret = btrfs_delete_raid_extent(trans, hole_start, hole_len); in test_punch_hole() 430 hole_start, hole_start + hole_len); in test_punch_hole() 479 ret = btrfs_get_raid_extent_offset(fs_info, hole_start, &hole_len, in test_punch_hole() 484 hole_start, hole_start + SZ_64K); in test_punch_hole()
|
| /linux/kernel/ |
| H A D | kexec_core.c | 421 unsigned long hole_start, hole_end, size; in kimage_alloc_crash_control_pages() local 426 hole_start = ALIGN(image->control_page, size); in kimage_alloc_crash_control_pages() 427 hole_end = hole_start + size - 1; in kimage_alloc_crash_control_pages() 441 if ((hole_end >= mstart) && (hole_start <= mend)) { in kimage_alloc_crash_control_pages() 443 hole_start = ALIGN(mend, size); in kimage_alloc_crash_control_pages() 444 hole_end = hole_start + size - 1; in kimage_alloc_crash_control_pages() 450 pages = pfn_to_page(hole_start >> PAGE_SHIFT); in kimage_alloc_crash_control_pages()
|
| /linux/fs/hugetlbfs/ |
| H A D | inode.c | 685 loff_t hole_start, hole_end; in hugetlbfs_punch_hole() local 690 hole_start = round_up(offset, hpage_size); in hugetlbfs_punch_hole() 704 if (offset < hole_start) in hugetlbfs_punch_hole() 706 offset, min(offset + len, hole_start)); in hugetlbfs_punch_hole() 709 if (hole_end > hole_start) { in hugetlbfs_punch_hole() 712 hole_start >> PAGE_SHIFT, in hugetlbfs_punch_hole() 717 if ((offset + len) > hole_end && (offset + len) > hole_start) in hugetlbfs_punch_hole() 724 if (hole_end > hole_start) in hugetlbfs_punch_hole() 725 remove_inode_hugepages(inode, hole_start, hole_end); in hugetlbfs_punch_hole()
|
| /linux/fs/btrfs/ |
| H A D | zoned.c | 1050 u64 btrfs_find_allocatable_zones(struct btrfs_device *device, u64 hole_start, in btrfs_find_allocatable_zones() argument 1056 u64 pos = hole_start; in btrfs_find_allocatable_zones() 1061 ASSERT(IS_ALIGNED(hole_start, zinfo->zone_size), in btrfs_find_allocatable_zones() 1062 "hole_start=%llu zinfo->zone_size=%llu", hole_start, zinfo->zone_size); in btrfs_find_allocatable_zones()
|
| H A D | inode.c | 5199 u64 hole_start = ALIGN(oldsize, fs_info->sectorsize); in btrfs_cont_expand() local 5215 if (size <= hole_start) in btrfs_cont_expand() 5218 btrfs_lock_and_flush_ordered_range(inode, hole_start, block_end - 1, in btrfs_cont_expand() 5220 cur_offset = hole_start; in btrfs_cont_expand() 5276 btrfs_unlock_extent(io_tree, hole_start, block_end - 1, &cached_state); in btrfs_cont_expand()
|