Searched refs:prev_end (Results 1 – 9 of 9) sorted by relevance
| /linux/drivers/gpu/drm/i915/ |
| H A D | i915_scatterlist.c | 86 u64 block_size, offset, prev_end; in i915_rsgt_from_mm_node() local 114 prev_end = (resource_size_t)-1; in i915_rsgt_from_mm_node() 121 if (offset != prev_end || sg->length >= max_segment) { in i915_rsgt_from_mm_node() 140 prev_end = offset; in i915_rsgt_from_mm_node() 176 resource_size_t prev_end; in i915_rsgt_from_buddy_resource() local 200 prev_end = (resource_size_t)-1; in i915_rsgt_from_buddy_resource() 211 if (offset != prev_end || sg->length >= max_segment) { in i915_rsgt_from_buddy_resource() 230 prev_end = offset; in i915_rsgt_from_buddy_resource()
|
| /linux/mm/ |
| H A D | mseal.c | 42 unsigned long prev_end = start; in range_contains_unmapped() local 46 if (vma->vm_start > prev_end) in range_contains_unmapped() 49 prev_end = vma->vm_end; in range_contains_unmapped() 52 return prev_end < end; in range_contains_unmapped()
|
| H A D | numa_memblks.c | 512 u64 prev_end; in numa_fill_memblks() local 542 prev_end = blk[0]->end; in numa_fill_memblks() 546 if (prev_end >= curr->start) { in numa_fill_memblks() 547 if (prev_end < curr->end) in numa_fill_memblks() 548 prev_end = curr->end; in numa_fill_memblks() 550 curr->start = prev_end; in numa_fill_memblks() 551 prev_end = curr->end; in numa_fill_memblks()
|
| H A D | memblock.c | 2150 unsigned long start, end, prev_end = 0; in free_unused_memmap() local 2167 start = min(start, ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap() 2180 if (prev_end && prev_end < start) in free_unused_memmap() 2181 free_memmap(prev_end, start); in free_unused_memmap() 2188 prev_end = pageblock_align(end); in free_unused_memmap() 2192 if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) { in free_unused_memmap() 2193 prev_end = pageblock_align(end); in free_unused_memmap() 2194 free_memmap(prev_end, ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap()
|
| /linux/drivers/parisc/ |
| H A D | iommu-helpers.h | 132 unsigned long prev_end, sg_start; in iommu_coalesce_chunks() local 134 prev_end = (unsigned long)sg_virt(startsg) + in iommu_coalesce_chunks() 159 if (unlikely((prev_end != sg_start) || in iommu_coalesce_chunks() 160 ((prev_end | sg_start) & ~PAGE_MASK))) in iommu_coalesce_chunks()
|
| /linux/drivers/iio/imu/bno055/ |
| H A D | bno055.c | 1451 int xfer_start, start, end, prev_end; in bno055_trigger_handler() local 1490 (prev_end <= BNO055_SCAN_QUATERNION)) ? 3 : 0; in bno055_trigger_handler() 1493 thr_hit = (start - prev_end + quat_extra_len) > in bno055_trigger_handler() 1504 prev_end - xfer_start, in bno055_trigger_handler() 1512 prev_end = end; in bno055_trigger_handler() 1521 prev_end - xfer_start, in bno055_trigger_handler()
|
| /linux/fs/btrfs/ |
| H A D | tree-checker.c | 331 u64 prev_end; in check_extent_data_item() local 335 prev_end = file_extent_end(leaf, prev_key, prev_fi); in check_extent_data_item() 336 if (unlikely(prev_end > key->offset)) { in check_extent_data_item() 339 prev_end, key->offset); in check_extent_data_item() 1612 u64 prev_end = prev_key->objectid; in check_extent_item() local 1615 prev_end += fs_info->nodesize; in check_extent_item() 1617 prev_end += prev_key->offset; in check_extent_item() 1619 if (unlikely(prev_end > key->objectid)) { in check_extent_item()
|
| /linux/fs/nfs/ |
| H A D | pagelist.c | 1049 size_t prev_end = prev->wb_pgbase + prev->wb_bytes; in nfs_page_is_contiguous() local 1054 return prev_end == nfs_page_max_length(prev); in nfs_page_is_contiguous() 1055 if (req->wb_pgbase == prev_end) { in nfs_page_is_contiguous()
|
| /linux/drivers/md/ |
| H A D | raid5.c | 5904 sector_t prev_end = end; in raid5_bitmap_sector() local 5931 prev_end = round_down(prev_end, sectors_per_chunk); in raid5_bitmap_sector() 5934 prev_end = raid5_compute_sector(conf, prev_end, 1, &dd_idx, NULL); in raid5_bitmap_sector() 5942 *sectors = max(end, prev_end) - *offset; in raid5_bitmap_sector()
|