| /linux/drivers/gpu/drm/i915/ |
| H A D | i915_scatterlist.c | 86 u64 block_size, offset, prev_end; in i915_rsgt_from_mm_node() local 114 prev_end = (resource_size_t)-1; in i915_rsgt_from_mm_node() 121 if (offset != prev_end || sg->length >= max_segment) { in i915_rsgt_from_mm_node() 140 prev_end = offset; in i915_rsgt_from_mm_node() 176 resource_size_t prev_end; in i915_rsgt_from_buddy_resource() local 200 prev_end = (resource_size_t)-1; in i915_rsgt_from_buddy_resource() 211 if (offset != prev_end || sg->length >= max_segment) { in i915_rsgt_from_buddy_resource() 230 prev_end = offset; in i915_rsgt_from_buddy_resource()
|
| /linux/mm/ |
| H A D | mseal.c | 42 unsigned long prev_end = start; in range_contains_unmapped() local 46 if (vma->vm_start > prev_end) in range_contains_unmapped() 49 prev_end = vma->vm_end; in range_contains_unmapped() 52 return prev_end < end; in range_contains_unmapped()
|
| H A D | numa_memblks.c | 514 u64 prev_end; in numa_fill_memblks() local 544 prev_end = blk[0]->end; in numa_fill_memblks() 548 if (prev_end >= curr->start) { in numa_fill_memblks() 549 if (prev_end < curr->end) in numa_fill_memblks() 550 prev_end = curr->end; in numa_fill_memblks() 552 curr->start = prev_end; in numa_fill_memblks() 553 prev_end = curr->end; in numa_fill_memblks()
|
| H A D | memblock.c | 2151 unsigned long start, end, prev_end = 0; in free_unused_memmap() 2168 start = min(start, ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap() 2181 if (prev_end && prev_end < start) in free_unused_memmap() 2182 free_memmap(prev_end, start); in free_unused_memmap() 2189 prev_end = pageblock_align(end); in free_unused_memmap() 2193 if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) { in free_unused_memmap() 2194 prev_end = pageblock_align(end); in free_unused_memmap() 2195 free_memmap(prev_end, ALIGN(prev_end, PAGES_PER_SECTIO in free_unused_memmap() 2150 unsigned long start, end, prev_end = 0; free_unused_memmap() local [all...] |
| /linux/drivers/parisc/ |
| H A D | iommu-helpers.h | 132 unsigned long prev_end, sg_start; in iommu_coalesce_chunks() local 134 prev_end = (unsigned long)sg_virt(startsg) + in iommu_coalesce_chunks() 159 if (unlikely((prev_end != sg_start) || in iommu_coalesce_chunks() 160 ((prev_end | sg_start) & ~PAGE_MASK))) in iommu_coalesce_chunks()
|
| /linux/arch/x86/virt/vmx/tdx/ |
| H A D | tdx.c | 736 u64 prev_end; in tdmr_populate_rsvd_holes() local 743 prev_end = tdmr->base; in tdmr_populate_rsvd_holes() 762 if (start <= prev_end) { in tdmr_populate_rsvd_holes() 763 prev_end = end; in tdmr_populate_rsvd_holes() 768 ret = tdmr_add_rsvd_area(tdmr, rsvd_idx, prev_end, in tdmr_populate_rsvd_holes() 769 start - prev_end, in tdmr_populate_rsvd_holes() 774 prev_end = end; in tdmr_populate_rsvd_holes() 778 if (prev_end < tdmr_end(tdmr)) { in tdmr_populate_rsvd_holes() 779 ret = tdmr_add_rsvd_area(tdmr, rsvd_idx, prev_end, in tdmr_populate_rsvd_holes() 780 tdmr_end(tdmr) - prev_end, in tdmr_populate_rsvd_holes()
|
| /linux/drivers/iio/imu/bno055/ |
| H A D | bno055.c | 1451 int xfer_start, start, end, prev_end; in bno055_trigger_handler() local 1490 (prev_end <= BNO055_SCAN_QUATERNION)) ? 3 : 0; in bno055_trigger_handler() 1493 thr_hit = (start - prev_end + quat_extra_len) > in bno055_trigger_handler() 1504 prev_end - xfer_start, in bno055_trigger_handler() 1512 prev_end = end; in bno055_trigger_handler() 1521 prev_end - xfer_start, in bno055_trigger_handler()
|
| /linux/fs/nfs/ |
| H A D | pagelist.c | 1050 size_t prev_end = prev->wb_pgbase + prev->wb_bytes; in nfs_page_is_contiguous() local 1055 return prev_end == nfs_page_max_length(prev); in nfs_page_is_contiguous() 1056 if (req->wb_pgbase == prev_end) { in nfs_page_is_contiguous()
|
| /linux/drivers/accel/habanalabs/common/ |
| H A D | memory.c | 601 u64 tmp_hint_addr, valid_start, valid_size, prev_start, prev_end, in get_va_block() local 710 prev_end = reserved_valid_start - 1; in get_va_block() 727 rc = add_va_block_locked(hdev, &va_range->list, prev_start, prev_end); in get_va_block()
|
| /linux/drivers/md/ |
| H A D | raid5.c | 5908 sector_t prev_end = end; in raid5_bitmap_sector() local 5935 prev_end = round_down(prev_end, sectors_per_chunk); in raid5_bitmap_sector() 5938 prev_end = raid5_compute_sector(conf, prev_end, 1, &dd_idx, NULL); in raid5_bitmap_sector() 5946 *sectors = max(end, prev_end) - *offset; in raid5_bitmap_sector()
|