| /linux/mm/ |
| H A D | vma_exec.c | 36 unsigned long new_start = old_start - shift; in relocate_vma_down() local 38 VMA_ITERATOR(vmi, mm, new_start); in relocate_vma_down() 39 VMG_STATE(vmg, mm, &vmi, new_start, old_end, 0, vma->vm_pgoff); in relocate_vma_down() 42 PAGETABLE_MOVE(pmc, vma, vma, old_start, new_start, length); in relocate_vma_down() 44 BUG_ON(new_start > new_end); in relocate_vma_down() 91 return vma_shrink(&vmi, vma, new_start, new_end, vma->vm_pgoff); in relocate_vma_down()
|
| H A D | readahead.c | 751 * @new_start: The revised start 767 loff_t new_start, size_t new_len) in readahead_expand() 776 new_index = new_start / PAGE_SIZE; in readahead_expand() 809 new_len += new_start - readahead_pos(ractl); in readahead_expand() 764 readahead_expand(struct readahead_control * ractl,loff_t new_start,size_t new_len) readahead_expand() argument
|
| H A D | vma.c | 3048 unsigned long new_start; in acct_stack_growth() local 3063 new_start = (vma->vm_flags & VM_GROWSUP) ? vma->vm_start : in acct_stack_growth() 3065 if (is_hugepage_only_range(vma->vm_mm, new_start, size)) in acct_stack_growth()
|
| /linux/arch/arm/mm/ |
| H A D | cache-l2x0.c | 1610 unsigned long new_start, new_end; in bcm_inv_range() local 1617 new_start = bcm_l2_phys_addr(start); in bcm_inv_range() 1622 l2c210_inv_range(new_start, new_end); in bcm_inv_range() 1629 l2c210_inv_range(new_start, in bcm_inv_range() 1637 unsigned long new_start, new_end; in bcm_clean_range() local 1644 new_start = bcm_l2_phys_addr(start); in bcm_clean_range() 1649 l2c210_clean_range(new_start, new_end); in bcm_clean_range() 1656 l2c210_clean_range(new_start, in bcm_clean_range() 1664 unsigned long new_start, new_end; in bcm_flush_range() local 1676 new_start = bcm_l2_phys_addr(start); in bcm_flush_range() [all …]
|
| /linux/drivers/nvdimm/ |
| H A D | badrange.c | 150 u64 new_start = clr_end + 1; in badrange_forget() local 151 u64 new_len = bre_end - new_start + 1; in badrange_forget() 154 alloc_and_append_badrange_entry(badrange, new_start, in badrange_forget()
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_buddy.c | 1000 u64 new_start; in drm_buddy_block_trim() local 1025 new_start = block_start; in drm_buddy_block_trim() 1027 new_start = *start; in drm_buddy_block_trim() 1029 if (new_start < block_start) in drm_buddy_block_trim() 1032 if (!IS_ALIGNED(new_start, mm->chunk_size)) in drm_buddy_block_trim() 1035 if (range_overflows(new_start, new_size, block_end)) in drm_buddy_block_trim() 1050 err = __alloc_range(mm, &dfs, new_start, new_size, blocks, NULL); in drm_buddy_block_trim()
|
| /linux/drivers/gpu/drm/radeon/ |
| H A D | radeon_ttm.c | 139 uint64_t old_start, new_start; in radeon_move_blit() local 147 new_start = (u64)new_mem->start << PAGE_SHIFT; in radeon_move_blit() 162 new_start += rdev->mc.vram_start; in radeon_move_blit() 165 new_start += rdev->mc.gtt_start; in radeon_move_blit() 179 fence = radeon_copy(rdev, old_start, new_start, num_pages, bo->base.resv); in radeon_move_blit()
|
| /linux/drivers/iommu/iommufd/ |
| H A D | io_pagetable.c | 1292 unsigned long new_start = iova + 1; in iopt_area_split() local 1311 if (new_start & (alignment - 1) || in iopt_area_split() 1312 iopt_area_start_byte(area, new_start) & (alignment - 1)) in iopt_area_split() 1347 (new_start - 1) - start_iova + 1, in iopt_area_split() 1352 rc = iopt_insert_area(iopt, rhs, area->pages, new_start, in iopt_area_split() 1353 iopt_area_start_byte(area, new_start), in iopt_area_split() 1354 last_iova - new_start + 1, area->iommu_prot); in iopt_area_split()
|
| /linux/drivers/md/dm-vdo/indexer/ |
| H A D | delta-index.c | 158 u64 new_start; in rebalance_delta_zone() local 163 new_start = delta_zone->new_offsets[first]; in rebalance_delta_zone() 164 if (delta_list->start != new_start) { in rebalance_delta_zone() 169 delta_list->start = new_start; in rebalance_delta_zone() 184 new_start = delta_zone->new_offsets[middle]; in rebalance_delta_zone() 190 if (new_start > delta_list->start) { in rebalance_delta_zone()
|
| /linux/fs/ext4/ |
| H A D | mballoc.c | 4332 ext4_mb_pa_rb_next_iter(ext4_lblk_t new_start, ext4_lblk_t cur_start, struct rb_node *node) in ext4_mb_pa_rb_next_iter() argument 4334 if (new_start < cur_start) in ext4_mb_pa_rb_next_iter() 4385 ext4_lblk_t new_start, tmp_pa_start, right_pa_start = -1; in ext4_mb_pa_adjust_overlap() local 4388 new_start = *start; in ext4_mb_pa_adjust_overlap() 4496 if (left_pa_end > new_start) in ext4_mb_pa_adjust_overlap() 4497 new_start = left_pa_end; in ext4_mb_pa_adjust_overlap() 4507 ext4_mb_pa_assert_overlap(ac, new_start, new_end); in ext4_mb_pa_adjust_overlap() 4509 *start = new_start; in ext4_mb_pa_adjust_overlap() 5232 ext4_lblk_t iter_start, new_start; in ext4_mb_pa_rb_insert() local 5240 new_start = new_pa->pa_lstart; in ext4_mb_pa_rb_insert() [all …]
|
| /linux/fs/orangefs/ |
| H A D | inode.c | 224 loff_t new_start = readahead_pos(rac); in orangefs_readahead() local 237 readahead_expand(rac, new_start, new_len); in orangefs_readahead()
|
| /linux/drivers/gpu/drm/i915/gt/uc/ |
| H A D | intel_guc_submission.c | 1182 __extend_last_switch(struct intel_guc *guc, u64 *prev_start, u32 new_start) in __extend_last_switch() argument 1187 if (new_start == lower_32_bits(*prev_start)) in __extend_last_switch() 1203 if (new_start < gt_stamp_last && in __extend_last_switch() 1204 (new_start - gt_stamp_last) <= POLL_TIME_CLKS) in __extend_last_switch() 1207 if (new_start > gt_stamp_last && in __extend_last_switch() 1208 (gt_stamp_last - new_start) <= POLL_TIME_CLKS && gt_stamp_hi) in __extend_last_switch() 1211 *prev_start = ((u64)gt_stamp_hi << 32) | new_start; in __extend_last_switch()
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_svm.c | 956 uint64_t new_start, uint64_t new_n, uint64_t *new_vram_pages) in svm_range_split_array() argument 967 d = (new_start - old_start) * size; in svm_range_split_array() 972 d = (new_start == old_start) ? new_n * size : 0; in svm_range_split_array() 1169 svm_range_split_head(struct svm_range *prange, uint64_t new_start, in svm_range_split_head() argument 1178 r = svm_range_split(prange, new_start, prange->last, &head); in svm_range_split_head()
|
| /linux/include/linux/ |
| H A D | pagemap.h | 1387 loff_t new_start, size_t new_len);
|
| /linux/drivers/net/ethernet/netronome/nfp/bpf/ |
| H A D | jit.c | 4330 s16 new_start = range_start; in nfp_bpf_opt_pkt_cache() local 4336 new_start = off; in nfp_bpf_opt_pkt_cache() 4349 if (new_end - new_start <= 64) { in nfp_bpf_opt_pkt_cache() 4351 range_start = new_start; in nfp_bpf_opt_pkt_cache()
|
| /linux/fs/ocfs2/ |
| H A D | alloc.c | 5835 unsigned int new_start) in ocfs2_truncate_log_can_coalesce() argument 5848 return current_tail == new_start; in ocfs2_truncate_log_can_coalesce()
|