/linux/mm/ |
H A D | nommu.c | 103 return vma->vm_end - vma->vm_start; in kobjsize() 422 BUG_ON(last->vm_end <= last->vm_start); in validate_nommu_regions() 423 BUG_ON(last->vm_top < last->vm_end); in validate_nommu_regions() 429 BUG_ON(region->vm_end <= region->vm_start); in validate_nommu_regions() 430 BUG_ON(region->vm_top < region->vm_end); in validate_nommu_regions() 575 vma_iter_config(&vmi, vma->vm_start, vma->vm_end); in delete_vma_from_mm() 670 if (vma->vm_end != end) in find_vma_exact() 889 vma->vm_region->vm_top = vma->vm_region->vm_end; in do_mmap_shared_file() 925 vma->vm_region->vm_top = vma->vm_region->vm_end; in do_mmap_private() 958 region->vm_end = region->vm_start + len; in do_mmap_private() [all …]
|
H A D | vma.c | 244 uprobe_munmap(vp->vma, vp->vma->vm_start, vp->vma->vm_end); in vma_prepare() 248 vp->adj_next->vm_end); in vma_prepare() 333 vp->remove->vm_end); in vma_complete() 341 WARN_ON_ONCE(vp->vma->vm_end < vp->remove->vm_end); in vma_complete() 375 return vmg->prev && vmg->prev->vm_end == vmg->start && in can_vma_merge_left() 435 unmap_vmas(&tlb, mas, vma, vma->vm_start, vma->vm_end, vma->vm_end, in unmap_region() 437 mas_set(mas, vma->vm_end); in unmap_region() 438 free_pgtables(&tlb, mas, vma, prev ? prev->vm_end : FIRST_USER_ADDRESS, in unmap_region() 458 WARN_ON(vma->vm_end <= addr); in __split_vma() 471 new->vm_end = addr; in __split_vma() [all …]
|
H A D | msync.c | 90 fend = fstart + (min(end, vma->vm_end) - start) - 1; in SYSCALL_DEFINE3() 91 start = vma->vm_end; in SYSCALL_DEFINE3() 107 vma = find_vma(mm, vma->vm_end); in SYSCALL_DEFINE3()
|
H A D | mremap.c | 728 if (!err && vma->vm_end != old_addr + old_len) in move_vma() 794 if (vma->vm_end > old_addr + old_len) in move_vma() 795 account_end = vma->vm_end; in move_vma() 823 vma->vm_end == (old_addr + old_len)) in move_vma() 895 if (old_len > vma->vm_end - addr) in resize_is_valid() 1026 unsigned long end = vma->vm_end + delta; in vma_expandable() 1028 if (end < vma->vm_end) /* overflow */ in vma_expandable() 1030 if (find_vma_intersection(vma->vm_mm, vma->vm_end, end)) in vma_expandable() 1172 if (old_len == vma->vm_end - addr) { in SYSCALL_DEFINE5() 1178 VMA_ITERATOR(vmi, mm, vma->vm_end); in SYSCALL_DEFINE5()
|
H A D | mmap.c | 979 populate_vma_page_range(prev, addr, prev->vm_end, NULL); in find_extend_vma_locked() 1173 if (start + size > vma->vm_end) { in SYSCALL_DEFINE5() 1174 VMA_ITERATOR(vmi, mm, vma->vm_end); in SYSCALL_DEFINE5() 1179 if (next->vm_start != prev->vm_end) in SYSCALL_DEFINE5() 1188 if (start + size <= next->vm_end) in SYSCALL_DEFINE5() 1294 vma_iter_set(&vmi, vma->vm_end); in exit_mmap() 1304 vma_iter_set(&vmi, vma->vm_end); in exit_mmap() 1332 if (find_vma_intersection(mm, vma->vm_start, vma->vm_end)) in insert_vm_struct() 1688 unsigned long old_end = vma->vm_end; in relocate_vma_down()
|
H A D | internal.h | 898 if (end > vma->vm_end) in folio_within_range() 899 end = vma->vm_end; in folio_within_range() 915 return folio_within_range(folio, vma, vma->vm_start, vma->vm_end); in folio_within_vma() 985 if (address < vma->vm_start || address >= vma->vm_end) in vma_address() 1013 if (address < vma->vm_start || address > vma->vm_end) in vma_address_end() 1014 address = vma->vm_end; in vma_address_end() 1418 vma->vm_end = end; in vma_set_range()
|
H A D | userfaultfd.c | 27 if (dst_end > dst_vma->vm_end) in validate_dst_vma() 911 VM_WARN_ONCE(start < dst_vma->vm_start || start + len > dst_vma->vm_end, in uffd_wp_range() 981 _end = min(dst_vma->vm_end, end); in mwriteprotect_range() 1510 if (src_start >= vma->vm_start && src_start < vma->vm_end) in find_vmas_mm_locked() 1540 if (src_start >= vma->vm_start && src_start < vma->vm_end) { in uffd_move_lock() 1727 if (src_start + len > src_vma->vm_end) in move_pages() 1732 if (dst_start + len > dst_vma->vm_end) in move_pages() 1954 vma_end = min(end, vma->vm_end); in userfaultfd_register_range() 1975 start = vma->vm_end; in userfaultfd_register_range() 2025 vma->vm_start, vma->vm_end); in userfaultfd_release_all()
|
H A D | madvise.c | 809 if (range.start >= vma->vm_end) in madvise_free_single_vma() 811 range.end = min(vma->vm_end, end_addr); in madvise_free_single_vma() 921 if (end > vma->vm_end) { in madvise_dontneed_free() 934 end = vma->vm_end; in madvise_dontneed_free() 1506 tmp = vma->vm_end; in madvise_walk_vmas() 1515 if (prev && start < prev->vm_end) in madvise_walk_vmas() 1516 start = prev->vm_end; in madvise_walk_vmas() 1520 vma = find_vma(mm, prev->vm_end); in madvise_walk_vmas()
|
/linux/tools/testing/selftests/bpf/progs/ |
H A D | iters_task_vma.c | 14 __u64 vm_end; member 35 vm_ranges[seen].vm_end = vma->vm_end; in iter_task_vma_for_each()
|
/linux/fs/proc/ |
H A D | task_nommu.c | 35 size += region->vm_end - region->vm_start; in task_mem() 37 size = vma->vm_end - vma->vm_start; in task_mem() 46 slack = region->vm_end - vma->vm_end; in task_mem() 89 vsize += vma->vm_end - vma->vm_start; in task_vsize() 109 size += region->vm_end - region->vm_start; in task_statm() 150 vma->vm_end, in nommu_vma_show()
|
/linux/include/trace/events/ |
H A D | mmap.h | 80 __field(unsigned long, vm_end) 87 __entry->vm_end = vma->vm_end - 1; 93 (unsigned long) __entry->vm_end
|
H A D | fs_dax.h | 17 __field(unsigned long, vm_end) 30 __entry->vm_end = vmf->vma->vm_end; 39 "%#lx vm_end %#lx pgoff %#lx max_pgoff %#lx %s", 47 __entry->vm_end,
|
/linux/scripts/coccinelle/api/ |
H A D | vma_pages.cocci | 22 * (vma->vm_end - vma->vm_start) >> PAGE_SHIFT 32 - ((vma->vm_end - vma->vm_start) >> PAGE_SHIFT) 44 (vma->vm_end@p - vma->vm_start) >> PAGE_SHIFT
|
/linux/arch/x86/um/ |
H A D | mem_32.c | 18 gate_vma.vm_end = FIXADDR_USER_END; in gate_vma_init() 49 return (addr >= vma->vm_start) && (addr < vma->vm_end); in in_gate_area()
|
/linux/drivers/media/common/videobuf2/ |
H A D | videobuf2-memops.c | 96 vma->vm_end); in vb2_common_vm_open() 114 vma->vm_end); in vb2_common_vm_close()
|
/linux/drivers/dax/ |
H A D | device.c | 35 if (vma->vm_start & mask || vma->vm_end & mask) { in check_vma() 38 current->comm, func, vma->vm_start, vma->vm_end, in check_vma() 162 (pmd_addr + PMD_SIZE) > vmf->vma->vm_end) in __dev_dax_pmd_fault() 207 (pud_addr + PUD_SIZE) > vmf->vma->vm_end) in __dev_dax_pud_fault()
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | file_ops.c | 315 vma->vm_end - vma->vm_start, vma->vm_flags); in mmap_cdbg() 415 if ((vma->vm_end - vma->vm_start) != memlen) { in hfi1_file_mmap() 417 (vma->vm_end - vma->vm_start), memlen); in hfi1_file_mmap() 434 vm_end_save = vma->vm_end; in hfi1_file_mmap() 435 vma->vm_end = vma->vm_start; in hfi1_file_mmap() 440 vma->vm_end += memlen; in hfi1_file_mmap() 447 vma->vm_end = vm_end_save; in hfi1_file_mmap() 453 vma->vm_end = vm_end_save; in hfi1_file_mmap() 553 if ((vma->vm_end - vma->vm_start) != memlen) { in hfi1_file_mmap() 556 (vma->vm_end - vma->vm_start), memlen); in hfi1_file_mmap()
|
/linux/drivers/sbus/char/ |
H A D | flash.c | 66 if (vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)) > size) in flash_mmap() 67 size = vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)); in flash_mmap()
|
/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_mman.c | 37 (vma->vm_end - vma->vm_start) == PAGE_ALIGN(size); in __vma_matches() 280 area->vm_start, area->vm_end - area->vm_start, in vm_fault_cpu() 303 unsigned long vm_start, vm_end, vma_size; /* user's memory parameters */ in set_address_limits() local 311 vm_end = area->vm_end >> PAGE_SHIFT; in set_address_limits() 325 end = min_t(long, end, vm_end); in set_address_limits()
|
/linux/kernel/bpf/ |
H A D | task_iter.c | 484 info->prev_vm_end = curr_vma->vm_end; in task_vma_seq_get_next() 530 curr_vma = find_vma(curr_mm, curr_vma->vm_end); in task_vma_seq_get_next() 543 curr_vma->vm_end == info->prev_vm_end) in task_vma_seq_get_next() 544 curr_vma = find_vma(curr_mm, curr_vma->vm_end); in task_vma_seq_get_next() 644 info->prev_vm_end = info->vma->vm_end; in task_vma_seq_stop() 777 if (vma && vma->vm_start <= start && vma->vm_end > start) { in BPF_CALL_5()
|
/linux/include/linux/ |
H A D | mm.h | 949 vma->vm_end > vma->vm_mm->start_brk; in vma_is_initial_heap() 964 vma->vm_end >= vma->vm_mm->start_stack; in vma_is_initial_stack() 1056 vmi->mas.last = vma->vm_end - 1; in vma_iter_bulk_store() 2386 vma->vm_end - vma->vm_start, NULL); in zap_vma_pages() 3480 unsigned long vm_end = vma->vm_end; in vm_end_gap() local 3483 vm_end += stack_guard_gap; in vm_end_gap() 3484 if (vm_end < vma->vm_end) in vm_end_gap() 3485 vm_end = -PAGE_SIZE; in vm_end_gap() 3487 return vm_end; in vm_end_gap() 3492 return (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; in vma_pages() [all …]
|
/linux/arch/nios2/kernel/ |
H A D | sys_nios2.c | 46 if (vma == NULL || addr < vma->vm_start || addr + len > vma->vm_end) { in sys_cacheflush()
|
/linux/drivers/infiniband/core/ |
H A D | ib_core_uverbs.c | 77 if (vma->vm_end - vma->vm_start != size) in rdma_user_mmap_io() 171 if (entry->npages * PAGE_SIZE != vma->vm_end - vma->vm_start) { in rdma_user_mmap_entry_get()
|
/linux/arch/x86/entry/vsyscall/ |
H A D | vsyscall_64.c | 287 .vm_end = VSYSCALL_ADDR + PAGE_SIZE, 311 return (addr >= vma->vm_start) && (addr < vma->vm_end); in in_gate_area()
|
/linux/drivers/infiniband/hw/qib/ |
H A D | qib_file_ops.c | 715 if ((vma->vm_end - vma->vm_start) > len) { in qib_mmap_mem() 718 vma->vm_end - vma->vm_start, len); in qib_mmap_mem() 763 if ((vma->vm_end - vma->vm_start) > sz) { in mmap_ureg() 766 vma->vm_end - vma->vm_start); in mmap_ureg() 775 vma->vm_end - vma->vm_start, in mmap_ureg() 795 if ((vma->vm_end - vma->vm_start) > (piocnt * dd->palign)) { in mmap_piobufs() 798 vma->vm_end - vma->vm_start); in mmap_piobufs() 820 vma->vm_end - vma->vm_start, in mmap_piobufs() 837 if ((vma->vm_end - vma->vm_start) > total_size) { in mmap_rcvegrbufs() 840 vma->vm_end - vma->vm_start, in mmap_rcvegrbufs() [all …]
|