| /linux/mm/ |
| H A D | nommu.c | 96 return vma->vm_end - vma->vm_start; in kobjsize() 448 BUG_ON(last->vm_end <= last->vm_start); in validate_nommu_regions() 455 BUG_ON(region->vm_end <= region->vm_start); in validate_nommu_regions() 457 BUG_ON(region->vm_start < last->vm_top); in validate_nommu_regions() 483 if (region->vm_start < pregion->vm_start) in add_nommu_region() 485 else if (region->vm_start > pregion->vm_start) in add_nommu_region() 536 if (region->vm_top > region->vm_start) in __put_nommu_region() 546 free_page_series(region->vm_start, region->vm_top); in __put_nommu_region() 599 VMA_ITERATOR(vmi, vma->vm_mm, vma->vm_start); in delete_vma_from_mm() 601 vma_iter_config(&vmi, vma->vm_start, vma->vm_end); in delete_vma_from_mm() [all …]
|
| H A D | vma.h | 213 .pg_end = _next ? ((struct vm_area_struct *)_next)->vm_start : \ 218 ((struct vm_area_struct *)_next)->vm_start : \ 233 return vma->vm_pgoff + PHYS_PFN(addr - vma->vm_start); in vma_pgoff_offset() 282 ((vmi->mas.index > vma->vm_start) || (vmi->mas.last < vma->vm_start))) in vma_iter_store_gfp() 285 __mas_set_range(&vmi->mas, vma->vm_start, vma->vm_end - 1); in vma_iter_store_gfp() 589 vmi->mas.index > vma->vm_start)) { in vma_iter_store_overwrite() 591 vmi->mas.index, vma->vm_start, vma->vm_start, in vma_iter_store_overwrite() 595 vmi->mas.last < vma->vm_start)) { in vma_iter_store_overwrite() 597 vmi->mas.last, vma->vm_start, vma->vm_start, vma->vm_end, in vma_iter_store_overwrite() 603 ((vmi->mas.index > vma->vm_start) || (vmi->mas.last < vma->vm_start))) in vma_iter_store_overwrite() [all …]
|
| H A D | vma.c | 291 uprobe_munmap(vp->vma, vp->vma->vm_start, vp->vma->vm_end); in vma_prepare() 294 uprobe_munmap(vp->adj_next, vp->adj_next->vm_start, in vma_prepare() 382 uprobe_munmap(vp->remove, vp->remove->vm_start, in vma_complete() 442 if (!next || vmg->end != next->vm_start || !can_vma_merge_before(vmg)) in can_vma_merge_right() 504 WARN_ON(vma->vm_start >= addr); in __split_vma() 520 new->vm_start = addr; in __split_vma() 521 new->vm_pgoff += ((addr - vma->vm_start) >> PAGE_SHIFT); in __split_vma() 525 vma_iter_config(vmi, new->vm_start, new->vm_end); in __split_vma() 554 vma_adjust_trans_huge(vma, vma->vm_start, addr, NULL); in __split_vma() 559 vma->vm_start = addr; in __split_vma() [all …]
|
| H A D | vma_exec.c | 33 unsigned long old_start = vma->vm_start; in relocate_vma_down() 76 next ? next->vm_start : USER_PGTABLES_CEILING); in relocate_vma_down() 85 next ? next->vm_start : USER_PGTABLES_CEILING); in relocate_vma_down() 140 vma->vm_start = vma->vm_end - PAGE_SIZE; in create_init_stack_vma()
|
| H A D | msync.c | 73 if (start < vma->vm_start) { in SYSCALL_DEFINE3() 76 start = vma->vm_start; in SYSCALL_DEFINE3() 88 fstart = (start - vma->vm_start) + in SYSCALL_DEFINE3()
|
| H A D | mremap.c | 651 if (!pmc->for_stack && vma->vm_start != addr_to_align) in can_align_down() 655 if (pmc->for_stack && addr_masked >= vma->vm_start) in can_align_down() 662 return find_vma_intersection(vma->vm_mm, addr_masked, vma->vm_start) == NULL; in can_align_down() 951 pgoff_t internal_pgoff = (vrm->addr - vma->vm_start) >> PAGE_SHIFT; in vrm_set_new_addr() 1051 if (vma->vm_start != old_addr) in prep_move_vma() 1090 unsigned long vm_start; in unmap_source_vma() local 1122 vm_start = vma->vm_start; in unmap_source_vma() 1163 if (vm_start < addr) { in unmap_source_vma() 1188 unsigned long internal_offset = vrm->addr - vrm->vma->vm_start; in copy_vma_and_data() 1254 unsigned long old_start = vrm->vma->vm_start; in dontunmap_complete() [all …]
|
| H A D | mlock.c | 533 if (start > vma->vm_start) in apply_vma_lock_flags() 537 tmp = vma->vm_start; in apply_vma_lock_flags() 542 if (vma->vm_start != tmp) in apply_vma_lock_flags() 587 if (start > vma->vm_start) in count_mm_mlocked_page_nr() 588 count -= (start - vma->vm_start); in count_mm_mlocked_page_nr() 590 count += end - vma->vm_start; in count_mm_mlocked_page_nr() 593 count += vma->vm_end - vma->vm_start; in count_mm_mlocked_page_nr() 734 error = mlock_fixup(&vmi, vma, &prev, vma->vm_start, vma->vm_end, in apply_mlockall_flags()
|
| /linux/tools/testing/vma/tests/ |
| H A D | merge.c | 118 ASSERT_EQ(vma->vm_start, 0); in test_simple_merge() 150 ASSERT_EQ(vma->vm_start, 0x1000); in test_simple_modify() 162 ASSERT_EQ(vma->vm_start, 0); in test_simple_modify() 171 ASSERT_EQ(vma->vm_start, 0x1000); in test_simple_modify() 180 ASSERT_EQ(vma->vm_start, 0x2000); in test_simple_modify() 208 ASSERT_EQ(vma->vm_start, 0); in test_simple_expand() 229 ASSERT_EQ(vma->vm_start, 0); in test_simple_shrink() 321 ASSERT_EQ(vma->vm_start, 0); in __test_merge_new() 340 ASSERT_EQ(vma->vm_start, 0); in __test_merge_new() 361 ASSERT_EQ(vma->vm_start, 0x6000); in __test_merge_new() [all …]
|
| H A D | mmap.c | 39 if (vma->vm_start == 0x300000) { in test_mmap_region_basic() 42 } else if (vma->vm_start == 0x24d000) { in test_mmap_region_basic()
|
| /linux/mm/damon/tests/ |
| H A D | vaddr-kunit.h | 28 mas_set_range(&mas, vmas[i].vm_start, vmas[i].vm_end - 1); in __link_vmas() 72 (struct vm_area_struct) {.vm_start = 10, .vm_end = 20}, in damon_test_three_regions_in_vmas() 73 (struct vm_area_struct) {.vm_start = 20, .vm_end = 25}, in damon_test_three_regions_in_vmas() 74 (struct vm_area_struct) {.vm_start = 200, .vm_end = 210}, in damon_test_three_regions_in_vmas() 75 (struct vm_area_struct) {.vm_start = 210, .vm_end = 220}, in damon_test_three_regions_in_vmas() 76 (struct vm_area_struct) {.vm_start = 300, .vm_end = 305}, in damon_test_three_regions_in_vmas() 77 (struct vm_area_struct) {.vm_start = 307, .vm_end = 330}, in damon_test_three_regions_in_vmas()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | iters_task_vma.c | 13 __u64 vm_start; member 34 vm_ranges[seen].vm_start = vma->vm_start; in iter_task_vma_for_each()
|
| H A D | find_vma_fail1.c | 5 #define vm_flags vm_start 17 vma->vm_start = 0xffffffffff600000; in write_vma()
|
| /linux/fs/proc/ |
| H A D | task_nommu.c | 35 size += region->vm_end - region->vm_start; in task_mem() 37 size = vma->vm_end - vma->vm_start; in task_mem() 89 vsize += vma->vm_end - vma->vm_start; in task_vsize() 109 size += region->vm_end - region->vm_start; in task_statm() 149 vma->vm_start, in nommu_vma_show() 184 *ppos = vma->vm_start; in proc_get_vma()
|
| /linux/drivers/android/ |
| H A D | binder_alloc.c | 65 return alloc->vm_start + alloc->buffer_size - buffer->user_data; in binder_alloc_buffer_size() 204 index = (page_addr - alloc->vm_start) / PAGE_SIZE; in binder_lru_freelist_add() 341 alloc->pid, addr - alloc->vm_start); in binder_install_single_page() 353 alloc->pid, __func__, addr - alloc->vm_start, ret); in binder_install_single_page() 375 index = (page_addr - alloc->vm_start) / PAGE_SIZE; in binder_install_buffer_pages() 404 index = (page_addr - alloc->vm_start) / PAGE_SIZE; in binder_lru_freelist_del() 757 BUG_ON(buffer->user_data < alloc->vm_start); in binder_free_buf_locked() 758 BUG_ON(buffer->user_data > alloc->vm_start + alloc->buffer_size); in binder_free_buf_locked() 817 (buffer->user_data - alloc->vm_start); in binder_alloc_get_page() 913 alloc->buffer_size = min_t(unsigned long, vma->vm_end - vma->vm_start, in binder_alloc_mmap_handler() [all …]
|
| /linux/scripts/coccinelle/api/ |
| H A D | vma_pages.cocci | 22 * (vma->vm_end - vma->vm_start) >> PAGE_SHIFT 32 - ((vma->vm_end - vma->vm_start) >> PAGE_SHIFT) 44 (vma->vm_end@p - vma->vm_start) >> PAGE_SHIFT
|
| /linux/arch/parisc/mm/ |
| H A D | fault.c | 130 if (tree->vm_start > addr) { 136 if (prev->vm_next->vm_start > addr) 262 vma->vm_start, vma->vm_end); in show_signal_msg() 296 if (!vma || address < vma->vm_start) { in do_page_fault() 377 address < vma->vm_start || address >= vma->vm_end) { in do_page_fault()
|
| /linux/drivers/gpu/drm/i915/gem/ |
| H A D | i915_gem_mman.c | 35 return vma->vm_start == addr && in __vma_matches() 36 (vma->vm_end - vma->vm_start) == PAGE_ALIGN(size); in __vma_matches() 282 area->vm_start, area->vm_end - area->vm_start, in vm_fault_cpu() 305 unsigned long vm_start, vm_end, vma_size; /* user's memory parameters */ in set_address_limits() local 312 vm_start = area->vm_start >> PAGE_SHIFT; in set_address_limits() 321 start = vm_start; in set_address_limits() 326 start = max_t(long, start, vm_start); in set_address_limits() 334 *pfn += (*start_vaddr - area->vm_start) >> PAGE_SHIFT; in set_address_limits() 360 page_offset = (vmf->address - area->vm_start) >> PAGE_SHIFT; in vm_fault_gtt() 520 addr -= area->vm_start; in vm_access()
|
| /linux/drivers/soc/qcom/ |
| H A D | rmtfs_mem.c | 136 if (vma->vm_end - vma->vm_start > rmtfs_mem->size) { in qcom_rmtfs_mem_mmap() 139 vma->vm_end, vma->vm_start, in qcom_rmtfs_mem_mmap() 140 (vma->vm_end - vma->vm_start), &rmtfs_mem->size); in qcom_rmtfs_mem_mmap() 146 vma->vm_start, in qcom_rmtfs_mem_mmap() 148 vma->vm_end - vma->vm_start, in qcom_rmtfs_mem_mmap()
|
| /linux/arch/arc/kernel/ |
| H A D | arc_hostlink.c | 22 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in arc_hl_mmap() 23 vma->vm_end - vma->vm_start, in arc_hl_mmap()
|
| /linux/arch/powerpc/platforms/book3s/ |
| H A D | vas-api.c | 438 fault = vmf_insert_pfn(vma, vma->vm_start, in vas_mmap_fault() 516 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) { in coproc_mmap() 518 (vma->vm_end - vma->vm_start), PAGE_SIZE); in coproc_mmap() 573 rc = remap_pfn_range(vma, vma->vm_start, pfn + vma->vm_pgoff, in coproc_mmap() 574 vma->vm_end - vma->vm_start, prot); in coproc_mmap() 577 vma->vm_start, rc); in coproc_mmap()
|
| /linux/drivers/accel/habanalabs/common/ |
| H A D | memory_mgr.c | 205 new_mmap_size = buf->real_mapped_size - (vma->vm_end - vma->vm_start); in hl_mmap_mem_buf_vm_close() 253 user_mem_size = vma->vm_end - vma->vm_start; in hl_mem_mgr_mmap() 262 if (!access_ok((void __user *)(uintptr_t)vma->vm_start, in hl_mem_mgr_mmap() 265 buf->behavior->topic, vma->vm_start); in hl_mem_mgr_mmap()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_mmio_gem.c | 144 if (vma->vm_end - vma->vm_start != base->size) in xe_mmio_gem_mmap() 185 unsigned long addr = vma->vm_start + i; in xe_mmio_gem_vm_fault_dummy_page() 216 unsigned long addr = vma->vm_start + i; in xe_mmio_gem_vm_fault()
|
| /linux/drivers/xen/xenfs/ |
| H A D | xenstored.c | 36 size_t size = vma->vm_end - vma->vm_start; in xsd_kva_mmap() 41 if (remap_pfn_range(vma, vma->vm_start, in xsd_kva_mmap()
|
| /linux/kernel/bpf/ |
| H A D | sysfs_btf.c | 23 size_t vm_size = vma->vm_end - vma->vm_start; in btf_sysfs_vmlinux_mmap() 43 return remap_pfn_range(vma, vma->vm_start, pfn, vm_size, vma->vm_page_prot); in btf_sysfs_vmlinux_mmap()
|
| /linux/include/trace/events/ |
| H A D | fs_dax.h | 16 __field(unsigned long, vm_start) 29 __entry->vm_start = vmf->vma->vm_start; 46 __entry->vm_start,
|