| /linux/rust/kernel/ |
| H A D | mm.rs | 244 pub fn vma_lookup(&self, vma_addr: usize) -> Option<&virt::VmaRef> { in vma_lookup() method 247 let vma = unsafe { bindings::vma_lookup(self.mm.as_raw(), vma_addr) }; in vma_lookup()
|
| /linux/rust/helpers/ |
| H A D | mm.c | 44 return vma_lookup(mm, addr); in rust_helper_vma_lookup()
|
| /linux/arch/arc/kernel/ |
| H A D | troubleshoot.c | 86 vma = vma_lookup(active_mm, address); in show_faulting_vma()
|
| /linux/mm/ |
| H A D | mincore.c | 253 vma = vma_lookup(current->mm, addr); in do_mincore()
|
| H A D | mmap_lock.c | 564 vma = vma_lookup(mm, addr);
|
| H A D | mremap.c | 1355 vrm->vma = vma_lookup(mm, vrm->addr); in shrink_vma() 1389 vrm->vma = vma_lookup(mm, vrm->addr); in mremap_to() 1935 vrm->vma = vma_lookup(current->mm, vrm->addr); in do_mremap()
|
| H A D | mmap.c | 1119 vma = vma_lookup(mm, start); in SYSCALL_DEFINE5() 1156 vma = vma_lookup(mm, start); in SYSCALL_DEFINE5()
|
| H A D | userfaultfd.c | 48 vma = vma_lookup(mm, addr); in find_vma_and_prepare_anon() 1591 vma = vma_lookup(mm, src_start); in find_vmas_mm_locked()
|
| H A D | migrate.c | 2286 vma = vma_lookup(mm, addr); in add_folio_for_migration() 2444 vma = vma_lookup(mm, addr); in do_pages_stat_array()
|
| H A D | gup.c | 1262 * This is "vma_lookup()", but with a warning if we would have 1269 return vma_lookup(mm, addr); in gup_vma_lookup() 1385 vma = vma_lookup(mm, start); in __get_user_pages()
|
| /linux/arch/parisc/mm/ |
| H A D | fault.c | 497 vma = vma_lookup(mm, address); in handle_nadtlb_fault()
|
| /linux/drivers/iommu/ |
| H A D | iommu-sva.c | 239 vma = vma_lookup(mm, prm->addr); in iommu_sva_handle_mm()
|
| /linux/arch/m68k/kernel/ |
| H A D | sys_m68k.c | 405 vma = vma_lookup(current->mm, addr); in sys_cacheflush()
|
| /linux/drivers/virt/acrn/ |
| H A D | mm.c | 177 vma = vma_lookup(current->mm, memmap->vma_base); in acrn_vm_ram_map()
|
| /linux/drivers/android/ |
| H A D | binder_alloc.c | 273 vma = vma_lookup(mm, addr); in binder_page_insert() 1161 vma = vma_lookup(mm, page_addr); in binder_alloc_free_page()
|
| /linux/arch/riscv/kvm/ |
| H A D | mmu.c | 483 vma = vma_lookup(current->mm, hva); in kvm_riscv_mmu_map()
|
| /linux/include/linux/ |
| H A D | mm.h | 2935 static inline struct vm_area_struct *vma_lookup(struct mm_struct *mm, 2977 vma = vma_lookup(mm, addr); in get_user_page_vma_remote() 3957 struct vm_area_struct *vma_lookup(struct mm_struct *mm, unsigned long addr) in vma_lookup() function 4090 struct vm_area_struct *vma = vma_lookup(mm, vm_start); in find_exact_vma()
|
| /linux/drivers/misc/sgi-gru/ |
| H A D | grufault.c | 52 vma = vma_lookup(current->mm, vaddr); in gru_find_vma()
|
| /linux/drivers/gpu/drm/i915/gem/selftests/ |
| H A D | i915_gem_mman.c | 926 area = vma_lookup(current->mm, addr); in __igt_mmap() 1194 area = vma_lookup(current->mm, addr); in __igt_mmap_migrate()
|
| /linux/lib/ |
| H A D | test_hmm.c | 1067 vma = vma_lookup(mm, addr); in dmirror_migrate_to_system() 1141 vma = vma_lookup(mm, addr); in dmirror_migrate_to_device()
|
| /linux/arch/loongarch/kernel/ |
| H A D | traps.c | 491 if (vma_lookup(current->mm, (unsigned long)fault_addr)) in process_fpemu_return()
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_svm.c | 1761 vma = vma_lookup(mm, addr); in svm_range_validate_and_map() 2808 vma = vma_lookup(p->mm, addr << PAGE_SHIFT); in svm_range_get_range_boundaries() 3168 vma = vma_lookup(mm, addr << PAGE_SHIFT); in svm_range_restore_pages() 3464 vma = vma_lookup(p->mm, start); in svm_range_is_valid()
|
| /linux/arch/powerpc/kvm/ |
| H A D | book3s_hv_uvmem.c | 629 vma = vma_lookup(kvm->mm, addr); in kvmppc_uvmem_drop_pages()
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_pagemap.c | 502 vas = vma_lookup(mm, start); in drm_pagemap_migrate_to_devmem()
|
| /linux/tools/testing/vma/include/ |
| H A D | dup.h | 1131 struct vm_area_struct *vma_lookup(struct mm_struct *mm, unsigned long addr) in vma_lookup() function
|