Searched refs:mmap_read_lock (Results 1 – 25 of 70) sorted by relevance
123
392 mmap_read_lock(current->mm); in sys_cacheflush()404 mmap_read_lock(current->mm); in sys_cacheflush()477 mmap_read_lock(mm); in sys_atomic_cmpxchg_32()552 mmap_read_lock(mm); in sys_atomic_cmpxchg_32()
62 mmap_read_lock(mm); in SYSCALL_DEFINE3()100 mmap_read_lock(mm); in SYSCALL_DEFINE3()
325 mmap_read_lock(mm); in madvise_willneed()913 mmap_read_lock(mm); in madvise_dontneed_free()971 mmap_read_lock(mm); in madvise_populate()1043 mmap_read_lock(mm); in madvise_remove()1648 mmap_read_lock(mm); in try_vma_read_lock()1781 mmap_read_lock(mm); in madvise_lock()
135 mmap_read_lock(mm); in uffd_lock_vma()175 mmap_read_lock(dst_mm); in uffd_mfill_lock()1042 mmap_read_lock(dst_mm); in mwriteprotect_range()1728 mmap_read_lock(mm); in uffd_move_lock()1770 mmap_read_lock(mm); in uffd_move_lock()
326 mmap_read_lock(current->mm); in SYSCALL_DEFINE3()
28 mmap_read_lock(mm); in task_mem()87 mmap_read_lock(mm); in task_vsize()103 mmap_read_lock(mm); in task_statm()
201 pub fn mmap_read_lock(&self) -> MmapReadGuard<'_> { in mmap_read_lock() method203 unsafe { bindings::mmap_read_lock(self.as_raw()) }; in mmap_read_lock()
589 static inline void mmap_read_lock(struct mm_struct *mm) in mmap_read_lock() function622 DEFINE_GUARD(mmap_read_lock, struct mm_struct *, in DEFINE_GUARD() argument623 mmap_read_lock(_T), mmap_read_unlock(_T)) in DEFINE_GUARD()
28 mmap_read_lock(mm); in rust_helper_mmap_read_lock()
72 mmap_read_lock(current->mm); in sys_cacheflush()
294 mmap_read_lock(mm); in do_page_fault()496 mmap_read_lock(mm); in handle_nadtlb_fault()
792 mmap_read_lock(range->gpusvm->mm); in drm_gpusvm_scan_mm()987 mmap_read_lock(mm); in drm_gpusvm_find_vma_start()1055 mmap_read_lock(mm); in drm_gpusvm_range_find_or_insert()1431 mmap_read_lock(mm); in drm_gpusvm_get_pages()1707 mmap_read_lock(mm); in drm_gpusvm_range_evict()
620 mmap_read_lock(kvm->mm); in kvmppc_uvmem_drop_pages()802 mmap_read_lock(kvm->mm); in kvmppc_uv_migrate_mem_slot()959 mmap_read_lock(kvm->mm); in kvmppc_h_svm_page_in()1066 mmap_read_lock(kvm->mm); in kvmppc_h_svm_page_out()
93 mmap_read_lock(mm); in do_page_fault()
93 mmap_read_lock(current->mm); in SYSCALL_DEFINE3()
85 mmap_read_lock(active_mm); in show_faulting_vma()
101 mmap_read_lock(current->mm); in set_segfault()
147 mmap_read_lock(mm); in do_page_fault()
110 mmap_read_lock(mm); in do_page_fault()
182 mmap_read_lock(current->mm); in kvm_arch_prepare_memory_region()454 mmap_read_lock(current->mm); in kvm_riscv_mmu_map()
66 mmap_read_lock(mm); in async_pf_execute()
105 mmap_read_lock(current->mm); in xdp_umem_pin_pages()
167 mmap_read_lock(mm); in nouveau_svmm_bind()612 mmap_read_lock(mm); in nouveau_atomic_range_fault()687 mmap_read_lock(mm); in nouveau_range_fault()
423 mmap_read_lock(mm->mm); in user_event_mm_fault_in()481 mmap_read_lock(mm->mm); in user_event_enabler_fault_fixup()616 mmap_read_lock(mm->mm); in user_event_enabler_update()914 mmap_read_lock(user_mm->mm); in user_event_enabler_create()2592 mmap_read_lock(user_mm->mm); in user_event_mm_clear_bit()
237 mmap_read_lock(mm); in iommu_sva_handle_mm()