Home
last modified time | relevance | path

Searched refs:mmap_lock (Results 1 – 25 of 31) sorted by relevance

12

/linux/drivers/media/common/videobuf2/
H A Dvideobuf2-core.c588 lockdep_assert_held(&q->mmap_lock); in __vb2_queue_free()
894 mutex_lock(&q->mmap_lock); in vb2_core_reqbufs()
906 mutex_unlock(&q->mmap_lock); in vb2_core_reqbufs()
927 mutex_lock(&q->mmap_lock); in vb2_core_reqbufs()
930 mutex_unlock(&q->mmap_lock); in vb2_core_reqbufs()
999 mutex_lock(&q->mmap_lock); in vb2_core_reqbufs()
1008 mutex_unlock(&q->mmap_lock); in vb2_core_reqbufs()
1011 mutex_unlock(&q->mmap_lock); in vb2_core_reqbufs()
1024 mutex_lock(&q->mmap_lock); in vb2_core_reqbufs()
1026 mutex_unlock(&q->mmap_lock); in vb2_core_reqbufs()
[all …]
/linux/io_uring/
H A Dmemmap.c261 lockdep_assert_held(&ctx->mmap_lock); in io_region_validate_mmap()
304 guard(mutex)(&ctx->mmap_lock); in io_uring_mmap()
336 guard(mutex)(&ctx->mmap_lock); in io_uring_get_unmapped_area()
386 guard(mutex)(&ctx->mmap_lock); in io_uring_get_unmapped_area()
H A Dkbuf.c95 guard(mutex)(&ctx->mmap_lock); in io_buffer_add_list()
462 scoped_guard(mutex, &ctx->mmap_lock) { in io_destroy_buffers()
475 scoped_guard(mutex, &ctx->mmap_lock) in io_destroy_bl()
716 scoped_guard(mutex, &ctx->mmap_lock) in io_unregister_pbuf_ring()
751 lockdep_assert_held(&ctx->mmap_lock); in io_pbuf_get_region()
H A Dzcrx.c641 lockdep_assert_held(&ctx->mmap_lock); in io_zcrx_get_region()
726 scoped_guard(mutex, &ctx->mmap_lock) { in import_zcrx()
739 scoped_guard(mutex, &ctx->mmap_lock) { in import_zcrx()
747 scoped_guard(mutex, &ctx->mmap_lock) in import_zcrx()
811 scoped_guard(mutex, &ctx->mmap_lock) { in io_register_zcrx_ifq()
852 scoped_guard(mutex, &ctx->mmap_lock) { in io_register_zcrx_ifq()
871 scoped_guard(mutex, &ctx->mmap_lock) in io_register_zcrx_ifq()
895 scoped_guard(mutex, &ctx->mmap_lock) { in io_unregister_zcrx_ifqs()
H A Dregister.c578 mutex_lock(&ctx->mmap_lock); in io_register_resize_rings()
652 mutex_unlock(&ctx->mmap_lock); in io_register_resize_rings()
/linux/kernel/bpf/
H A Dmmap_unlock_work.h60 rwsem_release(&mm->mmap_lock.dep_map, _RET_IP_); in bpf_mmap_unlock_mm()
/linux/mm/
H A Dinit-mm.c33 .mm_mt = MTREE_INIT_EXT(mm_mt, MM_MT_FLAGS, init_mm.mmap_lock),
H A DMakefile58 debug.o gup.o mmap_lock.o vma_init.o $(mmu-y)
H A Dmmu_notifier.c979 might_lock(&mm->mmap_lock); in mmu_interval_notifier_insert()
H A Dvma.c2125 down_write_nest_lock(&anon_vma->root->rwsem, &mm->mmap_lock); in vm_lock_anon_vma()
2155 down_write_nest_lock(&mapping->i_mmap_rwsem, &mm->mmap_lock); in vm_lock_mapping()
/linux/drivers/infiniband/hw/cxgb4/
H A Dprovider.c89 spin_lock_init(&context->mmap_lock); in c4iw_alloc_ucontext()
103 spin_lock(&context->mmap_lock); in c4iw_alloc_ucontext()
106 spin_unlock(&context->mmap_lock); in c4iw_alloc_ucontext()
H A Dcq.c1110 spin_lock(&ucontext->mmap_lock); in c4iw_create_cq()
1120 spin_unlock(&ucontext->mmap_lock); in c4iw_create_cq()
H A Dqp.c2265 spin_lock(&ucontext->mmap_lock); in c4iw_create_qp()
2282 spin_unlock(&ucontext->mmap_lock); in c4iw_create_qp()
2774 spin_lock(&ucontext->mmap_lock); in c4iw_create_srq()
2779 spin_unlock(&ucontext->mmap_lock); in c4iw_create_srq()
/linux/tools/perf/util/bpf_skel/vmlinux/
H A Dvmlinux.h97 struct rw_semaphore mmap_lock; member
/linux/tools/perf/util/bpf_skel/
H A Dlock_contention.bpf.c175 struct rw_semaphore mmap_lock; member
377 if (bpf_core_field_exists(mm_new->mmap_lock)) { in check_lock_type()
378 if (&mm_new->mmap_lock == (void *)lock) in check_lock_type()
/linux/Documentation/admin-guide/mm/
H A Dnuma_memory_policy.rst381 task's mm's mmap_lock for read during the query. The set_mempolicy() and
382 mbind() APIs [see below] always acquire the mmap_lock for write when
388 we hold them mmap_lock for read. Again, because replacing the task or vma
389 policy requires that the mmap_lock be held for write, the policy can't be
393 shared memory policy while another task, with a distinct mmap_lock, is
H A Dmultigen_lru.rst41 theoretically worsen lock contention (mmap_lock). If it is
/linux/include/media/
H A Dvideobuf2-core.h564 * @mmap_lock: private mutex used when buffers are allocated/freed/mmapped
636 struct mutex mmap_lock;
646 struct mutex mmap_lock; global() member
/linux/include/linux/
H A Dio_uring_types.h485 struct mutex mmap_lock; member
H A Dmm_types.h1196 struct rw_semaphore mmap_lock; member
/linux/drivers/gpu/drm/etnaviv/
H A Detnaviv_gem.c673 might_lock_read(&current->mm->mmap_lock); in etnaviv_gem_userptr_get_pages()
/linux/Documentation/filesystems/
H A Dlocking.rst604 ops mmap_lock PageLocked(page)
629 The mmap_lock may not be held when this method is called.
/linux/drivers/firmware/efi/
H A Defi.c69 .mm_mt = MTREE_INIT_EXT(mm_mt, MM_MT_FLAGS, efi_mm.mmap_lock),
/linux/Documentation/gpu/
H A Ddrm-vm-bind-locking.rst82 correspond to the mmap_lock. An rwsem allows several readers to walk
419 mmap_lock that is grabbed when resolving a CPU pagefault. This means
/linux/kernel/
H A Dfork.c1065 init_rwsem(&mm->mmap_lock); in mmap_init_lock()
1076 mt_set_external_lock(&mm->mm_mt, &mm->mmap_lock); in mm_init()

12