| /linux/fs/coda/ |
| H A D | file.c | 37 struct vm_operations_struct vm_ops; member 126 container_of(vma->vm_ops, struct coda_vm_ops, vm_ops); in coda_vm_open() 138 container_of(vma->vm_ops, struct coda_vm_ops, vm_ops); in coda_vm_close() 144 vma->vm_ops = cvm_ops->host_vm_ops; in coda_vm_close() 212 cvm_ops->host_vm_ops = vma->vm_ops; in coda_file_mmap() 213 if (vma->vm_ops) in coda_file_mmap() 214 cvm_ops->vm_ops = *vma->vm_ops; in coda_file_mmap() 216 cvm_ops->vm_ops.open = coda_vm_open; in coda_file_mmap() 217 cvm_ops->vm_ops.close = coda_vm_close; in coda_file_mmap() 221 vma->vm_ops = &cvm_ops->vm_ops; in coda_file_mmap()
|
| /linux/tools/testing/vma/ |
| H A D | vma.c | 468 const struct vm_operations_struct vm_ops = { in __test_merge_new() local 523 vma_a->vm_ops = &vm_ops; /* This should have no impact. */ in __test_merge_new() 564 vma_d->vm_ops = &vm_ops; /* This should have no impact. */ in __test_merge_new() 584 vma_d->vm_ops = NULL; /* This would otherwise degrade the merge. */ in __test_merge_new() 759 const struct vm_operations_struct vm_ops = { in test_vma_merge_with_close() local 834 vma_next->vm_ops = &vm_ops; in test_vma_merge_with_close() 859 vma->vm_ops = &vm_ops; in test_vma_merge_with_close() 887 vma->vm_ops = &vm_ops; in test_vma_merge_with_close() 916 vma->vm_ops = &vm_ops; in test_vma_merge_with_close() 942 vma_next->vm_ops = &vm_ops; in test_vma_merge_with_close() [all …]
|
| H A D | vma_internal.h | 627 const struct vm_operations_struct *vm_ops; member 701 const struct vm_operations_struct *vm_ops; member 941 vma->vm_ops = &vma_dummy_vm_ops; in vma_init() 1045 return !vma->vm_ops; in vma_is_anonymous() 1390 vma->vm_ops = NULL; in vma_set_anonymous()
|
| /linux/fs/kernfs/ |
| H A D | file.c | 375 if (!of->vm_ops) in kernfs_vma_open() 381 if (of->vm_ops->open) in kernfs_vma_open() 382 of->vm_ops->open(vma); in kernfs_vma_open() 393 if (!of->vm_ops) in kernfs_vma_fault() 400 if (of->vm_ops->fault) in kernfs_vma_fault() 401 ret = of->vm_ops->fault(vmf); in kernfs_vma_fault() 413 if (!of->vm_ops) in kernfs_vma_page_mkwrite() 420 if (of->vm_ops->page_mkwrite) in kernfs_vma_page_mkwrite() 421 ret = of->vm_ops->page_mkwrite(vmf); in kernfs_vma_page_mkwrite() 436 if (!of->vm_ops) in kernfs_vma_access() [all …]
|
| /linux/arch/um/kernel/ |
| H A D | tlb.c | 18 struct vm_ops { struct 53 struct vm_ops *ops) in update_pte_range() 94 struct vm_ops *ops) in update_pmd_range() 117 struct vm_ops *ops) in update_pud_range() 140 struct vm_ops *ops) in update_p4d_range() 164 struct vm_ops ops; in um_tlb_sync()
|
| /linux/ipc/ |
| H A D | shm.c | 89 const struct vm_operations_struct *vm_ops; member 311 if (sfd->vm_ops->open) in shm_open() 312 sfd->vm_ops->open(vma); in shm_open() 404 if (sfd->vm_ops->close) in shm_close() 405 sfd->vm_ops->close(vma); in shm_close() 545 return sfd->vm_ops->fault(vmf); in shm_fault() 553 if (sfd->vm_ops->may_split) in shm_may_split() 554 return sfd->vm_ops->may_split(vma, addr); in shm_may_split() 564 if (sfd->vm_ops->pagesize) in shm_pagesize() 565 return sfd->vm_ops->pagesize(vma); in shm_pagesize() [all …]
|
| /linux/mm/ |
| H A D | vma.c | 23 const struct vm_operations_struct *vm_ops; member 502 if (vma->vm_ops && vma->vm_ops->may_split) { in __split_vma() 503 err = vma->vm_ops->may_split(vma, addr); in __split_vma() 535 if (new->vm_ops && new->vm_ops->open) in __split_vma() 536 new->vm_ops->open(new); in __split_vma() 768 return !vma->vm_ops || !vma->vm_ops->close; in can_merge_remove_vma() 1869 if (new_vma->vm_ops && new_vma->vm_ops->open) in copy_vma() 1870 new_vma->vm_ops->open(new_vma); in copy_vma() 1992 static bool vm_ops_needs_writenotify(const struct vm_operations_struct *vm_ops) in vm_ops_needs_writenotify() argument 1994 return vm_ops && (vm_ops->page_mkwrite || vm_ops->pfn_mkwrite); in vm_ops_needs_writenotify() [all …]
|
| H A D | vma_init.c | 45 dest->vm_ops = src->vm_ops; in vm_area_init_from()
|
| H A D | secretmem.c | 132 desc->vm_ops = &secretmem_vm_ops; in secretmem_mmap_prepare() 139 return vma->vm_ops == &secretmem_vm_ops; in vma_is_secretmem()
|
| H A D | memory.c | 608 vma->vm_ops ? vma->vm_ops->fault : NULL, in print_bad_page_map() 690 if (vma->vm_ops && vma->vm_ops->find_normal_page) in __vm_normal_page() 691 return vma->vm_ops->find_normal_page(vma, addr); in __vm_normal_page() 2249 return vma->vm_ops && vma->vm_ops->pfn_mkwrite && in vm_mixed_zeropage_allowed() 3528 ret = vmf->vma->vm_ops->page_mkwrite(vmf); in do_page_mkwrite() 3556 bool page_mkwrite = vma->vm_ops && vma->vm_ops->page_mkwrite; in fault_dirty_shared_page() 3641 if (vma->vm_ops->map_pages || !(vmf->flags & FAULT_FLAG_VMA_LOCK)) in vmf_can_call_fault() 3893 if (vma->vm_ops && vma->vm_ops->pfn_mkwrite) { in wp_pfn_shared() 3902 ret = vma->vm_ops->pfn_mkwrite(vmf); in wp_pfn_shared() 3919 if (vma->vm_ops && vma->vm_ops->page_mkwrite) { in wp_page_shared() [all …]
|
| H A D | mremap.c | 1050 if (vma->vm_ops && vma->vm_ops->may_split) { in prep_move_vma() 1052 err = vma->vm_ops->may_split(vma, old_addr); in prep_move_vma() 1054 err = vma->vm_ops->may_split(vma, old_addr + old_len); in prep_move_vma() 1215 else if (vma->vm_ops && vma->vm_ops->mremap) in copy_vma_and_data() 1216 err = vma->vm_ops->mremap(new_vma); in copy_vma_and_data()
|
| H A D | internal.h | 177 vma->vm_ops = &vma_dummy_vm_ops; in mmap_file() 189 if (vma->vm_ops && vma->vm_ops->close) { in vma_close() 190 vma->vm_ops->close(vma); in vma_close() 196 vma->vm_ops = &vma_dummy_vm_ops; in vma_close()
|
| H A D | vma.h | 248 vma->vm_ops = desc->vm_ops; in set_vma_from_desc()
|
| H A D | mmap.c | 1457 vma->vm_ops = ops; in __install_special_mapping() 1479 vma->vm_ops == &special_mapping_vmops; in vma_is_special_mapping() 1807 if (tmp->vm_ops && tmp->vm_ops->open) in dup_mmap() 1808 tmp->vm_ops->open(tmp); in dup_mmap()
|
| /linux/arch/hexagon/kernel/ |
| H A D | Makefile | 14 obj-y += vm_entry.o vm_events.o vm_switch.o vm_ops.o vm_init_segtable.o
|
| /linux/include/linux/ |
| H A D | agp_backend.h | 56 const struct vm_operations_struct *vm_ops; member
|
| /linux/drivers/gpu/drm/msm/ |
| H A D | msm_gem_vma.c | 120 struct list_head vm_ops; member 473 list_add_tail(&op->node, &arg->job->vm_ops); in vm_op_enqueue() 708 while (!list_empty(&job->vm_ops)) { in msm_vma_job_run() 710 list_first_entry(&job->vm_ops, struct msm_vm_op, node); in msm_vma_job_run() 771 while (!list_empty(&job->vm_ops)) { in msm_vma_job_free() 773 list_first_entry(&job->vm_ops, struct msm_vm_op, node); in msm_vma_job_free() 974 INIT_LIST_HEAD(&job->vm_ops); in vm_bind_job_create()
|
| /linux/drivers/pci/ |
| H A D | mmap.c | 47 vma->vm_ops = &pci_phys_vm_ops; in pci_mmap_resource_range()
|
| /linux/drivers/media/common/videobuf2/ |
| H A D | videobuf2-vmalloc.c | 195 vma->vm_ops = &vb2_common_vm_ops; in vb2_vmalloc_mmap() 197 vma->vm_ops->open(vma); in vb2_vmalloc_mmap()
|
| /linux/drivers/gpu/drm/ttm/ |
| H A D | ttm_bo_vm.c | 502 if (!vma->vm_ops) in ttm_bo_mmap_obj() 503 vma->vm_ops = &ttm_bo_vm_ops; in ttm_bo_mmap_obj()
|
| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_mmap.c | 105 vma->vm_ops = &rxe_vm_ops; in rxe_mmap()
|
| /linux/fs/nilfs2/ |
| H A D | file.c | 131 desc->vm_ops = &nilfs_file_vm_ops; in nilfs_file_mmap_prepare()
|
| /linux/fs/ocfs2/ |
| H A D | mmap.c | 175 desc->vm_ops = &ocfs2_file_vm_ops; in ocfs2_mmap_prepare()
|
| /linux/drivers/infiniband/sw/rdmavt/ |
| H A D | mmap.c | 97 vma->vm_ops = &rvt_vm_ops; in rvt_mmap()
|
| /linux/drivers/xen/ |
| H A D | privcmd-buf.c | 161 vma->vm_ops = &privcmd_buf_vm_ops; in privcmd_buf_mmap()
|