/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_hmm.c | 66 static bool amdgpu_hmm_invalidate_gfx(struct mmu_interval_notifier *mni, in amdgpu_hmm_invalidate_gfx() argument 70 struct amdgpu_bo *bo = container_of(mni, struct amdgpu_bo, notifier); in amdgpu_hmm_invalidate_gfx() 79 mmu_interval_set_seq(mni, cur_seq); in amdgpu_hmm_invalidate_gfx() 103 static bool amdgpu_hmm_invalidate_hsa(struct mmu_interval_notifier *mni, in amdgpu_hmm_invalidate_hsa() argument 107 struct amdgpu_bo *bo = container_of(mni, struct amdgpu_bo, notifier); in amdgpu_hmm_invalidate_hsa() 112 amdgpu_amdkfd_evict_userptr(mni, cur_seq, bo->kfd_bo); in amdgpu_hmm_invalidate_hsa()
|
H A D | amdgpu_amdkfd.h | 196 int amdgpu_amdkfd_evict_userptr(struct mmu_interval_notifier *mni, 221 int amdgpu_amdkfd_evict_userptr(struct mmu_interval_notifier *mni, in amdgpu_amdkfd_evict_userptr() argument
|
H A D | amdgpu_amdkfd_gpuvm.c | 2507 int amdgpu_amdkfd_evict_userptr(struct mmu_interval_notifier *mni, in amdgpu_amdkfd_evict_userptr() argument 2520 mmu_interval_set_seq(mni, cur_seq); in amdgpu_amdkfd_evict_userptr() 2525 r = kgd2kfd_quiesce_mm(mni->mm, in amdgpu_amdkfd_evict_userptr()
|
/linux/samples/vfs/ |
H A D | mountinfo.c | 213 struct mnt_ns_info mni = { .size = MNT_NS_INFO_SIZE_VER0 }; in main() local 251 ret = ioctl(mntns, NS_MNT_GET_INFO, &mni); in main() 260 ret = dump_mounts(mni.mnt_ns_id); in main() 268 ret = ioctl(mntns, NS_MNT_GET_NEXT, &mni); in main()
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | user_exp_rcv.c | 23 static bool tid_rb_invalidate(struct mmu_interval_notifier *mni, 26 static bool tid_cover_invalidate(struct mmu_interval_notifier *mni, 892 static bool tid_rb_invalidate(struct mmu_interval_notifier *mni, in tid_rb_invalidate() argument 897 container_of(mni, struct tid_rb_node, notifier); in tid_rb_invalidate() 942 static bool tid_cover_invalidate(struct mmu_interval_notifier *mni, in tid_cover_invalidate() argument 947 container_of(mni, struct tid_user_buf, notifier); in tid_cover_invalidate() 952 mmu_interval_set_seq(mni, cur_seq); in tid_cover_invalidate()
|
/linux/lib/ |
H A D | test_hmm.c | 257 static bool dmirror_interval_invalidate(struct mmu_interval_notifier *mni, in dmirror_interval_invalidate() argument 261 struct dmirror *dmirror = container_of(mni, struct dmirror, notifier); in dmirror_interval_invalidate() 276 mmu_interval_set_seq(mni, cur_seq); in dmirror_interval_invalidate() 1077 static bool dmirror_snapshot_invalidate(struct mmu_interval_notifier *mni, in dmirror_snapshot_invalidate() argument 1082 container_of(mni, struct dmirror_interval, notifier); in dmirror_snapshot_invalidate() 1094 mmu_interval_set_seq(mni, cur_seq); in dmirror_snapshot_invalidate()
|
/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_userptr.c | 59 static bool i915_gem_userptr_invalidate(struct mmu_interval_notifier *mni, in i915_gem_userptr_invalidate() argument 63 mmu_interval_set_seq(mni, cur_seq); in i915_gem_userptr_invalidate()
|
/linux/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_svm.c | 73 svm_range_cpu_invalidate_pagetables(struct mmu_interval_notifier *mni, 2569 svm_range_cpu_invalidate_pagetables(struct mmu_interval_notifier *mni, in svm_range_cpu_invalidate_pagetables() argument 2579 if (!mmget_not_zero(mni->mm)) in svm_range_cpu_invalidate_pagetables() 2582 start = mni->interval_tree.start; in svm_range_cpu_invalidate_pagetables() 2583 last = mni->interval_tree.last; in svm_range_cpu_invalidate_pagetables() 2589 mni->interval_tree.start >> PAGE_SHIFT, in svm_range_cpu_invalidate_pagetables() 2590 mni->interval_tree.last >> PAGE_SHIFT, range->event); in svm_range_cpu_invalidate_pagetables() 2592 prange = container_of(mni, struct svm_range, notifier); in svm_range_cpu_invalidate_pagetables() 2595 mmu_interval_set_seq(mni, cur_seq); in svm_range_cpu_invalidate_pagetables() 2599 svm_range_unmap_from_cpu(mni->mm, prange, start, last); in svm_range_cpu_invalidate_pagetables() [all …]
|
/linux/drivers/accel/amdxdna/ |
H A D | amdxdna_gem.c | 93 static bool amdxdna_hmm_invalidate(struct mmu_interval_notifier *mni, in amdxdna_hmm_invalidate() argument 97 struct amdxdna_gem_obj *abo = container_of(mni, struct amdxdna_gem_obj, in amdxdna_hmm_invalidate()
|
/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_svm.c | 506 static bool nouveau_svm_range_invalidate(struct mmu_interval_notifier *mni, in nouveau_svm_range_invalidate() argument 511 container_of(mni, struct svm_notifier, notifier); in nouveau_svm_range_invalidate() 528 mmu_interval_set_seq(mni, cur_seq); in nouveau_svm_range_invalidate()
|
/linux/drivers/video/fbdev/sis/ |
H A D | sis_main.c | 576 switch(sisbios_mode[mode_idx].mode_no[ivideo->mni]) { in sisfb_verify_rate() 600 sisbios_mode[mode_idx].mode_no[ivideo->mni], in sisfb_verify_rate() 1234 ivideo->mode_no = sisbios_mode[ivideo->sisfb_mode_idx].mode_no[ivideo->mni]; in sisfb_do_set_var() 1255 ivideo->mode_no = sisbios_mode[ivideo->sisfb_mode_idx].mode_no[ivideo->mni]; in sisfb_do_set_var() 1532 if(ivideo->sisfb_lastrates[sisbios_mode[search_idx].mode_no[ivideo->mni]]) { in sisfb_check_var() 1534 ivideo->sisfb_lastrates[sisbios_mode[search_idx].mode_no[ivideo->mni]]; in sisfb_check_var() 1560 sisbios_mode[search_idx].mode_no[ivideo->mni], in sisfb_check_var() 1563 sisbios_mode[search_idx].mode_no[ivideo->mni], in sisfb_check_var() 5880 ivideo->mni = chipinfo->mni; in sisfb_probe() 6171 ((sisbios_mode[ivideo->sisfb_mode_idx].mode_no[ivideo->mni]) != 0xFF)) { in sisfb_probe() [all …]
|
H A D | sis_main.h | 84 int mni; member
|
H A D | sis.h | 510 int mni; /* Mode number index */ member
|
/linux/drivers/infiniband/hw/mlx5/ |
H A D | odp.c | 259 static bool mlx5_ib_invalidate_range(struct mmu_interval_notifier *mni, in mlx5_ib_invalidate_range() argument 264 container_of(mni, struct ib_umem_odp, notifier); in mlx5_ib_invalidate_range() 278 mmu_interval_set_seq(mni, cur_seq); in mlx5_ib_invalidate_range()
|
/linux/drivers/gpu/drm/xe/ |
H A D | xe_vm.c | 627 static bool vma_userptr_invalidate(struct mmu_interval_notifier *mni, in vma_userptr_invalidate() argument 631 struct xe_userptr_vma *uvma = container_of(mni, typeof(*uvma), userptr.notifier); in vma_userptr_invalidate() 646 mmu_interval_set_seq(mni, cur_seq); in vma_userptr_invalidate()
|