Lines Matching +full:- +full:group

1 // SPDX-License-Identifier: GPL-2.0-only
26 static struct vfio_device *vfio_device_get_from_name(struct vfio_group *group, in vfio_device_get_from_name() argument
29 struct vfio_device *it, *device = ERR_PTR(-ENODEV); in vfio_device_get_from_name()
31 mutex_lock(&group->device_lock); in vfio_device_get_from_name()
32 list_for_each_entry(it, &group->device_list, group_next) { in vfio_device_get_from_name()
35 if (it->ops->match) { in vfio_device_get_from_name()
36 ret = it->ops->match(it, buf); in vfio_device_get_from_name()
42 ret = !strcmp(dev_name(it->dev), buf); in vfio_device_get_from_name()
50 mutex_unlock(&group->device_lock); in vfio_device_get_from_name()
56 * VFIO Group fd, /dev/vfio/$GROUP
58 static bool vfio_group_has_iommu(struct vfio_group *group) in vfio_group_has_iommu() argument
60 lockdep_assert_held(&group->group_lock); in vfio_group_has_iommu()
65 WARN_ON(!group->container != !group->container_users); in vfio_group_has_iommu()
67 return group->container || group->iommufd; in vfio_group_has_iommu()
73 * the group, we know that still exists, therefore the only valid
74 * transition here is 1->0.
76 static int vfio_group_ioctl_unset_container(struct vfio_group *group) in vfio_group_ioctl_unset_container() argument
80 mutex_lock(&group->group_lock); in vfio_group_ioctl_unset_container()
81 if (!vfio_group_has_iommu(group)) { in vfio_group_ioctl_unset_container()
82 ret = -EINVAL; in vfio_group_ioctl_unset_container()
85 if (group->container) { in vfio_group_ioctl_unset_container()
86 if (group->container_users != 1) { in vfio_group_ioctl_unset_container()
87 ret = -EBUSY; in vfio_group_ioctl_unset_container()
90 vfio_group_detach_container(group); in vfio_group_ioctl_unset_container()
92 if (group->iommufd) { in vfio_group_ioctl_unset_container()
93 iommufd_ctx_put(group->iommufd); in vfio_group_ioctl_unset_container()
94 group->iommufd = NULL; in vfio_group_ioctl_unset_container()
98 mutex_unlock(&group->group_lock); in vfio_group_ioctl_unset_container()
102 static int vfio_group_ioctl_set_container(struct vfio_group *group, in vfio_group_ioctl_set_container() argument
111 return -EFAULT; in vfio_group_ioctl_set_container()
115 return -EBADF; in vfio_group_ioctl_set_container()
117 mutex_lock(&group->group_lock); in vfio_group_ioctl_set_container()
118 if (vfio_group_has_iommu(group)) { in vfio_group_ioctl_set_container()
119 ret = -EINVAL; in vfio_group_ioctl_set_container()
122 if (!group->iommu_group) { in vfio_group_ioctl_set_container()
123 ret = -ENODEV; in vfio_group_ioctl_set_container()
129 ret = vfio_container_attach_group(container, group); in vfio_group_ioctl_set_container()
136 group->type == VFIO_NO_IOMMU) in vfio_group_ioctl_set_container()
146 group->iommufd = iommufd; in vfio_group_ioctl_set_container()
151 ret = -EBADFD; in vfio_group_ioctl_set_container()
154 mutex_unlock(&group->group_lock); in vfio_group_ioctl_set_container()
160 spin_lock(&device->group->kvm_ref_lock); in vfio_device_group_get_kvm_safe()
161 vfio_device_get_kvm_safe(device, device->group->kvm); in vfio_device_group_get_kvm_safe()
162 spin_unlock(&device->group->kvm_ref_lock); in vfio_device_group_get_kvm_safe()
167 struct vfio_device *device = df->device; in vfio_df_group_open()
170 mutex_lock(&device->group->group_lock); in vfio_df_group_open()
171 if (!vfio_group_has_iommu(device->group)) { in vfio_df_group_open()
172 ret = -EINVAL; in vfio_df_group_open()
176 mutex_lock(&device->dev_set->lock); in vfio_df_group_open()
180 * associated with the group (if there is one) and obtain a reference in vfio_df_group_open()
184 if (device->open_count == 0) in vfio_df_group_open()
187 df->iommufd = device->group->iommufd; in vfio_df_group_open()
188 if (df->iommufd && vfio_device_is_noiommu(device) && device->open_count == 0) { in vfio_df_group_open()
195 vfio_iommufd_device_has_compat_ioas(device, df->iommufd)) { in vfio_df_group_open()
196 ret = -EPERM; in vfio_df_group_open()
205 if (df->iommufd && device->open_count == 1) { in vfio_df_group_open()
206 ret = vfio_iommufd_compat_attach_ioas(device, df->iommufd); in vfio_df_group_open()
215 smp_store_release(&df->access_granted, true); in vfio_df_group_open()
217 mutex_unlock(&device->dev_set->lock); in vfio_df_group_open()
218 mutex_unlock(&device->group->group_lock); in vfio_df_group_open()
224 df->iommufd = NULL; in vfio_df_group_open()
225 if (device->open_count == 0) in vfio_df_group_open()
227 mutex_unlock(&device->dev_set->lock); in vfio_df_group_open()
229 mutex_unlock(&device->group->group_lock); in vfio_df_group_open()
235 struct vfio_device *device = df->device; in vfio_df_group_close()
237 mutex_lock(&device->group->group_lock); in vfio_df_group_close()
238 mutex_lock(&device->dev_set->lock); in vfio_df_group_close()
241 df->iommufd = NULL; in vfio_df_group_close()
243 if (device->open_count == 0) in vfio_df_group_close()
246 mutex_unlock(&device->dev_set->lock); in vfio_df_group_close()
247 mutex_unlock(&device->group->group_lock); in vfio_df_group_close()
262 df->group = device->group; in vfio_device_open_file()
268 filep = anon_inode_getfile_fmode("[vfio-device]", &vfio_device_fops, in vfio_device_open_file()
279 filep->f_mapping = device->inode->i_mapping; in vfio_device_open_file()
281 if (device->group->type == VFIO_NO_IOMMU) in vfio_device_open_file()
282 dev_warn(device->dev, "vfio-noiommu device opened by user " in vfio_device_open_file()
283 "(%s:%d)\n", current->comm, task_pid_nr(current)); in vfio_device_open_file()
298 static int vfio_group_ioctl_get_device_fd(struct vfio_group *group, in vfio_group_ioctl_get_device_fd() argument
311 device = vfio_device_get_from_name(group, buf); in vfio_group_ioctl_get_device_fd()
338 static int vfio_group_ioctl_get_status(struct vfio_group *group, in vfio_group_ioctl_get_status() argument
345 return -EFAULT; in vfio_group_ioctl_get_status()
348 return -EINVAL; in vfio_group_ioctl_get_status()
352 mutex_lock(&group->group_lock); in vfio_group_ioctl_get_status()
353 if (!group->iommu_group) { in vfio_group_ioctl_get_status()
354 mutex_unlock(&group->group_lock); in vfio_group_ioctl_get_status()
355 return -ENODEV; in vfio_group_ioctl_get_status()
365 if (vfio_group_has_iommu(group)) in vfio_group_ioctl_get_status()
368 else if (!iommu_group_dma_owner_claimed(group->iommu_group)) in vfio_group_ioctl_get_status()
370 mutex_unlock(&group->group_lock); in vfio_group_ioctl_get_status()
373 return -EFAULT; in vfio_group_ioctl_get_status()
380 struct vfio_group *group = filep->private_data; in vfio_group_fops_unl_ioctl() local
385 return vfio_group_ioctl_get_device_fd(group, uarg); in vfio_group_fops_unl_ioctl()
387 return vfio_group_ioctl_get_status(group, uarg); in vfio_group_fops_unl_ioctl()
389 return vfio_group_ioctl_set_container(group, uarg); in vfio_group_fops_unl_ioctl()
391 return vfio_group_ioctl_unset_container(group); in vfio_group_fops_unl_ioctl()
393 return -ENOTTY; in vfio_group_fops_unl_ioctl()
399 struct vfio_group *group = device->group; in vfio_device_block_group() local
402 mutex_lock(&group->group_lock); in vfio_device_block_group()
403 if (group->opened_file) { in vfio_device_block_group()
404 ret = -EBUSY; in vfio_device_block_group()
408 group->cdev_device_open_cnt++; in vfio_device_block_group()
411 mutex_unlock(&group->group_lock); in vfio_device_block_group()
417 struct vfio_group *group = device->group; in vfio_device_unblock_group() local
419 mutex_lock(&group->group_lock); in vfio_device_unblock_group()
420 group->cdev_device_open_cnt--; in vfio_device_unblock_group()
421 mutex_unlock(&group->group_lock); in vfio_device_unblock_group()
426 struct vfio_group *group = in vfio_group_fops_open() local
427 container_of(inode->i_cdev, struct vfio_group, cdev); in vfio_group_fops_open()
430 mutex_lock(&group->group_lock); in vfio_group_fops_open()
434 * will be stable at 0 under the group rwsem in vfio_group_fops_open()
436 if (refcount_read(&group->drivers) == 0) { in vfio_group_fops_open()
437 ret = -ENODEV; in vfio_group_fops_open()
441 if (group->type == VFIO_NO_IOMMU && !capable(CAP_SYS_RAWIO)) { in vfio_group_fops_open()
442 ret = -EPERM; in vfio_group_fops_open()
446 if (group->cdev_device_open_cnt) { in vfio_group_fops_open()
447 ret = -EBUSY; in vfio_group_fops_open()
452 * Do we need multiple instances of the group open? Seems not. in vfio_group_fops_open()
454 if (group->opened_file) { in vfio_group_fops_open()
455 ret = -EBUSY; in vfio_group_fops_open()
458 group->opened_file = filep; in vfio_group_fops_open()
459 filep->private_data = group; in vfio_group_fops_open()
462 mutex_unlock(&group->group_lock); in vfio_group_fops_open()
468 struct vfio_group *group = filep->private_data; in vfio_group_fops_release() local
470 filep->private_data = NULL; in vfio_group_fops_release()
472 mutex_lock(&group->group_lock); in vfio_group_fops_release()
474 * Device FDs hold a group file reference, therefore the group release in vfio_group_fops_release()
477 WARN_ON(group->notifier.head); in vfio_group_fops_release()
478 if (group->container) in vfio_group_fops_release()
479 vfio_group_detach_container(group); in vfio_group_fops_release()
480 if (group->iommufd) { in vfio_group_fops_release()
481 iommufd_ctx_put(group->iommufd); in vfio_group_fops_release()
482 group->iommufd = NULL; in vfio_group_fops_release()
484 group->opened_file = NULL; in vfio_group_fops_release()
485 mutex_unlock(&group->group_lock); in vfio_group_fops_release()
498 * Group objects - create, release, get, put, search
503 struct vfio_group *group; in vfio_group_find_from_iommu() local
508 * group->iommu_group from the vfio.group_list cannot be NULL in vfio_group_find_from_iommu()
511 list_for_each_entry(group, &vfio.group_list, vfio_next) { in vfio_group_find_from_iommu()
512 if (group->iommu_group == iommu_group) in vfio_group_find_from_iommu()
513 return group; in vfio_group_find_from_iommu()
520 struct vfio_group *group = container_of(dev, struct vfio_group, dev); in vfio_group_release() local
522 mutex_destroy(&group->device_lock); in vfio_group_release()
523 mutex_destroy(&group->group_lock); in vfio_group_release()
524 WARN_ON(group->iommu_group); in vfio_group_release()
525 WARN_ON(group->cdev_device_open_cnt); in vfio_group_release()
526 ida_free(&vfio.group_ida, MINOR(group->dev.devt)); in vfio_group_release()
527 kfree(group); in vfio_group_release()
533 struct vfio_group *group; in vfio_group_alloc() local
536 group = kzalloc(sizeof(*group), GFP_KERNEL); in vfio_group_alloc()
537 if (!group) in vfio_group_alloc()
538 return ERR_PTR(-ENOMEM); in vfio_group_alloc()
542 kfree(group); in vfio_group_alloc()
546 device_initialize(&group->dev); in vfio_group_alloc()
547 group->dev.devt = MKDEV(MAJOR(vfio.group_devt), minor); in vfio_group_alloc()
548 group->dev.class = vfio.class; in vfio_group_alloc()
549 group->dev.release = vfio_group_release; in vfio_group_alloc()
550 cdev_init(&group->cdev, &vfio_group_fops); in vfio_group_alloc()
551 group->cdev.owner = THIS_MODULE; in vfio_group_alloc()
553 refcount_set(&group->drivers, 1); in vfio_group_alloc()
554 mutex_init(&group->group_lock); in vfio_group_alloc()
555 spin_lock_init(&group->kvm_ref_lock); in vfio_group_alloc()
556 INIT_LIST_HEAD(&group->device_list); in vfio_group_alloc()
557 mutex_init(&group->device_lock); in vfio_group_alloc()
558 group->iommu_group = iommu_group; in vfio_group_alloc()
561 group->type = type; in vfio_group_alloc()
562 BLOCKING_INIT_NOTIFIER_HEAD(&group->notifier); in vfio_group_alloc()
564 return group; in vfio_group_alloc()
570 struct vfio_group *group; in vfio_create_group() local
576 group = vfio_group_alloc(iommu_group, type); in vfio_create_group()
577 if (IS_ERR(group)) in vfio_create_group()
578 return group; in vfio_create_group()
580 err = dev_set_name(&group->dev, "%s%d", in vfio_create_group()
581 group->type == VFIO_NO_IOMMU ? "noiommu-" : "", in vfio_create_group()
588 err = cdev_device_add(&group->cdev, &group->dev); in vfio_create_group()
594 list_add(&group->vfio_next, &vfio.group_list); in vfio_create_group()
596 return group; in vfio_create_group()
599 put_device(&group->dev); in vfio_create_group()
607 struct vfio_group *group; in vfio_noiommu_group_alloc() local
614 ret = iommu_group_set_name(iommu_group, "vfio-noiommu"); in vfio_noiommu_group_alloc()
622 group = vfio_create_group(iommu_group, type); in vfio_noiommu_group_alloc()
624 if (IS_ERR(group)) { in vfio_noiommu_group_alloc()
625 ret = PTR_ERR(group); in vfio_noiommu_group_alloc()
629 return group; in vfio_noiommu_group_alloc()
638 static bool vfio_group_has_device(struct vfio_group *group, struct device *dev) in vfio_group_has_device() argument
642 mutex_lock(&group->device_lock); in vfio_group_has_device()
643 list_for_each_entry(device, &group->device_list, group_next) { in vfio_group_has_device()
644 if (device->dev == dev) { in vfio_group_has_device()
645 mutex_unlock(&group->device_lock); in vfio_group_has_device()
649 mutex_unlock(&group->device_lock); in vfio_group_has_device()
656 struct vfio_group *group; in vfio_group_find_or_alloc() local
661 * With noiommu enabled, create an IOMMU group for devices that in vfio_group_find_or_alloc()
666 group = vfio_noiommu_group_alloc(dev, VFIO_NO_IOMMU); in vfio_group_find_or_alloc()
667 if (!IS_ERR(group)) { in vfio_group_find_or_alloc()
669 dev_warn(dev, "Adding kernel taint for vfio-noiommu group on device\n"); in vfio_group_find_or_alloc()
671 return group; in vfio_group_find_or_alloc()
675 return ERR_PTR(-EINVAL); in vfio_group_find_or_alloc()
678 group = vfio_group_find_from_iommu(iommu_group); in vfio_group_find_or_alloc()
679 if (group) { in vfio_group_find_or_alloc()
680 if (WARN_ON(vfio_group_has_device(group, dev))) in vfio_group_find_or_alloc()
681 group = ERR_PTR(-EINVAL); in vfio_group_find_or_alloc()
683 refcount_inc(&group->drivers); in vfio_group_find_or_alloc()
685 group = vfio_create_group(iommu_group, VFIO_IOMMU); in vfio_group_find_or_alloc()
691 return group; in vfio_group_find_or_alloc()
697 struct vfio_group *group; in vfio_device_set_group() local
700 group = vfio_group_find_or_alloc(device->dev); in vfio_device_set_group()
702 group = vfio_noiommu_group_alloc(device->dev, type); in vfio_device_set_group()
704 if (IS_ERR(group)) in vfio_device_set_group()
705 return PTR_ERR(group); in vfio_device_set_group()
707 /* Our reference on group is moved to the device */ in vfio_device_set_group()
708 device->group = group; in vfio_device_set_group()
714 struct vfio_group *group = device->group; in vfio_device_remove_group() local
717 if (group->type == VFIO_NO_IOMMU || group->type == VFIO_EMULATED_IOMMU) in vfio_device_remove_group()
718 iommu_group_remove_device(device->dev); in vfio_device_remove_group()
721 if (!refcount_dec_and_mutex_lock(&group->drivers, &vfio.group_lock)) in vfio_device_remove_group()
723 list_del(&group->vfio_next); in vfio_device_remove_group()
726 * We could concurrently probe another driver in the group that might in vfio_device_remove_group()
731 cdev_device_del(&group->cdev, &group->dev); in vfio_device_remove_group()
733 mutex_lock(&group->group_lock); in vfio_device_remove_group()
738 * properly hold the group reference. in vfio_device_remove_group()
740 WARN_ON(!list_empty(&group->device_list)); in vfio_device_remove_group()
741 WARN_ON(group->notifier.head); in vfio_device_remove_group()
744 * Revoke all users of group->iommu_group. At this point we know there in vfio_device_remove_group()
748 if (group->container) in vfio_device_remove_group()
749 vfio_group_detach_container(group); in vfio_device_remove_group()
750 iommu_group = group->iommu_group; in vfio_device_remove_group()
751 group->iommu_group = NULL; in vfio_device_remove_group()
752 mutex_unlock(&group->group_lock); in vfio_device_remove_group()
756 put_device(&group->dev); in vfio_device_remove_group()
761 mutex_lock(&device->group->device_lock); in vfio_device_group_register()
762 list_add(&device->group_next, &device->group->device_list); in vfio_device_group_register()
763 mutex_unlock(&device->group->device_lock); in vfio_device_group_register()
768 mutex_lock(&device->group->device_lock); in vfio_device_group_unregister()
769 list_del(&device->group_next); in vfio_device_group_unregister()
770 mutex_unlock(&device->group->device_lock); in vfio_device_group_unregister()
775 struct vfio_group *group = device->group; in vfio_device_group_use_iommu() local
778 lockdep_assert_held(&group->group_lock); in vfio_device_group_use_iommu()
780 if (WARN_ON(!group->container)) in vfio_device_group_use_iommu()
781 return -EINVAL; in vfio_device_group_use_iommu()
783 ret = vfio_group_use_container(group); in vfio_device_group_use_iommu()
792 struct vfio_group *group = device->group; in vfio_device_group_unuse_iommu() local
794 lockdep_assert_held(&group->group_lock); in vfio_device_group_unuse_iommu()
796 if (WARN_ON(!group->container)) in vfio_device_group_unuse_iommu()
800 vfio_group_unuse_container(group); in vfio_device_group_unuse_iommu()
805 return device->group->container; in vfio_device_has_container()
810 struct vfio_group *group = file->private_data; in vfio_group_from_file() local
812 if (file->f_op != &vfio_group_fops) in vfio_group_from_file()
814 return group; in vfio_group_from_file()
818 * vfio_file_iommu_group - Return the struct iommu_group for the vfio group file
819 * @file: VFIO group file
822 * returns a reference on the group. This function is deprecated, only the SPAPR
827 struct vfio_group *group = vfio_group_from_file(file); in vfio_file_iommu_group() local
833 if (!group) in vfio_file_iommu_group()
836 mutex_lock(&group->group_lock); in vfio_file_iommu_group()
837 if (group->iommu_group) { in vfio_file_iommu_group()
838 iommu_group = group->iommu_group; in vfio_file_iommu_group()
841 mutex_unlock(&group->group_lock); in vfio_file_iommu_group()
847 * vfio_file_is_group - True if the file is a vfio group file
848 * @file: VFIO group file
856 bool vfio_group_enforced_coherent(struct vfio_group *group) in vfio_group_enforced_coherent() argument
864 * is set then the iommu_domain eventually attached to the device/group in vfio_group_enforced_coherent()
867 mutex_lock(&group->device_lock); in vfio_group_enforced_coherent()
868 list_for_each_entry(device, &group->device_list, group_next) { in vfio_group_enforced_coherent()
869 if (!device_iommu_capable(device->dev, in vfio_group_enforced_coherent()
875 mutex_unlock(&group->device_lock); in vfio_group_enforced_coherent()
879 void vfio_group_set_kvm(struct vfio_group *group, struct kvm *kvm) in vfio_group_set_kvm() argument
881 spin_lock(&group->kvm_ref_lock); in vfio_group_set_kvm()
882 group->kvm = kvm; in vfio_group_set_kvm()
883 spin_unlock(&group->kvm_ref_lock); in vfio_group_set_kvm()
887 * vfio_file_has_dev - True if the VFIO file is a handle for device
895 struct vfio_group *group = vfio_group_from_file(file); in vfio_file_has_dev() local
897 if (!group) in vfio_file_has_dev()
900 return group == device->group; in vfio_file_has_dev()
921 /* /dev/vfio/$GROUP */ in vfio_group_init()
928 vfio.class->devnode = vfio_devnode; in vfio_group_init()