Lines Matching full:group
142 * Group objects - create, release, get, put, search
147 struct vfio_group *group; in __vfio_group_get_from_iommu() local
150 * group->iommu_group from the vfio.group_list cannot be NULL in __vfio_group_get_from_iommu()
153 list_for_each_entry(group, &vfio.group_list, vfio_next) { in __vfio_group_get_from_iommu()
154 if (group->iommu_group == iommu_group) { in __vfio_group_get_from_iommu()
155 refcount_inc(&group->drivers); in __vfio_group_get_from_iommu()
156 return group; in __vfio_group_get_from_iommu()
165 struct vfio_group *group; in vfio_group_get_from_iommu() local
168 group = __vfio_group_get_from_iommu(iommu_group); in vfio_group_get_from_iommu()
170 return group; in vfio_group_get_from_iommu()
175 struct vfio_group *group = container_of(dev, struct vfio_group, dev); in vfio_group_release() local
177 mutex_destroy(&group->device_lock); in vfio_group_release()
178 mutex_destroy(&group->group_lock); in vfio_group_release()
179 WARN_ON(group->iommu_group); in vfio_group_release()
180 ida_free(&vfio.group_ida, MINOR(group->dev.devt)); in vfio_group_release()
181 kfree(group); in vfio_group_release()
187 struct vfio_group *group; in vfio_group_alloc() local
190 group = kzalloc(sizeof(*group), GFP_KERNEL); in vfio_group_alloc()
191 if (!group) in vfio_group_alloc()
196 kfree(group); in vfio_group_alloc()
200 device_initialize(&group->dev); in vfio_group_alloc()
201 group->dev.devt = MKDEV(MAJOR(vfio.group_devt), minor); in vfio_group_alloc()
202 group->dev.class = vfio.class; in vfio_group_alloc()
203 group->dev.release = vfio_group_release; in vfio_group_alloc()
204 cdev_init(&group->cdev, &vfio_group_fops); in vfio_group_alloc()
205 group->cdev.owner = THIS_MODULE; in vfio_group_alloc()
207 refcount_set(&group->drivers, 1); in vfio_group_alloc()
208 mutex_init(&group->group_lock); in vfio_group_alloc()
209 INIT_LIST_HEAD(&group->device_list); in vfio_group_alloc()
210 mutex_init(&group->device_lock); in vfio_group_alloc()
211 group->iommu_group = iommu_group; in vfio_group_alloc()
214 group->type = type; in vfio_group_alloc()
215 BLOCKING_INIT_NOTIFIER_HEAD(&group->notifier); in vfio_group_alloc()
217 return group; in vfio_group_alloc()
223 struct vfio_group *group; in vfio_create_group() local
227 group = vfio_group_alloc(iommu_group, type); in vfio_create_group()
228 if (IS_ERR(group)) in vfio_create_group()
229 return group; in vfio_create_group()
231 err = dev_set_name(&group->dev, "%s%d", in vfio_create_group()
232 group->type == VFIO_NO_IOMMU ? "noiommu-" : "", in vfio_create_group()
241 /* Did we race creating this group? */ in vfio_create_group()
246 err = cdev_device_add(&group->cdev, &group->dev); in vfio_create_group()
252 list_add(&group->vfio_next, &vfio.group_list); in vfio_create_group()
255 return group; in vfio_create_group()
260 put_device(&group->dev); in vfio_create_group()
266 struct vfio_group *group = device->group; in vfio_device_remove_group() local
269 if (group->type == VFIO_NO_IOMMU || group->type == VFIO_EMULATED_IOMMU) in vfio_device_remove_group()
273 if (!refcount_dec_and_mutex_lock(&group->drivers, &vfio.group_lock)) in vfio_device_remove_group()
275 list_del(&group->vfio_next); in vfio_device_remove_group()
278 * We could concurrently probe another driver in the group that might in vfio_device_remove_group()
283 cdev_device_del(&group->cdev, &group->dev); in vfio_device_remove_group()
285 mutex_lock(&group->group_lock); in vfio_device_remove_group()
290 * properly hold the group reference. in vfio_device_remove_group()
292 WARN_ON(!list_empty(&group->device_list)); in vfio_device_remove_group()
293 WARN_ON(group->notifier.head); in vfio_device_remove_group()
296 * Revoke all users of group->iommu_group. At this point we know there in vfio_device_remove_group()
300 if (group->container) in vfio_device_remove_group()
301 vfio_group_detach_container(group); in vfio_device_remove_group()
302 iommu_group = group->iommu_group; in vfio_device_remove_group()
303 group->iommu_group = NULL; in vfio_device_remove_group()
304 mutex_unlock(&group->group_lock); in vfio_device_remove_group()
308 put_device(&group->dev); in vfio_device_remove_group()
314 /* Device reference always implies a group reference */
326 static struct vfio_device *vfio_group_get_device(struct vfio_group *group, in vfio_group_get_device() argument
331 mutex_lock(&group->device_lock); in vfio_group_get_device()
332 list_for_each_entry(device, &group->device_list, group_next) { in vfio_group_get_device()
335 mutex_unlock(&group->device_lock); in vfio_group_get_device()
339 mutex_unlock(&group->device_lock); in vfio_group_get_device()
455 struct vfio_group *group; in vfio_noiommu_group_alloc() local
469 group = vfio_create_group(iommu_group, type); in vfio_noiommu_group_alloc()
470 if (IS_ERR(group)) { in vfio_noiommu_group_alloc()
471 ret = PTR_ERR(group); in vfio_noiommu_group_alloc()
475 return group; in vfio_noiommu_group_alloc()
487 struct vfio_group *group; in vfio_group_find_or_alloc() local
492 * With noiommu enabled, create an IOMMU group for devices that in vfio_group_find_or_alloc()
497 group = vfio_noiommu_group_alloc(dev, VFIO_NO_IOMMU); in vfio_group_find_or_alloc()
498 if (!IS_ERR(group)) { in vfio_group_find_or_alloc()
500 dev_warn(dev, "Adding kernel taint for vfio-noiommu group on device\n"); in vfio_group_find_or_alloc()
502 return group; in vfio_group_find_or_alloc()
518 group = vfio_group_get_from_iommu(iommu_group); in vfio_group_find_or_alloc()
519 if (!group) in vfio_group_find_or_alloc()
520 group = vfio_create_group(iommu_group, VFIO_IOMMU); in vfio_group_find_or_alloc()
524 return group; in vfio_group_find_or_alloc()
528 struct vfio_group *group) in __vfio_register_dev() argument
534 * In all cases group is the output of one of the group allocation in __vfio_register_dev()
535 * functions and we have group->drivers incremented for us. in __vfio_register_dev()
537 if (IS_ERR(group)) in __vfio_register_dev()
538 return PTR_ERR(group); in __vfio_register_dev()
547 existing_device = vfio_group_get_device(group, device->dev); in __vfio_register_dev()
550 * group->iommu_group is non-NULL because we hold the drivers in __vfio_register_dev()
553 dev_WARN(device->dev, "Device already exists on group %d\n", in __vfio_register_dev()
554 iommu_group_id(group->iommu_group)); in __vfio_register_dev()
560 /* Our reference on group is moved to the device */ in __vfio_register_dev()
561 device->group = group; in __vfio_register_dev()
574 mutex_lock(&group->device_lock); in __vfio_register_dev()
575 list_add(&device->group_next, &group->device_list); in __vfio_register_dev()
576 mutex_unlock(&group->device_lock); in __vfio_register_dev()
602 static struct vfio_device *vfio_device_get_from_name(struct vfio_group *group, in vfio_device_get_from_name() argument
607 mutex_lock(&group->device_lock); in vfio_device_get_from_name()
608 list_for_each_entry(it, &group->device_list, group_next) { in vfio_device_get_from_name()
626 mutex_unlock(&group->device_lock); in vfio_device_get_from_name()
636 struct vfio_group *group = device->group; in vfio_unregister_group_dev() local
664 mutex_lock(&group->device_lock); in vfio_unregister_group_dev()
666 mutex_unlock(&group->device_lock); in vfio_unregister_group_dev()
676 * VFIO Group fd, /dev/vfio/$GROUP
681 * the group, we know that still exists, therefore the only valid
684 static int vfio_group_ioctl_unset_container(struct vfio_group *group) in vfio_group_ioctl_unset_container() argument
688 mutex_lock(&group->group_lock); in vfio_group_ioctl_unset_container()
689 if (!group->container) { in vfio_group_ioctl_unset_container()
693 if (group->container_users != 1) { in vfio_group_ioctl_unset_container()
697 vfio_group_detach_container(group); in vfio_group_ioctl_unset_container()
700 mutex_unlock(&group->group_lock); in vfio_group_ioctl_unset_container()
704 static int vfio_group_ioctl_set_container(struct vfio_group *group, in vfio_group_ioctl_set_container() argument
719 mutex_lock(&group->group_lock); in vfio_group_ioctl_set_container()
720 if (group->container || WARN_ON(group->container_users)) { in vfio_group_ioctl_set_container()
724 if (!group->iommu_group) { in vfio_group_ioctl_set_container()
732 ret = vfio_container_attach_group(container, group); in vfio_group_ioctl_set_container()
737 mutex_unlock(&group->group_lock); in vfio_group_ioctl_set_container()
755 mutex_lock(&device->group->group_lock); in vfio_device_open()
757 mutex_unlock(&device->group->group_lock); in vfio_device_open()
770 * Here we pass the KVM pointer with the group under the read in vfio_device_open()
774 mutex_lock(&device->group->group_lock); in vfio_device_open()
775 device->kvm = device->group->kvm; in vfio_device_open()
783 mutex_unlock(&device->group->group_lock); in vfio_device_open()
805 if (device->group->type == VFIO_NO_IOMMU) in vfio_device_open()
816 mutex_lock(&device->group->group_lock); in vfio_device_open()
824 mutex_unlock(&device->group->group_lock); in vfio_device_open()
835 static int vfio_group_ioctl_get_device_fd(struct vfio_group *group, in vfio_group_ioctl_get_device_fd() argument
848 device = vfio_device_get_from_name(group, buf); in vfio_group_ioctl_get_device_fd()
875 static int vfio_group_ioctl_get_status(struct vfio_group *group, in vfio_group_ioctl_get_status() argument
889 mutex_lock(&group->group_lock); in vfio_group_ioctl_get_status()
890 if (!group->iommu_group) { in vfio_group_ioctl_get_status()
891 mutex_unlock(&group->group_lock); in vfio_group_ioctl_get_status()
895 if (group->container) in vfio_group_ioctl_get_status()
898 else if (!iommu_group_dma_owner_claimed(group->iommu_group)) in vfio_group_ioctl_get_status()
900 mutex_unlock(&group->group_lock); in vfio_group_ioctl_get_status()
910 struct vfio_group *group = filep->private_data; in vfio_group_fops_unl_ioctl() local
915 return vfio_group_ioctl_get_device_fd(group, uarg); in vfio_group_fops_unl_ioctl()
917 return vfio_group_ioctl_get_status(group, uarg); in vfio_group_fops_unl_ioctl()
919 return vfio_group_ioctl_set_container(group, uarg); in vfio_group_fops_unl_ioctl()
921 return vfio_group_ioctl_unset_container(group); in vfio_group_fops_unl_ioctl()
929 struct vfio_group *group = in vfio_group_fops_open() local
933 mutex_lock(&group->group_lock); in vfio_group_fops_open()
937 * will be stable at 0 under the group rwsem in vfio_group_fops_open()
939 if (refcount_read(&group->drivers) == 0) { in vfio_group_fops_open()
944 if (group->type == VFIO_NO_IOMMU && !capable(CAP_SYS_RAWIO)) { in vfio_group_fops_open()
950 * Do we need multiple instances of the group open? Seems not. in vfio_group_fops_open()
952 if (group->opened_file) { in vfio_group_fops_open()
956 group->opened_file = filep; in vfio_group_fops_open()
957 filep->private_data = group; in vfio_group_fops_open()
960 mutex_unlock(&group->group_lock); in vfio_group_fops_open()
966 struct vfio_group *group = filep->private_data; in vfio_group_fops_release() local
970 mutex_lock(&group->group_lock); in vfio_group_fops_release()
972 * Device FDs hold a group file reference, therefore the group release in vfio_group_fops_release()
975 WARN_ON(group->notifier.head); in vfio_group_fops_release()
976 if (group->container) in vfio_group_fops_release()
977 vfio_group_detach_container(group); in vfio_group_fops_release()
978 group->opened_file = NULL; in vfio_group_fops_release()
979 mutex_unlock(&group->group_lock); in vfio_group_fops_release()
1033 mutex_lock(&device->group->group_lock); in vfio_device_fops_release()
1040 mutex_unlock(&device->group->group_lock); in vfio_device_fops_release()
1582 * vfio_file_iommu_group - Return the struct iommu_group for the vfio group file
1583 * @file: VFIO group file
1586 * returns a reference on the group. This function is deprecated, only the SPAPR
1591 struct vfio_group *group = file->private_data; in vfio_file_iommu_group() local
1600 mutex_lock(&group->group_lock); in vfio_file_iommu_group()
1601 if (group->iommu_group) { in vfio_file_iommu_group()
1602 iommu_group = group->iommu_group; in vfio_file_iommu_group()
1605 mutex_unlock(&group->group_lock); in vfio_file_iommu_group()
1612 * @file: VFIO group file
1623 * @file: VFIO group file
1631 struct vfio_group *group = file->private_data; in vfio_file_enforced_coherent() local
1637 mutex_lock(&group->group_lock); in vfio_file_enforced_coherent()
1638 if (group->container) { in vfio_file_enforced_coherent()
1639 ret = vfio_container_ioctl_check_extension(group->container, in vfio_file_enforced_coherent()
1649 mutex_unlock(&group->group_lock); in vfio_file_enforced_coherent()
1656 * @file: VFIO group file
1660 * device->kvm if one was associated with the group.
1664 struct vfio_group *group = file->private_data; in vfio_file_set_kvm() local
1669 mutex_lock(&group->group_lock); in vfio_file_set_kvm()
1670 group->kvm = kvm; in vfio_file_set_kvm()
1671 mutex_unlock(&group->group_lock); in vfio_file_set_kvm()
1684 struct vfio_group *group = file->private_data; in vfio_file_has_dev() local
1689 return group == device->group; in vfio_file_has_dev()
1833 /* /dev/vfio/$GROUP */ in vfio_init()