Lines Matching refs:group
26 static struct vfio_device *vfio_device_get_from_name(struct vfio_group *group, in vfio_device_get_from_name() argument
31 mutex_lock(&group->device_lock); in vfio_device_get_from_name()
32 list_for_each_entry(it, &group->device_list, group_next) { in vfio_device_get_from_name()
50 mutex_unlock(&group->device_lock); in vfio_device_get_from_name()
58 static bool vfio_group_has_iommu(struct vfio_group *group) in vfio_group_has_iommu() argument
60 lockdep_assert_held(&group->group_lock); in vfio_group_has_iommu()
65 WARN_ON(!group->container != !group->container_users); in vfio_group_has_iommu()
67 return group->container || group->iommufd; in vfio_group_has_iommu()
76 static int vfio_group_ioctl_unset_container(struct vfio_group *group) in vfio_group_ioctl_unset_container() argument
80 mutex_lock(&group->group_lock); in vfio_group_ioctl_unset_container()
81 if (!vfio_group_has_iommu(group)) { in vfio_group_ioctl_unset_container()
85 if (group->container) { in vfio_group_ioctl_unset_container()
86 if (group->container_users != 1) { in vfio_group_ioctl_unset_container()
90 vfio_group_detach_container(group); in vfio_group_ioctl_unset_container()
92 if (group->iommufd) { in vfio_group_ioctl_unset_container()
93 iommufd_ctx_put(group->iommufd); in vfio_group_ioctl_unset_container()
94 group->iommufd = NULL; in vfio_group_ioctl_unset_container()
98 mutex_unlock(&group->group_lock); in vfio_group_ioctl_unset_container()
102 static int vfio_group_ioctl_set_container(struct vfio_group *group, in vfio_group_ioctl_set_container() argument
118 mutex_lock(&group->group_lock); in vfio_group_ioctl_set_container()
119 if (vfio_group_has_iommu(group)) { in vfio_group_ioctl_set_container()
123 if (!group->iommu_group) { in vfio_group_ioctl_set_container()
130 ret = vfio_container_attach_group(container, group); in vfio_group_ioctl_set_container()
137 group->type == VFIO_NO_IOMMU) in vfio_group_ioctl_set_container()
147 group->iommufd = iommufd; in vfio_group_ioctl_set_container()
155 mutex_unlock(&group->group_lock); in vfio_group_ioctl_set_container()
162 spin_lock(&device->group->kvm_ref_lock); in vfio_device_group_get_kvm_safe()
163 if (!device->group->kvm) in vfio_device_group_get_kvm_safe()
166 _vfio_device_get_kvm_safe(device, device->group->kvm); in vfio_device_group_get_kvm_safe()
169 spin_unlock(&device->group->kvm_ref_lock); in vfio_device_group_get_kvm_safe()
176 mutex_lock(&device->group->group_lock); in vfio_device_group_open()
177 if (!vfio_group_has_iommu(device->group)) { in vfio_device_group_open()
193 ret = vfio_device_open(device, device->group->iommufd); in vfio_device_group_open()
201 mutex_unlock(&device->group->group_lock); in vfio_device_group_open()
207 mutex_lock(&device->group->group_lock); in vfio_device_group_close()
210 vfio_device_close(device, device->group->iommufd); in vfio_device_group_close()
216 mutex_unlock(&device->group->group_lock); in vfio_device_group_close()
246 if (device->group->type == VFIO_NO_IOMMU) in vfio_device_open_file()
261 static int vfio_group_ioctl_get_device_fd(struct vfio_group *group, in vfio_group_ioctl_get_device_fd() argument
274 device = vfio_device_get_from_name(group, buf); in vfio_group_ioctl_get_device_fd()
301 static int vfio_group_ioctl_get_status(struct vfio_group *group, in vfio_group_ioctl_get_status() argument
315 mutex_lock(&group->group_lock); in vfio_group_ioctl_get_status()
316 if (!group->iommu_group) { in vfio_group_ioctl_get_status()
317 mutex_unlock(&group->group_lock); in vfio_group_ioctl_get_status()
328 if (vfio_group_has_iommu(group)) in vfio_group_ioctl_get_status()
331 else if (!iommu_group_dma_owner_claimed(group->iommu_group)) in vfio_group_ioctl_get_status()
333 mutex_unlock(&group->group_lock); in vfio_group_ioctl_get_status()
343 struct vfio_group *group = filep->private_data; in vfio_group_fops_unl_ioctl() local
348 return vfio_group_ioctl_get_device_fd(group, uarg); in vfio_group_fops_unl_ioctl()
350 return vfio_group_ioctl_get_status(group, uarg); in vfio_group_fops_unl_ioctl()
352 return vfio_group_ioctl_set_container(group, uarg); in vfio_group_fops_unl_ioctl()
354 return vfio_group_ioctl_unset_container(group); in vfio_group_fops_unl_ioctl()
362 struct vfio_group *group = in vfio_group_fops_open() local
366 mutex_lock(&group->group_lock); in vfio_group_fops_open()
372 if (refcount_read(&group->drivers) == 0) { in vfio_group_fops_open()
377 if (group->type == VFIO_NO_IOMMU && !capable(CAP_SYS_RAWIO)) { in vfio_group_fops_open()
385 if (group->opened_file) { in vfio_group_fops_open()
389 group->opened_file = filep; in vfio_group_fops_open()
390 filep->private_data = group; in vfio_group_fops_open()
393 mutex_unlock(&group->group_lock); in vfio_group_fops_open()
399 struct vfio_group *group = filep->private_data; in vfio_group_fops_release() local
403 mutex_lock(&group->group_lock); in vfio_group_fops_release()
408 WARN_ON(group->notifier.head); in vfio_group_fops_release()
409 if (group->container) in vfio_group_fops_release()
410 vfio_group_detach_container(group); in vfio_group_fops_release()
411 if (group->iommufd) { in vfio_group_fops_release()
412 iommufd_ctx_put(group->iommufd); in vfio_group_fops_release()
413 group->iommufd = NULL; in vfio_group_fops_release()
415 group->opened_file = NULL; in vfio_group_fops_release()
416 mutex_unlock(&group->group_lock); in vfio_group_fops_release()
434 struct vfio_group *group; in vfio_group_find_from_iommu() local
442 list_for_each_entry(group, &vfio.group_list, vfio_next) { in vfio_group_find_from_iommu()
443 if (group->iommu_group == iommu_group) in vfio_group_find_from_iommu()
444 return group; in vfio_group_find_from_iommu()
451 struct vfio_group *group = container_of(dev, struct vfio_group, dev); in vfio_group_release() local
453 mutex_destroy(&group->device_lock); in vfio_group_release()
454 mutex_destroy(&group->group_lock); in vfio_group_release()
455 WARN_ON(group->iommu_group); in vfio_group_release()
456 ida_free(&vfio.group_ida, MINOR(group->dev.devt)); in vfio_group_release()
457 kfree(group); in vfio_group_release()
463 struct vfio_group *group; in vfio_group_alloc() local
466 group = kzalloc(sizeof(*group), GFP_KERNEL); in vfio_group_alloc()
467 if (!group) in vfio_group_alloc()
472 kfree(group); in vfio_group_alloc()
476 device_initialize(&group->dev); in vfio_group_alloc()
477 group->dev.devt = MKDEV(MAJOR(vfio.group_devt), minor); in vfio_group_alloc()
478 group->dev.class = vfio.class; in vfio_group_alloc()
479 group->dev.release = vfio_group_release; in vfio_group_alloc()
480 cdev_init(&group->cdev, &vfio_group_fops); in vfio_group_alloc()
481 group->cdev.owner = THIS_MODULE; in vfio_group_alloc()
483 refcount_set(&group->drivers, 1); in vfio_group_alloc()
484 mutex_init(&group->group_lock); in vfio_group_alloc()
485 spin_lock_init(&group->kvm_ref_lock); in vfio_group_alloc()
486 INIT_LIST_HEAD(&group->device_list); in vfio_group_alloc()
487 mutex_init(&group->device_lock); in vfio_group_alloc()
488 group->iommu_group = iommu_group; in vfio_group_alloc()
491 group->type = type; in vfio_group_alloc()
492 BLOCKING_INIT_NOTIFIER_HEAD(&group->notifier); in vfio_group_alloc()
494 return group; in vfio_group_alloc()
500 struct vfio_group *group; in vfio_create_group() local
506 group = vfio_group_alloc(iommu_group, type); in vfio_create_group()
507 if (IS_ERR(group)) in vfio_create_group()
508 return group; in vfio_create_group()
510 err = dev_set_name(&group->dev, "%s%d", in vfio_create_group()
511 group->type == VFIO_NO_IOMMU ? "noiommu-" : "", in vfio_create_group()
518 err = cdev_device_add(&group->cdev, &group->dev); in vfio_create_group()
524 list_add(&group->vfio_next, &vfio.group_list); in vfio_create_group()
526 return group; in vfio_create_group()
529 put_device(&group->dev); in vfio_create_group()
537 struct vfio_group *group; in vfio_noiommu_group_alloc() local
552 group = vfio_create_group(iommu_group, type); in vfio_noiommu_group_alloc()
554 if (IS_ERR(group)) { in vfio_noiommu_group_alloc()
555 ret = PTR_ERR(group); in vfio_noiommu_group_alloc()
559 return group; in vfio_noiommu_group_alloc()
568 static bool vfio_group_has_device(struct vfio_group *group, struct device *dev) in vfio_group_has_device() argument
572 mutex_lock(&group->device_lock); in vfio_group_has_device()
573 list_for_each_entry(device, &group->device_list, group_next) { in vfio_group_has_device()
575 mutex_unlock(&group->device_lock); in vfio_group_has_device()
579 mutex_unlock(&group->device_lock); in vfio_group_has_device()
586 struct vfio_group *group; in vfio_group_find_or_alloc() local
596 group = vfio_noiommu_group_alloc(dev, VFIO_NO_IOMMU); in vfio_group_find_or_alloc()
597 if (!IS_ERR(group)) { in vfio_group_find_or_alloc()
601 return group; in vfio_group_find_or_alloc()
618 group = vfio_group_find_from_iommu(iommu_group); in vfio_group_find_or_alloc()
619 if (group) { in vfio_group_find_or_alloc()
620 if (WARN_ON(vfio_group_has_device(group, dev))) in vfio_group_find_or_alloc()
621 group = ERR_PTR(-EINVAL); in vfio_group_find_or_alloc()
623 refcount_inc(&group->drivers); in vfio_group_find_or_alloc()
625 group = vfio_create_group(iommu_group, VFIO_IOMMU); in vfio_group_find_or_alloc()
631 return group; in vfio_group_find_or_alloc()
637 struct vfio_group *group; in vfio_device_set_group() local
640 group = vfio_group_find_or_alloc(device->dev); in vfio_device_set_group()
642 group = vfio_noiommu_group_alloc(device->dev, type); in vfio_device_set_group()
644 if (IS_ERR(group)) in vfio_device_set_group()
645 return PTR_ERR(group); in vfio_device_set_group()
648 device->group = group; in vfio_device_set_group()
654 struct vfio_group *group = device->group; in vfio_device_remove_group() local
657 if (group->type == VFIO_NO_IOMMU || group->type == VFIO_EMULATED_IOMMU) in vfio_device_remove_group()
661 if (!refcount_dec_and_mutex_lock(&group->drivers, &vfio.group_lock)) in vfio_device_remove_group()
663 list_del(&group->vfio_next); in vfio_device_remove_group()
671 cdev_device_del(&group->cdev, &group->dev); in vfio_device_remove_group()
673 mutex_lock(&group->group_lock); in vfio_device_remove_group()
680 WARN_ON(!list_empty(&group->device_list)); in vfio_device_remove_group()
681 WARN_ON(group->notifier.head); in vfio_device_remove_group()
688 if (group->container) in vfio_device_remove_group()
689 vfio_group_detach_container(group); in vfio_device_remove_group()
690 iommu_group = group->iommu_group; in vfio_device_remove_group()
691 group->iommu_group = NULL; in vfio_device_remove_group()
692 mutex_unlock(&group->group_lock); in vfio_device_remove_group()
696 put_device(&group->dev); in vfio_device_remove_group()
701 mutex_lock(&device->group->device_lock); in vfio_device_group_register()
702 list_add(&device->group_next, &device->group->device_list); in vfio_device_group_register()
703 mutex_unlock(&device->group->device_lock); in vfio_device_group_register()
708 mutex_lock(&device->group->device_lock); in vfio_device_group_unregister()
710 mutex_unlock(&device->group->device_lock); in vfio_device_group_unregister()
715 struct vfio_group *group = device->group; in vfio_device_group_use_iommu() local
718 lockdep_assert_held(&group->group_lock); in vfio_device_group_use_iommu()
720 if (WARN_ON(!group->container)) in vfio_device_group_use_iommu()
723 ret = vfio_group_use_container(group); in vfio_device_group_use_iommu()
732 struct vfio_group *group = device->group; in vfio_device_group_unuse_iommu() local
734 lockdep_assert_held(&group->group_lock); in vfio_device_group_unuse_iommu()
736 if (WARN_ON(!group->container)) in vfio_device_group_unuse_iommu()
740 vfio_group_unuse_container(group); in vfio_device_group_unuse_iommu()
745 return device->group->container; in vfio_device_has_container()
758 struct vfio_group *group = file->private_data; in vfio_file_iommu_group() local
767 mutex_lock(&group->group_lock); in vfio_file_iommu_group()
768 if (group->iommu_group) { in vfio_file_iommu_group()
769 iommu_group = group->iommu_group; in vfio_file_iommu_group()
772 mutex_unlock(&group->group_lock); in vfio_file_iommu_group()
798 struct vfio_group *group = file->private_data; in vfio_file_enforced_coherent() local
811 mutex_lock(&group->device_lock); in vfio_file_enforced_coherent()
812 list_for_each_entry(device, &group->device_list, group_next) { in vfio_file_enforced_coherent()
819 mutex_unlock(&group->device_lock); in vfio_file_enforced_coherent()
834 struct vfio_group *group = file->private_data; in vfio_file_set_kvm() local
839 spin_lock(&group->kvm_ref_lock); in vfio_file_set_kvm()
840 group->kvm = kvm; in vfio_file_set_kvm()
841 spin_unlock(&group->kvm_ref_lock); in vfio_file_set_kvm()
854 struct vfio_group *group = file->private_data; in vfio_file_has_dev() local
859 return group == device->group; in vfio_file_has_dev()