Lines Matching refs:kv

121 	struct kvm_vfio *kv = dev->private;  in kvm_vfio_update_coherency()  local
125 mutex_lock(&kv->lock); in kvm_vfio_update_coherency()
127 list_for_each_entry(kvg, &kv->group_list, node) { in kvm_vfio_update_coherency()
134 if (noncoherent != kv->noncoherent) { in kvm_vfio_update_coherency()
135 kv->noncoherent = noncoherent; in kvm_vfio_update_coherency()
137 if (kv->noncoherent) in kvm_vfio_update_coherency()
143 mutex_unlock(&kv->lock); in kvm_vfio_update_coherency()
148 struct kvm_vfio *kv = dev->private; in kvm_vfio_group_add() local
163 mutex_lock(&kv->lock); in kvm_vfio_group_add()
165 list_for_each_entry(kvg, &kv->group_list, node) { in kvm_vfio_group_add()
179 list_add_tail(&kvg->node, &kv->group_list); in kvm_vfio_group_add()
183 mutex_unlock(&kv->lock); in kvm_vfio_group_add()
190 mutex_unlock(&kv->lock); in kvm_vfio_group_add()
198 struct kvm_vfio *kv = dev->private; in kvm_vfio_group_del() local
209 mutex_lock(&kv->lock); in kvm_vfio_group_del()
211 list_for_each_entry(kvg, &kv->group_list, node) { in kvm_vfio_group_del()
227 mutex_unlock(&kv->lock); in kvm_vfio_group_del()
241 struct kvm_vfio *kv = dev->private; in kvm_vfio_group_set_spapr_tce() local
255 mutex_lock(&kv->lock); in kvm_vfio_group_set_spapr_tce()
257 list_for_each_entry(kvg, &kv->group_list, node) { in kvm_vfio_group_set_spapr_tce()
275 mutex_unlock(&kv->lock); in kvm_vfio_group_set_spapr_tce()
341 struct kvm_vfio *kv = dev->private; in kvm_vfio_release() local
344 list_for_each_entry_safe(kvg, tmp, &kv->group_list, node) { in kvm_vfio_release()
357 kfree(kv); in kvm_vfio_release()
374 struct kvm_vfio *kv; in kvm_vfio_create() local
381 kv = kzalloc(sizeof(*kv), GFP_KERNEL_ACCOUNT); in kvm_vfio_create()
382 if (!kv) in kvm_vfio_create()
385 INIT_LIST_HEAD(&kv->group_list); in kvm_vfio_create()
386 mutex_init(&kv->lock); in kvm_vfio_create()
388 dev->private = kv; in kvm_vfio_create()