| /virt/kvm/ |
| A D | eventfd.c | 46 struct kvm *kvm = irqfd->kvm; in irqfd_inject() local 76 struct kvm *kvm; in irqfd_resampler_ack() local 81 kvm = resampler->kvm; in irqfd_resampler_ack() 95 struct kvm *kvm = resampler->kvm; in irqfd_resampler_shutdown() local 126 struct kvm *kvm = irqfd->kvm; in irqfd_shutdown() local 201 struct kvm *kvm = irqfd->kvm; in irqfd_wakeup() local 281 struct kvm *kvm; member 291 struct kvm *kvm = p->kvm; in kvm_irqfd_register() local 375 irqfd->kvm = kvm; in kvm_irqfd_assign() 427 resampler->kvm = kvm; in kvm_irqfd_assign() [all …]
|
| A D | kvm_main.c | 445 vcpu->kvm = kvm; in kvm_vcpu_init() 3695 struct kvm *kvm = vcpu->kvm; in kvm_vcpu_max_halt_poll_ns() local 3959 struct kvm *kvm = me->kvm; in kvm_vcpu_on_spin() local 4377 struct kvm *kvm = vcpu->kvm; in kvm_wait_for_vcpu_online() local 4741 struct kvm *kvm = dev->kvm; in kvm_device_release() local 4820 dev->kvm = kvm; in kvm_ioctl_create_device() 5471 struct kvm *kvm; in kvm_dev_ioctl_create_vm() local 6184 struct kvm *kvm; in vm_stat_get() local 6200 struct kvm *kvm; in vm_stat_clear() local 6220 struct kvm *kvm; in vcpu_stat_get() local [all …]
|
| A D | irqchip.c | 21 int kvm_irq_map_gsi(struct kvm *kvm, in kvm_irq_map_gsi() argument 28 irq_rt = srcu_dereference_check(kvm->irq_routing, &kvm->irq_srcu, in kvm_irq_map_gsi() 40 int kvm_irq_map_chip_pin(struct kvm *kvm, unsigned irqchip, unsigned pin) in kvm_irq_map_chip_pin() argument 44 irq_rt = srcu_dereference(kvm->irq_routing, &kvm->irq_srcu); in kvm_irq_map_chip_pin() 48 int kvm_send_userspace_msi(struct kvm *kvm, struct kvm_msi *msi) in kvm_send_userspace_msi() argument 70 int kvm_set_irq(struct kvm *kvm, int irq_source_id, u32 irq, int level, in kvm_set_irq() argument 119 void kvm_free_irq_routing(struct kvm *kvm) in kvm_free_irq_routing() argument 127 static int setup_routing_entry(struct kvm *kvm, in setup_routing_entry() argument 163 bool __weak kvm_arch_can_set_irq_routing(struct kvm *kvm) in kvm_arch_can_set_irq_routing() argument 168 int kvm_set_irq_routing(struct kvm *kvm, in kvm_set_irq_routing() argument [all …]
|
| A D | kvm_mm.h | 14 #define KVM_MMU_LOCK_INIT(kvm) rwlock_init(&(kvm)->mmu_lock) argument 15 #define KVM_MMU_LOCK(kvm) write_lock(&(kvm)->mmu_lock) argument 16 #define KVM_MMU_UNLOCK(kvm) write_unlock(&(kvm)->mmu_lock) argument 18 #define KVM_MMU_LOCK_INIT(kvm) spin_lock_init(&(kvm)->mmu_lock) argument 19 #define KVM_MMU_LOCK(kvm) spin_lock(&(kvm)->mmu_lock) argument 20 #define KVM_MMU_UNLOCK(kvm) spin_unlock(&(kvm)->mmu_lock) argument 59 void gfn_to_pfn_cache_invalidate_start(struct kvm *kvm, 63 static inline void gfn_to_pfn_cache_invalidate_start(struct kvm *kvm, in gfn_to_pfn_cache_invalidate_start() argument 72 int kvm_gmem_create(struct kvm *kvm, struct kvm_create_guest_memfd *args); 73 int kvm_gmem_bind(struct kvm *kvm, struct kvm_memory_slot *slot, [all …]
|
| A D | coalesced_mmio.c | 54 spin_lock(&dev->kvm->ring_lock); in coalesced_mmio_write() 95 int kvm_coalesced_mmio_init(struct kvm *kvm) in kvm_coalesced_mmio_init() argument 116 void kvm_coalesced_mmio_free(struct kvm *kvm) in kvm_coalesced_mmio_free() argument 118 if (kvm->coalesced_mmio_ring) in kvm_coalesced_mmio_free() 122 int kvm_vm_ioctl_register_coalesced_mmio(struct kvm *kvm, in kvm_vm_ioctl_register_coalesced_mmio() argument 137 dev->kvm = kvm; in kvm_vm_ioctl_register_coalesced_mmio() 140 mutex_lock(&kvm->slots_lock); in kvm_vm_ioctl_register_coalesced_mmio() 147 mutex_unlock(&kvm->slots_lock); in kvm_vm_ioctl_register_coalesced_mmio() 152 mutex_unlock(&kvm->slots_lock); in kvm_vm_ioctl_register_coalesced_mmio() 158 int kvm_vm_ioctl_unregister_coalesced_mmio(struct kvm *kvm, in kvm_vm_ioctl_unregister_coalesced_mmio() argument [all …]
|
| A D | coalesced_mmio.h | 21 struct kvm *kvm; member 25 int kvm_coalesced_mmio_init(struct kvm *kvm); 26 void kvm_coalesced_mmio_free(struct kvm *kvm); 27 int kvm_vm_ioctl_register_coalesced_mmio(struct kvm *kvm, 29 int kvm_vm_ioctl_unregister_coalesced_mmio(struct kvm *kvm, 34 static inline int kvm_coalesced_mmio_init(struct kvm *kvm) { return 0; } in kvm_coalesced_mmio_init() argument 35 static inline void kvm_coalesced_mmio_free(struct kvm *kvm) { } in kvm_coalesced_mmio_free() argument
|
| A D | dirty_ring.c | 14 int __weak kvm_cpu_dirty_log_size(struct kvm *kvm) in kvm_cpu_dirty_log_size() argument 19 u32 kvm_dirty_ring_get_rsvd_entries(struct kvm *kvm) in kvm_dirty_ring_get_rsvd_entries() argument 24 bool kvm_use_dirty_bitmap(struct kvm *kvm) in kvm_use_dirty_bitmap() argument 26 lockdep_assert_held(&kvm->slots_lock); in kvm_use_dirty_bitmap() 28 return !kvm->dirty_ring_size || kvm->dirty_ring_with_bitmap; in kvm_use_dirty_bitmap() 32 bool kvm_arch_allow_write_without_running_vcpu(struct kvm *kvm) in kvm_arch_allow_write_without_running_vcpu() argument 53 static void kvm_reset_dirty_gfn(struct kvm *kvm, u32 slot, u64 offset, u64 mask) in kvm_reset_dirty_gfn() argument 69 KVM_MMU_LOCK(kvm); in kvm_reset_dirty_gfn() 71 KVM_MMU_UNLOCK(kvm); in kvm_reset_dirty_gfn() 74 int kvm_dirty_ring_alloc(struct kvm *kvm, struct kvm_dirty_ring *ring, in kvm_dirty_ring_alloc() argument [all …]
|
| A D | guest_memfd.c | 11 struct kvm *kvm; member 110 struct kvm *kvm = gmem->kvm; in kvm_gmem_invalidate_begin() local 145 struct kvm *kvm = gmem->kvm; in kvm_gmem_invalidate_end() local 148 KVM_MMU_LOCK(kvm); in kvm_gmem_invalidate_end() 255 struct kvm *kvm = gmem->kvm; in kvm_gmem_release() local 295 kvm_put_kvm(kvm); in kvm_gmem_release() 434 kvm_get_kvm(kvm); in __kvm_gmem_create() 435 gmem->kvm = kvm; in __kvm_gmem_create() 464 int kvm_gmem_bind(struct kvm *kvm, struct kvm_memory_slot *slot, in kvm_gmem_bind() argument 484 if (gmem->kvm != kvm) in kvm_gmem_bind() [all …]
|
| A D | pfncache.c | 25 void gfn_to_pfn_cache_invalidate_start(struct kvm *kvm, unsigned long start, in gfn_to_pfn_cache_invalidate_start() argument 30 spin_lock(&kvm->gpc_lock); in gfn_to_pfn_cache_invalidate_start() 57 spin_unlock(&kvm->gpc_lock); in gfn_to_pfn_cache_invalidate_start() 127 static inline bool mmu_notifier_retry_cache(struct kvm *kvm, unsigned long mmu_seq) in mmu_notifier_retry_cache() argument 385 void kvm_gpc_init(struct gfn_to_pfn_cache *gpc, struct kvm *kvm) in kvm_gpc_init() argument 390 gpc->kvm = kvm; in kvm_gpc_init() 400 struct kvm *kvm = gpc->kvm; in __kvm_gpc_activate() local 411 spin_lock(&kvm->gpc_lock); in __kvm_gpc_activate() 413 spin_unlock(&kvm->gpc_lock); in __kvm_gpc_activate() 449 struct kvm *kvm = gpc->kvm; in kvm_gpc_deactivate() local [all …]
|
| A D | Makefile.kvm | 6 KVM ?= ../../../virt/kvm 8 kvm-y := $(KVM)/kvm_main.o $(KVM)/eventfd.o $(KVM)/binary_stats.o 9 kvm-$(CONFIG_KVM_VFIO) += $(KVM)/vfio.o 10 kvm-$(CONFIG_KVM_MMIO) += $(KVM)/coalesced_mmio.o 11 kvm-$(CONFIG_KVM_ASYNC_PF) += $(KVM)/async_pf.o 12 kvm-$(CONFIG_HAVE_KVM_IRQ_ROUTING) += $(KVM)/irqchip.o 13 kvm-$(CONFIG_HAVE_KVM_DIRTY_RING) += $(KVM)/dirty_ring.o 14 kvm-$(CONFIG_HAVE_KVM_PFNCACHE) += $(KVM)/pfncache.o 15 kvm-$(CONFIG_KVM_PRIVATE_MEM) += $(KVM)/guest_memfd.o
|
| A D | vfio.c | 38 static void kvm_vfio_file_set_kvm(struct file *file, struct kvm *kvm) in kvm_vfio_file_set_kvm() argument 40 void (*fn)(struct file *file, struct kvm *kvm); in kvm_vfio_file_set_kvm() 46 fn(file, kvm); in kvm_vfio_file_set_kvm() 100 static void kvm_spapr_tce_release_vfio_group(struct kvm *kvm, in kvm_spapr_tce_release_vfio_group() argument 137 kvm_arch_register_noncoherent_dma(dev->kvm); in kvm_vfio_update_coherency() 139 kvm_arch_unregister_noncoherent_dma(dev->kvm); in kvm_vfio_update_coherency() 178 kvm_vfio_file_set_kvm(kvf->file, dev->kvm); in kvm_vfio_file_add() 208 kvm_spapr_tce_release_vfio_group(dev->kvm, kvf); in kvm_vfio_file_del() 331 kvm_spapr_tce_release_vfio_group(dev->kvm, kvf); in kvm_vfio_release() 360 lockdep_assert_held(&dev->kvm->lock); in kvm_vfio_create() [all …]
|
| A D | async_pf.c | 50 struct mm_struct *mm = vcpu->kvm->mm; in async_pf_execute()
|