| /arch/arm64/kvm/ |
| A D | trace_arm.h | 94 TP_PROTO(unsigned int type, int vcpu_idx, int irq_num, int level), 95 TP_ARGS(type, vcpu_idx, irq_num, level), 99 __field( int, vcpu_idx ) 106 __entry->vcpu_idx = vcpu_idx; 115 __entry->type, __entry->vcpu_idx, __entry->irq_num, __entry->level)
|
| A D | arm.c | 607 if (*last_ran != vcpu->vcpu_idx) { in kvm_arch_vcpu_load() 609 *last_ran = vcpu->vcpu_idx; in kvm_arch_vcpu_load() 644 vcpu->vcpu_idx, vcpu->arch.hcr_el2); in kvm_arch_vcpu_load()
|
| /arch/x86/kvm/svm/ |
| A D | avic.c | 54 #define __AVIC_GATAG(vm_id, vcpu_idx) ((((vm_id) & AVIC_VM_ID_MASK) << AVIC_VM_ID_SHIFT) | \ argument 55 ((vcpu_idx) & AVIC_VCPU_IDX_MASK)) 56 #define AVIC_GATAG(vm_id, vcpu_idx) \ argument 58 u32 ga_tag = __AVIC_GATAG(vm_id, vcpu_idx); \ 60 WARN_ON_ONCE(AVIC_GATAG_TO_VCPUIDX(ga_tag) != (vcpu_idx)); \ 146 u32 vcpu_idx = AVIC_GATAG_TO_VCPUIDX(ga_tag); in avic_ga_log_notifier() local 148 pr_debug("SVM: %s: vm_id=%#x, vcpu_idx=%#x\n", __func__, vm_id, vcpu_idx); in avic_ga_log_notifier() 149 trace_kvm_avic_ga_log(vm_id, vcpu_idx); in avic_ga_log_notifier() 155 vcpu = kvm_get_vcpu(&kvm_svm->kvm, vcpu_idx); in avic_ga_log_notifier() 764 vcpu->vcpu_idx), in avic_pi_update_irte()
|
| /arch/x86/kvm/ |
| A D | xen.c | 114 e.vcpu_idx = vcpu->vcpu_idx; in kvm_xen_inject_timer_irqs() 136 e.vcpu_idx = vcpu->vcpu_idx; in xen_timer_callback() 1801 int vcpu_idx, idx, rc; in kvm_xen_set_evtchn_fast() local 1803 vcpu_idx = READ_ONCE(xe->vcpu_idx); in kvm_xen_set_evtchn_fast() 1804 if (vcpu_idx >= 0) in kvm_xen_set_evtchn_fast() 1805 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in kvm_xen_set_evtchn_fast() 1810 WRITE_ONCE(xe->vcpu_idx, vcpu->vcpu_idx); in kvm_xen_set_evtchn_fast() 2002 e->xen_evtchn.vcpu_idx = vcpu->vcpu_idx; in kvm_xen_setup_evtchn() 2004 e->xen_evtchn.vcpu_idx = -1; in kvm_xen_setup_evtchn() 2031 e.vcpu_idx = -1; in kvm_xen_hvm_evtchn_send() [all …]
|
| A D | hyperv.h | 92 return hv_vcpu ? hv_vcpu->vp_index : vcpu->vcpu_idx; in kvm_hv_get_vpindex() 322 return vcpu->vcpu_idx; in kvm_hv_get_vpindex()
|
| A D | lapic.c | 1372 int vcpu_idx; in kvm_bitmap_or_dest_vcpus() local 1384 vcpu_idx = dest_vcpu[i]->vcpu->vcpu_idx; in kvm_bitmap_or_dest_vcpus() 1385 __set_bit(vcpu_idx, vcpu_bitmap); in kvm_bitmap_or_dest_vcpus()
|
| A D | hyperv.c | 984 hv_vcpu->vp_index = vcpu->vcpu_idx; in kvm_hv_vcpu_init() 1545 if (hv_vcpu->vp_index == vcpu->vcpu_idx) in kvm_hv_set_msr() 1547 else if (new_vp_index == vcpu->vcpu_idx) in kvm_hv_set_msr()
|
| A D | x86.c | 12343 if (kvmclock_periodic_sync && vcpu->vcpu_idx == 0) in kvm_arch_vcpu_postcreate()
|
| /arch/arm64/kvm/hyp/nvhe/ |
| A D | pkvm.c | 238 unsigned int vcpu_idx) in pkvm_load_hyp_vcpu() argument 249 if (!hyp_vm || hyp_vm->kvm.created_vcpus <= vcpu_idx) in pkvm_load_hyp_vcpu() 252 hyp_vcpu = hyp_vm->vcpus[vcpu_idx]; in pkvm_load_hyp_vcpu() 463 hyp_vcpu->vcpu.vcpu_idx = READ_ONCE(host_vcpu->vcpu_idx); in init_pkvm_hyp_vcpu() 705 idx = hyp_vcpu->vcpu.vcpu_idx; in __pkvm_init_vcpu()
|
| A D | hyp-main.c | 167 DECLARE_REG(unsigned int, vcpu_idx, host_ctxt, 2); in handle___pkvm_vcpu_load() 174 hyp_vcpu = pkvm_load_hyp_vcpu(handle, vcpu_idx); in handle___pkvm_vcpu_load()
|
| /arch/arm64/kvm/hyp/include/nvhe/ |
| A D | pkvm.h | 77 unsigned int vcpu_idx);
|
| /arch/arm64/kvm/vgic/ |
| A D | vgic-debug.c | 206 id = vcpu->vcpu_idx; in print_header() 252 (irq->target_vcpu) ? irq->target_vcpu->vcpu_idx : -1, in print_irq_state() 264 (irq->vcpu) ? irq->vcpu->vcpu_idx : -1); in print_irq_state()
|
| A D | vgic.c | 446 trace_vgic_update_irq_pending(vcpu ? vcpu->vcpu_idx : 0, intid, level); in kvm_vgic_inject_irq()
|
| /arch/riscv/kvm/ |
| A D | aia_device.c | 161 unsigned long vcpu_idx, bool write) in aia_imsic_addr() argument 166 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in aia_imsic_addr() 528 vaia->hart_index = vcpu->vcpu_idx; in kvm_riscv_vcpu_aia_init()
|
| A D | vcpu.c | 184 if (vcpu->vcpu_idx != 0) in kvm_arch_vcpu_postcreate()
|
| /arch/s390/kvm/ |
| A D | interrupt.c | 411 set_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in __set_cpu_idle() 417 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in __unset_cpu_idle() 3071 int vcpu_idx, online_vcpus = atomic_read(&kvm->online_vcpus); in __airqs_kick_single_vcpu() local 3076 for_each_set_bit(vcpu_idx, kvm->arch.idle_mask, online_vcpus) { in __airqs_kick_single_vcpu() 3077 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in __airqs_kick_single_vcpu() 3083 if (test_and_set_bit(vcpu_idx, gi->kicked_mask)) in __airqs_kick_single_vcpu()
|
| A D | kvm-s390.h | 103 return test_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in is_vcpu_idle()
|
| A D | kvm-s390.c | 3623 unsigned long vcpu_idx; in sca_switch_to_extended() local 3643 kvm_for_each_vcpu(vcpu_idx, vcpu, kvm) { in sca_switch_to_extended() 4054 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.gisa_int.kicked_mask); in kvm_arch_vcpu_runnable() 4808 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.gisa_int.kicked_mask); in vcpu_pre_run()
|