| /linux/arch/x86/kvm/ |
| A D | trace.h | 28 __entry->vcpu_id = vcpu->vcpu_id; 1066 __entry->vcpu_id = vcpu_id; 1097 __entry->vcpu_id = vcpu_id; 1127 __entry->vcpu_id = vcpu_id; 1149 __entry->vcpu_id = vcpu_id; 1175 __entry->vcpu_id = vcpu_id; 1197 __entry->vcpu_id = vcpu_id; 1222 __entry->vcpu_id = vcpu_id; 1248 __entry->vcpu_id = vcpu_id; 1274 __entry->vcpu_id = vcpu_id; [all …]
|
| A D | ioapic.c | 123 old_val = test_bit(vcpu->vcpu_id, dest_map->map); in __rtc_irq_eoi_tracking_restore_one() 129 __set_bit(vcpu->vcpu_id, dest_map->map); in __rtc_irq_eoi_tracking_restore_one() 130 dest_map->vectors[vcpu->vcpu_id] = e->fields.vector; in __rtc_irq_eoi_tracking_restore_one() 133 __clear_bit(vcpu->vcpu_id, dest_map->map); in __rtc_irq_eoi_tracking_restore_one() 167 if (test_bit(vcpu->vcpu_id, dest_map->map) && in rtc_irq_eoi() 168 (vector == dest_map->vectors[vcpu->vcpu_id]) && in rtc_irq_eoi() 169 (test_and_clear_bit(vcpu->vcpu_id, in rtc_irq_eoi() 288 if (test_bit(vcpu->vcpu_id, dest_map->map)) in kvm_ioapic_scan_entry() 289 __set_bit(dest_map->vectors[vcpu->vcpu_id], in kvm_ioapic_scan_entry()
|
| A D | xen.c | 113 e.vcpu_id = vcpu->vcpu_id; in kvm_xen_inject_timer_irqs() 135 e.vcpu_id = vcpu->vcpu_id; in xen_timer_callback() 574 irq.dest_id = v->vcpu_id; in kvm_xen_inject_vcpu_vector() 1058 if (data->u.vcpu_id >= KVM_MAX_VCPUS) in kvm_xen_vcpu_set_attr() 1061 vcpu->arch.xen.vcpu_id = data->u.vcpu_id; in kvm_xen_vcpu_set_attr() 1180 data->u.vcpu_id = vcpu->arch.xen.vcpu_id; in kvm_xen_vcpu_get_attr() 1551 int vcpu_id, u64 param, u64 *r) in kvm_xen_hcall_vcpu_op() argument 1561 if (vcpu->arch.xen.vcpu_id != vcpu_id) { in kvm_xen_hcall_vcpu_op() 1593 if (vcpu->arch.xen.vcpu_id != vcpu_id) { in kvm_xen_hcall_vcpu_op() 1964 e.vcpu_id = uxe->vcpu; in kvm_xen_hvm_evtchn_send() [all …]
|
| /linux/arch/powerpc/kvm/ |
| A D | trace_hv.h | 253 __entry->vcpu_id = vcpu->vcpu_id; 260 __entry->vcpu_id, 278 __entry->vcpu_id = vcpu->vcpu_id; 286 __entry->vcpu_id, 311 __entry->vcpu_id = vcpu->vcpu_id; 341 __entry->vcpu_id = vcpu->vcpu_id; 367 __entry->vcpu_id = vcpu->vcpu_id; 393 __entry->vcpu_id = vcpu->vcpu_id; 487 __entry->vcpu_id = vcpu->vcpu_id; 506 __entry->vcpu_id = vcpu->vcpu_id; [all …]
|
| A D | guest-state-buffer.c | 27 unsigned long vcpu_id, gfp_t flags) in kvmppc_gsb_new() argument 42 gsb->vcpu_id = vcpu_id; in kvmppc_gsb_new() 592 rc = plpar_guest_set_state(hflags, gsb->guest_id, gsb->vcpu_id, in kvmppc_gsb_send() 617 rc = plpar_guest_get_state(hflags, gsb->guest_id, gsb->vcpu_id, in kvmppc_gsb_recv()
|
| /linux/tools/testing/selftests/kvm/lib/x86_64/ |
| A D | memstress.c | 18 void memstress_l2_guest_code(uint64_t vcpu_id) in memstress_l2_guest_code() argument 20 memstress_guest_code(vcpu_id); in memstress_l2_guest_code() 32 static void memstress_l1_guest_code(struct vmx_pages *vmx, uint64_t vcpu_id) in memstress_l1_guest_code() argument 44 *rsp = vcpu_id; in memstress_l1_guest_code() 85 int vcpu_id; in memstress_setup_nested() local 90 for (vcpu_id = 0; vcpu_id < nr_vcpus; vcpu_id++) { in memstress_setup_nested() 93 if (vcpu_id == 0) { in memstress_setup_nested() 107 vcpu_regs_get(vcpus[vcpu_id], ®s); in memstress_setup_nested() 109 vcpu_regs_set(vcpus[vcpu_id], ®s); in memstress_setup_nested() 110 vcpu_args_set(vcpus[vcpu_id], 2, vmx_gva, vcpu_id); in memstress_setup_nested()
|
| /linux/arch/loongarch/kvm/ |
| A D | trace.h | 22 __field(unsigned int, vcpu_id) 27 __entry->vcpu_id = vcpu->vcpu_id; 31 TP_printk("vcpu %u PC: 0x%08lx", __entry->vcpu_id, __entry->pc) 59 __field(unsigned int, vcpu_id) 65 __entry->vcpu_id = vcpu->vcpu_id; 71 __entry->vcpu_id, 93 __field(unsigned int, vcpu_id) 98 __entry->vcpu_id = vcpu->vcpu_id; 102 TP_printk("vcpu %u Inst word: 0x%08x", __entry->vcpu_id,
|
| /linux/arch/arm64/kvm/vgic/ |
| A D | trace.h | 11 TP_PROTO(unsigned long vcpu_id, __u32 irq, bool level), 12 TP_ARGS(vcpu_id, irq, level), 15 __field( unsigned long, vcpu_id ) 21 __entry->vcpu_id = vcpu_id; 27 __entry->vcpu_id, __entry->irq, __entry->level)
|
| A D | vgic-debug.c | 30 int vcpu_id; member 60 ++iter->vcpu_id < iter->nr_cpus) in iter_next() 114 iter->vcpu_id == iter->nr_cpus && in end_of_vgic() 283 if (iter->vcpu_id < iter->nr_cpus) in vgic_debug_show() 284 vcpu = kvm_get_vcpu(kvm, iter->vcpu_id); in vgic_debug_show()
|
| /linux/samples/acrn/ |
| A D | vm-sample.c | 46 int vcpu_id, ret; in main() local 82 regs.vcpu_id = 0; in main() 108 for (vcpu_id = 0; vcpu_id < vcpu_num; vcpu_id++) { in main() 109 io_req = &io_req_buf[vcpu_id]; in main() 121 notify.vcpu = vcpu_id; in main()
|
| /linux/drivers/virt/nitro_enclaves/ |
| A D | ne_misc_dev.c | 528 int core_id, u32 vcpu_id) in ne_set_enclave_threads_per_core() argument 532 if (core_id < 0 && vcpu_id == 0) { in ne_set_enclave_threads_per_core() 588 *vcpu_id = cpu; in ne_get_cpu_from_cpu_pool() 658 if (ne_donated_cpu(ne_enclave, vcpu_id)) { in ne_check_cpu_in_cpu_pool() 660 "CPU %d already used\n", vcpu_id); in ne_check_cpu_in_cpu_pool() 717 slot_add_vcpu_req.vcpu_id = vcpu_id; in ne_add_vcpu_ioctl() 1152 u32 vcpu_id = 0; in ne_enclave_ioctl() local 1154 if (copy_from_user(&vcpu_id, (void __user *)arg, sizeof(vcpu_id))) in ne_enclave_ioctl() 1178 if (!vcpu_id) { in ne_enclave_ioctl() 1196 vcpu_id, rc); in ne_enclave_ioctl() [all …]
|
| /linux/include/trace/events/ |
| A D | kvm.h | 350 TP_ARGS(grow, vcpu_id, new, old), 354 __field(unsigned int, vcpu_id) 361 __entry->vcpu_id = vcpu_id; 367 __entry->vcpu_id, 373 #define trace_kvm_halt_poll_ns_grow(vcpu_id, new, old) \ argument 374 trace_kvm_halt_poll_ns(true, vcpu_id, new, old) 375 #define trace_kvm_halt_poll_ns_shrink(vcpu_id, new, old) \ argument 376 trace_kvm_halt_poll_ns(false, vcpu_id, new, old) 431 __field(int, vcpu_id) 435 __entry->vcpu_id = vcpu->vcpu_id; [all …]
|
| /linux/arch/powerpc/include/asm/ |
| A D | plpar_wrappers.h | 349 unsigned long vcpu_id) in plpar_guest_create_vcpu() argument 354 rc = plpar_hcall_norets(H_GUEST_CREATE_VCPU, 0, guest_id, vcpu_id); in plpar_guest_create_vcpu() 371 unsigned long vcpu_id, in plpar_guest_set_state() argument 381 vcpu_id, data_buffer, data_size); in plpar_guest_set_state() 408 unsigned long vcpu_id, in plpar_guest_get_state() argument 418 vcpu_id, data_buffer, data_size); in plpar_guest_get_state() 444 unsigned long vcpu_id, int *trap, in plpar_guest_run_vcpu() argument 591 unsigned long vcpu_id) in plpar_guest_create_vcpu() argument 598 unsigned long vcpu_id, in plpar_guest_get_state() argument 608 unsigned long vcpu_id, in plpar_guest_set_state() argument [all …]
|
| /linux/arch/arm64/kvm/ |
| A D | trace_arm.h | 207 TP_PROTO(unsigned long vcpu_id, __u32 irq, int level), 208 TP_ARGS(vcpu_id, irq, level), 211 __field( unsigned long, vcpu_id ) 217 __entry->vcpu_id = vcpu_id; 223 __entry->vcpu_id, __entry->irq, __entry->level) 227 TP_PROTO(unsigned long vcpu_id, struct timer_map *map), 228 TP_ARGS(vcpu_id, map), 231 __field( unsigned long, vcpu_id ) 239 __entry->vcpu_id = vcpu_id; 250 __entry->vcpu_id,
|
| /linux/arch/x86/kvm/svm/ |
| A D | avic.c | 50 ((vcpu_id) & AVIC_VCPU_ID_MASK)) 51 #define AVIC_GATAG(vm_id, vcpu_id) \ argument 53 u32 ga_tag = __AVIC_GATAG(vm_id, vcpu_id); \ 55 WARN_ON_ONCE(AVIC_GATAG_TO_VCPUID(ga_tag) != (vcpu_id)); \ 149 u32 vcpu_id = AVIC_GATAG_TO_VCPUID(ga_tag); in avic_ga_log_notifier() local 152 trace_kvm_avic_ga_log(vm_id, vcpu_id); in avic_ga_log_notifier() 158 vcpu = kvm_get_vcpu_by_id(&kvm_svm->kvm, vcpu_id); in avic_ga_log_notifier() 280 int id = vcpu->vcpu_id; in avic_init_backing_page() 806 u32 vcpu_id = AVIC_GATAG_TO_VCPUID(pi->prev_ga_tag); in svm_ir_list_add() local 940 svm->vcpu.vcpu_id); in avic_pi_update_irte() [all …]
|
| /linux/arch/s390/kvm/ |
| A D | sigp.c | 39 VCPU_EVENT(vcpu, 4, "sensed status of cpu %x rc %x", dst_vcpu->vcpu_id, in __sigp_sense() 49 .u.emerg.code = vcpu->vcpu_id, in __inject_sigp_emergency() 56 dst_vcpu->vcpu_id); in __inject_sigp_emergency() 98 .u.extcall.code = vcpu->vcpu_id, in __sigp_external_call() 109 dst_vcpu->vcpu_id); in __sigp_external_call() 127 dst_vcpu->vcpu_id); in __sigp_stop() 146 dst_vcpu->vcpu_id); in __sigp_stop_and_store_status() 235 dst_vcpu->vcpu_id, rc); in __sigp_sense_running() 370 order_code, dst_vcpu->vcpu_id); in handle_sigp_dst()
|
| A D | interrupt.c | 56 sca->cpu[vcpu->vcpu_id].sigp_ctrl; in sca_ext_call_pending() 63 sca->cpu[vcpu->vcpu_id].sigp_ctrl; in sca_ext_call_pending() 85 &(sca->cpu[vcpu->vcpu_id].sigp_ctrl); in sca_inject_ext_call() 98 &(sca->cpu[vcpu->vcpu_id].sigp_ctrl); in sca_inject_ext_call() 130 &(sca->cpu[vcpu->vcpu_id].sigp_ctrl); in sca_clear_ext_call() 139 &(sca->cpu[vcpu->vcpu_id].sigp_ctrl); in sca_clear_ext_call() 548 trace_kvm_s390_deliver_interrupt(vcpu->vcpu_id, in __deliver_pfault_init() 716 trace_kvm_s390_deliver_interrupt(vcpu->vcpu_id, in __deliver_machine_check() 759 trace_kvm_s390_deliver_interrupt(vcpu->vcpu_id, in __deliver_set_prefix() 815 trace_kvm_s390_deliver_interrupt(vcpu->vcpu_id, in __deliver_external_call() [all …]
|
| /linux/tools/testing/selftests/kvm/x86_64/ |
| A D | hyperv_ipi.c | 56 u32 vcpu_id; in receiver_code() local 61 vcpu_id = rdmsr(HV_X64_MSR_VP_INDEX); in receiver_code() 64 ipis_rcvd[vcpu_id] = (u64)-1; in receiver_code() 72 u32 vcpu_id = rdmsr(HV_X64_MSR_VP_INDEX); in guest_ipi_handler() local 74 ipis_rcvd[vcpu_id]++; in guest_ipi_handler()
|
| A D | tsc_scaling_sync.c | 45 unsigned long vcpu_id = (unsigned long)_cpu_nr; in run_vcpu() local 53 vcpu = vm_vcpu_add(vm, vcpu_id, guest_code); in run_vcpu()
|
| /linux/tools/perf/ |
| A D | builtin-kvm.c | 694 int vcpu_id; member 735 if (vcpu_id < event->max_vcpu) in kvm_event_expand() 738 while (event->max_vcpu <= vcpu_id) in kvm_event_expand() 841 if (vcpu_id != -1) in kvm_event_rel_stddev() 842 kvm_stats = &event->vcpu[vcpu_id]; in kvm_event_rel_stddev() 849 struct kvm_event *event, int vcpu_id, in update_kvm_event() argument 856 if (vcpu_id == -1) { in update_kvm_event() 861 if (!kvm_event_expand(event, vcpu_id)) in update_kvm_event() 928 vcpu = vcpu_record->vcpu_id; in handle_end_event() 969 sample->time, sample->pid, vcpu_record->vcpu_id, in handle_end_event() [all …]
|
| /linux/tools/testing/selftests/kvm/lib/aarch64/ |
| A D | processor.c | 375 static struct kvm_vcpu *__aarch64_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id, in __aarch64_vcpu_add() argument 380 struct kvm_vcpu *vcpu = __vm_vcpu_add(vm, vcpu_id); in __aarch64_vcpu_add() 394 struct kvm_vcpu *aarch64_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id, in aarch64_vcpu_add() argument 397 struct kvm_vcpu *vcpu = __aarch64_vcpu_add(vm, vcpu_id, init); in aarch64_vcpu_add() 404 struct kvm_vcpu *vm_arch_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id) in vm_arch_vcpu_add() argument 406 return __aarch64_vcpu_add(vm, vcpu_id, NULL); in vm_arch_vcpu_add()
|
| /linux/arch/x86/kvm/mmu/ |
| A D | mmutrace.h | 262 __field(int, vcpu_id) 272 __entry->vcpu_id = vcpu->vcpu_id; 282 " new %llx spurious %d fixed %d", __entry->vcpu_id,
|
| /linux/arch/riscv/kvm/ |
| A D | vcpu_sbi_replace.c | 63 if (tmp->vcpu_id < hbase) in kvm_sbi_ext_ipi_handler() 65 if (!(hmask & (1UL << (tmp->vcpu_id - hbase)))) in kvm_sbi_ext_ipi_handler()
|
| /linux/arch/mips/kvm/ |
| A D | stats.c | 54 kvm_info("\nKVM VCPU[%d] COP0 Access Profile:\n", vcpu->vcpu_id); in kvm_mips_dump_stats()
|
| /linux/include/xen/interface/ |
| A D | xenpmu.h | 73 uint32_t vcpu_id; member
|