| /linux/tools/testing/selftests/kvm/ |
| A D | arch_timer.c | 49 unsigned int vcpu_idx = (unsigned long)arg; in test_vcpu_run() local 51 struct kvm_vcpu *vcpu = vcpus[vcpu_idx]; in test_vcpu_run() 53 struct test_vcpu_shared_data *shared_data = &vcpu_shared_data[vcpu_idx]; in test_vcpu_run() 59 __set_bit(vcpu_idx, vcpu_done_map); in test_vcpu_run() 69 vcpu_idx, shared_data->guest_stage, shared_data->nr_iter); in test_vcpu_run() 76 pr_info("PASS(vCPU-%d).\n", vcpu_idx); in test_vcpu_run() 98 static int test_migrate_vcpu(unsigned int vcpu_idx) in test_migrate_vcpu() argument 107 pr_debug("Migrating vCPU: %u to pCPU: %u\n", vcpu_idx, new_pcpu); in test_migrate_vcpu() 109 ret = pthread_setaffinity_np(pt_vcpu_run[vcpu_idx], in test_migrate_vcpu() 115 vcpu_idx, new_pcpu, ret); in test_migrate_vcpu()
|
| A D | access_tracking_perf_test.c | 129 int vcpu_idx = vcpu_args->vcpu_idx; in mark_vcpu_memory_idle() local 139 if (overlap_memory_access && vcpu_idx) in mark_vcpu_memory_idle() 171 vcpu_idx, no_pfn, pages); in mark_vcpu_memory_idle() 190 vcpu_idx, still_idle, pages); in mark_vcpu_memory_idle() 194 vcpu_idx, still_idle, pages); in mark_vcpu_memory_idle() 230 int vcpu_idx = vcpu_args->vcpu_idx; in vcpu_thread_main() local 244 vcpu_last_completed_iteration[vcpu_idx] = current_iteration; in vcpu_thread_main() 248 static void spin_wait_for_vcpu(int vcpu_idx, int target_iteration) in spin_wait_for_vcpu() argument 250 while (READ_ONCE(vcpu_last_completed_iteration[vcpu_idx]) != in spin_wait_for_vcpu()
|
| A D | dirty_log_perf_test.c | 71 int vcpu_idx = vcpu_args->vcpu_idx; in vcpu_worker() local 94 pr_debug("Got sync event from vCPU %d\n", vcpu_idx); in vcpu_worker() 95 vcpu_last_completed_iteration[vcpu_idx] = current_iteration; in vcpu_worker() 97 vcpu_idx, vcpu_last_completed_iteration[vcpu_idx]); in vcpu_worker() 103 vcpu_idx, current_iteration, ts_diff.tv_sec, in vcpu_worker() 107 vcpu_idx, current_iteration, ts_diff.tv_sec, in vcpu_worker() 120 avg = timespec_div(total, vcpu_last_completed_iteration[vcpu_idx]); in vcpu_worker() 122 vcpu_idx, pages_count, vcpu_last_completed_iteration[vcpu_idx], in vcpu_worker()
|
| A D | steal_time.c | 85 static void steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) in steal_time_dump() argument 87 struct kvm_steal_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]); in steal_time_dump() 89 ksft_print_msg("VCPU%d:\n", vcpu_idx); in steal_time_dump() 198 static void steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) in steal_time_dump() argument 200 struct st_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]); in steal_time_dump() 202 ksft_print_msg("VCPU%d:\n", vcpu_idx); in steal_time_dump() 289 static void steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) in steal_time_dump() argument 291 struct sta_struct *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]); in steal_time_dump() 294 pr_info("VCPU%d:\n", vcpu_idx); in steal_time_dump()
|
| A D | demand_paging_test.c | 35 int vcpu_idx = vcpu_args->vcpu_idx; in vcpu_worker() local 53 PER_VCPU_DEBUG("vCPU %d execution time: %ld.%.9lds\n", vcpu_idx, in vcpu_worker()
|
| /linux/tools/testing/selftests/kvm/lib/ |
| A D | memstress.c | 23 int vcpu_idx; member 47 void memstress_guest_code(uint32_t vcpu_idx) in memstress_guest_code() argument 50 struct memstress_vcpu_args *vcpu_args = &args->vcpu_args[vcpu_idx]; in memstress_guest_code() 58 rand_state = new_guest_random_state(guest_random_seed + vcpu_idx); in memstress_guest_code() 64 GUEST_ASSERT(vcpu_args->vcpu_idx == vcpu_idx); in memstress_guest_code() 101 vcpu_args->vcpu_idx = i; in memstress_setup_vcpus() 265 int vcpu_idx = vcpu->vcpu_idx; in vcpu_thread_main() local 268 kvm_pin_this_task_to_pcpu(memstress_args.vcpu_to_pcpu[vcpu_idx]); in vcpu_thread_main() 281 vcpu_thread_fn(&memstress_args.vcpu_args[vcpu_idx]); in vcpu_thread_main() 298 vcpu->vcpu_idx = i; in memstress_start_vcpu_threads()
|
| /linux/arch/arm64/kvm/hyp/nvhe/ |
| A D | pkvm.c | 266 unsigned int vcpu_idx) in pkvm_load_hyp_vcpu() argument 273 if (!hyp_vm || hyp_vm->nr_vcpus <= vcpu_idx) in pkvm_load_hyp_vcpu() 276 hyp_vcpu = hyp_vm->vcpus[vcpu_idx]; in pkvm_load_hyp_vcpu() 318 unsigned int vcpu_idx) in init_pkvm_hyp_vcpu() argument 325 if (host_vcpu->vcpu_idx != vcpu_idx) { in init_pkvm_hyp_vcpu() 334 hyp_vcpu->vcpu.vcpu_idx = vcpu_idx; in init_pkvm_hyp_vcpu()
|
| A D | hyp-main.c | 166 host_vcpu->vcpu_idx); in handle___kvm_vcpu_run()
|
| /linux/arch/arm64/kvm/ |
| A D | trace_arm.h | 94 TP_PROTO(unsigned int type, int vcpu_idx, int irq_num, int level), 95 TP_ARGS(type, vcpu_idx, irq_num, level), 99 __field( int, vcpu_idx ) 106 __entry->vcpu_idx = vcpu_idx; 115 __entry->type, __entry->vcpu_idx, __entry->irq_num, __entry->level)
|
| A D | pkvm.c | 173 if (WARN_ON(host_vcpu->vcpu_idx != idx)) { in __pkvm_create_hyp_vm()
|
| A D | arm.c | 596 if (*last_ran != vcpu->vcpu_idx) { in kvm_arch_vcpu_load() 598 *last_ran = vcpu->vcpu_idx; in kvm_arch_vcpu_load()
|
| /linux/tools/testing/selftests/kvm/x86_64/ |
| A D | dirty_log_page_splitting_test.c | 69 int vcpu_idx = vcpu_args->vcpu_idx; in vcpu_worker() local 78 vcpu_last_completed_iteration[vcpu_idx] = current_iteration; in vcpu_worker()
|
| /linux/arch/x86/kvm/ |
| A D | xen.c | 114 e.vcpu_idx = vcpu->vcpu_idx; in kvm_xen_inject_timer_irqs() 136 e.vcpu_idx = vcpu->vcpu_idx; in xen_timer_callback() 1746 int vcpu_idx, idx, rc; in kvm_xen_set_evtchn_fast() local 1748 vcpu_idx = READ_ONCE(xe->vcpu_idx); in kvm_xen_set_evtchn_fast() 1749 if (vcpu_idx >= 0) in kvm_xen_set_evtchn_fast() 1750 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in kvm_xen_set_evtchn_fast() 1755 WRITE_ONCE(xe->vcpu_idx, vcpu->vcpu_idx); in kvm_xen_set_evtchn_fast() 1936 e->xen_evtchn.vcpu_idx = vcpu->vcpu_idx; in kvm_xen_setup_evtchn() 1938 e->xen_evtchn.vcpu_idx = -1; in kvm_xen_setup_evtchn() 1965 e.vcpu_idx = -1; in kvm_xen_hvm_evtchn_send() [all …]
|
| A D | hyperv.h | 92 return hv_vcpu ? hv_vcpu->vp_index : vcpu->vcpu_idx; in kvm_hv_get_vpindex() 321 return vcpu->vcpu_idx; in kvm_hv_get_vpindex()
|
| A D | lapic.c | 1413 int vcpu_idx; in kvm_bitmap_or_dest_vcpus() local 1425 vcpu_idx = dest_vcpu[i]->vcpu->vcpu_idx; in kvm_bitmap_or_dest_vcpus() 1426 __set_bit(vcpu_idx, vcpu_bitmap); in kvm_bitmap_or_dest_vcpus()
|
| A D | hyperv.c | 981 hv_vcpu->vp_index = vcpu->vcpu_idx; in kvm_hv_vcpu_init() 1542 if (hv_vcpu->vp_index == vcpu->vcpu_idx) in kvm_hv_set_msr() 1544 else if (new_vp_index == vcpu->vcpu_idx) in kvm_hv_set_msr()
|
| /linux/arch/arm64/kvm/vgic/ |
| A D | vgic-debug.c | 206 id = vcpu->vcpu_idx; in print_header() 252 (irq->target_vcpu) ? irq->target_vcpu->vcpu_idx : -1, in print_irq_state() 264 (irq->vcpu) ? irq->vcpu->vcpu_idx : -1); in print_irq_state()
|
| A D | vgic.c | 438 trace_vgic_update_irq_pending(vcpu ? vcpu->vcpu_idx : 0, intid, level); in kvm_vgic_inject_irq()
|
| /linux/arch/arm64/kvm/hyp/include/nvhe/ |
| A D | pkvm.h | 70 unsigned int vcpu_idx);
|
| /linux/tools/testing/selftests/kvm/include/ |
| A D | memstress.h | 29 int vcpu_idx; member
|
| /linux/arch/riscv/kvm/ |
| A D | aia_device.c | 191 unsigned long vcpu_idx, bool write) in aia_imsic_addr() argument 196 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in aia_imsic_addr() 560 vaia->hart_index = vcpu->vcpu_idx; in kvm_riscv_vcpu_aia_init()
|
| A D | vcpu.c | 177 if (vcpu->vcpu_idx != 0) in kvm_arch_vcpu_postcreate()
|
| /linux/arch/s390/kvm/ |
| A D | interrupt.c | 418 set_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in __set_cpu_idle() 424 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in __unset_cpu_idle() 3070 int vcpu_idx, online_vcpus = atomic_read(&kvm->online_vcpus); in __airqs_kick_single_vcpu() local 3075 for_each_set_bit(vcpu_idx, kvm->arch.idle_mask, online_vcpus) { in __airqs_kick_single_vcpu() 3076 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in __airqs_kick_single_vcpu() 3082 if (test_and_set_bit(vcpu_idx, gi->kicked_mask)) in __airqs_kick_single_vcpu()
|
| A D | kvm-s390.h | 101 return test_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in is_vcpu_idle()
|
| /linux/include/linux/ |
| A D | kvm_host.h | 330 int vcpu_idx; /* index into kvm->vcpu_array */ member 653 int vcpu_idx; member
|