Lines Matching refs:hv_vcpu
69 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in to_hv_synic() local
71 return &hv_vcpu->synic; in to_hv_synic()
76 struct kvm_vcpu_hv *hv_vcpu = container_of(synic, struct kvm_vcpu_hv, synic); in hv_synic_to_vcpu() local
78 return hv_vcpu->vcpu; in hv_synic_to_vcpu()
88 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_get_vpindex() local
90 return hv_vcpu ? hv_vcpu->vp_index : vcpu->vcpu_idx; in kvm_hv_get_vpindex()
121 struct kvm_vcpu_hv *hv_vcpu; in hv_stimer_to_vcpu() local
123 hv_vcpu = container_of(stimer - stimer->index, struct kvm_vcpu_hv, in hv_stimer_to_vcpu()
125 return hv_vcpu->vcpu; in hv_stimer_to_vcpu()
130 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_has_stimer_pending() local
132 if (!hv_vcpu) in kvm_hv_has_stimer_pending()
135 return !bitmap_empty(hv_vcpu->stimer_pending_bitmap, in kvm_hv_has_stimer_pending()
145 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_invtsc_suppressed() local
154 if (!hv_vcpu || in kvm_hv_invtsc_suppressed()
155 !(hv_vcpu->cpuid_cache.features_eax & HV_ACCESS_TSC_INVARIANT)) in kvm_hv_invtsc_suppressed()
184 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_get_tlb_flush_fifo() local
188 return &hv_vcpu->tlb_flush_fifo[i]; in kvm_hv_get_tlb_flush_fifo()
205 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in guest_hv_cpuid_has_l2_tlb_flush() local
207 return hv_vcpu && in guest_hv_cpuid_has_l2_tlb_flush()
208 (hv_vcpu->cpuid_cache.nested_eax & HV_X64_NESTED_DIRECT_FLUSH); in guest_hv_cpuid_has_l2_tlb_flush()
213 struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu); in kvm_hv_is_tlb_flush_hcall() local
216 if (!hv_vcpu) in kvm_hv_is_tlb_flush_hcall()