Searched refs:is_hvm_vcpu (Results 1 – 25 of 27) sorted by relevance
12
17 (is_hvm_vcpu(current) ? \21 (is_hvm_vcpu(current) ? \25 (is_hvm_vcpu(current) ? \29 (is_hvm_vcpu(current) ? \33 (is_hvm_vcpu(current) ? \37 (is_hvm_vcpu(current) ? \
29 return (is_hvm_vcpu(v) ? hvm_local_events_need_delivery(v) : in local_events_need_delivery()
287 if ( !is_hvm_vcpu(v) ) in __core2_vpmu_save()295 if ( !is_hvm_vcpu(v) ) in core2_vpmu_save()343 if ( !is_hvm_vcpu(v) ) in __core2_vpmu_load()394 !(is_hvm_vcpu(v) in core2_vpmu_verify()455 if ( is_hvm_vcpu(v) ) in core2_vpmu_alloc_resource()521 if ( is_hvm_vcpu(current) && cpu_has_vmx_msr_bitmap ) in core2_vpmu_msr_common_check()615 if ( is_hvm_vcpu(v) ) in core2_vpmu_do_wrmsr()684 if ( is_hvm_vcpu(v) ) in core2_vpmu_do_wrmsr()703 if ( is_hvm_vcpu(v) ) in core2_vpmu_do_wrmsr()737 if ( is_hvm_vcpu(v) ) in core2_vpmu_do_rdmsr()[all …]
308 if ( !vpmu_is_set(vpmu, VPMU_RUNNING) && is_hvm_vcpu(v) && in amd_vpmu_save()370 if ( is_hvm_vcpu(v) && (type == MSR_TYPE_CTRL) && in amd_vpmu_do_wrmsr()384 if ( is_hvm_vcpu(v) && is_msr_bitmap_on(vpmu) ) in amd_vpmu_do_wrmsr()393 if ( is_hvm_vcpu(v) && is_msr_bitmap_on(vpmu) ) in amd_vpmu_do_wrmsr()436 if ( is_hvm_vcpu(v) && is_msr_bitmap_on(vpmu) ) in amd_vpmu_destroy()
238 if ( is_hvm_vcpu(sampled) ) in vpmu_do_interrupt()289 if ( !is_hvm_vcpu(sampled) ) in vpmu_do_interrupt()
174 (is_hvm_vcpu(v) ? __linear_l1_table : __sh_linear_l1_table) + \182 (is_hvm_vcpu(v) ? __linear_l2_table : __sh_linear_l2_table) + \191 (is_hvm_vcpu(v) ? __linear_l3_table : __sh_linear_l3_table) + \
4712 if ( (vaddr & (bytes - 1)) && !is_hvm_vcpu(v) ) in sh_x86_emulate_write()4754 if ( (vaddr & (bytes - 1)) && !is_hvm_vcpu(v) ) in sh_x86_emulate_cmpxchg()4827 if ( (sflags & _PAGE_GLOBAL) && !is_hvm_vcpu(v) ) in sh_audit_flags()
323 ASSERT(is_hvm_vcpu(v)); in shadow_init_emulation()365 ASSERT(is_hvm_vcpu(v)); in shadow_continue_emulation()
82 if ( !is_hvm_vcpu(curr) ) in xenoprofile_get_mode()
49 if (is_hvm_vcpu(vcpu)) in is_32bit_vcpu()
328 ASSERT(is_hvm_vcpu(v)); in athlon_check_ctrs()
52 if ( is_hvm_vcpu(v) ) in compat_vcpu_op()
57 (is_hvm_vcpu(v) ? hvm_get_guest_time(v) : (u64)get_s_time())184 if ( (v == NULL) || !is_hvm_vcpu(v) || (channel != 0) ) in pit_load_count()
44 ASSERT(is_hvm_vcpu(v)); in hvm_get_guest_time_fixed()
547 if ( is_hvm_vcpu(v) ) in arch_evtchn_inject()
182 if ( is_hvm_vcpu(current) ? hvm_guest_x86_mode(current) != 8 in tmem_get_tmemop_from_client()
886 #define is_hvm_vcpu(v) (is_hvm_domain(v->domain)) macro
243 if ( is_hvm_vcpu(v) ) in vmx_intr_assist()
791 if ( is_hvm_vcpu(current) ) in vmx_vmcs_exit()
132 ASSERT(is_hvm_vcpu(curr)); in vm_event_fill_regs()
391 if ( is_hvm_vcpu(v) ) in vcpu_destroy()1159 if ( is_hvm_vcpu(v) ) in arch_initialise_vcpu()1623 if ( cpu_has_xsaves && is_hvm_vcpu(n) ) in __context_switch()2059 if ( is_hvm_vcpu(v) ) in vcpu_mark_events_pending()
261 if ( is_hvm_vcpu(v) ) in show_guest_stack()728 system_state >= SYS_STATE_active && is_hvm_vcpu(curr) && in do_trap()
108 if ( guest_mode(regs) && is_hvm_vcpu(v) ) in show_registers()
86 is_hvm_vcpu(v) ? "HVM" : "PV", ctxt->caps, in vmce_restore_vcpu()
155 if ( is_hvm_vcpu(v) ) in elf_memcpy()
Completed in 55 milliseconds