| /arch/powerpc/mm/book3s64/ |
| A D | hash_64k.c | 45 unsigned long rflags, pa; in __hash_page_4K() local 77 rflags = htab_convert_pte_flags(subpg_pte, flags); in __hash_page_4K() 86 rflags = hash_page_do_lazy_icache(rflags, __pte(old_pte), trap); in __hash_page_4K() 120 ret = mmu_hash_ops.hpte_updatepp(gslot, rflags, vpn, in __hash_page_4K() 162 slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, 0, in __hash_page_4K() 172 rflags, HPTE_V_SECONDARY, in __hash_page_4K() 234 unsigned long rflags, pa; in __hash_page_64K() local 268 rflags = htab_convert_pte_flags(new_pte, flags); in __hash_page_64K() 273 rflags = hash_page_do_lazy_icache(rflags, __pte(old_pte), trap); in __hash_page_64K() 298 slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, 0, in __hash_page_64K() [all …]
|
| A D | hash_4k.c | 27 unsigned long rflags, pa; in __hash_page_4K() local 59 rflags = htab_convert_pte_flags(new_pte, flags); in __hash_page_4K() 64 rflags = hash_page_do_lazy_icache(rflags, __pte(old_pte), trap); in __hash_page_4K() 74 if (mmu_hash_ops.hpte_updatepp(gslot, rflags, vpn, MMU_PAGE_4K, in __hash_page_4K() 88 slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, 0, in __hash_page_4K() 96 rflags, in __hash_page_4K()
|
| A D | hugetlbpage.c | 27 unsigned long rflags, pa; in __hash_page_huge() local 80 rflags = htab_convert_pte_flags(new_pte, flags); in __hash_page_huge() 92 rflags = hash_page_do_lazy_icache(rflags, __pte(old_pte), trap); in __hash_page_huge() 100 if (mmu_hash_ops.hpte_updatepp(gslot, rflags, vpn, mmu_psize, in __hash_page_huge() 113 slot = hpte_insert_repeating(hash, vpn, pa, rflags, 0, in __hash_page_huge()
|
| A D | hash_hugepage.c | 27 unsigned long rflags, pa, hidx; in __hash_page_thp() local 60 rflags = htab_convert_pte_flags(new_pmd, flags); in __hash_page_thp() 109 ret = mmu_hash_ops.hpte_updatepp(slot, rflags, vpn, in __hash_page_thp() 137 slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, 0, in __hash_page_thp() 145 rflags, in __hash_page_thp()
|
| A D | hash_utils.c | 555 unsigned long rflags = 0; in htab_convert_pte_flags() local 559 rflags |= HPTE_R_N; in htab_convert_pte_flags() 573 rflags |= (HPTE_R_PP0 | 0x2); in htab_convert_pte_flags() 575 rflags |= 0x3; in htab_convert_pte_flags() 580 rflags |= 0x2; in htab_convert_pte_flags() 588 rflags |= 0x1; in htab_convert_pte_flags() 594 rflags |= HPTE_R_R; in htab_convert_pte_flags() 597 rflags |= HPTE_R_C; in htab_convert_pte_flags() 603 rflags |= HPTE_R_I; in htab_convert_pte_flags() 612 rflags |= HPTE_R_M; in htab_convert_pte_flags() [all …]
|
| A D | hash_native.c | 262 unsigned long pa, unsigned long rflags, in native_hpte_insert() argument 275 hpte_group, vpn, pa, rflags, vflags, psize); in native_hpte_insert() 296 hpte_r = hpte_encode_r(pa, psize, apsize) | rflags; in native_hpte_insert()
|
| /arch/powerpc/kvm/ |
| A D | book3s_64_mmu_host.c | 79 int rflags = 0x192; in kvmppc_mmu_map_page() local 126 rflags |= PP_RXRX; in kvmppc_mmu_map_page() 131 rflags |= HPTE_R_N; in kvmppc_mmu_map_page() 135 rflags |= pte_to_hpte_pkey_bits(0, HPTE_USE_KERNEL_KEY); in kvmppc_mmu_map_page() 136 rflags = (rflags & ~HPTE_R_WIMG) | orig_pte->wimg; in kvmppc_mmu_map_page() 167 ret = mmu_hash_ops.hpte_insert(hpteg, vpn, hpaddr, rflags, vflags, in kvmppc_mmu_map_page() 180 trace_kvm_book3s_64_mmu_map(rflags, hpteg, in kvmppc_mmu_map_page()
|
| A D | trace_pr.h | 32 TP_PROTO(int rflags, ulong hpteg, ulong va, kvm_pfn_t hpaddr, 34 TP_ARGS(rflags, hpteg, va, hpaddr, orig_pte), 47 __entry->flag_w = ((rflags & HPTE_R_PP) == 3) ? '-' : 'w'; 48 __entry->flag_x = (rflags & HPTE_R_N) ? '-' : 'x';
|
| /arch/x86/kvm/vmx/ |
| A D | sgx.c | 299 unsigned long sig_hva, secs_hva, token_hva, rflags; in handle_encls_einit() local 346 rflags = vmx_get_rflags(vcpu) & ~(X86_EFLAGS_CF | X86_EFLAGS_PF | in handle_encls_einit() 350 rflags |= X86_EFLAGS_ZF; in handle_encls_einit() 352 rflags &= ~X86_EFLAGS_ZF; in handle_encls_einit() 353 vmx_set_rflags(vcpu, rflags); in handle_encls_einit()
|
| A D | vmx.h | 215 ulong rflags; member 359 void vmx_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags);
|
| A D | x86_ops.h | 81 void vmx_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags);
|
| A D | vmx.c | 1483 unsigned long rflags, save_rflags; in vmx_get_rflags() local 1487 rflags = vmcs_readl(GUEST_RFLAGS); in vmx_get_rflags() 1489 rflags &= RMODE_GUEST_OWNED_EFLAGS_BITS; in vmx_get_rflags() 1493 vmx->rflags = rflags; in vmx_get_rflags() 1495 return vmx->rflags; in vmx_get_rflags() 1510 vmx->rflags = rflags; in vmx_set_rflags() 1511 vmcs_writel(GUEST_RFLAGS, rflags); in vmx_set_rflags() 1516 vmx->rflags = rflags; in vmx_set_rflags() 1518 vmx->rmode.save_rflags = rflags; in vmx_set_rflags() 1519 rflags |= X86_EFLAGS_IOPL | X86_EFLAGS_VM; in vmx_set_rflags() [all …]
|
| A D | main.c | 497 static void vt_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags) in vt_set_rflags() argument 502 vmx_set_rflags(vcpu, rflags); in vt_set_rflags()
|
| /arch/powerpc/platforms/ps3/ |
| A D | htab.c | 35 unsigned long pa, unsigned long rflags, unsigned long vflags, in ps3_hpte_insert() argument 53 hpte_r = hpte_encode_r(ps3_mm_phys_to_lpar(pa), psize, apsize) | rflags; in ps3_hpte_insert()
|
| /arch/x86/kvm/ |
| A D | mmu.h | 191 unsigned long rflags = kvm_x86_call(get_rflags)(vcpu); in permission_fault() local 206 bool not_smap = ((rflags & X86_EFLAGS_AC) | implicit_access) == X86_EFLAGS_AC; in permission_fault()
|
| A D | smm.h | 132 u64 rflags; member
|
| A D | smm.c | 101 CHECK_SMRAM64_OFFSET(rflags, 0xFF70); in check_smram_offsets() 238 smram->rflags = kvm_get_rflags(vcpu); in enter_smm_save_state_64() 523 ctxt->eflags = smstate->rflags | X86_EFLAGS_FIXED; in rsm_load_state_64()
|
| /arch/x86/include/asm/xen/ |
| A D | interface_64.h | 84 uint64_t rax, r11, rcx, flags, rip, cs, rflags, rsp, ss; member
|
| /arch/x86/kvm/svm/ |
| A D | nested.c | 156 if (!(svm->vmcb01.ptr->save.rflags & X86_EFLAGS_IF)) in recalc_intercepts() 639 kvm_set_rflags(vcpu, vmcb12->save.rflags | X86_EFLAGS_FIXED); in nested_vmcb02_prepare_save() 996 vmcb01->save.rflags = kvm_get_rflags(vcpu); in nested_svm_vmrun() 1038 to_save->rflags = from_save->rflags | X86_EFLAGS_FIXED; in svm_copy_vmrun_state() 1106 vmcb12->save.rflags = kvm_get_rflags(vcpu); in nested_svm_vmexit() 1218 kvm_set_rflags(vcpu, vmcb01->save.rflags); in nested_svm_vmexit() 1260 if (unlikely(vmcb01->save.rflags & X86_EFLAGS_TF)) in nested_svm_vmexit() 1833 (save->rflags & X86_EFLAGS_VM) || in svm_set_nested_state()
|
| A D | svm.c | 302 old_rflags = svm->vmcb->save.rflags; in __svm_skip_emulated_instruction() 308 svm->vmcb->save.rflags = old_rflags; in __svm_skip_emulated_instruction() 930 svm->vmcb->save.rflags &= ~X86_EFLAGS_TF; in disable_nmi_singlestep() 932 svm->vmcb->save.rflags &= ~X86_EFLAGS_RF; in disable_nmi_singlestep() 1470 unsigned long rflags = svm->vmcb->save.rflags; in svm_get_rflags() local 1475 rflags &= ~X86_EFLAGS_TF; in svm_get_rflags() 1477 rflags &= ~X86_EFLAGS_RF; in svm_get_rflags() 1479 return rflags; in svm_get_rflags() 1485 rflags |= (X86_EFLAGS_TF | X86_EFLAGS_RF); in svm_set_rflags() 1492 to_svm(vcpu)->vmcb->save.rflags = rflags; in svm_set_rflags() [all …]
|
| /arch/s390/include/asm/ |
| A D | qdio.h | 103 u32 rflags : 8; member
|
| /arch/x86/hyperv/ |
| A D | hv_vtl.c | 157 input->vp_context.rflags = 0x0000000000000002; in hv_vtl_bringup_vcpu()
|
| /arch/x86/include/asm/ |
| A D | svm.h | 336 u64 rflags; member 393 u64 rflags; member
|
| /arch/powerpc/platforms/pseries/ |
| A D | lpar.c | 759 unsigned long rflags, unsigned long vflags, in pSeries_lpar_hpte_insert() argument 770 hpte_group, vpn, pa, rflags, vflags, psize); in pSeries_lpar_hpte_insert() 773 hpte_r = hpte_encode_r(pa, psize, apsize) | rflags; in pSeries_lpar_hpte_insert()
|
| /arch/powerpc/include/asm/book3s/64/ |
| A D | mmu-hash.h | 151 unsigned long rflags,
|