| /arch/x86/kernel/fpu/ |
| A D | xstate.h | 149 #define XSTATE_XSAVE(st, lmask, hmask, err) \ argument 166 #define XSTATE_XRESTORE(st, lmask, hmask) \ argument 220 u32 hmask = mask >> 32; in os_xsave() local 240 u32 hmask = mask >> 32; in os_xrstor() local 251 u32 hmask = mask >> 32; in os_xrstor_supervisor() local 305 u32 hmask; in xsave_to_user_sigframe() local 313 hmask = mask >> 32; in xsave_to_user_sigframe() 317 XSTATE_OP(XSAVE, buf, lmask, hmask, err); in xsave_to_user_sigframe() 333 u32 hmask = mask >> 32; in xrstor_from_user_sigframe() local 339 XSTATE_OP(XRSTOR, xstate, lmask, hmask, err); in xrstor_from_user_sigframe() [all …]
|
| A D | xstate.c | 342 u32 hmask = mask >> 32; in os_xrstor_booting() local 346 XSTATE_OP(XRSTORS, xstate, lmask, hmask, err); in os_xrstor_booting() 348 XSTATE_OP(XRSTOR, xstate, lmask, hmask, err); in os_xrstor_booting()
|
| /arch/riscv/kvm/ |
| A D | vcpu_sbi_v01.c | 19 ulong hmask; in kvm_sbi_ext_v01_handler() local 51 hmask = kvm_riscv_vcpu_unpriv_read(vcpu, false, cp->a0, utrap); in kvm_sbi_ext_v01_handler() 53 hmask = (1UL << atomic_read(&kvm->online_vcpus)) - 1; in kvm_sbi_ext_v01_handler() 57 for_each_set_bit(i, &hmask, BITS_PER_LONG) { in kvm_sbi_ext_v01_handler() 73 hmask = kvm_riscv_vcpu_unpriv_read(vcpu, false, cp->a0, utrap); in kvm_sbi_ext_v01_handler() 75 hmask = (1UL << atomic_read(&kvm->online_vcpus)) - 1; in kvm_sbi_ext_v01_handler() 80 kvm_riscv_fence_i(vcpu->kvm, 0, hmask); in kvm_sbi_ext_v01_handler() 84 kvm_riscv_hfence_vvma_all(vcpu->kvm, 0, hmask, vmid); in kvm_sbi_ext_v01_handler() 86 kvm_riscv_hfence_vvma_gva(vcpu->kvm, 0, hmask, cp->a1, in kvm_sbi_ext_v01_handler() 91 kvm_riscv_hfence_vvma_asid_all(vcpu->kvm, 0, hmask, in kvm_sbi_ext_v01_handler() [all …]
|
| A D | tlb.c | 299 unsigned long hbase, unsigned long hmask, in make_xfence_request() argument 313 if (!(hmask & (1UL << (vcpu->vcpu_id - hbase)))) in make_xfence_request() 335 unsigned long hbase, unsigned long hmask) in kvm_riscv_fence_i() argument 342 unsigned long hbase, unsigned long hmask, in kvm_riscv_hfence_gvma_vmid_gpa() argument 354 make_xfence_request(kvm, hbase, hmask, KVM_REQ_HFENCE, in kvm_riscv_hfence_gvma_vmid_gpa() 359 unsigned long hbase, unsigned long hmask, in kvm_riscv_hfence_gvma_vmid_all() argument 366 make_xfence_request(kvm, hbase, hmask, KVM_REQ_HFENCE, in kvm_riscv_hfence_gvma_vmid_all() 371 unsigned long hbase, unsigned long hmask, in kvm_riscv_hfence_vvma_asid_gva() argument 389 unsigned long hbase, unsigned long hmask, in kvm_riscv_hfence_vvma_asid_all() argument 402 unsigned long hbase, unsigned long hmask, in kvm_riscv_hfence_vvma_gva() argument [all …]
|
| A D | vcpu_sbi_replace.c | 52 unsigned long hmask = cp->a0; in kvm_sbi_ext_ipi_handler() local 69 if (!(hmask & (1UL << hart_bit))) in kvm_sbi_ext_ipi_handler() 80 if (hbase != -1UL && (hmask ^ sentmask)) in kvm_sbi_ext_ipi_handler() 96 unsigned long hmask = cp->a0; in kvm_sbi_ext_rfence_handler() local 103 kvm_riscv_fence_i(vcpu->kvm, hbase, hmask); in kvm_sbi_ext_rfence_handler() 109 kvm_riscv_hfence_vvma_all(vcpu->kvm, hbase, hmask, vmid); in kvm_sbi_ext_rfence_handler() 111 kvm_riscv_hfence_vvma_gva(vcpu->kvm, hbase, hmask, in kvm_sbi_ext_rfence_handler() 118 kvm_riscv_hfence_vvma_asid_all(vcpu->kvm, hbase, hmask, in kvm_sbi_ext_rfence_handler() 121 kvm_riscv_hfence_vvma_asid_gva(vcpu->kvm, hbase, hmask, cp->a2, in kvm_sbi_ext_rfence_handler()
|
| /arch/riscv/include/asm/ |
| A D | kvm_tlb.h | 60 unsigned long hbase, unsigned long hmask); 62 unsigned long hbase, unsigned long hmask, 66 unsigned long hbase, unsigned long hmask, 69 unsigned long hbase, unsigned long hmask, 74 unsigned long hbase, unsigned long hmask, 77 unsigned long hbase, unsigned long hmask, 81 unsigned long hbase, unsigned long hmask,
|
| /arch/riscv/kernel/ |
| A D | sbi.c | 31 unsigned long hmask = 0; in __sbi_v01_cpumask_to_hartmask() local 46 hmask |= BIT(hartid); in __sbi_v01_cpumask_to_hartmask() 49 return hmask; in __sbi_v01_cpumask_to_hartmask() 251 __func__, hbase, hmask, result); in __sbi_rfence_v02_call() 269 if (hmask) { in __sbi_rfence_v02() 272 result = __sbi_rfence_v02_call(fid, hmask, in __sbi_rfence_v02() 276 hmask = 0; in __sbi_rfence_v02() 279 hmask <<= hbase - hartid; in __sbi_rfence_v02() 283 if (!hmask) { in __sbi_rfence_v02() 289 hmask |= BIT(hartid - hbase); in __sbi_rfence_v02() [all …]
|
| /arch/powerpc/mm/book3s32/ |
| A D | mmu.c | 395 unsigned int hmask = Hash_mask >> (16 - LG_HPTEG_SIZE); in MMU_init_hw_patch() local 414 modify_instruction_site(&patch__hash_page_B, 0xffff, hmask); in MMU_init_hw_patch() 415 modify_instruction_site(&patch__hash_page_C, 0xffff, hmask); in MMU_init_hw_patch() 423 modify_instruction_site(&patch__flush_hash_B, 0xffff, hmask); in MMU_init_hw_patch()
|
| /arch/s390/mm/ |
| A D | gmap.c | 2255 unsigned long hmask, unsigned long next, in __s390_enable_skey_hugetlb() argument
|