Searched refs:fst (Results 1 – 7 of 7) sorted by relevance
| /arch/loongarch/kernel/ |
| A D | fpu.S | 30 EX fst.d $f0, \base, (0 * FPU_REG_WIDTH) 31 EX fst.d $f1, \base, (1 * FPU_REG_WIDTH) 32 EX fst.d $f2, \base, (2 * FPU_REG_WIDTH) 33 EX fst.d $f3, \base, (3 * FPU_REG_WIDTH) 34 EX fst.d $f4, \base, (4 * FPU_REG_WIDTH) 35 EX fst.d $f5, \base, (5 * FPU_REG_WIDTH) 36 EX fst.d $f6, \base, (6 * FPU_REG_WIDTH) 37 EX fst.d $f7, \base, (7 * FPU_REG_WIDTH) 38 EX fst.d $f8, \base, (8 * FPU_REG_WIDTH) 39 EX fst.d $f9, \base, (9 * FPU_REG_WIDTH) [all …]
|
| /arch/loongarch/include/asm/ |
| A D | asmmacro.h | 134 fst.d $f0, \tmp, THREAD_FPR0 - THREAD_FPR0 135 fst.d $f1, \tmp, THREAD_FPR1 - THREAD_FPR0 136 fst.d $f2, \tmp, THREAD_FPR2 - THREAD_FPR0 137 fst.d $f3, \tmp, THREAD_FPR3 - THREAD_FPR0 138 fst.d $f4, \tmp, THREAD_FPR4 - THREAD_FPR0 139 fst.d $f5, \tmp, THREAD_FPR5 - THREAD_FPR0 140 fst.d $f6, \tmp, THREAD_FPR6 - THREAD_FPR0 141 fst.d $f7, \tmp, THREAD_FPR7 - THREAD_FPR0 142 fst.d $f8, \tmp, THREAD_FPR8 - THREAD_FPR0 143 fst.d $f9, \tmp, THREAD_FPR9 - THREAD_FPR0 [all …]
|
| /arch/arm64/kernel/ |
| A D | fpsimd.c | 631 static void __fpsimd_to_sve(void *sst, struct user_fpsimd_state const *fst, in __fpsimd_to_sve() argument 639 *p = arm64_cpu_to_le128(fst->vregs[i]); in __fpsimd_to_sve() 659 struct user_fpsimd_state const *fst = &task->thread.uw.fpsimd_state; in fpsimd_to_sve() local 665 __fpsimd_to_sve(sst, fst, vq); in fpsimd_to_sve() 683 struct user_fpsimd_state *fst = &task->thread.uw.fpsimd_state; in sve_to_fpsimd() local 694 fst->vregs[i] = arm64_le128_to_cpu(*p); in sve_to_fpsimd() 782 struct user_fpsimd_state const *fst = &task->thread.uw.fpsimd_state; in fpsimd_sync_to_effective_state_zeropad() local 790 __fpsimd_to_sve(sst, fst, vq); in fpsimd_sync_to_effective_state_zeropad()
|
| /arch/arm64/kvm/ |
| A D | at.c | 13 static void fail_s1_walk(struct s1_walk_result *wr, u8 fst, bool s1ptw) in fail_s1_walk() argument 15 wr->fst = fst; in fail_s1_walk() 761 par |= FIELD_PREP(SYS_PAR_EL1_FST, wr->fst); in compute_par_s1() 1298 u8 fst = FIELD_GET(SYS_PAR_EL1_FST, par); in par_check_s1_perm_fault() local 1300 return ((fst & ESR_ELx_FSC_TYPE) == ESR_ELx_FSC_PERM && in par_check_s1_perm_fault() 1306 u8 fst = FIELD_GET(SYS_PAR_EL1_FST, par); in par_check_s1_access_fault() local 1308 return ((fst & ESR_ELx_FSC_TYPE) == ESR_ELx_FSC_ACCESS && in par_check_s1_access_fault()
|
| A D | nested.c | 1322 esr |= FIELD_PREP(ESR_ELx_FSC, vt->wr.fst); in kvm_handle_vncr_abort()
|
| /arch/arm64/include/asm/ |
| A D | kvm_nested.h | 327 u8 fst; member
|
| /arch/x86/events/intel/ |
| A D | ds.c | 1773 bool fst = fl & (PERF_X86_EVENT_PEBS_ST | PERF_X86_EVENT_PEBS_HSW_PREC); in get_data_src() local 1781 else if (fst && (fl & PERF_X86_EVENT_PEBS_HSW_PREC)) in get_data_src() 1783 else if (fst) in get_data_src()
|
Completed in 27 milliseconds