/linux-6.3-rc2/arch/alpha/lib/ |
A D | ev67-strrchr.S | 37 insbl a1, 2, t5 # U : 0000000000ch0000 42 sll t5, 8, t3 # U : 00000000ch000000 46 or t5, t3, t3 # E : 00000000chch0000 53 lda t5, -1 # E : build garbage mask 56 mskqh t5, a0, t4 # E : Complete garbage mask 86 subq t4, 1, t5 # E : build a mask of the bytes up to... 87 or t4, t5, t4 # E : ... and including the null 102 lda t5, 0x3f($31) # E : 103 subq t5, t2, t5 # E : Normalize leading zero count 105 addq t6, t5, v0 # E : and add to quadword address
|
A D | strrchr.S | 24 sll a1, 8, t5 # e0 : replicate our test character 26 or t5, a1, a1 # e0 : 28 sll a1, 16, t5 # e0 : 30 or t5, a1, a1 # e0 : 32 sll a1, 32, t5 # e0 : 35 or t5, a1, a1 # .. e1 : character replication complete 58 subq t4, 1, t5 # e0 : build a mask of the bytes up to... 59 or t4, t5, t4 # e1 : ... and including the null
|
A D | ev67-strchr.S | 34 insbl a1, 1, t5 # U : 000000000000ch00 38 or t5, t3, a1 # E : 000000000000chch 44 inswl a1, 2, t5 # E : 00000000chch0000 48 or a3, t5, t5 # E : 0000chchchch0000 53 or t5, a1, a1 # E : chchchchchchchch
|
A D | strchr.S | 24 sll a1, 8, t5 # e0 : replicate the search character 26 or t5, a1, a1 # e0 : 28 sll a1, 16, t5 # e0 : 31 or t5, a1, a1 # .. e1 : 32 sll a1, 32, t5 # e0 : 34 or t5, a1, a1 # e0 :
|
A D | stxcpy.S | 239 and a1, 7, t5 # e0 : find src misalignment 256 cmplt t4, t5, t12 # e0 : 260 mskqh t2, t5, t2 # e0 : 275 and a1, 7, t5 # .. e1 : 278 srl t12, t5, t12 # e0 : adjust final null return value
|
A D | ev6-stxcpy.S | 269 and a1, 7, t5 # E : find src misalignment 287 cmplt t4, t5, t12 # E : 291 mskqh t2, t5, t2 # U : 304 and a1, 7, t5 # E : 308 srl t12, t5, t12 # U : adjust final null return value
|
/linux-6.3-rc2/arch/riscv/lib/ |
A D | memmove.S | 68 andi t5, t3, -SZREG 79 beq t5, t3, 1f 80 addi t5, t5, SZREG 165 addi a2, t5, -SZREG /* The other breakpoint for the unrolled loop*/ 213 bne t4, t5, 1b 215 mv t4, t5 /* Fix the dest pointer in case the loop was broken */ 248 bne t4, t5, 1b 262 beq t3, t5, 2f 268 bne t3, t5, 1b
|
A D | strncmp.S | 70 li t5, -1 85 bne t3, t5, 2f 87 bne t3, t5, 2f
|
/linux-6.3-rc2/arch/arm64/crypto/ |
A D | crct10dif-ce-core.S | 85 t5 .req v19 137 ext t5.8b, ad.8b, ad.8b, #2 // A2 142 pmull t5.8h, t5.8b, fold_consts.8b // H = A2*B 151 tbl t5.16b, {ad.16b}, perm2.16b // A2 156 pmull2 t5.8h, t5.16b, fold_consts.16b // H = A2*B 163 eor t5.16b, t5.16b, t7.16b // M = G + H 166 uzp1 t8.2d, t4.2d, t5.2d 167 uzp2 t4.2d, t4.2d, t5.2d 184 zip2 t5.2d, t8.2d, t4.2d 190 ext t5.16b, t5.16b, t5.16b, #14 [all …]
|
A D | ghash-ce-core.S | 27 t5 .req v12 73 ext t5.8b, \ad\().8b, \ad\().8b, #2 // A2 81 tbl t5.16b, {\ad\().16b}, perm2.16b // A2 102 pmull\t t5.8h, t5.\nb, \bd // H = A2*B 110 eor t5.16b, t5.16b, t6.16b // M = G + H 113 uzp1 t4.2d, t3.2d, t5.2d 114 uzp2 t3.2d, t3.2d, t5.2d 131 zip2 t5.2d, t4.2d, t3.2d 137 ext t5.16b, t5.16b, t5.16b, #14 141 eor t3.16b, t3.16b, t5.16b
|
/linux-6.3-rc2/lib/zlib_dfltcc/ |
A D | dfltcc_util.h | 39 size_t t5 = len2 ? *len2 : 0; in dfltcc() local 45 register size_t r5 __asm__("r5") = t5; in dfltcc() 60 t2 = r2; t3 = r3; t4 = r4; t5 = r5; in dfltcc() 69 *len2 = t5; in dfltcc()
|
/linux-6.3-rc2/arch/riscv/include/asm/ |
A D | compat.h | 51 compat_ulong_t t5; member 88 cregs->t5 = (compat_ulong_t) regs->t5; in regs_to_cregs() 125 regs->t5 = (unsigned long) cregs->t5; in cregs_to_regs()
|
/linux-6.3-rc2/crypto/ |
A D | ecc.c | 1085 u64 t5[ECC_MAX_DIGITS]; in ecc_point_double_jacobian() local 1133 vli_mod_sub(t5, t5, z1, curve_prime, ndigits); in ecc_point_double_jacobian() 1186 u64 t5[ECC_MAX_DIGITS]; in xycz_add() local 1193 vli_mod_square_fast(t5, t5, curve); in xycz_add() 1201 vli_mod_square_fast(t5, y2, curve); in xycz_add() 1204 vli_mod_sub(t5, t5, x1, curve_prime, ndigits); in xycz_add() 1206 vli_mod_sub(t5, t5, x2, curve_prime, ndigits); in xycz_add() 1218 vli_set(x2, t5, ndigits); in xycz_add() 1229 u64 t5[ECC_MAX_DIGITS]; in xycz_add_c() local 1238 vli_mod_square_fast(t5, t5, curve); in xycz_add_c() [all …]
|
/linux-6.3-rc2/arch/loongarch/mm/ |
A D | page.S | 49 ld.d t5, a1, 40 62 st.d t5, a0, 40 64 ld.d t5, a1, 104 77 st.d t5, a0, -24
|
/linux-6.3-rc2/scripts/ |
A D | makelst | 28 t5=`field 1 $t1` 29 t6=`printf "%lu" $((0x$t4 - 0x$t5))`
|
/linux-6.3-rc2/arch/mips/kernel/ |
A D | scall32-o32.S | 61 load_a4: user_lw(t5, 16(t0)) # argument #5 from usp 67 sw t5, 16(sp) # argument #5 to ksp 154 li t5, 0 193 lw t5, 24(sp) 196 sw t5, 20(sp)
|
/linux-6.3-rc2/arch/riscv/kernel/ |
A D | mcount.S | 96 REG_L t5, 0(t3) 97 bne t5, t4, do_trace 126 jalr t5
|
/linux-6.3-rc2/arch/ia64/lib/ |
A D | copy_page_mck.S | 81 #define t5 t1 // alias! macro 84 #define t9 t5 // alias! 153 (p[D]) ld8 t5 = [src0], 8 160 (p[D]) st8 [dst0] = t5, 8
|
A D | copy_page.S | 45 t5[PIPE_DEPTH], t6[PIPE_DEPTH], t7[PIPE_DEPTH], t8[PIPE_DEPTH] 82 (p[0]) ld8 t5[0]=[src1],16 83 (EPI) st8 [tgt1]=t5[PIPE_DEPTH-1],16
|
/linux-6.3-rc2/arch/x86/crypto/ |
A D | camellia-aesni-avx2-asm_64.S | 86 filter_8bit(x0, t5, t6, t7, t4); \ 87 filter_8bit(x7, t5, t6, t7, t4); \ 94 filter_8bit(x2, t5, t6, t7, t4); \ 95 filter_8bit(x5, t5, t6, t7, t4); \ 96 filter_8bit(x1, t5, t6, t7, t4); \ 103 vextracti128 $1, x5, t5##_x; \ 124 vaesenclast t4##_x, t5##_x, t5##_x; \ 125 vinserti128 $1, t5##_x, x5, x5; \ 168 vpsrldq $5, t0, t5; \ 177 vpshufb t7, t5, t5; \ [all …]
|
/linux-6.3-rc2/arch/riscv/kvm/ |
A D | vcpu_switch.S | 49 REG_L t5, (KVM_ARCH_GUEST_SEPC)(a0) 67 csrw CSR_SEPC, t5 105 REG_L t5, (KVM_ARCH_GUEST_T5)(a0) 149 REG_S t5, (KVM_ARCH_GUEST_T5)(a0) 157 REG_L t5, (KVM_ARCH_HOST_SSTATUS)(a0) 175 csrrw t5, CSR_SSTATUS, t5 182 REG_S t5, (KVM_ARCH_GUEST_SSTATUS)(a0)
|
/linux-6.3-rc2/arch/loongarch/lib/ |
A D | memcpy.S | 67 ld.d t5, a1, 40 75 st.d t5, a0, 40
|
A D | copy_user.S | 73 6: ld.d t5, a1, 40 81 14: st.d t5, a0, 40
|
A D | memmove.S | 94 ld.d t5, a1, -48 102 st.d t5, a0, -48
|
/linux-6.3-rc2/arch/loongarch/kernel/ |
A D | kprobes_trampoline.S | 23 cfi_st t5, PT_R17 58 cfi_ld t5, PT_R17
|