/linux-6.3-rc2/arch/arm/crypto/ |
A D | sha1-armv4-large.S | 77 mov r5,r5,ror#30 78 mov r6,r6,ror#30 242 mov r9,r9,ror#31 259 mov r9,r9,ror#31 276 mov r9,r9,ror#31 293 mov r9,r9,ror#31 314 mov r9,r9,ror#31 330 mov r9,r9,ror#31 346 mov r9,r9,ror#31 362 mov r9,r9,ror#31 [all …]
|
A D | chacha-scalar-core.S | 81 add \c1, \c1, \d1, ror #16 82 add \c2, \c2, \d2, ror #16 88 add \a1, \a1, \b1, ror #20 89 add \a2, \a2, \b2, ror #20 90 eor \d1, \a1, \d1, ror #16 91 eor \d2, \a2, \d2, ror #16 95 add \c1, \c1, \d1, ror #24 430 ror X12, X12, #drot 431 ror X13, X13, #drot 433 ror X14, X14, #drot [all …]
|
A D | blake2s-core.S | 84 add \c0, \c0, \d0, ror #16 85 add \c1, \c1, \d1, ror #16 95 add \a0, \a0, \b0, ror #12 96 add \a1, \a1, \b1, ror #12 101 eor \d0, \a0, \d0, ror#16 102 eor \d1, \a1, \d1, ror#16 105 add \c0, \c0, \d0, ror#8 106 add \c1, \c1, \d1, ror#8 109 eor \b0, \c0, \b0, ror#12 110 eor \b1, \c1, \b1, ror#12 [all …]
|
A D | aes-cipher-core.S | 60 eor \out0, \out0, t0, ror #24 82 eor \out1, \out1, t1, ror #24 83 eor \out0, \out0, t2, ror #16 85 eor \out1, \out1, \t3, ror #16 86 eor \out0, \out0, t0, ror #8 87 eor \out1, \out1, \t4, ror #8
|
A D | sha1-armv7-neon.S | 92 add e, e, a, ror #(32 - 5); \ 97 ror b, #(32 - 30); \ 106 add e, e, a, ror #(32 - 5); \ 110 ror b, #(32 - 30); \ 120 add e, e, a, ror #(32 - 5); \ 125 ror b, #(32 - 30); \
|
/linux-6.3-rc2/arch/arm/lib/ |
A D | bswapsdi2.S | 19 eor r3, r0, r0, ror #16 22 eor r0, r3, r0, ror #8 28 eor r3, ip, ip, ror #16 29 eor r1, r0, r0, ror #16 34 eor r1, r1, r0, ror #8 35 eor r0, r3, ip, ror #8
|
A D | csumpartial.S | 39 movne sum, sum, ror #8 74 movne r0, r0, ror #8 @ rotate checksum by 8 bits 105 movne sum, sum, ror #8
|
A D | io-readsw-armv4.S | 99 _BE_ONLY_( mov ip, ip, ror #8 ) 122 _BE_ONLY_( movne ip, ip, ror #8 )
|
/linux-6.3-rc2/arch/x86/crypto/ |
A D | sha256-ssse3-asm.S | 153 ror $(25-11), y0 # y0 = e >> (25-11) 156 ror $(22-13), y1 # y1 = a >> (22-13) 196 ror $(25-11), y0 # y0 = e >> (25-11) 199 ror $(22-13), y1 # y1 = a >> (22-13) 216 pxor XTMP2, XTMP1 # XTMP1 = W[-15] ror 7 ^ W[-15] ror 18 236 ror $(25-11), y0 # y0 = e >> (25-11) 239 ror $(22-13), y1 # y1 = a >> (22-13) 276 ror $(25-11), y0 # y0 = e >> (25-11) 279 ror $(22-13), y1 # y1 = a >> (22-13) 319 ror $(25-11), y0 # y0 = e >> (25-11) [all …]
|
A D | sha512-ssse3-asm.S | 121 ror $23, tmp0 # 41 # tmp = e ror 23 127 ror $4, tmp0 # 18 # tmp = ((e ror 23) ^ e) ror 4 128 xor e_64, tmp0 # tmp = (((e ror 23) ^ e) ror 4) ^ e 139 ror $5, tmp0 # 39 # tmp = a ror 5 142 ror $6, tmp0 # 34 # tmp = ((a ror 5) ^ a) ror 6 143 xor a_64, tmp0 # tmp = (((a ror 5) ^ a) ror 6) ^ a 185 ror $4, tmp0 # 18 203 ror $5, tmp0 # 39 206 ror $6, tmp0 # 34 230 ror $4, tmp0 # 18 [all …]
|
A D | twofish-i586-asm_32.S | 74 ror $16, b ## D;\ 77 ror $16, a ## D;\ 84 ror $15, b ## D;\ 112 ror $16, b ## D;\ 115 ror $16, a ## D;\ 122 ror $16, b ## D;\ 131 ror $1, c ## D;\ 149 ror $16, a ## D;\ 208 ror $1, d ## D; 225 ror $16, %eax [all …]
|
A D | twofish-x86_64-asm_64.S | 72 ror $16, b ## D;\ 75 ror $16, a ## D;\ 82 ror $15, b ## D;\ 110 ror $16, b ## D;\ 113 ror $16, a ## D;\ 128 ror $1, c ## D;\ 145 ror $16, a ## D;\ 148 ror $16, b ## D;\ 155 ror $15, a ## D;\ 181 ror $16, b ## D;\ [all …]
|
A D | sha512-avx-asm.S | 118 # shld is faster than ror on Sandybridge 127 RORQ tmp0, 23 # 41 # tmp = e ror 23 129 xor e_64, tmp0 # tmp = (e ror 23) ^ e 133 RORQ tmp0, 4 # 18 # tmp = ((e ror 23) ^ e) ror 4 134 xor e_64, tmp0 # tmp = (((e ror 23) ^ e) ror 4) ^ e 145 RORQ tmp0, 5 # 39 # tmp = a ror 5 146 xor a_64, tmp0 # tmp = (a ror 5) ^ a 148 RORQ tmp0, 6 # 34 # tmp = ((a ror 5) ^ a) ror 6 149 xor a_64, tmp0 # tmp = (((a ror 5) ^ a) ror 6) ^ a
|
A D | sha512-avx2-asm.S | 175 # Calculate w[t-15] ror 1 178 vpor YTMP2, YTMP3, YTMP3 # YTMP3 = W[-15] ror 1 221 # Calculate w[t-15] ror 8 224 vpor YTMP2, YTMP1, YTMP1 # YTMP1 = W[-15] ror 8 288 vpor YTMP1, YTMP3, YTMP3 # YTMP3 = W[-2] ror 19 {BABA} 292 vpor YTMP1, YTMP3, YTMP3 # YTMP3 = W[-2] ror 61 {BABA} 293 vpxor YTMP3, YTMP4, YTMP4 # YTMP4 = s1 = (W[-2] ror 19) ^ 294 # (W[-2] ror 61) ^ (W[-2] >> 6) {BABA} 344 vpor YTMP1, YTMP3, YTMP3 # YTMP3 = W[-2] ror 19 {DC--} 348 vpor YTMP1, YTMP3, YTMP3 # YTMP3 = W[-2] ror 61 {DC--} [all …]
|
A D | sha256-avx2-asm.S | 194 vpor XTMP2, XTMP3, XTMP3 # XTMP3 = W[-15] ror 7 234 vpxor XTMP2, XTMP3, XTMP3 # XTMP3 = W[-15] ror 7 ^ W[-15] ror 18 262 vpsrlq $19, XTMP2, XTMP3 # XTMP3 = W[-2] ror 19 {xBxA} 270 vpsrlq $17, XTMP2, XTMP2 # XTMP2 = W[-2] ror 17 {xBxA} 321 vpsrlq $19, XTMP2, XTMP3 # XTMP3 = W[-2] ror 19 {xDxC} 327 vpsrlq $17, XTMP2, XTMP2 # XTMP2 = W[-2] ror 17 {xDxC}
|
/linux-6.3-rc2/arch/arm64/crypto/ |
A D | chacha-neon-core.S | 300 ror a4, a4, #20 302 ror a5, a5, #20 304 ror a6, a6, #20 306 ror a7, a7, #20 367 ror a4, a4, #25 369 ror a5, a5, #25 371 ror a6, a6, #25 373 ror a7, a7, #25 434 ror a5, a5, #20 436 ror a6, a6, #20 [all …]
|
A D | aes-cipher-core.S | 66 eor \out0, \out0, w14, ror #24 67 eor \out1, \out1, w15, ror #24 68 eor \out0, \out0, w16, ror #16 69 eor \out1, \out1, w17, ror #16 70 eor \out0, \out0, \t0, ror #8 71 eor \out1, \out1, \t1, ror #8
|
/linux-6.3-rc2/arch/arc/lib/ |
A D | strlen.S | 18 ror r5,r4 28 ror r5,r4
|
A D | strchr-700.S | 30 ror r4,r3 55 ror r4,r3
|
A D | strcpy-700.S | 26 ror r12,r8
|
A D | strcmp-archs.S | 17 ror r11, r12
|
/linux-6.3-rc2/arch/arm64/include/asm/ |
A D | kvm_mmu.h | 68 ror \reg, \reg, #1 /* rotate to the first tag bit */ 71 ror \reg, \reg, #63 /* rotate back */
|
A D | asm-uaccess.h | 32 ror \tmp2, \tmp2, #16
|
/linux-6.3-rc2/arch/openrisc/ |
A D | Kconfig | 134 bool "Have instruction l.ror for rotate right" 137 This config enables gcc to generate l.ror instructions when compiling 142 l.ror rotate right instruction.
|
/linux-6.3-rc2/drivers/media/pci/cx23885/ |
A D | cx23888-ir.c | 514 int tsr, rsr, rto, ror, tse, rse, rte, roe, kror; in cx23888_ir_irq_handler() local 519 ror = stats & STATS_ROR; /* Rx FIFO Over Run */ in cx23888_ir_irq_handler() 529 rto ? "rto" : " ", ror ? "ror" : " ", in cx23888_ir_irq_handler() 594 if (roe && ror) { in cx23888_ir_irq_handler()
|