/linux-6.3-rc2/arch/powerpc/crypto/ |
A D | sha1-powerpc-asm.S | 58 xor r5,W((t)+4-3),W((t)+4-8); \ 60 xor W((t)+4),W((t)+4-16),W((t)+4-14); \ 62 xor W((t)+4),W((t)+4),r5; \ 64 rotlwi W((t)+4),W((t)+4),1 84 xor W((t)+4),W((t)+4-16),W((t)+4-14); \ 86 xor W((t)+4),W((t)+4),r5; \ 88 rotlwi W((t)+4),W((t)+4),1 97 xor r5,W((t)+4-3),W((t)+4-8); \ 99 xor W((t)+4),W((t)+4-16),W((t)+4-14); \ 103 xor W((t)+4),W((t)+4),r5; \ [all …]
|
/linux-6.3-rc2/lib/crypto/ |
A D | sha256.c | 62 W[I] = s1(W[I-2]) + W[I-7] + s0(W[I-15]) + W[I-16]; in BLEND_OP() 92 BLEND_OP(i + 0, W); in sha256_transform() 93 BLEND_OP(i + 1, W); in sha256_transform() 94 BLEND_OP(i + 2, W); in sha256_transform() 95 BLEND_OP(i + 3, W); in sha256_transform() 96 BLEND_OP(i + 4, W); in sha256_transform() 97 BLEND_OP(i + 5, W); in sha256_transform() 98 BLEND_OP(i + 6, W); in sha256_transform() 99 BLEND_OP(i + 7, W); in sha256_transform() 126 u32 W[64]; in sha256_update() local [all …]
|
A D | sha1.c | 40 #define setW(x, val) (*(volatile __u32 *)&W(x) = (val)) 42 #define setW(x, val) do { W(x) = (val); __asm__("":::"memory"); } while (0) 44 #define setW(x, val) (W(x) = (val)) 48 #define W(x) (array[(x)&15]) macro 55 #define SHA_MIX(t) rol32(W(t+13) ^ W(t+8) ^ W(t+2) ^ W(t), 1)
|
/linux-6.3-rc2/Documentation/translations/zh_CN/loongarch/ |
A D | introduction.rst | 205 ADD.W SUB.W ADDI.W ADD.D SUB.D ADDI.D 208 MUL.W MULH.W MULH.WU DIV.W DIV.WU MOD.W MOD.WU 215 SLL.W SRL.W SRA.W ROTR.W SLLI.W SRLI.W SRAI.W ROTRI.W 220 EXT.W.B EXT.W.H CLO.W CLO.D SLZ.W CLZ.D CTO.W CTO.D CTZ.W CTZ.D 221 BYTEPICK.W BYTEPICK.D BSTRINS.W BSTRINS.D BSTRPICK.W BSTRPICK.D 222 REVB.2H REVB.4H REVB.2W REVB.D REVH.2W REVH.D BITREV.4B BITREV.8B BITREV.W BITREV.D 231 LD.B LD.BU LD.H LD.HU LD.W LD.WU LD.D ST.B ST.H ST.W ST.D 233 LDPTR.W LDPTR.D STPTR.W STPTR.D 238 LL.W SC.W LL.D SC.D 239 AMSWAP.W AMSWAP.D AMADD.W AMADD.D AMAND.W AMAND.D AMOR.W AMOR.D AMXOR.W AMXOR.D [all …]
|
/linux-6.3-rc2/arch/x86/kernel/ |
A D | uprobes.c | 89 W(0x00, 1, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 1, 1, 1, 1, 1) | /* 00 */ 90 W(0x10, 1, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 1, 1, 1, 1, 0) , /* 10 */ 91 W(0x20, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) | /* 20 */ 92 W(0x30, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) , /* 30 */ 93 W(0x40, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) | /* 40 */ 94 W(0x50, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) , /* 50 */ 95 W(0x60, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0) | /* 60 */ 96 W(0x70, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) , /* 70 */ 97 W(0x80, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) | /* 80 */ 98 W(0x90, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) , /* 90 */ [all …]
|
/linux-6.3-rc2/arch/x86/crypto/ |
A D | sha1_ssse3_asm.S | 387 psrld $31, W 392 pslld $2, W 412 pxor W_minus_28, W # W is W_minus_32 before xor 419 psrld $30, W 503 vpxor W_minus_08, W, W 506 vpxor W_TMP1, W, W 510 vpsrld $31, W, W 526 vpxor W_minus_28, W, W # W is W_minus_32 before xor 529 vpxor W_TMP1, W, W 532 vpsrld $30, W, W [all …]
|
A D | sha512-ssse3-asm.S | 163 # Eg. XMM2=W[t-2] really means XMM2={W[t-2]|W[t-1]} 201 psrlq $(7-1), %xmm3 # XMM3 = ((W[t-15]>>1)^W[t-15])>>6 204 pxor %xmm2, %xmm0 # XMM0 = (((W[t-2]>>42)^W[t-2])>>13)^W[t-2] 207 pxor %xmm5, %xmm3 # XMM3 = (((W[t-15]>>1)^W[t-15])>>6)^W[t-15] 210 psrlq $6, %xmm0 # XMM0 = ((((W[t-2]>>42)^W[t-2])>>13)^W[t-2])>>6 213 psrlq $1, %xmm3 # XMM3 = (((W[t-15]>>1)^W[t-15])>>6)^W[t-15]>>1 229 pxor %xmm2, %xmm1 # XMM1 = (W[t-2] << 42)^W[t-2] 232 pxor %xmm5, %xmm4 # XMM4 = (W[t-15]<<7)^W[t-15] 249 paddq %xmm3, %xmm0 # XMM0 = s1(W[t-2]) + s0(W[t-15]) 252 paddq W_t(idx), %xmm0 # XMM0 = s1(W[t-2]) + s0(W[t-15]) + W[t-16] [all …]
|
A D | sha512-avx-asm.S | 78 # W[t] + K[t] | W[t+1] + K[t+1] 168 # Eg. XMM4=W[t-2] really means XMM4={W[t-2]|W[t-1]} 185 vpxor %xmm1, %xmm0, %xmm0 # XMM0 = W[t-2]>>61 ^ W[t-2]>>19 195 vpxor %xmm7, %xmm6, %xmm6 # XMM6 = W[t-15]>>1 ^ W[t-15]>>8 204 vpxor %xmm3, %xmm2, %xmm2 # XMM2 = W[t-2]>>6 ^ W[t-2]<<3 209 vpxor %xmm9, %xmm8, %xmm8 # XMM8 = W[t-15]>>7 ^ W[t-15]<<63 214 vpxor %xmm8, %xmm6, %xmm6 # XMM6 = W[t-15]>>1 ^ W[t-15]>>8 ^ 215 # W[t-15]>>7 ^ W[t-15]<<63 240 vpaddq %xmm6, %xmm0, %xmm0 # XMM0 = s1(W[t-2]) + W[t-16] + s0(W[t-15]) 242 vpaddq %xmm1, %xmm0, %xmm0 # XMM0 = W[t] = s1(W[t-2]) + W[t-7] + [all …]
|
A D | sha512-avx2-asm.S | 169 vpaddq Y_0, YTMP0, YTMP0 # YTMP0 = W[-7] + W[-16] 226 vpxor YTMP4, YTMP3, YTMP3 # YTMP3 = W[-15] ror 1 ^ W[-15] >> 7 231 vpaddq YTMP1, YTMP0, YTMP0 # YTMP0 = W[-16] + W[-7] + s0 233 vperm2f128 $0x0, YTMP0, YTMP0, Y_0 # Y_0 = W[-16] + W[-7] + s0 {BABA} 289 vpxor YTMP3, YTMP4, YTMP4 # YTMP4 = W[-2] ror 19 ^ W[-2] >> 6 {BABA} 294 # (W[-2] ror 61) ^ (W[-2] >> 6) {BABA} 297 vpaddq YTMP4, Y_0, Y_0 # Y_0 = {W[1], W[0], W[1], W[0]} 345 vpxor YTMP3, YTMP4, YTMP4 # YTMP4 = W[-2] ror 19 ^ W[-2] >> 6 {DC--} 350 # (W[-2] ror 61) ^ (W[-2] >> 6) {DC--} 354 vpaddq YTMP4, YTMP0, YTMP2 # YTMP2 = {W[3], W[2], --, --} [all …]
|
A D | sha256-ssse3-asm.S | 150 ## compute W[-16] + W[-7] 4 at a time 155 palignr $4, X2, XTMP0 # XTMP0 = W[-7] 163 paddd X0, XTMP0 # XTMP0 = W[-7] + W[-16] 168 palignr $4, X0, XTMP1 # XTMP1 = W[-15] 172 movdqa XTMP1, XTMP2 # XTMP2 = W[-15] 176 movdqa XTMP1, XTMP3 # XTMP3 = W[-15] 192 movdqa XTMP3, XTMP2 # XTMP2 = W[-15] 216 pxor XTMP2, XTMP1 # XTMP1 = W[-15] ror 7 ^ W[-15] ror 18 228 paddd XTMP1, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 265 paddd XTMP4, XTMP0 # XTMP0 = {..., ..., W[1], W[0]} [all …]
|
A D | sha256-avx-asm.S | 156 ## compute W[-16] + W[-7] 4 at a time 161 vpalignr $4, X2, X3, XTMP0 # XTMP0 = W[-7] 168 vpaddd X0, XTMP0, XTMP0 # XTMP0 = W[-7] + W[-16] 173 vpalignr $4, X0, X1, XTMP1 # XTMP1 = W[-15] 190 vpor XTMP2, XTMP3, XTMP3 # XTMP1 = W[-15] MY_ROR 7 204 vpsrld $3, XTMP1, XTMP4 # XTMP4 = W[-15] >> 3 216 vpxor XTMP2, XTMP3, XTMP3 # XTMP1 = W[-15] MY_ROR 7 ^ W[-15] MY_ROR 225 vpshufd $0b11111010, X3, XTMP2 # XTMP2 = W[-2] {BBAA} 228 vpaddd XTMP1, XTMP0, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 262 vpaddd XTMP4, XTMP0, XTMP0 # XTMP0 = {..., ..., W[1], W[0]} [all …]
|
/linux-6.3-rc2/tools/bpf/bpftool/bash-completion/ |
A D | bpftool | 22 COMPREPLY+=( $( compgen -W "$w" -- "$cur" ) ) 44 COMPREPLY+=( $( compgen -W "$*" -- "$cur" ) ) 49 COMPREPLY+=( $( compgen -W "$( bpftool -jp map 2>&1 | \ 57 COMPREPLY+=( $( compgen -W "$( bpftool -jp map 2>&1 | \ 64 COMPREPLY+=( $( compgen -W "$( bpftool -jp map 2>&1 | \ 72 COMPREPLY+=( $( compgen -W "$( bpftool -jp map 2>&1 | \ 79 COMPREPLY+=( $( compgen -W "$( bpftool -jp prog 2>&1 | \ 85 COMPREPLY+=( $( compgen -W "$( bpftool -jp prog 2>&1 | \ 97 COMPREPLY+=( $( compgen -W "$( bpftool -jp btf 2>&1 | \ 116 COMPREPLY+=( $( compgen -W "$maps" -- "$cur" ) ) [all …]
|
/linux-6.3-rc2/arch/arm/crypto/ |
A D | sha1-armv7-neon.S | 218 veor.32 W, W, W_m08; \ 222 veor W, W, tmp0; \ 225 vshl.u32 tmp0, W, #1; \ 229 vshr.u32 W, W, #31; \ 254 veor W, W_m28; \ 260 veor W, W_m16; \ 263 veor W, tmp0; \ 269 vshl.u32 tmp1, W, #2; \ 272 vshr.u32 tmp0, W, #30; \ 275 vorr W, tmp0, tmp1; \ [all …]
|
/linux-6.3-rc2/crypto/ |
A D | sha512_generic.c | 88 static inline void LOAD_OP(int I, u64 *W, const u8 *input) in LOAD_OP() argument 90 W[I] = get_unaligned_be64((__u64 *)input + I); in LOAD_OP() 93 static inline void BLEND_OP(int I, u64 *W) in BLEND_OP() argument 95 W[I & 15] += s1(W[(I-2) & 15]) + W[(I-7) & 15] + s0(W[(I-15) & 15]); in BLEND_OP() 104 u64 W[16]; in sha512_transform() local 118 LOAD_OP(i + j, W, input); in sha512_transform() 121 BLEND_OP(i + j, W); in sha512_transform() 126 t1 = h + e1(e) + Ch(e,f,g) + sha512_K[i ] + W[(i & 15)]; in sha512_transform() 128 t1 = g + e1(d) + Ch(d,e,f) + sha512_K[i+1] + W[(i & 15) + 1]; in sha512_transform() 130 t1 = f + e1(c) + Ch(c,d,e) + sha512_K[i+2] + W[(i & 15) + 2]; in sha512_transform() [all …]
|
A D | sm3.c | 64 #define W1(i) (W[i & 0x0f]) 65 #define W2(i) (W[i & 0x0f] = \ 66 P1(W[i & 0x0f] \ 67 ^ W[(i-9) & 0x0f] \ 69 ^ rol32(W[(i-13) & 0x0f], 7) \ 70 ^ W[(i-6) & 0x0f]) 173 sm3_transform(sctx, data, W); in sm3_block() 181 u32 W[16]; in sm3_update() local 206 memzero_explicit(W, sizeof(W)); in sm3_update() 221 u32 W[16]; in sm3_final() local [all …]
|
/linux-6.3-rc2/Documentation/loongarch/ |
A D | introduction.rst | 238 ADD.W SUB.W ADDI.W ADD.D SUB.D ADDI.D 241 MUL.W MULH.W MULH.WU DIV.W DIV.WU MOD.W MOD.WU 248 SLL.W SRL.W SRA.W ROTR.W SLLI.W SRLI.W SRAI.W ROTRI.W 253 EXT.W.B EXT.W.H CLO.W CLO.D SLZ.W CLZ.D CTO.W CTO.D CTZ.W CTZ.D 254 BYTEPICK.W BYTEPICK.D BSTRINS.W BSTRINS.D BSTRPICK.W BSTRPICK.D 255 REVB.2H REVB.4H REVB.2W REVB.D REVH.2W REVH.D BITREV.4B BITREV.8B BITREV.W BITREV.D 264 LD.B LD.BU LD.H LD.HU LD.W LD.WU LD.D ST.B ST.H ST.W ST.D 266 LDPTR.W LDPTR.D STPTR.W STPTR.D 271 LL.W SC.W LL.D SC.D 272 AMSWAP.W AMSWAP.D AMADD.W AMADD.D AMAND.W AMAND.D AMOR.W AMOR.D AMXOR.W AMXOR.D [all …]
|
/linux-6.3-rc2/arch/arm/lib/ |
A D | memmove.S | 82 6: W(nop) 83 W(ldr) r3, [r1, #-4]! 84 W(ldr) r4, [r1, #-4]! 85 W(ldr) r5, [r1, #-4]! 86 W(ldr) r6, [r1, #-4]! 87 W(ldr) r8, [r1, #-4]! 88 W(ldr) r9, [r1, #-4]! 89 W(ldr) lr, [r1, #-4]! 93 W(nop) 94 W(str) r3, [r0, #-4]! [all …]
|
/linux-6.3-rc2/tools/memory-model/ |
A D | linux-kernel.def | 34 cmpxchg(X,V,W) __cmpxchg{mb}(X,V,W) 35 cmpxchg_relaxed(X,V,W) __cmpxchg{once}(X,V,W) 36 cmpxchg_acquire(X,V,W) __cmpxchg{acquire}(X,V,W) 37 cmpxchg_release(X,V,W) __cmpxchg{release}(X,V,W) 108 atomic_cmpxchg(X,V,W) __cmpxchg{mb}(X,V,W) 109 atomic_cmpxchg_relaxed(X,V,W) __cmpxchg{once}(X,V,W) 110 atomic_cmpxchg_acquire(X,V,W) __cmpxchg{acquire}(X,V,W) 111 atomic_cmpxchg_release(X,V,W) __cmpxchg{release}(X,V,W)
|
/linux-6.3-rc2/arch/m68k/fpsp040/ |
A D | slogn.S | 436 |--LET V=U*U, W=V*V, CALCULATE 438 |--U + U*V*( [B1 + W*(B3 + W*B5)] + [V*(B2 + W*B4)] ) 443 fmulx %fp1,%fp1 | ...FP1 IS W 448 fmulx %fp1,%fp3 | ...W*B5 449 fmulx %fp1,%fp2 | ...W*B4 451 faddd LOGB3,%fp3 | ...B3+W*B5 452 faddd LOGB2,%fp2 | ...B2+W*B4 454 fmulx %fp3,%fp1 | ...W*(B3+W*B5), FP3 RELEASED 458 faddd LOGB1,%fp1 | ...B1+W*(B3+W*B5) 461 faddx %fp2,%fp1 | ...B1+W*(B3+W*B5) + V*(B2+W*B4), FP2 RELEASED [all …]
|
/linux-6.3-rc2/arch/arm/boot/compressed/ |
A D | head.S | 213 W(b) 1f 1014 W(b) __armv4_mmu_cache_on 1089 W(b) __fa526_cache_on 1091 W(b) __fa526_cache_flush 1420 W(b) . @ reset 1421 W(b) . @ undef 1425 W(b) . @ svc 1427 W(b) . @ pabort 1428 W(b) . @ dabort 1430 W(b) . @ irq [all …]
|
/linux-6.3-rc2/arch/mips/n64/ |
A D | init.c | 51 #define W 320 macro 82 .width = W, in n64_platform_init() 84 .stride = W * 2, in n64_platform_init() 122 orig = kzalloc(W * H * 2 + 63, GFP_DMA | GFP_KERNEL); in n64_platform_init() 141 res[0].end = phys + W * H * 2 - 1; in n64_platform_init() 149 #undef W
|
/linux-6.3-rc2/arch/arm/kernel/ |
A D | entry-armv.S | 1132 3: W(b) . + 4 1300 W(b) vector_rst 1301 W(b) vector_und 1304 W(ldr) pc, . 1305 W(b) vector_pabt 1308 W(b) vector_irq 1309 W(b) vector_fiq 1313 W(b) vector_rst 1317 W(ldr) pc, . 1325 W(b) vector_rst [all …]
|
/linux-6.3-rc2/tools/memory-model/Documentation/ |
A D | cheatsheet.txt | 3 C Self R W RMW Self R W DR DW RMW SV 11 Successful *_release() C Y Y Y W Y 13 smp_wmb() Y W Y Y W 28 W: Write, for example, WRITE_ONCE(), or write portion of RMW
|
/linux-6.3-rc2/drivers/atm/ |
A D | Kconfig | 86 when going from 8W to 16W bursts. 96 bool "Enable 8W TX bursts (recommended)" 103 bool "Enable 4W TX bursts (optional)" 107 this if you have disabled 8W bursts. Enabling 4W if 8W is also set 111 bool "Enable 2W TX bursts (optional)" 115 this if you have disabled 4W and 8W bursts. Enabling 2W if 4W or 8W 126 bool "Enable 8W RX bursts (discouraged)" 134 bool "Enable 4W RX bursts (recommended)" 138 default setting. Enabling 4W if 8W is also set may or may not 142 bool "Enable 2W RX bursts (optional)" [all …]
|
/linux-6.3-rc2/ |
A D | CREDITS | 344 W: http://www.mulix.org 581 W: https://pburton.com 1045 W: www.farsite.co.uk 1688 N: Rob W. W. Hooft 2037 W: http://bytesex.org 2071 W: http://www.xos.nl/ 2103 W: http://vtun.sf.net 2539 N: Mark W. McClelland 2742 W: http://namei.org/ 2902 W: https://ojeda.dev [all …]
|