/lib/crypto/arm64/ |
A D | poly1305-armv8.pl | 78 mov $s1,#0xfffffffc0fffffff 79 movk $s1,#0x0fff,lsl#48 84 and $r0,$r0,$s1 // &=0ffffffc0fffffff 85 and $s1,$s1,#-4 87 mov w#$s1,#-1 165 mul $t0,$h1,$s1 // h1*5*r1 166 umulh $t1,$h1,$s1 179 mul $t0,$h2,$s1 // h2*5*r1 279 mul $t0,$h1,$s1 // h1*5*r1 280 umulh $t1,$h1,$s1 [all …]
|
A D | sha256-ce.S | 44 .macro add_update, ev, rc, s0, s1, s2, s3 45 sha256su0 v\s0\().4s, v\s1\().4s 46 add_only \ev, \rc, \s1
|
A D | sha1-ce-core.S | 52 .macro add_update, op, ev, rc, s0, s1, s2, s3, dg1 53 sha1su0 v\s0\().4s, v\s1\().4s, v\s2\().4s 54 add_only \op, \ev, \rc, \s1, \dg1
|
/lib/tests/ |
A D | string_helpers_kunit.c | 117 .s1 = {{ 131 .s1 = {{ 145 .s1 = {{ 196 .s1 = {{ 215 .s1 = {{ 241 .s1 = {{ 258 .s1 = {{ 266 .s1 = {{ 307 .s1 = {{ 375 const struct test_string_1 *s1 = s2->s1; in test_string_find_match() local [all …]
|
/lib/ |
A D | string.c | 49 c1 = *s1++; in strncasecmp() 71 c1 = tolower(*s1++); in strcasecmp() 757 return (char *)s1; in strstr() 758 l1 = strlen(s1); in strstr() 761 if (!memcmp(s1, s2, l2)) in strstr() 762 return (char *)s1; in strstr() 763 s1++; in strstr() 783 return (char *)s1; in strnstr() 786 if (!memcmp(s1, s2, l2)) in strnstr() 787 return (char *)s1; in strnstr() [all …]
|
A D | random32.c | 55 state->s1 = TAUSWORTHE(state->s1, 6U, 13U, 4294967294U, 18U); in prandom_u32_state() 60 return (state->s1 ^ state->s2 ^ state->s3 ^ state->s4); in prandom_u32_state() 119 state->s1 = __seed(seeds[0], 2U); in prandom_seed_full_state() 251 state->s1 = __seed(LCG(seed), 2U); in prandom_state_selftest_seed() 252 state->s2 = __seed(LCG(state->s1), 8U); in prandom_state_selftest_seed()
|
A D | string_helpers.c | 889 bool sysfs_streq(const char *s1, const char *s2) in sysfs_streq() argument 891 while (*s1 && *s1 == *s2) { in sysfs_streq() 892 s1++; in sysfs_streq() 896 if (*s1 == *s2) in sysfs_streq() 898 if (!*s1 && *s2 == '\n' && !s2[1]) in sysfs_streq() 900 if (*s1 == '\n' && !s1[1] && !*s2) in sysfs_streq()
|
/lib/crypto/arm/ |
A D | sha512-armv4.pl | 559 my ($t0,$t1,$s0,$s1) = map("q$_",(12..15)); # temps 567 vshr.u64 $s1,@X[($i+7)%8],#@sigma1[2] 571 veor $s1,$t0 573 veor $s1,$t1 @ sigma1(X[i+14]) 575 vadd.i64 @X[$i%8],$s1 576 vshr.u64 $s1,$s0,#@sigma0[2] 580 veor $s1,$t0 584 veor $s1,$t1 @ sigma0(X[i+1]) 586 vadd.i64 @X[$i%8],$s1
|
A D | sha256-ce.S | 44 .macro add_update, ev, s0, s1, s2, s3 45 sha256su0.32 q\s0, q\s1 46 add_only \ev, \s1
|
A D | sha1-ce-core.S | 48 .macro add_update, op, ev, rc, s0, s1, s2, s3, dg1 49 sha1su0.32 q\s0, q\s1, q\s2 50 add_only \op, \ev, \rc, \s1, \dg1
|
A D | blake2s-core.S | 68 .macro _blake2s_quarterround a0, b0, c0, d0, a1, b1, c1, d1, s0, s1, s2, s3 91 ldr M_0, [sp, #32 + 4 * \s1] 128 .macro _blake2s_round s0, s1, s2, s3, s4, s5, s6, s7, \ 135 \s0, \s1, \s2, \s3
|
A D | poly1305-armv4.pl | 178 my ($s1,$s2,$s3)=($r1,$r2,$r3); 274 add $s1,$r1,$r1,lsr#2 292 add $s1,$r1,$r1,lsr#2 302 umlal r2,r3,$h4,$s1 303 umlal r0,r1,$h3,$s1
|
/lib/crypto/powerpc/ |
A D | poly1305-p10le_64.S | 25 …22/21 - this revison based on the above sum of products. Setup r^4, r^3, r^2, r and s3, s2, s1, s0 857 add 19, 21, 10 # s1: r19 - (r1 >> 2) *5 861 mtvsrdd 32+0, 9, 19 # r0, s1 863 mtvsrdd 32+2, 19, 25 # s1 871 # v0 = (r0, s1), v1 = (r1, r0), v2 = s1, v3 = r0 877 # d0 = h0 * r0 + h1 * s1 878 vmsumudm 7, 6, 0, 9 # h0 * r0, h1 * s1 880 # d1 = h0 * r1 + h1 * r0 + h2 * s1 882 vmsumudm 10, 8, 2, 11 # d1 += h2 * s1 917 # d0 = h0 * r0 + h1 * s1 [all …]
|
/lib/crypto/x86/ |
A D | poly1305-x86_64-cryptogams.pl | 211 mov $s1,%rax 219 imulq $s1,$h1 # h2*s1 346 mov $s1,$r1 347 shr \$2,$s1 689 mov $s1,$r1 690 mov $s1,%rax 691 shr \$2,$s1 792 mov $s1,$r1 794 shr \$2,$s1 1611 mov $s1,$r1 [all …]
|
A D | sha256-avx-asm.S | 154 ## compute s0 four at a time and s1 two at a time 223 ## compute low s1 253 vpxor XTMP2, XTMP4, XTMP4 # XTMP4 = s1 {xBxA} 257 vpshufb SHUF_00BA, XTMP4, XTMP4 # XTMP4 = s1 {00BA} 264 ## compute high s1 291 vpxor XTMP2, XTMP5, XTMP5 # XTMP5 = s1 {xDxC} 295 vpshufb SHUF_DC00, XTMP5, XTMP5 # XTMP5 = s1 {DC00}
|
A D | sha256-ssse3-asm.S | 148 ## compute s0 four at a time and s1 two at a time 223 ## compute low s1 256 pxor XTMP2, XTMP4 # XTMP4 = s1 {xBxA} 260 pshufb SHUF_00BA, XTMP4 # XTMP4 = s1 {00BA} 267 ## compute high s1 297 pxor XTMP2, X0 # X0 = s1 {xDxC} 301 pshufb SHUF_DC00, X0 # X0 = s1 {DC00}
|
A D | sha512-avx-asm.S | 223 vpxor %xmm2, %xmm0, %xmm0 # XMM0 = s1(W[t-2]) 227 vpaddq W_t(idx), %xmm0, %xmm0 # XMM0 = s1(W[t-2]) + W[t-16] 239 vpaddq %xmm6, %xmm0, %xmm0 # XMM0 = s1(W[t-2]) + W[t-16] + s0(W[t-15]) 241 vpaddq %xmm1, %xmm0, %xmm0 # XMM0 = W[t] = s1(W[t-2]) + W[t-7] +
|
A D | sha512-ssse3-asm.S | 240 pxor %xmm1, %xmm0 # XMM0 = s1(W[t-2]) 248 paddq %xmm3, %xmm0 # XMM0 = s1(W[t-2]) + s0(W[t-15]) 251 paddq W_t(idx), %xmm0 # XMM0 = s1(W[t-2]) + s0(W[t-15]) + W[t-16] 253 paddq %xmm1, %xmm0 # XMM0 = s1(W[t-2]) + W[t-7] + s0(W[t-15]) + W[t-16]
|
A D | chacha-ssse3-x86_64.S | 142 # o1 = i1 ^ (x1 + s1) 522 # x4[0-3] += s1[0] 523 # x5[0-3] += s1[1] 529 # x6[0-3] += s1[2] 530 # x7[0-3] += s1[3]
|
/lib/crypto/ |
A D | poly1305-donna64.c | 40 u64 s1, s2; in poly1305_core_blocks() local 58 s1 = key->precomputed_s.r64[0]; in poly1305_core_blocks() 76 d = (u128)h2 * s1; in poly1305_core_blocks()
|
A D | poly1305-donna32.c | 38 u32 s1, s2, s3, s4; in poly1305_core_blocks() local 54 s1 = key->precomputed_s.r[0]; in poly1305_core_blocks() 76 ((u64)h4 * s1); in poly1305_core_blocks()
|
/lib/crypto/mips/ |
A D | poly1305-mips.pl | 42 ($s0,$s1,$s2,$s3,$s4,$s5,$s6,$s7,$s8,$s9,$s10,$s11)=map("\$$_",(12..23)); 227 ($s0,$s1,$s2,$s3,$s4,$s5,$in0,$in1,$t2); 264 sd $s1,8($sp) 433 ld $s1,8($sp) 734 ($s0,$s1,$s2,$s3,$s4, $s5,$s6,$s7,$s8, $s9,$s10,$s11); 761 sw $s1, 4*1($sp) 927 maddu $rs1,$d3 # d3*s1 936 maddu $rs1,$h4 # h4*s1 985 multu ($rs1,$d3) # d3*s1 1025 multu ($rs1,$h4) # h4*s1 [all …]
|
A D | chacha-core.S | 28 #define T0 $s1 216 sw $s1, 4($sp) 295 lw $s1, 4($sp)
|
/lib/raid6/ |
A D | rvv.h | 26 int s1, int s2, \
|
/lib/crypto/riscv/ |
A D | chacha-riscv64-zvkb.S | 64 #define KEY1 s1 147 sd s1, 8(sp) 284 ld s1, 8(sp)
|