/lib/ |
A D | lru_cache.c | 288 return e && e->refcnt; in lc_is_used() 305 e->lc_number = e->lc_new_number = LC_FREE; in lc_del() 332 return e; in lc_prepare_for_change() 366 if (e) { in __lc_get() 367 if (e->lc_new_number != e->lc_number) { in __lc_get() 427 RETURN(e); in __lc_get() 532 e->lc_number = e->lc_new_number; in lc_committed() 554 BUG_ON(e->lc_number != e->lc_new_number); in lc_put() 596 if (e->lc_number != e->lc_new_number) in lc_seq_dump_details() 598 i, e->lc_number, e->lc_new_number, e->refcnt); in lc_seq_dump_details() [all …]
|
A D | inflate.c | 618 if ((e = (t = tl + ((unsigned)b & ml))->e) > 16) in inflate_codes() 623 e -= 16; in inflate_codes() 625 } while ((e = (t = t->v.t + ((unsigned)b & mask_bits[e]))->e) > 16); in inflate_codes() 644 NEEDBITS(e) in inflate_codes() 650 if ((e = (t = td + ((unsigned)b & md))->e) > 16) in inflate_codes() 657 } while ((e = (t = t->v.t + ((unsigned)b & mask_bits[e]))->e) > 16); in inflate_codes() 659 NEEDBITS(e) in inflate_codes() 666 n -= (e = (e = WSIZE - ((d &= WSIZE-1) > w ? d : w)) > n ? n : e); in inflate_codes() 1110 } while (!e); in inflate() 1160 e = 0; in makecrc() [all …]
|
A D | test_fpu_impl.c | 15 volatile double a, b, c, d, e, f, g; in test_fpu() local 25 e = a + b / 2; in test_fpu() 33 if (d > a && e > a && g > a) in test_fpu()
|
A D | test_hexdump.c | 167 int ae, he, e, f, r; in test_hexdump_overflow() local 184 e = ae; in test_hexdump_overflow() 186 e = he; in test_hexdump_overflow() 188 f = min_t(int, e + 1, buflen); in test_hexdump_overflow() 195 a = r == e && !memcmp(test, buf, TEST_HEXDUMP_BUF_SIZE); in test_hexdump_overflow() 203 pr_err("Expect: %d '%s'\n", e, test); in test_hexdump_overflow()
|
A D | test_bitmap.c | 910 unsigned int s, e; in test_for_each_set_bitrange() local 922 for_each_set_bitrange(s, e, orig, 500) in test_for_each_set_bitrange() 923 bitmap_set(copy, s, e-s); in test_for_each_set_bitrange() 932 unsigned int s, e; in test_for_each_clear_bitrange() local 944 for_each_clear_bitrange(s, e, orig, 500) in test_for_each_clear_bitrange() 945 bitmap_clear(copy, s, e-s); in test_for_each_clear_bitrange() 954 unsigned int wr, s, e; in test_for_each_set_bitrange_from() local 971 for_each_set_bitrange_from(s, e, orig, 500) in test_for_each_set_bitrange_from() 972 bitmap_set(copy, s, e - s); in test_for_each_set_bitrange_from() 984 unsigned int wr, s, e; in test_for_each_clear_bitrange_from() local [all …]
|
/lib/crypto/ |
A D | sm3.c | 46 h += GG ## i(e, f, g) + ss1 + (w1); \ 52 #define R1(a, b, c, d, e, f, g, h, t, w1, w2) \ argument 53 R(1, a, b, c, d, e, f, g, h, t, w1, w2) 54 #define R2(a, b, c, d, e, f, g, h, t, w1, w2) \ argument 55 R(2, a, b, c, d, e, f, g, h, t, w1, w2) 77 u32 a, b, c, d, e, f, g, h, ss1, ss2; in sm3_transform() local 83 e = sctx->state[4]; in sm3_transform() 88 R1(a, b, c, d, e, f, g, h, K[0], I(0), I(4)); in sm3_transform() 89 R1(d, a, b, c, h, e, f, g, K[1], I(1), I(5)); in sm3_transform() 90 R1(c, d, a, b, g, h, e, f, K[2], I(2), I(6)); in sm3_transform() [all …]
|
A D | sha256.c | 69 t1 = h + e1(e) + Ch(e, f, g) + sha256_K[i] + W[i]; \ 78 u32 a, b, c, d, e, f, g, h; in sha256_block_generic() local 110 e = state->h[4]; in sha256_block_generic() 117 SHA256_ROUND(i + 0, a, b, c, d, e, f, g, h); in sha256_block_generic() 118 SHA256_ROUND(i + 1, h, a, b, c, d, e, f, g); in sha256_block_generic() 119 SHA256_ROUND(i + 2, g, h, a, b, c, d, e, f); in sha256_block_generic() 120 SHA256_ROUND(i + 3, f, g, h, a, b, c, d, e); in sha256_block_generic() 121 SHA256_ROUND(i + 4, e, f, g, h, a, b, c, d); in sha256_block_generic() 122 SHA256_ROUND(i + 5, d, e, f, g, h, a, b, c); in sha256_block_generic() 123 SHA256_ROUND(i + 6, c, d, e, f, g, h, a, b); in sha256_block_generic() [all …]
|
A D | sha512.c | 79 u64 e = state->h[4]; in sha512_block_generic() local 97 t1 = h + e1(e) + Ch(e, f, g) + sha512_K[i] + W[(i & 15)]; in sha512_block_generic() 99 t1 = g + e1(d) + Ch(d, e, f) + sha512_K[i+1] + W[(i & 15) + 1]; in sha512_block_generic() 101 t1 = f + e1(c) + Ch(c, d, e) + sha512_K[i+2] + W[(i & 15) + 2]; in sha512_block_generic() 103 t1 = e + e1(b) + Ch(b, c, d) + sha512_K[i+3] + W[(i & 15) + 3]; in sha512_block_generic() 104 t2 = e0(f) + Maj(f, g, h); a += t1; e = t1 + t2; in sha512_block_generic() 106 t2 = e0(e) + Maj(e, f, g); h += t1; d = t1 + t2; in sha512_block_generic() 108 t2 = e0(d) + Maj(d, e, f); g += t1; c = t1 + t2; in sha512_block_generic() 110 t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2; in sha512_block_generic() 112 t2 = e0(b) + Maj(b, c, d); e += t1; a = t1 + t2; in sha512_block_generic() [all …]
|
/lib/crypto/x86/ |
A D | sha256-avx-asm.S | 157 mov e, y0 # y0 = e 162 xor e, y0 # y0 = e ^ (e >> (25-11)) 168 xor e, y0 # y0 = e ^ (e >> (11-6)) ^ (e >> (25-6)) 196 xor e, y0 # y0 = e ^ (e >> (25-11)) 205 xor e, y0 # y0 = e ^ (e >> (11-6)) ^ (e >> (25-6)) 234 xor e, y0 # y0 = e ^ (e >> (25-11)) 242 xor e, y0 # y0 = e ^ (e >> (11-6)) ^ (e >> (25-6)) 273 xor e, y0 # y0 = e ^ (e >> (25-11)) 280 xor e, y0 # y0 = e ^ (e >> (11-6)) ^ (e >> (25-6)) 310 mov e, y0 # y0 = e [all …]
|
A D | sha256-ssse3-asm.S | 151 mov e, y0 # y0 = e 156 xor e, y0 # y0 = e ^ (e >> (25-11)) 163 xor e, y0 # y0 = e ^ (e >> (11-6)) ^ (e >> (25-6)) 196 xor e, y0 # y0 = e ^ (e >> (25-11)) 205 xor e, y0 # y0 = e ^ (e >> (11-6)) ^ (e >> (25-6)) 237 xor e, y0 # y0 = e ^ (e >> (25-11)) 245 xor e, y0 # y0 = e ^ (e >> (11-6)) ^ (e >> (25-6)) 279 xor e, y0 # y0 = e ^ (e >> (25-11)) 286 xor e, y0 # y0 = e ^ (e >> (11-6)) ^ (e >> (25 320 xor e, y0 # y0 = e ^ (e >> (25-11)) [all …]
|
A D | sha256-avx2-asm.S | 157 rorx $25, e, y0 # y0 = e >> 25 # S1A 158 rorx $11, e, y1 # y1 = e >> 11 # S1B 169 rorx $6, e, y1 # y1 = (e >> 6) # S1 172 xor y1, y0 # y0 = (e>>25) ^ (e>>11) ^ (e>>6) # S1 220 xor y1, y0 # y0 = (e>>25) ^ (e>>11) ^ (e>>6) # S1 277 xor y1, y0 # y0 = (e>>25) ^ (e>>11) ^ (e>>6) # S1 327 xor y1, y0 # y0 = (e>>25) ^ (e>>11) ^ (e>>6) # S1 367 xor y1, y0 # y0 = (e>>25) ^ (e>>11) ^ (e>>6) # S1 405 xor y1, y0 # y0 = (e>>25) ^ (e>>11) ^ (e>>6) # S1 445 xor y1, y0 # y0 = (e>>25) ^ (e>>11) ^ (e>>6) # S1 [all …]
|
A D | sha512-avx2-asm.S | 182 rorx $41, e, y0 # y0 = e >> 41 # S1A 183 rorx $18, e, y1 # y1 = e >> 18 # S1B 191 rorx $14, e, y1 # y1 = (e >> 14) # S1 194 xor y1, y0 # y0 = (e>>41) ^ (e>>18) ^ (e>>14) # S1 257 xor y1, y0 # y0 = (e>>41) ^ (e>>18) ^ (e>>14) # S1 318 xor y1, y0 # y0 = (e>>41) ^ (e>>18) ^ (e>>14) # S1 376 xor y1, y0 # y0 = (e>>41) ^ (e>>18) ^ (e>>14) # S1 414 xor y1, y0 # y0 = (e>>41) ^ (e>>18) ^ (e>>14) # S1 452 xor y1, y0 # y0 = (e>>41) ^ (e>>18) ^ (e>>14) # S1 490 xor y1, y0 # y0 = (e>>41) ^ (e>>18) ^ (e>>14) # S1 [all …]
|
A D | sha512-avx-asm.S | 124 mov e_64, tmp0 # tmp = e 126 RORQ tmp0, 23 # 41 # tmp = e ror 23 127 and e_64, T1 # T1 = (f ^ g) & e 128 xor e_64, tmp0 # tmp = (e ror 23) ^ e 129 xor g_64, T1 # T1 = ((f ^ g) & e) ^ g = CH(e,f,g) 132 RORQ tmp0, 4 # 18 # tmp = ((e ror 23) ^ e) ror 4 133 xor e_64, tmp0 # tmp = (((e ror 23) ^ e) ror 4) ^ e 135 add h_64, T1 # T1 = CH(e,f,g) + W[t] + K[t] + h 136 RORQ tmp0, 14 # 14 # tmp = ((((e ror23)^e)ror4)^e)ror14 = S1(e) 137 add tmp0, T1 # T1 = CH(e,f,g) + W[t] + K[t] + S1(e) [all …]
|
A D | sha512-ssse3-asm.S | 118 mov e_64, tmp0 # tmp = e 120 ror $23, tmp0 # 41 # tmp = e ror 23 121 and e_64, T1 # T1 = (f ^ g) & e 122 xor e_64, tmp0 # tmp = (e ror 23) ^ e 123 xor g_64, T1 # T1 = ((f ^ g) & e) ^ g = CH(e,f,g) 126 ror $4, tmp0 # 18 # tmp = ((e ror 23) ^ e) ror 4 127 xor e_64, tmp0 # tmp = (((e ror 23) ^ e) ror 4) ^ e 129 add h_64, T1 # T1 = CH(e,f,g) + W[t] + K[t] + h 130 ror $14, tmp0 # 14 # tmp = ((((e ror23)^e)ror4)^e)ror14 = S1(e) 131 add tmp0, T1 # T1 = CH(e,f,g) + W[t] + K[t] + S1(e) [all …]
|
A D | sha1-ssse3-and-avx.S | 259 .macro RR F, a, b, c, d, e, round 260 add WK(\round), \e 264 add T1, \e 270 add \a, \e 274 mov \e, T1 275 SWAP_REG_NAMES \e, T1 281 # rotate: \a<=\d, \b<=\e, \c<=\a, \d<=\b, \e<=\c
|
/lib/crypto/powerpc/ |
A D | sha1-spe-asm.S | 113 add e,e,rT0; /* 1: E = E + A' */ \ 115 add e,e,w0; /* 1: E = E + W */ \ 117 add e,e,rT2; /* 1: E = E + F */ \ 119 add e,e,rK; /* 1: E = E + K */ \ 137 add e,e,rT1; /* 1: E = E + F */ \ 141 add e,e,rT2; /* 1: E = E + A' */ \ 146 add e,e,rT0; /* 1: E = E + WK */ \ 162 add e,e,rT2; /* 1: E = E + F */ \ 166 add e,e,rT2; /* 1: E = E + A' */ \ 171 add e,e,rT0; /* 1: E = E + WK */ \ [all …]
|
A D | sha256-spe-asm.S | 103 rotrwi rT0,e,6; /* 1: S1 = e rotr 6 */ \ 104 rotrwi rT1,e,11; /* 1: S1' = e rotr 11 */ \ 105 rotrwi rT2,e,25; /* 1: S1" = e rotr 25 */ \ 107 and rT3,e,f; /* 1: ch = e and f */ \ 109 andc rT1,g,e; /* 1: ch' = ~e and g */ \ 134 and rT3,d,e; /* 2: ch = e and f */ \ 157 rotrwi rT2,e,6; /* 1: S1 = e rotr 6 */ \ 159 rotrwi rT3,e,11; /* 1: S1' = e rotr 11 */ \ 169 and rT2,e,f; /* 1: ch = e and f */ \ 171 andc rT3,g,e; /* 1: ch' = ~e and g */ \ [all …]
|
/lib/tests/ |
A D | cmdline_kunit.c | 108 unsigned int n, const int *e) in cmdline_do_one_range_test() argument 116 KUNIT_EXPECT_EQ_MSG(test, r[0], e[0], "in test %u (parsed) expected %d numbers, got %d", in cmdline_do_one_range_test() 117 n, e[0], r[0]); in cmdline_do_one_range_test() 119 KUNIT_EXPECT_EQ_MSG(test, r[i], e[i], "in test %u at %u", n, i); in cmdline_do_one_range_test() 123 KUNIT_EXPECT_EQ_MSG(test, r[0], e[0], "in test %u (validated) expected %d numbers, got %d", in cmdline_do_one_range_test() 124 n, e[0], r[0]); in cmdline_do_one_range_test() 136 const int *e = cmdline_test_range_values[i]; in cmdline_test_range() local 138 cmdline_do_one_range_test(test, str, i, e); in cmdline_test_range()
|
/lib/zstd/common/ |
A D | error_private.h | 50 #define CHECK_V_F(e, f) \ argument 51 size_t const e = f; \ 53 if (ERR_isError(e)) \ 54 return e; \
|
/lib/crypto/mpi/ |
A D | mpi-pow.c | 154 mpi_limb_t e; in mpi_powm() local 164 e = ep[i]; in mpi_powm() 165 c = count_leading_zeros(e); in mpi_powm() 166 e = (e << c) << 1; /* shift the exp bits to the left, lose msb */ in mpi_powm() 214 if ((mpi_limb_signed_t) e < 0) { in mpi_powm() 240 e <<= 1; in mpi_powm() 248 e = ep[i]; in mpi_powm()
|
/lib/crypto/arm/ |
A D | sha256-armv4.pl | 72 my ($i,$a,$b,$c,$d,$e,$f,$g,$h) = @_; 80 eor $t0,$e,$e,ror#`$Sigma1[1]-$Sigma1[0]` 82 eor $t0,$t0,$e,ror#`$Sigma1[2]-$Sigma1[0]` @ Sigma1(e) 97 eor $t0,$e,$e,ror#`$Sigma1[1]-$Sigma1[0]` 99 eor $t0,$t0,$e,ror#`$Sigma1[2]-$Sigma1[0]` @ Sigma1(e) 108 and $t1,$t1,$e 140 my ($i,$a,$b,$c,$d,$e,$f,$g,$h) = @_; 156 eor $t0,$e,$e,ror#`$Sigma1[1]-$Sigma1[0]` @ from BODY_00_15 158 eor $t0,$t0,$e,ror#`$Sigma1[2]-$Sigma1[0]` @ Sigma1(e) 310 my ($a,$b,$c,$d,$e,$f,$g,$h); [all …]
|
A D | sha512-armv4.pl | 506 my ($a,$b,$c,$d,$e,$f,$g,$h)=@_; 510 vshr.u64 $t0,$e,#@Sigma1[0] @ $i 514 vshr.u64 $t1,$e,#@Sigma1[1] 518 vshr.u64 $t2,$e,#@Sigma1[2] 522 vsli.64 $t0,$e,#`64-@Sigma1[0]` 523 vsli.64 $t1,$e,#`64-@Sigma1[1]` 524 vmov $Ch,$e 525 vsli.64 $t2,$e,#`64-@Sigma1[2]` 561 my $e=@_[4]; # $e from NEON_00_15 581 vshr.u64 $d0,$e,#@Sigma1[0] @ from NEON_00_15 [all …]
|
A D | sha1-armv7-neon.S | 92 add e, e, a, ror #(32 - 5); \ 96 add e, e, RT1; \ 99 add e, e, RT0; 106 add e, e, a, ror #(32 - 5); \ 109 add e, e, RT3; \ 112 add e, e, RT0; \ 120 add e, e, a, ror #(32 - 5); \ 124 add e, e, RT0; \ 127 add e, e, RT1; 131 _R_F2(a,b,c,d,e,i,pre1,pre2,pre3,i16,\ [all …]
|
/lib/crypto/arm64/ |
A D | sha2-armv8.pl | 107 my ($i,$a,$b,$c,$d,$e,$f,$g,$h)=@_; 140 ror $t0,$e,#$Sigma1[0] 142 eor $T0,$e,$e,ror#`$Sigma1[2]-$Sigma1[1]` 143 and $t1,$f,$e 144 bic $t2,$g,$e 162 ror $t0,$e,#$Sigma1[0] 165 and $t1,$f,$e 167 bic $t2,$g,$e 170 eor $t0,$t0,$e,ror#$Sigma1[1] 481 my ($a,$b,$c,$d,$e,$f,$g,$h); [all …]
|
/lib/reed_solomon/ |
A D | test_rslib.c | 442 static int run_exercise(struct etab *e) in run_exercise() argument 444 int nn = (1 << e->symsize) - 1; in run_exercise() 445 int kk = nn - e->nroots; in run_exercise() 453 rsc = init_rs(e->symsize, e->genpoly, e->fcs, e->prim, e->nroots); in run_exercise() 475 retval |= exercise_rs(rsc, ws, len, e->ntrials); in run_exercise() 477 retval |= exercise_rs_bc(rsc, ws, len, e->ntrials); in run_exercise()
|