| /crypto/ |
| A D | rmd160.c | 64 ROUND(aa, bb, cc, dd, ee, F1, K1, in[0], 11); in rmd160_transform() 65 ROUND(ee, aa, bb, cc, dd, F1, K1, in[1], 14); in rmd160_transform() 66 ROUND(dd, ee, aa, bb, cc, F1, K1, in[2], 15); in rmd160_transform() 67 ROUND(cc, dd, ee, aa, bb, F1, K1, in[3], 12); in rmd160_transform() 68 ROUND(bb, cc, dd, ee, aa, F1, K1, in[4], 5); in rmd160_transform() 69 ROUND(aa, bb, cc, dd, ee, F1, K1, in[5], 8); in rmd160_transform() 70 ROUND(ee, aa, bb, cc, dd, F1, K1, in[6], 7); in rmd160_transform() 71 ROUND(dd, ee, aa, bb, cc, F1, K1, in[7], 9); in rmd160_transform() 72 ROUND(cc, dd, ee, aa, bb, F1, K1, in[8], 11); in rmd160_transform() 73 ROUND(bb, cc, dd, ee, aa, F1, K1, in[9], 13); in rmd160_transform() [all …]
|
| A D | md5.c | 35 #define MD5STEP(f, w, x, y, z, in, s) \ argument 47 MD5STEP(F1, a, b, c, d, in[0] + 0xd76aa478, 7); in md5_transform() 48 MD5STEP(F1, d, a, b, c, in[1] + 0xe8c7b756, 12); in md5_transform() 49 MD5STEP(F1, c, d, a, b, in[2] + 0x242070db, 17); in md5_transform() 50 MD5STEP(F1, b, c, d, a, in[3] + 0xc1bdceee, 22); in md5_transform() 51 MD5STEP(F1, a, b, c, d, in[4] + 0xf57c0faf, 7); in md5_transform() 52 MD5STEP(F1, d, a, b, c, in[5] + 0x4787c62a, 12); in md5_transform() 55 MD5STEP(F1, a, b, c, d, in[8] + 0x698098d8, 7); in md5_transform() 64 MD5STEP(F2, a, b, c, d, in[1] + 0xf61e2562, 5); in md5_transform() 65 MD5STEP(F2, d, a, b, c, in[6] + 0xc040b340, 9); in md5_transform() [all …]
|
| A D | md4.c | 76 ROUND1(a, b, c, d, in[0], 3); in md4_transform() 77 ROUND1(d, a, b, c, in[1], 7); in md4_transform() 80 ROUND1(a, b, c, d, in[4], 3); in md4_transform() 81 ROUND1(d, a, b, c, in[5], 7); in md4_transform() 84 ROUND1(a, b, c, d, in[8], 3); in md4_transform() 85 ROUND1(d, a, b, c, in[9], 7); in md4_transform() 93 ROUND2(a, b, c, d,in[ 0], 3); in md4_transform() 94 ROUND2(d, a, b, c, in[4], 5); in md4_transform() 95 ROUND2(c, d, a, b, in[8], 9); in md4_transform() 97 ROUND2(a, b, c, d, in[1], 3); in md4_transform() [all …]
|
| A D | scatterwalk.c | 114 scatterwalk_start(&walk.in, src); in memcpy_sglist() 172 walk->in.__addr = walk->out.addr; in skcipher_next_slow() 184 scatterwalk_map(&walk->in); in skcipher_next_copy() 186 scatterwalk_unmap(&walk->in); in skcipher_next_copy() 192 walk->in.__addr = tmp; in skcipher_next_copy() 203 diff |= (u8 *)(sg_page(walk->in.sg) + (walk->in.offset >> PAGE_SHIFT)) - in skcipher_next_fast() 207 walk->in.__addr = walk->out.__addr; in skcipher_next_fast() 211 scatterwalk_map(&walk->in); in skcipher_next_fast() 224 n = scatterwalk_clamp(&walk->in, n); in skcipher_walk_next() 327 scatterwalk_advance(&walk->in, n); in skcipher_walk_done() [all …]
|
| A D | aria_generic.c | 195 static void __aria_crypt(struct aria_ctx *ctx, u8 *out, const u8 *in, in __aria_crypt() argument 203 reg0 = get_unaligned_be32(&in[0]); in __aria_crypt() 204 reg1 = get_unaligned_be32(&in[4]); in __aria_crypt() 205 reg2 = get_unaligned_be32(&in[8]); in __aria_crypt() 206 reg3 = get_unaligned_be32(&in[12]); in __aria_crypt() 248 void aria_encrypt(void *_ctx, u8 *out, const u8 *in) in aria_encrypt() argument 252 __aria_crypt(ctx, out, in, ctx->enc_key); in aria_encrypt() 256 void aria_decrypt(void *_ctx, u8 *out, const u8 *in) in aria_decrypt() argument 260 __aria_crypt(ctx, out, in, ctx->dec_key); in aria_decrypt() 268 __aria_crypt(ctx, out, in, ctx->enc_key); in __aria_encrypt() [all …]
|
| A D | aes_generic.c | 1176 static void crypto_aes_encrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) in crypto_aes_encrypt() argument 1183 b0[0] = ctx->key_enc[0] ^ get_unaligned_le32(in); in crypto_aes_encrypt() 1184 b0[1] = ctx->key_enc[1] ^ get_unaligned_le32(in + 4); in crypto_aes_encrypt() 1185 b0[2] = ctx->key_enc[2] ^ get_unaligned_le32(in + 8); in crypto_aes_encrypt() 1186 b0[3] = ctx->key_enc[3] ^ get_unaligned_le32(in + 12); in crypto_aes_encrypt() 1246 static void crypto_aes_decrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) in crypto_aes_decrypt() argument 1253 b0[0] = ctx->key_dec[0] ^ get_unaligned_le32(in); in crypto_aes_decrypt() 1254 b0[1] = ctx->key_dec[1] ^ get_unaligned_le32(in + 4); in crypto_aes_decrypt() 1255 b0[2] = ctx->key_dec[2] ^ get_unaligned_le32(in + 8); in crypto_aes_decrypt() 1256 b0[3] = ctx->key_dec[3] ^ get_unaligned_le32(in + 12); in crypto_aes_decrypt()
|
| A D | seed.c | 363 static void seed_encrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) in seed_encrypt() argument 369 x1 = get_unaligned_be32(&in[0]); in seed_encrypt() 370 x2 = get_unaligned_be32(&in[4]); in seed_encrypt() 371 x3 = get_unaligned_be32(&in[8]); in seed_encrypt() 372 x4 = get_unaligned_be32(&in[12]); in seed_encrypt() 399 static void seed_decrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) in seed_decrypt() argument 405 x1 = get_unaligned_be32(&in[0]); in seed_decrypt() 406 x2 = get_unaligned_be32(&in[4]); in seed_decrypt() 407 x3 = get_unaligned_be32(&in[8]); in seed_decrypt() 408 x4 = get_unaligned_be32(&in[12]); in seed_decrypt()
|
| A D | sm4_generic.c | 41 static void sm4_encrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) in sm4_encrypt() argument 45 sm4_crypt_block(ctx->rkey_enc, out, in); in sm4_encrypt() 50 static void sm4_decrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) in sm4_decrypt() argument 54 sm4_crypt_block(ctx->rkey_dec, out, in); in sm4_decrypt()
|
| A D | sha512.c | 43 static int __crypto_sha512_import(struct __sha512_ctx *ctx, const void *in) in __crypto_sha512_import() argument 45 const u8 *p = in; in __crypto_sha512_import() 98 static int crypto_sha384_import(struct shash_desc *desc, const void *in) in crypto_sha384_import() argument 100 return __crypto_sha512_import(&SHA384_CTX(desc)->ctx, in); in crypto_sha384_import() 150 static int crypto_sha512_import(struct shash_desc *desc, const void *in) in crypto_sha512_import() argument 152 return __crypto_sha512_import(&SHA512_CTX(desc)->ctx, in); in crypto_sha512_import() 199 static int crypto_hmac_sha384_import(struct shash_desc *desc, const void *in) in crypto_hmac_sha384_import() argument 204 return __crypto_sha512_import(&ctx->ctx.sha_ctx, in); in crypto_hmac_sha384_import() 251 static int crypto_hmac_sha512_import(struct shash_desc *desc, const void *in) in crypto_hmac_sha512_import() argument 256 return __crypto_sha512_import(&ctx->ctx.sha_ctx, in); in crypto_hmac_sha512_import()
|
| A D | sha256.c | 43 static int __crypto_sha256_import(struct __sha256_ctx *ctx, const void *in) in __crypto_sha256_import() argument 45 const u8 *p = in; in __crypto_sha256_import() 96 static int crypto_sha224_import(struct shash_desc *desc, const void *in) in crypto_sha224_import() argument 98 return __crypto_sha256_import(&SHA224_CTX(desc)->ctx, in); in crypto_sha224_import() 144 static int crypto_sha256_import(struct shash_desc *desc, const void *in) in crypto_sha256_import() argument 146 return __crypto_sha256_import(&SHA256_CTX(desc)->ctx, in); in crypto_sha256_import() 193 static int crypto_hmac_sha224_import(struct shash_desc *desc, const void *in) in crypto_hmac_sha224_import() argument 198 return __crypto_sha256_import(&ctx->ctx.sha_ctx, in); in crypto_hmac_sha224_import() 245 static int crypto_hmac_sha256_import(struct shash_desc *desc, const void *in) in crypto_hmac_sha256_import() argument 250 return __crypto_sha256_import(&ctx->ctx.sha_ctx, in); in crypto_hmac_sha256_import()
|
| A D | aes_ti.c | 20 static void aesti_encrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) in aesti_encrypt() argument 31 aes_encrypt(ctx, out, in); in aesti_encrypt() 36 static void aesti_decrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) in aesti_decrypt() argument 47 aes_decrypt(ctx, out, in); in aesti_decrypt()
|
| A D | sha1.c | 42 static int __crypto_sha1_import(struct sha1_ctx *ctx, const void *in) in __crypto_sha1_import() argument 44 const u8 *p = in; in __crypto_sha1_import() 92 static int crypto_sha1_import(struct shash_desc *desc, const void *in) in crypto_sha1_import() argument 94 return __crypto_sha1_import(SHA1_CTX(desc), in); in crypto_sha1_import() 138 static int crypto_hmac_sha1_import(struct shash_desc *desc, const void *in) in crypto_hmac_sha1_import() argument 143 return __crypto_sha1_import(&ctx->sha_ctx, in); in crypto_hmac_sha1_import()
|
| A D | sm4.c | 160 void sm4_crypt_block(const u32 *rk, u8 *out, const u8 *in) in sm4_crypt_block() argument 164 x[0] = get_unaligned_be32(in + 0 * 4); in sm4_crypt_block() 165 x[1] = get_unaligned_be32(in + 1 * 4); in sm4_crypt_block() 166 x[2] = get_unaligned_be32(in + 2 * 4); in sm4_crypt_block() 167 x[3] = get_unaligned_be32(in + 3 * 4); in sm4_crypt_block()
|
| A D | shash.c | 249 static int __crypto_shash_import(struct shash_desc *desc, const void *in, in __crypto_shash_import() argument 251 const void *in)) in __crypto_shash_import() 268 memcpy(buf, in, ss); in __crypto_shash_import() 272 return import(desc, in); in __crypto_shash_import() 275 int crypto_shash_import_core(struct shash_desc *desc, const void *in) in crypto_shash_import_core() argument 277 return __crypto_shash_import(desc, in, in crypto_shash_import_core() 282 int crypto_shash_import(struct shash_desc *desc, const void *in) in crypto_shash_import() argument 287 err = __crypto_shash_import(desc, in, crypto_shash_alg(tfm)->import); in crypto_shash_import() 294 memcpy(buf + descsize - plen, in + ss - plen, plen); in crypto_shash_import() 457 static int shash_default_import_core(struct shash_desc *desc, const void *in) in shash_default_import_core() argument
|
| A D | aegis128-neon-inner.c | 247 const void *in = src; in crypto_aegis128_encrypt_chunk_neon() local 252 in = out = memcpy(buf + AEGIS_BLOCK_SIZE - size, src, size); in crypto_aegis128_encrypt_chunk_neon() 254 m = vqtbl1q_u8(vld1q_u8(in + size - AEGIS_BLOCK_SIZE), in crypto_aegis128_encrypt_chunk_neon() 293 const void *in = src; in crypto_aegis128_decrypt_chunk_neon() local 298 in = out = memcpy(buf + AEGIS_BLOCK_SIZE - size, src, size); in crypto_aegis128_decrypt_chunk_neon() 300 m = s ^ vqtbx1q_u8(s, vld1q_u8(in + size - AEGIS_BLOCK_SIZE), in crypto_aegis128_decrypt_chunk_neon()
|
| A D | Kconfig | 32 required if you want the system to operate in a FIPS 200 223 algorithm that executes in kernel threads. 265 algorithms in the kernel. It should not be enabled in production 302 defined in RFC7919. 407 in the NESSIE competition. 532 Keys are allowed to be from 0 to 256 bits in length, in steps 580 in the TEA algorithm. 656 Bernstein and further specified in RFC7539 for use in IETF protocols. 668 in some performance-sensitive scenarios. 1208 Hash_DRBG variant as defined in NIST SP800-90A. [all …]
|
| A D | blake2b_generic.c | 125 const u8 *in, unsigned int inlen) in crypto_blake2b_update_generic() argument 127 return crypto_blake2b_update_bo(desc, in, inlen, in crypto_blake2b_update_generic() 131 static int crypto_blake2b_finup_generic(struct shash_desc *desc, const u8 *in, in crypto_blake2b_finup_generic() argument 134 return crypto_blake2b_finup(desc, in, inlen, out, in crypto_blake2b_finup_generic()
|
| A D | acompress.c | 478 scatterwalk_advance(&walk->in, used); in acomp_walk_done_src() 480 scatterwalk_done_src(&walk->in, used); in acomp_walk_done_src() 508 walk->in.__addr = (void *)(((u8 *)walk->in.sg) + in acomp_walk_next_src() 509 walk->in.offset); in acomp_walk_next_src() 513 return slen ? scatterwalk_next(&walk->in, slen) : 0; in acomp_walk_next_src() 555 walk->in.sg = (void *)req->svirt; in acomp_walk_virt() 556 walk->in.offset = 0; in acomp_walk_virt() 558 scatterwalk_start(&walk->in, src); in acomp_walk_virt()
|
| A D | hmac.c | 89 static int hmac_import(struct shash_desc *pdesc, const void *in) in hmac_import() argument 96 return crypto_shash_import(desc, in); in hmac_import() 106 static int hmac_import_core(struct shash_desc *pdesc, const void *in) in hmac_import_core() argument 112 return crypto_shash_import_core(desc, in); in hmac_import_core() 308 static int hmac_import_ahash(struct ahash_request *preq, const void *in) in hmac_import_ahash() argument 315 return crypto_ahash_import(req, in); in hmac_import_ahash() 323 static int hmac_import_core_ahash(struct ahash_request *preq, const void *in) in hmac_import_core_ahash() argument 330 return crypto_ahash_import_core(req, in); in hmac_import_core_ahash()
|
| A D | skcipher.c | 58 scatterwalk_start(&walk->in, req->src); in skcipher_walk_virt() 89 scatterwalk_start_at_pos(&walk->in, req->src, req->assoclen); in skcipher_walk_aead_common() 225 static int crypto_lskcipher_import(struct skcipher_request *req, const void *in) in crypto_lskcipher_import() argument 232 memcpy(ivs + crypto_skcipher_ivsize(tfm), in, in crypto_lskcipher_import() 243 static int skcipher_noimport(struct skcipher_request *req, const void *in) in skcipher_noimport() argument 259 int crypto_skcipher_import(struct skcipher_request *req, const void *in) in crypto_skcipher_import() argument 265 return crypto_lskcipher_import(req, in); in crypto_skcipher_import() 266 return alg->import(req, in); in crypto_skcipher_import()
|
| A D | twofish_generic.c | 86 x = get_unaligned_le32(in + (n) * 4) ^ ctx->w[m] 95 static void twofish_encrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) in twofish_encrypt() argument 130 static void twofish_decrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) in twofish_decrypt() argument
|
| A D | camellia_generic.c | 986 static void camellia_encrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) in camellia_encrypt() argument 993 tmp[0] = get_unaligned_be32(in); in camellia_encrypt() 994 tmp[1] = get_unaligned_be32(in + 4); in camellia_encrypt() 995 tmp[2] = get_unaligned_be32(in + 8); in camellia_encrypt() 996 tmp[3] = get_unaligned_be32(in + 12); in camellia_encrypt() 1012 static void camellia_decrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) in camellia_decrypt() argument 1019 tmp[0] = get_unaligned_be32(in); in camellia_decrypt() 1020 tmp[1] = get_unaligned_be32(in + 4); in camellia_decrypt() 1021 tmp[2] = get_unaligned_be32(in + 8); in camellia_decrypt() 1022 tmp[3] = get_unaligned_be32(in + 12); in camellia_decrypt()
|
| A D | polyval-generic.c | 152 static int polyval_import(struct shash_desc *desc, const void *in) in polyval_import() argument 156 copy_and_reverse(dctx->buffer, in); in polyval_import()
|
| /crypto/asymmetric_keys/ |
| A D | public_key.c | 268 const void *in, void *out) in software_key_eds_op() argument 332 ret = crypto_akcipher_sync_encrypt(tfm, in, params->in_len, in software_key_eds_op() 338 ret = crypto_akcipher_sync_decrypt(tfm, in, params->in_len, in software_key_eds_op() 344 ret = crypto_sig_sign(sig, in, params->in_len, in software_key_eds_op()
|
| A D | asymmetric_keys.h | 17 const void *in, void *out);
|