/linux-6.3-rc2/include/crypto/ |
A D | sha256_base.h | 44 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_base_do_update() 48 if (unlikely((partial + len) >= SHA256_BLOCK_SIZE)) { in sha256_base_do_update() 52 int p = SHA256_BLOCK_SIZE - partial; in sha256_base_do_update() 61 blocks = len / SHA256_BLOCK_SIZE; in sha256_base_do_update() 62 len %= SHA256_BLOCK_SIZE; in sha256_base_do_update() 66 data += blocks * SHA256_BLOCK_SIZE; in sha256_base_do_update() 79 const int bit_offset = SHA256_BLOCK_SIZE - sizeof(__be64); in sha256_base_do_finalize() 82 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_base_do_finalize() 86 memset(sctx->buf + partial, 0x0, SHA256_BLOCK_SIZE - partial); in sha256_base_do_finalize()
|
A D | sha2.h | 15 #define SHA256_BLOCK_SIZE 64 macro 70 u8 buf[SHA256_BLOCK_SIZE];
|
/linux-6.3-rc2/net/mptcp/ |
A D | crypto.c | 46 u8 input[SHA256_BLOCK_SIZE + SHA256_DIGEST_SIZE]; in mptcp_crypto_hmac_sha() 58 memset(input, 0x36, SHA256_BLOCK_SIZE); in mptcp_crypto_hmac_sha() 64 memcpy(&input[SHA256_BLOCK_SIZE], msg, len); in mptcp_crypto_hmac_sha() 69 sha256(input, SHA256_BLOCK_SIZE + len, &input[SHA256_BLOCK_SIZE]); in mptcp_crypto_hmac_sha() 72 memset(input, 0x5C, SHA256_BLOCK_SIZE); in mptcp_crypto_hmac_sha() 78 sha256(input, SHA256_BLOCK_SIZE + SHA256_DIGEST_SIZE, hmac); in mptcp_crypto_hmac_sha()
|
/linux-6.3-rc2/arch/sparc/crypto/ |
A D | sha256_glue.c | 37 done = SHA256_BLOCK_SIZE - partial; in __sha256_sparc64_update() 41 if (len - done >= SHA256_BLOCK_SIZE) { in __sha256_sparc64_update() 42 const unsigned int rounds = (len - done) / SHA256_BLOCK_SIZE; in __sha256_sparc64_update() 45 done += rounds * SHA256_BLOCK_SIZE; in __sha256_sparc64_update() 55 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_sparc64_update() 58 if (partial + len < SHA256_BLOCK_SIZE) { in sha256_sparc64_update() 73 static const u8 padding[SHA256_BLOCK_SIZE] = { 0x80, }; in sha256_sparc64_final() 78 index = sctx->count % SHA256_BLOCK_SIZE; in sha256_sparc64_final() 79 padlen = (index < 56) ? (56 - index) : ((SHA256_BLOCK_SIZE+56) - index); in sha256_sparc64_final() 141 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/arch/mips/cavium-octeon/crypto/ |
A D | octeon-sha256.c | 74 partial = sctx->count % SHA256_BLOCK_SIZE; in __octeon_sha256_update() 79 if ((partial + len) >= SHA256_BLOCK_SIZE) { in __octeon_sha256_update() 83 done + SHA256_BLOCK_SIZE); in __octeon_sha256_update() 89 done += SHA256_BLOCK_SIZE; in __octeon_sha256_update() 91 } while (done + SHA256_BLOCK_SIZE <= len); in __octeon_sha256_update() 110 if ((sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) in octeon_sha256_update() 205 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/drivers/crypto/nx/ |
A D | nx-sha256.c | 22 u8 buf[SHA256_BLOCK_SIZE]; 73 u64 buf_len = (sctx->count % SHA256_BLOCK_SIZE); in nx_sha256_update() 81 total = (sctx->count % SHA256_BLOCK_SIZE) + len; in nx_sha256_update() 82 if (total < SHA256_BLOCK_SIZE) { in nx_sha256_update() 134 to_process = to_process & ~(SHA256_BLOCK_SIZE - 1); in nx_sha256_update() 168 } while (leftover >= SHA256_BLOCK_SIZE); in nx_sha256_update() 201 if (sctx->count >= SHA256_BLOCK_SIZE) { in nx_sha256_final() 214 len = sctx->count & (SHA256_BLOCK_SIZE - 1); in nx_sha256_final() 218 if (len != (sctx->count & (SHA256_BLOCK_SIZE - 1))) { in nx_sha256_final() 282 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/arch/arm64/crypto/ |
A D | sha256-glue.c | 77 .base.cra_blocksize = SHA256_BLOCK_SIZE, 111 chunk + sctx->count % SHA256_BLOCK_SIZE > SHA256_BLOCK_SIZE) in sha256_update_neon() 112 chunk = SHA256_BLOCK_SIZE - in sha256_update_neon() 113 sctx->count % SHA256_BLOCK_SIZE; in sha256_update_neon() 157 .base.cra_blocksize = SHA256_BLOCK_SIZE,
|
A D | sha2-ce-glue.c | 46 src += (blocks - rem) * SHA256_BLOCK_SIZE; in __sha2_ce_transform() 83 bool finalize = !sctx->sst.count && !(len % SHA256_BLOCK_SIZE) && len; in sha256_ce_finup() 150 .cra_blocksize = SHA256_BLOCK_SIZE, 167 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/arch/arm/crypto/ |
A D | sha2-ce-glue.c | 36 (sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) in sha2_ce_update() 79 .cra_blocksize = SHA256_BLOCK_SIZE, 93 .cra_blocksize = SHA256_BLOCK_SIZE,
|
A D | sha256_neon_glue.c | 33 (sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) in crypto_sha256_neon_update() 77 .cra_blocksize = SHA256_BLOCK_SIZE,
|
A D | sha256_glue.c | 68 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/arch/x86/crypto/ |
A D | sha256_ssse3_glue.c | 52 (sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) in _sha256_update() 112 .cra_blocksize = SHA256_BLOCK_SIZE, 177 .cra_blocksize = SHA256_BLOCK_SIZE, 253 .cra_blocksize = SHA256_BLOCK_SIZE, 328 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/drivers/crypto/ |
A D | padlock-sha.c | 155 leftover = ((state.count - 1) & (SHA256_BLOCK_SIZE - 1)) + 1; in padlock_sha256_finup() 156 space = SHA256_BLOCK_SIZE - leftover; in padlock_sha256_finup() 261 .cra_blocksize = SHA256_BLOCK_SIZE, 381 if ((partial + len) >= SHA256_BLOCK_SIZE) { in padlock_sha256_update_nano() 387 done + SHA256_BLOCK_SIZE); in padlock_sha256_update_nano() 392 done += SHA256_BLOCK_SIZE; in padlock_sha256_update_nano() 397 if (len - done >= SHA256_BLOCK_SIZE) { in padlock_sha256_update_nano() 488 .cra_blocksize = SHA256_BLOCK_SIZE,
|
A D | sahara.c | 35 #define SAHARA_MAX_SHA_BLOCK_SIZE SHA256_BLOCK_SIZE 1175 SHA_BUFFER_LEN + SHA256_BLOCK_SIZE); in sahara_sha_cra_init() 1262 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/drivers/crypto/qce/ |
A D | sha.h | 16 #define QCE_SHA_MAX_BLOCKSIZE SHA256_BLOCK_SIZE
|
A D | sha.c | 428 .blocksize = SHA256_BLOCK_SIZE, 446 .blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/crypto/ |
A D | sha256_generic.c | 72 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/arch/s390/crypto/ |
A D | sha256_s390.c | 74 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/drivers/crypto/allwinner/sun8i-ss/ |
A D | sun8i-ss.h | 244 u8 key[SHA256_BLOCK_SIZE];
|
/linux-6.3-rc2/arch/powerpc/crypto/ |
A D | sha256-spe-glue.c | 195 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/drivers/crypto/inside-secure/ |
A D | safexcel_hash.c | 1249 req->block_sz = SHA256_BLOCK_SIZE; in safexcel_sha256_init() 1285 .cra_blocksize = SHA256_BLOCK_SIZE, 1306 req->block_sz = SHA256_BLOCK_SIZE; in safexcel_sha224_init() 1369 req->len = SHA256_BLOCK_SIZE; in safexcel_hmac_sha224_init() 1370 req->processed = SHA256_BLOCK_SIZE; in safexcel_hmac_sha224_init() 1376 req->block_sz = SHA256_BLOCK_SIZE; in safexcel_hmac_sha224_init() 1441 req->len = SHA256_BLOCK_SIZE; in safexcel_hmac_sha256_init() 1442 req->processed = SHA256_BLOCK_SIZE; in safexcel_hmac_sha256_init() 1448 req->block_sz = SHA256_BLOCK_SIZE; in safexcel_hmac_sha256_init() 1486 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/drivers/crypto/axis/ |
A D | artpec6_crypto.c | 291 char partial_buffer[SHA256_BLOCK_SIZE]; 292 char partial_buffer_out[SHA256_BLOCK_SIZE]; 293 char key_buffer[SHA256_BLOCK_SIZE]; 294 char pad_buffer[SHA256_BLOCK_SIZE + 32]; 305 char partial_buffer[SHA256_BLOCK_SIZE]; 314 char hmac_key[SHA256_BLOCK_SIZE]; 2660 .cra_blocksize = SHA256_BLOCK_SIZE, 2685 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/drivers/crypto/keembay/ |
A D | keembay-ocs-hcu-core.c | 585 rctx->blk_sz = SHA256_BLOCK_SIZE; in kmb_ocs_hcu_init() 955 .cra_blocksize = SHA256_BLOCK_SIZE, 980 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/fs/verity/ |
A D | hash_algs.c | 18 .block_size = SHA256_BLOCK_SIZE,
|
/linux-6.3-rc2/drivers/crypto/aspeed/ |
A D | aspeed-hace-hash.c | 706 rctx->block_size = SHA256_BLOCK_SIZE; in aspeed_sham_init() 967 .cra_blocksize = SHA256_BLOCK_SIZE, 1091 .cra_blocksize = SHA256_BLOCK_SIZE,
|