| /drivers/crypto/nx/ |
| A D | nx-aes-xcbc.c | 24 u8 state[AES_BLOCK_SIZE]; 63 u8 keys[2][AES_BLOCK_SIZE]; in nx_xcbc_empty() 113 len = AES_BLOCK_SIZE; in nx_xcbc_empty() 117 if (len != AES_BLOCK_SIZE) in nx_xcbc_empty() 193 data_len = AES_BLOCK_SIZE; in nx_xcbc_update() 197 if (data_len != AES_BLOCK_SIZE) { in nx_xcbc_update() 234 } while (total >= AES_BLOCK_SIZE); in nx_xcbc_update() 283 len = AES_BLOCK_SIZE; in nx_xcbc_finup() 287 if (len != AES_BLOCK_SIZE) { in nx_xcbc_finup() 313 .digestsize = AES_BLOCK_SIZE, [all …]
|
| A D | nx-aes-gcm.c | 111 if (nbytes <= AES_BLOCK_SIZE) { in nx_gca() 152 AES_BLOCK_SIZE); in nx_gca() 221 csbcpb->cpb.aes_gcm.out_pat_or_mac, AES_BLOCK_SIZE); in gmac() 223 csbcpb->cpb.aes_gcm.out_s0, AES_BLOCK_SIZE); in gmac() 245 char out[AES_BLOCK_SIZE]; in gcm_empty() 263 len = AES_BLOCK_SIZE; in gcm_empty() 269 if (len != AES_BLOCK_SIZE) in gcm_empty() 375 csbcpb->cpb.aes_gcm.out_pat_or_mac, AES_BLOCK_SIZE); in gcm_aes_nx_crypt() 377 csbcpb->cpb.aes_gcm.out_s0, AES_BLOCK_SIZE); in gcm_aes_nx_crypt() 477 .maxauthsize = AES_BLOCK_SIZE, [all …]
|
| A D | nx-aes-ccm.c | 306 AES_BLOCK_SIZE); in generate_pat() 319 memcpy(out, result, AES_BLOCK_SIZE); in generate_pat() 378 memcpy(iv, csbcpb->cpb.aes_ccm.out_ctr, AES_BLOCK_SIZE); in ccm_nx_decrypt() 380 csbcpb->cpb.aes_ccm.out_pat_or_mac, AES_BLOCK_SIZE); in ccm_nx_decrypt() 382 csbcpb->cpb.aes_ccm.out_s0, AES_BLOCK_SIZE); in ccm_nx_decrypt() 447 memcpy(iv, csbcpb->cpb.aes_ccm.out_ctr, AES_BLOCK_SIZE); in ccm_nx_encrypt() 449 csbcpb->cpb.aes_ccm.out_pat_or_mac, AES_BLOCK_SIZE); in ccm_nx_encrypt() 451 csbcpb->cpb.aes_ccm.out_s0, AES_BLOCK_SIZE); in ccm_nx_encrypt() 533 .ivsize = AES_BLOCK_SIZE, 534 .maxauthsize = AES_BLOCK_SIZE, [all …]
|
| /drivers/crypto/ccp/ |
| A D | ccp-crypto.h | 109 u8 k1[AES_BLOCK_SIZE]; 110 u8 k2[AES_BLOCK_SIZE]; 115 u8 iv[AES_BLOCK_SIZE]; 118 u8 tag[AES_BLOCK_SIZE]; 122 u8 rfc3686_iv[AES_BLOCK_SIZE]; 142 u8 iv[AES_BLOCK_SIZE]; 146 u8 buf[AES_BLOCK_SIZE]; 150 u8 pad[AES_BLOCK_SIZE]; 158 u8 iv[AES_BLOCK_SIZE]; 161 u8 buf[AES_BLOCK_SIZE]; [all …]
|
| A D | ccp-crypto-aes.c | 34 memcpy(req->iv, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_complete() 80 (req->cryptlen & (AES_BLOCK_SIZE - 1))) in ccp_aes_crypt() 87 memcpy(rctx->iv, req->iv, AES_BLOCK_SIZE); in ccp_aes_crypt() 89 iv_len = AES_BLOCK_SIZE; in ccp_aes_crypt() 217 .base.cra_blocksize = AES_BLOCK_SIZE, 257 .blocksize = AES_BLOCK_SIZE, 266 .blocksize = AES_BLOCK_SIZE, 267 .ivsize = AES_BLOCK_SIZE, 276 .ivsize = AES_BLOCK_SIZE,
|
| A D | ccp-crypto-aes-xts.c | 70 memcpy(req->iv, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_xts_complete() 164 memcpy(rctx->iv, req->iv, AES_BLOCK_SIZE); in ccp_aes_xts_crypt() 165 sg_init_one(&rctx->iv_sg, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_xts_crypt() 177 rctx->cmd.u.xts.iv_len = AES_BLOCK_SIZE; in ccp_aes_xts_crypt() 249 alg->base.cra_blocksize = AES_BLOCK_SIZE; in ccp_register_aes_xts_alg() 260 alg->ivsize = AES_BLOCK_SIZE; in ccp_register_aes_xts_alg()
|
| A D | ccp-crypto-aes-galois.c | 108 rctx->iv[AES_BLOCK_SIZE - 1] = 1; in ccp_aes_gcm_crypt() 112 iv_len = AES_BLOCK_SIZE; in ccp_aes_gcm_crypt() 172 .maxauthsize = AES_BLOCK_SIZE, 178 .cra_blocksize = AES_BLOCK_SIZE, 203 .ivsize = AES_BLOCK_SIZE,
|
| /drivers/crypto/ |
| A D | padlock-aes.c | 329 .cra_blocksize = AES_BLOCK_SIZE, 359 nbytes / AES_BLOCK_SIZE); in ecb_aes_encrypt() 360 nbytes &= AES_BLOCK_SIZE - 1; in ecb_aes_encrypt() 384 nbytes / AES_BLOCK_SIZE); in ecb_aes_decrypt() 385 nbytes &= AES_BLOCK_SIZE - 1; in ecb_aes_decrypt() 398 .base.cra_blocksize = AES_BLOCK_SIZE, 425 nbytes / AES_BLOCK_SIZE); in cbc_aes_encrypt() 427 nbytes &= AES_BLOCK_SIZE - 1; in cbc_aes_encrypt() 451 nbytes / AES_BLOCK_SIZE); in cbc_aes_decrypt() 452 nbytes &= AES_BLOCK_SIZE - 1; in cbc_aes_decrypt() [all …]
|
| A D | atmel-aes.c | 146 u8 lastc[AES_BLOCK_SIZE]; 595 dd->datalen -= AES_BLOCK_SIZE; in atmel_aes_cpu_transfer() 851 case AES_BLOCK_SIZE: in atmel_aes_dma_start() 1211 .ivsize = AES_BLOCK_SIZE, 1225 .ivsize = AES_BLOCK_SIZE, 1751 .ivsize = AES_BLOCK_SIZE, 2039 .ivsize = AES_BLOCK_SIZE, 2055 .ivsize = AES_BLOCK_SIZE, 2071 .ivsize = AES_BLOCK_SIZE, 2087 .ivsize = AES_BLOCK_SIZE, [all …]
|
| A D | geode-aes.c | 173 geode_aes_crypt(tctx, in, out, AES_BLOCK_SIZE, NULL, in geode_encrypt() 188 geode_aes_crypt(tctx, in, out, AES_BLOCK_SIZE, NULL, in geode_decrypt() 224 .cra_blocksize = AES_BLOCK_SIZE, 286 round_down(nbytes, AES_BLOCK_SIZE), in geode_skcipher_crypt() 288 err = skcipher_walk_done(&walk, nbytes % AES_BLOCK_SIZE); in geode_skcipher_crypt() 321 .base.cra_blocksize = AES_BLOCK_SIZE, 332 .ivsize = AES_BLOCK_SIZE, 339 .base.cra_blocksize = AES_BLOCK_SIZE,
|
| A D | omap-aes-gcm.c | 50 alen = ALIGN(dd->assoc_len, AES_BLOCK_SIZE); in omap_aes_gcm_done_task() 51 clen = ALIGN(dd->total, AES_BLOCK_SIZE); in omap_aes_gcm_done_task() 109 alen = ALIGN(assoclen, AES_BLOCK_SIZE); in omap_aes_gcm_copy_buffers() 110 clen = ALIGN(cryptlen, AES_BLOCK_SIZE); in omap_aes_gcm_copy_buffers() 120 AES_BLOCK_SIZE, dd->in_sgl, in omap_aes_gcm_copy_buffers() 137 AES_BLOCK_SIZE, &dd->in_sgl[nsg], in omap_aes_gcm_copy_buffers() 163 AES_BLOCK_SIZE, &dd->out_sgl, in omap_aes_gcm_copy_buffers()
|
| /drivers/crypto/intel/keembay/ |
| A D | keembay-ocs-aes-core.c | 94 u8 last_ct_blk[AES_BLOCK_SIZE]; 102 u8 in_tag[AES_BLOCK_SIZE]; 105 u8 out_tag[AES_BLOCK_SIZE]; 227 u8 tmp_buf1[AES_BLOCK_SIZE], tmp_buf2[AES_BLOCK_SIZE]; in sg_swap_blocks() 619 if (iv_size != AES_BLOCK_SIZE) in kmb_ocs_aead_validate_input() 1296 .base.ivsize = AES_BLOCK_SIZE, 1318 .base.ivsize = AES_BLOCK_SIZE, 1341 .base.ivsize = AES_BLOCK_SIZE, 1385 .base.ivsize = AES_BLOCK_SIZE, 1406 .base.ivsize = AES_BLOCK_SIZE, [all …]
|
| /drivers/crypto/ccree/ |
| A D | cc_aead.h | 17 #define CCM_CONFIG_BUF_SIZE (AES_BLOCK_SIZE * 3) 52 u8 ctr_iv[AES_BLOCK_SIZE] ____cacheline_aligned; 55 u8 gcm_iv_inc1[AES_BLOCK_SIZE] ____cacheline_aligned; 56 u8 gcm_iv_inc2[AES_BLOCK_SIZE] ____cacheline_aligned; 57 u8 hkey[AES_BLOCK_SIZE] ____cacheline_aligned;
|
| A D | cc_cipher.c | 115 if (size >= AES_BLOCK_SIZE) in validate_data_size() 1014 .ivsize = AES_BLOCK_SIZE, 1032 .ivsize = AES_BLOCK_SIZE, 1068 .ivsize = AES_BLOCK_SIZE, 1086 .ivsize = AES_BLOCK_SIZE, 1104 .ivsize = AES_BLOCK_SIZE, 1126 .ivsize = AES_BLOCK_SIZE, 1143 .ivsize = AES_BLOCK_SIZE, 1177 .ivsize = AES_BLOCK_SIZE, 1194 .ivsize = AES_BLOCK_SIZE, [all …]
|
| A D | cc_aead.c | 1468 AES_BLOCK_SIZE, NS_BIT); in cc_ccm() 1507 AES_BLOCK_SIZE, NS_BIT); in cc_ccm() 1649 AES_BLOCK_SIZE, NS_BIT); in cc_set_ghash_desc() 1746 AES_BLOCK_SIZE, NS_BIT); in cc_proc_gcm_result() 2301 .blocksize = AES_BLOCK_SIZE, 2309 .ivsize = AES_BLOCK_SIZE, 2341 .blocksize = AES_BLOCK_SIZE, 2349 .ivsize = AES_BLOCK_SIZE, 2381 .blocksize = AES_BLOCK_SIZE, 2389 .ivsize = AES_BLOCK_SIZE, [all …]
|
| A D | cc_buffer_mgr.c | 309 (AES_BLOCK_SIZE + areq_ctx->ccm_hdr_size), in cc_set_aead_conf_buf() 472 AES_BLOCK_SIZE, DMA_BIDIRECTIONAL); in cc_unmap_aead_request() 477 AES_BLOCK_SIZE, DMA_TO_DEVICE); in cc_unmap_aead_request() 482 AES_BLOCK_SIZE, DMA_TO_DEVICE); in cc_unmap_aead_request() 487 AES_BLOCK_SIZE, DMA_TO_DEVICE); in cc_unmap_aead_request() 494 AES_BLOCK_SIZE, DMA_TO_DEVICE); in cc_unmap_aead_request() 995 AES_BLOCK_SIZE, addr); in cc_map_aead_request() 1013 AES_BLOCK_SIZE, areq_ctx->hkey); in cc_map_aead_request() 1020 AES_BLOCK_SIZE, DMA_TO_DEVICE); in cc_map_aead_request() 1030 AES_BLOCK_SIZE, DMA_TO_DEVICE); in cc_map_aead_request() [all …]
|
| /drivers/crypto/intel/qat/qat_common/ |
| A D | qat_algs.c | 481 memcpy(key_reverse + AES_BLOCK_SIZE, key - AES_BLOCK_SIZE, in qat_alg_xts_reverse_key() 482 AES_BLOCK_SIZE); in qat_alg_xts_reverse_key() 1282 .cra_blocksize = AES_BLOCK_SIZE, 1291 .ivsize = AES_BLOCK_SIZE, 1299 .cra_blocksize = AES_BLOCK_SIZE, 1308 .ivsize = AES_BLOCK_SIZE, 1316 .cra_blocksize = AES_BLOCK_SIZE, 1325 .ivsize = AES_BLOCK_SIZE, 1346 .ivsize = AES_BLOCK_SIZE, 1364 .ivsize = AES_BLOCK_SIZE, [all …]
|
| /drivers/crypto/starfive/ |
| A D | jh7110-aes.c | 200 u8 iv[AES_BLOCK_SIZE], b0[AES_BLOCK_SIZE]; in starfive_aes_ccm_init() 207 memcpy(b0, iv, AES_BLOCK_SIZE); in starfive_aes_ccm_init() 216 b0[AES_BLOCK_SIZE - 2] = textlen >> 8; in starfive_aes_ccm_init() 397 total_len -= AES_BLOCK_SIZE; in starfive_aes_ccm_write_adata() 1028 .cra_blocksize = AES_BLOCK_SIZE, 1044 .base.ivsize = AES_BLOCK_SIZE, 1051 .cra_blocksize = AES_BLOCK_SIZE, 1067 .base.ivsize = AES_BLOCK_SIZE, 1094 .base.maxauthsize = AES_BLOCK_SIZE, 1116 .base.ivsize = AES_BLOCK_SIZE, [all …]
|
| /drivers/crypto/caam/ |
| A D | caamalg_qi.c | 1494 .ivsize = AES_BLOCK_SIZE, 1542 .ivsize = AES_BLOCK_SIZE, 1584 .ivsize = AES_BLOCK_SIZE, 1662 .ivsize = AES_BLOCK_SIZE, 1684 .ivsize = AES_BLOCK_SIZE, 1706 .ivsize = AES_BLOCK_SIZE, 1728 .ivsize = AES_BLOCK_SIZE, 1750 .ivsize = AES_BLOCK_SIZE, 1772 .ivsize = AES_BLOCK_SIZE, 1794 .ivsize = AES_BLOCK_SIZE, [all …]
|
| /drivers/crypto/stm32/ |
| A D | stm32-cryp.c | 690 iv[AES_BLOCK_SIZE - 1] = 1; in stm32_cryp_ccm_init() 694 memcpy(b0, iv, AES_BLOCK_SIZE); in stm32_cryp_ccm_init() 1235 if (req->cryptlen % AES_BLOCK_SIZE) in stm32_cryp_aes_ecb_encrypt() 1246 if (req->cryptlen % AES_BLOCK_SIZE) in stm32_cryp_aes_ecb_decrypt() 1257 if (req->cryptlen % AES_BLOCK_SIZE) in stm32_cryp_aes_cbc_encrypt() 1268 if (req->cryptlen % AES_BLOCK_SIZE) in stm32_cryp_aes_cbc_decrypt() 2320 .ivsize = AES_BLOCK_SIZE, 2343 .ivsize = AES_BLOCK_SIZE, 2452 .base.maxauthsize = AES_BLOCK_SIZE, 2474 .base.ivsize = AES_BLOCK_SIZE, [all …]
|
| /drivers/crypto/cavium/nitrox/ |
| A D | nitrox_skcipher.c | 390 .cra_blocksize = AES_BLOCK_SIZE, 397 .ivsize = AES_BLOCK_SIZE, 409 .cra_blocksize = AES_BLOCK_SIZE, 416 .ivsize = AES_BLOCK_SIZE, 428 .cra_blocksize = AES_BLOCK_SIZE, 435 .ivsize = AES_BLOCK_SIZE, 466 .cra_blocksize = AES_BLOCK_SIZE, 473 .ivsize = AES_BLOCK_SIZE,
|
| /drivers/crypto/inside-secure/ |
| A D | safexcel_hash.c | 359 extra = AES_BLOCK_SIZE - cache_len; in safexcel_ahash_send_req() 385 cache_len = AES_BLOCK_SIZE; in safexcel_ahash_send_req() 1892 req->len = AES_BLOCK_SIZE; in safexcel_cbcmac_init() 1893 req->processed = AES_BLOCK_SIZE; in safexcel_cbcmac_init() 1897 req->digest_sz = AES_BLOCK_SIZE; in safexcel_cbcmac_init() 1898 req->block_sz = AES_BLOCK_SIZE; in safexcel_cbcmac_init() 1953 .digestsize = AES_BLOCK_SIZE, 1962 .cra_blocksize = AES_BLOCK_SIZE, 2036 .digestsize = AES_BLOCK_SIZE, 2074 memset(consts, 0, AES_BLOCK_SIZE); in safexcel_cmac_setkey() [all …]
|
| /drivers/crypto/inside-secure/eip93/ |
| A D | eip93-aead.c | 268 .ivsize = AES_BLOCK_SIZE, 279 .cra_blocksize = AES_BLOCK_SIZE, 296 .ivsize = AES_BLOCK_SIZE, 307 .cra_blocksize = AES_BLOCK_SIZE, 324 .ivsize = AES_BLOCK_SIZE, 335 .cra_blocksize = AES_BLOCK_SIZE, 352 .ivsize = AES_BLOCK_SIZE, 363 .cra_blocksize = AES_BLOCK_SIZE,
|
| /drivers/crypto/virtio/ |
| A D | virtio_crypto_skcipher_algs.c | 425 req->cryptlen - AES_BLOCK_SIZE, in __virtio_crypto_skcipher_do_req() 426 AES_BLOCK_SIZE, 0); in __virtio_crypto_skcipher_do_req() 477 if (req->cryptlen % AES_BLOCK_SIZE) in virtio_crypto_skcipher_encrypt() 500 if (req->cryptlen % AES_BLOCK_SIZE) in virtio_crypto_skcipher_decrypt() 556 req->cryptlen - AES_BLOCK_SIZE, in virtio_crypto_skcipher_finalize_req() 557 AES_BLOCK_SIZE, 0); in virtio_crypto_skcipher_finalize_req() 574 .base.cra_blocksize = AES_BLOCK_SIZE, 584 .ivsize = AES_BLOCK_SIZE,
|
| /drivers/crypto/tegra/ |
| A D | tegra-se-aes.c | 279 if (rctx->len % AES_BLOCK_SIZE) in tegra_aes_do_one_req() 280 rctx->len += AES_BLOCK_SIZE - (rctx->len % AES_BLOCK_SIZE); in tegra_aes_do_one_req() 527 .ivsize = AES_BLOCK_SIZE, 533 .cra_blocksize = AES_BLOCK_SIZE, 554 .cra_blocksize = AES_BLOCK_SIZE, 570 .ivsize = AES_BLOCK_SIZE, 592 .ivsize = AES_BLOCK_SIZE, 1919 .maxauthsize = AES_BLOCK_SIZE, 1940 .maxauthsize = AES_BLOCK_SIZE, 1941 .ivsize = AES_BLOCK_SIZE, [all …]
|