| /drivers/crypto/caam/ |
| A D | caamhash.c | 544 int src_nents; member 860 if (src_nents) { in ahash_update_ctx() 886 edesc->src_nents = src_nents; in ahash_update_ctx() 1026 if (src_nents) { in ahash_finup_ctx() 1049 edesc->src_nents = src_nents; in ahash_finup_ctx() 1100 if (src_nents) { in ahash_digest() 1119 edesc->src_nents = src_nents; in ahash_digest() 1262 edesc->src_nents = src_nents; in ahash_update_no_ctx() 1336 if (src_nents) { in ahash_finup_no_ctx() 1361 edesc->src_nents = src_nents; in ahash_finup_no_ctx() [all …]
|
| A D | caamalg_qi2.c | 158 if (src_nents) in caam_unmap() 402 if (src_nents) { in aead_edesc_alloc() 501 edesc->src_nents = src_nents; in aead_edesc_alloc() 1218 edesc->src_nents = src_nents; in skcipher_edesc_alloc() 3580 edesc->src_nents = src_nents; in ahash_update_ctx() 3744 if (src_nents) { in ahash_finup_ctx() 3762 edesc->src_nents = src_nents; in ahash_finup_ctx() 3855 edesc->src_nents = src_nents; in ahash_digest() 4032 edesc->src_nents = src_nents; in ahash_update_no_ctx() 4144 edesc->src_nents = src_nents; in ahash_finup_no_ctx() [all …]
|
| A D | caamalg_qi.c | 803 int src_nents; member 825 int src_nents; member 879 if (src_nents) in caam_unmap() 978 if (unlikely(src_nents < 0)) { in aead_edesc_alloc() 982 return ERR_PTR(src_nents); in aead_edesc_alloc() 1001 return ERR_PTR(src_nents); in aead_edesc_alloc() 1012 if (src_nents) { in aead_edesc_alloc() 1092 edesc->src_nents = src_nents; in aead_edesc_alloc() 1277 if (unlikely(src_nents < 0)) { in skcipher_edesc_alloc() 1280 return ERR_PTR(src_nents); in skcipher_edesc_alloc() [all …]
|
| A D | caamalg_qi2.h | 109 int src_nents; member 129 int src_nents; member 146 int src_nents; member
|
| A D | caamalg.c | 898 int src_nents; member 924 int src_nents; member 943 if (src_nents) in caam_unmap() 1326 if (unlikely(src_nents < 0)) { in aead_edesc_alloc() 1329 return ERR_PTR(src_nents); in aead_edesc_alloc() 1346 return ERR_PTR(src_nents); in aead_edesc_alloc() 1359 if (src_nents) { in aead_edesc_alloc() 1406 edesc->src_nents = src_nents; in aead_edesc_alloc() 1630 if (unlikely(src_nents < 0)) { in skcipher_edesc_alloc() 1633 return ERR_PTR(src_nents); in skcipher_edesc_alloc() [all …]
|
| A D | caampkc.c | 55 dma_unmap_sg(dev, req_ctx->fixup_src, edesc->src_nents, DMA_TO_DEVICE); in rsa_io_unmap() 262 int src_nents, dst_nents; in rsa_edesc_alloc() local 290 src_nents = sg_nents_for_len(req_ctx->fixup_src, in rsa_edesc_alloc() 294 mapped_src_nents = dma_map_sg(dev, req_ctx->fixup_src, src_nents, in rsa_edesc_alloc() 339 edesc->src_nents = src_nents; in rsa_edesc_alloc() 370 dma_unmap_sg(dev, req_ctx->fixup_src, src_nents, DMA_TO_DEVICE); in rsa_edesc_alloc()
|
| A D | caampkc.h | 132 int src_nents; member
|
| /drivers/crypto/qce/ |
| A D | aead.c | 289 rctx->dst_nents = rctx->src_nents + 1; in qce_aead_ccm_prepare_buf_assoclen() 291 rctx->dst_nents = rctx->src_nents; in qce_aead_ccm_prepare_buf_assoclen() 317 if (rctx->src_nents < 0) { in qce_aead_prepare_buf() 323 rctx->src_nents = rctx->dst_nents - 1; in qce_aead_prepare_buf() 354 rctx->dst_nents = rctx->src_nents; in qce_aead_ccm_prepare_buf() 418 int dst_nents, src_nents, ret; in qce_aead_async_req_handle() local 459 src_nents = dma_map_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_aead_async_req_handle() 460 if (src_nents < 0) { in qce_aead_async_req_handle() 461 ret = src_nents; in qce_aead_async_req_handle() 466 src_nents = dst_nents; in qce_aead_async_req_handle() [all …]
|
| A D | skcipher.c | 50 dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_skcipher_done() 75 int dst_nents, src_nents, ret; in qce_skcipher_async_req_handle() local 85 rctx->src_nents = sg_nents_for_len(req->src, req->cryptlen); in qce_skcipher_async_req_handle() 89 rctx->dst_nents = rctx->src_nents; in qce_skcipher_async_req_handle() 90 if (rctx->src_nents < 0) { in qce_skcipher_async_req_handle() 92 return rctx->src_nents; in qce_skcipher_async_req_handle() 133 src_nents = dma_map_sg(qce->dev, req->src, rctx->src_nents, dir_src); in qce_skcipher_async_req_handle() 134 if (!src_nents) { in qce_skcipher_async_req_handle() 141 src_nents = dst_nents - 1; in qce_skcipher_async_req_handle() 144 ret = qce_dma_prep_sgs(&qce->dma, rctx->src_sg, src_nents, in qce_skcipher_async_req_handle() [all …]
|
| A D | sha.c | 53 dma_unmap_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in qce_ahash_done() 93 rctx->src_nents = sg_nents_for_len(req->src, req->nbytes); in qce_ahash_async_req_handle() 94 if (rctx->src_nents < 0) { in qce_ahash_async_req_handle() 96 return rctx->src_nents; in qce_ahash_async_req_handle() 99 ret = dma_map_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in qce_ahash_async_req_handle() 111 ret = qce_dma_prep_sgs(&qce->dma, req->src, rctx->src_nents, in qce_ahash_async_req_handle() 129 dma_unmap_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in qce_ahash_async_req_handle()
|
| A D | cipher.h | 38 int src_nents; member
|
| A D | sha.h | 50 int src_nents; member
|
| A D | aead.h | 30 int src_nents; member
|
| /drivers/crypto/inside-secure/eip93/ |
| A D | eip93-common.c | 212 int src_nents, dst_nents; in check_valid_request() local 227 src_nents = sg_nents_for_len(src, totlen_src); in check_valid_request() 228 if (src_nents < 0) in check_valid_request() 229 return src_nents; in check_valid_request() 236 src_nents = max(src_nents, dst_nents); in check_valid_request() 237 dst_nents = src_nents; in check_valid_request() 242 if (unlikely(totlen_src && !src_nents)) in check_valid_request() 250 if (dst_nents == 1 && src_nents == 1) { in check_valid_request() 282 if (src_nents < 0) in check_valid_request() 283 return src_nents; in check_valid_request() [all …]
|
| A D | eip93-cipher.h | 41 int src_nents; member
|
| /drivers/crypto/virtio/ |
| A D | virtio_crypto_skcipher_algs.c | 331 int src_nents, dst_nents; in __virtio_crypto_skcipher_do_req() local 341 src_nents = sg_nents_for_len(req->src, req->cryptlen); in __virtio_crypto_skcipher_do_req() 342 if (src_nents < 0) { in __virtio_crypto_skcipher_do_req() 344 return src_nents; in __virtio_crypto_skcipher_do_req() 350 src_nents, dst_nents); in __virtio_crypto_skcipher_do_req() 353 sg_total = src_nents + dst_nents + 3; in __virtio_crypto_skcipher_do_req() 433 for (sg = req->src; src_nents; sg = sg_next(sg), src_nents--) in __virtio_crypto_skcipher_do_req()
|
| /drivers/crypto/marvell/cesa/ |
| A D | cipher.c | 66 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_cleanup() 69 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_cleanup() 100 len = mv_cesa_sg_copy_to_sram(engine, req->src, creq->src_nents, in mv_cesa_skcipher_std_step() 329 ret = dma_map_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_req_init() 341 ret = dma_map_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_req_init() 402 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_req_init() 435 creq->src_nents = sg_nents_for_len(req->src, req->cryptlen); in mv_cesa_skcipher_req_init() 436 if (creq->src_nents < 0) { in mv_cesa_skcipher_req_init() 438 return creq->src_nents; in mv_cesa_skcipher_req_init()
|
| A D | hash.c | 105 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, DMA_TO_DEVICE); in mv_cesa_ahash_dma_cleanup() 207 engine, req->src, creq->src_nents, in mv_cesa_ahash_std_step() 425 sg_pcopy_to_buffer(ahashreq->src, creq->src_nents, in mv_cesa_ahash_req_cleanup() 479 sg_pcopy_to_buffer(req->src, creq->src_nents, in mv_cesa_ahash_cache_req() 644 if (creq->src_nents) { in mv_cesa_ahash_dma_req_init() 645 ret = dma_map_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_ahash_dma_req_init() 751 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, DMA_TO_DEVICE); in mv_cesa_ahash_dma_req_init() 763 creq->src_nents = sg_nents_for_len(req->src, req->nbytes); in mv_cesa_ahash_req_init() 764 if (creq->src_nents < 0) { in mv_cesa_ahash_req_init() 766 return creq->src_nents; in mv_cesa_ahash_req_init()
|
| /drivers/crypto/aspeed/ |
| A D | aspeed-hace-crypto.c | 141 dma_unmap_sg(dev, req->src, rctx->src_nents, DMA_BIDIRECTIONAL); in aspeed_sk_transfer_sg() 143 dma_unmap_sg(dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in aspeed_sk_transfer_sg() 192 nbytes = sg_copy_to_buffer(in_sg, rctx->src_nents, in aspeed_sk_start() 197 "nb_in_sg", rctx->src_nents, in aspeed_sk_start() 243 rctx->src_nents, DMA_BIDIRECTIONAL); in aspeed_sk_start_sg() 252 rctx->src_nents, DMA_TO_DEVICE); in aspeed_sk_start_sg() 344 dma_unmap_sg(hace_dev->dev, req->src, rctx->src_nents, in aspeed_sk_start_sg() 350 dma_unmap_sg(hace_dev->dev, req->src, rctx->src_nents, in aspeed_sk_start_sg() 357 dma_unmap_sg(hace_dev->dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in aspeed_sk_start_sg() 381 rctx->src_nents = sg_nents(req->src); in aspeed_hace_skcipher_trigger()
|
| A D | aspeed-hace.h | 179 int src_nents; member 228 int src_nents; member
|
| A D | aspeed-hace-hash.c | 232 sg_len = dma_map_sg(hace_dev->dev, rctx->src_sg, rctx->src_nents, in aspeed_ahash_dma_prepare_sg() 316 dma_unmap_sg(hace_dev->dev, rctx->src_sg, rctx->src_nents, in aspeed_ahash_dma_prepare_sg() 389 dma_unmap_sg(hace_dev->dev, rctx->src_sg, rctx->src_nents, in aspeed_ahash_update_resume_sg() 514 rctx->src_nents = sg_nents_for_len(req->src, req->nbytes); in aspeed_sham_update()
|
| /drivers/crypto/ |
| A D | talitos.c | 967 unsigned int src_nents = edesc->src_nents ? : 1; in talitos_sg_unmap() local 977 if (src_nents == 1 || !is_sec1) in talitos_sg_unmap() 1220 sg_count = edesc->src_nents ?: 1; in ipsec_esp() 1348 if (src_nents < 0) { in talitos_edesc_alloc() 1352 src_nents = (src_nents == 1) ? 0 : src_nents; in talitos_edesc_alloc() 1353 dst_nents = dst ? src_nents : 0; in talitos_edesc_alloc() 1358 if (src_nents < 0) { in talitos_edesc_alloc() 1362 src_nents = (src_nents == 1) ? 0 : src_nents; in talitos_edesc_alloc() 1405 edesc->src_nents = src_nents; in talitos_edesc_alloc() 1593 sg_count = edesc->src_nents ?: 1; in common_nonsnoop() [all …]
|
| /drivers/crypto/intel/keembay/ |
| A D | keembay-ocs-aes-core.c | 85 int src_nents; member 382 dma_unmap_sg(dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in kmb_ocs_sk_dma_cleanup() 454 rctx->src_nents = sg_nents_for_len(req->src, req->cryptlen); in kmb_ocs_sk_prepare_notinplace() 455 if (rctx->src_nents < 0) in kmb_ocs_sk_prepare_notinplace() 460 rctx->src_nents, DMA_TO_DEVICE); in kmb_ocs_sk_prepare_notinplace() 689 dma_unmap_sg(dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in kmb_ocs_aead_dma_cleanup() 730 rctx->src_nents = sg_nents_for_len(req->src, in kmb_ocs_aead_dma_prepare() 732 if (rctx->src_nents < 0) in kmb_ocs_aead_dma_prepare() 758 sg_pcopy_to_buffer(req->src, rctx->src_nents, rctx->in_tag, in kmb_ocs_aead_dma_prepare() 852 rctx->src_nents, DMA_TO_DEVICE); in kmb_ocs_aead_dma_prepare() [all …]
|
| /drivers/crypto/ccp/ |
| A D | ccp-dmaengine.c | 356 unsigned int src_nents, in ccp_create_desc() argument 375 if (!dst_nents || !src_nents) in ccp_create_desc() 392 src_nents--; in ccp_create_desc() 393 if (!src_nents) in ccp_create_desc()
|
| /drivers/crypto/starfive/ |
| A D | jh7110-aes.c | 506 unsigned int len, src_nents, dst_nents; in starfive_aes_map_sg() local 512 src_nents = dma_map_sg(cryp->dev, stsg, 1, DMA_BIDIRECTIONAL); in starfive_aes_map_sg() 513 if (src_nents == 0) in starfive_aes_map_sg() 517 dst_nents = src_nents; in starfive_aes_map_sg() 529 src_nents = dma_map_sg(cryp->dev, stsg, 1, DMA_TO_DEVICE); in starfive_aes_map_sg() 530 if (src_nents == 0) in starfive_aes_map_sg()
|