Lines Matching refs:src_nents
151 struct scatterlist *dst, int src_nents, in caam_unmap() argument
157 if (src_nents) in caam_unmap()
158 dma_unmap_sg(dev, src, src_nents, DMA_TO_DEVICE); in caam_unmap()
162 dma_unmap_sg(dev, src, src_nents, DMA_BIDIRECTIONAL); in caam_unmap()
364 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in aead_edesc_alloc() local
385 src_nents = sg_nents_for_len(req->src, src_len); in aead_edesc_alloc()
386 if (unlikely(src_nents < 0)) { in aead_edesc_alloc()
390 return ERR_PTR(src_nents); in aead_edesc_alloc()
401 if (src_nents) { in aead_edesc_alloc()
402 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in aead_edesc_alloc()
418 dma_unmap_sg(dev, req->src, src_nents, in aead_edesc_alloc()
430 src_nents = sg_nents_for_len(req->src, src_len); in aead_edesc_alloc()
431 if (unlikely(src_nents < 0)) { in aead_edesc_alloc()
435 return ERR_PTR(src_nents); in aead_edesc_alloc()
438 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in aead_edesc_alloc()
478 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in aead_edesc_alloc()
493 caam_unmap(dev, req->src, req->dst, src_nents, in aead_edesc_alloc()
500 edesc->src_nents = src_nents; in aead_edesc_alloc()
517 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, in aead_edesc_alloc()
539 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, in aead_edesc_alloc()
1121 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in skcipher_edesc_alloc() local
1129 src_nents = sg_nents_for_len(req->src, req->cryptlen); in skcipher_edesc_alloc()
1130 if (unlikely(src_nents < 0)) { in skcipher_edesc_alloc()
1133 return ERR_PTR(src_nents); in skcipher_edesc_alloc()
1144 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in skcipher_edesc_alloc()
1155 dma_unmap_sg(dev, req->src, src_nents, DMA_TO_DEVICE); in skcipher_edesc_alloc()
1159 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in skcipher_edesc_alloc()
1188 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in skcipher_edesc_alloc()
1197 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in skcipher_edesc_alloc()
1210 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in skcipher_edesc_alloc()
1216 edesc->src_nents = src_nents; in skcipher_edesc_alloc()
1234 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, in skcipher_edesc_alloc()
1266 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in aead_unmap()
1278 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in skcipher_unmap()
1409 edesc->src_nents > 1 ? 100 : ivsize, 1); in skcipher_encrypt_done()
1447 edesc->src_nents > 1 ? 100 : ivsize, 1); in skcipher_decrypt_done()
3371 if (edesc->src_nents) in ahash_unmap()
3372 dma_unmap_sg(dev, req->src, edesc->src_nents, DMA_TO_DEVICE); in ahash_unmap()
3542 int src_nents, mapped_nents, qm_sg_bytes, qm_sg_src_index; in ahash_update_ctx() local
3553 src_nents = sg_nents_for_len(req->src, src_len); in ahash_update_ctx()
3554 if (src_nents < 0) { in ahash_update_ctx()
3556 return src_nents; in ahash_update_ctx()
3559 if (src_nents) { in ahash_update_ctx()
3560 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_update_ctx()
3573 dma_unmap_sg(ctx->dev, req->src, src_nents, in ahash_update_ctx()
3578 edesc->src_nents = src_nents; in ahash_update_ctx()
3730 int src_nents, mapped_nents; in ahash_finup_ctx() local
3736 src_nents = sg_nents_for_len(req->src, req->nbytes); in ahash_finup_ctx()
3737 if (src_nents < 0) { in ahash_finup_ctx()
3739 return src_nents; in ahash_finup_ctx()
3742 if (src_nents) { in ahash_finup_ctx()
3743 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_finup_ctx()
3756 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE); in ahash_finup_ctx()
3760 edesc->src_nents = src_nents; in ahash_finup_ctx()
3823 int src_nents, mapped_nents; in ahash_digest() local
3829 src_nents = sg_nents_for_len(req->src, req->nbytes); in ahash_digest()
3830 if (src_nents < 0) { in ahash_digest()
3832 return src_nents; in ahash_digest()
3835 if (src_nents) { in ahash_digest()
3836 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_digest()
3849 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE); in ahash_digest()
3853 edesc->src_nents = src_nents; in ahash_digest()
3994 int qm_sg_bytes, src_nents, mapped_nents; in ahash_update_no_ctx() local
4005 src_nents = sg_nents_for_len(req->src, src_len); in ahash_update_no_ctx()
4006 if (src_nents < 0) { in ahash_update_no_ctx()
4008 return src_nents; in ahash_update_no_ctx()
4011 if (src_nents) { in ahash_update_no_ctx()
4012 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_update_no_ctx()
4025 dma_unmap_sg(ctx->dev, req->src, src_nents, in ahash_update_no_ctx()
4030 edesc->src_nents = src_nents; in ahash_update_no_ctx()
4112 int qm_sg_bytes, src_nents, mapped_nents; in ahash_finup_no_ctx() local
4118 src_nents = sg_nents_for_len(req->src, req->nbytes); in ahash_finup_no_ctx()
4119 if (src_nents < 0) { in ahash_finup_no_ctx()
4121 return src_nents; in ahash_finup_no_ctx()
4124 if (src_nents) { in ahash_finup_no_ctx()
4125 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_finup_no_ctx()
4138 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE); in ahash_finup_no_ctx()
4142 edesc->src_nents = src_nents; in ahash_finup_no_ctx()
4211 int src_nents, mapped_nents; in ahash_update_first() local
4223 src_nents = sg_nents_for_len(req->src, src_len); in ahash_update_first()
4224 if (src_nents < 0) { in ahash_update_first()
4226 return src_nents; in ahash_update_first()
4229 if (src_nents) { in ahash_update_first()
4230 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_update_first()
4243 dma_unmap_sg(ctx->dev, req->src, src_nents, in ahash_update_first()
4248 edesc->src_nents = src_nents; in ahash_update_first()