Lines Matching refs:src_nents
152 struct scatterlist *dst, int src_nents, in caam_unmap() argument
158 if (src_nents) in caam_unmap()
159 dma_unmap_sg(dev, src, src_nents, DMA_TO_DEVICE); in caam_unmap()
163 dma_unmap_sg(dev, src, src_nents, DMA_BIDIRECTIONAL); in caam_unmap()
365 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in aead_edesc_alloc() local
386 src_nents = sg_nents_for_len(req->src, src_len); in aead_edesc_alloc()
387 if (unlikely(src_nents < 0)) { in aead_edesc_alloc()
391 return ERR_PTR(src_nents); in aead_edesc_alloc()
402 if (src_nents) { in aead_edesc_alloc()
403 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in aead_edesc_alloc()
419 dma_unmap_sg(dev, req->src, src_nents, in aead_edesc_alloc()
431 src_nents = sg_nents_for_len(req->src, src_len); in aead_edesc_alloc()
432 if (unlikely(src_nents < 0)) { in aead_edesc_alloc()
436 return ERR_PTR(src_nents); in aead_edesc_alloc()
439 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in aead_edesc_alloc()
479 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in aead_edesc_alloc()
494 caam_unmap(dev, req->src, req->dst, src_nents, in aead_edesc_alloc()
501 edesc->src_nents = src_nents; in aead_edesc_alloc()
518 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, in aead_edesc_alloc()
540 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, in aead_edesc_alloc()
1123 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in skcipher_edesc_alloc() local
1131 src_nents = sg_nents_for_len(req->src, req->cryptlen); in skcipher_edesc_alloc()
1132 if (unlikely(src_nents < 0)) { in skcipher_edesc_alloc()
1135 return ERR_PTR(src_nents); in skcipher_edesc_alloc()
1146 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in skcipher_edesc_alloc()
1157 dma_unmap_sg(dev, req->src, src_nents, DMA_TO_DEVICE); in skcipher_edesc_alloc()
1161 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in skcipher_edesc_alloc()
1190 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in skcipher_edesc_alloc()
1199 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in skcipher_edesc_alloc()
1212 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in skcipher_edesc_alloc()
1218 edesc->src_nents = src_nents; in skcipher_edesc_alloc()
1236 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, in skcipher_edesc_alloc()
1268 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in aead_unmap()
1280 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in skcipher_unmap()
1411 edesc->src_nents > 1 ? 100 : ivsize, 1); in skcipher_encrypt_done()
1449 edesc->src_nents > 1 ? 100 : ivsize, 1); in skcipher_decrypt_done()
3373 if (edesc->src_nents) in ahash_unmap()
3374 dma_unmap_sg(dev, req->src, edesc->src_nents, DMA_TO_DEVICE); in ahash_unmap()
3544 int src_nents, mapped_nents, qm_sg_bytes, qm_sg_src_index; in ahash_update_ctx() local
3555 src_nents = sg_nents_for_len(req->src, src_len); in ahash_update_ctx()
3556 if (src_nents < 0) { in ahash_update_ctx()
3558 return src_nents; in ahash_update_ctx()
3561 if (src_nents) { in ahash_update_ctx()
3562 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_update_ctx()
3575 dma_unmap_sg(ctx->dev, req->src, src_nents, in ahash_update_ctx()
3580 edesc->src_nents = src_nents; in ahash_update_ctx()
3732 int src_nents, mapped_nents; in ahash_finup_ctx() local
3738 src_nents = sg_nents_for_len(req->src, req->nbytes); in ahash_finup_ctx()
3739 if (src_nents < 0) { in ahash_finup_ctx()
3741 return src_nents; in ahash_finup_ctx()
3744 if (src_nents) { in ahash_finup_ctx()
3745 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_finup_ctx()
3758 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE); in ahash_finup_ctx()
3762 edesc->src_nents = src_nents; in ahash_finup_ctx()
3825 int src_nents, mapped_nents; in ahash_digest() local
3831 src_nents = sg_nents_for_len(req->src, req->nbytes); in ahash_digest()
3832 if (src_nents < 0) { in ahash_digest()
3834 return src_nents; in ahash_digest()
3837 if (src_nents) { in ahash_digest()
3838 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_digest()
3851 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE); in ahash_digest()
3855 edesc->src_nents = src_nents; in ahash_digest()
3996 int qm_sg_bytes, src_nents, mapped_nents; in ahash_update_no_ctx() local
4007 src_nents = sg_nents_for_len(req->src, src_len); in ahash_update_no_ctx()
4008 if (src_nents < 0) { in ahash_update_no_ctx()
4010 return src_nents; in ahash_update_no_ctx()
4013 if (src_nents) { in ahash_update_no_ctx()
4014 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_update_no_ctx()
4027 dma_unmap_sg(ctx->dev, req->src, src_nents, in ahash_update_no_ctx()
4032 edesc->src_nents = src_nents; in ahash_update_no_ctx()
4114 int qm_sg_bytes, src_nents, mapped_nents; in ahash_finup_no_ctx() local
4120 src_nents = sg_nents_for_len(req->src, req->nbytes); in ahash_finup_no_ctx()
4121 if (src_nents < 0) { in ahash_finup_no_ctx()
4123 return src_nents; in ahash_finup_no_ctx()
4126 if (src_nents) { in ahash_finup_no_ctx()
4127 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_finup_no_ctx()
4140 dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE); in ahash_finup_no_ctx()
4144 edesc->src_nents = src_nents; in ahash_finup_no_ctx()
4213 int src_nents, mapped_nents; in ahash_update_first() local
4225 src_nents = sg_nents_for_len(req->src, src_len); in ahash_update_first()
4226 if (src_nents < 0) { in ahash_update_first()
4228 return src_nents; in ahash_update_first()
4231 if (src_nents) { in ahash_update_first()
4232 mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents, in ahash_update_first()
4245 dma_unmap_sg(ctx->dev, req->src, src_nents, in ahash_update_first()
4250 edesc->src_nents = src_nents; in ahash_update_first()