| /drivers/crypto/amlogic/ |
| A D | amlogic-gxl-cipher.c | 28 struct scatterlist *dst_sg = areq->dst; in meson_cipher_need_fallback() local 33 if (sg_nents(src_sg) != sg_nents(dst_sg)) in meson_cipher_need_fallback() 40 while (src_sg && dst_sg) { in meson_cipher_need_fallback() 43 if ((dst_sg->length % 16) != 0) in meson_cipher_need_fallback() 45 if (src_sg->length != dst_sg->length) in meson_cipher_need_fallback() 49 if (!IS_ALIGNED(dst_sg->offset, sizeof(u32))) in meson_cipher_need_fallback() 52 dst_sg = sg_next(dst_sg); in meson_cipher_need_fallback() 95 struct scatterlist *dst_sg = areq->dst; in meson_cipher() local 204 dst_sg = areq->dst; in meson_cipher() 211 desc->t_dst = cpu_to_le32(sg_dma_address(dst_sg)); in meson_cipher() [all …]
|
| /drivers/crypto/ccp/ |
| A D | ccp-dmaengine.c | 353 struct scatterlist *dst_sg, in ccp_create_desc() argument 372 if (!dst_sg || !src_sg) in ccp_create_desc() 387 dst_len = sg_dma_len(dst_sg); in ccp_create_desc() 410 dst_sg = sg_next(dst_sg); in ccp_create_desc() 411 if (!dst_sg) in ccp_create_desc() 414 dst_len = sg_dma_len(dst_sg); in ccp_create_desc() 434 ccp_pt->dst_dma = sg_dma_address(dst_sg) + dst_offset; in ccp_create_desc() 485 struct scatterlist dst_sg, src_sg; in ccp_prep_dma_memcpy() local 491 sg_init_table(&dst_sg, 1); in ccp_prep_dma_memcpy() 492 sg_dma_address(&dst_sg) = dst; in ccp_prep_dma_memcpy() [all …]
|
| /drivers/mailbox/ |
| A D | bcm-flexrm-mailbox.c | 598 while (src_sg || dst_sg) { in flexrm_spu_estimate_nonheader_desc_count() 606 while (dst_target && dst_sg) { in flexrm_spu_estimate_nonheader_desc_count() 608 if (dst_sg->length < dst_target) in flexrm_spu_estimate_nonheader_desc_count() 609 dst_target -= dst_sg->length; in flexrm_spu_estimate_nonheader_desc_count() 612 dst_sg = sg_next(dst_sg); in flexrm_spu_estimate_nonheader_desc_count() 657 while (src_sg || dst_sg) { in flexrm_spu_write_descs() 674 while (dst_target && dst_sg) { in flexrm_spu_write_descs() 675 if (sg_dma_len(dst_sg) & 0xf) in flexrm_spu_write_descs() 677 sg_dma_len(dst_sg)); in flexrm_spu_write_descs() 680 sg_dma_len(dst_sg)/16); in flexrm_spu_write_descs() [all …]
|
| A D | bcm-pdc-mailbox.c | 270 struct scatterlist *dst_sg; member 623 dma_unmap_sg(dev, rx_ctx->dst_sg, sg_nents(rx_ctx->dst_sg), in pdc_receive_one() 810 static int pdc_rx_list_init(struct pdc_state *pdcs, struct scatterlist *dst_sg, in pdc_rx_list_init() argument 849 rx_ctx->dst_sg = dst_sg; in pdc_rx_list_init()
|
| /drivers/crypto/qce/ |
| A D | skcipher.c | 51 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_skcipher_done() 124 rctx->dst_sg = rctx->dst_tbl.sgl; in qce_skcipher_async_req_handle() 126 dst_nents = dma_map_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_skcipher_async_req_handle() 140 rctx->src_sg = rctx->dst_sg; in qce_skcipher_async_req_handle() 145 rctx->dst_sg, dst_nents, in qce_skcipher_async_req_handle() 164 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_skcipher_async_req_handle()
|
| A D | aead.c | 51 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_aead_done() 174 rctx->dst_sg = rctx->dst_tbl.sgl; in qce_aead_prepare_dst_buf() 175 rctx->dst_nents = sg_nents_for_len(rctx->dst_sg, totallen) + 1; in qce_aead_prepare_dst_buf() 292 rctx->dst_sg = rctx->src_sg; in qce_aead_ccm_prepare_buf_assoclen() 324 rctx->src_sg = rctx->dst_sg; in qce_aead_prepare_buf() 355 rctx->dst_sg = rctx->src_sg; in qce_aead_ccm_prepare_buf() 452 dst_nents = dma_map_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_aead_async_req_handle() 471 ret = qce_dma_prep_sgs(&qce->dma, rctx->src_sg, src_nents, rctx->dst_sg, dst_nents, in qce_aead_async_req_handle() 490 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_aead_async_req_handle()
|
| A D | cipher.h | 42 struct scatterlist *dst_sg; member
|
| A D | aead.h | 36 struct scatterlist *dst_sg; member
|
| /drivers/dma/ |
| A D | nbpfaxi.c | 933 struct scatterlist *src_sg, struct scatterlist *dst_sg, in nbpf_prep_sg() argument 946 mem_sg = dst_sg; in nbpf_prep_sg() 979 sg_dma_address(dst_sg), in nbpf_prep_sg() 990 dst_sg = sg_next(dst_sg); in nbpf_prep_sg() 991 mem_sg = direction == DMA_DEV_TO_MEM ? dst_sg : src_sg; in nbpf_prep_sg() 1006 struct scatterlist dst_sg; in nbpf_prep_memcpy() local 1009 sg_init_table(&dst_sg, 1); in nbpf_prep_memcpy() 1012 sg_dma_address(&dst_sg) = dst; in nbpf_prep_memcpy() 1015 sg_dma_len(&dst_sg) = len; in nbpf_prep_memcpy() 1021 return nbpf_prep_sg(chan, &src_sg, &dst_sg, 1, in nbpf_prep_memcpy()
|
| A D | ste_dma40.c | 2502 struct scatterlist dst_sg; in d40_prep_memcpy() local 2505 sg_init_table(&dst_sg, 1); in d40_prep_memcpy() 2508 sg_dma_address(&dst_sg) = dst; in d40_prep_memcpy() 2511 sg_dma_len(&dst_sg) = size; in d40_prep_memcpy() 2514 return d40_prep_sg(chan, &src_sg, &dst_sg, 1, in d40_prep_memcpy()
|
| /drivers/crypto/tegra/ |
| A D | tegra-se-aes.c | 58 struct scatterlist *dst_sg; member 793 scatterwalk_map_and_copy(rctx->outbuf.buf, rctx->dst_sg, in tegra_gcm_do_crypt() 818 scatterwalk_map_and_copy(rctx->outbuf.buf, rctx->dst_sg, in tegra_gcm_do_final() 1075 scatterwalk_map_and_copy(rctx->outbuf.buf + 16, rctx->dst_sg, in tegra_ccm_ctr_result() 1079 scatterwalk_map_and_copy(rctx->outbuf.buf, rctx->dst_sg, in tegra_ccm_ctr_result() 1099 sg = rctx->encrypt ? rctx->src_sg : rctx->dst_sg; in tegra_ccm_compute_auth() 1165 rctx->dst_sg = req->dst; in tegra_ccm_crypt_init() 1278 rctx->dst_sg = req->dst; in tegra_gcm_do_one_req()
|
| /drivers/crypto/bcm/ |
| A D | cipher.h | 294 struct scatterlist *dst_sg; member
|
| A D | cipher.c | 161 datalen = spu_msg_sg_add(&sg, &rctx->dst_sg, &rctx->dst_skip, in spu_skcipher_rx_sg_create() 349 rctx->dst_nents = spu_sg_count(rctx->dst_sg, rctx->dst_skip, chunksize); in handle_skcipher_req() 1105 datalen = spu_msg_sg_add(&sg, &rctx->dst_sg, &rctx->dst_skip, in spu_aead_rx_sg_create() 1345 rctx->dst_nents = spu_sg_count(rctx->dst_sg, rctx->dst_skip, chunksize); in handle_aead_req() 1737 rctx->dst_sg = req->dst; in skcipher_enqueue() 1924 rctx->dst_sg = NULL; in ahash_enqueue() 2649 rctx->dst_sg = rctx->src_sg; in aead_enqueue() 2657 if (spu_sg_at_offset(req->dst, req->assoclen, &rctx->dst_sg, in aead_enqueue() 2687 rctx->dst_sg, rctx->dst_skip); in aead_enqueue()
|