| /drivers/crypto/amlogic/ |
| A D | amlogic-gxl-cipher.c | 27 struct scatterlist *src_sg = areq->src; in meson_cipher_need_fallback() local 33 if (sg_nents(src_sg) != sg_nents(dst_sg)) in meson_cipher_need_fallback() 40 while (src_sg && dst_sg) { in meson_cipher_need_fallback() 41 if ((src_sg->length % 16) != 0) in meson_cipher_need_fallback() 45 if (src_sg->length != dst_sg->length) in meson_cipher_need_fallback() 51 src_sg = sg_next(src_sg); in meson_cipher_need_fallback() 203 src_sg = areq->src; in meson_cipher() 206 while (src_sg) { in meson_cipher() 212 todo = min(len, sg_dma_len(src_sg)); in meson_cipher() 218 if (!sg_next(src_sg)) in meson_cipher() [all …]
|
| /drivers/crypto/ccp/ |
| A D | ccp-dmaengine.c | 355 struct scatterlist *src_sg, in ccp_create_desc() argument 372 if (!dst_sg || !src_sg) in ccp_create_desc() 384 src_len = sg_dma_len(src_sg); in ccp_create_desc() 396 src_sg = sg_next(src_sg); in ccp_create_desc() 397 if (!src_sg) in ccp_create_desc() 400 src_len = sg_dma_len(src_sg); in ccp_create_desc() 433 ccp_pt->src_dma = sg_dma_address(src_sg) + src_offset; in ccp_create_desc() 485 struct scatterlist dst_sg, src_sg; in ccp_prep_dma_memcpy() local 495 sg_init_table(&src_sg, 1); in ccp_prep_dma_memcpy() 496 sg_dma_address(&src_sg) = src; in ccp_prep_dma_memcpy() [all …]
|
| /drivers/scsi/arm/ |
| A D | arm_scsi.h | 45 struct scatterlist *src_sg; in copy_SCp_to_sg() local 48 for_each_sg(sg_next(SCp->buffer), src_sg, bufs, i) in copy_SCp_to_sg() 49 *(++sg) = *src_sg; in copy_SCp_to_sg()
|
| /drivers/mailbox/ |
| A D | bcm-flexrm-mailbox.c | 598 while (src_sg || dst_sg) { in flexrm_spu_estimate_nonheader_desc_count() 599 if (src_sg) { in flexrm_spu_estimate_nonheader_desc_count() 601 dst_target = src_sg->length; in flexrm_spu_estimate_nonheader_desc_count() 602 src_sg = sg_next(src_sg); in flexrm_spu_estimate_nonheader_desc_count() 657 while (src_sg || dst_sg) { in flexrm_spu_write_descs() 658 if (src_sg) { in flexrm_spu_write_descs() 659 if (sg_dma_len(src_sg) & 0xf) in flexrm_spu_write_descs() 661 sg_dma_len(src_sg)); in flexrm_spu_write_descs() 664 sg_dma_len(src_sg)/16); in flexrm_spu_write_descs() 669 dst_target = sg_dma_len(src_sg); in flexrm_spu_write_descs() [all …]
|
| A D | bcm-pdc-mailbox.c | 392 struct scatterlist *src_sg[PDC_RING_ENTRIES]; member 607 dma_unmap_sg(dev, pdcs->src_sg[pdcs->txin], in pdc_receive_one() 608 sg_nents(pdcs->src_sg[pdcs->txin]), DMA_TO_DEVICE); in pdc_receive_one() 731 pdcs->src_sg[pdcs->txout] = sg; in pdc_tx_list_sg_add()
|
| /drivers/crypto/qce/ |
| A D | aead.c | 49 dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_aead_done() 277 rctx->src_sg = rctx->src_tbl.sgl; in qce_aead_ccm_prepare_buf_assoclen() 279 rctx->src_nents = sg_nents_for_len(rctx->src_sg, totallen); in qce_aead_ccm_prepare_buf_assoclen() 292 rctx->dst_sg = rctx->src_sg; in qce_aead_ccm_prepare_buf_assoclen() 321 rctx->src_sg = req->src; in qce_aead_prepare_buf() 324 rctx->src_sg = rctx->dst_sg; in qce_aead_prepare_buf() 347 rctx->src_sg = req->src; in qce_aead_ccm_prepare_buf() 353 rctx->src_sg = req->src; in qce_aead_ccm_prepare_buf() 355 rctx->dst_sg = rctx->src_sg; in qce_aead_ccm_prepare_buf() 459 src_nents = dma_map_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_aead_async_req_handle() [all …]
|
| A D | skcipher.c | 50 dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_skcipher_done() 138 rctx->src_sg = req->src; in qce_skcipher_async_req_handle() 140 rctx->src_sg = rctx->dst_sg; in qce_skcipher_async_req_handle() 144 ret = qce_dma_prep_sgs(&qce->dma, rctx->src_sg, src_nents, in qce_skcipher_async_req_handle()
|
| A D | cipher.h | 43 struct scatterlist *src_sg; member
|
| A D | aead.h | 37 struct scatterlist *src_sg; member
|
| /drivers/dma/ |
| A D | nbpfaxi.c | 933 struct scatterlist *src_sg, struct scatterlist *dst_sg, in nbpf_prep_sg() argument 952 mem_sg = src_sg; in nbpf_prep_sg() 959 mem_sg = src_sg; in nbpf_prep_sg() 978 sg_dma_address(src_sg), in nbpf_prep_sg() 988 src_sg = sg_next(src_sg); in nbpf_prep_sg() 991 mem_sg = direction == DMA_DEV_TO_MEM ? dst_sg : src_sg; in nbpf_prep_sg() 1007 struct scatterlist src_sg; in nbpf_prep_memcpy() local 1010 sg_init_table(&src_sg, 1); in nbpf_prep_memcpy() 1013 sg_dma_address(&src_sg) = src; in nbpf_prep_memcpy() 1016 sg_dma_len(&src_sg) = len; in nbpf_prep_memcpy() [all …]
|
| A D | ste_dma40.c | 2503 struct scatterlist src_sg; in d40_prep_memcpy() local 2506 sg_init_table(&src_sg, 1); in d40_prep_memcpy() 2509 sg_dma_address(&src_sg) = src; in d40_prep_memcpy() 2512 sg_dma_len(&src_sg) = size; in d40_prep_memcpy() 2514 return d40_prep_sg(chan, &src_sg, &dst_sg, 1, in d40_prep_memcpy()
|
| /drivers/crypto/aspeed/ |
| A D | aspeed-hace-hash.c | 185 scatterwalk_map_and_copy(hash_engine->ahash_src_addr, rctx->src_sg, in aspeed_ahash_dma_prepare() 232 sg_len = dma_map_sg(hace_dev->dev, rctx->src_sg, rctx->src_nents, in aspeed_ahash_dma_prepare_sg() 253 for_each_sg(rctx->src_sg, s, sg_len, i) { in aspeed_ahash_dma_prepare_sg() 316 dma_unmap_sg(hace_dev->dev, rctx->src_sg, rctx->src_nents, in aspeed_ahash_dma_prepare_sg() 389 dma_unmap_sg(hace_dev->dev, rctx->src_sg, rctx->src_nents, in aspeed_ahash_update_resume_sg() 512 rctx->src_sg = req->src; in aspeed_sham_update()
|
| A D | aspeed-hace.h | 178 struct scatterlist *src_sg; member
|
| /drivers/crypto/tegra/ |
| A D | tegra-se-aes.c | 57 struct scatterlist *src_sg; member 80 struct scatterlist *src_sg; member 763 rctx->src_sg, 0, rctx->assoclen, 0); in tegra_gcm_do_gmac() 1039 rctx->src_sg, 0, rctx->assoclen, 0); in tegra_ccm_format_blocks() 1099 sg = rctx->encrypt ? rctx->src_sg : rctx->dst_sg; in tegra_ccm_compute_auth() 1120 struct scatterlist *sg = rctx->src_sg; in tegra_ccm_do_ctr() 1164 rctx->src_sg = req->src; in tegra_ccm_crypt_init() 1277 rctx->src_sg = req->src; in tegra_gcm_do_one_req() 1592 rctx->src_sg = req->src; in tegra_cmac_do_update() 1604 rctx->src_sg, 0, req->nbytes, 0); in tegra_cmac_do_update() [all …]
|
| A D | tegra-se-hash.c | 33 struct scatterlist *src_sg; member 418 rctx->src_sg = req->src; in tegra_sha_do_update() 427 rctx->src_sg, 0, req->nbytes, 0); in tegra_sha_do_update() 443 rctx->src_sg, 0, req->nbytes - nresidue, 0); in tegra_sha_do_update() 445 scatterwalk_map_and_copy(rctx->residue.buf, rctx->src_sg, in tegra_sha_do_update()
|
| /drivers/crypto/bcm/ |
| A D | cipher.c | 225 datalen = spu_msg_sg_add(&sg, &rctx->src_sg, &rctx->src_skip, in spu_skcipher_tx_sg_create() 422 dump_sg(rctx->src_sg, rctx->src_skip, chunksize); in handle_skcipher_req() 606 datalen = spu_msg_sg_add(&sg, &rctx->src_sg, &rctx->src_skip, in spu_ahash_tx_sg_create() 769 rctx->src_nents = spu_sg_count(rctx->src_sg, rctx->src_skip, in handle_ahash_req() 854 dump_sg(rctx->src_sg, rctx->src_skip, new_data_len); in handle_ahash_req() 1461 dump_sg(rctx->src_sg, rctx->src_skip, chunksize); in handle_aead_req() 1734 rctx->src_sg = req->src; in skcipher_enqueue() 1921 rctx->src_sg = req->src; in ahash_enqueue() 2639 if (spu_sg_at_offset(req->src, req->assoclen, &rctx->src_sg, in aead_enqueue() 2649 rctx->dst_sg = rctx->src_sg; in aead_enqueue() [all …]
|
| A D | cipher.h | 286 struct scatterlist *src_sg; member
|