| /drivers/crypto/amlogic/ |
| A D | amlogic-gxl-cipher.c | 33 if (sg_nents(src_sg) != sg_nents(dst_sg)) in meson_cipher_need_fallback() 37 if (sg_nents(src_sg) > MAXDESC - 3 || sg_nents(dst_sg) > MAXDESC - 3) in meson_cipher_need_fallback() 178 nr_sgs = dma_map_sg(mc->dev, areq->src, sg_nents(areq->src), in meson_cipher() 187 nr_sgs = dma_map_sg(mc->dev, areq->src, sg_nents(areq->src), in meson_cipher() 194 nr_sgd = dma_map_sg(mc->dev, areq->dst, sg_nents(areq->dst), in meson_cipher() 239 dma_unmap_sg(mc->dev, areq->src, sg_nents(areq->src), DMA_BIDIRECTIONAL); in meson_cipher() 241 dma_unmap_sg(mc->dev, areq->src, sg_nents(areq->src), DMA_TO_DEVICE); in meson_cipher() 242 dma_unmap_sg(mc->dev, areq->dst, sg_nents(areq->dst), DMA_FROM_DEVICE); in meson_cipher()
|
| /drivers/infiniband/sw/rdmavt/ |
| A D | trace_mr.h | 144 TP_PROTO(struct ib_mr *ibmr, int sg_nents, unsigned int *sg_offset), 145 TP_ARGS(ibmr, sg_nents, sg_offset), 152 __field(int, sg_nents) 161 __entry->sg_nents = sg_nents; 171 __entry->sg_nents,
|
| A D | mr.h | 35 int sg_nents, unsigned int *sg_offset);
|
| A D | mr.c | 594 int sg_nents, unsigned int *sg_offset) in rvt_map_mr_sg() argument 601 ret = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, rvt_set_page); in rvt_map_mr_sg() 606 trace_rvt_map_mr_sg(ibmr, sg_nents, sg_offset); in rvt_map_mr_sg()
|
| /drivers/crypto/gemini/ |
| A D | sl3516-ce-cipher.c | 43 if (sg_nents(areq->src) > MAXDESC / 2) { in sl3516_ce_need_fallback() 48 if (sg_nents(areq->dst) > MAXDESC) { in sl3516_ce_need_fallback() 154 nr_sgs = dma_map_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() 163 nr_sgs = dma_map_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() 170 nr_sgd = dma_map_sg(ce->dev, areq->dst, sg_nents(areq->dst), in sl3516_ce_cipher() 251 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() 254 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() 256 dma_unmap_sg(ce->dev, areq->dst, sg_nents(areq->dst), in sl3516_ce_cipher()
|
| /drivers/crypto/intel/iaa/ |
| A D | iaa_crypto_main.c | 1236 dma_unmap_sg(dev, req->dst, sg_nents(req->dst), DMA_FROM_DEVICE); in iaa_remap_for_verify() 1237 dma_unmap_sg(dev, req->src, sg_nents(req->src), DMA_TO_DEVICE); in iaa_remap_for_verify() 1252 nr_sgs = dma_map_sg(dev, req->dst, sg_nents(req->dst), DMA_TO_DEVICE); in iaa_remap_for_verify() 1258 dma_unmap_sg(dev, req->src, sg_nents(req->src), DMA_FROM_DEVICE); in iaa_remap_for_verify() 1517 nr_sgs = dma_map_sg(dev, req->src, sg_nents(req->src), DMA_TO_DEVICE); in iaa_comp_acompress() 1560 dma_unmap_sg(dev, req->dst, sg_nents(req->dst), DMA_TO_DEVICE); in iaa_comp_acompress() 1561 dma_unmap_sg(dev, req->src, sg_nents(req->src), DMA_FROM_DEVICE); in iaa_comp_acompress() 1569 dma_unmap_sg(dev, req->dst, sg_nents(req->dst), DMA_FROM_DEVICE); in iaa_comp_acompress() 1571 dma_unmap_sg(dev, req->src, sg_nents(req->src), DMA_TO_DEVICE); in iaa_comp_acompress() 1649 dma_unmap_sg(dev, req->dst, sg_nents(req->dst), DMA_FROM_DEVICE); in iaa_comp_adecompress() [all …]
|
| /drivers/crypto/intel/qat/qat_common/ |
| A D | qat_bl.c | 62 int n = sg_nents(sgl); in __qat_bl_sgl_to_bufl() 130 int n_sglout = sg_nents(sglout); in __qat_bl_sgl_to_bufl() 200 n = sg_nents(sglout); in __qat_bl_sgl_to_bufl() 217 n = sg_nents(sgl); in __qat_bl_sgl_to_bufl()
|
| /drivers/crypto/cavium/nitrox/ |
| A D | nitrox_reqmgr.c | 61 dma_unmap_sg(dev, sr->in.sg, sg_nents(sr->in.sg), in softreq_unmap_sgbufs() 69 dma_unmap_sg(dev, sr->out.sg, sg_nents(sr->out.sg), in softreq_unmap_sgbufs() 165 nents = dma_map_sg(dev, req->src, sg_nents(req->src), in dma_map_inbufs() 182 dma_unmap_sg(dev, req->src, sg_nents(req->src), DMA_BIDIRECTIONAL); in dma_map_inbufs() 193 nents = dma_map_sg(dev, req->dst, sg_nents(req->dst), in dma_map_outbufs() 207 dma_unmap_sg(dev, req->dst, sg_nents(req->dst), DMA_BIDIRECTIONAL); in dma_map_outbufs()
|
| A D | nitrox_skcipher.c | 211 int nents = sg_nents(skreq->src) + 1; in alloc_src_sglist() 229 int nents = sg_nents(skreq->dst) + 3; in alloc_dst_sglist()
|
| /drivers/crypto/bcm/ |
| A D | util.c | 54 unsigned int nents = sg_nents(src); in sg_copy_part_to_buf() 73 unsigned int nents = sg_nents(dest); in sg_copy_part_from_buf() 96 int sg_nents = 0; in spu_sg_count() local 106 sg_nents++; in spu_sg_count() 111 return sg_nents; in spu_sg_count()
|
| /drivers/crypto/hisilicon/ |
| A D | sgl.c | 233 sg_n = sg_nents(sgl); in hisi_acc_sg_buf_map_to_hw_sgl() 289 dma_unmap_sg(dev, sgl, sg_nents(sgl), dir); in hisi_acc_sg_buf_unmap()
|
| /drivers/mailbox/ |
| A D | bcm-pdc-mailbox.c | 608 sg_nents(pdcs->src_sg[pdcs->txin]), DMA_TO_DEVICE); in pdc_receive_one() 623 dma_unmap_sg(dev, rx_ctx->dst_sg, sg_nents(rx_ctx->dst_sg), in pdc_receive_one() 717 num_desc = (u32)sg_nents(sg); in pdc_tx_list_sg_add() 884 num_desc = (u32)sg_nents(sg); in pdc_rx_list_sg_add() 1205 src_nent = sg_nents(mssg->spu.src); in pdc_send_data() 1212 dst_nent = sg_nents(mssg->spu.dst); in pdc_send_data()
|
| A D | bcm-flexrm-mailbox.c | 623 rc = dma_map_sg(dev, msg->spu.src, sg_nents(msg->spu.src), in flexrm_spu_dma_map() 628 rc = dma_map_sg(dev, msg->spu.dst, sg_nents(msg->spu.dst), in flexrm_spu_dma_map() 631 dma_unmap_sg(dev, msg->spu.src, sg_nents(msg->spu.src), in flexrm_spu_dma_map() 641 dma_unmap_sg(dev, msg->spu.dst, sg_nents(msg->spu.dst), in flexrm_spu_dma_unmap() 643 dma_unmap_sg(dev, msg->spu.src, sg_nents(msg->spu.src), in flexrm_spu_dma_unmap()
|
| /drivers/crypto/allwinner/sun4i-ss/ |
| A D | sun4i-ss-cipher.c | 80 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_opti_poll() 108 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_opti_poll() 272 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_cipher_poll() 326 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_cipher_poll()
|
| A D | sun4i-ss-hash.c | 220 copied = sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in sun4i_hash() 272 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_hash()
|
| /drivers/infiniband/hw/vmw_pvrdma/ |
| A D | pvrdma_mr.c | 318 int pvrdma_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, in pvrdma_map_mr_sg() argument 327 ret = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, pvrdma_set_page); in pvrdma_map_mr_sg()
|
| /drivers/infiniband/hw/ocrdma/ |
| A D | ocrdma_verbs.h | 105 int ocrdma_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents,
|
| /drivers/usb/storage/ |
| A D | protocol.c | 131 nents = sg_nents(sg); in usb_stor_access_xfer_buf()
|
| /drivers/crypto/hisilicon/sec/ |
| A D | sec_algs.c | 460 sg_nents(sec_req_el->sgl_out), in sec_skcipher_alg_callback() 467 sg_nents(sec_req_el->sgl_in), in sec_skcipher_alg_callback() 730 sec_req->len_in = sg_nents(skreq->src); in sec_alg_skcipher_crypto() 744 sec_req->len_out = sg_nents(skreq->dst); in sec_alg_skcipher_crypto()
|
| /drivers/crypto/chelsio/ |
| A D | chcr_algo.c | 185 sg_pcopy_to_buffer(req->src, sg_nents(req->src), temp, in chcr_verify_tag() 890 sg_nents(wrparam->req->src), wrparam->req->iv, 16, in create_cipher_wr() 1732 sg_pcopy_to_buffer(req->src, sg_nents(req->src), in chcr_ahash_update() 2698 sg_pcopy_to_buffer(req->src, sg_nents(req->src), in chcr_add_aead_src_ent() 2744 sg_pcopy_to_buffer(req->src, sg_nents(req->src), in chcr_add_cipher_src_ent() 2793 sg_nents(reqctx->hctx_wr.srcsg), buf, in chcr_add_hash_src_ent() 2832 dma_unmap_sg(dev, req->src, sg_nents(req->src), in chcr_hash_dma_unmap() 2856 dma_unmap_sg(dev, req->src, sg_nents(req->src), in chcr_cipher_dma_map() 2871 dma_unmap_sg(dev, req->src, sg_nents(req->src), in chcr_cipher_dma_unmap() 2874 dma_unmap_sg(dev, req->src, sg_nents(req->src), in chcr_cipher_dma_unmap() [all …]
|
| /drivers/crypto/allwinner/sun8i-ss/ |
| A D | sun8i-ss-hash.c | 365 if (sg_nents(areq->src) > MAX_SG - 1) { in sun8i_ss_hash_need_fallback() 525 nr_sgs = dma_map_sg(ss->dev, areq->src, sg_nents(areq->src), DMA_TO_DEVICE); in sun8i_ss_hash_run() 691 dma_unmap_sg(ss->dev, areq->src, sg_nents(areq->src), in sun8i_ss_hash_run()
|
| /drivers/infiniband/hw/qedr/ |
| A D | verbs.h | 86 int sg_nents, unsigned int *sg_offset);
|
| /drivers/infiniband/hw/mlx4/ |
| A D | mr.c | 442 int mlx4_ib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, in mlx4_ib_map_mr_sg() argument 453 rc = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, mlx4_set_page); in mlx4_ib_map_mr_sg()
|
| /drivers/crypto/virtio/ |
| A D | virtio_crypto_akcipher_algs.c | 92 sg_copy_from_buffer(akcipher_req->dst, sg_nents(akcipher_req->dst), in virtio_crypto_dataq_akcipher_callback() 235 sg_copy_to_buffer(req->src, sg_nents(req->src), src_buf, req->src_len); in __virtio_crypto_akcipher_do_req()
|
| /drivers/infiniband/ulp/isert/ |
| A D | ib_isert.c | 1051 int imm_data, imm_data_len, unsol_data, sg_nents, rc; in isert_handle_scsi_cmd() local 1078 sg_nents = max(1UL, DIV_ROUND_UP(imm_data_len, PAGE_SIZE)); in isert_handle_scsi_cmd() 1079 sg_copy_from_buffer(cmd->se_cmd.t_data_sg, sg_nents, in isert_handle_scsi_cmd() 1082 sg_nents, imm_data_len); in isert_handle_scsi_cmd() 1122 int rc, sg_nents, sg_off, page_off; in isert_handle_iscsi_dataout() local 1145 sg_nents = max(1UL, DIV_ROUND_UP(unsol_data_len, PAGE_SIZE)); in isert_handle_iscsi_dataout() 1157 sg_nents, isert_get_data(rx_desc), unsol_data_len); in isert_handle_iscsi_dataout() 1159 sg_copy_from_buffer(sg_start, sg_nents, isert_get_data(rx_desc), in isert_handle_iscsi_dataout()
|