Home
last modified time | relevance | path

Searched refs:sg_nents (Results 1 – 25 of 84) sorted by relevance

1234

/drivers/crypto/amlogic/
A Damlogic-gxl-cipher.c33 if (sg_nents(src_sg) != sg_nents(dst_sg)) in meson_cipher_need_fallback()
37 if (sg_nents(src_sg) > MAXDESC - 3 || sg_nents(dst_sg) > MAXDESC - 3) in meson_cipher_need_fallback()
178 nr_sgs = dma_map_sg(mc->dev, areq->src, sg_nents(areq->src), in meson_cipher()
187 nr_sgs = dma_map_sg(mc->dev, areq->src, sg_nents(areq->src), in meson_cipher()
194 nr_sgd = dma_map_sg(mc->dev, areq->dst, sg_nents(areq->dst), in meson_cipher()
239 dma_unmap_sg(mc->dev, areq->src, sg_nents(areq->src), DMA_BIDIRECTIONAL); in meson_cipher()
241 dma_unmap_sg(mc->dev, areq->src, sg_nents(areq->src), DMA_TO_DEVICE); in meson_cipher()
242 dma_unmap_sg(mc->dev, areq->dst, sg_nents(areq->dst), DMA_FROM_DEVICE); in meson_cipher()
/drivers/infiniband/sw/rdmavt/
A Dtrace_mr.h144 TP_PROTO(struct ib_mr *ibmr, int sg_nents, unsigned int *sg_offset),
145 TP_ARGS(ibmr, sg_nents, sg_offset),
152 __field(int, sg_nents)
161 __entry->sg_nents = sg_nents;
171 __entry->sg_nents,
A Dmr.h35 int sg_nents, unsigned int *sg_offset);
A Dmr.c594 int sg_nents, unsigned int *sg_offset) in rvt_map_mr_sg() argument
601 ret = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, rvt_set_page); in rvt_map_mr_sg()
606 trace_rvt_map_mr_sg(ibmr, sg_nents, sg_offset); in rvt_map_mr_sg()
/drivers/crypto/gemini/
A Dsl3516-ce-cipher.c43 if (sg_nents(areq->src) > MAXDESC / 2) { in sl3516_ce_need_fallback()
48 if (sg_nents(areq->dst) > MAXDESC) { in sl3516_ce_need_fallback()
154 nr_sgs = dma_map_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher()
163 nr_sgs = dma_map_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher()
170 nr_sgd = dma_map_sg(ce->dev, areq->dst, sg_nents(areq->dst), in sl3516_ce_cipher()
251 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher()
254 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher()
256 dma_unmap_sg(ce->dev, areq->dst, sg_nents(areq->dst), in sl3516_ce_cipher()
/drivers/crypto/intel/iaa/
A Diaa_crypto_main.c1236 dma_unmap_sg(dev, req->dst, sg_nents(req->dst), DMA_FROM_DEVICE); in iaa_remap_for_verify()
1237 dma_unmap_sg(dev, req->src, sg_nents(req->src), DMA_TO_DEVICE); in iaa_remap_for_verify()
1252 nr_sgs = dma_map_sg(dev, req->dst, sg_nents(req->dst), DMA_TO_DEVICE); in iaa_remap_for_verify()
1258 dma_unmap_sg(dev, req->src, sg_nents(req->src), DMA_FROM_DEVICE); in iaa_remap_for_verify()
1517 nr_sgs = dma_map_sg(dev, req->src, sg_nents(req->src), DMA_TO_DEVICE); in iaa_comp_acompress()
1560 dma_unmap_sg(dev, req->dst, sg_nents(req->dst), DMA_TO_DEVICE); in iaa_comp_acompress()
1561 dma_unmap_sg(dev, req->src, sg_nents(req->src), DMA_FROM_DEVICE); in iaa_comp_acompress()
1569 dma_unmap_sg(dev, req->dst, sg_nents(req->dst), DMA_FROM_DEVICE); in iaa_comp_acompress()
1571 dma_unmap_sg(dev, req->src, sg_nents(req->src), DMA_TO_DEVICE); in iaa_comp_acompress()
1649 dma_unmap_sg(dev, req->dst, sg_nents(req->dst), DMA_FROM_DEVICE); in iaa_comp_adecompress()
[all …]
/drivers/crypto/intel/qat/qat_common/
A Dqat_bl.c62 int n = sg_nents(sgl); in __qat_bl_sgl_to_bufl()
130 int n_sglout = sg_nents(sglout); in __qat_bl_sgl_to_bufl()
200 n = sg_nents(sglout); in __qat_bl_sgl_to_bufl()
217 n = sg_nents(sgl); in __qat_bl_sgl_to_bufl()
/drivers/crypto/cavium/nitrox/
A Dnitrox_reqmgr.c61 dma_unmap_sg(dev, sr->in.sg, sg_nents(sr->in.sg), in softreq_unmap_sgbufs()
69 dma_unmap_sg(dev, sr->out.sg, sg_nents(sr->out.sg), in softreq_unmap_sgbufs()
165 nents = dma_map_sg(dev, req->src, sg_nents(req->src), in dma_map_inbufs()
182 dma_unmap_sg(dev, req->src, sg_nents(req->src), DMA_BIDIRECTIONAL); in dma_map_inbufs()
193 nents = dma_map_sg(dev, req->dst, sg_nents(req->dst), in dma_map_outbufs()
207 dma_unmap_sg(dev, req->dst, sg_nents(req->dst), DMA_BIDIRECTIONAL); in dma_map_outbufs()
A Dnitrox_skcipher.c211 int nents = sg_nents(skreq->src) + 1; in alloc_src_sglist()
229 int nents = sg_nents(skreq->dst) + 3; in alloc_dst_sglist()
/drivers/crypto/bcm/
A Dutil.c54 unsigned int nents = sg_nents(src); in sg_copy_part_to_buf()
73 unsigned int nents = sg_nents(dest); in sg_copy_part_from_buf()
96 int sg_nents = 0; in spu_sg_count() local
106 sg_nents++; in spu_sg_count()
111 return sg_nents; in spu_sg_count()
/drivers/crypto/hisilicon/
A Dsgl.c233 sg_n = sg_nents(sgl); in hisi_acc_sg_buf_map_to_hw_sgl()
289 dma_unmap_sg(dev, sgl, sg_nents(sgl), dir); in hisi_acc_sg_buf_unmap()
/drivers/mailbox/
A Dbcm-pdc-mailbox.c608 sg_nents(pdcs->src_sg[pdcs->txin]), DMA_TO_DEVICE); in pdc_receive_one()
623 dma_unmap_sg(dev, rx_ctx->dst_sg, sg_nents(rx_ctx->dst_sg), in pdc_receive_one()
717 num_desc = (u32)sg_nents(sg); in pdc_tx_list_sg_add()
884 num_desc = (u32)sg_nents(sg); in pdc_rx_list_sg_add()
1205 src_nent = sg_nents(mssg->spu.src); in pdc_send_data()
1212 dst_nent = sg_nents(mssg->spu.dst); in pdc_send_data()
A Dbcm-flexrm-mailbox.c623 rc = dma_map_sg(dev, msg->spu.src, sg_nents(msg->spu.src), in flexrm_spu_dma_map()
628 rc = dma_map_sg(dev, msg->spu.dst, sg_nents(msg->spu.dst), in flexrm_spu_dma_map()
631 dma_unmap_sg(dev, msg->spu.src, sg_nents(msg->spu.src), in flexrm_spu_dma_map()
641 dma_unmap_sg(dev, msg->spu.dst, sg_nents(msg->spu.dst), in flexrm_spu_dma_unmap()
643 dma_unmap_sg(dev, msg->spu.src, sg_nents(msg->spu.src), in flexrm_spu_dma_unmap()
/drivers/crypto/allwinner/sun4i-ss/
A Dsun4i-ss-cipher.c80 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_opti_poll()
108 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_opti_poll()
272 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_cipher_poll()
326 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_cipher_poll()
A Dsun4i-ss-hash.c220 copied = sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in sun4i_hash()
272 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_hash()
/drivers/infiniband/hw/vmw_pvrdma/
A Dpvrdma_mr.c318 int pvrdma_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, in pvrdma_map_mr_sg() argument
327 ret = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, pvrdma_set_page); in pvrdma_map_mr_sg()
/drivers/infiniband/hw/ocrdma/
A Docrdma_verbs.h105 int ocrdma_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents,
/drivers/usb/storage/
A Dprotocol.c131 nents = sg_nents(sg); in usb_stor_access_xfer_buf()
/drivers/crypto/hisilicon/sec/
A Dsec_algs.c460 sg_nents(sec_req_el->sgl_out), in sec_skcipher_alg_callback()
467 sg_nents(sec_req_el->sgl_in), in sec_skcipher_alg_callback()
730 sec_req->len_in = sg_nents(skreq->src); in sec_alg_skcipher_crypto()
744 sec_req->len_out = sg_nents(skreq->dst); in sec_alg_skcipher_crypto()
/drivers/crypto/chelsio/
A Dchcr_algo.c185 sg_pcopy_to_buffer(req->src, sg_nents(req->src), temp, in chcr_verify_tag()
890 sg_nents(wrparam->req->src), wrparam->req->iv, 16, in create_cipher_wr()
1732 sg_pcopy_to_buffer(req->src, sg_nents(req->src), in chcr_ahash_update()
2698 sg_pcopy_to_buffer(req->src, sg_nents(req->src), in chcr_add_aead_src_ent()
2744 sg_pcopy_to_buffer(req->src, sg_nents(req->src), in chcr_add_cipher_src_ent()
2793 sg_nents(reqctx->hctx_wr.srcsg), buf, in chcr_add_hash_src_ent()
2832 dma_unmap_sg(dev, req->src, sg_nents(req->src), in chcr_hash_dma_unmap()
2856 dma_unmap_sg(dev, req->src, sg_nents(req->src), in chcr_cipher_dma_map()
2871 dma_unmap_sg(dev, req->src, sg_nents(req->src), in chcr_cipher_dma_unmap()
2874 dma_unmap_sg(dev, req->src, sg_nents(req->src), in chcr_cipher_dma_unmap()
[all …]
/drivers/crypto/allwinner/sun8i-ss/
A Dsun8i-ss-hash.c365 if (sg_nents(areq->src) > MAX_SG - 1) { in sun8i_ss_hash_need_fallback()
525 nr_sgs = dma_map_sg(ss->dev, areq->src, sg_nents(areq->src), DMA_TO_DEVICE); in sun8i_ss_hash_run()
691 dma_unmap_sg(ss->dev, areq->src, sg_nents(areq->src), in sun8i_ss_hash_run()
/drivers/infiniband/hw/qedr/
A Dverbs.h86 int sg_nents, unsigned int *sg_offset);
/drivers/infiniband/hw/mlx4/
A Dmr.c442 int mlx4_ib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, in mlx4_ib_map_mr_sg() argument
453 rc = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, mlx4_set_page); in mlx4_ib_map_mr_sg()
/drivers/crypto/virtio/
A Dvirtio_crypto_akcipher_algs.c92 sg_copy_from_buffer(akcipher_req->dst, sg_nents(akcipher_req->dst), in virtio_crypto_dataq_akcipher_callback()
235 sg_copy_to_buffer(req->src, sg_nents(req->src), src_buf, req->src_len); in __virtio_crypto_akcipher_do_req()
/drivers/infiniband/ulp/isert/
A Dib_isert.c1051 int imm_data, imm_data_len, unsol_data, sg_nents, rc; in isert_handle_scsi_cmd() local
1078 sg_nents = max(1UL, DIV_ROUND_UP(imm_data_len, PAGE_SIZE)); in isert_handle_scsi_cmd()
1079 sg_copy_from_buffer(cmd->se_cmd.t_data_sg, sg_nents, in isert_handle_scsi_cmd()
1082 sg_nents, imm_data_len); in isert_handle_scsi_cmd()
1122 int rc, sg_nents, sg_off, page_off; in isert_handle_iscsi_dataout() local
1145 sg_nents = max(1UL, DIV_ROUND_UP(unsol_data_len, PAGE_SIZE)); in isert_handle_iscsi_dataout()
1157 sg_nents, isert_get_data(rx_desc), unsol_data_len); in isert_handle_iscsi_dataout()
1159 sg_copy_from_buffer(sg_start, sg_nents, isert_get_data(rx_desc), in isert_handle_iscsi_dataout()

Completed in 87 milliseconds

1234