Home
last modified time | relevance | path

Searched refs:sg_list (Results 1 – 25 of 148) sorted by relevance

123456

/linux/drivers/virt/
A Dfsl_hypervisor.c153 struct fh_sg_list *sg_list = NULL; in ioctl_memcpy() local
264 sg_list[0].target = param.remote_paddr; in ioctl_memcpy()
266 sg_list[0].source = param.remote_paddr; in ioctl_memcpy()
272 count = param.count - sg_list[0].size; in ioctl_memcpy()
277 sg_list[i].source = page_to_phys(pages[i]); in ioctl_memcpy()
278 sg_list[i].target = remote_paddr; in ioctl_memcpy()
281 sg_list[i].source = remote_paddr; in ioctl_memcpy()
282 sg_list[i].target = page_to_phys(pages[i]); in ioctl_memcpy()
286 remote_paddr += sg_list[i].size; in ioctl_memcpy()
287 count -= sg_list[i].size; in ioctl_memcpy()
[all …]
/linux/drivers/scsi/qla2xxx/
A Dqla_bsg.c52 bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free()
57 bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free()
206 bsg_job->reply_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd()
963 bsg_job->reply_payload.sg_list, in qla2x00_process_loopback()
967 bsg_job->request_payload.sg_list, in qla2x00_process_loopback()
1158 bsg_job->reply_payload.sg_list, in qla84xx_mgmt_cmd()
1208 bsg_job->request_payload.sg_list, in qla84xx_mgmt_cmd()
1890 bsg_job->request_payload.sg_list, in qla24xx_process_bidir_cmd()
1958 bsg_job->reply_payload.sg_list, in qla24xx_process_bidir_cmd()
2022 bsg_job->reply_payload.sg_list, in qlafx00_mgmt_cmd()
[all …]
/linux/drivers/tee/amdtee/
A Dcall.c315 cmd->sg_list.count = count; in handle_map_shmem()
320 cmd->sg_list.buf[i].hi_addr = upper_32_bits(paddr); in handle_map_shmem()
321 cmd->sg_list.buf[i].low_addr = lower_32_bits(paddr); in handle_map_shmem()
322 cmd->sg_list.buf[i].size = start[i].size; in handle_map_shmem()
323 cmd->sg_list.size += cmd->sg_list.buf[i].size; in handle_map_shmem()
326 cmd->sg_list.buf[i].hi_addr); in handle_map_shmem()
328 cmd->sg_list.buf[i].low_addr); in handle_map_shmem()
329 pr_debug("buf[%d]:size = 0x%x\n", i, cmd->sg_list.buf[i].size); in handle_map_shmem()
330 pr_debug("list size = 0x%x\n", cmd->sg_list.size); in handle_map_shmem()
/linux/block/
A Dbsg-lib.c163 kfree(job->request_payload.sg_list); in bsg_teardown_job()
164 kfree(job->reply_payload.sg_list); in bsg_teardown_job()
218 buf->sg_list = kmalloc(sz, GFP_KERNEL); in bsg_map_buffer()
219 if (!buf->sg_list) in bsg_map_buffer()
221 sg_init_table(buf->sg_list, req->nr_phys_segments); in bsg_map_buffer()
222 buf->sg_cnt = blk_rq_map_sg(req->q, req, buf->sg_list); in bsg_map_buffer()
256 kfree(job->request_payload.sg_list); in bsg_prepare_job()
/linux/drivers/infiniband/sw/rdmavt/
A Dqp.c1831 wqe->sg_list[i].addr = wr->sg_list[i].addr; in rvt_post_recv()
1832 wqe->sg_list[i].length = wr->sg_list[i].length; in rvt_post_recv()
1833 wqe->sg_list[i].lkey = wr->sg_list[i].lkey; in rvt_post_recv()
2058 &wr->sg_list[i], acc); in rvt_post_one_wr()
2238 wqe->sg_list[i].addr = wr->sg_list[i].addr; in rvt_post_srq_recv()
2239 wqe->sg_list[i].length = wr->sg_list[i].length; in rvt_post_srq_recv()
2240 wqe->sg_list[i].lkey = wr->sg_list[i].lkey; in rvt_post_srq_recv()
2982 sqp->s_sge.sg_list = wqe->sg_list + 1; in rvt_ruc_loopback()
3050 qp->r_sge.sg_list = NULL; in rvt_ruc_loopback()
3064 sqp->s_sge.sg_list = NULL; in rvt_ruc_loopback()
[all …]
A Drc.c164 ss->sge = wqe->sg_list[0]; in rvt_restart_sge()
165 ss->sg_list = wqe->sg_list + 1; in rvt_restart_sge()
/linux/drivers/dma/
A Dimx-dma.c159 struct scatterlist *sg_list; member
791 kfree(imxdmac->sg_list); in imxdma_free_chan_resources()
792 imxdmac->sg_list = NULL; in imxdma_free_chan_resources()
868 kfree(imxdmac->sg_list); in imxdma_prep_dma_cyclic()
870 imxdmac->sg_list = kcalloc(periods + 1, in imxdma_prep_dma_cyclic()
872 if (!imxdmac->sg_list) in imxdma_prep_dma_cyclic()
875 sg_init_table(imxdmac->sg_list, periods); in imxdma_prep_dma_cyclic()
878 sg_assign_page(&imxdmac->sg_list[i], NULL); in imxdma_prep_dma_cyclic()
879 imxdmac->sg_list[i].offset = 0; in imxdma_prep_dma_cyclic()
886 sg_chain(imxdmac->sg_list, periods + 1, imxdmac->sg_list); in imxdma_prep_dma_cyclic()
[all …]
/linux/include/rdma/
A Drdmavt_mr.h77 struct rvt_sge *sg_list; /* next SGE to be used if any */ member
98 ss->sge = *ss->sg_list++; in rvt_put_ss()
126 *sge = *ss->sg_list++; in rvt_update_sge()
/linux/drivers/infiniband/hw/qib/
A Dqib_ud.c171 ssge.sg_list = swqe->sg_list + 1; in qib_ud_loopback()
172 ssge.sge = *swqe->sg_list; in qib_ud_loopback()
184 *sge = *ssge.sg_list++; in qib_ud_loopback()
318 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_ud_req()
319 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_ud_req()
A Dqib_uc.c99 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_uc_req()
100 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_uc_req()
419 qp->r_sge.sg_list = NULL; in qib_uc_rcv()
/linux/drivers/infiniband/hw/qedr/
A Dqedr_roce_cm.c110 qp->rqe_wr_id[qp->rq.gsi_cons].sg_list[0].length = in qedr_ll2_complete_rx_packet()
408 send_size += swr->sg_list[i].length; in qedr_gsi_build_header()
533 packet->payload[i].baddr = swr->sg_list[i].addr; in qedr_gsi_build_packet()
534 packet->payload[i].len = swr->sg_list[i].length; in qedr_gsi_build_packet()
641 wr->sg_list[0].addr, in qedr_gsi_post_recv()
642 wr->sg_list[0].length, in qedr_gsi_post_recv()
654 qp->rqe_wr_id[qp->rq.prod].sg_list[0] = wr->sg_list[0]; in qedr_gsi_post_recv()
692 wc[i].byte_len = qp->rqe_wr_id[qp->rq.cons].sg_list[0].length; in qedr_gsi_poll_cq()
/linux/drivers/scsi/aacraid/
A Dcommctrl.c494 void *sg_list[HBA_MAX_SG_EMBEDDED]; in aac_send_raw_srb() local
519 memset(sg_list, 0, sizeof(sg_list)); /* cleanup may take issue */ in aac_send_raw_srb()
553 if (user_srbcmd->sg.count > ARRAY_SIZE(sg_list)) { in aac_send_raw_srb()
681 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
742 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
797 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
854 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
895 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
935 if (copy_to_user(sg_user[i], sg_list[i], sg_count[i])) { in aac_send_raw_srb()
990 kfree(sg_list[i]); in aac_send_raw_srb()
/linux/fs/ksmbd/
A Dtransport_rdma.c201 struct scatterlist sg_list[0]; member
644 wr.sg_list = &recvmsg->sge; in smb_direct_post_recv()
1086 if (!sg_list) in get_sg_list()
1088 sg_set_page(sg_list, page, len, offset); in get_sg_list()
1089 sg_list = sg_next(sg_list); in get_sg_list()
1100 struct scatterlist *sg_list, int nentries, in get_mapped_sg_list() argument
1124 msg->wr.sg_list = &msg->sge[0]; in post_sendmsg()
1323 msg->sg_list, msg->sgt.nents, dir); in read_write_done()
1360 msg->sgt.sgl = &msg->sg_list[0]; in smb_direct_rdma_xmit()
1363 msg->sg_list, SG_CHUNK_SIZE); in smb_direct_rdma_xmit()
[all …]
/linux/drivers/infiniband/hw/bnxt_re/
A Dib_verbs.c117 total += sg_list[i].size; in bnxt_re_build_sgl()
1761 bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, wr->num_sge); in bnxt_re_post_srq_recv()
2241 wqe->sg_list[i].addr = wqe->sg_list[i - 1].addr; in bnxt_re_build_qp1_send_v2()
2242 wqe->sg_list[i].lkey = wqe->sg_list[i - 1].lkey; in bnxt_re_build_qp1_send_v2()
2243 wqe->sg_list[i].size = wqe->sg_list[i - 1].size; in bnxt_re_build_qp1_send_v2()
2505 wr->sg_list[i].addr; in bnxt_re_copy_inline_data()
2532 payload_sz = bnxt_re_build_sgl(wr->sg_list, wqe->sg_list, in bnxt_re_copy_wr_payload()
2719 bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, wr->num_sge); in bnxt_re_post_recv_shadow_qp()
2758 payload_sz = bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, in bnxt_re_post_recv()
3223 rwr.sg_list = r_sge; in bnxt_re_process_raw_qp_pkt_rx()
[all …]
/linux/drivers/net/ethernet/marvell/octeontx2/nic/
A Dotx2_txrx.h72 struct sg_list { struct
93 struct sg_list *sg; argument
A Dotx2_txrx.c96 static void otx2_dma_unmap_skb_frags(struct otx2_nic *pfvf, struct sg_list *sg) in otx2_dma_unmap_skb_frags()
112 struct sg_list *sg; in otx2_xdp_snd_pkt_handler()
135 struct sg_list *sg; in otx2_snd_pkt_handler()
730 struct sg_list *sg = &sq->sg[sqe]; in otx2_dma_map_tso_skb()
762 struct sg_list *sg = &sq->sg[sqe]; in otx2_tso_frag_dma_addr()
777 struct sg_list *list, int *offset) in otx2_sqe_tso_add_sg()
814 struct sg_list list; in otx2_sq_append_tso()
1054 struct sg_list *sg; in otx2_cleanup_tx_cqes()
/linux/drivers/infiniband/hw/cxgb4/
A Dqp.c424 if ((plen + wr->sg_list[i].length) > max) in build_immd()
427 plen += wr->sg_list[i].length; in build_immd()
428 rem = wr->sg_list[i].length; in build_immd()
468 if ((plen + sg_list[i].length) < plen) in build_isgl()
470 plen += sg_list[i].length; in build_isgl()
472 sg_list[i].length); in build_isgl()
475 *flitp = cpu_to_be64(sg_list[i].addr); in build_isgl()
537 wr->sg_list, wr->num_sge, &plen); in build_rdma_send()
588 wr->sg_list, wr->num_sge, &plen); in build_rdma_write()
665 if (wr->num_sge && wr->sg_list[0].length) { in build_rdma_read()
[all …]
/linux/drivers/infiniband/core/
A Duverbs_std_types_mr.c54 struct ib_sge *sg_list; in UVERBS_HANDLER() local
77 sg_list = uverbs_attr_get_alloced_ptr(attrs, in UVERBS_HANDLER()
79 return ib_dev->ops.advise_mr(pd, advice, flags, sg_list, num_sge, in UVERBS_HANDLER()
A Dmad.c884 mad_send_wr->sg_list[0].length = hdr_len; in ib_create_send_mad()
890 mad_send_wr->sg_list[1].length = data_len; in ib_create_send_mad()
892 mad_send_wr->sg_list[1].length = mad_size - hdr_len; in ib_create_send_mad()
899 mad_send_wr->send_wr.wr.sg_list = mad_send_wr->sg_list; in ib_create_send_mad()
1014 sge = mad_send_wr->sg_list; in ib_send_mad()
2302 mad_send_wr->sg_list[0].length, DMA_TO_DEVICE); in ib_mad_send_done()
2679 struct ib_sge sg_list; in ib_mad_post_receive_mads() local
2688 recv_wr.sg_list = &sg_list; in ib_mad_post_receive_mads()
2704 sg_list.length = mad_priv_dma_size(mad_priv); in ib_mad_post_receive_mads()
2710 sg_list.addr))) { in ib_mad_post_receive_mads()
[all …]
/linux/drivers/xen/
A Defi.c219 unsigned long count, unsigned long sg_list) in xen_efi_update_capsule() argument
229 efi_data(op).u.update_capsule.sg_list = sg_list; in xen_efi_update_capsule()
/linux/drivers/net/ethernet/ibm/ehea/
A Dehea_qmr.h106 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member
115 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES-1]; member
132 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member
/linux/arch/powerpc/include/asm/
A Dfsl_hcalls.h358 unsigned int target, phys_addr_t sg_list, unsigned int count) in fh_partition_memcpy() argument
370 r5 = (uint32_t) sg_list; in fh_partition_memcpy()
373 r6 = sg_list >> 32; in fh_partition_memcpy()
/linux/drivers/infiniband/hw/mlx5/
A Dodp.c1748 struct ib_sge *sg_list, u32 num_sge) in init_prefetch_work() argument
1758 mr = get_prefetchable_mr(pd, advice, sg_list[i].lkey); in init_prefetch_work()
1763 work->frags[i].io_virt = sg_list[i].addr; in init_prefetch_work()
1764 work->frags[i].length = sg_list[i].length; in init_prefetch_work()
1773 u32 pf_flags, struct ib_sge *sg_list, in mlx5_ib_prefetch_sg_list() argument
1783 mr = get_prefetchable_mr(pd, advice, sg_list[i].lkey); in mlx5_ib_prefetch_sg_list()
1786 ret = pagefault_mr(mr, sg_list[i].addr, sg_list[i].length, in mlx5_ib_prefetch_sg_list()
1801 u32 flags, struct ib_sge *sg_list, u32 num_sge) in mlx5_ib_advise_mr_prefetch() argument
1814 return mlx5_ib_prefetch_sg_list(pd, advice, pf_flags, sg_list, in mlx5_ib_advise_mr_prefetch()
1821 rc = init_prefetch_work(pd, advice, pf_flags, work, sg_list, num_sge); in mlx5_ib_advise_mr_prefetch()
/linux/drivers/scsi/ufs/
A Dufs_bsg.c70 sg_copy_to_buffer(job->request_payload.sg_list, in ufs_bsg_alloc_desc_buffer()
148 sg_copy_from_buffer(job->request_payload.sg_list, in ufs_bsg_request()
/linux/drivers/scsi/qla4xxx/
A Dql4_bsg.c62 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_read_flash()
122 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla4xxx_update_flash()
186 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_get_acb_state()
257 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_read_nvram()
321 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla4xxx_update_nvram()
435 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_bsg_get_acb()

Completed in 75 milliseconds

123456