| /linux/drivers/virt/ |
| A D | fsl_hypervisor.c | 153 struct fh_sg_list *sg_list = NULL; in ioctl_memcpy() local 264 sg_list[0].target = param.remote_paddr; in ioctl_memcpy() 266 sg_list[0].source = param.remote_paddr; in ioctl_memcpy() 272 count = param.count - sg_list[0].size; in ioctl_memcpy() 277 sg_list[i].source = page_to_phys(pages[i]); in ioctl_memcpy() 278 sg_list[i].target = remote_paddr; in ioctl_memcpy() 281 sg_list[i].source = remote_paddr; in ioctl_memcpy() 282 sg_list[i].target = page_to_phys(pages[i]); in ioctl_memcpy() 286 remote_paddr += sg_list[i].size; in ioctl_memcpy() 287 count -= sg_list[i].size; in ioctl_memcpy() [all …]
|
| /linux/drivers/ufs/core/ |
| A D | ufs_bsg.c | 57 sg_copy_to_buffer(job->request_payload.sg_list, in ufs_bsg_alloc_desc_buffer() 73 struct scatterlist *sg_list = NULL; in ufs_bsg_exec_advanced_rpmb_req() local 111 sg_cnt = dma_map_sg(hba->host->dma_dev, payload->sg_list, payload->sg_cnt, dir); in ufs_bsg_exec_advanced_rpmb_req() 114 sg_list = payload->sg_list; in ufs_bsg_exec_advanced_rpmb_req() 120 &rpmb_reply->ehs_rsp, sg_cnt, sg_list, dir); in ufs_bsg_exec_advanced_rpmb_req() 123 dma_unmap_sg(hba->host->dma_dev, payload->sg_list, payload->sg_cnt, dir); in ufs_bsg_exec_advanced_rpmb_req() 166 sg_copy_from_buffer(job->request_payload.sg_list, in ufs_bsg_request()
|
| /linux/drivers/scsi/qla2xxx/ |
| A D | qla_bsg.c | 53 bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free() 58 bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free() 207 bsg_job->reply_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd() 969 bsg_job->reply_payload.sg_list, in qla2x00_process_loopback() 973 bsg_job->request_payload.sg_list, in qla2x00_process_loopback() 1164 bsg_job->reply_payload.sg_list, in qla84xx_mgmt_cmd() 1214 bsg_job->request_payload.sg_list, in qla84xx_mgmt_cmd() 1896 bsg_job->request_payload.sg_list, in qla24xx_process_bidir_cmd() 1964 bsg_job->reply_payload.sg_list, in qla24xx_process_bidir_cmd() 2028 bsg_job->reply_payload.sg_list, in qlafx00_mgmt_cmd() [all …]
|
| /linux/drivers/infiniband/hw/erdma/ |
| A D | erdma_qp.c | 221 bytes += send_wr->sg_list[i].length; in fill_inline_data() 225 remain_size = send_wr->sg_list[i].length; in fill_inline_data() 232 (void *)(uintptr_t)send_wr->sg_list[i].addr + in fill_inline_data() 273 bytes += send_wr->sg_list[i].length; in fill_sgl() 274 memcpy(sgl + sgl_offset, &send_wr->sg_list[i], in fill_sgl() 361 cpu_to_le32(lower_32_bits(send_wr->sg_list[0].addr)); in erdma_push_one_sqe() 363 cpu_to_le32(upper_32_bits(send_wr->sg_list[0].addr)); in erdma_push_one_sqe() 453 sge->addr = cpu_to_le64(send_wr->sg_list[0].addr); in erdma_push_one_sqe() 454 sge->key = cpu_to_le32(send_wr->sg_list[0].lkey); in erdma_push_one_sqe() 553 rqe->stag = cpu_to_le32(recv_wr->sg_list[0].lkey); in erdma_post_recv_one() [all …]
|
| /linux/drivers/tee/amdtee/ |
| A D | call.c | 315 cmd->sg_list.count = count; in handle_map_shmem() 320 cmd->sg_list.buf[i].hi_addr = upper_32_bits(paddr); in handle_map_shmem() 321 cmd->sg_list.buf[i].low_addr = lower_32_bits(paddr); in handle_map_shmem() 322 cmd->sg_list.buf[i].size = start[i].size; in handle_map_shmem() 323 cmd->sg_list.size += cmd->sg_list.buf[i].size; in handle_map_shmem() 326 cmd->sg_list.buf[i].hi_addr); in handle_map_shmem() 328 cmd->sg_list.buf[i].low_addr); in handle_map_shmem() 329 pr_debug("buf[%d]:size = 0x%x\n", i, cmd->sg_list.buf[i].size); in handle_map_shmem() 330 pr_debug("list size = 0x%x\n", cmd->sg_list.size); in handle_map_shmem()
|
| /linux/block/ |
| A D | bsg-lib.c | 163 kfree(job->request_payload.sg_list); in bsg_teardown_job() 164 kfree(job->reply_payload.sg_list); in bsg_teardown_job() 218 buf->sg_list = kmalloc(sz, GFP_KERNEL); in bsg_map_buffer() 219 if (!buf->sg_list) in bsg_map_buffer() 221 sg_init_table(buf->sg_list, req->nr_phys_segments); in bsg_map_buffer() 222 buf->sg_cnt = blk_rq_map_sg(req->q, req, buf->sg_list); in bsg_map_buffer() 256 kfree(job->request_payload.sg_list); in bsg_prepare_job()
|
| /linux/drivers/infiniband/sw/rdmavt/ |
| A D | rc.c | 164 ss->sge = wqe->sg_list[0]; in rvt_restart_sge() 165 ss->sg_list = wqe->sg_list + 1; in rvt_restart_sge()
|
| A D | qp.c | 1829 wqe->sg_list[i].addr = wr->sg_list[i].addr; in rvt_post_recv() 1830 wqe->sg_list[i].length = wr->sg_list[i].length; in rvt_post_recv() 1831 wqe->sg_list[i].lkey = wr->sg_list[i].lkey; in rvt_post_recv() 2056 &wr->sg_list[i], acc); in rvt_post_one_wr() 2236 wqe->sg_list[i].addr = wr->sg_list[i].addr; in rvt_post_srq_recv() 2237 wqe->sg_list[i].length = wr->sg_list[i].length; in rvt_post_srq_recv() 2238 wqe->sg_list[i].lkey = wr->sg_list[i].lkey; in rvt_post_srq_recv() 2980 sqp->s_sge.sg_list = wqe->sg_list + 1; in rvt_ruc_loopback() 3048 qp->r_sge.sg_list = NULL; in rvt_ruc_loopback() 3062 sqp->s_sge.sg_list = NULL; in rvt_ruc_loopback() [all …]
|
| /linux/include/rdma/ |
| A D | rdmavt_mr.h | 77 struct rvt_sge *sg_list; /* next SGE to be used if any */ member 98 ss->sge = *ss->sg_list++; in rvt_put_ss() 126 *sge = *ss->sg_list++; in rvt_update_sge()
|
| /linux/drivers/dma/ |
| A D | imx-dma.c | 159 struct scatterlist *sg_list; member 787 kfree(imxdmac->sg_list); in imxdma_free_chan_resources() 788 imxdmac->sg_list = NULL; in imxdma_free_chan_resources() 864 kfree(imxdmac->sg_list); in imxdma_prep_dma_cyclic() 866 imxdmac->sg_list = kcalloc(periods + 1, in imxdma_prep_dma_cyclic() 868 if (!imxdmac->sg_list) in imxdma_prep_dma_cyclic() 871 sg_init_table(imxdmac->sg_list, periods); in imxdma_prep_dma_cyclic() 874 sg_assign_page(&imxdmac->sg_list[i], NULL); in imxdma_prep_dma_cyclic() 875 imxdmac->sg_list[i].offset = 0; in imxdma_prep_dma_cyclic() 882 sg_chain(imxdmac->sg_list, periods + 1, imxdmac->sg_list); in imxdma_prep_dma_cyclic() [all …]
|
| /linux/drivers/infiniband/hw/qib/ |
| A D | qib_ud.c | 171 ssge.sg_list = swqe->sg_list + 1; in qib_ud_loopback() 172 ssge.sge = *swqe->sg_list; in qib_ud_loopback() 184 *sge = *ssge.sg_list++; in qib_ud_loopback() 318 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_ud_req() 319 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_ud_req()
|
| A D | qib_uc.c | 99 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_uc_req() 100 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_uc_req() 419 qp->r_sge.sg_list = NULL; in qib_uc_rcv()
|
| /linux/drivers/infiniband/hw/qedr/ |
| A D | qedr_roce_cm.c | 110 qp->rqe_wr_id[qp->rq.gsi_cons].sg_list[0].length = in qedr_ll2_complete_rx_packet() 407 send_size += swr->sg_list[i].length; in qedr_gsi_build_header() 532 packet->payload[i].baddr = swr->sg_list[i].addr; in qedr_gsi_build_packet() 533 packet->payload[i].len = swr->sg_list[i].length; in qedr_gsi_build_packet() 640 wr->sg_list[0].addr, in qedr_gsi_post_recv() 641 wr->sg_list[0].length, in qedr_gsi_post_recv() 653 qp->rqe_wr_id[qp->rq.prod].sg_list[0] = wr->sg_list[0]; in qedr_gsi_post_recv() 691 wc[i].byte_len = qp->rqe_wr_id[qp->rq.cons].sg_list[0].length; in qedr_gsi_poll_cq()
|
| /linux/drivers/scsi/aacraid/ |
| A D | commctrl.c | 494 void *sg_list[HBA_MAX_SG_EMBEDDED]; in aac_send_raw_srb() local 519 memset(sg_list, 0, sizeof(sg_list)); /* cleanup may take issue */ in aac_send_raw_srb() 553 if (user_srbcmd->sg.count > ARRAY_SIZE(sg_list)) { in aac_send_raw_srb() 681 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 742 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 797 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 854 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 895 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 935 if (copy_to_user(sg_user[i], sg_list[i], sg_count[i])) { in aac_send_raw_srb() 990 kfree(sg_list[i]); in aac_send_raw_srb()
|
| /linux/drivers/net/ethernet/marvell/octeon_ep/ |
| A D | octep_ctrl_net.c | 51 msg->sg_list[0].msg = buf; in init_send_req() 52 msg->sg_list[0].sz = msg->hdr.s.sz; in init_send_req() 309 msg->sg_list[0].msg, in process_mbox_resp() 325 req = (struct octep_ctrl_net_f2h_req *)msg->sg_list[0].msg; in process_mbox_notify() 367 msg.sg_list[0].sz = msg_sz; in octep_ctrl_net_recv_fw_messages() 368 msg.sg_list[0].msg = &data; in octep_ctrl_net_recv_fw_messages()
|
| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| A D | otx2_txrx.h | 77 struct sg_list { struct 99 struct sg_list *sg; argument
|
| /linux/drivers/firmware/efi/ |
| A D | runtime-wrappers.c | 102 unsigned long sg_list; member 283 args->UPDATE_CAPSULE.sg_list); in efi_call_rts() 523 unsigned long sg_list) in virt_efi_update_capsule() argument 532 status = efi_queue_work(UPDATE_CAPSULE, capsules, count, sg_list); in virt_efi_update_capsule()
|
| /linux/drivers/infiniband/hw/cxgb4/ |
| A D | qp.c | 424 if ((plen + wr->sg_list[i].length) > max) in build_immd() 427 plen += wr->sg_list[i].length; in build_immd() 428 rem = wr->sg_list[i].length; in build_immd() 468 if ((plen + sg_list[i].length) < plen) in build_isgl() 470 plen += sg_list[i].length; in build_isgl() 472 sg_list[i].length); in build_isgl() 475 *flitp = cpu_to_be64(sg_list[i].addr); in build_isgl() 537 wr->sg_list, wr->num_sge, &plen); in build_rdma_send() 588 wr->sg_list, wr->num_sge, &plen); in build_rdma_write() 665 if (wr->num_sge && wr->sg_list[0].length) { in build_rdma_read() [all …]
|
| /linux/drivers/net/ethernet/ibm/ehea/ |
| A D | ehea_qmr.h | 106 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member 115 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES-1]; member 132 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member
|
| /linux/drivers/infiniband/core/ |
| A D | uverbs_std_types_mr.c | 54 struct ib_sge *sg_list; in UVERBS_HANDLER() local 77 sg_list = uverbs_attr_get_alloced_ptr(attrs, in UVERBS_HANDLER() 79 return ib_dev->ops.advise_mr(pd, advice, flags, sg_list, num_sge, in UVERBS_HANDLER()
|
| /linux/arch/powerpc/include/asm/ |
| A D | fsl_hcalls.h | 358 unsigned int target, phys_addr_t sg_list, unsigned int count) in fh_partition_memcpy() argument 370 r5 = (uint32_t) sg_list; in fh_partition_memcpy() 373 r6 = sg_list >> 32; in fh_partition_memcpy()
|
| /linux/drivers/infiniband/hw/bnxt_re/ |
| A D | ib_verbs.c | 135 total += sg_list[i].size; in bnxt_re_build_sgl() 1938 bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, wr->num_sge); in bnxt_re_post_srq_recv() 2422 wqe->sg_list[i].addr = wqe->sg_list[i - 1].addr; in bnxt_re_build_qp1_send_v2() 2423 wqe->sg_list[i].lkey = wqe->sg_list[i - 1].lkey; in bnxt_re_build_qp1_send_v2() 2424 wqe->sg_list[i].size = wqe->sg_list[i - 1].size; in bnxt_re_build_qp1_send_v2() 2675 wr->sg_list[i].addr; in bnxt_re_copy_inline_data() 2702 payload_sz = bnxt_re_build_sgl(wr->sg_list, wqe->sg_list, in bnxt_re_copy_wr_payload() 2903 bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, wr->num_sge); in bnxt_re_post_recv_shadow_qp() 2942 payload_sz = bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, in bnxt_re_post_recv() 3526 rwr.sg_list = r_sge; in bnxt_re_process_raw_qp_pkt_rx() [all …]
|
| /linux/include/uapi/rdma/ |
| A D | rvt-abi.h | 47 struct rvt_wqe_sge sg_list[]; member
|
| /linux/fs/smb/server/ |
| A D | transport_rdma.c | 204 struct scatterlist sg_list[]; member 657 wr.sg_list = &recvmsg->sge; in smb_direct_post_recv() 1098 if (!sg_list) in get_sg_list() 1100 sg_set_page(sg_list, page, len, offset); in get_sg_list() 1101 sg_list = sg_next(sg_list); in get_sg_list() 1112 struct scatterlist *sg_list, int nentries, in get_mapped_sg_list() argument 1117 npages = get_sg_list(buf, size, sg_list, nentries); in get_mapped_sg_list() 1120 return ib_dma_map_sg(device, sg_list, npages, dir); in get_mapped_sg_list() 1136 msg->wr.sg_list = &msg->sge[0]; in post_sendmsg() 1421 msg->sgt.sgl = &msg->sg_list[0]; in smb_direct_rdma_xmit() [all …]
|
| /linux/include/linux/ |
| A D | bsg-lib.h | 26 struct scatterlist *sg_list; member
|