| /drivers/dma/sh/ |
| A D | usb-dmac.c | 55 unsigned int sg_len; member 269 desc = kzalloc(struct_size(desc, sg, sg_len), gfp); in usb_dmac_desc_alloc() 273 desc->sg_allocated_len = sg_len; in usb_dmac_desc_alloc() 299 unsigned int sg_len, gfp_t gfp) in usb_dmac_desc_get() argument 307 if (sg_len <= desc->sg_allocated_len) { in usb_dmac_desc_get() 316 if (!usb_dmac_desc_alloc(chan, sg_len, gfp)) { in usb_dmac_desc_get() 425 if (!sg_len) { in usb_dmac_prep_slave_sg() 436 desc->sg_len = sg_len; in usb_dmac_prep_slave_sg() 437 for_each_sg(sgl, sg, sg_len, i) { in usb_dmac_prep_slave_sg() 519 for (i = desc->sg_index + 1; i < desc->sg_len; i++) in usb_dmac_chan_get_residue() [all …]
|
| A D | shdma-base.c | 571 for_each_sg(sgl, sg, sg_len, i) in shdma_prep_sg() 588 for_each_sg(sgl, sg, sg_len, i) { in shdma_prep_sg() 677 if (slave_id < 0 || !sg_len) { in shdma_prep_slave_sg() 679 __func__, sg_len, slave_id); in shdma_prep_slave_sg() 685 return shdma_prep_sg(schan, sgl, sg_len, &slave_addr, in shdma_prep_slave_sg() 700 unsigned int sg_len = buf_len / period_len; in shdma_prep_dma_cyclic() local 711 if (sg_len > SHDMA_MAX_SG_LEN) { in shdma_prep_dma_cyclic() 713 sg_len, SHDMA_MAX_SG_LEN); in shdma_prep_dma_cyclic() 735 sg_init_table(sgl, sg_len); in shdma_prep_dma_cyclic() 737 for (i = 0; i < sg_len; i++) { in shdma_prep_dma_cyclic() [all …]
|
| A D | rcar-dmac.c | 929 unsigned int sg_len, dma_addr_t dev_addr, in rcar_dmac_chan_prep_sg() argument 964 for_each_sg(sgl, sg, sg_len, i) { in rcar_dmac_chan_prep_sg() 1201 unsigned int sg_len, enum dma_transfer_direction dir, in rcar_dmac_prep_slave_sg() argument 1210 __func__, sg_len, rchan->mid_rid); in rcar_dmac_prep_slave_sg() 1231 unsigned int sg_len; in rcar_dmac_prep_dma_cyclic() local 1245 sg_len = buf_len / period_len; in rcar_dmac_prep_dma_cyclic() 1246 if (sg_len > RCAR_DMAC_MAX_SG_LEN) { in rcar_dmac_prep_dma_cyclic() 1249 rchan->index, sg_len, RCAR_DMAC_MAX_SG_LEN); in rcar_dmac_prep_dma_cyclic() 1257 sgl = kmalloc_array(sg_len, sizeof(*sgl), GFP_NOWAIT); in rcar_dmac_prep_dma_cyclic() 1261 sg_init_table(sgl, sg_len); in rcar_dmac_prep_dma_cyclic() [all …]
|
| /drivers/mmc/core/ |
| A D | mmc_test.c | 82 unsigned int sg_len; member 242 mrq->data->sg_len = sg_len; in mmc_test_prepare_mrq() 421 *sg_len = 0; in mmc_test_map_sg() 440 *sg_len += 1; in mmc_test_map_sg() 473 *sg_len = 0; in mmc_test_map_sg_max_scatter() 495 *sg_len += 1; in mmc_test_map_sg_max_scatter() 1412 if (!err && sg_len != t->sg_len) in mmc_test_area_map() 2262 .sg_len = sg_len, in mmc_test_profile_sglen_wr_blocking_perf() 2281 .sg_len = sg_len, in mmc_test_profile_sglen_wr_nonblock_perf() 2300 .sg_len = sg_len, in mmc_test_profile_sglen_r_blocking_perf() [all …]
|
| A D | sdio_ops.c | 158 data.sg_len = nents; in mmc_io_rw_extended() 160 for_each_sg(data.sg, sg_ptr, data.sg_len, i) { in mmc_io_rw_extended() 167 data.sg_len = 1; in mmc_io_rw_extended()
|
| /drivers/mmc/host/ |
| A D | uniphier-sd.c | 109 dma_unmap_sg(mmc_dev(host->mmc), host->sg_ptr, host->sg_len, in uniphier_sd_external_dma_callback() 138 int sg_len; in uniphier_sd_external_dma_start() local 151 sg_len = dma_map_sg(mmc_dev(host->mmc), host->sg_ptr, host->sg_len, in uniphier_sd_external_dma_start() 153 if (sg_len == 0) in uniphier_sd_external_dma_start() 156 desc = dmaengine_prep_slave_sg(priv->chan, host->sg_ptr, sg_len, in uniphier_sd_external_dma_start() 173 dma_unmap_sg(mmc_dev(host->mmc), host->sg_ptr, host->sg_len, in uniphier_sd_external_dma_start() 260 int sg_len; in uniphier_sd_internal_dma_start() local 265 if (WARN_ON(host->sg_len != 1)) in uniphier_sd_internal_dma_start() 279 sg_len = dma_map_sg(mmc_dev(host->mmc), sg, 1, priv->dma_dir); in uniphier_sd_internal_dma_start() 280 if (sg_len == 0) in uniphier_sd_internal_dma_start()
|
| A D | au1xmmc.c | 380 int sg_len, max, count; in au1xmmc_send_pio() local 398 max = (sg_len > host->pio.len) ? host->pio.len : sg_len; in au1xmmc_send_pio() 418 if (count == sg_len) { in au1xmmc_send_pio() 436 int max, count, sg_len = 0; in au1xmmc_receive_pio() local 456 if (sg_len < max) in au1xmmc_receive_pio() 457 max = sg_len; in au1xmmc_receive_pio() 497 if (sg_len && count == sg_len) { in au1xmmc_receive_pio() 618 data->sg_len, host->dma.dir); in au1xmmc_prepare_data() 634 int sg_len = sg->length; in au1xmmc_prepare_data() local 636 int len = (datalen > sg_len) ? sg_len : datalen; in au1xmmc_prepare_data() [all …]
|
| A D | mxs-mmc.c | 147 data->sg_len, ssp->dma_dir); in mxs_mmc_request_done() 220 unsigned int sg_len; in mxs_mmc_prep_dma() local 225 data->sg_len, ssp->dma_dir); in mxs_mmc_prep_dma() 227 sg_len = data->sg_len; in mxs_mmc_prep_dma() 231 sg_len = SSP_PIO_NUM; in mxs_mmc_prep_dma() 235 sgl, sg_len, ssp->slave_dirn, flags); in mxs_mmc_prep_dma() 242 data->sg_len, ssp->dma_dir); in mxs_mmc_prep_dma() 349 unsigned int sg_len = data->sg_len; in mxs_mmc_adtc() local 394 for_each_sg(sgl, sg, sg_len, i) in mxs_mmc_adtc()
|
| A D | renesas_sdhi_sys_dmac.c | 139 host->sg_ptr, host->sg_len, in renesas_sdhi_sys_dmac_dma_callback() 143 host->sg_ptr, host->sg_len, in renesas_sdhi_sys_dmac_dma_callback() 167 for_each_sg(sg, sg_tmp, host->sg_len, i) { in renesas_sdhi_sys_dmac_start_dma_rx() 176 if ((!aligned && (host->sg_len > 1 || sg->length > PAGE_SIZE || in renesas_sdhi_sys_dmac_start_dma_rx() 192 ret = dma_map_sg(chan->device->dev, sg, host->sg_len, DMA_FROM_DEVICE); in renesas_sdhi_sys_dmac_start_dma_rx() 239 for_each_sg(sg, sg_tmp, host->sg_len, i) { in renesas_sdhi_sys_dmac_start_dma_tx() 248 if ((!aligned && (host->sg_len > 1 || sg->length > PAGE_SIZE || in renesas_sdhi_sys_dmac_start_dma_tx() 268 ret = dma_map_sg(chan->device->dev, sg, host->sg_len, DMA_TO_DEVICE); in renesas_sdhi_sys_dmac_start_dma_tx()
|
| A D | mmci_stm32_sdmmc.c | 94 for_each_sg(data->sg, sg, data->sg_len - 1, i) { in sdmmc_idma_validate_data() 140 sg_copy_to_buffer(data->sg, data->sg_len, in _sdmmc_idma_prep_data() 149 data->sg_len, in _sdmmc_idma_prep_data() 179 sg_copy_from_buffer(data->sg, data->sg_len, in sdmmc_idma_unprep_data() 183 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, in sdmmc_idma_unprep_data() 231 if (!host->variant->dma_lli || data->sg_len == 1 || in sdmmc_idma_start() 247 for_each_sg(data->sg, sg, data->sg_len, i) { in sdmmc_idma_start() 256 desc[data->sg_len - 1].idmalar &= ~MMCI_STM32_ULA; in sdmmc_idma_start() 282 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, in sdmmc_idma_error()
|
| A D | tifm_sd.c | 105 int sg_len; member 174 if (host->sg_pos == host->sg_len) in tifm_sd_transfer_data() 181 if (host->sg_pos == host->sg_len) { in tifm_sd_transfer_data() 237 if (host->sg_pos == host->sg_len) in tifm_sd_bounce_block() 269 if (host->sg_pos == host->sg_len) in tifm_sd_set_dma_data() 275 if (host->sg_pos == host->sg_len) in tifm_sd_set_dma_data() 283 if (host->sg_pos == host->sg_len) in tifm_sd_set_dma_data() 658 host->sg_len = r_data->sg_len; in tifm_sd_request() 672 host->sg_len = tifm_map_sg(sock, r_data->sg, in tifm_sd_request() 673 r_data->sg_len, in tifm_sd_request() [all …]
|
| A D | omap.c | 153 unsigned int sg_len; member 436 dma_unmap_sg(dev, data->sg, host->sg_len, dma_data_dir); in mmc_omap_release_dma() 462 host->sg_len = 0; in mmc_omap_xfer_done() 519 host->sg_len = 0; in mmc_omap_abort_xfer() 956 unsigned sg_len; in mmc_omap_prepare_data() local 977 sg_len = (data->blocks == 1) ? 1 : data->sg_len; in mmc_omap_prepare_data() 980 for_each_sg(data->sg, sg, sg_len, i) { in mmc_omap_prepare_data() 1040 host->sg_len = dma_map_sg(c->device->dev, data->sg, sg_len, in mmc_omap_prepare_data() 1042 if (host->sg_len == 0) in mmc_omap_prepare_data() 1045 tx = dmaengine_prep_slave_sg(c, data->sg, host->sg_len, in mmc_omap_prepare_data() [all …]
|
| A D | cavium.c | 377 dma_unmap_sg(host->dev, data->sg, data->sg_len, get_dma_dir(data)); in finish_dma_single() 397 dma_unmap_sg(host->dev, data->sg, data->sg_len, get_dma_dir(data)); in finish_dma_sg() 403 if (host->use_sg && data->sg_len > 1) in finish_dma() 521 count = dma_map_sg(host->dev, data->sg, data->sg_len, in prepare_dma_single() 558 count = dma_map_sg(host->dev, data->sg, data->sg_len, in prepare_dma_sg() 611 dma_unmap_sg(host->dev, data->sg, data->sg_len, get_dma_dir(data)); in prepare_dma_sg() 619 if (host->use_sg && data->sg_len > 1) in prepare_dma() 657 if (!mrq->data || !mrq->data->sg || !mrq->data->sg_len || in cvm_mmc_dma_request() 709 sg_miter_start(&host->smi, mrq->data->sg, mrq->data->sg_len, in do_read_request() 722 sg_miter_start(smi, mrq->data->sg, mrq->data->sg_len, SG_MITER_FROM_SG); in do_write_request()
|
| A D | davinci_mmc.c | 193 unsigned int sg_len; member 409 host->sg_len, in mmc_davinci_send_dma_request() 431 host->sg_len, in mmc_davinci_send_dma_request() 456 host->sg_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, in mmc_davinci_start_dma_transfer() 460 for (i = 0; i < host->sg_len; i++) { in mmc_davinci_start_dma_transfer() 463 data->sg, data->sg_len, in mmc_davinci_start_dma_transfer() 570 host->sg_len = data->sg_len; in mmc_davinci_prepare_data() 571 sg_miter_start(&host->sg_miter, data->sg, data->sg_len, flags); in mmc_davinci_prepare_data() 776 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, in mmc_davinci_xfer_done()
|
| A D | dw_mmc.c | 442 data->sg_len, in dw_mci_dma_cleanup() 485 data->sg_len, in dw_mci_dmac_complete_dma() 883 return data->sg_len; in dw_mci_pre_dma_transfer() 903 data->sg_len, in dw_mci_pre_dma_transfer() 905 if (sg_len == 0) in dw_mci_pre_dma_transfer() 910 return sg_len; in dw_mci_pre_dma_transfer() 943 data->sg_len, in dw_mci_post_req() 1084 int sg_len; in dw_mci_submit_data_dma() local 1094 if (sg_len < 0) { in dw_mci_submit_data_dma() 1096 return sg_len; in dw_mci_submit_data_dma() [all …]
|
| A D | sh_mmcif.c | 294 ret = dma_map_sg(chan->device->dev, sg, data->sg_len, in sh_mmcif_start_dma_rx() 310 __func__, data->sg_len, ret, cookie); in sh_mmcif_start_dma_rx() 331 desc, cookie, data->sg_len); in sh_mmcif_start_dma_rx() 344 ret = dma_map_sg(chan->device->dev, sg, data->sg_len, in sh_mmcif_start_dma_tx() 360 __func__, data->sg_len, ret, cookie); in sh_mmcif_start_dma_tx() 660 if (!data->sg_len || !data->sg->length) in sh_mmcif_multi_read() 666 sg_miter_start(sgm, data->sg, data->sg_len, in sh_mmcif_multi_read() 770 if (!data->sg_len || !data->sg->length) in sh_mmcif_multi_write() 776 sg_miter_start(sgm, data->sg, data->sg_len, in sh_mmcif_multi_write() 1179 data->sg, data->sg_len, in sh_mmcif_end_cmd() [all …]
|
| /drivers/dma/ |
| A D | loongson1-apb-dma.c | 217 struct scatterlist *sgl, unsigned int sg_len, in ls1x_dma_prep_lli() argument 244 for_each_sg(sgl, sg, sg_len, i) { in ls1x_dma_prep_lli() 297 unsigned int sg_len, enum dma_transfer_direction dir, in ls1x_dma_prep_slave_sg() argument 303 sg_len, flags, dmaengine_get_direction_text(dir)); in ls1x_dma_prep_slave_sg() 309 if (ls1x_dma_prep_lli(dchan, desc, sgl, sg_len, dir, false)) { in ls1x_dma_prep_slave_sg() 324 unsigned int sg_len; in ls1x_dma_prep_dma_cyclic() local 337 sg_len = buf_len / period_len; in ls1x_dma_prep_dma_cyclic() 338 sgl = kmalloc_array(sg_len, sizeof(*sgl), GFP_NOWAIT); in ls1x_dma_prep_dma_cyclic() 342 sg_init_table(sgl, sg_len); in ls1x_dma_prep_dma_cyclic() 343 for (i = 0; i < sg_len; ++i) { in ls1x_dma_prep_dma_cyclic() [all …]
|
| A D | st_fdma.c | 239 int sg_len) in st_fdma_alloc_desc() argument 249 fdesc->n_nodes = sg_len; in st_fdma_alloc_desc() 250 for (i = 0; i < sg_len; i++) { in st_fdma_alloc_desc() 455 int sg_len, i; in st_fdma_prep_dma_cyclic() local 475 sg_len = len / period_len; in st_fdma_prep_dma_cyclic() 476 fdesc = st_fdma_alloc_desc(fchan, sg_len); in st_fdma_prep_dma_cyclic() 484 for (i = 0; i < sg_len; i++) { in st_fdma_prep_dma_cyclic() 487 hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc; in st_fdma_prep_dma_cyclic() 518 fchan = st_fdma_prep_common(chan, sg_len, direction); in st_fdma_prep_slave_sg() 525 fdesc = st_fdma_alloc_desc(fchan, sg_len); in st_fdma_prep_slave_sg() [all …]
|
| A D | milbeaut-hdmac.c | 62 unsigned int sg_len; member 193 if (md->sg_cur >= md->sg_len) { in milbeaut_hdmac_interrupt() 257 unsigned int sg_len, in milbeaut_hdmac_prep_slave_sg() argument 272 md->sgl = kcalloc(sg_len, sizeof(*sgl), GFP_NOWAIT); in milbeaut_hdmac_prep_slave_sg() 278 for (i = 0; i < sg_len; i++) in milbeaut_hdmac_prep_slave_sg() 281 md->sg_len = sg_len; in milbeaut_hdmac_prep_slave_sg() 371 for (i = md->sg_cur; i < md->sg_len; i++) in milbeaut_hdmac_tx_status()
|
| A D | fsl-edma-common.c | 559 int sg_len) in fsl_edma_alloc_desc() argument 569 fsl_desc->n_tcds = sg_len; in fsl_edma_alloc_desc() 570 for (i = 0; i < sg_len; i++) { in fsl_edma_alloc_desc() 595 int sg_len, i; in fsl_edma_prep_dma_cyclic() local 606 sg_len = buf_len / period_len; in fsl_edma_prep_dma_cyclic() 607 fsl_desc = fsl_edma_alloc_desc(fsl_chan, sg_len); in fsl_edma_prep_dma_cyclic() 628 for (i = 0; i < sg_len; i++) { in fsl_edma_prep_dma_cyclic() 633 last_sg = fsl_desc->tcd[(i + 1) % sg_len].ptcd; in fsl_edma_prep_dma_cyclic() 685 fsl_desc = fsl_edma_alloc_desc(fsl_chan, sg_len); in fsl_edma_prep_slave_sg() 703 for_each_sg(sgl, sg, sg_len, i) { in fsl_edma_prep_slave_sg() [all …]
|
| A D | uniphier-mdmac.c | 54 unsigned int sg_len; member 209 if (md->sg_cur >= md->sg_len) { in uniphier_mdmac_interrupt() 231 unsigned int sg_len, in uniphier_mdmac_prep_slave_sg() argument 246 md->sg_len = sg_len; in uniphier_mdmac_prep_slave_sg() 319 for (i = md->sg_cur; i < md->sg_len; i++) in uniphier_mdmac_tx_status()
|
| A D | ste_dma40_ll.c | 267 int sg_len, in d40_phy_sg_to_lli() argument 285 for_each_sg(sg, current_sg, sg_len, i) { in d40_phy_sg_to_lli() 292 if (i == sg_len - 1) in d40_phy_sg_to_lli() 419 int sg_len, in d40_log_sg_to_lli() argument 434 for_each_sg(sg, current_sg, sg_len, i) { in d40_log_sg_to_lli()
|
| A D | at_hdmac.c | 979 size_t sg_len; in atc_prep_dma_memcpy() local 998 desc->sglen = sg_len; in atc_prep_dma_memcpy() 1178 value, sg_len, flags); in atc_prep_dma_memset_sg() 1180 if (unlikely(!sgl || !sg_len)) { in atc_prep_dma_memset_sg() 1197 desc->sglen = sg_len; in atc_prep_dma_memset_sg() 1199 for_each_sg(sgl, sg, sg_len, i) { in atc_prep_dma_memset_sg() 1268 sg_len, in atc_prep_slave_sg() 1272 if (unlikely(!atslave || !sg_len)) { in atc_prep_slave_sg() 1280 desc->sglen = sg_len; in atc_prep_slave_sg() 1297 for_each_sg(sgl, sg, sg_len, i) { in atc_prep_slave_sg() [all …]
|
| /drivers/crypto/nx/ |
| A D | nx.c | 83 unsigned int sg_len = 0; in nx_build_sg_list() local 108 for (sg = sg_head; sg_len < *len; sg++) { in nx_build_sg_list() 117 sg_len += sg->len; in nx_build_sg_list() 120 is_vmalloc_addr(start_addr + sg_len)) { in nx_build_sg_list() 122 start_addr + sg_len)); in nx_build_sg_list() 123 end_addr = sg_addr + *len - sg_len; in nx_build_sg_list() 131 *len = sg_len; in nx_build_sg_list()
|
| /drivers/crypto/chelsio/ |
| A D | chcr_algo.c | 1709 if (params.sg_len > req->nbytes) in chcr_ahash_update() 1710 params.sg_len = req->nbytes; in chcr_ahash_update() 1711 params.sg_len = rounddown(params.sg_len + req_ctx->reqlen, bs) - in chcr_ahash_update() 1785 params.sg_len = 0; in chcr_ahash_final() 1882 params.sg_len = rounddown(params.sg_len + req_ctx->reqlen, bs) in chcr_ahash_finup() 1889 params.sg_len = req->nbytes; in chcr_ahash_finup() 1892 params.sg_len; in chcr_ahash_finup() 1986 params.sg_len = rounddown(params.sg_len, bs); in chcr_ahash_digest() 1989 params.sg_len = req->nbytes; in chcr_ahash_digest() 2064 params.sg_len = rounddown(params.sg_len, bs); in chcr_ahash_continue() [all …]
|