| /drivers/parisc/ |
| A D | iommu-helpers.h | 22 unsigned long dma_offset = 0, dma_len = 0; in iommu_fill_pdir() local 47 BUG_ON(pdirp && (dma_len != sg_dma_len(dma_sg))); in iommu_fill_pdir() 51 dma_len = sg_dma_len(startsg); in iommu_fill_pdir() 106 unsigned long dma_offset, dma_len; /* start/len of DMA stream */ in iommu_coalesce_chunks() local 120 dma_len = startsg->length; in iommu_coalesce_chunks() 149 if (unlikely(ALIGN(dma_len + dma_offset + startsg->length, IOVP_SIZE) > in iommu_coalesce_chunks() 163 dma_len += startsg->length; in iommu_coalesce_chunks() 171 sg_dma_len(contig_sg) = dma_len; in iommu_coalesce_chunks() 172 dma_len = ALIGN(dma_len + dma_offset, IOVP_SIZE); in iommu_coalesce_chunks() 175 | (iommu_alloc_range(ioc, dev, dma_len) << IOVP_SHIFT) in iommu_coalesce_chunks()
|
| /drivers/net/ethernet/netronome/nfp/nfdk/ |
| A D | dp.c | 324 dma_len -= 1; in nfp_nfdk_tx() 347 dma_len -= tmp_dlen; in nfp_nfdk_tx() 358 while (dma_len > 0) { in nfp_nfdk_tx() 359 dma_len -= 1; in nfp_nfdk_tx() 964 dma_len = pkt_len; in nfp_nfdk_tx_xdp_buf() 973 dma_len -= 1; in nfp_nfdk_tx_xdp_buf() 983 dma_len -= tmp_dlen; in nfp_nfdk_tx_xdp_buf() 988 dma_len -= 1; in nfp_nfdk_tx_xdp_buf() 1387 dma_len -= 1; in nfp_nfdk_ctrl_tx_one() 1397 dma_len -= tmp_dlen; in nfp_nfdk_ctrl_tx_one() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/en/ |
| A D | xdp.h | 201 u32 dma_len = xdptxd->len; in mlx5e_xdp_mpwqe_add_dseg() local 204 session->bytes_count += dma_len; in mlx5e_xdp_mpwqe_add_dseg() 206 if (session->inline_on && dma_len <= MLX5E_XDP_INLINE_WQE_SZ_THRSD) { in mlx5e_xdp_mpwqe_add_dseg() 209 u16 ds_len = sizeof(*inline_dseg) + dma_len; in mlx5e_xdp_mpwqe_add_dseg() 212 inline_dseg->byte_count = cpu_to_be32(dma_len | MLX5_INLINE_SEG); in mlx5e_xdp_mpwqe_add_dseg() 213 memcpy(inline_dseg->data, xdptxd->data, dma_len); in mlx5e_xdp_mpwqe_add_dseg() 221 dseg->byte_count = cpu_to_be32(dma_len); in mlx5e_xdp_mpwqe_add_dseg()
|
| /drivers/net/wireless/broadcom/brcm80211/brcmsmac/ |
| A D | brcms_trace_brcmsmac_tx.h | 76 u16 dma_len), 78 dma_len), 85 __field(u16, dma_len) 93 __entry->dma_len = dma_len; 98 __entry->ampdu_frames, __entry->dma_len)
|
| A D | ampdu.c | 477 session->dma_len = 0; in brcms_c_ampdu_reset_session() 536 session->dma_len += p->len; in brcms_c_ampdu_add_frame() 612 u16 dma_len = session->dma_len; in brcms_c_ampdu_finalize() local 736 dma_len = min(dma_len, f->ampdu_pld_size); in brcms_c_ampdu_finalize() 737 txh->PreloadSize = cpu_to_le16(dma_len); in brcms_c_ampdu_finalize()
|
| /drivers/i2c/busses/ |
| A D | i2c-stm32.c | 88 dma->dma_len = 0; in stm32_i2c_dma_free() 117 dma->dma_len = len; in stm32_i2c_prep_dma_xfer() 119 dma->dma_buf = dma_map_single(dev, buf, dma->dma_len, in stm32_i2c_prep_dma_xfer() 127 dma->dma_len, in stm32_i2c_prep_dma_xfer() 151 dma_unmap_single(dev, dma->dma_buf, dma->dma_len, in stm32_i2c_prep_dma_xfer()
|
| A D | i2c-imx-lpi2c.c | 156 unsigned int dma_len; member 634 time = 8 * lpi2c_imx->dma->dma_len * 1000 / lpi2c_imx->bitrate; in lpi2c_imx_dma_timeout_calculate() 646 u16 rx_remain = dma->dma_len; in lpi2c_imx_alloc_rx_cmd_buf() 694 dma->dma_len, dma->dma_data_dir); in lpi2c_dma_unmap() 786 dma->dma_buf, dma->dma_len, dma->dma_data_dir); in lpi2c_dma_submit() 793 dma->dma_len, dma->dma_transfer_dir, in lpi2c_dma_submit() 853 cmd_num = DIV_ROUND_UP(dma->dma_len, CHUNK_DATA); in lpi2c_imx_dma_burst_num_calculate() 857 dma->dma_len); in lpi2c_imx_dma_burst_num_calculate() 860 dma->dma_len); in lpi2c_imx_dma_burst_num_calculate() 957 dma->dma_len = msg->len; in lpi2c_imx_dma_xfer()
|
| /drivers/net/ethernet/sfc/ |
| A D | tx_tso.c | 107 unsigned int dma_len; in efx_tx_queue_insert() local 121 dma_len = tx_queue->efx->type->tx_limit_len(tx_queue, in efx_tx_queue_insert() 125 if (dma_len >= len) in efx_tx_queue_insert() 128 buffer->len = dma_len; in efx_tx_queue_insert() 130 dma_addr += dma_len; in efx_tx_queue_insert() 131 len -= dma_len; in efx_tx_queue_insert()
|
| A D | tx_common.c | 333 unsigned int dma_len; in efx_tx_map_chunk() local 340 dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len); in efx_tx_map_chunk() 342 dma_len = len; in efx_tx_map_chunk() 344 buffer->len = dma_len; in efx_tx_map_chunk() 347 len -= dma_len; in efx_tx_map_chunk() 348 dma_addr += dma_len; in efx_tx_map_chunk()
|
| /drivers/net/ethernet/sfc/siena/ |
| A D | tx_common.c | 293 unsigned int dma_len; in efx_siena_tx_map_chunk() local 300 dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len); in efx_siena_tx_map_chunk() 302 dma_len = len; in efx_siena_tx_map_chunk() 304 buffer->len = dma_len; in efx_siena_tx_map_chunk() 307 len -= dma_len; in efx_siena_tx_map_chunk() 308 dma_addr += dma_len; in efx_siena_tx_map_chunk()
|
| /drivers/scsi/ |
| A D | am53c974.c | 243 static u32 pci_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) in pci_esp_dma_length_limit() argument 255 if (dma_len > (1U << dma_limit)) in pci_esp_dma_length_limit() 256 dma_len = (1U << dma_limit); in pci_esp_dma_length_limit() 262 end = base + dma_len; in pci_esp_dma_length_limit() 265 dma_len = end - base; in pci_esp_dma_length_limit() 267 return dma_len; in pci_esp_dma_length_limit()
|
| A D | zorro_esp.c | 219 u32 dma_len) in zorro_esp_dma_length_limit() argument 221 return dma_len > (1U << 16) ? (1U << 16) : dma_len; in zorro_esp_dma_length_limit() 225 u32 dma_len) in fastlane_esp_dma_length_limit() argument 228 return dma_len > 0xfffc ? 0xfffc : dma_len; in fastlane_esp_dma_length_limit()
|
| A D | mac53c94.c | 364 u32 dma_len; in set_dma_cmds() local 378 dma_len = sg_dma_len(scl); in set_dma_cmds() 379 if (dma_len > 0xffff) in set_dma_cmds() 381 total += dma_len; in set_dma_cmds() 382 dcmds->req_count = cpu_to_le16(dma_len); in set_dma_cmds()
|
| A D | esp_scsi.c | 526 if (dma_len > (1U << 24)) in esp_dma_length_limit() 527 dma_len = (1U << 24); in esp_dma_length_limit() 537 if (dma_len > (1U << 16)) in esp_dma_length_limit() 538 dma_len = (1U << 16); in esp_dma_length_limit() 544 end = base + dma_len; in esp_dma_length_limit() 547 dma_len = end - base; in esp_dma_length_limit() 549 return dma_len; in esp_dma_length_limit() 1754 dma_len); in esp_process_event() 1756 dma_len = esp_dma_length_limit(esp, dma_addr, dma_len); in esp_process_event() 1760 if (!dma_len) { in esp_process_event() [all …]
|
| A D | mac_esp.c | 247 static u32 mac_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) in mac_esp_dma_length_limit() argument 249 return dma_len > 0xFFFF ? 0xFFFF : dma_len; in mac_esp_dma_length_limit()
|
| /drivers/net/wireless/mediatek/mt7601u/ |
| A D | dma.c | 121 u16 dma_len = get_unaligned_le16(data); in mt7601u_rx_next_seg_len() local 124 WARN_ON_ONCE(!dma_len) || in mt7601u_rx_next_seg_len() 125 WARN_ON_ONCE(dma_len + MT_DMA_HDRS > data_len) || in mt7601u_rx_next_seg_len() 126 WARN_ON_ONCE(dma_len & 0x3) || in mt7601u_rx_next_seg_len() 127 WARN_ON_ONCE(dma_len < min_seg_len)) in mt7601u_rx_next_seg_len() 130 return MT_DMA_HDRS + dma_len; in mt7601u_rx_next_seg_len()
|
| /drivers/crypto/marvell/octeontx2/ |
| A D | otx2_cpt_reqmgr.h | 157 u32 dma_len; member 192 info->dma_len, DMA_BIDIRECTIONAL); in otx2_cpt_info_destroy() 422 info->dma_len = total_mem_len - info_len; in cn10k_sgv2_info_create() 424 info->dma_len, DMA_BIDIRECTIONAL); in cn10k_sgv2_info_create() 531 info->dma_len = total_mem_len - info_len; in otx2_sg_info_create() 533 info->dma_len, DMA_BIDIRECTIONAL); in otx2_sg_info_create()
|
| /drivers/net/ethernet/chelsio/cxgb/ |
| A D | sge.c | 156 DEFINE_DMA_UNMAP_LEN(dma_len); 162 DEFINE_DMA_UNMAP_LEN(dma_len); 507 dma_unmap_len(ce, dma_len), DMA_FROM_DEVICE); in free_freelQ_buffers() 621 if (likely(dma_unmap_len(ce, dma_len))) { in free_cmdQ_buffers() 624 dma_unmap_len(ce, dma_len), in free_cmdQ_buffers() 843 dma_unmap_len_set(ce, dma_len, dma_len); in refill_free_list() 1055 dma_unmap_len(ce, dma_len), in get_packet() 1060 dma_unmap_len(ce, dma_len), in get_packet() 1073 dma_unmap_len(ce, dma_len), DMA_FROM_DEVICE); in get_packet() 1178 dma_unmap_len_set(ce1, dma_len, 0); in write_large_page_tx_descs() [all …]
|
| /drivers/nvme/host/ |
| A D | apple.c | 362 if (iod->dma_len) { in apple_nvme_unmap_data() 401 int dma_len = sg_dma_len(sg); in apple_nvme_setup_prps() local 415 dma_len -= (NVME_CTRL_PAGE_SIZE - offset); in apple_nvme_setup_prps() 416 if (dma_len) { in apple_nvme_setup_prps() 421 dma_len = sg_dma_len(sg); in apple_nvme_setup_prps() 460 dma_len -= NVME_CTRL_PAGE_SIZE; in apple_nvme_setup_prps() 465 if (dma_len > 0) in apple_nvme_setup_prps() 467 if (unlikely(dma_len < 0)) in apple_nvme_setup_prps() 471 dma_len = sg_dma_len(sg); in apple_nvme_setup_prps() 499 iod->dma_len = bv->bv_len; in apple_nvme_setup_prp_simple() [all …]
|
| /drivers/net/ethernet/sfc/falcon/ |
| A D | tx.c | 179 unsigned int dma_len; in ef4_tx_map_chunk() local 184 dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len); in ef4_tx_map_chunk() 186 buffer->len = dma_len; in ef4_tx_map_chunk() 189 len -= dma_len; in ef4_tx_map_chunk() 190 dma_addr += dma_len; in ef4_tx_map_chunk()
|
| /drivers/crypto/ |
| A D | talitos.c | 496 (edesc->buf + edesc->dma_len))->hdr; 1007 if (edesc->dma_len) in ipsec_esp_unmap() 1058 oicv = edesc->buf + edesc->dma_len; in ipsec_esp_decrypt_swauth_done() 1305 edesc->dma_len, in ipsec_esp() 1385 alloc_len += dma_len; in talitos_edesc_alloc() 1387 dma_len = 0; in talitos_edesc_alloc() 1408 edesc->dma_len = dma_len; in talitos_edesc_alloc() 1409 if (dma_len) in talitos_edesc_alloc() 1411 edesc->dma_len, in talitos_edesc_alloc() 1541 if (edesc->dma_len) in common_nonsnoop_unmap() [all …]
|
| /drivers/gpu/drm/kmb/ |
| A D | kmb_plane.c | 352 unsigned int dma_len; in kmb_plane_atomic_update() local 395 dma_len = (width * height * fb->format->cpp[0]); in kmb_plane_atomic_update() 396 drm_dbg(&kmb->drm, "dma_len=%d ", dma_len); in kmb_plane_atomic_update() 397 kmb_write_lcd(kmb, LCD_LAYERn_DMA_LEN(plane_id), dma_len); in kmb_plane_atomic_update() 398 kmb_write_lcd(kmb, LCD_LAYERn_DMA_LEN_SHADOW(plane_id), dma_len); in kmb_plane_atomic_update()
|
| /drivers/ata/ |
| A D | pata_arasan_cf.c | 434 u32 xfer_cnt, sglen, dma_len, xfer_ctr; in sg_xfer() local 474 dma_len = min(xfer_cnt, FIFO_SIZE); in sg_xfer() 475 ret = dma_xfer(acdev, src, dest, dma_len); in sg_xfer() 482 src += dma_len; in sg_xfer() 484 dest += dma_len; in sg_xfer() 486 sglen -= dma_len; in sg_xfer() 487 xfer_cnt -= dma_len; in sg_xfer()
|
| /drivers/crypto/marvell/octeontx/ |
| A D | otx_cptvf_reqmgr.h | 186 u32 dma_len; member 198 info->dma_len, DMA_BIDIRECTIONAL); in do_request_cleanup()
|
| /drivers/mmc/host/ |
| A D | omap_hsmmc.c | 165 unsigned int dma_len; member 183 unsigned int dma_len; member 1196 int dma_len; in omap_hsmmc_pre_dma_transfer() local 1208 dma_len = dma_map_sg(chan->device->dev, data->sg, data->sg_len, in omap_hsmmc_pre_dma_transfer() 1212 dma_len = host->next_data.dma_len; in omap_hsmmc_pre_dma_transfer() 1213 host->next_data.dma_len = 0; in omap_hsmmc_pre_dma_transfer() 1217 if (dma_len == 0) in omap_hsmmc_pre_dma_transfer() 1221 next->dma_len = dma_len; in omap_hsmmc_pre_dma_transfer() 1224 host->dma_len = dma_len; in omap_hsmmc_pre_dma_transfer()
|