| /drivers/net/ethernet/cisco/enic/ |
| A D | enic_res.h | 34 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_ex() argument 45 (u64)dma_addr | VNIC_PADDR_TARGET, in enic_queue_wq_desc_ex() 60 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_cont() argument 63 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_cont() 72 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc() 79 void *os_buf, dma_addr_t dma_addr, unsigned int len, in enic_queue_wq_desc_csum() argument 83 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_csum() 95 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_csum_l4() 106 enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len, in enic_queue_wq_desc_tso() 114 dma_addr_t dma_addr, unsigned int len) in enic_queue_rq_desc() argument [all …]
|
| /drivers/gpu/drm/ |
| A D | drm_pagemap.c | 218 dma_addr_t *dma_addr, in drm_pagemap_migrate_map_pages() argument 235 if (dma_mapping_error(dev, dma_addr[i])) in drm_pagemap_migrate_map_pages() 254 dma_addr_t *dma_addr, in drm_pagemap_migrate_unmap_pages() argument 261 if (!dma_addr[i] || dma_mapping_error(dev, dma_addr[i])) in drm_pagemap_migrate_unmap_pages() 264 dma_unmap_page(dev, dma_addr[i], PAGE_SIZE, dir); in drm_pagemap_migrate_unmap_pages() 317 dma_addr_t *dma_addr; in drm_pagemap_migrate_to_devmem() local 512 dma_addr_t *dma_addr; in drm_pagemap_evict_to_ram() local 531 dma_addr = buf + (2 * sizeof(*src) * npages); in drm_pagemap_evict_to_ram() 555 err = ops->copy_to_ram(pages, dma_addr, npages); in drm_pagemap_evict_to_ram() 615 dma_addr_t *dma_addr; in __drm_pagemap_migrate_to_ram() local [all …]
|
| /drivers/crypto/marvell/octeontx2/ |
| A D | otx2_cpt_reqmgr.h | 107 dma_addr_t dma_addr; member 197 if (req->out[i].dma_addr) in otx2_cpt_info_destroy() 199 req->out[i].dma_addr, in otx2_cpt_info_destroy() 205 if (req->in[i].dma_addr) in otx2_cpt_info_destroy() 207 req->in[i].dma_addr, in otx2_cpt_info_destroy() 274 if (list[j].dma_addr) { in setup_sgio_components() 275 dma_unmap_single(&pdev->dev, list[j].dma_addr, in setup_sgio_components() 279 list[j].dma_addr = 0; in setup_sgio_components() 339 if (list[j].dma_addr) { in sgv2io_components_setup() 340 dma_unmap_single(&pdev->dev, list[j].dma_addr, in sgv2io_components_setup() [all …]
|
| /drivers/net/ethernet/apm/xgene-v2/ |
| A D | ring.c | 24 next_dma = ring->dma_addr + (offset * XGENE_ENET_DESC_SIZE); in xge_setup_desc() 37 dma_addr_t dma_addr = ring->dma_addr; in xge_update_tx_desc_addr() local 39 xge_wr_csr(pdata, DMATXDESCL, dma_addr); in xge_update_tx_desc_addr() 40 xge_wr_csr(pdata, DMATXDESCH, upper_32_bits(dma_addr)); in xge_update_tx_desc_addr() 49 dma_addr_t dma_addr = ring->dma_addr; in xge_update_rx_desc_addr() local 51 xge_wr_csr(pdata, DMARXDESCL, dma_addr); in xge_update_rx_desc_addr() 52 xge_wr_csr(pdata, DMARXDESCH, upper_32_bits(dma_addr)); in xge_update_rx_desc_addr()
|
| A D | main.c | 72 dma_addr_t dma_addr; in xge_refill_buffers() local 92 ring->pkt_info[tail].dma_addr = dma_addr; in xge_refill_buffers() 209 tx_ring->pkt_info[tail].dma_addr = dma_addr; in xge_start_xmit() 240 dma_addr_t dma_addr; in xge_txc_poll() local 262 dma_addr = tx_ring->pkt_info[head].dma_addr; in xge_txc_poll() 290 dma_addr_t dma_addr; in xge_rx_poll() local 314 dma_addr = rx_ring->pkt_info[head].dma_addr; in xge_rx_poll() 373 dma_addr_t dma_addr; in xge_free_buffers() local 378 dma_addr = ring->pkt_info[i].dma_addr; in xge_free_buffers() 555 dma_addr_t dma_addr; in xge_free_pending_skb() local [all …]
|
| /drivers/media/platform/mediatek/vcodec/decoder/vdec/ |
| A D | vdec_h264_req_multi_if.c | 138 u64 dma_addr; member 494 inst->vsi_ext->bs.dma_addr = (u64)bs->dma_addr; in vdec_h264_slice_setup_lat_buffer_ext() 499 inst->vsi_ext->mv_buf_dma[i].dma_addr = mem->dma_addr; in vdec_h264_slice_setup_lat_buffer_ext() 502 inst->vsi_ext->ube.dma_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr; in vdec_h264_slice_setup_lat_buffer_ext() 508 inst->vsi_ext->err_map.dma_addr = lat_buf->wdma_err_addr.dma_addr; in vdec_h264_slice_setup_lat_buffer_ext() 511 inst->vsi_ext->slice_bc.dma_addr = lat_buf->slice_bc_addr.dma_addr; in vdec_h264_slice_setup_lat_buffer_ext() 554 inst->vsi_core_ext->err_map.dma_addr = lat_buf->wdma_err_addr.dma_addr; in vdec_h264_slice_setup_core_buffer_ext() 557 inst->vsi_core_ext->slice_bc.dma_addr = lat_buf->slice_bc_addr.dma_addr; in vdec_h264_slice_setup_core_buffer_ext() 568 inst->vsi_core_ext->mv_buf_dma[i].dma_addr = mem->dma_addr; in vdec_h264_slice_setup_core_buffer_ext() 1035 inst->vsi_ctx_ext.bs.dma_addr = (u64)bs->dma_addr; in vdec_h264_slice_single_decode_ext() [all …]
|
| A D | vdec_hevc_req_multi_if.c | 740 inst->vsi->bs.dma_addr = (u64)bs->dma_addr; in vdec_hevc_slice_setup_lat_buffer() 761 inst->vsi->mv_buf_dma[i].dma_addr = mem->dma_addr; in vdec_hevc_slice_setup_lat_buffer() 765 inst->vsi->ube.dma_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr; in vdec_hevc_slice_setup_lat_buffer() 768 inst->vsi->err_map.dma_addr = lat_buf->wdma_err_addr.dma_addr; in vdec_hevc_slice_setup_lat_buffer() 771 inst->vsi->slice_bc.dma_addr = lat_buf->slice_bc_addr.dma_addr; in vdec_hevc_slice_setup_lat_buffer() 778 share_info->trans.dma_addr = inst->vsi->trans.dma_addr; in vdec_hevc_slice_setup_lat_buffer() 831 inst->vsi_core->err_map.dma_addr = lat_buf->wdma_err_addr.dma_addr; in vdec_hevc_slice_setup_core_buffer() 834 inst->vsi_core->slice_bc.dma_addr = lat_buf->slice_bc_addr.dma_addr; in vdec_hevc_slice_setup_core_buffer() 837 inst->vsi_core->trans.dma_addr = share_info->trans.dma_addr; in vdec_hevc_slice_setup_core_buffer() 840 inst->vsi_core->wrap.dma_addr = inst->wrap_addr.dma_addr; in vdec_hevc_slice_setup_core_buffer() [all …]
|
| /drivers/net/ethernet/netronome/nfp/nfdk/ |
| A D | dp.c | 267 dma_addr_t dma_addr; in nfp_nfdk_tx() local 320 txbuf->dma_addr = dma_addr; in nfp_nfdk_tx() 348 dma_addr += tmp_dlen + 1; in nfp_nfdk_tx() 379 txbuf->dma_addr = dma_addr; in nfp_nfdk_tx() 619 rx_ring->rxbufs[wr_idx].dma_addr = dma_addr; in nfp_nfdk_rx_give_one() 909 dma_addr_t dma_addr; in nfp_nfdk_tx_xdp_buf() local 956 txbuf[1].dma_addr = rxbuf->dma_addr; in nfp_nfdk_tx_xdp_buf() 965 dma_addr = rxbuf->dma_addr + dma_off; in nfp_nfdk_tx_xdp_buf() 984 dma_addr += tmp_dlen + 1; in nfp_nfdk_tx_xdp_buf() 1326 dma_addr_t dma_addr; in nfp_nfdk_ctrl_tx_one() local [all …]
|
| /drivers/gpu/drm/xe/ |
| A D | xe_res_cursor.h | 59 const struct drm_pagemap_device_addr *dma_addr; member 95 cur->dma_addr = NULL; in xe_res_first() 192 cur->dma_addr = addr; in __xe_res_dma_next() 216 cur->dma_addr = NULL; in xe_res_first_sg() 236 XE_WARN_ON(!dma_addr); in xe_res_first_dma() 243 cur->dma_seg_size = PAGE_SIZE << dma_addr->order; in xe_res_first_dma() 246 cur->dma_addr = dma_addr; in xe_res_first_dma() 278 if (cur->dma_addr) { in xe_res_next() 325 if (cur->dma_addr) in xe_res_dma() 342 if (cur->dma_addr) in xe_res_is_vram() [all …]
|
| /drivers/net/ethernet/freescale/dpaa2/ |
| A D | dpaa2-eth-trace.h | 116 dma_addr_t dma_addr, 122 TP_ARGS(netdev, vaddr, size, dma_addr, map_size, bpid), 132 __field(dma_addr_t, dma_addr) 144 __entry->dma_addr = dma_addr; 157 &__entry->dma_addr, 167 dma_addr_t dma_addr, 171 TP_ARGS(netdev, vaddr, size, dma_addr, map_size, bpid) 179 dma_addr_t dma_addr, 183 TP_ARGS(netdev, vaddr, size, dma_addr, map_size, bpid)
|
| /drivers/net/ethernet/qualcomm/emac/ |
| A D | emac-mac.c | 602 tpbuf->dma_addr = 0; in emac_tx_q_descs_free() 638 rfbuf->dma_addr = 0; in emac_rx_q_free_descs() 687 tx_q->tpd.dma_addr = ring_header->dma_addr + ring_header->used; in emac_tx_q_desc_alloc() 731 rx_q->rrd.dma_addr = ring_header->dma_addr + ring_header->used; in emac_rx_descs_alloc() 735 rx_q->rfd.dma_addr = ring_header->dma_addr + ring_header->used; in emac_rx_descs_alloc() 1118 rfbuf->dma_addr = 0; in emac_mac_rx_process() 1193 tpbuf->dma_addr = 0; in emac_mac_tx_process() 1355 tpbuf->dma_addr); in emac_tx_fill_tpd() 1376 tpbuf->dma_addr); in emac_tx_fill_tpd() 1396 tpbuf->dma_addr); in emac_tx_fill_tpd() [all …]
|
| /drivers/net/ethernet/sfc/ |
| A D | tx_tso.c | 65 dma_addr_t dma_addr; member 103 dma_addr_t dma_addr, unsigned int len, in efx_tx_queue_insert() argument 119 buffer->dma_addr = dma_addr; in efx_tx_queue_insert() 122 dma_addr, len); in efx_tx_queue_insert() 130 dma_addr += dma_len; in efx_tx_queue_insert() 175 dma_addr_t dma_addr; in tso_start() local 200 st->header_dma_addr = dma_addr; in tso_start() 202 st->dma_addr = dma_addr + header_len; in tso_start() 216 st->dma_addr = st->unmap_addr; in tso_get_fragment() 270 st->dma_addr += n; in tso_fill_packet_with_fragment() [all …]
|
| A D | mcdi_functions.c | 80 dma_addr_t dma_addr; in efx_mcdi_ev_init() local 115 dma_addr = channel->eventq.dma_addr; in efx_mcdi_ev_init() 118 dma_addr += EFX_BUF_SIZE; in efx_mcdi_ev_init() 172 dma_addr_t dma_addr; in efx_mcdi_tx_init() local 185 dma_addr = tx_queue->txd.dma_addr; in efx_mcdi_tx_init() 188 tx_queue->queue, entries, (u64)dma_addr); in efx_mcdi_tx_init() 192 dma_addr += EFX_BUF_SIZE; in efx_mcdi_tx_init() 285 dma_addr_t dma_addr; in efx_mcdi_rx_init() local 309 dma_addr = rx_queue->rxd.dma_addr; in efx_mcdi_rx_init() 312 efx_rx_queue_index(rx_queue), entries, (u64)dma_addr); in efx_mcdi_rx_init() [all …]
|
| A D | tx_common.c | 329 dma_addr_t dma_addr, size_t len) in efx_tx_map_chunk() argument 345 buffer->dma_addr = dma_addr; in efx_tx_map_chunk() 348 dma_addr += dma_len; in efx_tx_map_chunk() 375 dma_addr_t dma_addr, unmap_addr; in efx_tx_map_data() local 387 unmap_addr = dma_addr; in efx_tx_map_data() 389 if (unlikely(dma_mapping_error(dma_dev, dma_addr))) in efx_tx_map_data() 400 efx_tx_map_chunk(tx_queue, dma_addr, header_len); in efx_tx_map_data() 402 dma_addr += header_len; in efx_tx_map_data() 411 buffer = efx_tx_map_chunk(tx_queue, dma_addr, len); in efx_tx_map_data() 418 buffer->dma_offset = buffer->dma_addr - unmap_addr; in efx_tx_map_data() [all …]
|
| /drivers/net/ethernet/sfc/siena/ |
| A D | tx_common.c | 289 dma_addr_t dma_addr, size_t len) in efx_siena_tx_map_chunk() argument 305 buffer->dma_addr = dma_addr; in efx_siena_tx_map_chunk() 308 dma_addr += dma_len; in efx_siena_tx_map_chunk() 335 dma_addr_t dma_addr, unmap_addr; in efx_siena_tx_map_data() local 347 unmap_addr = dma_addr; in efx_siena_tx_map_data() 349 if (unlikely(dma_mapping_error(dma_dev, dma_addr))) in efx_siena_tx_map_data() 362 dma_addr += header_len; in efx_siena_tx_map_data() 378 buffer->dma_offset = buffer->dma_addr - unmap_addr; in efx_siena_tx_map_data() 392 dma_addr = skb_frag_dma_map(dma_dev, fragment, 0, len, in efx_siena_tx_map_data() 396 unmap_addr = dma_addr; in efx_siena_tx_map_data() [all …]
|
| /drivers/net/ethernet/netronome/nfp/nfd3/ |
| A D | dp.c | 266 dma_addr_t dma_addr; in nfp_nfd3_tx() local 312 txbuf->dma_addr = dma_addr; in nfp_nfd3_tx() 357 tx_ring->txbufs[wr_idx].dma_addr = dma_addr; in nfp_nfd3_tx() 476 tx_buf->dma_addr = 0; in nfp_nfd3_tx_complete() 599 rx_ring->rxbufs[wr_idx].dma_addr = dma_addr; in nfp_nfd3_rx_give_one() 609 dma_addr + dp->rx_dma_off); in nfp_nfd3_rx_give_one() 871 txbuf->dma_addr = rxbuf->dma_addr; in nfp_nfd3_tx_xdp_buf() 1045 rxbuf->dma_addr); in nfp_nfd3_rx() 1058 rxbuf->dma_addr); in nfp_nfd3_rx() 1215 dma_addr_t dma_addr; in nfp_nfd3_ctrl_tx_one() local [all …]
|
| /drivers/accel/ivpu/ |
| A D | ivpu_mmu_context.c | 49 dma_addr_t dma_addr; in ivpu_pgtable_alloc_page() local 60 if (dma_mapping_error(vdev->drm.dev, dma_addr)) in ivpu_pgtable_alloc_page() 68 *dma = dma_addr; in ivpu_pgtable_alloc_page() 240 u64 vpu_addr, dma_addr_t dma_addr, u64 prot) in ivpu_mmu_context_map_page() argument 268 pte[pte_idx] = dma_addr | prot; in ivpu_mmu_context_map_page() 275 dma_addr_t dma_addr, u64 prot) in ivpu_mmu_context_map_cont_64k() argument 280 drm_WARN_ON(&vdev->drm, !IS_ALIGNED(dma_addr, size)); in ivpu_mmu_context_map_cont_64k() 292 dma_addr += IVPU_MMU_PAGE_SIZE; in ivpu_mmu_context_map_cont_64k() 330 dma_addr += map_size; in ivpu_mmu_context_map_pages() 461 ctx->id, dma_addr, vpu_addr, size); in ivpu_mmu_context_map_sgt() [all …]
|
| /drivers/media/pci/solo6x10/ |
| A D | solo6x10-p2m.c | 32 dma_addr_t dma_addr; in solo_p2m_dma() local 40 dma_addr = dma_map_single(&solo_dev->pdev->dev, sys_addr, size, in solo_p2m_dma() 42 if (dma_mapping_error(&solo_dev->pdev->dev, dma_addr)) in solo_p2m_dma() 45 ret = solo_p2m_dma_t(solo_dev, wr, dma_addr, ext_addr, size, in solo_p2m_dma() 48 dma_unmap_single(&solo_dev->pdev->dev, dma_addr, size, in solo_p2m_dma() 93 desc[1].dma_addr); in solo_p2m_dma_desc() 125 dma_addr_t dma_addr, u32 ext_addr, u32 size, in solo_p2m_fill_desc() argument 128 WARN_ON_ONCE(dma_addr & 0x03); in solo_p2m_fill_desc() 141 desc->dma_addr = dma_addr; in solo_p2m_fill_desc() 146 dma_addr_t dma_addr, u32 ext_addr, u32 size, in solo_p2m_dma_t() argument [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/en/ |
| A D | xdp.c | 68 dma_addr_t dma_addr; in mlx5e_xmit_xdp_buff() local 103 xdptxd->dma_addr = dma_addr; in mlx5e_xmit_xdp_buff() 115 (union mlx5e_xdp_info) { .frame.dma_addr = dma_addr }); in mlx5e_xmit_xdp_buff() 145 xdptxd->dma_addr = dma_addr; in mlx5e_xmit_xdp_buff() 541 dma_addr_t dma_addr = xdptxd->dma_addr; in mlx5e_xmit_xdp_frame() local 603 dma_addr += inline_hdr_sz; in mlx5e_xmit_xdp_frame() 670 dma_addr_t dma_addr; in mlx5e_free_xdpsq_desc() local 675 dma_addr = xdpi.frame.dma_addr; in mlx5e_free_xdpsq_desc() 677 dma_unmap_single(sq->pdev, dma_addr, in mlx5e_free_xdpsq_desc() 688 dma_addr = xdpi.frame.dma_addr; in mlx5e_free_xdpsq_desc() [all …]
|
| /drivers/media/platform/mediatek/vcodec/encoder/ |
| A D | venc_vpu_if.c | 270 if ((frm_buf->fb_addr[0].dma_addr % 16 == 0) && in vpu_enc_encode_32bits() 271 (frm_buf->fb_addr[1].dma_addr % 16 == 0) && in vpu_enc_encode_32bits() 272 (frm_buf->fb_addr[2].dma_addr % 16 == 0)) { in vpu_enc_encode_32bits() 282 out.base.bs_addr = bs_buf->dma_addr; in vpu_enc_encode_32bits() 316 if ((frm_buf->fb_addr[0].dma_addr % 16 == 0) && in vpu_enc_encode_34bits() 317 (frm_buf->fb_addr[1].dma_addr % 16 == 0) && in vpu_enc_encode_34bits() 318 (frm_buf->fb_addr[2].dma_addr % 16 == 0)) { in vpu_enc_encode_34bits() 319 out.input_addr[0] = frm_buf->fb_addr[0].dma_addr; in vpu_enc_encode_34bits() 320 out.input_addr[1] = frm_buf->fb_addr[1].dma_addr; in vpu_enc_encode_34bits() 321 out.input_addr[2] = frm_buf->fb_addr[2].dma_addr; in vpu_enc_encode_34bits() [all …]
|
| /drivers/xen/ |
| A D | swiotlb-xen.c | 72 dma_addr_t dma_addr) in xen_dma_to_phys() argument 74 return xen_bus_to_phys(dev, dma_to_phys(dev, dma_addr)); in xen_dma_to_phys() 100 dma_addr_t dma_addr) in xen_swiotlb_find_pool() argument 292 phys_addr_t paddr = xen_dma_to_phys(dev, dma_addr); in xen_swiotlb_sync_single_for_cpu() 296 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr)))) in xen_swiotlb_sync_single_for_cpu() 299 xen_dma_sync_for_cpu(dev, dma_addr, size, dir); in xen_swiotlb_sync_single_for_cpu() 302 pool = xen_swiotlb_find_pool(dev, dma_addr); in xen_swiotlb_sync_single_for_cpu() 311 phys_addr_t paddr = xen_dma_to_phys(dev, dma_addr); in xen_swiotlb_sync_single_for_device() 314 pool = xen_swiotlb_find_pool(dev, dma_addr); in xen_swiotlb_sync_single_for_device() 319 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr)))) in xen_swiotlb_sync_single_for_device() [all …]
|
| /drivers/net/ethernet/sfc/falcon/ |
| A D | tx.c | 38 buffer->dma_addr = page_buf->dma_addr + offset; in ef4_tx_get_copy_buffer() 174 dma_addr_t dma_addr, in ef4_tx_map_chunk() argument 187 buffer->dma_addr = dma_addr; in ef4_tx_map_chunk() 190 dma_addr += dma_len; in ef4_tx_map_chunk() 204 dma_addr_t dma_addr, unmap_addr; in ef4_tx_map_data() local 216 unmap_addr = dma_addr; in ef4_tx_map_data() 218 if (unlikely(dma_mapping_error(dma_dev, dma_addr))) in ef4_tx_map_data() 226 buffer = ef4_tx_map_chunk(tx_queue, dma_addr, len); in ef4_tx_map_data() 233 buffer->dma_offset = buffer->dma_addr - unmap_addr; in ef4_tx_map_data() 247 dma_addr = skb_frag_dma_map(dma_dev, fragment, in ef4_tx_map_data() [all …]
|
| /drivers/usb/musb/ |
| A D | tusb6010_omap.c | 41 dma_addr_t dma_addr; member 140 dma_unmap_single(dev, chdat->dma_addr, in tusb_omap_dma_cb() 145 dma_unmap_single(dev, chdat->dma_addr, in tusb_omap_dma_cb() 182 u8 rndis_mode, dma_addr_t dma_addr, u32 len) in tusb_omap_dma_program() argument 211 if (dma_addr & 0x2) in tusb_omap_dma_program() 256 chdat->dma_addr = dma_addr; in tusb_omap_dma_program() 262 dma_map_single(dev, phys_to_virt(dma_addr), len, in tusb_omap_dma_program() 266 dma_map_single(dev, phys_to_virt(dma_addr), len, in tusb_omap_dma_program() 273 if ((dma_addr & 0x3) == 0) { in tusb_omap_dma_program() 287 chdat->epnum, chdat->tx ? "tx" : "rx", &dma_addr, in tusb_omap_dma_program() [all …]
|
| /drivers/crypto/caam/ |
| A D | key_gen.c | 50 dma_addr_t dma_addr; in gen_split_key() local 75 dma_addr = dma_map_single(jrdev, key_out, local_max, DMA_BIDIRECTIONAL); in gen_split_key() 76 if (dma_mapping_error(jrdev, dma_addr)) { in gen_split_key() 82 append_key(desc, dma_addr, keylen, CLASS_2 | KEY_DEST_CLASS_REG); in gen_split_key() 100 append_fifo_store(desc, dma_addr, adata->keylen, in gen_split_key() 121 dma_unmap_single(jrdev, dma_addr, local_max, DMA_BIDIRECTIONAL); in gen_split_key()
|
| /drivers/crypto/intel/qat/qat_common/ |
| A D | adf_rl_admin.c | 12 prep_admin_req_msg(struct rl_sla *sla, dma_addr_t dma_addr, in prep_admin_req_msg() argument 17 req->init_cfg_ptr = dma_addr; in prep_admin_req_msg() 76 dma_addr_t dma_addr; in adf_rl_send_admin_add_update_msg() local 80 &dma_addr, GFP_KERNEL); in adf_rl_send_admin_add_update_msg() 85 prep_admin_req_msg(sla, dma_addr, fw_params, &req, is_update); in adf_rl_send_admin_add_update_msg() 89 dma_addr); in adf_rl_send_admin_add_update_msg()
|