| /drivers/gpu/drm/nouveau/ |
| A D | nouveau_uvmm.c | 183 u64 addr = uvma->va.va.addr; in nouveau_uvma_vmm_put() 184 u64 range = uvma->va.va.range; in nouveau_uvma_vmm_put() 193 u64 addr = uvma->va.va.addr; in nouveau_uvma_map() 195 u64 range = uvma->va.va.range; in nouveau_uvma_map() 204 u64 addr = uvma->va.va.addr; in nouveau_uvma_unmap() 205 u64 range = uvma->va.va.range; in nouveau_uvma_unmap() 510 u64 ustart = va->va.addr; in nouveau_uvmm_sm_prepare_unwind() 528 u64 ustart = va->va.addr; in nouveau_uvmm_sm_prepare_unwind() 652 u64 ustart = va->va.addr; in nouveau_uvmm_sm_prepare() 815 u64 end = uvma->va.va.addr + uvma->va.va.range; in op_remap() [all …]
|
| /drivers/scsi/qedi/ |
| A D | qedi_dbg.c | 14 va_list va; in qedi_dbg_err() local 20 vaf.va = &va; in qedi_dbg_err() 28 va_end(va); in qedi_dbg_err() 35 va_list va; in qedi_dbg_warn() local 41 vaf.va = &va; in qedi_dbg_warn() 53 va_end(va); in qedi_dbg_warn() 60 va_list va; in qedi_dbg_notice() local 66 vaf.va = &va; in qedi_dbg_notice() 79 va_end(va); in qedi_dbg_notice() 86 va_list va; in qedi_dbg_info() local [all …]
|
| /drivers/gpu/drm/ |
| A D | drm_gpuvm.c | 1762 u64 addr = va->va.addr; in drm_gpuva_insert() 1763 u64 range = va->va.range; in drm_gpuva_insert() 1919 if (va->va.addr != addr || in drm_gpuva_find() 1920 va->va.range != range) in drm_gpuva_find() 2096 op.unmap.va = va; in op_unmap_cb() 2118 u64 addr = va->va.addr; in __drm_gpuvm_sm_map() 2149 .va = va, in __drm_gpuvm_sm_map() 2227 .va = va, in __drm_gpuvm_sm_map() 2261 u64 addr = va->va.addr; in __drm_gpuvm_sm_unmap() 2782 op->prefetch.va = va; in drm_gpuvm_prefetch_ops_create() [all …]
|
| /drivers/scsi/qedf/ |
| A D | qedf_dbg.c | 13 va_list va; in qedf_dbg_err() local 19 vaf.va = &va; in qedf_dbg_err() 27 va_end(va); in qedf_dbg_err() 34 va_list va; in qedf_dbg_warn() local 40 vaf.va = &va; in qedf_dbg_warn() 52 va_end(va); in qedf_dbg_warn() 59 va_list va; in qedf_dbg_notice() local 65 vaf.va = &va; in qedf_dbg_notice() 78 va_end(va); in qedf_dbg_notice() 85 va_list va; in qedf_dbg_info() local [all …]
|
| /drivers/gpu/drm/i915/gt/ |
| A D | selftest_tlb.c | 33 struct i915_vma *va, in pte_tlbinv() argument 68 va->size, align); in pte_tlbinv() 75 if (vb != va) { in pte_tlbinv() 93 if (va != vb) in pte_tlbinv() 97 va->page_sizes.phys, va->page_sizes.sg, in pte_tlbinv() 148 if (va == vb) { in pte_tlbinv() 153 va->page_sizes.phys, va->page_sizes.sg); in pte_tlbinv() 189 if (vb != va) in pte_tlbinv() 284 if (IS_ERR(va)) { in mem_tlbinv() 320 if (BIT_ULL(bit) < i915_vm_obj_min_alignment(va->vm, va->obj)) in mem_tlbinv() [all …]
|
| /drivers/dio/ |
| A D | dio.c | 127 void *va; in dio_find() local 146 iounmap(va); in dio_find() 150 prid = DIO_ID(va); in dio_find() 153 secid = DIO_SECID(va); in dio_find() 160 iounmap(va); in dio_find() 200 u_char *va; in dio_init() local 219 iounmap(va); in dio_init() 227 iounmap(va); in dio_init() 241 prid = DIO_ID(va); in dio_init() 244 secid = DIO_SECID(va); in dio_init() [all …]
|
| /drivers/gpu/drm/imagination/ |
| A D | pvr_vm.c | 118 #define to_pvr_vm_gpuva(va) container_of_const(va, struct pvr_vm_gpuva, base) argument 358 op->map.va.addr); in pvr_vm_gpuva_map() 386 int err = pvr_mmu_unmap(ctx->mmu_op_ctx, op->unmap.va->va.addr, in pvr_vm_gpuva_unmap() 387 op->unmap.va->va.range); in pvr_vm_gpuva_unmap() 853 struct drm_gpuva *va; in pvr_vm_unmap() local 859 if (va) { in pvr_vm_unmap() 862 va->va.addr, va->va.range); in pvr_vm_unmap() 891 if (!va) in pvr_vm_unmap_all() 897 va->va.addr, va->va.range)); in pvr_vm_unmap_all() 1155 if (!va) in pvr_vm_find_gem_object() [all …]
|
| /drivers/media/platform/mediatek/vcodec/decoder/vdec/ |
| A D | vdec_vp9_if.c | 37 unsigned long va; member 391 if (mem->va) in vp9_alloc_work_buf() 403 vsi->mv_buf.va = (unsigned long)mem->va; in vp9_alloc_work_buf() 409 if (mem->va) in vp9_alloc_work_buf() 420 vsi->seg_id_buf.va = (unsigned long)mem->va; in vp9_alloc_work_buf() 560 inst = mem.va; in vp9_alloc_inst() 571 if (mem.va) in vp9_free_inst() 667 inst->vsi->mv_buf.va = (unsigned long)inst->mv_buf.va; in vp9_reset() 672 inst->vsi->seg_id_buf.va = (unsigned long)inst->seg_id_buf.va; in vp9_reset() 765 if (mem->va) in vdec_vp9_deinit() [all …]
|
| /drivers/infiniband/sw/rxe/ |
| A D | rxe_mr.c | 252 void *va; in rxe_mr_copy_xarray() local 261 va = kmap_local_page(page); in rxe_mr_copy_xarray() 266 kunmap_local(va); in rxe_mr_copy_xarray() 283 u8 *va; in rxe_mr_copy_dma() local 296 kunmap_local(va); in rxe_mr_copy_dma() 434 u8 *va; in rxe_mr_flush_pmem_iova() local 451 kunmap_local(va); in rxe_mr_flush_pmem_iova() 492 u64 *va; in rxe_mr_do_atomic_op() local 537 kunmap_local(va); in rxe_mr_do_atomic_op() 546 u64 *va; in rxe_mr_do_atomic_write() local [all …]
|
| /drivers/video/fbdev/matrox/ |
| A D | matroxfb_base.h | 128 return readb(va.vaddr + offs); in mga_readb() 132 writeb(value, va.vaddr + offs); in mga_writeb() 136 writew(value, va.vaddr + offs); in mga_writew() 140 return readl(va.vaddr + offs); in mga_readl() 144 writel(value, va.vaddr + offs); in mga_writel() 156 iowrite32_rep(va.vaddr, src, len >> 2); in mga_memcpy_toio() 158 u_int32_t __iomem* addr = va.vaddr; in mga_memcpy_toio() 178 static inline void vaddr_add(vaddr_t* va, unsigned long offs) { in vaddr_add() argument 179 va->vaddr += offs; in vaddr_add() 182 static inline void __iomem* vaddr_va(vaddr_t va) { in vaddr_va() argument [all …]
|
| /drivers/tee/amdtee/ |
| A D | shm_pool.c | 15 unsigned long va; in pool_op_alloc() local 22 va = __get_free_pages(GFP_KERNEL | __GFP_ZERO, order); in pool_op_alloc() 23 if (!va) in pool_op_alloc() 26 shm->kaddr = (void *)va; in pool_op_alloc() 27 shm->paddr = __psp_pa((void *)va); in pool_op_alloc() 33 free_pages(va, order); in pool_op_alloc()
|
| /drivers/net/ethernet/ti/icssg/ |
| A D | icssg_config.c | 168 void __iomem *smem = prueth->shram.va; in icssg_miig_queues_init() 258 p = emac->dram.va + MGR_R30_CMD_OFFSET; in emac_r30_cmd_init() 270 p = emac->dram.va + MGR_R30_CMD_OFFSET; in emac_r30_is_done() 303 bpool_cfg = emac->dram.va + BUFFER_POOL_0_ADDR_OFFSET; in prueth_fw_offload_buffer_setup() 341 rxq_ctx = emac->dram.va + HOST_RX_Q_EXP_CONTEXT_OFFSET; in prueth_fw_offload_buffer_setup() 362 rxq_ctx = emac->dram.va + DEFAULT_MSMC_Q_OFFSET; in prueth_fw_offload_buffer_setup() 390 bpool_cfg = emac->dram.va + BUFFER_POOL_0_ADDR_OFFSET; in prueth_emac_buffer_setup() 405 bpool_cfg = emac->dram.va + BUFFER_POOL_0_ADDR_OFFSET; in prueth_emac_buffer_setup() 448 rxq_ctx = emac->dram.va + DEFAULT_MSMC_Q_OFFSET; in prueth_emac_buffer_setup() 592 p = emac->dram.va + MGR_R30_CMD_OFFSET; in icssg_set_port_state() [all …]
|
| /drivers/gpu/drm/lima/ |
| A D | lima_vm.c | 29 #define LIMA_PDE(va) (va >> LIMA_VM_PD_SHIFT) argument 30 #define LIMA_PTE(va) ((va & LIMA_VM_PT_MASK) >> LIMA_VM_PT_SHIFT) argument 31 #define LIMA_PBE(va) (va >> LIMA_VM_PB_SHIFT) argument 32 #define LIMA_BTE(va) ((va & LIMA_VM_BT_MASK) >> LIMA_VM_BT_SHIFT) argument 47 static int lima_vm_map_page(struct lima_vm *vm, dma_addr_t pa, u32 va) in lima_vm_map_page() argument 49 u32 pbe = LIMA_PBE(va); in lima_vm_map_page() 50 u32 bte = LIMA_BTE(va); in lima_vm_map_page() 81 list_for_each_entry(bo_va, &bo->va, list) { in lima_vm_bo_find() 138 list_add_tail(&bo_va->list, &bo->va); in lima_vm_bo_add()
|
| /drivers/infiniband/hw/irdma/ |
| A D | puda.c | 147 if (!buf_mem.va) in irdma_puda_alloc_buf() 150 buf = buf_mem.va; in irdma_puda_alloc_buf() 153 if (!buf->mem.va) in irdma_puda_alloc_buf() 158 kfree(buf->mem.va); in irdma_puda_alloc_buf() 162 buf->buf_mem.va = buf_mem.va; in irdma_puda_alloc_buf() 168 kfree(buf_mem.va); in irdma_puda_alloc_buf() 182 kfree(buf->mem.va); in irdma_puda_dele_buf() 659 if (!rsrc->qpmem.va) in irdma_puda_qp_create() 957 kfree(vmem->va); in irdma_puda_dele_rsrc() 1024 if (!vmem->va) in irdma_puda_create_rsrc() [all …]
|
| A D | hmc.c | 347 if (!mem || !mem->va) { in irdma_finish_del_sd_reg() 352 mem->va = NULL; in irdma_finish_del_sd_reg() 492 if (!dma_mem.va) in irdma_add_sd_table_entry() 500 if (!vmem->va) { in irdma_add_sd_table_entry() 502 dma_mem.va, dma_mem.pa); in irdma_add_sd_table_entry() 503 dma_mem.va = NULL; in irdma_add_sd_table_entry() 576 if (!page->va) in irdma_add_pd_table_entry() 586 pd_addr = pd_table->pd_page_addr.va; in irdma_add_pd_table_entry() 640 pd_addr = pd_table->pd_page_addr.va; in irdma_remove_pd_bp() 647 if (!mem || !mem->va) in irdma_remove_pd_bp() [all …]
|
| A D | hw.c | 601 cqp->sq.va = NULL; in irdma_destroy_cqp() 616 vfree(aeq->mem.va); in irdma_destroy_virt_aeq() 901 va = (unsigned long)rf->obj_next.va; in irdma_obj_aligned_mem() 902 newva = va; in irdma_obj_aligned_mem() 905 extra = newva - va; in irdma_obj_aligned_mem() 906 memptr->va = (u8 *)va + extra; in irdma_obj_aligned_mem() 912 rf->obj_next.va = (u8 *)memptr->va + size; in irdma_obj_aligned_mem() 951 if (!cqp->sq.va) { in irdma_create_cqp() 1013 cqp->sq.va = NULL; in irdma_create_cqp() 1349 if (!aeq->mem.va) in irdma_create_virt_aeq() [all …]
|
| /drivers/firmware/efi/ |
| A D | mokvar-table.c | 104 void *va = NULL; in efi_mokvar_table_init() local 136 if (va) in efi_mokvar_table_init() 137 early_memunmap(va, sizeof(*mokvar_entry)); in efi_mokvar_table_init() 138 va = early_memremap(efi.mokvar_table + cur_offset, sizeof(*mokvar_entry)); in efi_mokvar_table_init() 139 if (!va) { in efi_mokvar_table_init() 144 mokvar_entry = va; in efi_mokvar_table_init() 174 if (va) in efi_mokvar_table_init() 175 early_memunmap(va, sizeof(*mokvar_entry)); in efi_mokvar_table_init()
|
| /drivers/tee/ |
| A D | tee_shm_pool.c | 15 unsigned long va; in pool_op_gen_alloc() local 21 va = gen_pool_alloc_algo(genpool, s, gen_pool_first_fit_align, &data); in pool_op_gen_alloc() 22 if (!va) in pool_op_gen_alloc() 25 memset((void *)va, 0, s); in pool_op_gen_alloc() 26 shm->kaddr = (void *)va; in pool_op_gen_alloc() 27 shm->paddr = gen_pool_virt_to_phys(genpool, va); in pool_op_gen_alloc()
|
| /drivers/infiniband/hw/usnic/ |
| A D | usnic_uiom.c | 194 long unsigned va, size; in usnic_uiom_unmap_sorted_intervals() local 197 va = interval->start << PAGE_SHIFT; in usnic_uiom_unmap_sorted_intervals() 201 usnic_dbg("va 0x%lx size 0x%lx", va, PAGE_SIZE); in usnic_uiom_unmap_sorted_intervals() 202 iommu_unmap(pd->domain, va, PAGE_SIZE); in usnic_uiom_unmap_sorted_intervals() 203 va += PAGE_SIZE; in usnic_uiom_unmap_sorted_intervals() 220 vpn_start = (uiomr->va & PAGE_MASK) >> PAGE_SHIFT; in __usnic_uiom_reg_release() 251 long int va = uiomr->va & PAGE_MASK; in usnic_uiom_map_sorted_intervals() local 261 if ((va >> PAGE_SHIFT) < interval_node->start) in usnic_uiom_map_sorted_intervals() 266 va_start = va; in usnic_uiom_map_sorted_intervals() 286 va_start = va; in usnic_uiom_map_sorted_intervals() [all …]
|
| /drivers/gpu/drm/panthor/ |
| A D | panthor_mmu.c | 181 } va; member 1024 if (va != PANTHOR_VM_KERNEL_AUTO_VA && !IS_ALIGNED(va, vm_pgsz)) in panthor_vm_alloc_va() 1210 op_ctx->va.addr = va; in panthor_vm_prepare_map_op_ctx() 1316 op_ctx->va.addr = va; in panthor_vm_prepare_unmap_op_ctx() 1322 if (va != ALIGN(va, SZ_2M)) in panthor_vm_prepare_unmap_op_ctx() 1325 if (va + size != ALIGN(va + size, SZ_2M) && in panthor_vm_prepare_unmap_op_ctx() 1326 ALIGN(va + size, SZ_2M) != ALIGN(va, SZ_2M)) in panthor_vm_prepare_unmap_op_ctx() 1394 *bo_offset = vma->base.gem.offset + (va - vma->base.va.addr); in panthor_vm_get_bo_for_va() 2178 ret = drm_gpuvm_sm_map(&vm->base, vm, op->va.addr, op->va.range, in panthor_vm_exec_op() 2183 ret = drm_gpuvm_sm_unmap(&vm->base, vm, op->va.addr, op->va.range); in panthor_vm_exec_op() [all …]
|
| /drivers/net/wireless/ath/wil6210/ |
| A D | txrx_edma.c | 52 if (!wil->srings[i].va) in wil_find_free_sring() 65 if (!sring || !sring->va) in wil_sring_free() 75 sring->va = NULL; in wil_sring_free() 97 if (!sring->va) in wil_sring_alloc() 165 &ring->va[i].rx.enhanced; in wil_ring_alloc_skb_edma() 390 if (!ring->va) in wil_ring_alloc_desc_ring() 395 ring->edma_rx_swtail.va = in wil_ring_alloc_desc_ring() 410 ring->va = NULL; in wil_ring_alloc_desc_ring() 424 if (!ring->va) in wil_ring_free_edma() 476 ring->va = NULL; in wil_ring_free_edma() [all …]
|
| A D | debug.c | 17 vaf.va = &args; in __wil_err() 33 vaf.va = &args; in __wil_err_ratelimited() 49 vaf.va = &args; in wil_dbg_ratelimited() 62 vaf.va = &args; in __wil_info() 75 vaf.va = &args; in wil_dbg_trace()
|
| /drivers/remoteproc/ |
| A D | pru_rproc.c | 656 void *va = NULL; in pru_d_da_to_va() local 670 va = (__force void *)(dram0.va + offset); in pru_d_da_to_va() 674 va = (__force void *)(dram1.va + offset); in pru_d_da_to_va() 678 va = (__force void *)(shrd_ram.va + offset); in pru_d_da_to_va() 681 return va; in pru_d_da_to_va() 694 void *va = NULL; in pru_i_da_to_va() local 719 va = (__force void *)(pru->mem_regions[PRU_IOMEM_IRAM].va + in pru_i_da_to_va() 723 return va; in pru_i_da_to_va() 742 void *va; in pru_da_to_va() local 745 va = pru_i_da_to_va(pru, da, len); in pru_da_to_va() [all …]
|
| /drivers/gpu/drm/i915/ |
| A D | intel_memory_region.c | 37 u8 __iomem *va, int pagesize, in __iopagetest() argument 44 memset_io(va, value, pagesize); /* or GPF! */ in __iopagetest() 47 result[0] = ioread8(va); in __iopagetest() 48 result[1] = ioread8(va + byte); in __iopagetest() 49 result[2] = ioread8(va + pagesize - 1); in __iopagetest() 66 void __iomem *va; in iopagetest() local 70 va = ioremap_wc(mem->io.start + offset, PAGE_SIZE); in iopagetest() 71 if (!va) { in iopagetest() 79 err = __iopagetest(mem, va, PAGE_SIZE, val[i], offset, caller); in iopagetest() 83 err = __iopagetest(mem, va, PAGE_SIZE, ~val[i], offset, caller); in iopagetest() [all …]
|
| /drivers/media/platform/mediatek/vcodec/common/ |
| A D | mtk_vcodec_util.c | 66 mem->va = dma_alloc_attrs(&plat_dev->dev, mem->size, &mem->dma_addr, in mtk_vcodec_mem_alloc() 68 if (!mem->va) { in mtk_vcodec_mem_alloc() 74 mtk_v4l2_debug(plat_dev, 3, "[%d] - va = %p dma = 0x%lx size = 0x%zx", id, mem->va, in mtk_vcodec_mem_alloc() 99 if (!mem->va) { in mtk_vcodec_mem_free() 106 mtk_v4l2_debug(plat_dev, 3, "[%d] - va = %p dma = 0x%lx size = 0x%zx", id, mem->va, in mtk_vcodec_mem_free() 109 dma_free_coherent(&plat_dev->dev, mem->size, mem->va, mem->dma_addr); in mtk_vcodec_mem_free() 110 mem->va = NULL; in mtk_vcodec_mem_free()
|