Home
last modified time | relevance | path

Searched refs:va (Results 1 – 25 of 362) sorted by relevance

12345678910>>...15

/drivers/gpu/drm/nouveau/
A Dnouveau_uvmm.c183 u64 addr = uvma->va.va.addr; in nouveau_uvma_vmm_put()
184 u64 range = uvma->va.va.range; in nouveau_uvma_vmm_put()
193 u64 addr = uvma->va.va.addr; in nouveau_uvma_map()
195 u64 range = uvma->va.va.range; in nouveau_uvma_map()
204 u64 addr = uvma->va.va.addr; in nouveau_uvma_unmap()
205 u64 range = uvma->va.va.range; in nouveau_uvma_unmap()
510 u64 ustart = va->va.addr; in nouveau_uvmm_sm_prepare_unwind()
528 u64 ustart = va->va.addr; in nouveau_uvmm_sm_prepare_unwind()
652 u64 ustart = va->va.addr; in nouveau_uvmm_sm_prepare()
815 u64 end = uvma->va.va.addr + uvma->va.va.range; in op_remap()
[all …]
/drivers/scsi/qedi/
A Dqedi_dbg.c14 va_list va; in qedi_dbg_err() local
20 vaf.va = &va; in qedi_dbg_err()
28 va_end(va); in qedi_dbg_err()
35 va_list va; in qedi_dbg_warn() local
41 vaf.va = &va; in qedi_dbg_warn()
53 va_end(va); in qedi_dbg_warn()
60 va_list va; in qedi_dbg_notice() local
66 vaf.va = &va; in qedi_dbg_notice()
79 va_end(va); in qedi_dbg_notice()
86 va_list va; in qedi_dbg_info() local
[all …]
/drivers/gpu/drm/
A Ddrm_gpuvm.c1762 u64 addr = va->va.addr; in drm_gpuva_insert()
1763 u64 range = va->va.range; in drm_gpuva_insert()
1919 if (va->va.addr != addr || in drm_gpuva_find()
1920 va->va.range != range) in drm_gpuva_find()
2096 op.unmap.va = va; in op_unmap_cb()
2118 u64 addr = va->va.addr; in __drm_gpuvm_sm_map()
2149 .va = va, in __drm_gpuvm_sm_map()
2227 .va = va, in __drm_gpuvm_sm_map()
2261 u64 addr = va->va.addr; in __drm_gpuvm_sm_unmap()
2782 op->prefetch.va = va; in drm_gpuvm_prefetch_ops_create()
[all …]
/drivers/scsi/qedf/
A Dqedf_dbg.c13 va_list va; in qedf_dbg_err() local
19 vaf.va = &va; in qedf_dbg_err()
27 va_end(va); in qedf_dbg_err()
34 va_list va; in qedf_dbg_warn() local
40 vaf.va = &va; in qedf_dbg_warn()
52 va_end(va); in qedf_dbg_warn()
59 va_list va; in qedf_dbg_notice() local
65 vaf.va = &va; in qedf_dbg_notice()
78 va_end(va); in qedf_dbg_notice()
85 va_list va; in qedf_dbg_info() local
[all …]
/drivers/gpu/drm/i915/gt/
A Dselftest_tlb.c33 struct i915_vma *va, in pte_tlbinv() argument
68 va->size, align); in pte_tlbinv()
75 if (vb != va) { in pte_tlbinv()
93 if (va != vb) in pte_tlbinv()
97 va->page_sizes.phys, va->page_sizes.sg, in pte_tlbinv()
148 if (va == vb) { in pte_tlbinv()
153 va->page_sizes.phys, va->page_sizes.sg); in pte_tlbinv()
189 if (vb != va) in pte_tlbinv()
284 if (IS_ERR(va)) { in mem_tlbinv()
320 if (BIT_ULL(bit) < i915_vm_obj_min_alignment(va->vm, va->obj)) in mem_tlbinv()
[all …]
/drivers/dio/
A Ddio.c127 void *va; in dio_find() local
146 iounmap(va); in dio_find()
150 prid = DIO_ID(va); in dio_find()
153 secid = DIO_SECID(va); in dio_find()
160 iounmap(va); in dio_find()
200 u_char *va; in dio_init() local
219 iounmap(va); in dio_init()
227 iounmap(va); in dio_init()
241 prid = DIO_ID(va); in dio_init()
244 secid = DIO_SECID(va); in dio_init()
[all …]
/drivers/gpu/drm/imagination/
A Dpvr_vm.c118 #define to_pvr_vm_gpuva(va) container_of_const(va, struct pvr_vm_gpuva, base) argument
358 op->map.va.addr); in pvr_vm_gpuva_map()
386 int err = pvr_mmu_unmap(ctx->mmu_op_ctx, op->unmap.va->va.addr, in pvr_vm_gpuva_unmap()
387 op->unmap.va->va.range); in pvr_vm_gpuva_unmap()
853 struct drm_gpuva *va; in pvr_vm_unmap() local
859 if (va) { in pvr_vm_unmap()
862 va->va.addr, va->va.range); in pvr_vm_unmap()
891 if (!va) in pvr_vm_unmap_all()
897 va->va.addr, va->va.range)); in pvr_vm_unmap_all()
1155 if (!va) in pvr_vm_find_gem_object()
[all …]
/drivers/media/platform/mediatek/vcodec/decoder/vdec/
A Dvdec_vp9_if.c37 unsigned long va; member
391 if (mem->va) in vp9_alloc_work_buf()
403 vsi->mv_buf.va = (unsigned long)mem->va; in vp9_alloc_work_buf()
409 if (mem->va) in vp9_alloc_work_buf()
420 vsi->seg_id_buf.va = (unsigned long)mem->va; in vp9_alloc_work_buf()
560 inst = mem.va; in vp9_alloc_inst()
571 if (mem.va) in vp9_free_inst()
667 inst->vsi->mv_buf.va = (unsigned long)inst->mv_buf.va; in vp9_reset()
672 inst->vsi->seg_id_buf.va = (unsigned long)inst->seg_id_buf.va; in vp9_reset()
765 if (mem->va) in vdec_vp9_deinit()
[all …]
/drivers/infiniband/sw/rxe/
A Drxe_mr.c252 void *va; in rxe_mr_copy_xarray() local
261 va = kmap_local_page(page); in rxe_mr_copy_xarray()
266 kunmap_local(va); in rxe_mr_copy_xarray()
283 u8 *va; in rxe_mr_copy_dma() local
296 kunmap_local(va); in rxe_mr_copy_dma()
434 u8 *va; in rxe_mr_flush_pmem_iova() local
451 kunmap_local(va); in rxe_mr_flush_pmem_iova()
492 u64 *va; in rxe_mr_do_atomic_op() local
537 kunmap_local(va); in rxe_mr_do_atomic_op()
546 u64 *va; in rxe_mr_do_atomic_write() local
[all …]
/drivers/video/fbdev/matrox/
A Dmatroxfb_base.h128 return readb(va.vaddr + offs); in mga_readb()
132 writeb(value, va.vaddr + offs); in mga_writeb()
136 writew(value, va.vaddr + offs); in mga_writew()
140 return readl(va.vaddr + offs); in mga_readl()
144 writel(value, va.vaddr + offs); in mga_writel()
156 iowrite32_rep(va.vaddr, src, len >> 2); in mga_memcpy_toio()
158 u_int32_t __iomem* addr = va.vaddr; in mga_memcpy_toio()
178 static inline void vaddr_add(vaddr_t* va, unsigned long offs) { in vaddr_add() argument
179 va->vaddr += offs; in vaddr_add()
182 static inline void __iomem* vaddr_va(vaddr_t va) { in vaddr_va() argument
[all …]
/drivers/tee/amdtee/
A Dshm_pool.c15 unsigned long va; in pool_op_alloc() local
22 va = __get_free_pages(GFP_KERNEL | __GFP_ZERO, order); in pool_op_alloc()
23 if (!va) in pool_op_alloc()
26 shm->kaddr = (void *)va; in pool_op_alloc()
27 shm->paddr = __psp_pa((void *)va); in pool_op_alloc()
33 free_pages(va, order); in pool_op_alloc()
/drivers/net/ethernet/ti/icssg/
A Dicssg_config.c168 void __iomem *smem = prueth->shram.va; in icssg_miig_queues_init()
258 p = emac->dram.va + MGR_R30_CMD_OFFSET; in emac_r30_cmd_init()
270 p = emac->dram.va + MGR_R30_CMD_OFFSET; in emac_r30_is_done()
303 bpool_cfg = emac->dram.va + BUFFER_POOL_0_ADDR_OFFSET; in prueth_fw_offload_buffer_setup()
341 rxq_ctx = emac->dram.va + HOST_RX_Q_EXP_CONTEXT_OFFSET; in prueth_fw_offload_buffer_setup()
362 rxq_ctx = emac->dram.va + DEFAULT_MSMC_Q_OFFSET; in prueth_fw_offload_buffer_setup()
390 bpool_cfg = emac->dram.va + BUFFER_POOL_0_ADDR_OFFSET; in prueth_emac_buffer_setup()
405 bpool_cfg = emac->dram.va + BUFFER_POOL_0_ADDR_OFFSET; in prueth_emac_buffer_setup()
448 rxq_ctx = emac->dram.va + DEFAULT_MSMC_Q_OFFSET; in prueth_emac_buffer_setup()
592 p = emac->dram.va + MGR_R30_CMD_OFFSET; in icssg_set_port_state()
[all …]
/drivers/gpu/drm/lima/
A Dlima_vm.c29 #define LIMA_PDE(va) (va >> LIMA_VM_PD_SHIFT) argument
30 #define LIMA_PTE(va) ((va & LIMA_VM_PT_MASK) >> LIMA_VM_PT_SHIFT) argument
31 #define LIMA_PBE(va) (va >> LIMA_VM_PB_SHIFT) argument
32 #define LIMA_BTE(va) ((va & LIMA_VM_BT_MASK) >> LIMA_VM_BT_SHIFT) argument
47 static int lima_vm_map_page(struct lima_vm *vm, dma_addr_t pa, u32 va) in lima_vm_map_page() argument
49 u32 pbe = LIMA_PBE(va); in lima_vm_map_page()
50 u32 bte = LIMA_BTE(va); in lima_vm_map_page()
81 list_for_each_entry(bo_va, &bo->va, list) { in lima_vm_bo_find()
138 list_add_tail(&bo_va->list, &bo->va); in lima_vm_bo_add()
/drivers/infiniband/hw/irdma/
A Dpuda.c147 if (!buf_mem.va) in irdma_puda_alloc_buf()
150 buf = buf_mem.va; in irdma_puda_alloc_buf()
153 if (!buf->mem.va) in irdma_puda_alloc_buf()
158 kfree(buf->mem.va); in irdma_puda_alloc_buf()
162 buf->buf_mem.va = buf_mem.va; in irdma_puda_alloc_buf()
168 kfree(buf_mem.va); in irdma_puda_alloc_buf()
182 kfree(buf->mem.va); in irdma_puda_dele_buf()
659 if (!rsrc->qpmem.va) in irdma_puda_qp_create()
957 kfree(vmem->va); in irdma_puda_dele_rsrc()
1024 if (!vmem->va) in irdma_puda_create_rsrc()
[all …]
A Dhmc.c347 if (!mem || !mem->va) { in irdma_finish_del_sd_reg()
352 mem->va = NULL; in irdma_finish_del_sd_reg()
492 if (!dma_mem.va) in irdma_add_sd_table_entry()
500 if (!vmem->va) { in irdma_add_sd_table_entry()
502 dma_mem.va, dma_mem.pa); in irdma_add_sd_table_entry()
503 dma_mem.va = NULL; in irdma_add_sd_table_entry()
576 if (!page->va) in irdma_add_pd_table_entry()
586 pd_addr = pd_table->pd_page_addr.va; in irdma_add_pd_table_entry()
640 pd_addr = pd_table->pd_page_addr.va; in irdma_remove_pd_bp()
647 if (!mem || !mem->va) in irdma_remove_pd_bp()
[all …]
A Dhw.c601 cqp->sq.va = NULL; in irdma_destroy_cqp()
616 vfree(aeq->mem.va); in irdma_destroy_virt_aeq()
901 va = (unsigned long)rf->obj_next.va; in irdma_obj_aligned_mem()
902 newva = va; in irdma_obj_aligned_mem()
905 extra = newva - va; in irdma_obj_aligned_mem()
906 memptr->va = (u8 *)va + extra; in irdma_obj_aligned_mem()
912 rf->obj_next.va = (u8 *)memptr->va + size; in irdma_obj_aligned_mem()
951 if (!cqp->sq.va) { in irdma_create_cqp()
1013 cqp->sq.va = NULL; in irdma_create_cqp()
1349 if (!aeq->mem.va) in irdma_create_virt_aeq()
[all …]
/drivers/firmware/efi/
A Dmokvar-table.c104 void *va = NULL; in efi_mokvar_table_init() local
136 if (va) in efi_mokvar_table_init()
137 early_memunmap(va, sizeof(*mokvar_entry)); in efi_mokvar_table_init()
138 va = early_memremap(efi.mokvar_table + cur_offset, sizeof(*mokvar_entry)); in efi_mokvar_table_init()
139 if (!va) { in efi_mokvar_table_init()
144 mokvar_entry = va; in efi_mokvar_table_init()
174 if (va) in efi_mokvar_table_init()
175 early_memunmap(va, sizeof(*mokvar_entry)); in efi_mokvar_table_init()
/drivers/tee/
A Dtee_shm_pool.c15 unsigned long va; in pool_op_gen_alloc() local
21 va = gen_pool_alloc_algo(genpool, s, gen_pool_first_fit_align, &data); in pool_op_gen_alloc()
22 if (!va) in pool_op_gen_alloc()
25 memset((void *)va, 0, s); in pool_op_gen_alloc()
26 shm->kaddr = (void *)va; in pool_op_gen_alloc()
27 shm->paddr = gen_pool_virt_to_phys(genpool, va); in pool_op_gen_alloc()
/drivers/infiniband/hw/usnic/
A Dusnic_uiom.c194 long unsigned va, size; in usnic_uiom_unmap_sorted_intervals() local
197 va = interval->start << PAGE_SHIFT; in usnic_uiom_unmap_sorted_intervals()
201 usnic_dbg("va 0x%lx size 0x%lx", va, PAGE_SIZE); in usnic_uiom_unmap_sorted_intervals()
202 iommu_unmap(pd->domain, va, PAGE_SIZE); in usnic_uiom_unmap_sorted_intervals()
203 va += PAGE_SIZE; in usnic_uiom_unmap_sorted_intervals()
220 vpn_start = (uiomr->va & PAGE_MASK) >> PAGE_SHIFT; in __usnic_uiom_reg_release()
251 long int va = uiomr->va & PAGE_MASK; in usnic_uiom_map_sorted_intervals() local
261 if ((va >> PAGE_SHIFT) < interval_node->start) in usnic_uiom_map_sorted_intervals()
266 va_start = va; in usnic_uiom_map_sorted_intervals()
286 va_start = va; in usnic_uiom_map_sorted_intervals()
[all …]
/drivers/gpu/drm/panthor/
A Dpanthor_mmu.c181 } va; member
1024 if (va != PANTHOR_VM_KERNEL_AUTO_VA && !IS_ALIGNED(va, vm_pgsz)) in panthor_vm_alloc_va()
1210 op_ctx->va.addr = va; in panthor_vm_prepare_map_op_ctx()
1316 op_ctx->va.addr = va; in panthor_vm_prepare_unmap_op_ctx()
1322 if (va != ALIGN(va, SZ_2M)) in panthor_vm_prepare_unmap_op_ctx()
1325 if (va + size != ALIGN(va + size, SZ_2M) && in panthor_vm_prepare_unmap_op_ctx()
1326 ALIGN(va + size, SZ_2M) != ALIGN(va, SZ_2M)) in panthor_vm_prepare_unmap_op_ctx()
1394 *bo_offset = vma->base.gem.offset + (va - vma->base.va.addr); in panthor_vm_get_bo_for_va()
2178 ret = drm_gpuvm_sm_map(&vm->base, vm, op->va.addr, op->va.range, in panthor_vm_exec_op()
2183 ret = drm_gpuvm_sm_unmap(&vm->base, vm, op->va.addr, op->va.range); in panthor_vm_exec_op()
[all …]
/drivers/net/wireless/ath/wil6210/
A Dtxrx_edma.c52 if (!wil->srings[i].va) in wil_find_free_sring()
65 if (!sring || !sring->va) in wil_sring_free()
75 sring->va = NULL; in wil_sring_free()
97 if (!sring->va) in wil_sring_alloc()
165 &ring->va[i].rx.enhanced; in wil_ring_alloc_skb_edma()
390 if (!ring->va) in wil_ring_alloc_desc_ring()
395 ring->edma_rx_swtail.va = in wil_ring_alloc_desc_ring()
410 ring->va = NULL; in wil_ring_alloc_desc_ring()
424 if (!ring->va) in wil_ring_free_edma()
476 ring->va = NULL; in wil_ring_free_edma()
[all …]
A Ddebug.c17 vaf.va = &args; in __wil_err()
33 vaf.va = &args; in __wil_err_ratelimited()
49 vaf.va = &args; in wil_dbg_ratelimited()
62 vaf.va = &args; in __wil_info()
75 vaf.va = &args; in wil_dbg_trace()
/drivers/remoteproc/
A Dpru_rproc.c656 void *va = NULL; in pru_d_da_to_va() local
670 va = (__force void *)(dram0.va + offset); in pru_d_da_to_va()
674 va = (__force void *)(dram1.va + offset); in pru_d_da_to_va()
678 va = (__force void *)(shrd_ram.va + offset); in pru_d_da_to_va()
681 return va; in pru_d_da_to_va()
694 void *va = NULL; in pru_i_da_to_va() local
719 va = (__force void *)(pru->mem_regions[PRU_IOMEM_IRAM].va + in pru_i_da_to_va()
723 return va; in pru_i_da_to_va()
742 void *va; in pru_da_to_va() local
745 va = pru_i_da_to_va(pru, da, len); in pru_da_to_va()
[all …]
/drivers/gpu/drm/i915/
A Dintel_memory_region.c37 u8 __iomem *va, int pagesize, in __iopagetest() argument
44 memset_io(va, value, pagesize); /* or GPF! */ in __iopagetest()
47 result[0] = ioread8(va); in __iopagetest()
48 result[1] = ioread8(va + byte); in __iopagetest()
49 result[2] = ioread8(va + pagesize - 1); in __iopagetest()
66 void __iomem *va; in iopagetest() local
70 va = ioremap_wc(mem->io.start + offset, PAGE_SIZE); in iopagetest()
71 if (!va) { in iopagetest()
79 err = __iopagetest(mem, va, PAGE_SIZE, val[i], offset, caller); in iopagetest()
83 err = __iopagetest(mem, va, PAGE_SIZE, ~val[i], offset, caller); in iopagetest()
[all …]
/drivers/media/platform/mediatek/vcodec/common/
A Dmtk_vcodec_util.c66 mem->va = dma_alloc_attrs(&plat_dev->dev, mem->size, &mem->dma_addr, in mtk_vcodec_mem_alloc()
68 if (!mem->va) { in mtk_vcodec_mem_alloc()
74 mtk_v4l2_debug(plat_dev, 3, "[%d] - va = %p dma = 0x%lx size = 0x%zx", id, mem->va, in mtk_vcodec_mem_alloc()
99 if (!mem->va) { in mtk_vcodec_mem_free()
106 mtk_v4l2_debug(plat_dev, 3, "[%d] - va = %p dma = 0x%lx size = 0x%zx", id, mem->va, in mtk_vcodec_mem_free()
109 dma_free_coherent(&plat_dev->dev, mem->size, mem->va, mem->dma_addr); in mtk_vcodec_mem_free()
110 mem->va = NULL; in mtk_vcodec_mem_free()

Completed in 84 milliseconds

12345678910>>...15