| /drivers/gpu/drm/msm/ |
| A D | msm_gem_shrinker.c | 255 unsigned idx, unmapped = 0; in msm_gem_shrinker_vmap() local 258 for (idx = 0; lrus[idx] && unmapped < vmap_shrink_limit; idx++) { in msm_gem_shrinker_vmap() 259 unmapped += drm_gem_lru_scan(lrus[idx], in msm_gem_shrinker_vmap() 260 vmap_shrink_limit - unmapped, in msm_gem_shrinker_vmap() 266 *(unsigned long *)ptr += unmapped; in msm_gem_shrinker_vmap() 268 if (unmapped > 0) in msm_gem_shrinker_vmap() 269 trace_msm_gem_purge_vmaps(unmapped); in msm_gem_shrinker_vmap()
|
| A D | msm_gpu_trace.h | 143 TP_PROTO(u32 unmapped), 144 TP_ARGS(unmapped), 146 __field(u32, unmapped) 149 __entry->unmapped = unmapped; 151 TP_printk("Purging %u vmaps", __entry->unmapped)
|
| A D | msm_iommu.c | 108 ssize_t unmapped; in msm_iommu_pagetable_unmap() local 112 unmapped = ops->unmap_pages(ops, iova, pgsize, count, NULL); in msm_iommu_pagetable_unmap() 113 if (unmapped <= 0) { in msm_iommu_pagetable_unmap() 120 unmapped = PAGE_SIZE; in msm_iommu_pagetable_unmap() 123 iova += unmapped; in msm_iommu_pagetable_unmap() 124 size -= unmapped; in msm_iommu_pagetable_unmap()
|
| /drivers/iommu/amd/ |
| A D | io_pgtable_v2.c | 292 unsigned long unmapped = 0; in iommu_v2_unmap_pages() local 299 while (unmapped < size) { in iommu_v2_unmap_pages() 302 return unmapped; in iommu_v2_unmap_pages() 307 unmapped += unmap_size; in iommu_v2_unmap_pages() 310 return unmapped; in iommu_v2_unmap_pages()
|
| A D | io_pgtable.c | 404 unsigned long long unmapped; in iommu_v1_unmap_pages() local 411 unmapped = 0; in iommu_v1_unmap_pages() 413 while (unmapped < size) { in iommu_v1_unmap_pages() 422 return unmapped; in iommu_v1_unmap_pages() 426 unmapped += unmap_size; in iommu_v1_unmap_pages() 429 return unmapped; in iommu_v1_unmap_pages()
|
| /drivers/vfio/ |
| A D | vfio_iommu_type1.c | 1006 size_t unmapped = 0; in unmap_unpin_fast() local 1013 if (!unmapped) { in unmap_unpin_fast() 1021 *iova += unmapped; in unmap_unpin_fast() 1036 return unmapped; in unmap_unpin_fast() 1046 if (unmapped) { in unmap_unpin_slow() 1051 *iova += unmapped; in unmap_unpin_slow() 1054 return unmapped; in unmap_unpin_slow() 1090 size_t unmapped, len; in vfio_unmap_unpin() local 1118 if (!unmapped) { in vfio_unmap_unpin() 1406 unmapped = 0; in vfio_dma_do_unmap() [all …]
|
| /drivers/staging/media/ipu3/ |
| A D | ipu3-mmu.c | 383 size_t unmapped_page, unmapped = 0; in imgu_mmu_unmap() local 402 while (unmapped < size) { in imgu_mmu_unmap() 411 unmapped += unmapped_page; in imgu_mmu_unmap() 416 return unmapped; in imgu_mmu_unmap()
|
| /drivers/iommu/ |
| A D | virtio-iommu.c | 364 size_t unmapped = 0; in viommu_del_mappings() local 384 unmapped += mapping->iova.last - mapping->iova.start + 1; in viommu_del_mappings() 391 return unmapped; in viommu_del_mappings() 889 size_t unmapped; in viommu_unmap_pages() local 894 unmapped = viommu_del_mappings(vdomain, iova, iova + size - 1); in viommu_unmap_pages() 895 if (unmapped < size) in viommu_unmap_pages() 900 return unmapped; in viommu_unmap_pages() 906 .virt_end = cpu_to_le64(iova + unmapped - 1), in viommu_unmap_pages() 910 return ret ? 0 : unmapped; in viommu_unmap_pages()
|
| A D | io-pgtable-arm-v7s.c | 627 size_t unmapped = 0, ret; in arm_v7s_unmap_pages() local 637 unmapped += pgsize; in arm_v7s_unmap_pages() 641 return unmapped; in arm_v7s_unmap_pages()
|
| A D | dma-iommu.c | 815 size_t unmapped; in __iommu_dma_unmap() local 822 unmapped = iommu_unmap_fast(domain, dma_addr, size, &iotlb_gather); in __iommu_dma_unmap() 823 WARN_ON(unmapped != size); in __iommu_dma_unmap() 2031 size_t unmapped; in __iommu_dma_iova_unlink() local 2042 unmapped = iommu_unmap_fast(domain, addr, size, &iotlb_gather); in __iommu_dma_iova_unlink() 2043 WARN_ON(unmapped != size); in __iommu_dma_iova_unlink()
|
| A D | iommu.c | 2560 size_t unmapped_page, unmapped = 0; in __iommu_unmap() local 2590 while (unmapped < size) { in __iommu_unmap() 2593 pgsize = iommu_pgsize(domain, iova, iova, size - unmapped, &count); in __iommu_unmap() 2602 unmapped += unmapped_page; in __iommu_unmap() 2605 trace_unmap(orig_iova, size, unmapped); in __iommu_unmap() 2606 return unmapped; in __iommu_unmap()
|
| /drivers/iommu/iommufd/ |
| A D | vfio_compat.c | 209 unsigned long unmapped = 0; in iommufd_vfio_unmap_dma() local 228 rc = iopt_unmap_all(&ioas->iopt, &unmapped); in iommufd_vfio_unmap_dma() 245 &unmapped); in iommufd_vfio_unmap_dma() 247 unmap.size = unmapped; in iommufd_vfio_unmap_dma()
|
| A D | ioas.c | 350 unsigned long unmapped = 0; in iommufd_ioas_unmap() local 358 rc = iopt_unmap_all(&ioas->iopt, &unmapped); in iommufd_ioas_unmap() 367 &unmapped); in iommufd_ioas_unmap() 372 cmd->length = unmapped; in iommufd_ioas_unmap()
|
| A D | io_pagetable.c | 705 unsigned long last, unsigned long *unmapped) in iopt_unmap_iova_range() argument 787 if (unmapped) in iopt_unmap_iova_range() 788 *unmapped = unmapped_bytes; in iopt_unmap_iova_range() 803 unsigned long length, unsigned long *unmapped) in iopt_unmap_iova() argument 813 return iopt_unmap_iova_range(iopt, iova, iova_last, unmapped); in iopt_unmap_iova() 816 int iopt_unmap_all(struct io_pagetable *iopt, unsigned long *unmapped) in iopt_unmap_all() argument 820 rc = iopt_unmap_iova_range(iopt, 0, ULONG_MAX, unmapped); in iopt_unmap_all()
|
| A D | iommufd_private.h | 118 unsigned long length, unsigned long *unmapped); 119 int iopt_unmap_all(struct io_pagetable *iopt, unsigned long *unmapped);
|
| /drivers/media/platform/qcom/venus/ |
| A D | firmware.c | 178 size_t unmapped; in venus_shutdown_no_tz() local 199 unmapped = iommu_unmap(iommu, VENUS_FW_START_ADDR, mapped); in venus_shutdown_no_tz() 201 if (unmapped != mapped) in venus_shutdown_no_tz()
|
| /drivers/gpu/drm/etnaviv/ |
| A D | etnaviv_mmu.c | 19 size_t unmapped_page, unmapped = 0; in etnaviv_context_unmap() local 22 while (unmapped < size) { in etnaviv_context_unmap() 29 unmapped += unmapped_page; in etnaviv_context_unmap()
|
| /drivers/iommu/riscv/ |
| A D | iommu.c | 1252 size_t unmapped = 0; in riscv_iommu_unmap_pages() local 1255 while (unmapped < size) { in riscv_iommu_unmap_pages() 1258 return unmapped; in riscv_iommu_unmap_pages() 1262 return unmapped; in riscv_iommu_unmap_pages() 1272 unmapped += pte_size; in riscv_iommu_unmap_pages() 1275 return unmapped; in riscv_iommu_unmap_pages()
|
| /drivers/accel/amdxdna/ |
| A D | amdxdna_gem.h | 20 bool unmapped; member
|
| A D | amdxdna_gem.c | 127 if (!mapp->unmapped) { in amdxdna_hmm_invalidate() 129 mapp->unmapped = true; in amdxdna_hmm_invalidate() 150 if (!mapp->unmapped) { in amdxdna_hmm_unregister() 152 mapp->unmapped = true; in amdxdna_hmm_unregister()
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_umsch_mm.h | 112 uint32_t unmapped; member
|
| /drivers/dax/ |
| A D | Kconfig | 56 instance is created to access that unmapped-by-default address range.
|
| /drivers/remoteproc/ |
| A D | remoteproc_core.c | 1238 size_t unmapped; in rproc_resource_cleanup() local 1240 unmapped = iommu_unmap(rproc->domain, entry->da, entry->len); in rproc_resource_cleanup() 1241 if (unmapped != entry->len) { in rproc_resource_cleanup() 1244 unmapped); in rproc_resource_cleanup()
|
| /drivers/gpu/drm/ |
| A D | drm_gpusvm.c | 1314 if (flags.unmapped) { in drm_gpusvm_range_get_pages() 1564 range->flags.unmapped = true; in drm_gpusvm_range_set_unmapped()
|
| /drivers/net/ethernet/broadcom/bnx2x/ |
| A D | bnx2x_dcb.c | 351 u32 unmapped = (1 << MAX_PFC_PRIORITIES) - 1; /* all ones */ in bnx2x_dcbx_map_nw() local 359 unmapped &= ~(1 << ttp[i]); in bnx2x_dcbx_map_nw() 366 "cos %d extended with 0x%08x\n", i, unmapped); in bnx2x_dcbx_map_nw() 367 cos_params[i].pri_bitmask |= unmapped; in bnx2x_dcbx_map_nw()
|