| /drivers/block/xen-blkback/ |
| A D | blkback.c | 650 struct page **unmap_pages) in xen_blkbk_unmap_prepare() argument 661 unmap_pages[invcount] = pages[i]->page; in xen_blkbk_unmap_prepare() 711 req->unmap, req->unmap_pages); in xen_blkbk_unmap_and_respond() 717 work->pages = req->unmap_pages; in xen_blkbk_unmap_and_respond() 736 struct page *unmap_pages[BLKIF_MAX_SEGMENTS_PER_REQUEST]; in xen_blkbk_unmap() local 744 unmap, unmap_pages); in xen_blkbk_unmap() 746 ret = gnttab_unmap_refs(unmap, NULL, unmap_pages, invcount); in xen_blkbk_unmap() 748 gnttab_page_cache_put(&ring->free_pages, unmap_pages, in xen_blkbk_unmap()
|
| A D | common.h | 358 struct page *unmap_pages[MAX_INDIRECT_SEGMENTS]; member
|
| /drivers/iommu/arm/arm-smmu/ |
| A D | qcom_iommu.c | 465 ret = ops->unmap_pages(ops, iova, pgsize, pgcount, gather); in qcom_iommu_unmap() 603 .unmap_pages = qcom_iommu_unmap,
|
| A D | arm-smmu.c | 1296 ret = ops->unmap_pages(ops, iova, pgsize, pgcount, iotlb_gather); in arm_smmu_unmap_pages() 1649 .unmap_pages = arm_smmu_unmap_pages,
|
| /drivers/iommu/ |
| A D | msm_iommu.c | 519 ret = priv->iop->unmap_pages(priv->iop, iova, pgsize, pgcount, gather); in msm_iommu_unmap() 699 .unmap_pages = msm_iommu_unmap,
|
| A D | io-pgtable-arm-v7s.c | 715 .unmap_pages = arm_v7s_unmap_pages, in arm_v7s_alloc_pgtable() 878 if (ops->unmap_pages(ops, iova, size, 1, NULL) != size) in arm_v7s_do_selftests()
|
| A D | ipmmu-vmsa.c | 685 return domain->iop->unmap_pages(domain->iop, iova, pgsize, pgcount, gather); in ipmmu_unmap() 890 .unmap_pages = ipmmu_unmap,
|
| A D | io-pgtable-arm.c | 951 .unmap_pages = arm_lpae_unmap_pages, in arm_lpae_alloc_pgtable() 1381 if (ops->unmap_pages(ops, iova, size, 1, NULL) != size) in arm_lpae_run_tests() 1412 if (ops->unmap_pages(ops, iova, size, 1, NULL) != size) in arm_lpae_run_tests()
|
| A D | io-pgtable-dart.c | 376 .unmap_pages = dart_unmap_pages, in dart_alloc_pgtable()
|
| A D | sprd-iommu.c | 419 .unmap_pages = sprd_iommu_unmap,
|
| A D | apple-dart.c | 555 return ops->unmap_pages(ops, iova, pgsize, pgcount, gather); in apple_dart_unmap_pages() 998 .unmap_pages = apple_dart_unmap_pages,
|
| A D | mtk_iommu_v1.c | 587 .unmap_pages = mtk_iommu_v1_unmap,
|
| A D | mtk_iommu.c | 818 return dom->iop->unmap_pages(dom->iop, iova, pgsize, pgcount, gather); in mtk_iommu_unmap() 1024 .unmap_pages = mtk_iommu_unmap,
|
| A D | sun50i-iommu.c | 858 .unmap_pages = sun50i_iommu_unmap,
|
| A D | s390-iommu.c | 1166 .unmap_pages = s390_iommu_unmap_pages, \
|
| A D | tegra-smmu.c | 1010 .unmap_pages = tegra_smmu_unmap,
|
| A D | virtio-iommu.c | 1102 .unmap_pages = viommu_unmap_pages,
|
| A D | rockchip-iommu.c | 1179 .unmap_pages = rk_iommu_unmap,
|
| A D | iommu.c | 2567 if (WARN_ON(!ops->unmap_pages || domain->pgsize_bitmap == 0UL)) in __iommu_unmap() 2594 unmapped_page = ops->unmap_pages(domain, iova, pgsize, count, iotlb_gather); in __iommu_unmap()
|
| A D | exynos-iommu.c | 1488 .unmap_pages = exynos_iommu_unmap,
|
| /drivers/iommu/amd/ |
| A D | io_pgtable_v2.c | 357 pgtable->pgtbl.ops.unmap_pages = iommu_v2_unmap_pages; in v2_alloc_pgtable()
|
| A D | io_pgtable.c | 550 pgtable->pgtbl.ops.unmap_pages = iommu_v1_unmap_pages; in v1_alloc_pgtable()
|
| A D | iommu.c | 2806 r = (ops->unmap_pages) ? ops->unmap_pages(ops, iova, pgsize, pgcount, NULL) : 0; in amd_iommu_unmap_pages() 3058 .unmap_pages = amd_iommu_unmap_pages,
|
| /drivers/gpu/drm/msm/ |
| A D | msm_iommu.c | 112 unmapped = ops->unmap_pages(ops, iova, pgsize, count, NULL); in msm_iommu_pagetable_unmap()
|
| /drivers/gpu/drm/panfrost/ |
| A D | panfrost_mmu.c | 475 unmapped_page = ops->unmap_pages(ops, iova, pgsize, pgcount, NULL); in panfrost_mmu_unmap()
|