/linux/drivers/block/xen-blkback/ |
H A D | blkback.c | 650 struct page **unmap_pages) in xen_blkbk_unmap_prepare() argument 661 unmap_pages[invcount] = pages[i]->page; in xen_blkbk_unmap_prepare() 711 req->unmap, req->unmap_pages); in xen_blkbk_unmap_and_respond() 717 work->pages = req->unmap_pages; in xen_blkbk_unmap_and_respond() 736 struct page *unmap_pages[BLKIF_MAX_SEGMENTS_PER_REQUEST]; in xen_blkbk_unmap() local 744 unmap, unmap_pages); in xen_blkbk_unmap() 746 ret = gnttab_unmap_refs(unmap, NULL, unmap_pages, invcount); in xen_blkbk_unmap() 748 gnttab_page_cache_put(&ring->free_pages, unmap_pages, in xen_blkbk_unmap()
|
H A D | common.h | 358 struct page *unmap_pages[MAX_INDIRECT_SEGMENTS]; member
|
/linux/drivers/iommu/arm/arm-smmu/ |
H A D | qcom_iommu.c | 466 ret = ops->unmap_pages(ops, iova, pgsize, pgcount, gather); in qcom_iommu_unmap() 605 .unmap_pages = qcom_iommu_unmap,
|
H A D | arm-smmu.c | 1286 ret = ops->unmap_pages(ops, iova, pgsize, pgcount, iotlb_gather); in arm_smmu_unmap_pages() 1641 .unmap_pages = arm_smmu_unmap_pages,
|
/linux/drivers/iommu/ |
H A D | msm_iommu.c | 519 ret = priv->iop->unmap_pages(priv->iop, iova, pgsize, pgcount, gather); in msm_iommu_unmap() 700 .unmap_pages = msm_iommu_unmap,
|
H A D | ipmmu-vmsa.c | 684 return domain->iop->unmap_pages(domain->iop, iova, pgsize, pgcount, gather); in ipmmu_unmap() 890 .unmap_pages = ipmmu_unmap,
|
H A D | io-pgtable-dart.c | 384 .unmap_pages = dart_unmap_pages, in dart_alloc_pgtable()
|
H A D | sprd-iommu.c | 418 .unmap_pages = sprd_iommu_unmap,
|
H A D | apple-dart.c | 552 return ops->unmap_pages(ops, iova, pgsize, pgcount, gather); in apple_dart_unmap_pages() 994 .unmap_pages = apple_dart_unmap_pages,
|
H A D | mtk_iommu_v1.c | 587 .unmap_pages = mtk_iommu_v1_unmap,
|
H A D | mtk_iommu.c | 820 return dom->iop->unmap_pages(dom->iop, iova, pgsize, pgcount, gather); in mtk_iommu_unmap() 1027 .unmap_pages = mtk_iommu_unmap,
|
H A D | sun50i-iommu.c | 857 .unmap_pages = sun50i_iommu_unmap,
|
H A D | iommu.c | 2520 if (WARN_ON(!ops->unmap_pages || domain->pgsize_bitmap == 0UL)) in __iommu_unmap() 2547 unmapped_page = ops->unmap_pages(domain, iova, pgsize, count, iotlb_gather); in __iommu_unmap()
|
H A D | tegra-smmu.c | 1005 .unmap_pages = tegra_smmu_unmap,
|
H A D | virtio-iommu.c | 1075 .unmap_pages = viommu_unmap_pages,
|
H A D | exynos-iommu.c | 1485 .unmap_pages = exynos_iommu_unmap,
|
H A D | omap-iommu.c | 1742 .unmap_pages = omap_iommu_unmap,
|
/linux/drivers/gpu/drm/msm/ |
H A D | msm_iommu.c | 102 unmapped = ops->unmap_pages(ops, iova, pgsize, count, NULL); in msm_iommu_pagetable_unmap()
|
/linux/drivers/gpu/drm/panfrost/ |
H A D | panfrost_mmu.c | 372 unmapped_page = ops->unmap_pages(ops, iova, pgsize, pgcount, NULL); in panfrost_mmu_unmap()
|
/linux/drivers/iommu/amd/ |
H A D | iommu.c | 2754 r = (ops->unmap_pages) ? ops->unmap_pages(ops, iova, pgsize, pgcount, NULL) : 0; in amd_iommu_unmap_pages() 3037 .unmap_pages = amd_iommu_unmap_pages,
|
/linux/drivers/gpu/drm/panthor/ |
H A D | panthor_mmu.c | 920 unmapped_sz = ops->unmap_pages(ops, iova + offset, pgsize, pgcount, NULL); in panthor_vm_unmap_pages()
|
/linux/drivers/iommu/arm/arm-smmu-v3/ |
H A D | arm-smmu-v3.c | 3285 return ops->unmap_pages(ops, iova, pgsize, pgcount, gather); in arm_smmu_unmap_pages() 3651 .unmap_pages = arm_smmu_unmap_pages,
|
/linux/drivers/iommu/intel/ |
H A D | iommu.c | 4474 .unmap_pages = intel_iommu_unmap_pages,
|