| /linux/tools/testing/selftests/vfio/lib/ |
| H A D | iommu.c | 152 static int __vfio_iommu_unmap(int fd, u64 iova, u64 size, u32 flags, u64 *unmapped) in __vfio_iommu_unmap() argument 164 if (unmapped) in __vfio_iommu_unmap() 165 *unmapped = args.size; in __vfio_iommu_unmap() 171 u64 *unmapped) in vfio_iommu_unmap() argument 174 region->size, 0, unmapped); in vfio_iommu_unmap() 177 static int __iommufd_unmap(int fd, u64 iova, u64 length, u32 ioas_id, u64 *unmapped) in __iommufd_unmap() argument 189 if (unmapped) in __iommufd_unmap() 190 *unmapped = args.length; in __iommufd_unmap() 196 u64 *unmapped) in iommufd_unmap() argument 199 iommu->ioas_id, unmapped); in iommufd_unmap() [all …]
|
| /linux/tools/testing/selftests/vfio/ |
| H A D | vfio_dma_mapping_test.c | 141 u64 unmapped; in TEST_F() local 193 rc = __iommu_unmap(self->iommu, ®ion, &unmapped); in TEST_F() 195 ASSERT_EQ(unmapped, region.size); in TEST_F() 263 u64 unmapped; in TEST_F() local 269 rc = __iommu_unmap(self->iommu, region, &unmapped); in TEST_F() 271 ASSERT_EQ(unmapped, region->size); in TEST_F() 277 u64 unmapped; in TEST_F() local 283 rc = __iommu_unmap_all(self->iommu, &unmapped); in TEST_F() 285 ASSERT_EQ(unmapped, region->size); in TEST_F()
|
| /linux/include/trace/events/ |
| H A D | huge_memory.h | 58 int referenced, int none_or_zero, int status, int unmapped), 60 TP_ARGS(mm, folio, referenced, none_or_zero, status, unmapped), 68 __field(int, unmapped) 77 __entry->unmapped = unmapped; 80 TP_printk("mm=%p, scan_pfn=0x%lx, referenced=%d, none_or_zero=%d, status=%s, unmapped=%d", 86 __entry->unmapped)
|
| /linux/tools/testing/selftests/vfio/lib/include/libvfio/ |
| H A D | iommu.h | 45 int __iommu_unmap(struct iommu *iommu, struct dma_region *region, u64 *unmapped); 52 int __iommu_unmap_all(struct iommu *iommu, u64 *unmapped);
|
| /linux/drivers/accel/rocket/ |
| H A D | rocket_gem.c | 18 size_t unmapped; in rocket_gem_bo_free() local 22 unmapped = iommu_unmap(bo->domain->domain, bo->mm.start, bo->size); in rocket_gem_bo_free() 23 drm_WARN_ON(obj->dev, unmapped != bo->size); in rocket_gem_bo_free()
|
| /linux/drivers/staging/media/ipu3/ |
| H A D | ipu3-mmu.c | 383 size_t unmapped_page, unmapped = 0; in imgu_mmu_unmap() local 402 while (unmapped < size) { in imgu_mmu_unmap() 411 unmapped += unmapped_page; in imgu_mmu_unmap() 416 return unmapped; in imgu_mmu_unmap()
|
| /linux/drivers/vfio/ |
| H A D | vfio_iommu_type1.c | 1099 size_t unmapped = 0; in unmap_unpin_fast() local 1103 unmapped = iommu_unmap_fast(domain->domain, iova, len, in unmap_unpin_fast() 1106 if (!unmapped) { in unmap_unpin_fast() 1111 entry->len = unmapped; in unmap_unpin_fast() 1122 if (*unmapped_cnt >= VFIO_IOMMU_TLB_SYNC_MAX || !unmapped) { in unmap_unpin_fast() 1128 return unmapped; in unmap_unpin_fast() 1136 size_t unmapped = iommu_unmap(domain->domain, iova, len); in unmap_unpin_slow() local 1138 if (unmapped) { in unmap_unpin_slow() 1141 unmapped >> PAGE_SHIFT, in unmap_unpin_slow() 1145 return unmapped; in unmap_unpin_slow() [all …]
|
| /linux/drivers/gpu/drm/msm/ |
| H A D | msm_iommu.c | 110 ssize_t unmapped; in msm_iommu_pagetable_unmap() local 114 unmapped = ops->unmap_pages(ops, iova, pgsize, count, NULL); in msm_iommu_pagetable_unmap() 115 if (unmapped <= 0) { in msm_iommu_pagetable_unmap() 122 unmapped = PAGE_SIZE; in msm_iommu_pagetable_unmap() 125 iova += unmapped; in msm_iommu_pagetable_unmap() 126 size -= unmapped; in msm_iommu_pagetable_unmap()
|
| /linux/mm/ |
| H A D | migrate_device.c | 255 unsigned long addr = start, unmapped = 0; in migrate_vma_collect_pmd() local 322 if (unmapped) in migrate_vma_collect_pmd() 365 if (unmapped) in migrate_vma_collect_pmd() 473 unmapped++; in migrate_vma_collect_pmd() 485 if (unmapped) in migrate_vma_collect_pmd() 574 unsigned long unmapped = 0; in migrate_device_unmap() local 585 unmapped++; in migrate_device_unmap() 629 unmapped++; in migrate_device_unmap() 651 return unmapped; in migrate_device_unmap() 681 * and unmapped, chec [all...] |
| H A D | khugepaged.c | 1079 int referenced, int unmapped, struct collapse_control *cc) in collapse_huge_page() argument 1118 if (unmapped) { in collapse_huge_page() 1245 int node = NUMA_NO_NODE, unmapped = 0; in hpage_collapse_scan_pmd() local 1277 ++unmapped; in hpage_collapse_scan_pmd() 1279 unmapped <= khugepaged_max_ptes_swap) { in hpage_collapse_scan_pmd() 1384 (unmapped && referenced < HPAGE_PMD_NR / 2))) { in hpage_collapse_scan_pmd() 1393 unmapped, cc); in hpage_collapse_scan_pmd() 1399 none_or_zero, result, unmapped); in hpage_collapse_scan_pmd()
|
| /linux/drivers/iommu/ |
| H A D | virtio-iommu.c | 364 size_t unmapped = 0; in viommu_del_mappings() local 384 unmapped += mapping->iova.last - mapping->iova.start + 1; in viommu_del_mappings() 391 return unmapped; in viommu_del_mappings() 891 size_t unmapped; in viommu_unmap_pages() local 896 unmapped = viommu_del_mappings(vdomain, iova, iova + size - 1); in viommu_unmap_pages() 897 if (unmapped < size) in viommu_unmap_pages() 902 return unmapped; in viommu_unmap_pages() 908 .virt_end = cpu_to_le64(iova + unmapped - 1), in viommu_unmap_pages() 912 return ret ? 0 : unmapped; in viommu_unmap_pages()
|
| H A D | io-pgtable-arm-v7s.c | 627 size_t unmapped = 0, ret; in arm_v7s_unmap_pages() local 637 unmapped += pgsize; in arm_v7s_unmap_pages() 641 return unmapped; in arm_v7s_unmap_pages()
|
| H A D | dma-iommu.c | 820 size_t unmapped; in __iommu_dma_unmap() local 827 unmapped = iommu_unmap_fast(domain, dma_addr, size, &iotlb_gather); in __iommu_dma_unmap() 828 WARN_ON(unmapped != size); in __iommu_dma_unmap() 2032 size_t unmapped; in __iommu_dma_iova_unlink() local 2044 unmapped = iommu_unmap_fast(domain, addr, size, &iotlb_gather); in __iommu_dma_iova_unlink() 2045 WARN_ON(unmapped != size); in __iommu_dma_iova_unlink()
|
| /linux/drivers/iommu/iommufd/ |
| H A D | vfio_compat.c | 209 unsigned long unmapped = 0; in iommufd_vfio_unmap_dma() local 228 rc = iopt_unmap_all(&ioas->iopt, &unmapped); in iommufd_vfio_unmap_dma() 245 &unmapped); in iommufd_vfio_unmap_dma() 247 unmap.size = unmapped; in iommufd_vfio_unmap_dma()
|
| H A D | ioas.c | 344 unsigned long unmapped = 0; in iommufd_ioas_unmap() local 352 rc = iopt_unmap_all(&ioas->iopt, &unmapped); in iommufd_ioas_unmap() 361 &unmapped); in iommufd_ioas_unmap() 364 if (!unmapped) { in iommufd_ioas_unmap() 370 cmd->length = unmapped; in iommufd_ioas_unmap()
|
| H A D | io_pagetable.c | 742 unsigned long last, unsigned long *unmapped) in iopt_unmap_iova_range() argument 823 if (unmapped) in iopt_unmap_iova_range() 824 *unmapped = unmapped_bytes; in iopt_unmap_iova_range() 839 unsigned long length, unsigned long *unmapped) in iopt_unmap_iova() argument 849 return iopt_unmap_iova_range(iopt, iova, iova_last, unmapped); in iopt_unmap_iova() 852 int iopt_unmap_all(struct io_pagetable *iopt, unsigned long *unmapped) in iopt_unmap_all() argument 855 return iopt_unmap_iova_range(iopt, 0, ULONG_MAX, unmapped); in iopt_unmap_all()
|
| /linux/Documentation/features/vm/TLB/ |
| H A D | arch-support.txt | 4 # description: arch supports deferral of TLB flush until multiple pages are unmapped
|
| /linux/drivers/gpu/drm/etnaviv/ |
| H A D | etnaviv_mmu.c | 21 size_t unmapped_page, unmapped = 0; in etnaviv_context_unmap() local 24 while (unmapped < size) { in etnaviv_context_unmap() 31 unmapped += unmapped_page; in etnaviv_context_unmap()
|
| /linux/drivers/iommu/riscv/ |
| H A D | iommu.c | 1254 size_t unmapped = 0; in riscv_iommu_unmap_pages() local 1257 while (unmapped < size) { in riscv_iommu_unmap_pages() 1260 return unmapped; in riscv_iommu_unmap_pages() 1264 return unmapped; in riscv_iommu_unmap_pages() 1274 unmapped += pte_size; in riscv_iommu_unmap_pages() 1277 return unmapped; in riscv_iommu_unmap_pages()
|
| /linux/drivers/accel/amdxdna/ |
| H A D | amdxdna_gem.c | 129 if (!mapp->unmapped) { in amdxdna_hmm_invalidate() 131 mapp->unmapped = true; in amdxdna_hmm_invalidate() 152 if (!mapp->unmapped) { in amdxdna_hmm_unregister() 154 mapp->unmapped = true; in amdxdna_hmm_unregister()
|
| /linux/Documentation/userspace-api/ |
| H A D | tee.rst | 16 any longer it should be unmapped with munmap() to allow the reuse of
|
| /linux/Documentation/networking/device_drivers/ethernet/marvell/ |
| H A D | octeontx2.rst | 173 - Error due to operation of unmapped PF. 187 - Error due to unmapped slot. 237 - Receive packet on an unmapped PF. 249 - Error due to unmapped slot. 291 Rx on unmapped PF_FUNC
|
| /linux/arch/arm64/kvm/hyp/ |
| H A D | pgtable.c | 479 u64 *unmapped = ctx->arg; in hyp_unmap_walker() local 501 *unmapped += granule; in hyp_unmap_walker() 516 u64 unmapped = 0; in kvm_pgtable_hyp_unmap() local 519 .arg = &unmapped, in kvm_pgtable_hyp_unmap() 527 return unmapped; in kvm_pgtable_hyp_unmap()
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-class-rnbd-server | 32 When the device is unmapped by that client, the directory will be removed.
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_umsch_mm.h | 112 uint32_t unmapped; member
|