| /linux/tools/testing/selftests/vfio/lib/ |
| H A D | iommu.c | 152 static int __vfio_iommu_unmap(int fd, u64 iova, u64 size, u32 flags, u64 *unmapped) in __vfio_iommu_unmap() argument 164 if (unmapped) in __vfio_iommu_unmap() 165 *unmapped = args.size; in __vfio_iommu_unmap() 171 u64 *unmapped) in vfio_iommu_unmap() argument 174 region->size, 0, unmapped); in vfio_iommu_unmap() 177 static int __iommufd_unmap(int fd, u64 iova, u64 length, u32 ioas_id, u64 *unmapped) in __iommufd_unmap() argument 189 if (unmapped) in __iommufd_unmap() 190 *unmapped = args.length; in __iommufd_unmap() 196 u64 *unmapped) in iommufd_unmap() argument 199 iommu->ioas_id, unmapped); in iommufd_unmap() [all …]
|
| /linux/tools/testing/selftests/vfio/ |
| H A D | vfio_dma_mapping_test.c | 141 u64 unmapped; in TEST_F() local 193 rc = __iommu_unmap(self->iommu, ®ion, &unmapped); in TEST_F() 195 ASSERT_EQ(unmapped, region.size); in TEST_F() 196 printf("Unmapped IOVA 0x%lx\n", region.iova); in TEST_F() 263 u64 unmapped; in TEST_F() local 269 rc = __iommu_unmap(self->iommu, region, &unmapped); in TEST_F() 271 ASSERT_EQ(unmapped, region->size); in TEST_F() 277 u64 unmapped; in TEST_F() local 283 rc = __iommu_unmap_all(self->iommu, &unmapped); in TEST_F() 285 ASSERT_EQ(unmapped, region->size); in TEST_F()
|
| H A D | vfio_pci_driver_test.c | 152 * accesses to unmapped IOVAs as errors. in TEST_F() 166 * accesses to unmapped IOVAs as errors. in TEST_F()
|
| /linux/drivers/accel/rocket/ |
| H A D | rocket_gem.c | 18 size_t unmapped; in rocket_gem_bo_free() local 22 unmapped = iommu_unmap(bo->domain->domain, bo->mm.start, bo->size); in rocket_gem_bo_free() 23 drm_WARN_ON(obj->dev, unmapped != bo->size); in rocket_gem_bo_free()
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_exec.c | 35 * mapping. If such memory backed mappings are unmapped the kernel will make 38 * will result in those memory backed mappings being unmapped first. 48 * backed mappings being mapped and unmapped, either within a single or multiple 54 * of the previously unmapped sparse mapping within the same VM_BIND ioctl 56 * range of the previously unmapped sparse mapping within the same VM_BIND
|
| /linux/tools/testing/selftests/vfio/lib/include/libvfio/ |
| H A D | iommu.h | 45 int __iommu_unmap(struct iommu *iommu, struct dma_region *region, u64 *unmapped); 52 int __iommu_unmap_all(struct iommu *iommu, u64 *unmapped);
|
| /linux/mm/ |
| H A D | mseal.c | 19 * mseal() disallows an input range which contain unmapped ranges (VMA holes). 21 * It disallows unmapped regions from start to end whether they exist at the 31 * Does the [start, end) range contain any unmapped memory?
|
| H A D | migrate_device.c | 249 unsigned long addr = start, unmapped = 0; in migrate_vma_collect_pmd() local 316 if (unmapped) in migrate_vma_collect_pmd() 359 if (unmapped) in migrate_vma_collect_pmd() 467 unmapped++; in migrate_vma_collect_pmd() 479 if (unmapped) in migrate_vma_collect_pmd() 568 unsigned long unmapped = 0; in migrate_device_unmap() local 579 unmapped++; in migrate_device_unmap() 623 unmapped++; in migrate_device_unmap() 645 return unmapped; in migrate_device_unmap() 675 * and unmapped, check whether each page is pinned or not. Pages that aren't [all …]
|
| /linux/include/linux/dma/ |
| H A D | k3-psil.h | 56 * @flow_start: PKDMA flow range start of mapped channel. Unmapped 58 * @flow_num: PKDMA flow count of mapped channel. Unmapped channels
|
| /linux/drivers/iommu/ |
| H A D | virtio-iommu.c | 359 * On success, returns the number of unmapped bytes 364 size_t unmapped = 0; in viommu_del_mappings() local 384 unmapped += mapping->iova.last - mapping->iova.start + 1; in viommu_del_mappings() 391 return unmapped; in viommu_del_mappings() 891 size_t unmapped; in viommu_unmap_pages() local 896 unmapped = viommu_del_mappings(vdomain, iova, iova + size - 1); in viommu_unmap_pages() 897 if (unmapped < size) in viommu_unmap_pages() 902 return unmapped; in viommu_unmap_pages() 908 .virt_end = cpu_to_le64(iova + unmapped - 1), in viommu_unmap_pages() 912 return ret ? 0 : unmapped; in viommu_unmap_pages()
|
| /linux/drivers/staging/media/ipu3/ |
| H A D | ipu3-mmu.c | 383 size_t unmapped_page, unmapped = 0; in imgu_mmu_unmap() local 402 while (unmapped < size) { in imgu_mmu_unmap() 407 dev_dbg(mmu->dev, "unmapped: iova 0x%lx size 0x%zx\n", in imgu_mmu_unmap() 411 unmapped += unmapped_page; in imgu_mmu_unmap() 416 return unmapped; in imgu_mmu_unmap()
|
| /linux/Documentation/translations/zh_CN/mm/damon/ |
| H A D | design.rst | 57 <BIG UNMAPPED REGION 1> 61 <BIG UNMAPPED REGION 2>
|
| /linux/lib/ |
| H A D | devres.c | 70 * Managed ioremap(). Map is automatically unmapped on driver detach. 85 * Managed ioremap_uc(). Map is automatically unmapped on driver detach. 100 * Managed ioremap_wc(). Map is automatically unmapped on driver detach. 275 * Managed ioport_map(). Map is automatically unmapped on driver
|
| /linux/arch/hexagon/include/asm/ |
| H A D | processor.h | 47 * Apparently the convention is that ld.so will ask for "unmapped" private 52 * you have to kick the "unmapped" base requests higher up.
|
| /linux/drivers/irqchip/ |
| H A D | irq-ti-sci-inta.c | 90 * unmapped event sources. 91 * Unmapped Events are not part of the Global Event Map and 95 * generating Unmapped Event, we must use the INTA's TI-SCI 127 * For devices sending Unmapped Events we must use the INTA's TI-SCI in ti_sci_inta_get_dev_id() 619 count = of_count_phandle_with_args(node, "ti,unmapped-event-sources", NULL); in ti_sci_inta_get_unmapped_sources() 630 of_for_each_phandle(&it, err, node, "ti,unmapped-event-sources", NULL, 0) { in ti_sci_inta_get_unmapped_sources()
|
| /linux/tools/testing/selftests/powerpc/signal/ |
| H A D | sigreturn_vdso.c | 106 printf("Unmapped VDSO\n"); in test_sigreturn_vdso() 119 printf("Signal delivered OK with VDSO unmapped\n"); in test_sigreturn_vdso()
|
| /linux/drivers/iommu/iommufd/ |
| H A D | ioas.c | 344 unsigned long unmapped = 0; in iommufd_ioas_unmap() local 352 rc = iopt_unmap_all(&ioas->iopt, &unmapped); in iommufd_ioas_unmap() 361 &unmapped); in iommufd_ioas_unmap() 364 if (!unmapped) { in iommufd_ioas_unmap() 370 cmd->length = unmapped; in iommufd_ioas_unmap()
|
| /linux/drivers/gpu/drm/msm/ |
| H A D | msm_iommu.c | 110 ssize_t unmapped; in msm_iommu_pagetable_unmap() local 114 unmapped = ops->unmap_pages(ops, iova, pgsize, count, NULL); in msm_iommu_pagetable_unmap() 115 if (unmapped <= 0) { in msm_iommu_pagetable_unmap() 122 unmapped = PAGE_SIZE; in msm_iommu_pagetable_unmap() 125 iova += unmapped; in msm_iommu_pagetable_unmap() 126 size -= unmapped; in msm_iommu_pagetable_unmap()
|
| /linux/arch/mips/include/asm/octeon/ |
| H A D | cvmx-address.h | 108 /* mapped or unmapped virtual address */ 120 /* mapped or unmapped virtual address */ 128 * physical address accessed through xkphys unmapped virtual
|
| /linux/arch/powerpc/platforms/powernv/ |
| H A D | vas-debug.c | 36 /* ensure window is not unmapped */ in info_show() 63 /* ensure window is not unmapped */ in hvwc_show()
|
| /linux/tools/testing/selftests/namespaces/ |
| H A D | listns_efault_test.c | 29 * into unmapped memory), listns() returns EINVAL. 54 * - Second page: will be unmapped to trigger EFAULT in TEST() 67 * unmapped page. in TEST() 93 * 2. Fail with EFAULT when trying to write the second ID (unmapped page) in TEST()
|
| /linux/Documentation/devicetree/bindings/memory-controllers/ |
| H A D | st,stm32mp25-omm.yaml | 95 - 000: OCTOSPI1 (256 Mbytes), OCTOSPI2 unmapped 99 - 1xx: OCTOSPI1 unmapped, OCTOSPI2 (256 Mbytes)
|
| /linux/include/linux/ |
| H A D | highmem-internal.h | 262 * @__addr: Virtual address to be unmapped 268 * Mappings should be unmapped in the reverse order that they were mapped.
|
| /linux/Documentation/features/vm/TLB/ |
| H A D | arch-support.txt | 4 # description: arch supports deferral of TLB flush until multiple pages are unmapped
|
| /linux/tools/lib/perf/ |
| H A D | mmap.c | 184 * Check if event was unmapped due to a POLLHUP/POLLERR. in perf_mmap__read_init() 201 * Check if event was unmapped due to a POLLHUP/POLLERR. in perf_mmap__read_done() 277 * Check if event was unmapped due to a POLLHUP/POLLERR. in perf_mmap__read_event()
|