| /linux/tools/testing/selftests/vfio/lib/ |
| H A D | iommu.c | 153 static int __vfio_iommu_unmap(int fd, u64 iova, u64 size, u32 flags, u64 *unmapped) in __vfio_iommu_unmap() argument 165 if (unmapped) in __vfio_iommu_unmap() 166 *unmapped = args.size; in __vfio_iommu_unmap() 172 u64 *unmapped) in vfio_iommu_unmap() argument 175 region->size, 0, unmapped); in vfio_iommu_unmap() 178 static int __iommufd_unmap(int fd, u64 iova, u64 length, u32 ioas_id, u64 *unmapped) in __iommufd_unmap() argument 190 if (unmapped) in __iommufd_unmap() 191 *unmapped = args.length; in __iommufd_unmap() 197 u64 *unmapped) in iommufd_unmap() argument 200 iommu->ioas_id, unmapped); in iommufd_unmap() [all …]
|
| /linux/tools/testing/selftests/vfio/ |
| H A D | vfio_dma_mapping_test.c | 142 u64 unmapped; in TEST_F() local 198 rc = __iommu_unmap(self->iommu, ®ion, &unmapped); in TEST_F() 200 ASSERT_EQ(unmapped, region.size); in TEST_F() 201 printf("Unmapped IOVA 0x%lx\n", region.iova); in TEST_F() 268 u64 unmapped; in TEST_F() local 274 rc = __iommu_unmap(self->iommu, region, &unmapped); in TEST_F() 276 ASSERT_EQ(unmapped, region->size); in TEST_F() 282 u64 unmapped; in TEST_F() local 288 rc = __iommu_unmap_all(self->iommu, &unmapped); in TEST_F() 290 ASSERT_EQ(unmapped, region->size); in TEST_F()
|
| /linux/include/trace/events/ |
| H A D | huge_memory.h | 57 int referenced, int none_or_zero, int status, int unmapped), 59 TP_ARGS(mm, folio, referenced, none_or_zero, status, unmapped), 67 __field(int, unmapped) 76 __entry->unmapped = unmapped; 79 TP_printk("mm=%p, scan_pfn=0x%lx, referenced=%d, none_or_zero=%d, status=%s, unmapped=%d", 85 __entry->unmapped)
|
| /linux/drivers/scsi/lpfc/ |
| H A D | lpfc_disc.h | 229 * nodes transition from the unmapped to the mapped list. 248 * and put on the unmapped list. For ADISC processing, the node is taken off 249 * the ADISC list and placed on either the mapped or unmapped list (depending 250 * on its previous state). Once on the unmapped list, a PRLI is issued and the 253 * node, the node is taken off the unmapped list. The binding list is checked 255 * assignment is unsuccessful, the node is left on the unmapped list. If 260 * For a Link Down, all nodes on the ADISC, PLOGI, unmapped or mapped 265 * For a Link Up or RSCN, all nodes will move from the mapped / unmapped lists 282 * unmapped lists.
|
| /linux/mm/damon/ |
| H A D | vaddr.c | 108 * Find three regions separated by two biggest unmapped regions 114 * separated by the two biggest unmapped regions in the space. Please refer to 203 * is actually mapped to the memory and accessed, monitoring the unmapped 208 * with the noise by simply identifying the unmapped areas as a region that 210 * unmapped areas inside will make the adaptive mechanism quite complex. That 211 * said, too huge unmapped areas inside the monitoring target should be removed 216 * between the three regions are the two biggest unmapped areas in the given 218 * end of the mappings and the two biggest unmapped areas of the address space. 227 * region and the stack will be two biggest unmapped regions. Because these 229 * two biggest unmapped regions will be sufficient to make a trade-off. [all …]
|
| /linux/include/drm/ |
| H A D | drm_gpusvm.h | 112 * @unmapped: Flag indicating if the pages has been unmapped 113 * @partial_unmap: Flag indicating if the pages has been partially unmapped 124 u16 unmapped : 1; member 142 * @flags.unmapped: Flag indicating if the range has been unmapped 143 * @flags.partial_unmap: Flag indicating if the range has been partially unmapped
|
| /linux/drivers/vfio/ |
| H A D | vfio_iommu_type1.c | 1099 size_t unmapped = 0; in unmap_unpin_fast() local 1103 unmapped = iommu_unmap_fast(domain->domain, iova, len, in unmap_unpin_fast() 1106 if (!unmapped) { in unmap_unpin_fast() 1111 entry->len = unmapped; in unmap_unpin_fast() 1122 if (*unmapped_cnt >= VFIO_IOMMU_TLB_SYNC_MAX || !unmapped) { in unmap_unpin_fast() 1128 return unmapped; in unmap_unpin_fast() 1136 size_t unmapped = iommu_unmap(domain->domain, iova, len); in unmap_unpin_slow() local 1138 if (unmapped) { in unmap_unpin_slow() 1141 unmapped >> PAGE_SHIFT, in unmap_unpin_slow() 1145 return unmapped; in unmap_unpin_slow() [all …]
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_exec.c | 35 * mapping. If such memory backed mappings are unmapped the kernel will make 38 * will result in those memory backed mappings being unmapped first. 48 * backed mappings being mapped and unmapped, either within a single or multiple 54 * of the previously unmapped sparse mapping within the same VM_BIND ioctl 56 * range of the previously unmapped sparse mapping within the same VM_BIND
|
| /linux/tools/testing/selftests/vfio/lib/include/libvfio/ |
| H A D | iommu.h | 45 int __iommu_unmap(struct iommu *iommu, struct dma_region *region, u64 *unmapped); 52 int __iommu_unmap_all(struct iommu *iommu, u64 *unmapped);
|
| /linux/drivers/media/platform/chips-media/wave5/ |
| H A D | wave5-vdi.c | 91 dev_err(vpu_dev->dev, "%s: unable to clear unmapped buffer\n", __func__); in wave5_vdi_clear_memory() 103 dev_err(vpu_dev->dev, "%s: unable to write to unmapped buffer\n", __func__); in wave5_vdi_write_memory() 142 dev_err(vpu_dev->dev, "%s: requested free of unmapped buffer\n", __func__); in wave5_vdi_free_dma_memory()
|
| /linux/mm/ |
| H A D | mseal.c | 19 * mseal() disallows an input range which contain unmapped ranges (VMA holes). 21 * It disallows unmapped regions from start to end whether they exist at the 31 * Does the [start, end) range contain any unmapped memory?
|
| H A D | migrate_device.c | 255 unsigned long addr = start, unmapped = 0; in migrate_vma_collect_pmd() local 322 if (unmapped) in migrate_vma_collect_pmd() 365 if (unmapped) in migrate_vma_collect_pmd() 473 unmapped++; in migrate_vma_collect_pmd() 485 if (unmapped) in migrate_vma_collect_pmd() 574 unsigned long unmapped = 0; in migrate_device_unmap() local 585 unmapped++; in migrate_device_unmap() 629 unmapped++; in migrate_device_unmap() 651 return unmapped; in migrate_device_unmap() 681 * and unmapped, check whether each page is pinned or not. Pages that aren't [all …]
|
| /linux/include/linux/dma/ |
| H A D | k3-psil.h | 56 * @flow_start: PKDMA flow range start of mapped channel. Unmapped 58 * @flow_num: PKDMA flow count of mapped channel. Unmapped channels
|
| /linux/drivers/iommu/ |
| H A D | virtio-iommu.c | 359 * On success, returns the number of unmapped bytes 364 size_t unmapped = 0; in viommu_del_mappings() local 384 unmapped += mapping->iova.last - mapping->iova.start + 1; in viommu_del_mappings() 391 return unmapped; in viommu_del_mappings() 891 size_t unmapped; in viommu_unmap_pages() local 896 unmapped = viommu_del_mappings(vdomain, iova, iova + size - 1); in viommu_unmap_pages() 897 if (unmapped < size) in viommu_unmap_pages() 902 return unmapped; in viommu_unmap_pages() 908 .virt_end = cpu_to_le64(iova + unmapped - 1), in viommu_unmap_pages() 912 return ret ? 0 : unmapped; in viommu_unmap_pages()
|
| /linux/drivers/staging/media/ipu3/ |
| H A D | ipu3-mmu.c | 383 size_t unmapped_page, unmapped = 0; in imgu_mmu_unmap() local 402 while (unmapped < size) { in imgu_mmu_unmap() 407 dev_dbg(mmu->dev, "unmapped: iova 0x%lx size 0x%zx\n", in imgu_mmu_unmap() 411 unmapped += unmapped_page; in imgu_mmu_unmap() 416 return unmapped; in imgu_mmu_unmap()
|
| /linux/mm/damon/tests/ |
| H A D | vaddr-kunit.h | 45 * regions should not include the two biggest unmapped areas in the original 48 * Because these two unmapped areas are very huge but obviously never accessed, 53 * unmapped areas. After that, based on the information, it constructs the 61 * and end with 305. The process also has three unmapped areas, 25-200, 63 * unmapped areas, and thus it should be converted to three regions of 10-25,
|
| /linux/lib/ |
| H A D | devres.c | 70 * Managed ioremap(). Map is automatically unmapped on driver detach. 85 * Managed ioremap_uc(). Map is automatically unmapped on driver detach. 100 * Managed ioremap_wc(). Map is automatically unmapped on driver detach. 275 * Managed ioport_map(). Map is automatically unmapped on driver
|
| /linux/Documentation/translations/zh_CN/mm/damon/ |
| H A D | design.rst | 57 <BIG UNMAPPED REGION 1> 61 <BIG UNMAPPED REGION 2>
|
| /linux/arch/hexagon/include/asm/ |
| H A D | processor.h | 47 * Apparently the convention is that ld.so will ask for "unmapped" private 52 * you have to kick the "unmapped" base requests higher up.
|
| /linux/drivers/irqchip/ |
| H A D | irq-ti-sci-inta.c | 90 * unmapped event sources. 91 * Unmapped Events are not part of the Global Event Map and 95 * generating Unmapped Event, we must use the INTA's TI-SCI 127 * For devices sending Unmapped Events we must use the INTA's TI-SCI in ti_sci_inta_get_dev_id() 619 count = of_count_phandle_with_args(node, "ti,unmapped-event-sources", NULL); in ti_sci_inta_get_unmapped_sources() 630 of_for_each_phandle(&it, err, node, "ti,unmapped-event-sources", NULL, 0) { in ti_sci_inta_get_unmapped_sources()
|
| /linux/Documentation/mm/damon/ |
| H A D | design.rst | 96 mapped to the physical memory and accessed. Thus, tracking the unmapped 100 cases. That said, too huge unmapped areas inside the monitoring target should 105 gaps between the three regions are the two biggest unmapped areas in the given 106 address space. The two biggest unmapped areas would be the gap between the 113 <BIG UNMAPPED REGION 1> 117 <BIG UNMAPPED REGION 2> 268 virtual memory could be dynamically mapped and unmapped. Physical memory could 696 - unmapped 697 - Applied to pages that unmapped.
|
| /linux/drivers/iommu/iommufd/ |
| H A D | io_pagetable.c | 738 unsigned long last, unsigned long *unmapped) in iopt_unmap_iova_range() argument 819 if (unmapped) in iopt_unmap_iova_range() 820 *unmapped = unmapped_bytes; in iopt_unmap_iova_range() 829 * @unmapped: Return number of bytes unmapped 835 unsigned long length, unsigned long *unmapped) in iopt_unmap_iova() argument 845 return iopt_unmap_iova_range(iopt, iova, iova_last, unmapped); in iopt_unmap_iova() 848 int iopt_unmap_all(struct io_pagetable *iopt, unsigned long *unmapped) in iopt_unmap_all() argument 851 return iopt_unmap_iova_range(iopt, 0, ULONG_MAX, unmapped); in iopt_unmap_all() 965 * will be unmapped from the domain. The domain must already be removed from the
|
| H A D | ioas.c | 344 unsigned long unmapped = 0; in iommufd_ioas_unmap() local 352 rc = iopt_unmap_all(&ioas->iopt, &unmapped); in iommufd_ioas_unmap() 361 &unmapped); in iommufd_ioas_unmap() 364 if (!unmapped) { in iommufd_ioas_unmap() 370 cmd->length = unmapped; in iommufd_ioas_unmap()
|
| /linux/tools/testing/selftests/powerpc/signal/ |
| H A D | sigreturn_vdso.c | 106 printf("Unmapped VDSO\n"); in test_sigreturn_vdso() 119 printf("Signal delivered OK with VDSO unmapped\n"); in test_sigreturn_vdso()
|
| /linux/Documentation/networking/device_drivers/ethernet/marvell/ |
| H A D | octeontx2.rst | 173 - Error due to operation of unmapped PF. 187 - Error due to unmapped slot. 237 - Receive packet on an unmapped PF. 249 - Error due to unmapped slot. 291 Rx on unmapped PF_FUNC
|