/linux/tools/testing/selftests/dma/ |
H A D | dma_map_benchmark.c | 33 int granule = 1; in main() local 58 granule = atoi(optarg); in main() 95 if (granule < 1 || granule > 1024) { in main() 113 map.granule = granule; in main() 121 threads, seconds, node, dir[directions], granule); in main()
|
/linux/Documentation/virt/kvm/arm/ |
H A D | hypercalls.rst | 67 | | | | memory protection granule in bytes | 74 permissions. The size of the region is equal to the memory protection granule 100 memory protection granule advertised by ``ARM_SMCCC_KVM_FUNC_HYP_MEMINFO``. 125 region is equal to the memory protection granule advertised by
|
/linux/drivers/iommu/arm/arm-smmu/ |
H A D | qcom_iommu.c | 148 size_t granule, bool leaf, void *cookie) in qcom_iommu_tlb_inv_range_nosync() argument 164 iova += granule; in qcom_iommu_tlb_inv_range_nosync() 165 } while (s -= granule); in qcom_iommu_tlb_inv_range_nosync() 170 size_t granule, void *cookie) in qcom_iommu_tlb_flush_walk() argument 172 qcom_iommu_tlb_inv_range_nosync(iova, size, granule, false, cookie); in qcom_iommu_tlb_flush_walk() 177 unsigned long iova, size_t granule, in qcom_iommu_tlb_add_page() argument 180 qcom_iommu_tlb_inv_range_nosync(iova, granule, granule, true, cookie); in qcom_iommu_tlb_add_page()
|
H A D | arm-smmu.c | 279 size_t granule, void *cookie, int reg) in arm_smmu_tlb_inv_range_s1() argument 294 iova += granule; in arm_smmu_tlb_inv_range_s1() 295 } while (size -= granule); in arm_smmu_tlb_inv_range_s1() 301 iova += granule >> 12; in arm_smmu_tlb_inv_range_s1() 302 } while (size -= granule); in arm_smmu_tlb_inv_range_s1() 307 size_t granule, void *cookie, int reg) in arm_smmu_tlb_inv_range_s2() argument 322 iova += granule >> 12; in arm_smmu_tlb_inv_range_s2() 323 } while (size -= granule); in arm_smmu_tlb_inv_range_s2() 327 size_t granule, void *cookie) in arm_smmu_tlb_inv_walk_s1() argument 335 arm_smmu_tlb_inv_range_s1(iova, size, granule, cookie, in arm_smmu_tlb_inv_walk_s1() [all …]
|
/linux/drivers/iommu/ |
H A D | msm_iommu.c | 139 size_t granule, bool leaf, void *cookie) in __flush_iotlb_range() argument 159 iova += granule; in __flush_iotlb_range() 160 } while (temp_size -= granule); in __flush_iotlb_range() 171 size_t granule, void *cookie) in __flush_iotlb_walk() argument 173 __flush_iotlb_range(iova, size, granule, false, cookie); in __flush_iotlb_walk() 177 unsigned long iova, size_t granule, void *cookie) in __flush_iotlb_page() argument 179 __flush_iotlb_range(iova, granule, granule, true, cookie); in __flush_iotlb_page()
|
H A D | virtio-iommu.c | 389 unsigned long granule = 1UL << __ffs(vdomain->domain.pgsize_bitmap); in viommu_domain_map_identity() local 391 iova = ALIGN(iova, granule); in viommu_domain_map_identity() 392 limit = ALIGN_DOWN(limit + 1, granule) - 1; in viommu_domain_map_identity() 395 u64 resv_start = ALIGN_DOWN(resv->start, granule); in viommu_domain_map_identity() 396 u64 resv_end = ALIGN(resv->start + resv->length, granule) - 1; in viommu_domain_map_identity()
|
H A D | dma-iommu.c | 371 return cookie->iovad.granule; in cookie_msi_granule() 449 if (cookie->type == IOMMU_DMA_IOVA_COOKIE && cookie->iovad.granule) { in iommu_put_dma_cookie() 502 start += iovad->granule; in cookie_init_hw_msi_region() 703 if (1UL << order != iovad->granule || in iommu_dma_init_domain()
|
H A D | ipmmu-vmsa.c | 317 size_t granule, void *cookie) in ipmmu_tlb_flush() argument
|
/linux/kernel/dma/ |
H A D | map_benchmark.c | 39 int npages = map->bparam.granule; in map_benchmark_thread() 236 if (map->bparam.granule < 1 || map->bparam.granule > 1024) { in map_benchmark_ioctl()
|
/linux/include/linux/ |
H A D | map_benchmark.h | 29 __u32 granule; /* how many PAGE_SIZE will do map/unmap once a time */ member
|
/linux/drivers/gpu/drm/msm/ |
H A D | msm_iommu.c | 220 size_t granule, void *cookie) in msm_iommu_tlb_flush_walk() argument 230 pagetable->tlb->tlb_flush_walk(iova, size, granule, (void *)adreno_smmu->cookie); in msm_iommu_tlb_flush_walk() 236 unsigned long iova, size_t granule, void *cookie) in msm_iommu_tlb_add_page() argument
|
/linux/Documentation/arch/arm64/ |
H A D | memory-tagging-extension.rst | 19 allocation tag for each 16-byte granule in the physical address space. 197 4-bit tag per byte and correspond to a 16-byte MTE tag granule in the 200 **Note**: If ``addr`` is not aligned to a 16-byte granule, the kernel 247 in a byte. With the tag granule of 16 bytes, a 4K page requires 128
|
/linux/arch/powerpc/boot/dts/ |
H A D | microwatt.dts | 78 reservation-granule-size = <64>;
|
/linux/drivers/iommu/arm/arm-smmu-v3/ |
H A D | arm-smmu-v3.c | 2263 size_t granule, in __arm_smmu_tlb_inv_range() argument 2268 size_t inv_range = granule; in __arm_smmu_tlb_inv_range() 2292 cmd->tlbi.ttl = 4 - ((ilog2(granule) - 3) / (tg - 3)); in __arm_smmu_tlb_inv_range() 2333 size_t granule, bool leaf, in arm_smmu_tlb_inv_range_domain() argument 2350 __arm_smmu_tlb_inv_range(&cmd, iova, size, granule, smmu_domain); in arm_smmu_tlb_inv_range_domain() 2369 size_t granule, bool leaf, in arm_smmu_tlb_inv_range_asid() argument 2381 __arm_smmu_tlb_inv_range(&cmd, iova, size, granule, smmu_domain); in arm_smmu_tlb_inv_range_asid() 2385 unsigned long iova, size_t granule, in arm_smmu_tlb_inv_page_nosync() argument 2391 iommu_iotlb_gather_add_page(domain, gather, iova, granule); in arm_smmu_tlb_inv_page_nosync() 2395 size_t granule, void *cookie) in arm_smmu_tlb_inv_walk() argument [all …]
|
/linux/drivers/gpu/drm/panfrost/ |
H A D | panfrost_mmu.c | 393 static void mmu_tlb_flush_walk(unsigned long iova, size_t size, size_t granule, in mmu_tlb_flush_walk() argument
|
/linux/Documentation/admin-guide/mm/ |
H A D | nommu-mmap.rst | 118 granule but will only discard the excess if appropriately configured as
|
/linux/drivers/gpu/drm/panthor/ |
H A D | panthor_mmu.c | 1637 static void mmu_tlb_flush_walk(unsigned long iova, size_t size, size_t granule, void *cookie) in mmu_tlb_flush_walk() argument
|
/linux/arch/arm64/ |
H A D | Kconfig | 1330 Page size (translation granule) configuration.
|