1b77cf11fSRob Herring /* SPDX-License-Identifier: GPL-2.0 */ 2b77cf11fSRob Herring #ifndef __IO_PGTABLE_H 3b77cf11fSRob Herring #define __IO_PGTABLE_H 4a2d3a382SWill Deacon 5b77cf11fSRob Herring #include <linux/bitops.h> 6a2d3a382SWill Deacon #include <linux/iommu.h> 7b77cf11fSRob Herring 8b77cf11fSRob Herring /* 9b77cf11fSRob Herring * Public API for use by IOMMU drivers 10b77cf11fSRob Herring */ 11b77cf11fSRob Herring enum io_pgtable_fmt { 12b77cf11fSRob Herring ARM_32_LPAE_S1, 13b77cf11fSRob Herring ARM_32_LPAE_S2, 14b77cf11fSRob Herring ARM_64_LPAE_S1, 15b77cf11fSRob Herring ARM_64_LPAE_S2, 16b77cf11fSRob Herring ARM_V7S, 17d08d42deSRob Herring ARM_MALI_LPAE, 18c9b258c6SSuravee Suthikulpanit AMD_IOMMU_V1, 19aaac38f6SVasant Hegde AMD_IOMMU_V2, 20892384cdSSven Peter APPLE_DART, 21dc09fe1cSSven Peter APPLE_DART2, 22b77cf11fSRob Herring IO_PGTABLE_NUM_FMTS, 23b77cf11fSRob Herring }; 24b77cf11fSRob Herring 25b77cf11fSRob Herring /** 26298f7889SWill Deacon * struct iommu_flush_ops - IOMMU callbacks for TLB and page table management. 27b77cf11fSRob Herring * 28b77cf11fSRob Herring * @tlb_flush_all: Synchronously invalidate the entire TLB context. 293445545bSWill Deacon * @tlb_flush_walk: Synchronously invalidate all intermediate TLB state 303445545bSWill Deacon * (sometimes referred to as the "walk cache") for a virtual 313445545bSWill Deacon * address range. 32abfd6fe0SWill Deacon * @tlb_add_page: Optional callback to queue up leaf TLB invalidation for a 333951c41aSWill Deacon * single page. IOMMUs that cannot batch TLB invalidation 343951c41aSWill Deacon * operations efficiently will typically issue them here, but 353951c41aSWill Deacon * others may decide to update the iommu_iotlb_gather structure 36aae4c8e2STom Murphy * and defer the invalidation until iommu_iotlb_sync() instead. 37b77cf11fSRob Herring * 38b77cf11fSRob Herring * Note that these can all be called in atomic context and must therefore 39b77cf11fSRob Herring * not block. 40b77cf11fSRob Herring */ 41298f7889SWill Deacon struct iommu_flush_ops { 42b77cf11fSRob Herring void (*tlb_flush_all)(void *cookie); 433445545bSWill Deacon void (*tlb_flush_walk)(unsigned long iova, size_t size, size_t granule, 443445545bSWill Deacon void *cookie); 453951c41aSWill Deacon void (*tlb_add_page)(struct iommu_iotlb_gather *gather, 463951c41aSWill Deacon unsigned long iova, size_t granule, void *cookie); 47b77cf11fSRob Herring }; 48b77cf11fSRob Herring 49b77cf11fSRob Herring /** 50b77cf11fSRob Herring * struct io_pgtable_cfg - Configuration data for a set of page tables. 51b77cf11fSRob Herring * 52b77cf11fSRob Herring * @quirks: A bitmap of hardware quirks that require some special 53b77cf11fSRob Herring * action by the low-level page table allocator. 54b77cf11fSRob Herring * @pgsize_bitmap: A bitmap of page sizes supported by this set of page 55b77cf11fSRob Herring * tables. 56b77cf11fSRob Herring * @ias: Input address (iova) size, in bits. 57b77cf11fSRob Herring * @oas: Output address (paddr) size, in bits. 584f41845bSWill Deacon * @coherent_walk A flag to indicate whether or not page table walks made 594f41845bSWill Deacon * by the IOMMU are coherent with the CPU caches. 60b77cf11fSRob Herring * @tlb: TLB management callbacks for this set of tables. 61b77cf11fSRob Herring * @iommu_dev: The device representing the DMA configuration for the 62b77cf11fSRob Herring * page table walker. 63b77cf11fSRob Herring */ 64b77cf11fSRob Herring struct io_pgtable_cfg { 65b77cf11fSRob Herring /* 66b77cf11fSRob Herring * IO_PGTABLE_QUIRK_ARM_NS: (ARM formats) Set NS and NSTABLE bits in 67b77cf11fSRob Herring * stage 1 PTEs, for hardware which insists on validating them 68b77cf11fSRob Herring * even in non-secure state where they should normally be ignored. 69b77cf11fSRob Herring * 70b77cf11fSRob Herring * IO_PGTABLE_QUIRK_NO_PERMS: Ignore the IOMMU_READ, IOMMU_WRITE and 71b77cf11fSRob Herring * IOMMU_NOEXEC flags and map everything with full access, for 72b77cf11fSRob Herring * hardware which does not implement the permissions of a given 73b77cf11fSRob Herring * format, and/or requires some format-specific default value. 74b77cf11fSRob Herring * 754c019de6SYong Wu * IO_PGTABLE_QUIRK_ARM_MTK_EXT: (ARM v7s format) MediaTek IOMMUs extend 7640596d2fSYong Wu * to support up to 35 bits PA where the bit32, bit33 and bit34 are 7740596d2fSYong Wu * encoded in the bit9, bit4 and bit5 of the PTE respectively. 78b77cf11fSRob Herring * 79bfdd2313SYunfei Wang * IO_PGTABLE_QUIRK_ARM_MTK_TTBR_EXT: (ARM v7s format) MediaTek IOMMUs 80bfdd2313SYunfei Wang * extend the translation table base support up to 35 bits PA, the 81bfdd2313SYunfei Wang * encoding format is same with IO_PGTABLE_QUIRK_ARM_MTK_EXT. 82bfdd2313SYunfei Wang * 83db690301SRobin Murphy * IO_PGTABLE_QUIRK_ARM_TTBR1: (ARM LPAE format) Configure the table 84db690301SRobin Murphy * for use in the upper half of a split address space. 85e67890c9SSai Prakash Ranjan * 86e67890c9SSai Prakash Ranjan * IO_PGTABLE_QUIRK_ARM_OUTER_WBWA: Override the outer-cacheability 87e67890c9SSai Prakash Ranjan * attributes set in the TCR for a non-coherent page-table walker. 88b77cf11fSRob Herring */ 89b77cf11fSRob Herring #define IO_PGTABLE_QUIRK_ARM_NS BIT(0) 90b77cf11fSRob Herring #define IO_PGTABLE_QUIRK_NO_PERMS BIT(1) 9173d50811SYong Wu #define IO_PGTABLE_QUIRK_ARM_MTK_EXT BIT(3) 92bfdd2313SYunfei Wang #define IO_PGTABLE_QUIRK_ARM_MTK_TTBR_EXT BIT(4) 93db690301SRobin Murphy #define IO_PGTABLE_QUIRK_ARM_TTBR1 BIT(5) 94e67890c9SSai Prakash Ranjan #define IO_PGTABLE_QUIRK_ARM_OUTER_WBWA BIT(6) 95b77cf11fSRob Herring unsigned long quirks; 96b77cf11fSRob Herring unsigned long pgsize_bitmap; 97b77cf11fSRob Herring unsigned int ias; 98b77cf11fSRob Herring unsigned int oas; 994f41845bSWill Deacon bool coherent_walk; 100298f7889SWill Deacon const struct iommu_flush_ops *tlb; 101b77cf11fSRob Herring struct device *iommu_dev; 102b77cf11fSRob Herring 103*17b226dcSBoris Brezillon /** 104*17b226dcSBoris Brezillon * @alloc: Custom page allocator. 105*17b226dcSBoris Brezillon * 106*17b226dcSBoris Brezillon * Optional hook used to allocate page tables. If this function is NULL, 107*17b226dcSBoris Brezillon * @free must be NULL too. 108*17b226dcSBoris Brezillon * 109*17b226dcSBoris Brezillon * Memory returned should be zeroed and suitable for dma_map_single() and 110*17b226dcSBoris Brezillon * virt_to_phys(). 111*17b226dcSBoris Brezillon * 112*17b226dcSBoris Brezillon * Not all formats support custom page allocators. Before considering 113*17b226dcSBoris Brezillon * passing a non-NULL value, make sure the chosen page format supports 114*17b226dcSBoris Brezillon * this feature. 115*17b226dcSBoris Brezillon */ 116*17b226dcSBoris Brezillon void *(*alloc)(void *cookie, size_t size, gfp_t gfp); 117*17b226dcSBoris Brezillon 118*17b226dcSBoris Brezillon /** 119*17b226dcSBoris Brezillon * @free: Custom page de-allocator. 120*17b226dcSBoris Brezillon * 121*17b226dcSBoris Brezillon * Optional hook used to free page tables allocated with the @alloc 122*17b226dcSBoris Brezillon * hook. Must be non-NULL if @alloc is not NULL, must be NULL 123*17b226dcSBoris Brezillon * otherwise. 124*17b226dcSBoris Brezillon */ 125*17b226dcSBoris Brezillon void (*free)(void *cookie, void *pages, size_t size); 126*17b226dcSBoris Brezillon 127b77cf11fSRob Herring /* Low-level data specific to the table format */ 128b77cf11fSRob Herring union { 129b77cf11fSRob Herring struct { 130d1e5f26fSRobin Murphy u64 ttbr; 131fb485eb1SRobin Murphy struct { 132fb485eb1SRobin Murphy u32 ips:3; 133fb485eb1SRobin Murphy u32 tg:2; 134fb485eb1SRobin Murphy u32 sh:2; 135fb485eb1SRobin Murphy u32 orgn:2; 136fb485eb1SRobin Murphy u32 irgn:2; 137fb485eb1SRobin Murphy u32 tsz:6; 138fb485eb1SRobin Murphy } tcr; 139205577abSRobin Murphy u64 mair; 140b77cf11fSRob Herring } arm_lpae_s1_cfg; 141b77cf11fSRob Herring 142b77cf11fSRob Herring struct { 143b77cf11fSRob Herring u64 vttbr; 144ac4b80e5SWill Deacon struct { 145ac4b80e5SWill Deacon u32 ps:3; 146ac4b80e5SWill Deacon u32 tg:2; 147ac4b80e5SWill Deacon u32 sh:2; 148ac4b80e5SWill Deacon u32 orgn:2; 149ac4b80e5SWill Deacon u32 irgn:2; 150ac4b80e5SWill Deacon u32 sl:2; 151ac4b80e5SWill Deacon u32 tsz:6; 152ac4b80e5SWill Deacon } vtcr; 153b77cf11fSRob Herring } arm_lpae_s2_cfg; 154b77cf11fSRob Herring 155b77cf11fSRob Herring struct { 156d1e5f26fSRobin Murphy u32 ttbr; 157b77cf11fSRob Herring u32 tcr; 158b77cf11fSRob Herring u32 nmrr; 159b77cf11fSRob Herring u32 prrr; 160b77cf11fSRob Herring } arm_v7s_cfg; 161d08d42deSRob Herring 162d08d42deSRob Herring struct { 163d08d42deSRob Herring u64 transtab; 164d08d42deSRob Herring u64 memattr; 165d08d42deSRob Herring } arm_mali_lpae_cfg; 166892384cdSSven Peter 167892384cdSSven Peter struct { 168892384cdSSven Peter u64 ttbr[4]; 169892384cdSSven Peter u32 n_ttbrs; 170892384cdSSven Peter } apple_dart_cfg; 171b77cf11fSRob Herring }; 172b77cf11fSRob Herring }; 173b77cf11fSRob Herring 174b77cf11fSRob Herring /** 175b77cf11fSRob Herring * struct io_pgtable_ops - Page table manipulation API for IOMMU drivers. 176b77cf11fSRob Herring * 177ca073b55SIsaac J. Manjarres * @map_pages: Map a physically contiguous range of pages of the same size. 178374c1559SIsaac J. Manjarres * @unmap_pages: Unmap a range of virtually contiguous pages of the same size. 179b77cf11fSRob Herring * @iova_to_phys: Translate iova to physical address. 180b77cf11fSRob Herring * 181b77cf11fSRob Herring * These functions map directly onto the iommu_ops member functions with 182b77cf11fSRob Herring * the same names. 183b77cf11fSRob Herring */ 184b77cf11fSRob Herring struct io_pgtable_ops { 185ca073b55SIsaac J. Manjarres int (*map_pages)(struct io_pgtable_ops *ops, unsigned long iova, 186ca073b55SIsaac J. Manjarres phys_addr_t paddr, size_t pgsize, size_t pgcount, 187ca073b55SIsaac J. Manjarres int prot, gfp_t gfp, size_t *mapped); 188374c1559SIsaac J. Manjarres size_t (*unmap_pages)(struct io_pgtable_ops *ops, unsigned long iova, 189374c1559SIsaac J. Manjarres size_t pgsize, size_t pgcount, 190374c1559SIsaac J. Manjarres struct iommu_iotlb_gather *gather); 191b77cf11fSRob Herring phys_addr_t (*iova_to_phys)(struct io_pgtable_ops *ops, 192b77cf11fSRob Herring unsigned long iova); 193750e2e90SJoao Martins int (*read_and_clear_dirty)(struct io_pgtable_ops *ops, 194750e2e90SJoao Martins unsigned long iova, size_t size, 195750e2e90SJoao Martins unsigned long flags, 196750e2e90SJoao Martins struct iommu_dirty_bitmap *dirty); 197b77cf11fSRob Herring }; 198b77cf11fSRob Herring 199b77cf11fSRob Herring /** 200b77cf11fSRob Herring * alloc_io_pgtable_ops() - Allocate a page table allocator for use by an IOMMU. 201b77cf11fSRob Herring * 202b77cf11fSRob Herring * @fmt: The page table format. 203b77cf11fSRob Herring * @cfg: The page table configuration. This will be modified to represent 204b77cf11fSRob Herring * the configuration actually provided by the allocator (e.g. the 205b77cf11fSRob Herring * pgsize_bitmap may be restricted). 206b77cf11fSRob Herring * @cookie: An opaque token provided by the IOMMU driver and passed back to 207b77cf11fSRob Herring * the callback routines in cfg->tlb. 208b77cf11fSRob Herring */ 209b77cf11fSRob Herring struct io_pgtable_ops *alloc_io_pgtable_ops(enum io_pgtable_fmt fmt, 210b77cf11fSRob Herring struct io_pgtable_cfg *cfg, 211b77cf11fSRob Herring void *cookie); 212b77cf11fSRob Herring 213b77cf11fSRob Herring /** 214b77cf11fSRob Herring * free_io_pgtable_ops() - Free an io_pgtable_ops structure. The caller 215b77cf11fSRob Herring * *must* ensure that the page table is no longer 216b77cf11fSRob Herring * live, but the TLB can be dirty. 217b77cf11fSRob Herring * 218b77cf11fSRob Herring * @ops: The ops returned from alloc_io_pgtable_ops. 219b77cf11fSRob Herring */ 220b77cf11fSRob Herring void free_io_pgtable_ops(struct io_pgtable_ops *ops); 221b77cf11fSRob Herring 222b77cf11fSRob Herring 223b77cf11fSRob Herring /* 224b77cf11fSRob Herring * Internal structures for page table allocator implementations. 225b77cf11fSRob Herring */ 226b77cf11fSRob Herring 227b77cf11fSRob Herring /** 228b77cf11fSRob Herring * struct io_pgtable - Internal structure describing a set of page tables. 229b77cf11fSRob Herring * 230b77cf11fSRob Herring * @fmt: The page table format. 231b77cf11fSRob Herring * @cookie: An opaque token provided by the IOMMU driver and passed back to 232b77cf11fSRob Herring * any callback routines. 233b77cf11fSRob Herring * @cfg: A copy of the page table configuration. 234b77cf11fSRob Herring * @ops: The page table operations in use for this set of page tables. 235b77cf11fSRob Herring */ 236b77cf11fSRob Herring struct io_pgtable { 237b77cf11fSRob Herring enum io_pgtable_fmt fmt; 238b77cf11fSRob Herring void *cookie; 239b77cf11fSRob Herring struct io_pgtable_cfg cfg; 240b77cf11fSRob Herring struct io_pgtable_ops ops; 241b77cf11fSRob Herring }; 242b77cf11fSRob Herring 243b77cf11fSRob Herring #define io_pgtable_ops_to_pgtable(x) container_of((x), struct io_pgtable, ops) 244b77cf11fSRob Herring 245b77cf11fSRob Herring static inline void io_pgtable_tlb_flush_all(struct io_pgtable *iop) 246b77cf11fSRob Herring { 24777e0992aSYong Wu if (iop->cfg.tlb && iop->cfg.tlb->tlb_flush_all) 248b77cf11fSRob Herring iop->cfg.tlb->tlb_flush_all(iop->cookie); 249b77cf11fSRob Herring } 250b77cf11fSRob Herring 25110b7a7d9SWill Deacon static inline void 25210b7a7d9SWill Deacon io_pgtable_tlb_flush_walk(struct io_pgtable *iop, unsigned long iova, 25310b7a7d9SWill Deacon size_t size, size_t granule) 254b77cf11fSRob Herring { 25577e0992aSYong Wu if (iop->cfg.tlb && iop->cfg.tlb->tlb_flush_walk) 25610b7a7d9SWill Deacon iop->cfg.tlb->tlb_flush_walk(iova, size, granule, iop->cookie); 257b77cf11fSRob Herring } 258b77cf11fSRob Herring 25910b7a7d9SWill Deacon static inline void 2603951c41aSWill Deacon io_pgtable_tlb_add_page(struct io_pgtable *iop, 2613951c41aSWill Deacon struct iommu_iotlb_gather * gather, unsigned long iova, 262abfd6fe0SWill Deacon size_t granule) 263b77cf11fSRob Herring { 26477e0992aSYong Wu if (iop->cfg.tlb && iop->cfg.tlb->tlb_add_page) 2653951c41aSWill Deacon iop->cfg.tlb->tlb_add_page(gather, iova, granule, iop->cookie); 266b77cf11fSRob Herring } 267b77cf11fSRob Herring 268b77cf11fSRob Herring /** 269*17b226dcSBoris Brezillon * enum io_pgtable_caps - IO page table backend capabilities. 270*17b226dcSBoris Brezillon */ 271*17b226dcSBoris Brezillon enum io_pgtable_caps { 272*17b226dcSBoris Brezillon /** @IO_PGTABLE_CAP_CUSTOM_ALLOCATOR: Backend accepts custom page table allocators. */ 273*17b226dcSBoris Brezillon IO_PGTABLE_CAP_CUSTOM_ALLOCATOR = BIT(0), 274*17b226dcSBoris Brezillon }; 275*17b226dcSBoris Brezillon 276*17b226dcSBoris Brezillon /** 277b77cf11fSRob Herring * struct io_pgtable_init_fns - Alloc/free a set of page tables for a 278b77cf11fSRob Herring * particular format. 279b77cf11fSRob Herring * 280b77cf11fSRob Herring * @alloc: Allocate a set of page tables described by cfg. 281b77cf11fSRob Herring * @free: Free the page tables associated with iop. 282*17b226dcSBoris Brezillon * @caps: Combination of @io_pgtable_caps flags encoding the backend capabilities. 283b77cf11fSRob Herring */ 284b77cf11fSRob Herring struct io_pgtable_init_fns { 285b77cf11fSRob Herring struct io_pgtable *(*alloc)(struct io_pgtable_cfg *cfg, void *cookie); 286b77cf11fSRob Herring void (*free)(struct io_pgtable *iop); 287*17b226dcSBoris Brezillon u32 caps; 288b77cf11fSRob Herring }; 289b77cf11fSRob Herring 290b77cf11fSRob Herring extern struct io_pgtable_init_fns io_pgtable_arm_32_lpae_s1_init_fns; 291b77cf11fSRob Herring extern struct io_pgtable_init_fns io_pgtable_arm_32_lpae_s2_init_fns; 292b77cf11fSRob Herring extern struct io_pgtable_init_fns io_pgtable_arm_64_lpae_s1_init_fns; 293b77cf11fSRob Herring extern struct io_pgtable_init_fns io_pgtable_arm_64_lpae_s2_init_fns; 294b77cf11fSRob Herring extern struct io_pgtable_init_fns io_pgtable_arm_v7s_init_fns; 295d08d42deSRob Herring extern struct io_pgtable_init_fns io_pgtable_arm_mali_lpae_init_fns; 296c9b258c6SSuravee Suthikulpanit extern struct io_pgtable_init_fns io_pgtable_amd_iommu_v1_init_fns; 297aaac38f6SVasant Hegde extern struct io_pgtable_init_fns io_pgtable_amd_iommu_v2_init_fns; 298892384cdSSven Peter extern struct io_pgtable_init_fns io_pgtable_apple_dart_init_fns; 299b77cf11fSRob Herring 300b77cf11fSRob Herring #endif /* __IO_PGTABLE_H */ 301