1 /* SPDX-License-Identifier: GPL-2.0-only */ 2 /* 3 * Copyright (C) 2013 Red Hat 4 * Author: Rob Clark <robdclark@gmail.com> 5 */ 6 7 #ifndef __MSM_MMU_H__ 8 #define __MSM_MMU_H__ 9 10 #include <linux/iommu.h> 11 12 struct msm_mmu_prealloc; 13 struct msm_mmu; 14 struct msm_gpu; 15 16 struct msm_mmu_funcs { 17 void (*detach)(struct msm_mmu *mmu); 18 void (*prealloc_count)(struct msm_mmu *mmu, struct msm_mmu_prealloc *p, 19 uint64_t iova, size_t len); 20 int (*prealloc_allocate)(struct msm_mmu *mmu, struct msm_mmu_prealloc *p); 21 void (*prealloc_cleanup)(struct msm_mmu *mmu, struct msm_mmu_prealloc *p); 22 int (*map)(struct msm_mmu *mmu, uint64_t iova, struct sg_table *sgt, 23 size_t off, size_t len, int prot); 24 int (*unmap)(struct msm_mmu *mmu, uint64_t iova, size_t len); 25 void (*destroy)(struct msm_mmu *mmu); 26 void (*set_stall)(struct msm_mmu *mmu, bool enable); 27 }; 28 29 enum msm_mmu_type { 30 MSM_MMU_GPUMMU, 31 MSM_MMU_IOMMU, 32 MSM_MMU_IOMMU_PAGETABLE, 33 }; 34 35 /** 36 * struct msm_mmu_prealloc - Tracking for pre-allocated pages for MMU updates. 37 */ 38 struct msm_mmu_prealloc { 39 /** @count: Number of pages reserved. */ 40 uint32_t count; 41 /** @ptr: Index of first unused page in @pages */ 42 uint32_t ptr; 43 /** 44 * @pages: Array of pages preallocated for MMU table updates. 45 * 46 * After a VM operation, there might be free pages remaining in this 47 * array (since the amount allocated is a worst-case). These are 48 * returned to the pt_cache at mmu->prealloc_cleanup(). 49 */ 50 void **pages; 51 }; 52 53 struct msm_mmu { 54 const struct msm_mmu_funcs *funcs; 55 struct device *dev; 56 int (*handler)(void *arg, unsigned long iova, int flags, void *data); 57 void *arg; 58 enum msm_mmu_type type; 59 60 /** 61 * @prealloc: pre-allocated pages for pgtable 62 * 63 * Set while a VM_BIND job is running, serialized under 64 * msm_gem_vm::mmu_lock. 65 */ 66 struct msm_mmu_prealloc *prealloc; 67 }; 68 69 static inline void msm_mmu_init(struct msm_mmu *mmu, struct device *dev, 70 const struct msm_mmu_funcs *funcs, enum msm_mmu_type type) 71 { 72 mmu->dev = dev; 73 mmu->funcs = funcs; 74 mmu->type = type; 75 } 76 77 struct msm_mmu *msm_iommu_new(struct device *dev, unsigned long quirks); 78 struct msm_mmu *msm_iommu_gpu_new(struct device *dev, struct msm_gpu *gpu, unsigned long quirks); 79 struct msm_mmu *msm_iommu_disp_new(struct device *dev, unsigned long quirks); 80 81 static inline void msm_mmu_set_fault_handler(struct msm_mmu *mmu, void *arg, 82 int (*handler)(void *arg, unsigned long iova, int flags, void *data)) 83 { 84 mmu->arg = arg; 85 mmu->handler = handler; 86 } 87 88 struct msm_mmu *msm_iommu_pagetable_create(struct msm_mmu *parent, bool kernel_managed); 89 90 int msm_iommu_pagetable_params(struct msm_mmu *mmu, phys_addr_t *ttbr, 91 int *asid); 92 int msm_iommu_pagetable_walk(struct msm_mmu *mmu, unsigned long iova, uint64_t ptes[4]); 93 struct iommu_domain_geometry *msm_iommu_get_geometry(struct msm_mmu *mmu); 94 95 #endif /* __MSM_MMU_H__ */ 96