1 /* SPDX-License-Identifier: MIT */ 2 #ifndef __NVKM_MMU_H__ 3 #define __NVKM_MMU_H__ 4 #include <core/subdev.h> 5 #include <subdev/gsp.h> 6 7 struct nvkm_vma { 8 struct list_head head; 9 struct rb_node tree; 10 u64 addr; 11 u64 size:50; 12 bool mapref:1; /* PTs (de)referenced on (un)map (vs pre-allocated). */ 13 bool sparse:1; /* Unmapped PDEs/PTEs will not trigger MMU faults. */ 14 #define NVKM_VMA_PAGE_NONE 7 15 u8 page:3; /* Requested page type (index, or NONE for automatic). */ 16 u8 refd:3; /* Current page type (index, or NONE for unreferenced). */ 17 bool used:1; /* Region allocated. */ 18 bool part:1; /* Region was split from an allocated region by map(). */ 19 bool busy:1; /* Region busy (for temporarily preventing user access). */ 20 bool mapped:1; /* Region contains valid pages. */ 21 bool no_comp:1; /* Force no memory compression. */ 22 struct nvkm_memory *memory; /* Memory currently mapped into VMA. */ 23 struct nvkm_tags *tags; /* Compression tag reference. */ 24 }; 25 26 struct nvkm_vmm { 27 const struct nvkm_vmm_func *func; 28 struct nvkm_mmu *mmu; 29 const char *name; 30 u32 debug; 31 struct kref kref; 32 33 struct { 34 struct mutex vmm; 35 struct mutex ref; 36 struct mutex map; 37 } mutex; 38 39 u64 start; 40 u64 limit; 41 struct { 42 struct { 43 u64 addr; 44 u64 size; 45 } p; 46 struct { 47 u64 addr; 48 u64 size; 49 } n; 50 bool raw; 51 } managed; 52 53 struct nvkm_vmm_pt *pd; 54 struct list_head join; 55 56 struct list_head list; 57 struct rb_root free; 58 struct rb_root root; 59 60 bool bootstrapped; 61 atomic_t engref[NVKM_SUBDEV_NR]; 62 63 dma_addr_t null; 64 void *nullp; 65 66 bool replay; 67 68 struct { 69 u64 bar2_pdb; 70 71 struct nvkm_gsp_client client; 72 struct nvkm_gsp_device device; 73 struct nvkm_gsp_object object; 74 75 struct nvkm_vma *rsvd; 76 } rm; 77 }; 78 79 int nvkm_vmm_new(struct nvkm_device *, u64 addr, u64 size, void *argv, u32 argc, 80 struct lock_class_key *, const char *name, struct nvkm_vmm **); 81 struct nvkm_vmm *nvkm_vmm_ref(struct nvkm_vmm *); 82 void nvkm_vmm_unref(struct nvkm_vmm **); 83 int nvkm_vmm_boot(struct nvkm_vmm *); 84 int nvkm_vmm_join(struct nvkm_vmm *, struct nvkm_memory *inst); 85 void nvkm_vmm_part(struct nvkm_vmm *, struct nvkm_memory *inst); 86 int nvkm_vmm_get(struct nvkm_vmm *, u8 page, u64 size, struct nvkm_vma **); 87 void nvkm_vmm_put(struct nvkm_vmm *, struct nvkm_vma **); 88 89 struct nvkm_vmm_map { 90 struct nvkm_memory *memory; 91 u64 offset; 92 93 struct nvkm_mm_node *mem; 94 struct scatterlist *sgl; 95 dma_addr_t *dma; 96 u64 *pfn; 97 u64 off; 98 99 const struct nvkm_vmm_page *page; 100 101 bool no_comp; 102 struct nvkm_tags *tags; 103 u64 next; 104 u64 type; 105 u64 ctag; 106 }; 107 108 int nvkm_vmm_map(struct nvkm_vmm *, struct nvkm_vma *, void *argv, u32 argc, 109 struct nvkm_vmm_map *); 110 void nvkm_vmm_unmap(struct nvkm_vmm *, struct nvkm_vma *); 111 112 struct nvkm_memory *nvkm_umem_search(struct nvkm_client *, u64); 113 struct nvkm_vmm *nvkm_uvmm_search(struct nvkm_client *, u64 handle); 114 115 struct nvkm_mmu { 116 const struct nvkm_mmu_func *func; 117 struct nvkm_subdev subdev; 118 119 u8 dma_bits; 120 121 int heap_nr; 122 struct { 123 #define NVKM_MEM_VRAM 0x01 124 #define NVKM_MEM_HOST 0x02 125 #define NVKM_MEM_COMP 0x04 126 #define NVKM_MEM_DISP 0x08 127 u8 type; 128 u64 size; 129 } heap[4]; 130 131 int type_nr; 132 struct { 133 #define NVKM_MEM_KIND 0x10 134 #define NVKM_MEM_MAPPABLE 0x20 135 #define NVKM_MEM_COHERENT 0x40 136 #define NVKM_MEM_UNCACHED 0x80 137 u8 type; 138 u8 heap; 139 } type[16]; 140 141 struct nvkm_vmm *vmm; 142 143 struct { 144 struct mutex mutex; 145 struct list_head list; 146 } ptc, ptp; 147 148 struct mutex mutex; /* serialises mmu invalidations */ 149 150 struct nvkm_device_oclass user; 151 }; 152 153 int nv04_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 154 int nv41_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 155 int nv44_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 156 int nv50_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 157 int g84_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 158 int mcp77_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 159 int gf100_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 160 int gk104_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 161 int gk20a_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 162 int gm200_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 163 int gm20b_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 164 int gp100_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 165 int gp10b_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 166 int gv100_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 167 int tu102_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); 168 #endif 169