1 /*-
2 * SPDX-License-Identifier: BSD-3-Clause AND BSD-4-Clause
3 *
4 * Copyright (C) 2006 Semihalf, Marian Balakowicz <m8@semihalf.com>
5 * All rights reserved.
6 *
7 * Adapted for Freescale's e500 core CPUs.
8 *
9 * Redistribution and use in source and binary forms, with or without
10 * modification, are permitted provided that the following conditions
11 * are met:
12 * 1. Redistributions of source code must retain the above copyright
13 * notice, this list of conditions and the following disclaimer.
14 * 2. Redistributions in binary form must reproduce the above copyright
15 * notice, this list of conditions and the following disclaimer in the
16 * documentation and/or other materials provided with the distribution.
17 * 3. The name of the author may not be used to endorse or promote products
18 * derived from this software without specific prior written permission.
19 *
20 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
21 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
22 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN
23 * NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
24 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
25 * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
26 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
27 * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
28 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
29 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
30 */
31 /*-
32 * Copyright (C) 1995, 1996 Wolfgang Solfrank.
33 * Copyright (C) 1995, 1996 TooLs GmbH.
34 * All rights reserved.
35 *
36 * Redistribution and use in source and binary forms, with or without
37 * modification, are permitted provided that the following conditions
38 * are met:
39 * 1. Redistributions of source code must retain the above copyright
40 * notice, this list of conditions and the following disclaimer.
41 * 2. Redistributions in binary form must reproduce the above copyright
42 * notice, this list of conditions and the following disclaimer in the
43 * documentation and/or other materials provided with the distribution.
44 * 3. All advertising materials mentioning features or use of this software
45 * must display the following acknowledgement:
46 * This product includes software developed by TooLs GmbH.
47 * 4. The name of TooLs GmbH may not be used to endorse or promote products
48 * derived from this software without specific prior written permission.
49 *
50 * THIS SOFTWARE IS PROVIDED BY TOOLS GMBH ``AS IS'' AND ANY EXPRESS OR
51 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
52 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
53 * IN NO EVENT SHALL TOOLS GMBH BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
54 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
55 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
56 * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
57 * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
58 * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
59 * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
60 *
61 * from: $NetBSD: pmap.h,v 1.17 2000/03/30 16:18:24 jdolecek Exp $
62 */
63
64 #ifndef _MACHINE_PMAP_H_
65 #define _MACHINE_PMAP_H_
66
67 #include <sys/queue.h>
68 #include <sys/tree.h>
69 #include <sys/_cpuset.h>
70 #include <sys/_lock.h>
71 #include <sys/_mutex.h>
72 #include <machine/sr.h>
73 #include <machine/pte.h>
74 #include <machine/slb.h>
75 #include <machine/tlb.h>
76 #include <machine/vmparam.h>
77 #ifdef __powerpc64__
78 #include <vm/_vm_radix.h>
79 #endif
80
81 /*
82 * The radix page table structure is described by levels 1-4.
83 * See Fig 33. on p. 1002 of Power ISA v3.0B
84 *
85 * Page directories and tables must be size aligned.
86 */
87
88 /* Root page directory - 64k -- each entry covers 512GB */
89 typedef uint64_t pml1_entry_t;
90 /* l2 page directory - 4k -- each entry covers 1GB */
91 typedef uint64_t pml2_entry_t;
92 /* l3 page directory - 4k -- each entry covers 2MB */
93 typedef uint64_t pml3_entry_t;
94 /* l4 page directory - 256B/4k -- each entry covers 64k/4k */
95 typedef uint64_t pml4_entry_t;
96
97 typedef uint64_t pt_entry_t;
98
99 struct pmap;
100 typedef struct pmap *pmap_t;
101
102 #define PMAP_ENTER_QUICK_LOCKED 0x10000000
103
104 #if !defined(NPMAPS)
105 #define NPMAPS 32768
106 #endif /* !defined(NPMAPS) */
107
108 struct slbtnode;
109
110 struct pvo_entry {
111 LIST_ENTRY(pvo_entry) pvo_vlink; /* Link to common virt page */
112 #ifndef __powerpc64__
113 LIST_ENTRY(pvo_entry) pvo_olink; /* Link to overflow entry */
114 #endif
115 union {
116 RB_ENTRY(pvo_entry) pvo_plink; /* Link to pmap entries */
117 SLIST_ENTRY(pvo_entry) pvo_dlink; /* Link to delete enty */
118 };
119 struct {
120 #ifndef __powerpc64__
121 /* 32-bit fields */
122 pte_t pte;
123 #endif
124 /* 64-bit fields */
125 uintptr_t slot;
126 vm_paddr_t pa;
127 vm_prot_t prot;
128 } pvo_pte;
129 pmap_t pvo_pmap; /* Owning pmap */
130 vm_offset_t pvo_vaddr; /* VA of entry */
131 uint64_t pvo_vpn; /* Virtual page number */
132 };
133 LIST_HEAD(pvo_head, pvo_entry);
134 SLIST_HEAD(pvo_dlist, pvo_entry);
135 RB_HEAD(pvo_tree, pvo_entry);
136 int pvo_vaddr_compare(struct pvo_entry *, struct pvo_entry *);
137 RB_PROTOTYPE(pvo_tree, pvo_entry, pvo_plink, pvo_vaddr_compare);
138
139 /* Used by 32-bit PMAP */
140 #define PVO_PTEGIDX_MASK 0x007UL /* which PTEG slot */
141 #define PVO_PTEGIDX_VALID 0x008UL /* slot is valid */
142 /* Used by 64-bit PMAP */
143 #define PVO_HID 0x008UL /* PVO entry in alternate hash*/
144 /* Used by both */
145 #define PVO_WIRED 0x010UL /* PVO entry is wired */
146 #define PVO_MANAGED 0x020UL /* PVO entry is managed */
147 #define PVO_BOOTSTRAP 0x080UL /* PVO entry allocated during
148 bootstrap */
149 #define PVO_DEAD 0x100UL /* waiting to be deleted */
150 #define PVO_LARGE 0x200UL /* large page */
151 #define PVO_VADDR(pvo) ((pvo)->pvo_vaddr & ~ADDR_POFF)
152 #define PVO_PTEGIDX_GET(pvo) ((pvo)->pvo_vaddr & PVO_PTEGIDX_MASK)
153 #define PVO_PTEGIDX_ISSET(pvo) ((pvo)->pvo_vaddr & PVO_PTEGIDX_VALID)
154 #define PVO_PTEGIDX_CLR(pvo) \
155 ((void)((pvo)->pvo_vaddr &= ~(PVO_PTEGIDX_VALID|PVO_PTEGIDX_MASK)))
156 #define PVO_PTEGIDX_SET(pvo, i) \
157 ((void)((pvo)->pvo_vaddr |= (i)|PVO_PTEGIDX_VALID))
158 #define PVO_VSID(pvo) ((pvo)->pvo_vpn >> 16)
159
160 struct pmap {
161 struct pmap_statistics pm_stats;
162 struct mtx pm_mtx;
163 cpuset_t pm_active;
164 union {
165 struct {
166 #ifdef __powerpc64__
167 struct slbtnode *pm_slb_tree_root;
168 struct slb **pm_slb;
169 int pm_slb_len;
170 #else
171 register_t pm_sr[16];
172 #endif
173
174 struct pmap *pmap_phys;
175 struct pvo_tree pmap_pvo;
176 };
177 #ifdef __powerpc64__
178 /* Radix support */
179 struct {
180 pml1_entry_t *pm_pml1; /* KVA of root page directory */
181 struct vm_radix pm_radix; /* spare page table pages */
182 TAILQ_HEAD(,pv_chunk) pm_pvchunk; /* list of mappings in pmap */
183 uint64_t pm_pid; /* PIDR value */
184 int pm_flags;
185 };
186 #endif
187 struct {
188 /* TID to identify this pmap entries in TLB */
189 tlbtid_t pm_tid[MAXCPU];
190
191 #ifdef __powerpc64__
192 /*
193 * Page table directory,
194 * array of pointers to page directories.
195 */
196 pte_t ****pm_root;
197 #else
198 /*
199 * Page table directory,
200 * array of pointers to page tables.
201 */
202 pte_t **pm_pdir;
203
204 /* List of allocated ptbl bufs (ptbl kva regions). */
205 TAILQ_HEAD(, ptbl_buf) pm_ptbl_list;
206 #endif
207 };
208 } __aligned(CACHE_LINE_SIZE);
209 };
210
211 /*
212 * pv_entries are allocated in chunks per-process. This avoids the
213 * need to track per-pmap assignments.
214 */
215 #define _NPCPV 126
216 #define _NPCM howmany(_NPCPV, 64)
217
218 #define PV_CHUNK_HEADER \
219 pmap_t pc_pmap; \
220 TAILQ_ENTRY(pv_chunk) pc_list; \
221 uint64_t pc_map[_NPCM]; /* bitmap; 1 = free */ \
222 TAILQ_ENTRY(pv_chunk) pc_lru;
223
224 struct pv_entry {
225 pmap_t pv_pmap;
226 vm_offset_t pv_va;
227 TAILQ_ENTRY(pv_entry) pv_link;
228 };
229 typedef struct pv_entry *pv_entry_t;
230
231 struct pv_chunk_header {
232 PV_CHUNK_HEADER
233 };
234 struct pv_chunk {
235 PV_CHUNK_HEADER
236 uint64_t reserved;
237 struct pv_entry pc_pventry[_NPCPV];
238 };
239
240 struct md_page {
241 union {
242 struct {
243 volatile int32_t mdpg_attrs;
244 vm_memattr_t mdpg_cache_attrs;
245 struct pvo_head mdpg_pvoh;
246 int pv_gen; /* (p) */
247 };
248 struct {
249 int pv_tracked;
250 };
251 };
252 TAILQ_HEAD(, pv_entry) pv_list; /* (p) */
253 };
254
255 #ifdef AIM
256 #define pmap_page_get_memattr(m) ((m)->md.mdpg_cache_attrs)
257 #else
258 #define pmap_page_get_memattr(m) VM_MEMATTR_DEFAULT
259 #endif /* AIM */
260
261 /*
262 * Return the VSID corresponding to a given virtual address.
263 * If no VSID is currently defined, it will allocate one, and add
264 * it to a free slot if available.
265 *
266 * NB: The PMAP MUST be locked already.
267 */
268 uint64_t va_to_vsid(pmap_t pm, vm_offset_t va);
269
270 /* Lock-free, non-allocating lookup routines */
271 uint64_t kernel_va_to_slbv(vm_offset_t va);
272 struct slb *user_va_to_slb_entry(pmap_t pm, vm_offset_t va);
273
274 uint64_t allocate_user_vsid(pmap_t pm, uint64_t esid, int large);
275 void free_vsid(pmap_t pm, uint64_t esid, int large);
276 void slb_insert_user(pmap_t pm, struct slb *slb);
277 void slb_insert_kernel(uint64_t slbe, uint64_t slbv);
278
279 struct slbtnode *slb_alloc_tree(void);
280 void slb_free_tree(pmap_t pm);
281 struct slb **slb_alloc_user_cache(void);
282 void slb_free_user_cache(struct slb **);
283
284 extern struct pmap kernel_pmap_store;
285 #define kernel_pmap (&kernel_pmap_store)
286
287 #ifdef _KERNEL
288
289 #define PMAP_LOCK(pmap) mtx_lock(&(pmap)->pm_mtx)
290 #define PMAP_LOCK_ASSERT(pmap, type) \
291 mtx_assert(&(pmap)->pm_mtx, (type))
292 #define PMAP_LOCK_DESTROY(pmap) mtx_destroy(&(pmap)->pm_mtx)
293 #define PMAP_LOCK_INIT(pmap) mtx_init(&(pmap)->pm_mtx, \
294 (pmap == kernel_pmap) ? "kernelpmap" : \
295 "pmap", NULL, MTX_DEF | MTX_DUPOK)
296 #define PMAP_LOCKED(pmap) mtx_owned(&(pmap)->pm_mtx)
297 #define PMAP_MTX(pmap) (&(pmap)->pm_mtx)
298 #define PMAP_TRYLOCK(pmap) mtx_trylock(&(pmap)->pm_mtx)
299 #define PMAP_UNLOCK(pmap) mtx_unlock(&(pmap)->pm_mtx)
300
301 #define pmap_page_is_write_mapped(m) (((m)->a.flags & PGA_WRITEABLE) != 0)
302
303 #define pmap_vm_page_alloc_check(m)
304
305 void pmap_bootstrap(vm_offset_t, vm_offset_t);
306 void pmap_kenter(vm_offset_t va, vm_paddr_t pa);
307 void pmap_kenter_attr(vm_offset_t va, vm_paddr_t pa, vm_memattr_t);
308 void pmap_kremove(vm_offset_t);
309 void *pmap_mapdev(vm_paddr_t, vm_size_t);
310 void *pmap_mapdev_attr(vm_paddr_t, vm_size_t, vm_memattr_t);
311 void pmap_unmapdev(void *, vm_size_t);
312 void pmap_page_set_memattr(vm_page_t, vm_memattr_t);
313 int pmap_change_attr(vm_offset_t, vm_size_t, vm_memattr_t);
314 int pmap_map_user_ptr(pmap_t pm, volatile const void *uaddr,
315 void **kaddr, size_t ulen, size_t *klen);
316 int pmap_decode_kernel_ptr(vm_offset_t addr, int *is_user,
317 vm_offset_t *decoded_addr);
318 void pmap_deactivate(struct thread *);
319 vm_paddr_t pmap_kextract(vm_offset_t);
320 int pmap_dev_direct_mapped(vm_paddr_t, vm_size_t);
321 bool pmap_mmu_install(char *name, int prio);
322 void pmap_mmu_init(void);
323 const char *pmap_mmu_name(void);
324 bool pmap_ps_enabled(pmap_t pmap);
325 int pmap_nofault(pmap_t pmap, vm_offset_t va, vm_prot_t flags);
326 bool pmap_page_is_mapped(vm_page_t m);
327 #define pmap_map_delete(pmap, sva, eva) pmap_remove(pmap, sva, eva)
328
329 void pmap_page_array_startup(long count);
330
331 #define vtophys(va) pmap_kextract((vm_offset_t)(va))
332
333 extern vm_offset_t virtual_avail;
334 extern vm_offset_t virtual_end;
335 extern caddr_t crashdumpmap;
336
337 extern vm_offset_t msgbuf_phys;
338
339 extern int pmap_bootstrapped;
340 extern int radix_mmu;
341 extern int superpages_enabled;
342
343 #ifdef AIM
344 void pmap_early_io_map_init(void);
345 #endif
346 vm_offset_t pmap_early_io_map(vm_paddr_t pa, vm_size_t size);
347 void pmap_early_io_unmap(vm_offset_t va, vm_size_t size);
348 void pmap_track_page(pmap_t pmap, vm_offset_t va);
349 void pmap_page_print_mappings(vm_page_t m);
350 void pmap_tlbie_all(void);
351
352 static inline int
pmap_vmspace_copy(pmap_t dst_pmap __unused,pmap_t src_pmap __unused)353 pmap_vmspace_copy(pmap_t dst_pmap __unused, pmap_t src_pmap __unused)
354 {
355
356 return (0);
357 }
358
359 #endif
360
361 #endif /* !_MACHINE_PMAP_H_ */
362