1 /*- 2 * Copyright (C) 2006 Semihalf, Marian Balakowicz <m8@semihalf.com> 3 * All rights reserved. 4 * 5 * Adapted for Freescale's e500 core CPUs. 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice, this list of conditions and the following disclaimer. 12 * 2. Redistributions in binary form must reproduce the above copyright 13 * notice, this list of conditions and the following disclaimer in the 14 * documentation and/or other materials provided with the distribution. 15 * 3. The name of the author may not be used to endorse or promote products 16 * derived from this software without specific prior written permission. 17 * 18 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 19 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 20 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN 21 * NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, 22 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED 23 * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR 24 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF 25 * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING 26 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS 27 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 28 * 29 * $FreeBSD$ 30 */ 31 /*- 32 * Copyright (C) 1995, 1996 Wolfgang Solfrank. 33 * Copyright (C) 1995, 1996 TooLs GmbH. 34 * All rights reserved. 35 * 36 * Redistribution and use in source and binary forms, with or without 37 * modification, are permitted provided that the following conditions 38 * are met: 39 * 1. Redistributions of source code must retain the above copyright 40 * notice, this list of conditions and the following disclaimer. 41 * 2. Redistributions in binary form must reproduce the above copyright 42 * notice, this list of conditions and the following disclaimer in the 43 * documentation and/or other materials provided with the distribution. 44 * 3. All advertising materials mentioning features or use of this software 45 * must display the following acknowledgement: 46 * This product includes software developed by TooLs GmbH. 47 * 4. The name of TooLs GmbH may not be used to endorse or promote products 48 * derived from this software without specific prior written permission. 49 * 50 * THIS SOFTWARE IS PROVIDED BY TOOLS GMBH ``AS IS'' AND ANY EXPRESS OR 51 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 52 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 53 * IN NO EVENT SHALL TOOLS GMBH BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, 54 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, 55 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; 56 * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, 57 * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR 58 * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF 59 * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 60 * 61 * from: $NetBSD: pmap.h,v 1.17 2000/03/30 16:18:24 jdolecek Exp $ 62 */ 63 64 #ifndef _MACHINE_PMAP_H_ 65 #define _MACHINE_PMAP_H_ 66 67 #include <sys/queue.h> 68 #include <sys/tree.h> 69 #include <sys/_lock.h> 70 #include <sys/_mutex.h> 71 #include <machine/sr.h> 72 #include <machine/pte.h> 73 #include <machine/slb.h> 74 #include <machine/tlb.h> 75 76 struct pmap_md { 77 u_int md_index; 78 vm_paddr_t md_paddr; 79 vm_offset_t md_vaddr; 80 vm_size_t md_size; 81 }; 82 83 #if defined(AIM) 84 85 #if !defined(NPMAPS) 86 #define NPMAPS 32768 87 #endif /* !defined(NPMAPS) */ 88 89 struct slbtnode; 90 91 struct pmap { 92 struct mtx pm_mtx; 93 94 #ifdef __powerpc64__ 95 struct slbtnode *pm_slb_tree_root; 96 struct slb **pm_slb; 97 int pm_slb_len; 98 #else 99 register_t pm_sr[16]; 100 #endif 101 cpumask_t pm_active; 102 uint32_t pm_gen_count; /* generation count (pmap lock dropped) */ 103 u_int pm_retries; 104 105 struct pmap *pmap_phys; 106 struct pmap_statistics pm_stats; 107 }; 108 109 typedef struct pmap *pmap_t; 110 111 struct pvo_entry { 112 LIST_ENTRY(pvo_entry) pvo_vlink; /* Link to common virt page */ 113 LIST_ENTRY(pvo_entry) pvo_olink; /* Link to overflow entry */ 114 union { 115 struct pte pte; /* 32 bit PTE */ 116 struct lpte lpte; /* 64 bit PTE */ 117 } pvo_pte; 118 pmap_t pvo_pmap; /* Owning pmap */ 119 vm_offset_t pvo_vaddr; /* VA of entry */ 120 uint64_t pvo_vpn; /* Virtual page number */ 121 }; 122 LIST_HEAD(pvo_head, pvo_entry); 123 124 #define PVO_PTEGIDX_MASK 0x007UL /* which PTEG slot */ 125 #define PVO_PTEGIDX_VALID 0x008UL /* slot is valid */ 126 #define PVO_WIRED 0x010UL /* PVO entry is wired */ 127 #define PVO_MANAGED 0x020UL /* PVO entry is managed */ 128 #define PVO_EXECUTABLE 0x040UL /* PVO entry is executable */ 129 #define PVO_BOOTSTRAP 0x080UL /* PVO entry allocated during 130 bootstrap */ 131 #define PVO_FAKE 0x100UL /* fictitious phys page */ 132 #define PVO_LARGE 0x200UL /* large page */ 133 #define PVO_VADDR(pvo) ((pvo)->pvo_vaddr & ~ADDR_POFF) 134 #define PVO_ISFAKE(pvo) ((pvo)->pvo_vaddr & PVO_FAKE) 135 #define PVO_PTEGIDX_GET(pvo) ((pvo)->pvo_vaddr & PVO_PTEGIDX_MASK) 136 #define PVO_PTEGIDX_ISSET(pvo) ((pvo)->pvo_vaddr & PVO_PTEGIDX_VALID) 137 #define PVO_PTEGIDX_CLR(pvo) \ 138 ((void)((pvo)->pvo_vaddr &= ~(PVO_PTEGIDX_VALID|PVO_PTEGIDX_MASK))) 139 #define PVO_PTEGIDX_SET(pvo, i) \ 140 ((void)((pvo)->pvo_vaddr |= (i)|PVO_PTEGIDX_VALID)) 141 #define PVO_VSID(pvo) ((pvo)->pvo_vpn >> 16) 142 143 struct md_page { 144 u_int64_t mdpg_attrs; 145 vm_memattr_t mdpg_cache_attrs; 146 struct pvo_head mdpg_pvoh; 147 }; 148 149 #define pmap_page_get_memattr(m) ((m)->md.mdpg_cache_attrs) 150 #define pmap_page_is_mapped(m) (!LIST_EMPTY(&(m)->md.mdpg_pvoh)) 151 152 /* 153 * Return the VSID corresponding to a given virtual address. 154 * If no VSID is currently defined, it will allocate one, and add 155 * it to a free slot if available. 156 * 157 * NB: The PMAP MUST be locked already. 158 */ 159 uint64_t va_to_vsid(pmap_t pm, vm_offset_t va); 160 161 /* Lock-free, non-allocating lookup routines */ 162 uint64_t kernel_va_to_slbv(vm_offset_t va); 163 struct slb *user_va_to_slb_entry(pmap_t pm, vm_offset_t va); 164 165 uint64_t allocate_user_vsid(pmap_t pm, uint64_t esid, int large); 166 void free_vsid(pmap_t pm, uint64_t esid, int large); 167 void slb_insert_user(pmap_t pm, struct slb *slb); 168 void slb_insert_kernel(uint64_t slbe, uint64_t slbv); 169 170 struct slbtnode *slb_alloc_tree(void); 171 void slb_free_tree(pmap_t pm); 172 struct slb **slb_alloc_user_cache(void); 173 void slb_free_user_cache(struct slb **); 174 175 #else 176 177 struct pmap { 178 struct mtx pm_mtx; /* pmap mutex */ 179 tlbtid_t pm_tid[MAXCPU]; /* TID to identify this pmap entries in TLB */ 180 cpumask_t pm_active; /* active on cpus */ 181 int pm_refs; /* ref count */ 182 struct pmap_statistics pm_stats; /* pmap statistics */ 183 184 /* Page table directory, array of pointers to page tables. */ 185 pte_t *pm_pdir[PDIR_NENTRIES]; 186 187 /* generation count (pmap lock dropped) */ 188 uint32_t pm_gen_count; 189 u_int pm_retries; 190 191 /* List of allocated ptbl bufs (ptbl kva regions). */ 192 TAILQ_HEAD(, ptbl_buf) pm_ptbl_list; 193 }; 194 typedef struct pmap *pmap_t; 195 196 struct pv_entry { 197 pmap_t pv_pmap; 198 vm_offset_t pv_va; 199 TAILQ_ENTRY(pv_entry) pv_link; 200 }; 201 typedef struct pv_entry *pv_entry_t; 202 203 struct md_page { 204 TAILQ_HEAD(, pv_entry) pv_list; 205 }; 206 207 #define pmap_page_get_memattr(m) VM_MEMATTR_DEFAULT 208 #define pmap_page_is_mapped(m) (!TAILQ_EMPTY(&(m)->md.pv_list)) 209 210 #endif /* AIM */ 211 212 extern struct pmap kernel_pmap_store; 213 #define kernel_pmap (&kernel_pmap_store) 214 215 #ifdef _KERNEL 216 217 #define PMAP_LOCK(pmap) mtx_lock(&(pmap)->pm_mtx) 218 #define PMAP_LOCK_ASSERT(pmap, type) \ 219 mtx_assert(&(pmap)->pm_mtx, (type)) 220 #define PMAP_LOCK_DESTROY(pmap) mtx_destroy(&(pmap)->pm_mtx) 221 #define PMAP_LOCK_INIT(pmap) mtx_init(&(pmap)->pm_mtx, "pmap", \ 222 NULL, MTX_DEF) 223 #define PMAP_LOCKED(pmap) mtx_owned(&(pmap)->pm_mtx) 224 #define PMAP_MTX(pmap) (&(pmap)->pm_mtx) 225 #define PMAP_TRYLOCK(pmap) mtx_trylock(&(pmap)->pm_mtx) 226 #define PMAP_UNLOCK(pmap) mtx_unlock(&(pmap)->pm_mtx) 227 228 void pmap_bootstrap(vm_offset_t, vm_offset_t); 229 void pmap_kenter(vm_offset_t va, vm_offset_t pa); 230 void pmap_kenter_attr(vm_offset_t va, vm_offset_t pa, vm_memattr_t); 231 void pmap_kremove(vm_offset_t); 232 void *pmap_mapdev(vm_offset_t, vm_size_t); 233 void *pmap_mapdev_attr(vm_offset_t, vm_size_t, vm_memattr_t); 234 void pmap_unmapdev(vm_offset_t, vm_size_t); 235 void pmap_page_set_memattr(vm_page_t, vm_memattr_t); 236 void pmap_deactivate(struct thread *); 237 vm_offset_t pmap_kextract(vm_offset_t); 238 int pmap_dev_direct_mapped(vm_offset_t, vm_size_t); 239 boolean_t pmap_mmu_install(char *name, int prio); 240 241 #define vtophys(va) pmap_kextract((vm_offset_t)(va)) 242 243 #define PHYS_AVAIL_SZ 128 244 extern vm_offset_t phys_avail[PHYS_AVAIL_SZ]; 245 extern vm_offset_t virtual_avail; 246 extern vm_offset_t virtual_end; 247 248 extern vm_offset_t msgbuf_phys; 249 250 extern int pmap_bootstrapped; 251 252 extern vm_offset_t pmap_dumpsys_map(struct pmap_md *, vm_size_t, vm_size_t *); 253 extern void pmap_dumpsys_unmap(struct pmap_md *, vm_size_t, vm_offset_t); 254 255 extern struct pmap_md *pmap_scan_md(struct pmap_md *); 256 257 #endif 258 259 #endif /* !_MACHINE_PMAP_H_ */ 260