xref: /freebsd/sys/powerpc/aim/slb.c (revision 1f4bcc459a76b7aa664f3fd557684cd0ba6da352)
1 /*-
2  * Copyright (c) 2010 Nathan Whitehorn
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  *
9  * 1. Redistributions of source code must retain the above copyright
10  *    notice, this list of conditions and the following disclaimer.
11  * 2. Redistributions in binary form must reproduce the above copyright
12  *    notice, this list of conditions and the following disclaimer in the
13  *    documentation and/or other materials provided with the distribution.
14  *
15  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
16  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
17  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
18  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
19  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
20  * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
21  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
22  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
23  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
24  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
25  *
26  * $FreeBSD$
27  */
28 
29 #include <sys/param.h>
30 #include <sys/kernel.h>
31 #include <sys/lock.h>
32 #include <sys/malloc.h>
33 #include <sys/mutex.h>
34 #include <sys/proc.h>
35 #include <sys/systm.h>
36 
37 #include <vm/vm.h>
38 #include <vm/pmap.h>
39 #include <vm/uma.h>
40 #include <vm/vm.h>
41 #include <vm/vm_map.h>
42 #include <vm/vm_page.h>
43 #include <vm/vm_pageout.h>
44 
45 #include <machine/md_var.h>
46 #include <machine/platform.h>
47 #include <machine/pmap.h>
48 #include <machine/vmparam.h>
49 
50 uintptr_t moea64_get_unique_vsid(void);
51 void moea64_release_vsid(uint64_t vsid);
52 static void slb_zone_init(void *);
53 
54 static uma_zone_t slbt_zone;
55 static uma_zone_t slb_cache_zone;
56 int n_slbs = 64;
57 
58 SYSINIT(slb_zone_init, SI_SUB_KMEM, SI_ORDER_ANY, slb_zone_init, NULL);
59 
60 struct slbtnode {
61 	uint16_t	ua_alloc;
62 	uint8_t		ua_level;
63 	/* Only 36 bits needed for full 64-bit address space. */
64 	uint64_t	ua_base;
65 	union {
66 		struct slbtnode	*ua_child[16];
67 		struct slb	slb_entries[16];
68 	} u;
69 };
70 
71 /*
72  * For a full 64-bit address space, there are 36 bits in play in an
73  * esid, so 8 levels, with the leaf being at level 0.
74  *
75  * |3333|3322|2222|2222|1111|1111|11  |    |    |  esid
76  * |5432|1098|7654|3210|9876|5432|1098|7654|3210|  bits
77  * +----+----+----+----+----+----+----+----+----+--------
78  * | 8  | 7  | 6  | 5  | 4  | 3  | 2  | 1  | 0  | level
79  */
80 #define UAD_ROOT_LEVEL  8
81 #define UAD_LEAF_LEVEL  0
82 
83 static inline int
84 esid2idx(uint64_t esid, int level)
85 {
86 	int shift;
87 
88 	shift = level * 4;
89 	return ((esid >> shift) & 0xF);
90 }
91 
92 /*
93  * The ua_base field should have 0 bits after the first 4*(level+1)
94  * bits; i.e. only
95  */
96 #define uad_baseok(ua)                          \
97 	(esid2base(ua->ua_base, ua->ua_level) == ua->ua_base)
98 
99 
100 static inline uint64_t
101 esid2base(uint64_t esid, int level)
102 {
103 	uint64_t mask;
104 	int shift;
105 
106 	shift = (level + 1) * 4;
107 	mask = ~((1ULL << shift) - 1);
108 	return (esid & mask);
109 }
110 
111 /*
112  * Allocate a new leaf node for the specified esid/vmhandle from the
113  * parent node.
114  */
115 static struct slb *
116 make_new_leaf(uint64_t esid, uint64_t slbv, struct slbtnode *parent)
117 {
118 	struct slbtnode *child;
119 	struct slb *retval;
120 	int idx;
121 
122 	idx = esid2idx(esid, parent->ua_level);
123 	KASSERT(parent->u.ua_child[idx] == NULL, ("Child already exists!"));
124 
125 	/* unlock and M_WAITOK and loop? */
126 	child = uma_zalloc(slbt_zone, M_NOWAIT | M_ZERO);
127 	KASSERT(child != NULL, ("unhandled NULL case"));
128 
129 	child->ua_level = UAD_LEAF_LEVEL;
130 	child->ua_base = esid2base(esid, child->ua_level);
131 	idx = esid2idx(esid, child->ua_level);
132 	child->u.slb_entries[idx].slbv = slbv;
133 	child->u.slb_entries[idx].slbe = (esid << SLBE_ESID_SHIFT) | SLBE_VALID;
134 	setbit(&child->ua_alloc, idx);
135 
136 	retval = &child->u.slb_entries[idx];
137 
138 	/*
139 	 * The above stores must be visible before the next one, so
140 	 * that a lockless searcher always sees a valid path through
141 	 * the tree.
142 	 */
143 	powerpc_lwsync();
144 
145 	idx = esid2idx(esid, parent->ua_level);
146 	parent->u.ua_child[idx] = child;
147 	setbit(&parent->ua_alloc, idx);
148 
149 	return (retval);
150 }
151 
152 /*
153  * Allocate a new intermediate node to fit between the parent and
154  * esid.
155  */
156 static struct slbtnode*
157 make_intermediate(uint64_t esid, struct slbtnode *parent)
158 {
159 	struct slbtnode *child, *inter;
160 	int idx, level;
161 
162 	idx = esid2idx(esid, parent->ua_level);
163 	child = parent->u.ua_child[idx];
164 	KASSERT(esid2base(esid, child->ua_level) != child->ua_base,
165 	    ("No need for an intermediate node?"));
166 
167 	/*
168 	 * Find the level where the existing child and our new esid
169 	 * meet.  It must be lower than parent->ua_level or we would
170 	 * have chosen a different index in parent.
171 	 */
172 	level = child->ua_level + 1;
173 	while (esid2base(esid, level) !=
174 	    esid2base(child->ua_base, level))
175 		level++;
176 	KASSERT(level < parent->ua_level,
177 	    ("Found splitting level %d for %09jx and %09jx, "
178 	    "but it's the same as %p's",
179 	    level, esid, child->ua_base, parent));
180 
181 	/* unlock and M_WAITOK and loop? */
182 	inter = uma_zalloc(slbt_zone, M_NOWAIT | M_ZERO);
183 	KASSERT(inter != NULL, ("unhandled NULL case"));
184 
185 	/* Set up intermediate node to point to child ... */
186 	inter->ua_level = level;
187 	inter->ua_base = esid2base(esid, inter->ua_level);
188 	idx = esid2idx(child->ua_base, inter->ua_level);
189 	inter->u.ua_child[idx] = child;
190 	setbit(&inter->ua_alloc, idx);
191 	powerpc_lwsync();
192 
193 	/* Set up parent to point to intermediate node ... */
194 	idx = esid2idx(inter->ua_base, parent->ua_level);
195 	parent->u.ua_child[idx] = inter;
196 	setbit(&parent->ua_alloc, idx);
197 
198 	return (inter);
199 }
200 
201 uint64_t
202 kernel_va_to_slbv(vm_offset_t va)
203 {
204 	uint64_t slbv;
205 
206 	/* Set kernel VSID to deterministic value */
207 	slbv = (KERNEL_VSID((uintptr_t)va >> ADDR_SR_SHFT)) << SLBV_VSID_SHIFT;
208 
209 	/* Figure out if this is a large-page mapping */
210 	if (hw_direct_map && va < VM_MIN_KERNEL_ADDRESS) {
211 		/*
212 		 * XXX: If we have set up a direct map, assumes
213 		 * all physical memory is mapped with large pages.
214 		 */
215 		if (mem_valid(va, 0) == 0)
216 			slbv |= SLBV_L;
217 	}
218 
219 	return (slbv);
220 }
221 
222 struct slb *
223 user_va_to_slb_entry(pmap_t pm, vm_offset_t va)
224 {
225 	uint64_t esid = va >> ADDR_SR_SHFT;
226 	struct slbtnode *ua;
227 	int idx;
228 
229 	ua = pm->pm_slb_tree_root;
230 
231 	for (;;) {
232 		KASSERT(uad_baseok(ua), ("uad base %016jx level %d bad!",
233 		    ua->ua_base, ua->ua_level));
234 		idx = esid2idx(esid, ua->ua_level);
235 
236 		/*
237 		 * This code is specific to ppc64 where a load is
238 		 * atomic, so no need for atomic_load macro.
239 		 */
240 		if (ua->ua_level == UAD_LEAF_LEVEL)
241 			return ((ua->u.slb_entries[idx].slbe & SLBE_VALID) ?
242 			    &ua->u.slb_entries[idx] : NULL);
243 
244 		/*
245 		 * The following accesses are implicitly ordered under the POWER
246 		 * ISA by load dependencies (the store ordering is provided by
247 		 * the powerpc_lwsync() calls elsewhere) and so are run without
248 		 * barriers.
249 		 */
250 		ua = ua->u.ua_child[idx];
251 		if (ua == NULL ||
252 		    esid2base(esid, ua->ua_level) != ua->ua_base)
253 			return (NULL);
254 	}
255 
256 	return (NULL);
257 }
258 
259 uint64_t
260 va_to_vsid(pmap_t pm, vm_offset_t va)
261 {
262 	struct slb *entry;
263 
264 	/* Shortcut kernel case */
265 	if (pm == kernel_pmap)
266 		return (KERNEL_VSID((uintptr_t)va >> ADDR_SR_SHFT));
267 
268 	/*
269 	 * If there is no vsid for this VA, we need to add a new entry
270 	 * to the PMAP's segment table.
271 	 */
272 
273 	entry = user_va_to_slb_entry(pm, va);
274 
275 	if (entry == NULL)
276 		return (allocate_user_vsid(pm,
277 		    (uintptr_t)va >> ADDR_SR_SHFT, 0));
278 
279 	return ((entry->slbv & SLBV_VSID_MASK) >> SLBV_VSID_SHIFT);
280 }
281 
282 uint64_t
283 allocate_user_vsid(pmap_t pm, uint64_t esid, int large)
284 {
285 	uint64_t vsid, slbv;
286 	struct slbtnode *ua, *next, *inter;
287 	struct slb *slb;
288 	int idx;
289 
290 	KASSERT(pm != kernel_pmap, ("Attempting to allocate a kernel VSID"));
291 
292 	PMAP_LOCK_ASSERT(pm, MA_OWNED);
293 	vsid = moea64_get_unique_vsid();
294 
295 	slbv = vsid << SLBV_VSID_SHIFT;
296 	if (large)
297 		slbv |= SLBV_L;
298 
299 	ua = pm->pm_slb_tree_root;
300 
301 	/* Descend to the correct leaf or NULL pointer. */
302 	for (;;) {
303 		KASSERT(uad_baseok(ua),
304 		   ("uad base %09jx level %d bad!", ua->ua_base, ua->ua_level));
305 		idx = esid2idx(esid, ua->ua_level);
306 
307 		if (ua->ua_level == UAD_LEAF_LEVEL) {
308 			ua->u.slb_entries[idx].slbv = slbv;
309 			eieio();
310 			ua->u.slb_entries[idx].slbe = (esid << SLBE_ESID_SHIFT)
311 			    | SLBE_VALID;
312 			setbit(&ua->ua_alloc, idx);
313 			slb = &ua->u.slb_entries[idx];
314 			break;
315 		}
316 
317 		next = ua->u.ua_child[idx];
318 		if (next == NULL) {
319 			slb = make_new_leaf(esid, slbv, ua);
320 			break;
321                 }
322 
323 		/*
324 		 * Check if the next item down has an okay ua_base.
325 		 * If not, we need to allocate an intermediate node.
326 		 */
327 		if (esid2base(esid, next->ua_level) != next->ua_base) {
328 			inter = make_intermediate(esid, ua);
329 			slb = make_new_leaf(esid, slbv, inter);
330 			break;
331 		}
332 
333 		ua = next;
334 	}
335 
336 	/*
337 	 * Someone probably wants this soon, and it may be a wired
338 	 * SLB mapping, so pre-spill this entry.
339 	 */
340 	eieio();
341 	slb_insert_user(pm, slb);
342 
343 	return (vsid);
344 }
345 
346 void
347 free_vsid(pmap_t pm, uint64_t esid, int large)
348 {
349 	struct slbtnode *ua;
350 	int idx;
351 
352 	PMAP_LOCK_ASSERT(pm, MA_OWNED);
353 
354 	ua = pm->pm_slb_tree_root;
355 	/* Descend to the correct leaf. */
356 	for (;;) {
357 		KASSERT(uad_baseok(ua),
358 		   ("uad base %09jx level %d bad!", ua->ua_base, ua->ua_level));
359 
360 		idx = esid2idx(esid, ua->ua_level);
361 		if (ua->ua_level == UAD_LEAF_LEVEL) {
362 			ua->u.slb_entries[idx].slbv = 0;
363 			eieio();
364 			ua->u.slb_entries[idx].slbe = 0;
365 			clrbit(&ua->ua_alloc, idx);
366 			return;
367 		}
368 
369 		ua = ua->u.ua_child[idx];
370 		if (ua == NULL ||
371 		    esid2base(esid, ua->ua_level) != ua->ua_base) {
372 			/* Perhaps just return instead of assert? */
373 			KASSERT(0,
374 			    ("Asked to remove an entry that was never inserted!"));
375 			return;
376 		}
377 	}
378 }
379 
380 static void
381 free_slb_tree_node(struct slbtnode *ua)
382 {
383 	int idx;
384 
385 	for (idx = 0; idx < 16; idx++) {
386 		if (ua->ua_level != UAD_LEAF_LEVEL) {
387 			if (ua->u.ua_child[idx] != NULL)
388 				free_slb_tree_node(ua->u.ua_child[idx]);
389 		} else {
390 			if (ua->u.slb_entries[idx].slbv != 0)
391 				moea64_release_vsid(ua->u.slb_entries[idx].slbv
392 				    >> SLBV_VSID_SHIFT);
393 		}
394 	}
395 
396 	uma_zfree(slbt_zone, ua);
397 }
398 
399 void
400 slb_free_tree(pmap_t pm)
401 {
402 
403 	free_slb_tree_node(pm->pm_slb_tree_root);
404 }
405 
406 struct slbtnode *
407 slb_alloc_tree(void)
408 {
409 	struct slbtnode *root;
410 
411 	root = uma_zalloc(slbt_zone, M_NOWAIT | M_ZERO);
412 	root->ua_level = UAD_ROOT_LEVEL;
413 
414 	return (root);
415 }
416 
417 /* Lock entries mapping kernel text and stacks */
418 
419 void
420 slb_insert_kernel(uint64_t slbe, uint64_t slbv)
421 {
422 	struct slb *slbcache;
423 	int i;
424 
425 	/* We don't want to be preempted while modifying the kernel map */
426 	critical_enter();
427 
428 	slbcache = PCPU_GET(slb);
429 
430 	/* Check for an unused slot, abusing the user slot as a full flag */
431 	if (slbcache[USER_SLB_SLOT].slbe == 0) {
432 		for (i = 0; i < n_slbs; i++) {
433 			if (i == USER_SLB_SLOT)
434 				continue;
435 			if (!(slbcache[i].slbe & SLBE_VALID))
436 				goto fillkernslb;
437 		}
438 
439 		if (i == n_slbs)
440 			slbcache[USER_SLB_SLOT].slbe = 1;
441 	}
442 
443 	i = mftb() % n_slbs;
444 	if (i == USER_SLB_SLOT)
445 			i = (i+1) % n_slbs;
446 
447 fillkernslb:
448 	KASSERT(i != USER_SLB_SLOT,
449 	    ("Filling user SLB slot with a kernel mapping"));
450 	slbcache[i].slbv = slbv;
451 	slbcache[i].slbe = slbe | (uint64_t)i;
452 
453 	/* If it is for this CPU, put it in the SLB right away */
454 	if (pmap_bootstrapped) {
455 		/* slbie not required */
456 		__asm __volatile ("slbmte %0, %1" ::
457 		    "r"(slbcache[i].slbv), "r"(slbcache[i].slbe));
458 	}
459 
460 	critical_exit();
461 }
462 
463 void
464 slb_insert_user(pmap_t pm, struct slb *slb)
465 {
466 	int i;
467 
468 	PMAP_LOCK_ASSERT(pm, MA_OWNED);
469 
470 	if (pm->pm_slb_len < n_slbs) {
471 		i = pm->pm_slb_len;
472 		pm->pm_slb_len++;
473 	} else {
474 		i = mftb() % n_slbs;
475 	}
476 
477 	/* Note that this replacement is atomic with respect to trap_subr */
478 	pm->pm_slb[i] = slb;
479 }
480 
481 static void *
482 slb_uma_real_alloc(uma_zone_t zone, vm_size_t bytes, u_int8_t *flags, int wait)
483 {
484 	static vm_offset_t realmax = 0;
485 	void *va;
486 	vm_page_t m;
487 	int pflags;
488 
489 	if (realmax == 0)
490 		realmax = platform_real_maxaddr();
491 
492 	*flags = UMA_SLAB_PRIV;
493 	pflags = malloc2vm_flags(wait) | VM_ALLOC_NOOBJ | VM_ALLOC_WIRED;
494 
495 	for (;;) {
496 		m = vm_page_alloc_contig(NULL, 0, pflags, 1, 0, realmax,
497 		    PAGE_SIZE, PAGE_SIZE, VM_MEMATTR_DEFAULT);
498 		if (m == NULL) {
499 			if (wait & M_NOWAIT)
500 				return (NULL);
501 			VM_WAIT;
502 		} else
503                         break;
504         }
505 
506 	va = (void *) VM_PAGE_TO_PHYS(m);
507 
508 	if (!hw_direct_map)
509 		pmap_kenter((vm_offset_t)va, VM_PAGE_TO_PHYS(m));
510 
511 	if ((wait & M_ZERO) && (m->flags & PG_ZERO) == 0)
512 		bzero(va, PAGE_SIZE);
513 
514 	return (va);
515 }
516 
517 static void
518 slb_zone_init(void *dummy)
519 {
520 
521 	slbt_zone = uma_zcreate("SLB tree node", sizeof(struct slbtnode),
522 	    NULL, NULL, NULL, NULL, UMA_ALIGN_PTR, UMA_ZONE_VM);
523 	slb_cache_zone = uma_zcreate("SLB cache",
524 	    (n_slbs + 1)*sizeof(struct slb *), NULL, NULL, NULL, NULL,
525 	    UMA_ALIGN_PTR, UMA_ZONE_VM);
526 
527 	if (platform_real_maxaddr() != VM_MAX_ADDRESS) {
528 		uma_zone_set_allocf(slb_cache_zone, slb_uma_real_alloc);
529 		uma_zone_set_allocf(slbt_zone, slb_uma_real_alloc);
530 	}
531 }
532 
533 struct slb **
534 slb_alloc_user_cache(void)
535 {
536 	return (uma_zalloc(slb_cache_zone, M_ZERO));
537 }
538 
539 void
540 slb_free_user_cache(struct slb **slb)
541 {
542 	uma_zfree(slb_cache_zone, slb);
543 }
544