xref: /titanic_44/usr/src/uts/sun4u/opl/io/dr_mem.c (revision d0538f66491267879b7418b21ad78e3dcc2dcc83)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2007 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 #pragma ident	"%Z%%M%	%I%	%E% SMI"
27 
28 /*
29  * DR memory support routines.
30  */
31 
32 #include <sys/note.h>
33 #include <sys/debug.h>
34 #include <sys/types.h>
35 #include <sys/errno.h>
36 #include <sys/param.h>
37 #include <sys/dditypes.h>
38 #include <sys/kmem.h>
39 #include <sys/conf.h>
40 #include <sys/ddi.h>
41 #include <sys/sunddi.h>
42 #include <sys/sunndi.h>
43 #include <sys/ddi_impldefs.h>
44 #include <sys/ndi_impldefs.h>
45 #include <sys/sysmacros.h>
46 #include <sys/machsystm.h>
47 #include <sys/spitregs.h>
48 #include <sys/cpuvar.h>
49 #include <sys/promif.h>
50 #include <vm/seg_kmem.h>
51 #include <sys/lgrp.h>
52 #include <sys/platform_module.h>
53 
54 #include <vm/page.h>
55 
56 #include <sys/dr.h>
57 #include <sys/dr_util.h>
58 #include <sys/drmach.h>
59 #include <sys/kobj.h>
60 
61 extern struct memlist	*phys_install;
62 extern vnode_t		*retired_pages;
63 
64 /* TODO: push this reference below drmach line */
65 extern int		kcage_on;
66 
67 /* for the DR*INTERNAL_ERROR macros.  see sys/dr.h. */
68 static char *dr_ie_fmt = "dr_mem.c %d";
69 
70 typedef enum {
71 	DR_TP_INVALID = -1,
72 	DR_TP_SAME,
73 	DR_TP_LARGE,
74 	DR_TP_NONRELOC,
75 	DR_TP_FLOATING
76 } dr_target_pref_t;
77 
78 static int		dr_post_detach_mem_unit(dr_mem_unit_t *mp);
79 static int		dr_reserve_mem_spans(memhandle_t *mhp,
80 				struct memlist *mlist);
81 static int		dr_select_mem_target(dr_handle_t *hp,
82 				dr_mem_unit_t *mp, struct memlist *ml);
83 static void		dr_init_mem_unit_data(dr_mem_unit_t *mp);
84 static struct memlist	*dr_memlist_del_retired_pages(struct memlist *ml);
85 static dr_target_pref_t	dr_get_target_preference(dr_handle_t *hp,
86 				dr_mem_unit_t *t_mp, dr_mem_unit_t *s_mp,
87 				struct memlist *s_ml, struct memlist *x_ml,
88 				struct memlist *b_ml);
89 
90 static int		memlist_canfit(struct memlist *s_mlist,
91 				struct memlist *t_mlist);
92 static int		dr_del_mlist_query(struct memlist *mlist,
93 				memquery_t *mp);
94 static struct memlist	*dr_get_copy_mlist(struct memlist *s_ml,
95 				struct memlist *t_ml, dr_mem_unit_t *s_mp,
96 				dr_mem_unit_t *t_mp);
97 static struct memlist	*dr_get_nonreloc_mlist(struct memlist *s_ml,
98 				dr_mem_unit_t *s_mp);
99 static int		dr_memlist_canfit(struct memlist *s_mlist,
100 				struct memlist *t_mlist, dr_mem_unit_t *s_mp,
101 				dr_mem_unit_t *t_mp);
102 
103 /*
104  * dr_mem_unit_t.sbm_flags
105  */
106 #define	DR_MFLAG_RESERVED	0x01	/* mem unit reserved for delete */
107 #define	DR_MFLAG_SOURCE		0x02	/* source brd of copy/rename op */
108 #define	DR_MFLAG_TARGET		0x04	/* target brd of copy/rename op */
109 #define	DR_MFLAG_RELOWNER	0x20	/* memory release (delete) owner */
110 #define	DR_MFLAG_RELDONE	0x40	/* memory release (delete) done */
111 
112 /* helper macros */
113 #define	_ptob64(p) ((uint64_t)(p) << PAGESHIFT)
114 #define	_b64top(b) ((pgcnt_t)((b) >> PAGESHIFT))
115 
116 static struct memlist *
117 dr_get_memlist(dr_mem_unit_t *mp)
118 {
119 	struct memlist	*mlist = NULL;
120 	sbd_error_t	*err;
121 	static fn_t	f = "dr_get_memlist";
122 
123 	PR_MEM("%s for %s...\n", f, mp->sbm_cm.sbdev_path);
124 
125 	/*
126 	 * Return cached memlist, if present.
127 	 * This memlist will be present following an
128 	 * unconfigure (a.k.a: detach) of this memunit.
129 	 * It should only be used in the case were a configure
130 	 * is bringing this memunit back in without going
131 	 * through the disconnect and connect states.
132 	 */
133 	if (mp->sbm_mlist) {
134 		PR_MEM("%s: found cached memlist\n", f);
135 
136 		mlist = memlist_dup(mp->sbm_mlist);
137 	} else {
138 		uint64_t basepa = _ptob64(mp->sbm_basepfn);
139 
140 		/* attempt to construct a memlist using phys_install */
141 
142 		/* round down to slice base address */
143 		basepa &= ~(mp->sbm_slice_size - 1);
144 
145 		/* get a copy of phys_install to edit */
146 		memlist_read_lock();
147 		mlist = memlist_dup(phys_install);
148 		memlist_read_unlock();
149 
150 		/* trim lower irrelevant span */
151 		if (mlist)
152 			mlist = memlist_del_span(mlist, 0ull, basepa);
153 
154 		/* trim upper irrelevant span */
155 		if (mlist) {
156 			uint64_t endpa;
157 
158 			basepa += mp->sbm_slice_size;
159 			endpa = _ptob64(physmax + 1);
160 			if (endpa > basepa)
161 				mlist = memlist_del_span(
162 				    mlist, basepa,
163 				    endpa - basepa);
164 		}
165 
166 		if (mlist) {
167 			/* successfully built a memlist */
168 			PR_MEM("%s: derived memlist from phys_install\n", f);
169 		}
170 
171 		/* if no mlist yet, try platform layer */
172 		if (!mlist) {
173 			err = drmach_mem_get_memlist(
174 			    mp->sbm_cm.sbdev_id, &mlist);
175 			if (err) {
176 				DRERR_SET_C(&mp->sbm_cm.sbdev_error, &err);
177 				mlist = NULL; /* paranoia */
178 			}
179 		}
180 	}
181 
182 	PR_MEM("%s: memlist for %s\n", f, mp->sbm_cm.sbdev_path);
183 	PR_MEMLIST_DUMP(mlist);
184 
185 	return (mlist);
186 }
187 
188 typedef struct {
189 	kcondvar_t cond;
190 	kmutex_t lock;
191 	int error;
192 	int done;
193 } dr_release_mem_sync_t;
194 
195 /*
196  * Memory has been logically removed by the time this routine is called.
197  */
198 static void
199 dr_mem_del_done(void *arg, int error)
200 {
201 	dr_release_mem_sync_t *ds = arg;
202 
203 	mutex_enter(&ds->lock);
204 	ds->error = error;
205 	ds->done = 1;
206 	cv_signal(&ds->cond);
207 	mutex_exit(&ds->lock);
208 }
209 
210 /*
211  * When we reach here the memory being drained should have
212  * already been reserved in dr_pre_release_mem().
213  * Our only task here is to kick off the "drain" and wait
214  * for it to finish.
215  */
216 void
217 dr_release_mem(dr_common_unit_t *cp)
218 {
219 	dr_mem_unit_t	*mp = (dr_mem_unit_t *)cp;
220 	int		err;
221 	dr_release_mem_sync_t rms;
222 	static fn_t	f = "dr_release_mem";
223 
224 	/* check that this memory unit has been reserved */
225 	if (!(mp->sbm_flags & DR_MFLAG_RELOWNER)) {
226 		DR_DEV_INTERNAL_ERROR(&mp->sbm_cm);
227 		return;
228 	}
229 
230 	bzero((void *) &rms, sizeof (rms));
231 
232 	mutex_init(&rms.lock, NULL, MUTEX_DRIVER, NULL);
233 	cv_init(&rms.cond, NULL, CV_DRIVER, NULL);
234 
235 	mutex_enter(&rms.lock);
236 	err = kphysm_del_start(mp->sbm_memhandle,
237 	    dr_mem_del_done, (void *) &rms);
238 	if (err == KPHYSM_OK) {
239 		/* wait for completion or interrupt */
240 		while (!rms.done) {
241 			if (cv_wait_sig(&rms.cond, &rms.lock) == 0) {
242 				/* then there is a pending UNIX signal */
243 				(void) kphysm_del_cancel(mp->sbm_memhandle);
244 
245 				/* wait for completion */
246 				while (!rms.done)
247 					cv_wait(&rms.cond, &rms.lock);
248 			}
249 		}
250 		/* get the result of the memory delete operation */
251 		err = rms.error;
252 	}
253 	mutex_exit(&rms.lock);
254 
255 	cv_destroy(&rms.cond);
256 	mutex_destroy(&rms.lock);
257 
258 	if (err != KPHYSM_OK) {
259 		int e_code;
260 
261 		switch (err) {
262 			case KPHYSM_ENOWORK:
263 				e_code = ESBD_NOERROR;
264 				break;
265 
266 			case KPHYSM_EHANDLE:
267 			case KPHYSM_ESEQUENCE:
268 				e_code = ESBD_INTERNAL;
269 				break;
270 
271 			case KPHYSM_ENOTVIABLE:
272 				e_code = ESBD_MEM_NOTVIABLE;
273 				break;
274 
275 			case KPHYSM_EREFUSED:
276 				e_code = ESBD_MEM_REFUSED;
277 				break;
278 
279 			case KPHYSM_ENONRELOC:
280 				e_code = ESBD_MEM_NONRELOC;
281 				break;
282 
283 			case KPHYSM_ECANCELLED:
284 				e_code = ESBD_MEM_CANCELLED;
285 				break;
286 
287 			case KPHYSM_ERESOURCE:
288 				e_code = ESBD_MEMFAIL;
289 				break;
290 
291 			default:
292 				cmn_err(CE_WARN,
293 				    "%s: unexpected kphysm error code %d,"
294 				    " id 0x%p",
295 				    f, err, mp->sbm_cm.sbdev_id);
296 
297 				e_code = ESBD_IO;
298 				break;
299 		}
300 
301 		if (e_code != ESBD_NOERROR) {
302 			dr_dev_err(CE_WARN, &mp->sbm_cm, e_code);
303 		}
304 	}
305 }
306 
307 void
308 dr_attach_mem(dr_handle_t *hp, dr_common_unit_t *cp)
309 {
310 	_NOTE(ARGUNUSED(hp))
311 
312 	dr_mem_unit_t	*mp = (dr_mem_unit_t *)cp;
313 	struct memlist	*ml, *mc;
314 	sbd_error_t	*err;
315 	static fn_t	f = "dr_attach_mem";
316 
317 	PR_MEM("%s...\n", f);
318 
319 	dr_lock_status(hp->h_bd);
320 	err = drmach_configure(cp->sbdev_id, 0);
321 	dr_unlock_status(hp->h_bd);
322 	if (err) {
323 		DRERR_SET_C(&cp->sbdev_error, &err);
324 		return;
325 	}
326 
327 	ml = dr_get_memlist(mp);
328 	for (mc = ml; mc; mc = mc->next) {
329 		int		 rv;
330 		sbd_error_t	*err;
331 
332 		rv = kphysm_add_memory_dynamic(
333 		    (pfn_t)(mc->address >> PAGESHIFT),
334 		    (pgcnt_t)(mc->size >> PAGESHIFT));
335 		if (rv != KPHYSM_OK) {
336 			/*
337 			 * translate kphysm error and
338 			 * store in devlist error
339 			 */
340 			switch (rv) {
341 			case KPHYSM_ERESOURCE:
342 				rv = ESBD_NOMEM;
343 				break;
344 
345 			case KPHYSM_EFAULT:
346 				rv = ESBD_FAULT;
347 				break;
348 
349 			default:
350 				rv = ESBD_INTERNAL;
351 				break;
352 			}
353 
354 			if (rv == ESBD_INTERNAL) {
355 				DR_DEV_INTERNAL_ERROR(&mp->sbm_cm);
356 			} else
357 				dr_dev_err(CE_WARN, &mp->sbm_cm, rv);
358 			break;
359 		}
360 
361 		err = drmach_mem_add_span(
362 		    mp->sbm_cm.sbdev_id, mc->address, mc->size);
363 		if (err) {
364 			DRERR_SET_C(&mp->sbm_cm.sbdev_error, &err);
365 			break;
366 		}
367 	}
368 
369 	memlist_delete(ml);
370 
371 	/* back out if configure failed */
372 	if (mp->sbm_cm.sbdev_error != NULL) {
373 		dr_lock_status(hp->h_bd);
374 		err = drmach_unconfigure(cp->sbdev_id, 0);
375 		if (err)
376 			sbd_err_clear(&err);
377 		dr_unlock_status(hp->h_bd);
378 	}
379 }
380 
381 static struct memlist *
382 dr_memlist_del_retired_pages(struct memlist *mlist)
383 {
384 	page_t		*pp;
385 	pfn_t		pfn;
386 	kmutex_t	*vphm;
387 	vnode_t		*vp = retired_pages;
388 	static fn_t	f = "dr_memlist_del_retired_pages";
389 
390 	vphm = page_vnode_mutex(vp);
391 	mutex_enter(vphm);
392 
393 	PR_MEM("%s\n", f);
394 
395 	if ((pp = vp->v_pages) == NULL) {
396 		mutex_exit(vphm);
397 		return (mlist);
398 	}
399 
400 	do {
401 		ASSERT(pp != NULL);
402 		ASSERT(pp->p_vnode == retired_pages);
403 
404 		if (!page_try_reclaim_lock(pp, SE_SHARED, SE_RETIRED))
405 			continue;
406 
407 		pfn = page_pptonum(pp);
408 
409 		ASSERT((pp->p_offset >> PAGESHIFT) == pfn);
410 		/*
411 		 * Page retirement currently breaks large pages into PAGESIZE
412 		 * pages. If this changes, need to remove the assert and deal
413 		 * with different page sizes.
414 		 */
415 		ASSERT(pp->p_szc == 0);
416 
417 		if (address_in_memlist(mlist, ptob(pfn), PAGESIZE)) {
418 			mlist = memlist_del_span(mlist, ptob(pfn), PAGESIZE);
419 			PR_MEM("deleted retired page 0x%lx (pfn 0x%lx) "
420 			    "from memlist\n", ptob(pfn), pfn);
421 		}
422 
423 		page_unlock(pp);
424 	} while ((pp = pp->p_vpnext) != vp->v_pages);
425 
426 	mutex_exit(vphm);
427 
428 	return (mlist);
429 }
430 
431 static int
432 dr_move_memory(dr_handle_t *hp, dr_mem_unit_t *s_mp, dr_mem_unit_t *t_mp)
433 {
434 	int		rv = -1;
435 	time_t		 copytime;
436 	drmachid_t	 cr_id;
437 	dr_sr_handle_t	*srhp = NULL;
438 	dr_board_t	*t_bp, *s_bp;
439 	struct memlist	*c_ml, *d_ml;
440 	sbd_error_t	*err;
441 	static fn_t	 f = "dr_move_memory";
442 
443 	PR_MEM("%s: (INLINE) moving memory from %s to %s\n",
444 	    f,
445 	    s_mp->sbm_cm.sbdev_path,
446 	    t_mp->sbm_cm.sbdev_path);
447 
448 	ASSERT(s_mp->sbm_flags & DR_MFLAG_SOURCE);
449 	ASSERT(s_mp->sbm_peer == t_mp);
450 	ASSERT(s_mp->sbm_mlist);
451 
452 	ASSERT(t_mp->sbm_flags & DR_MFLAG_TARGET);
453 	ASSERT(t_mp->sbm_peer == s_mp);
454 
455 	/*
456 	 * create a memlist of spans to copy by removing
457 	 * the spans that have been deleted, if any, from
458 	 * the full source board memlist.  s_mp->sbm_del_mlist
459 	 * will be NULL if there were no spans deleted from
460 	 * the source board.
461 	 */
462 	c_ml = memlist_dup(s_mp->sbm_mlist);
463 	d_ml = s_mp->sbm_del_mlist;
464 	while (d_ml != NULL) {
465 		c_ml = memlist_del_span(c_ml, d_ml->address, d_ml->size);
466 		d_ml = d_ml->next;
467 	}
468 
469 	/*
470 	 * Remove retired pages from the copy list. The page content
471 	 * need not be copied since the pages are no longer in use.
472 	 */
473 	PR_MEM("%s: copy list before removing retired pages (if any):\n", f);
474 	PR_MEMLIST_DUMP(c_ml);
475 
476 	c_ml = dr_memlist_del_retired_pages(c_ml);
477 
478 	PR_MEM("%s: copy list after removing retired pages:\n", f);
479 	PR_MEMLIST_DUMP(c_ml);
480 
481 	/*
482 	 * With parallel copy, it shouldn't make a difference which
483 	 * CPU is the actual master during copy-rename since all
484 	 * CPUs participate in the parallel copy anyway.
485 	 */
486 	affinity_set(CPU_CURRENT);
487 
488 	err = drmach_copy_rename_init(
489 	    t_mp->sbm_cm.sbdev_id, s_mp->sbm_cm.sbdev_id, c_ml, &cr_id);
490 	if (err) {
491 		DRERR_SET_C(&s_mp->sbm_cm.sbdev_error, &err);
492 		affinity_clear();
493 		memlist_delete(c_ml);
494 		return (-1);
495 	}
496 
497 	srhp = dr_get_sr_handle(hp);
498 	ASSERT(srhp);
499 
500 	copytime = lbolt;
501 
502 	/* Quiesce the OS.  */
503 	if (dr_suspend(srhp)) {
504 		cmn_err(CE_WARN, "%s: failed to quiesce OS"
505 		    " for copy-rename", f);
506 
507 		err = drmach_copy_rename_fini(cr_id);
508 		if (err) {
509 			/*
510 			 * no error is expected since the program has
511 			 * not yet run.
512 			 */
513 
514 			/* catch this in debug kernels */
515 			ASSERT(0);
516 
517 			sbd_err_clear(&err);
518 		}
519 
520 		/* suspend error reached via hp */
521 		s_mp->sbm_cm.sbdev_error = hp->h_err;
522 		hp->h_err = NULL;
523 		goto done;
524 	}
525 
526 	drmach_copy_rename(cr_id);
527 
528 	/* Resume the OS.  */
529 	dr_resume(srhp);
530 
531 	copytime = lbolt - copytime;
532 
533 	if (err = drmach_copy_rename_fini(cr_id))
534 		goto done;
535 
536 	/*
537 	 * Rename memory for lgroup.
538 	 * Source and target board numbers are packaged in arg.
539 	 */
540 	s_bp = s_mp->sbm_cm.sbdev_bp;
541 	t_bp = t_mp->sbm_cm.sbdev_bp;
542 
543 	lgrp_plat_config(LGRP_CONFIG_MEM_RENAME,
544 	    (uintptr_t)(s_bp->b_num | (t_bp->b_num << 16)));
545 
546 
547 	PR_MEM("%s: copy-rename elapsed time = %ld ticks (%ld secs)\n",
548 	    f, copytime, copytime / hz);
549 
550 	rv = 0;
551 done:
552 	if (srhp)
553 		dr_release_sr_handle(srhp);
554 	if (err)
555 		DRERR_SET_C(&s_mp->sbm_cm.sbdev_error, &err);
556 	affinity_clear();
557 
558 	return (rv);
559 }
560 
561 /*
562  * If detaching node contains memory that is "non-permanent"
563  * then the memory adr's are simply cleared.  If the memory
564  * is non-relocatable, then do a copy-rename.
565  */
566 void
567 dr_detach_mem(dr_handle_t *hp, dr_common_unit_t *cp)
568 {
569 	int			rv = 0;
570 	dr_mem_unit_t		*s_mp = (dr_mem_unit_t *)cp;
571 	dr_mem_unit_t		*t_mp;
572 	dr_state_t		state;
573 	static fn_t		f = "dr_detach_mem";
574 
575 	PR_MEM("%s...\n", f);
576 
577 	/* lookup target mem unit and target board structure, if any */
578 	if (s_mp->sbm_flags & DR_MFLAG_SOURCE) {
579 		t_mp = s_mp->sbm_peer;
580 		ASSERT(t_mp != NULL);
581 		ASSERT(t_mp->sbm_peer == s_mp);
582 	} else {
583 		t_mp = NULL;
584 	}
585 
586 	/* verify mem unit's state is UNREFERENCED */
587 	state = s_mp->sbm_cm.sbdev_state;
588 	if (state != DR_STATE_UNREFERENCED) {
589 		dr_dev_err(CE_IGNORE, &s_mp->sbm_cm, ESBD_STATE);
590 		return;
591 	}
592 
593 	/* verify target mem unit's state is UNREFERENCED, if any */
594 	if (t_mp != NULL) {
595 		state = t_mp->sbm_cm.sbdev_state;
596 		if (state != DR_STATE_UNREFERENCED) {
597 			dr_dev_err(CE_IGNORE, &t_mp->sbm_cm, ESBD_STATE);
598 			return;
599 		}
600 	}
601 
602 	/*
603 	 * If there is no target board (no copy/rename was needed), then
604 	 * we're done!
605 	 */
606 	if (t_mp == NULL) {
607 		sbd_error_t *err;
608 		/*
609 		 * Reprogram interconnect hardware and disable
610 		 * memory controllers for memory node that's going away.
611 		 */
612 
613 		err = drmach_mem_disable(s_mp->sbm_cm.sbdev_id);
614 		if (err) {
615 			DRERR_SET_C(&s_mp->sbm_cm.sbdev_error, &err);
616 			rv = -1;
617 		}
618 	} else {
619 		rv = dr_move_memory(hp, s_mp, t_mp);
620 		PR_MEM("%s: %s memory COPY-RENAME (board %d -> %d)\n",
621 		    f,
622 		    rv ? "FAILED" : "COMPLETED",
623 		    s_mp->sbm_cm.sbdev_bp->b_num,
624 		    t_mp->sbm_cm.sbdev_bp->b_num);
625 
626 		if (rv != 0)
627 			(void) dr_cancel_mem(s_mp);
628 	}
629 
630 	if (rv == 0) {
631 		sbd_error_t *err;
632 
633 		dr_lock_status(hp->h_bd);
634 		err = drmach_unconfigure(s_mp->sbm_cm.sbdev_id, 0);
635 		dr_unlock_status(hp->h_bd);
636 		if (err)
637 			sbd_err_clear(&err);
638 	}
639 }
640 
641 /*
642  * This routine acts as a wrapper for kphysm_del_span_query in order to
643  * support potential memory holes in a board's physical address space.
644  * It calls kphysm_del_span_query for each node in a memlist and accumulates
645  * the results in *mp.
646  */
647 static int
648 dr_del_mlist_query(struct memlist *mlist, memquery_t *mp)
649 {
650 	struct memlist	*ml;
651 	int		 rv = 0;
652 
653 
654 	if (mlist == NULL)
655 		cmn_err(CE_WARN, "dr_del_mlist_query: mlist=NULL\n");
656 
657 	mp->phys_pages = 0;
658 	mp->managed = 0;
659 	mp->nonrelocatable = 0;
660 	mp->first_nonrelocatable = (pfn_t)-1;	/* XXX */
661 	mp->last_nonrelocatable = 0;
662 
663 	for (ml = mlist; ml; ml = ml->next) {
664 		memquery_t mq;
665 
666 		rv = kphysm_del_span_query(
667 		    _b64top(ml->address), _b64top(ml->size), &mq);
668 		if (rv)
669 			break;
670 
671 		mp->phys_pages += mq.phys_pages;
672 		mp->managed += mq.managed;
673 		mp->nonrelocatable += mq.nonrelocatable;
674 
675 		if (mq.nonrelocatable != 0) {
676 			if (mq.first_nonrelocatable < mp->first_nonrelocatable)
677 				mp->first_nonrelocatable =
678 				    mq.first_nonrelocatable;
679 			if (mq.last_nonrelocatable > mp->last_nonrelocatable)
680 				mp->last_nonrelocatable =
681 				    mq.last_nonrelocatable;
682 		}
683 	}
684 
685 	if (mp->nonrelocatable == 0)
686 		mp->first_nonrelocatable = 0;	/* XXX */
687 
688 	return (rv);
689 }
690 
691 /*
692  * NOTE: This routine is only partially smart about multiple
693  *	 mem-units.  Need to make mem-status structure smart
694  *	 about them also.
695  */
696 int
697 dr_mem_status(dr_handle_t *hp, dr_devset_t devset, sbd_dev_stat_t *dsp)
698 {
699 	int		m, mix;
700 	memdelstat_t	mdst;
701 	memquery_t	mq;
702 	dr_board_t	*bp;
703 	dr_mem_unit_t	*mp;
704 	sbd_mem_stat_t	*msp;
705 	static fn_t	f = "dr_mem_status";
706 
707 	bp = hp->h_bd;
708 	devset &= DR_DEVS_PRESENT(bp);
709 
710 	for (m = mix = 0; m < MAX_MEM_UNITS_PER_BOARD; m++) {
711 		int		rv;
712 		sbd_error_t	*err;
713 		drmach_status_t	 pstat;
714 		dr_mem_unit_t	*p_mp;
715 
716 		if (DEVSET_IN_SET(devset, SBD_COMP_MEM, m) == 0)
717 			continue;
718 
719 		mp = dr_get_mem_unit(bp, m);
720 
721 		if (mp->sbm_cm.sbdev_state == DR_STATE_EMPTY) {
722 			/* present, but not fully initialized */
723 			continue;
724 		}
725 
726 		if (mp->sbm_cm.sbdev_id == (drmachid_t)0)
727 			continue;
728 
729 		/* fetch platform status */
730 		err = drmach_status(mp->sbm_cm.sbdev_id, &pstat);
731 		if (err) {
732 			DRERR_SET_C(&mp->sbm_cm.sbdev_error, &err);
733 			continue;
734 		}
735 
736 		msp = &dsp->d_mem;
737 		bzero((caddr_t)msp, sizeof (*msp));
738 
739 		strncpy(msp->ms_cm.c_id.c_name, pstat.type,
740 		    sizeof (msp->ms_cm.c_id.c_name));
741 		msp->ms_cm.c_id.c_type = mp->sbm_cm.sbdev_type;
742 		msp->ms_cm.c_id.c_unit = SBD_NULL_UNIT;
743 		msp->ms_cm.c_cond = mp->sbm_cm.sbdev_cond;
744 		msp->ms_cm.c_busy = mp->sbm_cm.sbdev_busy | pstat.busy;
745 		msp->ms_cm.c_time = mp->sbm_cm.sbdev_time;
746 		msp->ms_cm.c_ostate = mp->sbm_cm.sbdev_ostate;
747 
748 		msp->ms_totpages = mp->sbm_npages;
749 		msp->ms_basepfn = mp->sbm_basepfn;
750 		msp->ms_pageslost = mp->sbm_pageslost;
751 		msp->ms_cage_enabled = kcage_on;
752 
753 		if (mp->sbm_flags & DR_MFLAG_RESERVED)
754 			p_mp = mp->sbm_peer;
755 		else
756 			p_mp = NULL;
757 
758 		if (p_mp == NULL) {
759 			msp->ms_peer_is_target = 0;
760 			msp->ms_peer_ap_id[0] = '\0';
761 		} else if (p_mp->sbm_flags & DR_MFLAG_RESERVED) {
762 			char *path = kmem_alloc(MAXPATHLEN, KM_SLEEP);
763 			char *minor;
764 
765 			/*
766 			 * b_dip doesn't have to be held for ddi_pathname()
767 			 * because the board struct (dr_board_t) will be
768 			 * destroyed before b_dip detaches.
769 			 */
770 			(void) ddi_pathname(bp->b_dip, path);
771 			minor = strchr(p_mp->sbm_cm.sbdev_path, ':');
772 
773 			snprintf(msp->ms_peer_ap_id,
774 			    sizeof (msp->ms_peer_ap_id), "%s%s",
775 			    path, (minor == NULL) ? "" : minor);
776 
777 			kmem_free(path, MAXPATHLEN);
778 
779 			if (p_mp->sbm_flags & DR_MFLAG_TARGET)
780 				msp->ms_peer_is_target = 1;
781 		}
782 
783 		if (mp->sbm_flags & DR_MFLAG_RELOWNER)
784 			rv = kphysm_del_status(mp->sbm_memhandle, &mdst);
785 		else
786 			rv = KPHYSM_EHANDLE;	/* force 'if' to fail */
787 
788 		if (rv == KPHYSM_OK) {
789 			/*
790 			 * Any pages above managed is "free",
791 			 * i.e. it's collected.
792 			 */
793 			msp->ms_detpages += (uint_t)(mdst.collected +
794 			    mdst.phys_pages - mdst.managed);
795 		} else {
796 			/*
797 			 * If we're UNREFERENCED or UNCONFIGURED,
798 			 * then the number of detached pages is
799 			 * however many pages are on the board.
800 			 * I.e. detached = not in use by OS.
801 			 */
802 			switch (msp->ms_cm.c_ostate) {
803 			/*
804 			 * changed to use cfgadm states
805 			 *
806 			 * was:
807 			 *	case DR_STATE_UNREFERENCED:
808 			 *	case DR_STATE_UNCONFIGURED:
809 			 */
810 			case SBD_STAT_UNCONFIGURED:
811 				msp->ms_detpages = msp->ms_totpages;
812 				break;
813 
814 			default:
815 				break;
816 			}
817 		}
818 
819 		/*
820 		 * kphysm_del_span_query can report non-reloc pages = total
821 		 * pages for memory that is not yet configured
822 		 */
823 		if (mp->sbm_cm.sbdev_state != DR_STATE_UNCONFIGURED) {
824 			struct memlist *ml;
825 
826 			ml = dr_get_memlist(mp);
827 			rv = ml ? dr_del_mlist_query(ml, &mq) : -1;
828 			memlist_delete(ml);
829 
830 			if (rv == KPHYSM_OK) {
831 				msp->ms_managed_pages = mq.managed;
832 				msp->ms_noreloc_pages = mq.nonrelocatable;
833 				msp->ms_noreloc_first =
834 				    mq.first_nonrelocatable;
835 				msp->ms_noreloc_last =
836 				    mq.last_nonrelocatable;
837 				msp->ms_cm.c_sflags = 0;
838 				if (mq.nonrelocatable) {
839 					SBD_SET_SUSPEND(SBD_CMD_UNCONFIGURE,
840 					    msp->ms_cm.c_sflags);
841 				}
842 			} else {
843 				PR_MEM("%s: kphysm_del_span_query() = %d\n",
844 				    f, rv);
845 			}
846 		}
847 
848 		/*
849 		 * Check source unit state during copy-rename
850 		 */
851 		if ((mp->sbm_flags & DR_MFLAG_SOURCE) &&
852 		    (mp->sbm_cm.sbdev_state == DR_STATE_UNREFERENCED ||
853 		    mp->sbm_cm.sbdev_state == DR_STATE_RELEASE))
854 			msp->ms_cm.c_ostate = SBD_STAT_CONFIGURED;
855 
856 		mix++;
857 		dsp++;
858 	}
859 
860 	return (mix);
861 }
862 
863 int
864 dr_pre_attach_mem(dr_handle_t *hp, dr_common_unit_t **devlist, int devnum)
865 {
866 	_NOTE(ARGUNUSED(hp))
867 
868 	int		err_flag = 0;
869 	int		d;
870 	sbd_error_t	*err;
871 	static fn_t	f = "dr_pre_attach_mem";
872 
873 	PR_MEM("%s...\n", f);
874 
875 	for (d = 0; d < devnum; d++) {
876 		dr_mem_unit_t	*mp = (dr_mem_unit_t *)devlist[d];
877 		dr_state_t	state;
878 
879 		cmn_err(CE_CONT, "OS configure %s", mp->sbm_cm.sbdev_path);
880 
881 		state = mp->sbm_cm.sbdev_state;
882 		switch (state) {
883 		case DR_STATE_UNCONFIGURED:
884 			PR_MEM("%s: recovering from UNCONFIG for %s\n",
885 			    f,
886 			    mp->sbm_cm.sbdev_path);
887 
888 			/* use memlist cached by dr_post_detach_mem_unit */
889 			ASSERT(mp->sbm_mlist != NULL);
890 			PR_MEM("%s: re-configuring cached memlist for %s:\n",
891 			    f, mp->sbm_cm.sbdev_path);
892 			PR_MEMLIST_DUMP(mp->sbm_mlist);
893 
894 			/* kphysm del handle should be have been freed */
895 			ASSERT((mp->sbm_flags & DR_MFLAG_RELOWNER) == 0);
896 
897 			/*FALLTHROUGH*/
898 
899 		case DR_STATE_CONNECTED:
900 			PR_MEM("%s: reprogramming mem hardware on %s\n",
901 			    f, mp->sbm_cm.sbdev_bp->b_path);
902 
903 			PR_MEM("%s: enabling %s\n",
904 			    f, mp->sbm_cm.sbdev_path);
905 
906 			err = drmach_mem_enable(mp->sbm_cm.sbdev_id);
907 			if (err) {
908 				DRERR_SET_C(&mp->sbm_cm.sbdev_error, &err);
909 				err_flag = 1;
910 			}
911 			break;
912 
913 		default:
914 			dr_dev_err(CE_WARN, &mp->sbm_cm, ESBD_STATE);
915 			err_flag = 1;
916 			break;
917 		}
918 
919 		/* exit for loop if error encountered */
920 		if (err_flag)
921 			break;
922 	}
923 
924 	return (err_flag ? -1 : 0);
925 }
926 
927 static void
928 dr_update_mc_memory()
929 {
930 	void		(*mc_update_mlist)(void);
931 
932 	/*
933 	 * mc-opl is configured during drmach_mem_new but the memory
934 	 * has not been added to phys_install at that time.
935 	 * we must inform mc-opl to update the mlist after we
936 	 * attach or detach a system board.
937 	 */
938 
939 	mc_update_mlist = (void (*)(void))
940 	    modgetsymvalue("opl_mc_update_mlist", 0);
941 
942 	if (mc_update_mlist != NULL) {
943 		(*mc_update_mlist)();
944 	}
945 }
946 
947 int
948 dr_post_attach_mem(dr_handle_t *hp, dr_common_unit_t **devlist, int devnum)
949 {
950 	_NOTE(ARGUNUSED(hp))
951 
952 	int		d;
953 	static fn_t	f = "dr_post_attach_mem";
954 
955 	PR_MEM("%s...\n", f);
956 
957 	for (d = 0; d < devnum; d++) {
958 		dr_mem_unit_t	*mp = (dr_mem_unit_t *)devlist[d];
959 		struct memlist	*mlist, *ml;
960 
961 		mlist = dr_get_memlist(mp);
962 		if (mlist == NULL) {
963 			/* OPL supports memoryless board */
964 			continue;
965 		}
966 
967 		/*
968 		 * Verify the memory really did successfully attach
969 		 * by checking for its existence in phys_install.
970 		 */
971 		memlist_read_lock();
972 		if (memlist_intersect(phys_install, mlist) == 0) {
973 			memlist_read_unlock();
974 
975 			DR_DEV_INTERNAL_ERROR(&mp->sbm_cm);
976 
977 			PR_MEM("%s: %s memlist not in phys_install",
978 			    f, mp->sbm_cm.sbdev_path);
979 
980 			memlist_delete(mlist);
981 			continue;
982 		}
983 		memlist_read_unlock();
984 
985 		for (ml = mlist; ml != NULL; ml = ml->next) {
986 			sbd_error_t *err;
987 
988 			err = drmach_mem_add_span(
989 			    mp->sbm_cm.sbdev_id,
990 			    ml->address,
991 			    ml->size);
992 			if (err)
993 				DRERR_SET_C(&mp->sbm_cm.sbdev_error, &err);
994 		}
995 
996 		memlist_delete(mlist);
997 
998 		/*
999 		 * Destroy cached memlist, if any.
1000 		 * There will be a cached memlist in sbm_mlist if
1001 		 * this board is being configured directly after
1002 		 * an unconfigure.
1003 		 * To support this transition, dr_post_detach_mem
1004 		 * left a copy of the last known memlist in sbm_mlist.
1005 		 * This memlist could differ from any derived from
1006 		 * hardware if while this memunit was last configured
1007 		 * the system detected and deleted bad pages from
1008 		 * phys_install.  The location of those bad pages
1009 		 * will be reflected in the cached memlist.
1010 		 */
1011 		if (mp->sbm_mlist) {
1012 			memlist_delete(mp->sbm_mlist);
1013 			mp->sbm_mlist = NULL;
1014 		}
1015 	}
1016 
1017 	dr_update_mc_memory();
1018 
1019 	return (0);
1020 }
1021 
1022 int
1023 dr_pre_detach_mem(dr_handle_t *hp, dr_common_unit_t **devlist, int devnum)
1024 {
1025 	_NOTE(ARGUNUSED(hp))
1026 
1027 	int d;
1028 
1029 	for (d = 0; d < devnum; d++) {
1030 		dr_mem_unit_t *mp = (dr_mem_unit_t *)devlist[d];
1031 
1032 		cmn_err(CE_CONT, "OS unconfigure %s", mp->sbm_cm.sbdev_path);
1033 	}
1034 
1035 	return (0);
1036 }
1037 
1038 int
1039 dr_post_detach_mem(dr_handle_t *hp, dr_common_unit_t **devlist, int devnum)
1040 {
1041 	_NOTE(ARGUNUSED(hp))
1042 
1043 	int		d, rv;
1044 	static fn_t	f = "dr_post_detach_mem";
1045 
1046 	PR_MEM("%s...\n", f);
1047 
1048 	rv = 0;
1049 	for (d = 0; d < devnum; d++) {
1050 		dr_mem_unit_t	*mp = (dr_mem_unit_t *)devlist[d];
1051 
1052 		ASSERT(mp->sbm_cm.sbdev_bp == hp->h_bd);
1053 
1054 		if (dr_post_detach_mem_unit(mp))
1055 			rv = -1;
1056 	}
1057 	dr_update_mc_memory();
1058 
1059 	return (rv);
1060 }
1061 
1062 static void
1063 dr_add_memory_spans(dr_mem_unit_t *mp, struct memlist *ml)
1064 {
1065 	static fn_t	f = "dr_add_memory_spans";
1066 
1067 	PR_MEM("%s...", f);
1068 	PR_MEMLIST_DUMP(ml);
1069 
1070 #ifdef DEBUG
1071 	memlist_read_lock();
1072 	if (memlist_intersect(phys_install, ml)) {
1073 		PR_MEM("%s:WARNING: memlist intersects with phys_install\n", f);
1074 	}
1075 	memlist_read_unlock();
1076 #endif
1077 
1078 	for (; ml; ml = ml->next) {
1079 		pfn_t		 base;
1080 		pgcnt_t		 npgs;
1081 		int		 rv;
1082 		sbd_error_t	*err;
1083 
1084 		base = _b64top(ml->address);
1085 		npgs = _b64top(ml->size);
1086 
1087 		rv = kphysm_add_memory_dynamic(base, npgs);
1088 
1089 		err = drmach_mem_add_span(
1090 		    mp->sbm_cm.sbdev_id,
1091 		    ml->address,
1092 		    ml->size);
1093 
1094 		if (err)
1095 			DRERR_SET_C(&mp->sbm_cm.sbdev_error, &err);
1096 
1097 		if (rv != KPHYSM_OK) {
1098 			cmn_err(CE_WARN, "%s:"
1099 			    " unexpected kphysm_add_memory_dynamic"
1100 			    " return value %d;"
1101 			    " basepfn=0x%lx, npages=%ld\n",
1102 			    f, rv, base, npgs);
1103 
1104 			continue;
1105 		}
1106 	}
1107 }
1108 
1109 static int
1110 memlist_touch(struct memlist *ml, uint64_t add)
1111 {
1112 	while (ml != NULL) {
1113 		if ((add == ml->address) ||
1114 		    (add == (ml->address + ml->size)))
1115 			return (1);
1116 		ml = ml->next;
1117 	}
1118 	return (0);
1119 }
1120 
1121 static sbd_error_t *
1122 dr_process_excess_mlist(dr_mem_unit_t *s_mp,
1123 	dr_mem_unit_t *t_mp, struct memlist *t_excess_mlist)
1124 {
1125 	struct memlist	*ml;
1126 	sbd_error_t	*err;
1127 	static fn_t	f = "dr_process_excess_mlist";
1128 	uint64_t	new_pa, nbytes;
1129 	int rv;
1130 
1131 	err = NULL;
1132 
1133 	/*
1134 	 * After the small <-> big copy-rename,
1135 	 * the original address space for the
1136 	 * source board may have excess to be
1137 	 * deleted. This is a case different
1138 	 * from the big->small excess source
1139 	 * memory case listed below.
1140 	 * Remove s_mp->sbm_del_mlist from
1141 	 * the kernel cage glist.
1142 	 */
1143 	for (ml = s_mp->sbm_del_mlist; ml;
1144 	    ml = ml->next) {
1145 		PR_MEM("%s: delete small<->big copy-"
1146 		    "rename source excess memory", f);
1147 		PR_MEMLIST_DUMP(ml);
1148 
1149 		err = drmach_mem_del_span(
1150 		    s_mp->sbm_cm.sbdev_id,
1151 		    ml->address, ml->size);
1152 		if (err)
1153 			DRERR_SET_C(&s_mp->
1154 			    sbm_cm.sbdev_error, &err);
1155 		ASSERT(err == NULL);
1156 	}
1157 
1158 	PR_MEM("%s: adding back remaining portion"
1159 	    " of %s, memlist:\n",
1160 	    f, t_mp->sbm_cm.sbdev_path);
1161 	PR_MEMLIST_DUMP(t_excess_mlist);
1162 
1163 	for (ml = t_excess_mlist; ml; ml = ml->next) {
1164 		struct memlist ml0;
1165 
1166 		ml0.address = ml->address;
1167 		ml0.size = ml->size;
1168 		ml0.next = ml0.prev = NULL;
1169 
1170 		/*
1171 		 * If the memory object is 256 MB aligned (max page size
1172 		 * on OPL, it will not be coalesced to the adjacent memory
1173 		 * chunks.  The coalesce logic assumes contiguous page
1174 		 * structures for contiguous memory and we hit panic.
1175 		 * For anything less than 256 MB alignment, we have
1176 		 * to make sure that it is not adjacent to anything.
1177 		 * If the new chunk is adjacent to phys_install, we
1178 		 * truncate it to 4MB boundary.  4 MB is somewhat
1179 		 * arbitrary.  However we do not want to create
1180 		 * very small segments because they can cause problem.
1181 		 * The extreme case of 8K segment will fail
1182 		 * kphysm_add_memory_dynamic(), e.g.
1183 		 */
1184 		if ((ml->address & (MH_MPSS_ALIGNMENT - 1)) ||
1185 		    (ml->size & (MH_MPSS_ALIGNMENT - 1))) {
1186 
1187 		memlist_read_lock();
1188 		rv = memlist_touch(phys_install, ml0.address);
1189 		memlist_read_unlock();
1190 
1191 		if (rv) {
1192 			new_pa = roundup(ml0.address + 1, MH_MIN_ALIGNMENT);
1193 			nbytes = (new_pa -  ml0.address);
1194 			if (nbytes >= ml0.size) {
1195 				t_mp->sbm_dyn_segs =
1196 				    memlist_del_span(t_mp->sbm_dyn_segs,
1197 				    ml0.address, ml0.size);
1198 				continue;
1199 			}
1200 			t_mp->sbm_dyn_segs =
1201 			    memlist_del_span(t_mp->sbm_dyn_segs,
1202 			    ml0.address, nbytes);
1203 			ml0.size -= nbytes;
1204 			ml0.address = new_pa;
1205 		}
1206 
1207 		if (ml0.size == 0) {
1208 			continue;
1209 		}
1210 
1211 		memlist_read_lock();
1212 		rv = memlist_touch(phys_install, ml0.address + ml0.size);
1213 		memlist_read_unlock();
1214 
1215 		if (rv) {
1216 			new_pa = rounddown(ml0.address + ml0.size - 1,
1217 			    MH_MIN_ALIGNMENT);
1218 			nbytes = (ml0.address + ml0.size - new_pa);
1219 			if (nbytes >= ml0.size) {
1220 				t_mp->sbm_dyn_segs =
1221 				    memlist_del_span(t_mp->sbm_dyn_segs,
1222 				    ml0.address, ml0.size);
1223 				continue;
1224 			}
1225 			t_mp->sbm_dyn_segs =
1226 			    memlist_del_span(t_mp->sbm_dyn_segs,
1227 			    new_pa, nbytes);
1228 			ml0.size -= nbytes;
1229 		}
1230 
1231 		if (ml0.size > 0) {
1232 			dr_add_memory_spans(s_mp, &ml0);
1233 		}
1234 		} else if (ml0.size > 0) {
1235 			dr_add_memory_spans(s_mp, &ml0);
1236 		}
1237 	}
1238 	memlist_delete(t_excess_mlist);
1239 	return (err);
1240 }
1241 
1242 static int
1243 dr_post_detach_mem_unit(dr_mem_unit_t *s_mp)
1244 {
1245 	uint64_t	sz = s_mp->sbm_slice_size;
1246 	uint64_t	sm = sz - 1;
1247 	/* old and new below refer to PAs before and after copy-rename */
1248 	uint64_t	s_old_basepa, s_new_basepa;
1249 	uint64_t	t_old_basepa, t_new_basepa;
1250 	dr_mem_unit_t	*t_mp, *x_mp;
1251 	drmach_mem_info_t	minfo;
1252 	struct memlist	*ml;
1253 	struct memlist	*t_excess_mlist;
1254 	int		rv;
1255 	int		s_excess_mem_deleted = 0;
1256 	sbd_error_t	*err;
1257 	static fn_t	f = "dr_post_detach_mem_unit";
1258 
1259 	PR_MEM("%s...\n", f);
1260 
1261 	/* s_mp->sbm_del_mlist could be NULL, meaning no deleted spans */
1262 	PR_MEM("%s: %s: deleted memlist (EMPTY maybe okay):\n",
1263 	    f, s_mp->sbm_cm.sbdev_path);
1264 	PR_MEMLIST_DUMP(s_mp->sbm_del_mlist);
1265 
1266 	/* sanity check */
1267 	ASSERT(s_mp->sbm_del_mlist == NULL ||
1268 	    (s_mp->sbm_flags & DR_MFLAG_RELDONE) != 0);
1269 
1270 	if (s_mp->sbm_flags & DR_MFLAG_SOURCE) {
1271 		t_mp = s_mp->sbm_peer;
1272 		ASSERT(t_mp != NULL);
1273 		ASSERT(t_mp->sbm_flags & DR_MFLAG_TARGET);
1274 		ASSERT(t_mp->sbm_peer == s_mp);
1275 
1276 		ASSERT(t_mp->sbm_flags & DR_MFLAG_RELDONE);
1277 		ASSERT(t_mp->sbm_del_mlist);
1278 
1279 		PR_MEM("%s: target %s: deleted memlist:\n",
1280 		    f, t_mp->sbm_cm.sbdev_path);
1281 		PR_MEMLIST_DUMP(t_mp->sbm_del_mlist);
1282 	} else {
1283 		/* this is no target unit */
1284 		t_mp = NULL;
1285 	}
1286 
1287 	/*
1288 	 * Verify the memory really did successfully detach
1289 	 * by checking for its non-existence in phys_install.
1290 	 */
1291 	rv = 0;
1292 	memlist_read_lock();
1293 	if (s_mp->sbm_flags & DR_MFLAG_RELDONE) {
1294 		x_mp = s_mp;
1295 		rv = memlist_intersect(phys_install, x_mp->sbm_del_mlist);
1296 	}
1297 	if (rv == 0 && t_mp && (t_mp->sbm_flags & DR_MFLAG_RELDONE)) {
1298 		x_mp = t_mp;
1299 		rv = memlist_intersect(phys_install, x_mp->sbm_del_mlist);
1300 	}
1301 	memlist_read_unlock();
1302 
1303 	if (rv) {
1304 		/* error: memlist still in phys_install */
1305 		DR_DEV_INTERNAL_ERROR(&x_mp->sbm_cm);
1306 	}
1307 
1308 	/*
1309 	 * clean mem unit state and bail out if an error has been recorded.
1310 	 */
1311 	rv = 0;
1312 	if (s_mp->sbm_cm.sbdev_error) {
1313 		PR_MEM("%s: %s flags=%x", f,
1314 		    s_mp->sbm_cm.sbdev_path, s_mp->sbm_flags);
1315 		DR_DEV_CLR_UNREFERENCED(&s_mp->sbm_cm);
1316 		DR_DEV_CLR_RELEASED(&s_mp->sbm_cm);
1317 		dr_device_transition(&s_mp->sbm_cm, DR_STATE_CONFIGURED);
1318 		rv = -1;
1319 	}
1320 	if (t_mp != NULL && t_mp->sbm_cm.sbdev_error != NULL) {
1321 		PR_MEM("%s: %s flags=%x", f,
1322 		    s_mp->sbm_cm.sbdev_path, s_mp->sbm_flags);
1323 		DR_DEV_CLR_UNREFERENCED(&t_mp->sbm_cm);
1324 		DR_DEV_CLR_RELEASED(&t_mp->sbm_cm);
1325 		dr_device_transition(&t_mp->sbm_cm, DR_STATE_CONFIGURED);
1326 		rv = -1;
1327 	}
1328 	if (rv)
1329 		goto cleanup;
1330 
1331 	s_old_basepa = _ptob64(s_mp->sbm_basepfn);
1332 	err = drmach_mem_get_info(s_mp->sbm_cm.sbdev_id, &minfo);
1333 	ASSERT(err == NULL);
1334 	s_new_basepa = minfo.mi_basepa;
1335 
1336 	PR_MEM("%s:s_old_basepa: 0x%lx\n", f, s_old_basepa);
1337 	PR_MEM("%s:s_new_basepa: 0x%lx\n", f, s_new_basepa);
1338 
1339 	if (t_mp != NULL) {
1340 		struct memlist *s_copy_mlist;
1341 
1342 		t_old_basepa = _ptob64(t_mp->sbm_basepfn);
1343 		err = drmach_mem_get_info(t_mp->sbm_cm.sbdev_id, &minfo);
1344 		ASSERT(err == NULL);
1345 		t_new_basepa = minfo.mi_basepa;
1346 
1347 		PR_MEM("%s:t_old_basepa: 0x%lx\n", f, t_old_basepa);
1348 		PR_MEM("%s:t_new_basepa: 0x%lx\n", f, t_new_basepa);
1349 
1350 		/*
1351 		 * Construct copy list with original source addresses.
1352 		 * Used to add back excess target mem.
1353 		 */
1354 		s_copy_mlist = memlist_dup(s_mp->sbm_mlist);
1355 		for (ml = s_mp->sbm_del_mlist; ml; ml = ml->next) {
1356 			s_copy_mlist = memlist_del_span(s_copy_mlist,
1357 			    ml->address, ml->size);
1358 		}
1359 
1360 		PR_MEM("%s: source copy list:\n:", f);
1361 		PR_MEMLIST_DUMP(s_copy_mlist);
1362 
1363 		/*
1364 		 * We had to swap mem-units, so update
1365 		 * memlists accordingly with new base
1366 		 * addresses.
1367 		 */
1368 		for (ml = t_mp->sbm_mlist; ml; ml = ml->next) {
1369 			ml->address -= t_old_basepa;
1370 			ml->address += t_new_basepa;
1371 		}
1372 
1373 		/*
1374 		 * There is no need to explicitly rename the target delete
1375 		 * memlist, because sbm_del_mlist and sbm_mlist always
1376 		 * point to the same memlist for a copy/rename operation.
1377 		 */
1378 		ASSERT(t_mp->sbm_del_mlist == t_mp->sbm_mlist);
1379 
1380 		PR_MEM("%s: renamed target memlist and delete memlist:\n", f);
1381 		PR_MEMLIST_DUMP(t_mp->sbm_mlist);
1382 
1383 		for (ml = s_mp->sbm_mlist; ml; ml = ml->next) {
1384 			ml->address -= s_old_basepa;
1385 			ml->address += s_new_basepa;
1386 		}
1387 
1388 		PR_MEM("%s: renamed source memlist:\n", f);
1389 		PR_MEMLIST_DUMP(s_mp->sbm_mlist);
1390 		PR_MEM("%s: source dyn seg memlist:\n", f);
1391 		PR_MEMLIST_DUMP(s_mp->sbm_dyn_segs);
1392 
1393 		/*
1394 		 * Keep track of dynamically added segments
1395 		 * since they cannot be split if we need to delete
1396 		 * excess source memory later for this board.
1397 		 */
1398 		if (t_mp->sbm_dyn_segs)
1399 			memlist_delete(t_mp->sbm_dyn_segs);
1400 		t_mp->sbm_dyn_segs = s_mp->sbm_dyn_segs;
1401 		s_mp->sbm_dyn_segs = NULL;
1402 
1403 		/*
1404 		 * Add back excess target memory.
1405 		 * Subtract out the portion of the target memory
1406 		 * node that was taken over by the source memory
1407 		 * node.
1408 		 */
1409 		t_excess_mlist = memlist_dup(t_mp->sbm_mlist);
1410 		for (ml = s_copy_mlist; ml; ml = ml->next) {
1411 			t_excess_mlist =
1412 			    memlist_del_span(t_excess_mlist,
1413 			    ml->address, ml->size);
1414 		}
1415 		PR_MEM("%s: excess memlist:\n", f);
1416 		PR_MEMLIST_DUMP(t_excess_mlist);
1417 
1418 		/*
1419 		 * Update dynamically added segs
1420 		 */
1421 		for (ml = s_mp->sbm_del_mlist; ml; ml = ml->next) {
1422 			t_mp->sbm_dyn_segs =
1423 			    memlist_del_span(t_mp->sbm_dyn_segs,
1424 			    ml->address, ml->size);
1425 		}
1426 		for (ml = t_excess_mlist; ml; ml = ml->next) {
1427 			t_mp->sbm_dyn_segs =
1428 			    memlist_cat_span(t_mp->sbm_dyn_segs,
1429 			    ml->address, ml->size);
1430 		}
1431 		PR_MEM("%s: %s: updated dynamic seg list:\n",
1432 		    f, t_mp->sbm_cm.sbdev_path);
1433 		PR_MEMLIST_DUMP(t_mp->sbm_dyn_segs);
1434 
1435 		if (t_excess_mlist != NULL) {
1436 			err = dr_process_excess_mlist(s_mp, t_mp,
1437 			    t_excess_mlist);
1438 			s_excess_mem_deleted = 1;
1439 		}
1440 
1441 		memlist_delete(s_copy_mlist);
1442 
1443 #ifdef DEBUG
1444 		/*
1445 		 * s_mp->sbm_del_mlist may still needed
1446 		 */
1447 		PR_MEM("%s: source delete memeory flag %d",
1448 		    f, s_excess_mem_deleted);
1449 		PR_MEM("%s: source delete memlist", f);
1450 		PR_MEMLIST_DUMP(s_mp->sbm_del_mlist);
1451 #endif
1452 
1453 	}
1454 
1455 	if (t_mp != NULL) {
1456 		/* delete target's entire address space */
1457 		err = drmach_mem_del_span(
1458 		    t_mp->sbm_cm.sbdev_id, t_old_basepa & ~ sm, sz);
1459 		if (err)
1460 			DRERR_SET_C(&t_mp->sbm_cm.sbdev_error, &err);
1461 		ASSERT(err == NULL);
1462 
1463 		/*
1464 		 * After the copy/rename, the original address space
1465 		 * for the source board (which is now located on the
1466 		 * target board) may now have some excess to be deleted.
1467 		 * Those excess memory on the source board are kept in
1468 		 * source board's sbm_del_mlist
1469 		 */
1470 		for (ml = s_mp->sbm_del_mlist; !s_excess_mem_deleted && ml;
1471 		    ml = ml->next) {
1472 			PR_MEM("%s: delete source excess memory", f);
1473 			PR_MEMLIST_DUMP(ml);
1474 
1475 			err = drmach_mem_del_span(s_mp->sbm_cm.sbdev_id,
1476 			    ml->address, ml->size);
1477 			if (err)
1478 				DRERR_SET_C(&s_mp->sbm_cm.sbdev_error, &err);
1479 			ASSERT(err == NULL);
1480 		}
1481 
1482 	} else {
1483 		/* delete board's entire address space */
1484 		err = drmach_mem_del_span(s_mp->sbm_cm.sbdev_id,
1485 		    s_old_basepa & ~ sm, sz);
1486 		if (err)
1487 			DRERR_SET_C(&s_mp->sbm_cm.sbdev_error, &err);
1488 		ASSERT(err == NULL);
1489 	}
1490 
1491 cleanup:
1492 	/* clean up target mem unit */
1493 	if (t_mp != NULL) {
1494 		memlist_delete(t_mp->sbm_del_mlist);
1495 		/* no need to delete sbm_mlist, it shares sbm_del_mlist */
1496 
1497 		t_mp->sbm_del_mlist = NULL;
1498 		t_mp->sbm_mlist = NULL;
1499 		t_mp->sbm_peer = NULL;
1500 		t_mp->sbm_flags = 0;
1501 		t_mp->sbm_cm.sbdev_busy = 0;
1502 		dr_init_mem_unit_data(t_mp);
1503 
1504 	}
1505 	if (t_mp != NULL && t_mp->sbm_cm.sbdev_error == NULL) {
1506 		/*
1507 		 * now that copy/rename has completed, undo this
1508 		 * work that was done in dr_release_mem_done.
1509 		 */
1510 		DR_DEV_CLR_UNREFERENCED(&t_mp->sbm_cm);
1511 		DR_DEV_CLR_RELEASED(&t_mp->sbm_cm);
1512 		dr_device_transition(&t_mp->sbm_cm, DR_STATE_CONFIGURED);
1513 	}
1514 
1515 	/*
1516 	 * clean up (source) board's mem unit structure.
1517 	 * NOTE: sbm_mlist is retained if no error has been record (in other
1518 	 * words, when s_mp->sbm_cm.sbdev_error is NULL). This memlist is
1519 	 * referred to elsewhere as the cached memlist.  The cached memlist
1520 	 * is used to re-attach (configure back in) this memunit from the
1521 	 * unconfigured state.  The memlist is retained because it may
1522 	 * represent bad pages that were detected while the memory was
1523 	 * configured into the OS.  The OS deletes bad pages from phys_install.
1524 	 * Those deletes, if any, will be represented in the cached mlist.
1525 	 */
1526 	if (s_mp->sbm_del_mlist && s_mp->sbm_del_mlist != s_mp->sbm_mlist)
1527 		memlist_delete(s_mp->sbm_del_mlist);
1528 
1529 	if (s_mp->sbm_cm.sbdev_error && s_mp->sbm_mlist) {
1530 		memlist_delete(s_mp->sbm_mlist);
1531 		s_mp->sbm_mlist = NULL;
1532 	}
1533 
1534 	if (s_mp->sbm_dyn_segs != NULL && s_mp->sbm_cm.sbdev_error == 0) {
1535 		memlist_delete(s_mp->sbm_dyn_segs);
1536 		s_mp->sbm_dyn_segs = NULL;
1537 	}
1538 
1539 	s_mp->sbm_del_mlist = NULL;
1540 	s_mp->sbm_peer = NULL;
1541 	s_mp->sbm_flags = 0;
1542 	s_mp->sbm_cm.sbdev_busy = 0;
1543 	dr_init_mem_unit_data(s_mp);
1544 
1545 	PR_MEM("%s: cached memlist for %s:", f, s_mp->sbm_cm.sbdev_path);
1546 	PR_MEMLIST_DUMP(s_mp->sbm_mlist);
1547 
1548 	return (0);
1549 }
1550 
1551 /*
1552  * Successful return from this function will have the memory
1553  * handle in bp->b_dev[..mem-unit...].sbm_memhandle allocated
1554  * and waiting.  This routine's job is to select the memory that
1555  * actually has to be released (detached) which may not necessarily
1556  * be the same memory node that came in in devlist[],
1557  * i.e. a copy-rename is needed.
1558  */
1559 int
1560 dr_pre_release_mem(dr_handle_t *hp, dr_common_unit_t **devlist, int devnum)
1561 {
1562 	int		d;
1563 	int		err_flag = 0;
1564 	static fn_t	f = "dr_pre_release_mem";
1565 
1566 	PR_MEM("%s...\n", f);
1567 
1568 	for (d = 0; d < devnum; d++) {
1569 		dr_mem_unit_t	*mp = (dr_mem_unit_t *)devlist[d];
1570 		int		rv;
1571 		memquery_t	mq;
1572 		struct memlist	*ml;
1573 
1574 		if (mp->sbm_cm.sbdev_error) {
1575 			err_flag = 1;
1576 			continue;
1577 		} else if (!kcage_on) {
1578 			dr_dev_err(CE_WARN, &mp->sbm_cm, ESBD_KCAGE_OFF);
1579 			err_flag = 1;
1580 			continue;
1581 		}
1582 
1583 		if (mp->sbm_flags & DR_MFLAG_RESERVED) {
1584 			/*
1585 			 * Board is currently involved in a delete
1586 			 * memory operation. Can't detach this guy until
1587 			 * that operation completes.
1588 			 */
1589 			dr_dev_err(CE_WARN, &mp->sbm_cm, ESBD_INVAL);
1590 			err_flag = 1;
1591 			break;
1592 		}
1593 
1594 		/* flags should be clean at this time */
1595 		ASSERT(mp->sbm_flags == 0);
1596 
1597 		ASSERT(mp->sbm_mlist == NULL);
1598 		ASSERT(mp->sbm_del_mlist == NULL);
1599 		if (mp->sbm_mlist != NULL) {
1600 			memlist_delete(mp->sbm_mlist);
1601 			mp->sbm_mlist = NULL;
1602 		}
1603 
1604 		ml = dr_get_memlist(mp);
1605 		if (ml == NULL) {
1606 			err_flag = 1;
1607 			PR_MEM("%s: no memlist found for %s\n",
1608 			    f, mp->sbm_cm.sbdev_path);
1609 			continue;
1610 		}
1611 
1612 		/*
1613 		 * Check whether the detaching memory requires a
1614 		 * copy-rename.
1615 		 */
1616 		ASSERT(mp->sbm_npages != 0);
1617 
1618 		rv = dr_del_mlist_query(ml, &mq);
1619 		if (rv != KPHYSM_OK) {
1620 			memlist_delete(ml);
1621 			DR_DEV_INTERNAL_ERROR(&mp->sbm_cm);
1622 			err_flag = 1;
1623 			break;
1624 		}
1625 
1626 		if (mq.nonrelocatable != 0) {
1627 			if (!(dr_cmd_flags(hp) &
1628 			    (SBD_FLAG_FORCE | SBD_FLAG_QUIESCE_OKAY))) {
1629 				memlist_delete(ml);
1630 				/* caller wasn't prompted for a suspend */
1631 				dr_dev_err(CE_WARN, &mp->sbm_cm,
1632 				    ESBD_QUIESCE_REQD);
1633 				err_flag = 1;
1634 				break;
1635 			}
1636 		}
1637 
1638 		/* allocate a kphysm handle */
1639 		rv = kphysm_del_gethandle(&mp->sbm_memhandle);
1640 		if (rv != KPHYSM_OK) {
1641 			memlist_delete(ml);
1642 
1643 			DR_DEV_INTERNAL_ERROR(&mp->sbm_cm);
1644 			err_flag = 1;
1645 			break;
1646 		}
1647 		mp->sbm_flags |= DR_MFLAG_RELOWNER;
1648 
1649 		if ((mq.nonrelocatable != 0) ||
1650 		    dr_reserve_mem_spans(&mp->sbm_memhandle, ml)) {
1651 			/*
1652 			 * Either the detaching memory node contains
1653 			 * non-reloc memory or we failed to reserve the
1654 			 * detaching memory node (which did _not_ have
1655 			 * any non-reloc memory, i.e. some non-reloc mem
1656 			 * got onboard).
1657 			 */
1658 
1659 			if (dr_select_mem_target(hp, mp, ml)) {
1660 				int rv;
1661 
1662 				/*
1663 				 * We had no luck locating a target
1664 				 * memory node to be the recipient of
1665 				 * the non-reloc memory on the node
1666 				 * we're trying to detach.
1667 				 * Clean up be disposing the mem handle
1668 				 * and the mem list.
1669 				 */
1670 				rv = kphysm_del_release(mp->sbm_memhandle);
1671 				if (rv != KPHYSM_OK) {
1672 					/*
1673 					 * can do nothing but complain
1674 					 * and hope helpful for debug
1675 					 */
1676 					cmn_err(CE_WARN, "%s: unexpected"
1677 					    " kphysm_del_release return"
1678 					    " value %d",
1679 					    f, rv);
1680 				}
1681 				mp->sbm_flags &= ~DR_MFLAG_RELOWNER;
1682 
1683 				memlist_delete(ml);
1684 
1685 				/* make sure sbm_flags is clean */
1686 				ASSERT(mp->sbm_flags == 0);
1687 
1688 				dr_dev_err(CE_WARN,
1689 				    &mp->sbm_cm, ESBD_NO_TARGET);
1690 
1691 				err_flag = 1;
1692 				break;
1693 			}
1694 
1695 			/*
1696 			 * ml is not memlist_delete'd here because
1697 			 * it has been assigned to mp->sbm_mlist
1698 			 * by dr_select_mem_target.
1699 			 */
1700 		} else {
1701 			/* no target needed to detach this board */
1702 			mp->sbm_flags |= DR_MFLAG_RESERVED;
1703 			mp->sbm_peer = NULL;
1704 			mp->sbm_del_mlist = ml;
1705 			mp->sbm_mlist = ml;
1706 			mp->sbm_cm.sbdev_busy = 1;
1707 		}
1708 #ifdef DEBUG
1709 		ASSERT(mp->sbm_mlist != NULL);
1710 
1711 		if (mp->sbm_flags & DR_MFLAG_SOURCE) {
1712 			PR_MEM("%s: release of %s requires copy/rename;"
1713 			    " selected target board %s\n",
1714 			    f,
1715 			    mp->sbm_cm.sbdev_path,
1716 			    mp->sbm_peer->sbm_cm.sbdev_path);
1717 		} else {
1718 			PR_MEM("%s: copy/rename not required to release %s\n",
1719 			    f, mp->sbm_cm.sbdev_path);
1720 		}
1721 
1722 		ASSERT(mp->sbm_flags & DR_MFLAG_RELOWNER);
1723 		ASSERT(mp->sbm_flags & DR_MFLAG_RESERVED);
1724 #endif
1725 	}
1726 
1727 	return (err_flag ? -1 : 0);
1728 }
1729 
1730 void
1731 dr_release_mem_done(dr_common_unit_t *cp)
1732 {
1733 	dr_mem_unit_t	*s_mp = (dr_mem_unit_t *)cp;
1734 	dr_mem_unit_t *t_mp, *mp;
1735 	int		rv;
1736 	static fn_t	f = "dr_release_mem_done";
1737 
1738 	/*
1739 	 * This unit will be flagged with DR_MFLAG_SOURCE, if it
1740 	 * has a target unit.
1741 	 */
1742 	if (s_mp->sbm_flags & DR_MFLAG_SOURCE) {
1743 		t_mp = s_mp->sbm_peer;
1744 		ASSERT(t_mp != NULL);
1745 		ASSERT(t_mp->sbm_peer == s_mp);
1746 		ASSERT(t_mp->sbm_flags & DR_MFLAG_TARGET);
1747 		ASSERT(t_mp->sbm_flags & DR_MFLAG_RESERVED);
1748 	} else {
1749 		/* this is no target unit */
1750 		t_mp = NULL;
1751 	}
1752 
1753 	/* free delete handle */
1754 	ASSERT(s_mp->sbm_flags & DR_MFLAG_RELOWNER);
1755 	ASSERT(s_mp->sbm_flags & DR_MFLAG_RESERVED);
1756 	rv = kphysm_del_release(s_mp->sbm_memhandle);
1757 	if (rv != KPHYSM_OK) {
1758 		/*
1759 		 * can do nothing but complain
1760 		 * and hope helpful for debug
1761 		 */
1762 		cmn_err(CE_WARN, "%s: unexpected kphysm_del_release"
1763 		    " return value %d", f, rv);
1764 	}
1765 	s_mp->sbm_flags &= ~DR_MFLAG_RELOWNER;
1766 
1767 	/*
1768 	 * If an error was encountered during release, clean up
1769 	 * the source (and target, if present) unit data.
1770 	 */
1771 /* XXX Can we know that sbdev_error was encountered during release? */
1772 	if (s_mp->sbm_cm.sbdev_error != NULL) {
1773 
1774 		if (t_mp != NULL) {
1775 			ASSERT(t_mp->sbm_del_mlist == t_mp->sbm_mlist);
1776 			t_mp->sbm_del_mlist = NULL;
1777 
1778 			if (t_mp->sbm_mlist != NULL) {
1779 				memlist_delete(t_mp->sbm_mlist);
1780 				t_mp->sbm_mlist = NULL;
1781 			}
1782 
1783 			t_mp->sbm_peer = NULL;
1784 			t_mp->sbm_flags = 0;
1785 			t_mp->sbm_cm.sbdev_busy = 0;
1786 		}
1787 
1788 		if (s_mp->sbm_del_mlist != s_mp->sbm_mlist)
1789 			memlist_delete(s_mp->sbm_del_mlist);
1790 		s_mp->sbm_del_mlist = NULL;
1791 
1792 		if (s_mp->sbm_mlist != NULL) {
1793 			memlist_delete(s_mp->sbm_mlist);
1794 			s_mp->sbm_mlist = NULL;
1795 		}
1796 
1797 		s_mp->sbm_peer = NULL;
1798 		s_mp->sbm_flags = 0;
1799 		s_mp->sbm_cm.sbdev_busy = 0;
1800 
1801 		/* bail out */
1802 		return;
1803 	}
1804 
1805 	DR_DEV_SET_RELEASED(&s_mp->sbm_cm);
1806 	dr_device_transition(&s_mp->sbm_cm, DR_STATE_RELEASE);
1807 
1808 	if (t_mp != NULL) {
1809 		/*
1810 		 * the kphysm delete operation that drained the source
1811 		 * board also drained this target board.  Since the source
1812 		 * board drain is now known to have succeeded, we know this
1813 		 * target board is drained too.
1814 		 *
1815 		 * because DR_DEV_SET_RELEASED and dr_device_transition
1816 		 * is done here, the dr_release_dev_done should not
1817 		 * fail.
1818 		 */
1819 		DR_DEV_SET_RELEASED(&t_mp->sbm_cm);
1820 		dr_device_transition(&t_mp->sbm_cm, DR_STATE_RELEASE);
1821 
1822 		/*
1823 		 * NOTE: do not transition target's board state,
1824 		 * even if the mem-unit was the last configure
1825 		 * unit of the board.  When copy/rename completes
1826 		 * this mem-unit will transitioned back to
1827 		 * the configured state.  In the meantime, the
1828 		 * board's must remain as is.
1829 		 */
1830 	}
1831 
1832 	/* if board(s) had deleted memory, verify it is gone */
1833 	rv = 0;
1834 	memlist_read_lock();
1835 	if (s_mp->sbm_del_mlist != NULL) {
1836 		mp = s_mp;
1837 		rv = memlist_intersect(phys_install, mp->sbm_del_mlist);
1838 	}
1839 	if (rv == 0 && t_mp && t_mp->sbm_del_mlist != NULL) {
1840 		mp = t_mp;
1841 		rv = memlist_intersect(phys_install, mp->sbm_del_mlist);
1842 	}
1843 	memlist_read_unlock();
1844 	if (rv) {
1845 		cmn_err(CE_WARN, "%s: %smem-unit (%d.%d): "
1846 		    "deleted memory still found in phys_install",
1847 		    f,
1848 		    (mp == t_mp ? "target " : ""),
1849 		    mp->sbm_cm.sbdev_bp->b_num,
1850 		    mp->sbm_cm.sbdev_unum);
1851 
1852 		DR_DEV_INTERNAL_ERROR(&s_mp->sbm_cm);
1853 		return;
1854 	}
1855 
1856 	s_mp->sbm_flags |= DR_MFLAG_RELDONE;
1857 	if (t_mp != NULL)
1858 		t_mp->sbm_flags |= DR_MFLAG_RELDONE;
1859 
1860 	/* this should not fail */
1861 	if (dr_release_dev_done(&s_mp->sbm_cm) != 0) {
1862 		/* catch this in debug kernels */
1863 		ASSERT(0);
1864 		return;
1865 	}
1866 
1867 	PR_MEM("%s: marking %s release DONE\n",
1868 	    f, s_mp->sbm_cm.sbdev_path);
1869 
1870 	s_mp->sbm_cm.sbdev_ostate = SBD_STAT_UNCONFIGURED;
1871 
1872 	if (t_mp != NULL) {
1873 		/* should not fail */
1874 		rv = dr_release_dev_done(&t_mp->sbm_cm);
1875 		if (rv != 0) {
1876 			/* catch this in debug kernels */
1877 			ASSERT(0);
1878 			return;
1879 		}
1880 
1881 		PR_MEM("%s: marking %s release DONE\n",
1882 		    f, t_mp->sbm_cm.sbdev_path);
1883 
1884 		t_mp->sbm_cm.sbdev_ostate = SBD_STAT_UNCONFIGURED;
1885 	}
1886 }
1887 
1888 /*ARGSUSED*/
1889 int
1890 dr_disconnect_mem(dr_mem_unit_t *mp)
1891 {
1892 	static fn_t	f = "dr_disconnect_mem";
1893 	update_membounds_t umb;
1894 
1895 #ifdef DEBUG
1896 	int state = mp->sbm_cm.sbdev_state;
1897 	ASSERT(state == DR_STATE_CONNECTED ||
1898 	    state == DR_STATE_UNCONFIGURED);
1899 #endif
1900 
1901 	PR_MEM("%s...\n", f);
1902 
1903 	if (mp->sbm_del_mlist && mp->sbm_del_mlist != mp->sbm_mlist)
1904 		memlist_delete(mp->sbm_del_mlist);
1905 	mp->sbm_del_mlist = NULL;
1906 
1907 	if (mp->sbm_mlist) {
1908 		memlist_delete(mp->sbm_mlist);
1909 		mp->sbm_mlist = NULL;
1910 	}
1911 
1912 	/*
1913 	 * Remove memory from lgroup
1914 	 * For now, only board info is required.
1915 	 */
1916 	umb.u_board = mp->sbm_cm.sbdev_bp->b_num;
1917 	umb.u_base = (uint64_t)-1;
1918 	umb.u_len = (uint64_t)-1;
1919 
1920 	lgrp_plat_config(LGRP_CONFIG_MEM_DEL, (uintptr_t)&umb);
1921 
1922 	return (0);
1923 }
1924 
1925 int
1926 dr_cancel_mem(dr_mem_unit_t *s_mp)
1927 {
1928 	dr_mem_unit_t	*t_mp;
1929 	dr_state_t	state;
1930 	static fn_t	f = "dr_cancel_mem";
1931 
1932 	state = s_mp->sbm_cm.sbdev_state;
1933 
1934 	if (s_mp->sbm_flags & DR_MFLAG_TARGET) {
1935 		/* must cancel source board, not target board */
1936 		/* TODO: set error */
1937 		return (-1);
1938 	} else if (s_mp->sbm_flags & DR_MFLAG_SOURCE) {
1939 		t_mp = s_mp->sbm_peer;
1940 		ASSERT(t_mp != NULL);
1941 		ASSERT(t_mp->sbm_peer == s_mp);
1942 
1943 		/* must always match the source board's state */
1944 		/* TODO: is this assertion correct? */
1945 		ASSERT(t_mp->sbm_cm.sbdev_state == state);
1946 	} else {
1947 		/* this is no target unit */
1948 		t_mp = NULL;
1949 	}
1950 
1951 	switch (state) {
1952 	case DR_STATE_UNREFERENCED:	/* state set by dr_release_dev_done */
1953 		ASSERT((s_mp->sbm_flags & DR_MFLAG_RELOWNER) == 0);
1954 
1955 		if (t_mp != NULL && t_mp->sbm_del_mlist != NULL) {
1956 			PR_MEM("%s: undoing target %s memory delete\n",
1957 			    f, t_mp->sbm_cm.sbdev_path);
1958 			dr_add_memory_spans(t_mp, t_mp->sbm_del_mlist);
1959 
1960 			DR_DEV_CLR_UNREFERENCED(&t_mp->sbm_cm);
1961 		}
1962 
1963 		if (s_mp->sbm_del_mlist != NULL) {
1964 			PR_MEM("%s: undoing %s memory delete\n",
1965 			    f, s_mp->sbm_cm.sbdev_path);
1966 
1967 			dr_add_memory_spans(s_mp, s_mp->sbm_del_mlist);
1968 		}
1969 
1970 		/*FALLTHROUGH*/
1971 
1972 /* TODO: should no longer be possible to see the release state here */
1973 	case DR_STATE_RELEASE:	/* state set by dr_release_mem_done */
1974 
1975 		ASSERT((s_mp->sbm_flags & DR_MFLAG_RELOWNER) == 0);
1976 
1977 		if (t_mp != NULL) {
1978 			ASSERT(t_mp->sbm_del_mlist == t_mp->sbm_mlist);
1979 			t_mp->sbm_del_mlist = NULL;
1980 
1981 			if (t_mp->sbm_mlist != NULL) {
1982 				memlist_delete(t_mp->sbm_mlist);
1983 				t_mp->sbm_mlist = NULL;
1984 			}
1985 
1986 			t_mp->sbm_peer = NULL;
1987 			t_mp->sbm_flags = 0;
1988 			t_mp->sbm_cm.sbdev_busy = 0;
1989 			dr_init_mem_unit_data(t_mp);
1990 
1991 			DR_DEV_CLR_RELEASED(&t_mp->sbm_cm);
1992 
1993 			dr_device_transition(
1994 			    &t_mp->sbm_cm, DR_STATE_CONFIGURED);
1995 		}
1996 
1997 		if (s_mp->sbm_del_mlist != s_mp->sbm_mlist)
1998 			memlist_delete(s_mp->sbm_del_mlist);
1999 		s_mp->sbm_del_mlist = NULL;
2000 
2001 		if (s_mp->sbm_mlist != NULL) {
2002 			memlist_delete(s_mp->sbm_mlist);
2003 			s_mp->sbm_mlist = NULL;
2004 		}
2005 
2006 		s_mp->sbm_peer = NULL;
2007 		s_mp->sbm_flags = 0;
2008 		s_mp->sbm_cm.sbdev_busy = 0;
2009 		dr_init_mem_unit_data(s_mp);
2010 
2011 		return (0);
2012 
2013 	default:
2014 		PR_MEM("%s: WARNING unexpected state (%d) for %s\n",
2015 		    f, (int)state, s_mp->sbm_cm.sbdev_path);
2016 
2017 		return (-1);
2018 	}
2019 	/*NOTREACHED*/
2020 }
2021 
2022 void
2023 dr_init_mem_unit(dr_mem_unit_t *mp)
2024 {
2025 	dr_state_t	new_state;
2026 
2027 
2028 	if (DR_DEV_IS_ATTACHED(&mp->sbm_cm)) {
2029 		new_state = DR_STATE_CONFIGURED;
2030 		mp->sbm_cm.sbdev_cond = SBD_COND_OK;
2031 	} else if (DR_DEV_IS_PRESENT(&mp->sbm_cm)) {
2032 		new_state = DR_STATE_CONNECTED;
2033 		mp->sbm_cm.sbdev_cond = SBD_COND_OK;
2034 	} else if (mp->sbm_cm.sbdev_id != (drmachid_t)0) {
2035 		new_state = DR_STATE_OCCUPIED;
2036 	} else {
2037 		new_state = DR_STATE_EMPTY;
2038 	}
2039 
2040 	if (DR_DEV_IS_PRESENT(&mp->sbm_cm))
2041 		dr_init_mem_unit_data(mp);
2042 
2043 	/* delay transition until fully initialized */
2044 	dr_device_transition(&mp->sbm_cm, new_state);
2045 }
2046 
2047 static void
2048 dr_init_mem_unit_data(dr_mem_unit_t *mp)
2049 {
2050 	drmachid_t	id = mp->sbm_cm.sbdev_id;
2051 	drmach_mem_info_t	minfo;
2052 	sbd_error_t	*err;
2053 	static fn_t	f = "dr_init_mem_unit_data";
2054 	update_membounds_t umb;
2055 
2056 	PR_MEM("%s...\n", f);
2057 
2058 	/* a little sanity checking */
2059 	ASSERT(mp->sbm_peer == NULL);
2060 	ASSERT(mp->sbm_flags == 0);
2061 
2062 	if (err = drmach_mem_get_info(id, &minfo)) {
2063 		DRERR_SET_C(&mp->sbm_cm.sbdev_error, &err);
2064 		return;
2065 	}
2066 	mp->sbm_basepfn = _b64top(minfo.mi_basepa);
2067 	mp->sbm_npages = _b64top(minfo.mi_size);
2068 	mp->sbm_alignment_mask = _b64top(minfo.mi_alignment_mask);
2069 	mp->sbm_slice_size = minfo.mi_slice_size;
2070 
2071 	/*
2072 	 * Add memory to lgroup
2073 	 */
2074 	umb.u_board = mp->sbm_cm.sbdev_bp->b_num;
2075 	umb.u_base = (uint64_t)mp->sbm_basepfn << MMU_PAGESHIFT;
2076 	umb.u_len = (uint64_t)mp->sbm_npages << MMU_PAGESHIFT;
2077 
2078 	lgrp_plat_config(LGRP_CONFIG_MEM_ADD, (uintptr_t)&umb);
2079 
2080 	PR_MEM("%s: %s (basepfn = 0x%lx, npgs = %ld)\n",
2081 	    f, mp->sbm_cm.sbdev_path, mp->sbm_basepfn, mp->sbm_npages);
2082 }
2083 
2084 static int
2085 dr_reserve_mem_spans(memhandle_t *mhp, struct memlist *ml)
2086 {
2087 	int		err;
2088 	pfn_t		base;
2089 	pgcnt_t		npgs;
2090 	struct memlist	*mc;
2091 	static fn_t	f = "dr_reserve_mem_spans";
2092 
2093 	PR_MEM("%s...\n", f);
2094 
2095 	/*
2096 	 * Walk the supplied memlist scheduling each span for removal
2097 	 * with kphysm_del_span.  It is possible that a span may intersect
2098 	 * an area occupied by the cage.
2099 	 */
2100 	for (mc = ml; mc != NULL; mc = mc->next) {
2101 		base = _b64top(mc->address);
2102 		npgs = _b64top(mc->size);
2103 
2104 		err = kphysm_del_span(*mhp, base, npgs);
2105 		if (err != KPHYSM_OK) {
2106 			cmn_err(CE_WARN, "%s memory reserve failed."
2107 			    " unexpected kphysm_del_span return value %d;"
2108 			    " basepfn=0x%lx npages=%ld",
2109 			    f, err, base, npgs);
2110 
2111 			return (-1);
2112 		}
2113 	}
2114 
2115 	return (0);
2116 }
2117 
2118 #define	DR_SMT_NPREF_SETS	6
2119 #define	DR_SMT_NUNITS_PER_SET	MAX_BOARDS * MAX_MEM_UNITS_PER_BOARD
2120 
2121 /* debug counters */
2122 int dr_smt_realigned;
2123 int dr_smt_preference[DR_SMT_NPREF_SETS];
2124 
2125 #ifdef DEBUG
2126 uint_t dr_ignore_board; /* if bit[bnum-1] set, board won't be candidate */
2127 #endif
2128 
2129 /*
2130  * Find and reserve a copy/rename target board suitable for the
2131  * given source board.
2132  * All boards in the system are examined and categorized in relation to
2133  * their memory size versus the source board's memory size.  Order of
2134  * preference is:
2135  *	1st copy all source, source/target same size
2136  *	2nd copy all source, larger target
2137  * 	3rd copy nonrelocatable source span
2138  */
2139 static int
2140 dr_select_mem_target(dr_handle_t *hp,
2141 	dr_mem_unit_t *s_mp, struct memlist *s_ml)
2142 {
2143 	dr_target_pref_t preference; /* lower value is higher preference */
2144 	int		idx;
2145 	dr_mem_unit_t	**sets;
2146 
2147 	int		t_bd;
2148 	int		t_unit;
2149 	int		rv;
2150 	dr_board_t	*s_bp, *t_bp;
2151 	dr_mem_unit_t	*t_mp, *c_mp;
2152 	struct memlist	*d_ml, *t_ml, *ml, *b_ml, *x_ml = NULL;
2153 	memquery_t	s_mq = {0};
2154 	static fn_t	f = "dr_select_mem_target";
2155 
2156 	PR_MEM("%s...\n", f);
2157 
2158 	ASSERT(s_ml != NULL);
2159 
2160 	sets = GETSTRUCT(dr_mem_unit_t *, DR_SMT_NUNITS_PER_SET *
2161 	    DR_SMT_NPREF_SETS);
2162 
2163 	s_bp = hp->h_bd;
2164 	/* calculate the offset into the slice of the last source board pfn */
2165 	ASSERT(s_mp->sbm_npages != 0);
2166 
2167 	/*
2168 	 * Find non-relocatable span on source board.
2169 	 */
2170 	rv = kphysm_del_span_query(s_mp->sbm_basepfn, s_mp->sbm_npages, &s_mq);
2171 	if (rv != KPHYSM_OK) {
2172 		PR_MEM("%s: %s: unexpected kphysm_del_span_query"
2173 		    " return value %d; basepfn 0x%lx, npages %ld\n",
2174 		    f, s_mp->sbm_cm.sbdev_path, rv, s_mp->sbm_basepfn,
2175 		    s_mp->sbm_npages);
2176 		return (-1);
2177 	}
2178 
2179 	ASSERT(s_mq.phys_pages != 0);
2180 	ASSERT(s_mq.nonrelocatable != 0);
2181 
2182 	PR_MEM("%s: %s: nonrelocatable span (0x%lx..0x%lx)\n", f,
2183 	    s_mp->sbm_cm.sbdev_path, s_mq.first_nonrelocatable,
2184 	    s_mq.last_nonrelocatable);
2185 
2186 	/* break down s_ml if it contains dynamic segments */
2187 	b_ml = memlist_dup(s_ml);
2188 
2189 	for (ml = s_mp->sbm_dyn_segs; ml; ml = ml->next) {
2190 		b_ml = memlist_del_span(b_ml, ml->address, ml->size);
2191 		b_ml = memlist_cat_span(b_ml, ml->address, ml->size);
2192 	}
2193 
2194 
2195 	/*
2196 	 * Make one pass through all memory units on all boards
2197 	 * and categorize them with respect to the source board.
2198 	 */
2199 	for (t_bd = 0; t_bd < MAX_BOARDS; t_bd++) {
2200 		/*
2201 		 * The board structs are a contiguous array
2202 		 * so we take advantage of that to find the
2203 		 * correct board struct pointer for a given
2204 		 * board number.
2205 		 */
2206 		t_bp = dr_lookup_board(t_bd);
2207 
2208 		/* source board can not be its own target */
2209 		if (s_bp->b_num == t_bp->b_num)
2210 			continue;
2211 
2212 		for (t_unit = 0; t_unit < MAX_MEM_UNITS_PER_BOARD; t_unit++) {
2213 
2214 			t_mp = dr_get_mem_unit(t_bp, t_unit);
2215 
2216 			/* this memory node must be attached */
2217 			if (!DR_DEV_IS_ATTACHED(&t_mp->sbm_cm))
2218 				continue;
2219 
2220 			/* source unit can not be its own target */
2221 			if (s_mp == t_mp) {
2222 				/* catch this is debug kernels */
2223 				ASSERT(0);
2224 				continue;
2225 			}
2226 
2227 			/*
2228 			 * this memory node must not already be reserved
2229 			 * by some other memory delete operation.
2230 			 */
2231 			if (t_mp->sbm_flags & DR_MFLAG_RESERVED)
2232 				continue;
2233 
2234 			/* get target board memlist */
2235 			t_ml = dr_get_memlist(t_mp);
2236 			if (t_ml == NULL) {
2237 				cmn_err(CE_WARN, "%s: no memlist for"
2238 				    " mem-unit %d, board %d", f,
2239 				    t_mp->sbm_cm.sbdev_bp->b_num,
2240 				    t_mp->sbm_cm.sbdev_unum);
2241 				continue;
2242 			}
2243 
2244 			preference = dr_get_target_preference(hp, t_mp, s_mp,
2245 			    t_ml, s_ml, b_ml);
2246 
2247 			memlist_delete(t_ml);
2248 
2249 			if (preference == DR_TP_INVALID)
2250 				continue;
2251 
2252 			dr_smt_preference[preference]++;
2253 
2254 			/* calculate index to start of preference set */
2255 			idx  = DR_SMT_NUNITS_PER_SET * preference;
2256 			/* calculate offset to respective element */
2257 			idx += t_bd * MAX_MEM_UNITS_PER_BOARD + t_unit;
2258 
2259 			ASSERT(idx < DR_SMT_NUNITS_PER_SET * DR_SMT_NPREF_SETS);
2260 			sets[idx] = t_mp;
2261 		}
2262 	}
2263 
2264 	if (b_ml != NULL)
2265 		memlist_delete(b_ml);
2266 
2267 	/*
2268 	 * NOTE: this would be a good place to sort each candidate
2269 	 * set in to some desired order, e.g. memory size in ascending
2270 	 * order.  Without an additional sorting step here, the order
2271 	 * within a set is ascending board number order.
2272 	 */
2273 
2274 	c_mp = NULL;
2275 	x_ml = NULL;
2276 	t_ml = NULL;
2277 	for (idx = 0; idx < DR_SMT_NUNITS_PER_SET * DR_SMT_NPREF_SETS; idx++) {
2278 		memquery_t mq;
2279 
2280 		preference = (dr_target_pref_t)(idx / DR_SMT_NUNITS_PER_SET);
2281 
2282 		ASSERT(preference != DR_TP_INVALID);
2283 
2284 		/* cleanup t_ml after previous pass */
2285 		if (t_ml != NULL) {
2286 			memlist_delete(t_ml);
2287 			t_ml = NULL;
2288 		}
2289 
2290 		/* get candidate target board mem unit */
2291 		t_mp = sets[idx];
2292 		if (t_mp == NULL)
2293 			continue;
2294 
2295 		/* get target board memlist */
2296 		t_ml = dr_get_memlist(t_mp);
2297 		if (t_ml == NULL) {
2298 			cmn_err(CE_WARN, "%s: no memlist for"
2299 			    " mem-unit %d, board %d",
2300 			    f,
2301 			    t_mp->sbm_cm.sbdev_bp->b_num,
2302 			    t_mp->sbm_cm.sbdev_unum);
2303 
2304 			continue;
2305 		}
2306 
2307 		PR_MEM("%s: checking for no-reloc in %s, "
2308 		    " basepfn=0x%lx, npages=%ld\n",
2309 		    f,
2310 		    t_mp->sbm_cm.sbdev_path,
2311 		    t_mp->sbm_basepfn,
2312 		    t_mp->sbm_npages);
2313 
2314 		rv = dr_del_mlist_query(t_ml, &mq);
2315 		if (rv != KPHYSM_OK) {
2316 			PR_MEM("%s: kphysm_del_span_query:"
2317 			    " unexpected return value %d\n", f, rv);
2318 
2319 			continue;
2320 		}
2321 
2322 		if (mq.nonrelocatable != 0) {
2323 			PR_MEM("%s: candidate %s has"
2324 			    " nonrelocatable span [0x%lx..0x%lx]\n",
2325 			    f,
2326 			    t_mp->sbm_cm.sbdev_path,
2327 			    mq.first_nonrelocatable,
2328 			    mq.last_nonrelocatable);
2329 
2330 			continue;
2331 		}
2332 
2333 #ifdef DEBUG
2334 		/*
2335 		 * This is a debug tool for excluding certain boards
2336 		 * from being selected as a target board candidate.
2337 		 * dr_ignore_board is only tested by this driver.
2338 		 * It must be set with adb, obp, /etc/system or your
2339 		 * favorite debugger.
2340 		 */
2341 		if (dr_ignore_board &
2342 		    (1 << (t_mp->sbm_cm.sbdev_bp->b_num - 1))) {
2343 			PR_MEM("%s: dr_ignore_board flag set,"
2344 			    " ignoring %s as candidate\n",
2345 			    f, t_mp->sbm_cm.sbdev_path);
2346 			continue;
2347 		}
2348 #endif
2349 
2350 		/*
2351 		 * Reserve excess source board memory, if any.
2352 		 *
2353 		 * Only the nonrelocatable source span will be copied
2354 		 * so schedule the rest of the source mem to be deleted.
2355 		 */
2356 		switch (preference) {
2357 		case DR_TP_NONRELOC:
2358 			/*
2359 			 * Get source copy memlist and use it to construct
2360 			 * delete memlist.
2361 			 */
2362 			d_ml = memlist_dup(s_ml);
2363 			x_ml = dr_get_copy_mlist(s_ml, t_ml, s_mp, t_mp);
2364 
2365 			/* XXX */
2366 			ASSERT(d_ml != NULL);
2367 			ASSERT(x_ml != NULL);
2368 
2369 			for (ml = x_ml; ml != NULL; ml = ml->next) {
2370 				d_ml = memlist_del_span(d_ml, ml->address,
2371 				    ml->size);
2372 			}
2373 
2374 			PR_MEM("%s: %s: reserving src brd memlist:\n", f,
2375 			    s_mp->sbm_cm.sbdev_path);
2376 			PR_MEMLIST_DUMP(d_ml);
2377 
2378 			/* reserve excess spans */
2379 			if (dr_reserve_mem_spans(&s_mp->sbm_memhandle,
2380 			    d_ml) != 0) {
2381 				/* likely more non-reloc pages appeared */
2382 				/* TODO: restart from top? */
2383 				continue;
2384 			}
2385 			break;
2386 		default:
2387 			d_ml = NULL;
2388 			break;
2389 		}
2390 
2391 		s_mp->sbm_flags |= DR_MFLAG_RESERVED;
2392 
2393 		/*
2394 		 * reserve all memory on target board.
2395 		 * NOTE: source board's memhandle is used.
2396 		 *
2397 		 * If this succeeds (eq 0), then target selection is
2398 		 * complete and all unwanted memory spans, both source and
2399 		 * target, have been reserved.  Loop is terminated.
2400 		 */
2401 		if (dr_reserve_mem_spans(&s_mp->sbm_memhandle, t_ml) == 0) {
2402 			PR_MEM("%s: %s: target board memory reserved\n",
2403 			    f, t_mp->sbm_cm.sbdev_path);
2404 
2405 			/* a candidate target board is now reserved */
2406 			t_mp->sbm_flags |= DR_MFLAG_RESERVED;
2407 			c_mp = t_mp;
2408 
2409 			/* *** EXITING LOOP *** */
2410 			break;
2411 		}
2412 
2413 		/* did not successfully reserve the target board. */
2414 		PR_MEM("%s: could not reserve target %s\n",
2415 		    f, t_mp->sbm_cm.sbdev_path);
2416 
2417 		/*
2418 		 * NOTE: an undo of the dr_reserve_mem_span work
2419 		 * will happen automatically when the memhandle
2420 		 * (s_mp->sbm_memhandle) is kphysm_del_release'd.
2421 		 */
2422 
2423 		s_mp->sbm_flags &= ~DR_MFLAG_RESERVED;
2424 	}
2425 
2426 	/* clean up after memlist editing logic */
2427 	if (x_ml != NULL)
2428 		memlist_delete(x_ml);
2429 
2430 	FREESTRUCT(sets, dr_mem_unit_t *, DR_SMT_NUNITS_PER_SET *
2431 	    DR_SMT_NPREF_SETS);
2432 
2433 	/*
2434 	 * c_mp will be NULL when the entire sets[] array
2435 	 * has been searched without reserving a target board.
2436 	 */
2437 	if (c_mp == NULL) {
2438 		PR_MEM("%s: %s: target selection failed.\n",
2439 		    f, s_mp->sbm_cm.sbdev_path);
2440 
2441 		if (t_ml != NULL)
2442 			memlist_delete(t_ml);
2443 
2444 		return (-1);
2445 	}
2446 
2447 	PR_MEM("%s: found target %s for source %s\n",
2448 	    f,
2449 	    c_mp->sbm_cm.sbdev_path,
2450 	    s_mp->sbm_cm.sbdev_path);
2451 
2452 	s_mp->sbm_peer = c_mp;
2453 	s_mp->sbm_flags |= DR_MFLAG_SOURCE;
2454 	s_mp->sbm_del_mlist = d_ml;	/* spans to be deleted, if any */
2455 	s_mp->sbm_mlist = s_ml;
2456 	s_mp->sbm_cm.sbdev_busy = 1;
2457 
2458 	c_mp->sbm_peer = s_mp;
2459 	c_mp->sbm_flags |= DR_MFLAG_TARGET;
2460 	c_mp->sbm_del_mlist = t_ml;	/* spans to be deleted */
2461 	c_mp->sbm_mlist = t_ml;
2462 	c_mp->sbm_cm.sbdev_busy = 1;
2463 
2464 	return (0);
2465 }
2466 
2467 /*
2468  * Returns target preference rank:
2469  *     -1 not a valid copy-rename target board
2470  *	0 copy all source, source/target same size
2471  *	1 copy all source, larger target
2472  * 	2 copy nonrelocatable source span
2473  */
2474 static dr_target_pref_t
2475 dr_get_target_preference(dr_handle_t *hp,
2476     dr_mem_unit_t *t_mp, dr_mem_unit_t *s_mp,
2477     struct memlist *t_ml, struct memlist *s_ml,
2478     struct memlist *b_ml)
2479 {
2480 	dr_target_pref_t preference;
2481 	struct memlist *s_nonreloc_ml = NULL;
2482 	drmachid_t t_id;
2483 	static fn_t	f = "dr_get_target_preference";
2484 
2485 	t_id = t_mp->sbm_cm.sbdev_bp->b_id;
2486 
2487 	/*
2488 	 * Can the entire source board be copied?
2489 	 */
2490 	if (dr_memlist_canfit(s_ml, t_ml, s_mp, t_mp)) {
2491 		if (s_mp->sbm_npages == t_mp->sbm_npages)
2492 			preference = DR_TP_SAME;	/* same size */
2493 		else
2494 			preference = DR_TP_LARGE;	/* larger target */
2495 	} else {
2496 		/*
2497 		 * Entire source won't fit so try non-relocatable memory only
2498 		 * (target aligned).
2499 		 */
2500 		s_nonreloc_ml = dr_get_nonreloc_mlist(b_ml, s_mp);
2501 		if (s_nonreloc_ml == NULL) {
2502 			PR_MEM("%s: dr_get_nonreloc_mlist failed\n", f);
2503 			preference = DR_TP_INVALID;
2504 		}
2505 		if (dr_memlist_canfit(s_nonreloc_ml, t_ml, s_mp, t_mp))
2506 			preference = DR_TP_NONRELOC;
2507 		else
2508 			preference = DR_TP_INVALID;
2509 	}
2510 
2511 	if (s_nonreloc_ml != NULL)
2512 		memlist_delete(s_nonreloc_ml);
2513 
2514 	/*
2515 	 * Force floating board preference lower than all other boards
2516 	 * if the force flag is present; otherwise disallow the board.
2517 	 */
2518 	if ((preference != DR_TP_INVALID) && drmach_board_is_floating(t_id)) {
2519 		if (dr_cmd_flags(hp) & SBD_FLAG_FORCE)
2520 			preference += DR_TP_FLOATING;
2521 		else
2522 			preference = DR_TP_INVALID;
2523 	}
2524 
2525 	PR_MEM("%s: %s preference=%d\n", f, t_mp->sbm_cm.sbdev_path,
2526 	    preference);
2527 
2528 	return (preference);
2529 }
2530 
2531 /*
2532  * Create a memlist representing the source memory that will be copied to
2533  * the target board.  The memory to be copied is the maximum amount that
2534  * will fit on the target board.
2535  */
2536 static struct memlist *
2537 dr_get_copy_mlist(struct memlist *s_mlist, struct memlist *t_mlist,
2538     dr_mem_unit_t *s_mp, dr_mem_unit_t *t_mp)
2539 {
2540 	struct memlist	*t_ml, *s_copy_ml, *s_del_ml, *ml, *x_ml;
2541 	uint64_t	s_slice_mask, s_slice_base;
2542 	uint64_t	t_slice_mask, t_slice_base;
2543 	static fn_t	f = "dr_get_copy_mlist";
2544 
2545 	ASSERT(s_mlist != NULL);
2546 	ASSERT(t_mlist != NULL);
2547 	ASSERT(t_mp->sbm_slice_size == s_mp->sbm_slice_size);
2548 
2549 	s_slice_mask = s_mp->sbm_slice_size - 1;
2550 	s_slice_base = s_mlist->address & ~s_slice_mask;
2551 
2552 	t_slice_mask = t_mp->sbm_slice_size - 1;
2553 	t_slice_base = t_mlist->address & ~t_slice_mask;
2554 
2555 	t_ml = memlist_dup(t_mlist);
2556 	s_del_ml = memlist_dup(s_mlist);
2557 	s_copy_ml = memlist_dup(s_mlist);
2558 
2559 	/* XXX */
2560 	ASSERT(t_ml != NULL);
2561 	ASSERT(s_del_ml != NULL);
2562 	ASSERT(s_copy_ml != NULL);
2563 
2564 	/*
2565 	 * To construct the source copy memlist:
2566 	 *
2567 	 * The target memlist is converted to the post-rename
2568 	 * source addresses.  This is the physical address range
2569 	 * the target will have after the copy-rename.  Overlaying
2570 	 * and deleting this from the current source memlist will
2571 	 * give the source delete memlist.  The copy memlist is
2572 	 * the reciprocal of the source delete memlist.
2573 	 */
2574 	for (ml = t_ml; ml != NULL; ml = ml->next) {
2575 		/*
2576 		 * Normalize relative to target slice base PA
2577 		 * in order to preseve slice offsets.
2578 		 */
2579 		ml->address -= t_slice_base;
2580 		/*
2581 		 * Convert to source slice PA address.
2582 		 */
2583 		ml->address += s_slice_base;
2584 	}
2585 
2586 	for (ml = t_ml; ml != NULL; ml = ml->next) {
2587 		s_del_ml = memlist_del_span(s_del_ml, ml->address, ml->size);
2588 	}
2589 
2590 	/*
2591 	 * Expand the delete mlist to fully include any dynamic segments
2592 	 * it intersects with.
2593 	 */
2594 	for (x_ml = NULL, ml = s_del_ml; ml != NULL; ml = ml->next) {
2595 		uint64_t del_base = ml->address;
2596 		uint64_t del_end = ml->address + ml->size;
2597 		struct memlist *dyn;
2598 
2599 		for (dyn = s_mp->sbm_dyn_segs; dyn != NULL; dyn = dyn->next) {
2600 			uint64_t dyn_base = dyn->address;
2601 			uint64_t dyn_end = dyn->address + dyn->size;
2602 
2603 			if (del_base > dyn_base && del_base < dyn_end)
2604 				del_base = dyn_base;
2605 
2606 			if (del_end > dyn_base && del_end < dyn_end)
2607 				del_end = dyn_end;
2608 		}
2609 
2610 		x_ml = memlist_cat_span(x_ml, del_base, del_end - del_base);
2611 	}
2612 
2613 	memlist_delete(s_del_ml);
2614 	s_del_ml = x_ml;
2615 
2616 	for (ml = s_del_ml; ml != NULL; ml = ml->next) {
2617 		s_copy_ml = memlist_del_span(s_copy_ml, ml->address, ml->size);
2618 	}
2619 
2620 	PR_MEM("%s: source delete mlist\n", f);
2621 	PR_MEMLIST_DUMP(s_del_ml);
2622 
2623 	PR_MEM("%s: source copy mlist\n", f);
2624 	PR_MEMLIST_DUMP(s_copy_ml);
2625 
2626 	memlist_delete(t_ml);
2627 	memlist_delete(s_del_ml);
2628 
2629 	return (s_copy_ml);
2630 }
2631 
2632 /*
2633  * Scan the non-relocatable spans on the source memory
2634  * and construct a minimum mlist that includes all non-reloc
2635  * memory subject to target alignment, and dynamic segment
2636  * constraints where only whole dynamic segments may be deleted.
2637  */
2638 static struct memlist *
2639 dr_get_nonreloc_mlist(struct memlist *s_ml, dr_mem_unit_t *s_mp)
2640 {
2641 	struct memlist	*x_ml = NULL;
2642 	struct memlist	*ml;
2643 	static fn_t	f = "dr_get_nonreloc_mlist";
2644 
2645 	PR_MEM("%s: checking for split of dyn seg list:\n", f);
2646 	PR_MEMLIST_DUMP(s_mp->sbm_dyn_segs);
2647 
2648 	for (ml = s_ml; ml; ml = ml->next) {
2649 		int rv;
2650 		uint64_t nr_base, nr_end;
2651 		memquery_t mq;
2652 		struct memlist *dyn;
2653 
2654 		rv = kphysm_del_span_query(
2655 		    _b64top(ml->address), _b64top(ml->size), &mq);
2656 		if (rv) {
2657 			memlist_delete(x_ml);
2658 			return (NULL);
2659 		}
2660 
2661 		if (mq.nonrelocatable == 0)
2662 			continue;
2663 
2664 		PR_MEM("%s: non-reloc span: 0x%lx, 0x%lx (%lx, %lx)\n", f,
2665 		    _ptob64(mq.first_nonrelocatable),
2666 		    _ptob64(mq.last_nonrelocatable),
2667 		    mq.first_nonrelocatable,
2668 		    mq.last_nonrelocatable);
2669 
2670 		/*
2671 		 * Align the span at both ends to allow for possible
2672 		 * cage expansion.
2673 		 */
2674 		nr_base = _ptob64(mq.first_nonrelocatable);
2675 		nr_end = _ptob64(mq.last_nonrelocatable + 1);
2676 
2677 		PR_MEM("%s: adjusted non-reloc span: 0x%lx, 0x%lx\n",
2678 		    f, nr_base, nr_end);
2679 
2680 		/*
2681 		 * Expand the non-reloc span to fully include any
2682 		 * dynamic segments it intersects with.
2683 		 */
2684 		for (dyn = s_mp->sbm_dyn_segs; dyn != NULL; dyn = dyn->next) {
2685 			uint64_t dyn_base = dyn->address;
2686 			uint64_t dyn_end = dyn->address + dyn->size;
2687 
2688 			if (nr_base > dyn_base && nr_base < dyn_end)
2689 				nr_base = dyn_base;
2690 
2691 			if (nr_end > dyn_base && nr_end < dyn_end)
2692 				nr_end = dyn_end;
2693 		}
2694 
2695 		x_ml = memlist_cat_span(x_ml, nr_base, nr_end - nr_base);
2696 	}
2697 
2698 	if (x_ml == NULL) {
2699 		PR_MEM("%s: source didn't have any non-reloc pages!\n", f);
2700 		return (NULL);
2701 	}
2702 
2703 	PR_MEM("%s: %s: edited source memlist:\n", f, s_mp->sbm_cm.sbdev_path);
2704 	PR_MEMLIST_DUMP(x_ml);
2705 
2706 	return (x_ml);
2707 }
2708 
2709 /*
2710  * Check if source memlist can fit in target memlist while maintaining
2711  * relative offsets within board.
2712  */
2713 static int
2714 dr_memlist_canfit(struct memlist *s_mlist, struct memlist *t_mlist,
2715     dr_mem_unit_t *s_mp, dr_mem_unit_t *t_mp)
2716 {
2717 	int		canfit = 0;
2718 	struct memlist	*s_ml, *t_ml, *ml;
2719 	uint64_t	s_slice_mask, t_slice_mask;
2720 	static fn_t	f = "dr_mlist_canfit";
2721 
2722 	s_ml = memlist_dup(s_mlist);
2723 	t_ml = memlist_dup(t_mlist);
2724 
2725 	if (s_ml == NULL || t_ml == NULL) {
2726 		cmn_err(CE_WARN, "%s: memlist_dup failed\n", f);
2727 		goto done;
2728 	}
2729 
2730 	s_slice_mask = s_mp->sbm_slice_size - 1;
2731 	t_slice_mask = t_mp->sbm_slice_size - 1;
2732 
2733 	/*
2734 	 * Normalize to slice relative offsets.
2735 	 */
2736 	for (ml = s_ml; ml; ml = ml->next)
2737 		ml->address &= s_slice_mask;
2738 
2739 	for (ml = t_ml; ml; ml = ml->next)
2740 		ml->address &= t_slice_mask;
2741 
2742 	canfit = memlist_canfit(s_ml, t_ml);
2743 done:
2744 	memlist_delete(s_ml);
2745 	memlist_delete(t_ml);
2746 
2747 	return (canfit);
2748 }
2749 
2750 /*
2751  * Memlist support.
2752  */
2753 
2754 /*
2755  * Determine whether the source memlist (s_mlist) will
2756  * fit into the target memlist (t_mlist) in terms of
2757  * size and holes.  Assumes the caller has normalized the
2758  * memlist physical addresses for comparison.
2759  */
2760 static int
2761 memlist_canfit(struct memlist *s_mlist, struct memlist *t_mlist)
2762 {
2763 	int		rv = 0;
2764 	struct memlist	*s_ml, *t_ml;
2765 
2766 	if ((s_mlist == NULL) || (t_mlist == NULL))
2767 		return (0);
2768 
2769 	s_ml = s_mlist;
2770 	for (t_ml = t_mlist; t_ml && s_ml; t_ml = t_ml->next) {
2771 		uint64_t	s_start, s_end;
2772 		uint64_t	t_start, t_end;
2773 
2774 		t_start = t_ml->address;
2775 		t_end = t_start + t_ml->size;
2776 
2777 		for (; s_ml; s_ml = s_ml->next) {
2778 			s_start = s_ml->address;
2779 			s_end = s_start + s_ml->size;
2780 
2781 			if ((s_start < t_start) || (s_end > t_end))
2782 				break;
2783 		}
2784 	}
2785 
2786 	/*
2787 	 * If we ran out of source memlist chunks that mean
2788 	 * we found a home for all of them.
2789 	 */
2790 	if (s_ml == NULL)
2791 		rv = 1;
2792 
2793 	return (rv);
2794 }
2795