xref: /illumos-gate/usr/src/lib/libdtrace/common/dt_aggregate.c (revision e4d060fb4c00d44cd578713eb9a921f594b733b8)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 #pragma ident	"%Z%%M%	%I%	%E% SMI"
28 
29 #include <stdlib.h>
30 #include <strings.h>
31 #include <errno.h>
32 #include <unistd.h>
33 #include <dt_impl.h>
34 #include <assert.h>
35 #include <alloca.h>
36 #include <limits.h>
37 
38 #define	DTRACE_AHASHSIZE	32779		/* big 'ol prime */
39 
40 /*
41  * Because qsort(3C) does not allow an argument to be passed to a comparison
42  * function, the variables that affect comparison must regrettably be global;
43  * they are protected by a global static lock, dt_qsort_lock.
44  */
45 static pthread_mutex_t dt_qsort_lock = PTHREAD_MUTEX_INITIALIZER;
46 
47 static int dt_revsort;
48 static int dt_keysort;
49 static int dt_keypos;
50 
51 #define	DT_LESSTHAN	(dt_revsort == 0 ? -1 : 1)
52 #define	DT_GREATERTHAN	(dt_revsort == 0 ? 1 : -1)
53 
54 static void
55 dt_aggregate_count(int64_t *existing, int64_t *new, size_t size)
56 {
57 	int i;
58 
59 	for (i = 0; i < size / sizeof (int64_t); i++)
60 		existing[i] = existing[i] + new[i];
61 }
62 
63 static int
64 dt_aggregate_countcmp(int64_t *lhs, int64_t *rhs)
65 {
66 	int64_t lvar = *lhs;
67 	int64_t rvar = *rhs;
68 
69 	if (lvar < rvar)
70 		return (DT_LESSTHAN);
71 
72 	if (lvar > rvar)
73 		return (DT_GREATERTHAN);
74 
75 	return (0);
76 }
77 
78 /*ARGSUSED*/
79 static void
80 dt_aggregate_min(int64_t *existing, int64_t *new, size_t size)
81 {
82 	if (*new < *existing)
83 		*existing = *new;
84 }
85 
86 /*ARGSUSED*/
87 static void
88 dt_aggregate_max(int64_t *existing, int64_t *new, size_t size)
89 {
90 	if (*new > *existing)
91 		*existing = *new;
92 }
93 
94 static int
95 dt_aggregate_averagecmp(int64_t *lhs, int64_t *rhs)
96 {
97 	int64_t lavg = lhs[0] ? (lhs[1] / lhs[0]) : 0;
98 	int64_t ravg = rhs[0] ? (rhs[1] / rhs[0]) : 0;
99 
100 	if (lavg < ravg)
101 		return (DT_LESSTHAN);
102 
103 	if (lavg > ravg)
104 		return (DT_GREATERTHAN);
105 
106 	return (0);
107 }
108 
109 static int
110 dt_aggregate_stddevcmp(int64_t *lhs, int64_t *rhs)
111 {
112 	uint64_t lsd = dt_stddev((uint64_t *)lhs, 1);
113 	uint64_t rsd = dt_stddev((uint64_t *)rhs, 1);
114 
115 	if (lsd < rsd)
116 		return (DT_LESSTHAN);
117 
118 	if (lsd > rsd)
119 		return (DT_GREATERTHAN);
120 
121 	return (0);
122 }
123 
124 /*ARGSUSED*/
125 static void
126 dt_aggregate_lquantize(int64_t *existing, int64_t *new, size_t size)
127 {
128 	int64_t arg = *existing++;
129 	uint16_t levels = DTRACE_LQUANTIZE_LEVELS(arg);
130 	int i;
131 
132 	for (i = 0; i <= levels + 1; i++)
133 		existing[i] = existing[i] + new[i + 1];
134 }
135 
136 static long double
137 dt_aggregate_lquantizedsum(int64_t *lquanta)
138 {
139 	int64_t arg = *lquanta++;
140 	int32_t base = DTRACE_LQUANTIZE_BASE(arg);
141 	uint16_t step = DTRACE_LQUANTIZE_STEP(arg);
142 	uint16_t levels = DTRACE_LQUANTIZE_LEVELS(arg), i;
143 	long double total = (long double)lquanta[0] * (long double)(base - 1);
144 
145 	for (i = 0; i < levels; base += step, i++)
146 		total += (long double)lquanta[i + 1] * (long double)base;
147 
148 	return (total + (long double)lquanta[levels + 1] *
149 	    (long double)(base + 1));
150 }
151 
152 static int64_t
153 dt_aggregate_lquantizedzero(int64_t *lquanta)
154 {
155 	int64_t arg = *lquanta++;
156 	int32_t base = DTRACE_LQUANTIZE_BASE(arg);
157 	uint16_t step = DTRACE_LQUANTIZE_STEP(arg);
158 	uint16_t levels = DTRACE_LQUANTIZE_LEVELS(arg), i;
159 
160 	if (base - 1 == 0)
161 		return (lquanta[0]);
162 
163 	for (i = 0; i < levels; base += step, i++) {
164 		if (base != 0)
165 			continue;
166 
167 		return (lquanta[i + 1]);
168 	}
169 
170 	if (base + 1 == 0)
171 		return (lquanta[levels + 1]);
172 
173 	return (0);
174 }
175 
176 static int
177 dt_aggregate_lquantizedcmp(int64_t *lhs, int64_t *rhs)
178 {
179 	long double lsum = dt_aggregate_lquantizedsum(lhs);
180 	long double rsum = dt_aggregate_lquantizedsum(rhs);
181 	int64_t lzero, rzero;
182 
183 	if (lsum < rsum)
184 		return (DT_LESSTHAN);
185 
186 	if (lsum > rsum)
187 		return (DT_GREATERTHAN);
188 
189 	/*
190 	 * If they're both equal, then we will compare based on the weights at
191 	 * zero.  If the weights at zero are equal (or if zero is not within
192 	 * the range of the linear quantization), then this will be judged a
193 	 * tie and will be resolved based on the key comparison.
194 	 */
195 	lzero = dt_aggregate_lquantizedzero(lhs);
196 	rzero = dt_aggregate_lquantizedzero(rhs);
197 
198 	if (lzero < rzero)
199 		return (DT_LESSTHAN);
200 
201 	if (lzero > rzero)
202 		return (DT_GREATERTHAN);
203 
204 	return (0);
205 }
206 
207 static int
208 dt_aggregate_quantizedcmp(int64_t *lhs, int64_t *rhs)
209 {
210 	int nbuckets = DTRACE_QUANTIZE_NBUCKETS, i;
211 	long double ltotal = 0, rtotal = 0;
212 	int64_t lzero, rzero;
213 
214 	for (i = 0; i < nbuckets; i++) {
215 		int64_t bucketval = DTRACE_QUANTIZE_BUCKETVAL(i);
216 
217 		if (bucketval == 0) {
218 			lzero = lhs[i];
219 			rzero = rhs[i];
220 		}
221 
222 		ltotal += (long double)bucketval * (long double)lhs[i];
223 		rtotal += (long double)bucketval * (long double)rhs[i];
224 	}
225 
226 	if (ltotal < rtotal)
227 		return (DT_LESSTHAN);
228 
229 	if (ltotal > rtotal)
230 		return (DT_GREATERTHAN);
231 
232 	/*
233 	 * If they're both equal, then we will compare based on the weights at
234 	 * zero.  If the weights at zero are equal, then this will be judged a
235 	 * tie and will be resolved based on the key comparison.
236 	 */
237 	if (lzero < rzero)
238 		return (DT_LESSTHAN);
239 
240 	if (lzero > rzero)
241 		return (DT_GREATERTHAN);
242 
243 	return (0);
244 }
245 
246 static void
247 dt_aggregate_usym(dtrace_hdl_t *dtp, uint64_t *data)
248 {
249 	uint64_t pid = data[0];
250 	uint64_t *pc = &data[1];
251 	struct ps_prochandle *P;
252 	GElf_Sym sym;
253 
254 	if (dtp->dt_vector != NULL)
255 		return;
256 
257 	if ((P = dt_proc_grab(dtp, pid, PGRAB_RDONLY | PGRAB_FORCE, 0)) == NULL)
258 		return;
259 
260 	dt_proc_lock(dtp, P);
261 
262 	if (Plookup_by_addr(P, *pc, NULL, 0, &sym) == 0)
263 		*pc = sym.st_value;
264 
265 	dt_proc_unlock(dtp, P);
266 	dt_proc_release(dtp, P);
267 }
268 
269 static void
270 dt_aggregate_umod(dtrace_hdl_t *dtp, uint64_t *data)
271 {
272 	uint64_t pid = data[0];
273 	uint64_t *pc = &data[1];
274 	struct ps_prochandle *P;
275 	const prmap_t *map;
276 
277 	if (dtp->dt_vector != NULL)
278 		return;
279 
280 	if ((P = dt_proc_grab(dtp, pid, PGRAB_RDONLY | PGRAB_FORCE, 0)) == NULL)
281 		return;
282 
283 	dt_proc_lock(dtp, P);
284 
285 	if ((map = Paddr_to_map(P, *pc)) != NULL)
286 		*pc = map->pr_vaddr;
287 
288 	dt_proc_unlock(dtp, P);
289 	dt_proc_release(dtp, P);
290 }
291 
292 static void
293 dt_aggregate_sym(dtrace_hdl_t *dtp, uint64_t *data)
294 {
295 	GElf_Sym sym;
296 	uint64_t *pc = data;
297 
298 	if (dtrace_lookup_by_addr(dtp, *pc, &sym, NULL) == 0)
299 		*pc = sym.st_value;
300 }
301 
302 static void
303 dt_aggregate_mod(dtrace_hdl_t *dtp, uint64_t *data)
304 {
305 	uint64_t *pc = data;
306 	dt_module_t *dmp;
307 
308 	if (dtp->dt_vector != NULL) {
309 		/*
310 		 * We don't have a way of just getting the module for a
311 		 * vectored open, and it doesn't seem to be worth defining
312 		 * one.  This means that use of mod() won't get true
313 		 * aggregation in the postmortem case (some modules may
314 		 * appear more than once in aggregation output).  It seems
315 		 * unlikely that anyone will ever notice or care...
316 		 */
317 		return;
318 	}
319 
320 	for (dmp = dt_list_next(&dtp->dt_modlist); dmp != NULL;
321 	    dmp = dt_list_next(dmp)) {
322 		if (*pc - dmp->dm_text_va < dmp->dm_text_size) {
323 			*pc = dmp->dm_text_va;
324 			return;
325 		}
326 	}
327 }
328 
329 static dtrace_aggvarid_t
330 dt_aggregate_aggvarid(dt_ahashent_t *ent)
331 {
332 	dtrace_aggdesc_t *agg = ent->dtahe_data.dtada_desc;
333 	caddr_t data = ent->dtahe_data.dtada_data;
334 	dtrace_recdesc_t *rec = agg->dtagd_rec;
335 
336 	/*
337 	 * First, we'll check the variable ID in the aggdesc.  If it's valid,
338 	 * we'll return it.  If not, we'll use the compiler-generated ID
339 	 * present as the first record.
340 	 */
341 	if (agg->dtagd_varid != DTRACE_AGGVARIDNONE)
342 		return (agg->dtagd_varid);
343 
344 	agg->dtagd_varid = *((dtrace_aggvarid_t *)(uintptr_t)(data +
345 	    rec->dtrd_offset));
346 
347 	return (agg->dtagd_varid);
348 }
349 
350 
351 static int
352 dt_aggregate_snap_cpu(dtrace_hdl_t *dtp, processorid_t cpu)
353 {
354 	dtrace_epid_t id;
355 	uint64_t hashval;
356 	size_t offs, roffs, size, ndx;
357 	int i, j, rval;
358 	caddr_t addr, data;
359 	dtrace_recdesc_t *rec;
360 	dt_aggregate_t *agp = &dtp->dt_aggregate;
361 	dtrace_aggdesc_t *agg;
362 	dt_ahash_t *hash = &agp->dtat_hash;
363 	dt_ahashent_t *h;
364 	dtrace_bufdesc_t b = agp->dtat_buf, *buf = &b;
365 	dtrace_aggdata_t *aggdata;
366 	int flags = agp->dtat_flags;
367 
368 	buf->dtbd_cpu = cpu;
369 
370 	if (dt_ioctl(dtp, DTRACEIOC_AGGSNAP, buf) == -1) {
371 		if (errno == ENOENT) {
372 			/*
373 			 * If that failed with ENOENT, it may be because the
374 			 * CPU was unconfigured.  This is okay; we'll just
375 			 * do nothing but return success.
376 			 */
377 			return (0);
378 		}
379 
380 		return (dt_set_errno(dtp, errno));
381 	}
382 
383 	if (buf->dtbd_drops != 0) {
384 		if (dt_handle_cpudrop(dtp, cpu,
385 		    DTRACEDROP_AGGREGATION, buf->dtbd_drops) == -1)
386 			return (-1);
387 	}
388 
389 	if (buf->dtbd_size == 0)
390 		return (0);
391 
392 	if (hash->dtah_hash == NULL) {
393 		size_t size;
394 
395 		hash->dtah_size = DTRACE_AHASHSIZE;
396 		size = hash->dtah_size * sizeof (dt_ahashent_t *);
397 
398 		if ((hash->dtah_hash = malloc(size)) == NULL)
399 			return (dt_set_errno(dtp, EDT_NOMEM));
400 
401 		bzero(hash->dtah_hash, size);
402 	}
403 
404 	for (offs = 0; offs < buf->dtbd_size; ) {
405 		/*
406 		 * We're guaranteed to have an ID.
407 		 */
408 		id = *((dtrace_epid_t *)((uintptr_t)buf->dtbd_data +
409 		    (uintptr_t)offs));
410 
411 		if (id == DTRACE_AGGIDNONE) {
412 			/*
413 			 * This is filler to assure proper alignment of the
414 			 * next record; we simply ignore it.
415 			 */
416 			offs += sizeof (id);
417 			continue;
418 		}
419 
420 		if ((rval = dt_aggid_lookup(dtp, id, &agg)) != 0)
421 			return (rval);
422 
423 		addr = buf->dtbd_data + offs;
424 		size = agg->dtagd_size;
425 		hashval = 0;
426 
427 		for (j = 0; j < agg->dtagd_nrecs - 1; j++) {
428 			rec = &agg->dtagd_rec[j];
429 			roffs = rec->dtrd_offset;
430 
431 			switch (rec->dtrd_action) {
432 			case DTRACEACT_USYM:
433 				dt_aggregate_usym(dtp,
434 				    /* LINTED - alignment */
435 				    (uint64_t *)&addr[roffs]);
436 				break;
437 
438 			case DTRACEACT_UMOD:
439 				dt_aggregate_umod(dtp,
440 				    /* LINTED - alignment */
441 				    (uint64_t *)&addr[roffs]);
442 				break;
443 
444 			case DTRACEACT_SYM:
445 				/* LINTED - alignment */
446 				dt_aggregate_sym(dtp, (uint64_t *)&addr[roffs]);
447 				break;
448 
449 			case DTRACEACT_MOD:
450 				/* LINTED - alignment */
451 				dt_aggregate_mod(dtp, (uint64_t *)&addr[roffs]);
452 				break;
453 
454 			default:
455 				break;
456 			}
457 
458 			for (i = 0; i < rec->dtrd_size; i++)
459 				hashval += addr[roffs + i];
460 		}
461 
462 		ndx = hashval % hash->dtah_size;
463 
464 		for (h = hash->dtah_hash[ndx]; h != NULL; h = h->dtahe_next) {
465 			if (h->dtahe_hashval != hashval)
466 				continue;
467 
468 			if (h->dtahe_size != size)
469 				continue;
470 
471 			aggdata = &h->dtahe_data;
472 			data = aggdata->dtada_data;
473 
474 			for (j = 0; j < agg->dtagd_nrecs - 1; j++) {
475 				rec = &agg->dtagd_rec[j];
476 				roffs = rec->dtrd_offset;
477 
478 				for (i = 0; i < rec->dtrd_size; i++)
479 					if (addr[roffs + i] != data[roffs + i])
480 						goto hashnext;
481 			}
482 
483 			/*
484 			 * We found it.  Now we need to apply the aggregating
485 			 * action on the data here.
486 			 */
487 			rec = &agg->dtagd_rec[agg->dtagd_nrecs - 1];
488 			roffs = rec->dtrd_offset;
489 			/* LINTED - alignment */
490 			h->dtahe_aggregate((int64_t *)&data[roffs],
491 			    /* LINTED - alignment */
492 			    (int64_t *)&addr[roffs], rec->dtrd_size);
493 
494 			/*
495 			 * If we're keeping per CPU data, apply the aggregating
496 			 * action there as well.
497 			 */
498 			if (aggdata->dtada_percpu != NULL) {
499 				data = aggdata->dtada_percpu[cpu];
500 
501 				/* LINTED - alignment */
502 				h->dtahe_aggregate((int64_t *)data,
503 				    /* LINTED - alignment */
504 				    (int64_t *)&addr[roffs], rec->dtrd_size);
505 			}
506 
507 			goto bufnext;
508 hashnext:
509 			continue;
510 		}
511 
512 		/*
513 		 * If we're here, we couldn't find an entry for this record.
514 		 */
515 		if ((h = malloc(sizeof (dt_ahashent_t))) == NULL)
516 			return (dt_set_errno(dtp, EDT_NOMEM));
517 		bzero(h, sizeof (dt_ahashent_t));
518 		aggdata = &h->dtahe_data;
519 
520 		if ((aggdata->dtada_data = malloc(size)) == NULL) {
521 			free(h);
522 			return (dt_set_errno(dtp, EDT_NOMEM));
523 		}
524 
525 		bcopy(addr, aggdata->dtada_data, size);
526 		aggdata->dtada_size = size;
527 		aggdata->dtada_desc = agg;
528 		aggdata->dtada_handle = dtp;
529 		(void) dt_epid_lookup(dtp, agg->dtagd_epid,
530 		    &aggdata->dtada_edesc, &aggdata->dtada_pdesc);
531 		aggdata->dtada_normal = 1;
532 
533 		h->dtahe_hashval = hashval;
534 		h->dtahe_size = size;
535 		(void) dt_aggregate_aggvarid(h);
536 
537 		rec = &agg->dtagd_rec[agg->dtagd_nrecs - 1];
538 
539 		if (flags & DTRACE_A_PERCPU) {
540 			int max_cpus = agp->dtat_maxcpu;
541 			caddr_t *percpu = malloc(max_cpus * sizeof (caddr_t));
542 
543 			if (percpu == NULL) {
544 				free(aggdata->dtada_data);
545 				free(h);
546 				return (dt_set_errno(dtp, EDT_NOMEM));
547 			}
548 
549 			for (j = 0; j < max_cpus; j++) {
550 				percpu[j] = malloc(rec->dtrd_size);
551 
552 				if (percpu[j] == NULL) {
553 					while (--j >= 0)
554 						free(percpu[j]);
555 
556 					free(aggdata->dtada_data);
557 					free(h);
558 					return (dt_set_errno(dtp, EDT_NOMEM));
559 				}
560 
561 				if (j == cpu) {
562 					bcopy(&addr[rec->dtrd_offset],
563 					    percpu[j], rec->dtrd_size);
564 				} else {
565 					bzero(percpu[j], rec->dtrd_size);
566 				}
567 			}
568 
569 			aggdata->dtada_percpu = percpu;
570 		}
571 
572 		switch (rec->dtrd_action) {
573 		case DTRACEAGG_MIN:
574 			h->dtahe_aggregate = dt_aggregate_min;
575 			break;
576 
577 		case DTRACEAGG_MAX:
578 			h->dtahe_aggregate = dt_aggregate_max;
579 			break;
580 
581 		case DTRACEAGG_LQUANTIZE:
582 			h->dtahe_aggregate = dt_aggregate_lquantize;
583 			break;
584 
585 		case DTRACEAGG_COUNT:
586 		case DTRACEAGG_SUM:
587 		case DTRACEAGG_AVG:
588 		case DTRACEAGG_STDDEV:
589 		case DTRACEAGG_QUANTIZE:
590 			h->dtahe_aggregate = dt_aggregate_count;
591 			break;
592 
593 		default:
594 			return (dt_set_errno(dtp, EDT_BADAGG));
595 		}
596 
597 		if (hash->dtah_hash[ndx] != NULL)
598 			hash->dtah_hash[ndx]->dtahe_prev = h;
599 
600 		h->dtahe_next = hash->dtah_hash[ndx];
601 		hash->dtah_hash[ndx] = h;
602 
603 		if (hash->dtah_all != NULL)
604 			hash->dtah_all->dtahe_prevall = h;
605 
606 		h->dtahe_nextall = hash->dtah_all;
607 		hash->dtah_all = h;
608 bufnext:
609 		offs += agg->dtagd_size;
610 	}
611 
612 	return (0);
613 }
614 
615 int
616 dtrace_aggregate_snap(dtrace_hdl_t *dtp)
617 {
618 	int i, rval;
619 	dt_aggregate_t *agp = &dtp->dt_aggregate;
620 	hrtime_t now = gethrtime();
621 	dtrace_optval_t interval = dtp->dt_options[DTRACEOPT_AGGRATE];
622 
623 	if (dtp->dt_lastagg != 0) {
624 		if (now - dtp->dt_lastagg < interval)
625 			return (0);
626 
627 		dtp->dt_lastagg += interval;
628 	} else {
629 		dtp->dt_lastagg = now;
630 	}
631 
632 	if (!dtp->dt_active)
633 		return (dt_set_errno(dtp, EINVAL));
634 
635 	if (agp->dtat_buf.dtbd_size == 0)
636 		return (0);
637 
638 	for (i = 0; i < agp->dtat_ncpus; i++) {
639 		if (rval = dt_aggregate_snap_cpu(dtp, agp->dtat_cpus[i]))
640 			return (rval);
641 	}
642 
643 	return (0);
644 }
645 
646 static int
647 dt_aggregate_hashcmp(const void *lhs, const void *rhs)
648 {
649 	dt_ahashent_t *lh = *((dt_ahashent_t **)lhs);
650 	dt_ahashent_t *rh = *((dt_ahashent_t **)rhs);
651 	dtrace_aggdesc_t *lagg = lh->dtahe_data.dtada_desc;
652 	dtrace_aggdesc_t *ragg = rh->dtahe_data.dtada_desc;
653 
654 	if (lagg->dtagd_nrecs < ragg->dtagd_nrecs)
655 		return (DT_LESSTHAN);
656 
657 	if (lagg->dtagd_nrecs > ragg->dtagd_nrecs)
658 		return (DT_GREATERTHAN);
659 
660 	return (0);
661 }
662 
663 static int
664 dt_aggregate_varcmp(const void *lhs, const void *rhs)
665 {
666 	dt_ahashent_t *lh = *((dt_ahashent_t **)lhs);
667 	dt_ahashent_t *rh = *((dt_ahashent_t **)rhs);
668 	dtrace_aggvarid_t lid, rid;
669 
670 	lid = dt_aggregate_aggvarid(lh);
671 	rid = dt_aggregate_aggvarid(rh);
672 
673 	if (lid < rid)
674 		return (DT_LESSTHAN);
675 
676 	if (lid > rid)
677 		return (DT_GREATERTHAN);
678 
679 	return (0);
680 }
681 
682 static int
683 dt_aggregate_keycmp(const void *lhs, const void *rhs)
684 {
685 	dt_ahashent_t *lh = *((dt_ahashent_t **)lhs);
686 	dt_ahashent_t *rh = *((dt_ahashent_t **)rhs);
687 	dtrace_aggdesc_t *lagg = lh->dtahe_data.dtada_desc;
688 	dtrace_aggdesc_t *ragg = rh->dtahe_data.dtada_desc;
689 	dtrace_recdesc_t *lrec, *rrec;
690 	char *ldata, *rdata;
691 	int rval, i, j, keypos, nrecs;
692 
693 	if ((rval = dt_aggregate_hashcmp(lhs, rhs)) != 0)
694 		return (rval);
695 
696 	nrecs = lagg->dtagd_nrecs - 1;
697 	assert(nrecs == ragg->dtagd_nrecs - 1);
698 
699 	keypos = dt_keypos + 1 >= nrecs ? 0 : dt_keypos;
700 
701 	for (i = 1; i < nrecs; i++) {
702 		uint64_t lval, rval;
703 		int ndx = i + keypos;
704 
705 		if (ndx >= nrecs)
706 			ndx = ndx - nrecs + 1;
707 
708 		lrec = &lagg->dtagd_rec[ndx];
709 		rrec = &ragg->dtagd_rec[ndx];
710 
711 		ldata = lh->dtahe_data.dtada_data + lrec->dtrd_offset;
712 		rdata = rh->dtahe_data.dtada_data + rrec->dtrd_offset;
713 
714 		if (lrec->dtrd_size < rrec->dtrd_size)
715 			return (DT_LESSTHAN);
716 
717 		if (lrec->dtrd_size > rrec->dtrd_size)
718 			return (DT_GREATERTHAN);
719 
720 		switch (lrec->dtrd_size) {
721 		case sizeof (uint64_t):
722 			/* LINTED - alignment */
723 			lval = *((uint64_t *)ldata);
724 			/* LINTED - alignment */
725 			rval = *((uint64_t *)rdata);
726 			break;
727 
728 		case sizeof (uint32_t):
729 			/* LINTED - alignment */
730 			lval = *((uint32_t *)ldata);
731 			/* LINTED - alignment */
732 			rval = *((uint32_t *)rdata);
733 			break;
734 
735 		case sizeof (uint16_t):
736 			/* LINTED - alignment */
737 			lval = *((uint16_t *)ldata);
738 			/* LINTED - alignment */
739 			rval = *((uint16_t *)rdata);
740 			break;
741 
742 		case sizeof (uint8_t):
743 			lval = *((uint8_t *)ldata);
744 			rval = *((uint8_t *)rdata);
745 			break;
746 
747 		default:
748 			switch (lrec->dtrd_action) {
749 			case DTRACEACT_UMOD:
750 			case DTRACEACT_UADDR:
751 			case DTRACEACT_USYM:
752 				for (j = 0; j < 2; j++) {
753 					/* LINTED - alignment */
754 					lval = ((uint64_t *)ldata)[j];
755 					/* LINTED - alignment */
756 					rval = ((uint64_t *)rdata)[j];
757 
758 					if (lval < rval)
759 						return (DT_LESSTHAN);
760 
761 					if (lval > rval)
762 						return (DT_GREATERTHAN);
763 				}
764 
765 				break;
766 
767 			default:
768 				for (j = 0; j < lrec->dtrd_size; j++) {
769 					lval = ((uint8_t *)ldata)[j];
770 					rval = ((uint8_t *)rdata)[j];
771 
772 					if (lval < rval)
773 						return (DT_LESSTHAN);
774 
775 					if (lval > rval)
776 						return (DT_GREATERTHAN);
777 				}
778 			}
779 
780 			continue;
781 		}
782 
783 		if (lval < rval)
784 			return (DT_LESSTHAN);
785 
786 		if (lval > rval)
787 			return (DT_GREATERTHAN);
788 	}
789 
790 	return (0);
791 }
792 
793 static int
794 dt_aggregate_valcmp(const void *lhs, const void *rhs)
795 {
796 	dt_ahashent_t *lh = *((dt_ahashent_t **)lhs);
797 	dt_ahashent_t *rh = *((dt_ahashent_t **)rhs);
798 	dtrace_aggdesc_t *lagg = lh->dtahe_data.dtada_desc;
799 	dtrace_aggdesc_t *ragg = rh->dtahe_data.dtada_desc;
800 	caddr_t ldata = lh->dtahe_data.dtada_data;
801 	caddr_t rdata = rh->dtahe_data.dtada_data;
802 	dtrace_recdesc_t *lrec, *rrec;
803 	int64_t *laddr, *raddr;
804 	int rval, i;
805 
806 	if ((rval = dt_aggregate_hashcmp(lhs, rhs)) != 0)
807 		return (rval);
808 
809 	if (lagg->dtagd_nrecs > ragg->dtagd_nrecs)
810 		return (DT_GREATERTHAN);
811 
812 	if (lagg->dtagd_nrecs < ragg->dtagd_nrecs)
813 		return (DT_LESSTHAN);
814 
815 	for (i = 0; i < lagg->dtagd_nrecs; i++) {
816 		lrec = &lagg->dtagd_rec[i];
817 		rrec = &ragg->dtagd_rec[i];
818 
819 		if (lrec->dtrd_offset < rrec->dtrd_offset)
820 			return (DT_LESSTHAN);
821 
822 		if (lrec->dtrd_offset > rrec->dtrd_offset)
823 			return (DT_GREATERTHAN);
824 
825 		if (lrec->dtrd_action < rrec->dtrd_action)
826 			return (DT_LESSTHAN);
827 
828 		if (lrec->dtrd_action > rrec->dtrd_action)
829 			return (DT_GREATERTHAN);
830 	}
831 
832 	laddr = (int64_t *)(uintptr_t)(ldata + lrec->dtrd_offset);
833 	raddr = (int64_t *)(uintptr_t)(rdata + rrec->dtrd_offset);
834 
835 	switch (lrec->dtrd_action) {
836 	case DTRACEAGG_AVG:
837 		rval = dt_aggregate_averagecmp(laddr, raddr);
838 		break;
839 
840 	case DTRACEAGG_STDDEV:
841 		rval = dt_aggregate_stddevcmp(laddr, raddr);
842 		break;
843 
844 	case DTRACEAGG_QUANTIZE:
845 		rval = dt_aggregate_quantizedcmp(laddr, raddr);
846 		break;
847 
848 	case DTRACEAGG_LQUANTIZE:
849 		rval = dt_aggregate_lquantizedcmp(laddr, raddr);
850 		break;
851 
852 	case DTRACEAGG_COUNT:
853 	case DTRACEAGG_SUM:
854 	case DTRACEAGG_MIN:
855 	case DTRACEAGG_MAX:
856 		rval = dt_aggregate_countcmp(laddr, raddr);
857 		break;
858 
859 	default:
860 		assert(0);
861 	}
862 
863 	return (rval);
864 }
865 
866 static int
867 dt_aggregate_valkeycmp(const void *lhs, const void *rhs)
868 {
869 	int rval;
870 
871 	if ((rval = dt_aggregate_valcmp(lhs, rhs)) != 0)
872 		return (rval);
873 
874 	/*
875 	 * If we're here, the values for the two aggregation elements are
876 	 * equal.  We already know that the key layout is the same for the two
877 	 * elements; we must now compare the keys themselves as a tie-breaker.
878 	 */
879 	return (dt_aggregate_keycmp(lhs, rhs));
880 }
881 
882 static int
883 dt_aggregate_keyvarcmp(const void *lhs, const void *rhs)
884 {
885 	int rval;
886 
887 	if ((rval = dt_aggregate_keycmp(lhs, rhs)) != 0)
888 		return (rval);
889 
890 	return (dt_aggregate_varcmp(lhs, rhs));
891 }
892 
893 static int
894 dt_aggregate_varkeycmp(const void *lhs, const void *rhs)
895 {
896 	int rval;
897 
898 	if ((rval = dt_aggregate_varcmp(lhs, rhs)) != 0)
899 		return (rval);
900 
901 	return (dt_aggregate_keycmp(lhs, rhs));
902 }
903 
904 static int
905 dt_aggregate_valvarcmp(const void *lhs, const void *rhs)
906 {
907 	int rval;
908 
909 	if ((rval = dt_aggregate_valkeycmp(lhs, rhs)) != 0)
910 		return (rval);
911 
912 	return (dt_aggregate_varcmp(lhs, rhs));
913 }
914 
915 static int
916 dt_aggregate_varvalcmp(const void *lhs, const void *rhs)
917 {
918 	int rval;
919 
920 	if ((rval = dt_aggregate_varcmp(lhs, rhs)) != 0)
921 		return (rval);
922 
923 	return (dt_aggregate_valkeycmp(lhs, rhs));
924 }
925 
926 static int
927 dt_aggregate_keyvarrevcmp(const void *lhs, const void *rhs)
928 {
929 	return (dt_aggregate_keyvarcmp(rhs, lhs));
930 }
931 
932 static int
933 dt_aggregate_varkeyrevcmp(const void *lhs, const void *rhs)
934 {
935 	return (dt_aggregate_varkeycmp(rhs, lhs));
936 }
937 
938 static int
939 dt_aggregate_valvarrevcmp(const void *lhs, const void *rhs)
940 {
941 	return (dt_aggregate_valvarcmp(rhs, lhs));
942 }
943 
944 static int
945 dt_aggregate_varvalrevcmp(const void *lhs, const void *rhs)
946 {
947 	return (dt_aggregate_varvalcmp(rhs, lhs));
948 }
949 
950 static int
951 dt_aggregate_bundlecmp(const void *lhs, const void *rhs)
952 {
953 	dt_ahashent_t **lh = *((dt_ahashent_t ***)lhs);
954 	dt_ahashent_t **rh = *((dt_ahashent_t ***)rhs);
955 	int i, rval;
956 
957 	if (dt_keysort) {
958 		/*
959 		 * If we're sorting on keys, we need to scan until we find the
960 		 * last entry -- that's the representative key.  (The order of
961 		 * the bundle is values followed by key to accommodate the
962 		 * default behavior of sorting by value.)  If the keys are
963 		 * equal, we'll fall into the value comparison loop, below.
964 		 */
965 		for (i = 0; lh[i + 1] != NULL; i++)
966 			continue;
967 
968 		assert(i != 0);
969 		assert(rh[i + 1] == NULL);
970 
971 		if ((rval = dt_aggregate_keycmp(&lh[i], &rh[i])) != 0)
972 			return (rval);
973 	}
974 
975 	for (i = 0; ; i++) {
976 		if (lh[i + 1] == NULL) {
977 			/*
978 			 * All of the values are equal; if we're sorting on
979 			 * keys, then we're only here because the keys were
980 			 * found to be equal and these records are therefore
981 			 * equal.  If we're not sorting on keys, we'll use the
982 			 * key comparison from the representative key as the
983 			 * tie-breaker.
984 			 */
985 			if (dt_keysort)
986 				return (0);
987 
988 			assert(i != 0);
989 			assert(rh[i + 1] == NULL);
990 			return (dt_aggregate_keycmp(&lh[i], &rh[i]));
991 		} else {
992 			if ((rval = dt_aggregate_valcmp(&lh[i], &rh[i])) != 0)
993 				return (rval);
994 		}
995 	}
996 }
997 
998 int
999 dt_aggregate_go(dtrace_hdl_t *dtp)
1000 {
1001 	dt_aggregate_t *agp = &dtp->dt_aggregate;
1002 	dtrace_optval_t size, cpu;
1003 	dtrace_bufdesc_t *buf = &agp->dtat_buf;
1004 	int rval, i;
1005 
1006 	assert(agp->dtat_maxcpu == 0);
1007 	assert(agp->dtat_ncpu == 0);
1008 	assert(agp->dtat_cpus == NULL);
1009 
1010 	agp->dtat_maxcpu = dt_sysconf(dtp, _SC_CPUID_MAX) + 1;
1011 	agp->dtat_ncpu = dt_sysconf(dtp, _SC_NPROCESSORS_MAX);
1012 	agp->dtat_cpus = malloc(agp->dtat_ncpu * sizeof (processorid_t));
1013 
1014 	if (agp->dtat_cpus == NULL)
1015 		return (dt_set_errno(dtp, EDT_NOMEM));
1016 
1017 	/*
1018 	 * Use the aggregation buffer size as reloaded from the kernel.
1019 	 */
1020 	size = dtp->dt_options[DTRACEOPT_AGGSIZE];
1021 
1022 	rval = dtrace_getopt(dtp, "aggsize", &size);
1023 	assert(rval == 0);
1024 
1025 	if (size == 0 || size == DTRACEOPT_UNSET)
1026 		return (0);
1027 
1028 	buf = &agp->dtat_buf;
1029 	buf->dtbd_size = size;
1030 
1031 	if ((buf->dtbd_data = malloc(buf->dtbd_size)) == NULL)
1032 		return (dt_set_errno(dtp, EDT_NOMEM));
1033 
1034 	/*
1035 	 * Now query for the CPUs enabled.
1036 	 */
1037 	rval = dtrace_getopt(dtp, "cpu", &cpu);
1038 	assert(rval == 0 && cpu != DTRACEOPT_UNSET);
1039 
1040 	if (cpu != DTRACE_CPUALL) {
1041 		assert(cpu < agp->dtat_ncpu);
1042 		agp->dtat_cpus[agp->dtat_ncpus++] = (processorid_t)cpu;
1043 
1044 		return (0);
1045 	}
1046 
1047 	agp->dtat_ncpus = 0;
1048 	for (i = 0; i < agp->dtat_maxcpu; i++) {
1049 		if (dt_status(dtp, i) == -1)
1050 			continue;
1051 
1052 		agp->dtat_cpus[agp->dtat_ncpus++] = i;
1053 	}
1054 
1055 	return (0);
1056 }
1057 
1058 static int
1059 dt_aggwalk_rval(dtrace_hdl_t *dtp, dt_ahashent_t *h, int rval)
1060 {
1061 	dt_aggregate_t *agp = &dtp->dt_aggregate;
1062 	dtrace_aggdata_t *data;
1063 	dtrace_aggdesc_t *aggdesc;
1064 	dtrace_recdesc_t *rec;
1065 	int i;
1066 
1067 	switch (rval) {
1068 	case DTRACE_AGGWALK_NEXT:
1069 		break;
1070 
1071 	case DTRACE_AGGWALK_CLEAR: {
1072 		uint32_t size, offs = 0;
1073 
1074 		aggdesc = h->dtahe_data.dtada_desc;
1075 		rec = &aggdesc->dtagd_rec[aggdesc->dtagd_nrecs - 1];
1076 		size = rec->dtrd_size;
1077 		data = &h->dtahe_data;
1078 
1079 		if (rec->dtrd_action == DTRACEAGG_LQUANTIZE) {
1080 			offs = sizeof (uint64_t);
1081 			size -= sizeof (uint64_t);
1082 		}
1083 
1084 		bzero(&data->dtada_data[rec->dtrd_offset] + offs, size);
1085 
1086 		if (data->dtada_percpu == NULL)
1087 			break;
1088 
1089 		for (i = 0; i < dtp->dt_aggregate.dtat_maxcpu; i++)
1090 			bzero(data->dtada_percpu[i] + offs, size);
1091 		break;
1092 	}
1093 
1094 	case DTRACE_AGGWALK_ERROR:
1095 		/*
1096 		 * We assume that errno is already set in this case.
1097 		 */
1098 		return (dt_set_errno(dtp, errno));
1099 
1100 	case DTRACE_AGGWALK_ABORT:
1101 		return (dt_set_errno(dtp, EDT_DIRABORT));
1102 
1103 	case DTRACE_AGGWALK_DENORMALIZE:
1104 		h->dtahe_data.dtada_normal = 1;
1105 		return (0);
1106 
1107 	case DTRACE_AGGWALK_NORMALIZE:
1108 		if (h->dtahe_data.dtada_normal == 0) {
1109 			h->dtahe_data.dtada_normal = 1;
1110 			return (dt_set_errno(dtp, EDT_BADRVAL));
1111 		}
1112 
1113 		return (0);
1114 
1115 	case DTRACE_AGGWALK_REMOVE: {
1116 		dtrace_aggdata_t *aggdata = &h->dtahe_data;
1117 		int i, max_cpus = agp->dtat_maxcpu;
1118 
1119 		/*
1120 		 * First, remove this hash entry from its hash chain.
1121 		 */
1122 		if (h->dtahe_prev != NULL) {
1123 			h->dtahe_prev->dtahe_next = h->dtahe_next;
1124 		} else {
1125 			dt_ahash_t *hash = &agp->dtat_hash;
1126 			size_t ndx = h->dtahe_hashval % hash->dtah_size;
1127 
1128 			assert(hash->dtah_hash[ndx] == h);
1129 			hash->dtah_hash[ndx] = h->dtahe_next;
1130 		}
1131 
1132 		if (h->dtahe_next != NULL)
1133 			h->dtahe_next->dtahe_prev = h->dtahe_prev;
1134 
1135 		/*
1136 		 * Now remove it from the list of all hash entries.
1137 		 */
1138 		if (h->dtahe_prevall != NULL) {
1139 			h->dtahe_prevall->dtahe_nextall = h->dtahe_nextall;
1140 		} else {
1141 			dt_ahash_t *hash = &agp->dtat_hash;
1142 
1143 			assert(hash->dtah_all == h);
1144 			hash->dtah_all = h->dtahe_nextall;
1145 		}
1146 
1147 		if (h->dtahe_nextall != NULL)
1148 			h->dtahe_nextall->dtahe_prevall = h->dtahe_prevall;
1149 
1150 		/*
1151 		 * We're unlinked.  We can safely destroy the data.
1152 		 */
1153 		if (aggdata->dtada_percpu != NULL) {
1154 			for (i = 0; i < max_cpus; i++)
1155 				free(aggdata->dtada_percpu[i]);
1156 			free(aggdata->dtada_percpu);
1157 		}
1158 
1159 		free(aggdata->dtada_data);
1160 		free(h);
1161 
1162 		return (0);
1163 	}
1164 
1165 	default:
1166 		return (dt_set_errno(dtp, EDT_BADRVAL));
1167 	}
1168 
1169 	return (0);
1170 }
1171 
1172 void
1173 dt_aggregate_qsort(dtrace_hdl_t *dtp, void *base, size_t nel, size_t width,
1174     int (*compar)(const void *, const void *))
1175 {
1176 	int rev = dt_revsort, key = dt_keysort, keypos = dt_keypos;
1177 	dtrace_optval_t keyposopt = dtp->dt_options[DTRACEOPT_AGGSORTKEYPOS];
1178 
1179 	dt_revsort = (dtp->dt_options[DTRACEOPT_AGGSORTREV] != DTRACEOPT_UNSET);
1180 	dt_keysort = (dtp->dt_options[DTRACEOPT_AGGSORTKEY] != DTRACEOPT_UNSET);
1181 
1182 	if (keyposopt != DTRACEOPT_UNSET && keyposopt <= INT_MAX) {
1183 		dt_keypos = (int)keyposopt;
1184 	} else {
1185 		dt_keypos = 0;
1186 	}
1187 
1188 	if (compar == NULL) {
1189 		if (!dt_keysort) {
1190 			compar = dt_aggregate_varvalcmp;
1191 		} else {
1192 			compar = dt_aggregate_varkeycmp;
1193 		}
1194 	}
1195 
1196 	qsort(base, nel, width, compar);
1197 
1198 	dt_revsort = rev;
1199 	dt_keysort = key;
1200 	dt_keypos = keypos;
1201 }
1202 
1203 int
1204 dtrace_aggregate_walk(dtrace_hdl_t *dtp, dtrace_aggregate_f *func, void *arg)
1205 {
1206 	dt_ahashent_t *h, *next;
1207 	dt_ahash_t *hash = &dtp->dt_aggregate.dtat_hash;
1208 
1209 	for (h = hash->dtah_all; h != NULL; h = next) {
1210 		/*
1211 		 * dt_aggwalk_rval() can potentially remove the current hash
1212 		 * entry; we need to load the next hash entry before calling
1213 		 * into it.
1214 		 */
1215 		next = h->dtahe_nextall;
1216 
1217 		if (dt_aggwalk_rval(dtp, h, func(&h->dtahe_data, arg)) == -1)
1218 			return (-1);
1219 	}
1220 
1221 	return (0);
1222 }
1223 
1224 static int
1225 dt_aggregate_walk_sorted(dtrace_hdl_t *dtp,
1226     dtrace_aggregate_f *func, void *arg,
1227     int (*sfunc)(const void *, const void *))
1228 {
1229 	dt_aggregate_t *agp = &dtp->dt_aggregate;
1230 	dt_ahashent_t *h, **sorted;
1231 	dt_ahash_t *hash = &agp->dtat_hash;
1232 	size_t i, nentries = 0;
1233 
1234 	for (h = hash->dtah_all; h != NULL; h = h->dtahe_nextall)
1235 		nentries++;
1236 
1237 	sorted = dt_alloc(dtp, nentries * sizeof (dt_ahashent_t *));
1238 
1239 	if (sorted == NULL)
1240 		return (-1);
1241 
1242 	for (h = hash->dtah_all, i = 0; h != NULL; h = h->dtahe_nextall)
1243 		sorted[i++] = h;
1244 
1245 	(void) pthread_mutex_lock(&dt_qsort_lock);
1246 
1247 	if (sfunc == NULL) {
1248 		dt_aggregate_qsort(dtp, sorted, nentries,
1249 		    sizeof (dt_ahashent_t *), NULL);
1250 	} else {
1251 		/*
1252 		 * If we've been explicitly passed a sorting function,
1253 		 * we'll use that -- ignoring the values of the "aggsortrev",
1254 		 * "aggsortkey" and "aggsortkeypos" options.
1255 		 */
1256 		qsort(sorted, nentries, sizeof (dt_ahashent_t *), sfunc);
1257 	}
1258 
1259 	(void) pthread_mutex_unlock(&dt_qsort_lock);
1260 
1261 	for (i = 0; i < nentries; i++) {
1262 		h = sorted[i];
1263 
1264 		if (dt_aggwalk_rval(dtp, h, func(&h->dtahe_data, arg)) == -1) {
1265 			dt_free(dtp, sorted);
1266 			return (-1);
1267 		}
1268 	}
1269 
1270 	dt_free(dtp, sorted);
1271 	return (0);
1272 }
1273 
1274 int
1275 dtrace_aggregate_walk_sorted(dtrace_hdl_t *dtp,
1276     dtrace_aggregate_f *func, void *arg)
1277 {
1278 	return (dt_aggregate_walk_sorted(dtp, func, arg, NULL));
1279 }
1280 
1281 int
1282 dtrace_aggregate_walk_keysorted(dtrace_hdl_t *dtp,
1283     dtrace_aggregate_f *func, void *arg)
1284 {
1285 	return (dt_aggregate_walk_sorted(dtp, func,
1286 	    arg, dt_aggregate_varkeycmp));
1287 }
1288 
1289 int
1290 dtrace_aggregate_walk_valsorted(dtrace_hdl_t *dtp,
1291     dtrace_aggregate_f *func, void *arg)
1292 {
1293 	return (dt_aggregate_walk_sorted(dtp, func,
1294 	    arg, dt_aggregate_varvalcmp));
1295 }
1296 
1297 int
1298 dtrace_aggregate_walk_keyvarsorted(dtrace_hdl_t *dtp,
1299     dtrace_aggregate_f *func, void *arg)
1300 {
1301 	return (dt_aggregate_walk_sorted(dtp, func,
1302 	    arg, dt_aggregate_keyvarcmp));
1303 }
1304 
1305 int
1306 dtrace_aggregate_walk_valvarsorted(dtrace_hdl_t *dtp,
1307     dtrace_aggregate_f *func, void *arg)
1308 {
1309 	return (dt_aggregate_walk_sorted(dtp, func,
1310 	    arg, dt_aggregate_valvarcmp));
1311 }
1312 
1313 int
1314 dtrace_aggregate_walk_keyrevsorted(dtrace_hdl_t *dtp,
1315     dtrace_aggregate_f *func, void *arg)
1316 {
1317 	return (dt_aggregate_walk_sorted(dtp, func,
1318 	    arg, dt_aggregate_varkeyrevcmp));
1319 }
1320 
1321 int
1322 dtrace_aggregate_walk_valrevsorted(dtrace_hdl_t *dtp,
1323     dtrace_aggregate_f *func, void *arg)
1324 {
1325 	return (dt_aggregate_walk_sorted(dtp, func,
1326 	    arg, dt_aggregate_varvalrevcmp));
1327 }
1328 
1329 int
1330 dtrace_aggregate_walk_keyvarrevsorted(dtrace_hdl_t *dtp,
1331     dtrace_aggregate_f *func, void *arg)
1332 {
1333 	return (dt_aggregate_walk_sorted(dtp, func,
1334 	    arg, dt_aggregate_keyvarrevcmp));
1335 }
1336 
1337 int
1338 dtrace_aggregate_walk_valvarrevsorted(dtrace_hdl_t *dtp,
1339     dtrace_aggregate_f *func, void *arg)
1340 {
1341 	return (dt_aggregate_walk_sorted(dtp, func,
1342 	    arg, dt_aggregate_valvarrevcmp));
1343 }
1344 
1345 int
1346 dtrace_aggregate_walk_joined(dtrace_hdl_t *dtp, dtrace_aggvarid_t *aggvars,
1347     int naggvars, dtrace_aggregate_walk_joined_f *func, void *arg)
1348 {
1349 	dt_aggregate_t *agp = &dtp->dt_aggregate;
1350 	dt_ahashent_t *h, **sorted = NULL, ***bundle, **nbundle;
1351 	const dtrace_aggdata_t **data;
1352 	dt_ahashent_t *zaggdata = NULL;
1353 	dt_ahash_t *hash = &agp->dtat_hash;
1354 	size_t nentries = 0, nbundles = 0, start, zsize = 0, bundlesize;
1355 	dtrace_aggvarid_t max = 0, aggvar;
1356 	int rval = -1, *map, *remap = NULL;
1357 	int i, j;
1358 	dtrace_optval_t sortpos = dtp->dt_options[DTRACEOPT_AGGSORTPOS];
1359 
1360 	/*
1361 	 * If the sorting position is greater than the number of aggregation
1362 	 * variable IDs, we silently set it to 0.
1363 	 */
1364 	if (sortpos == DTRACEOPT_UNSET || sortpos >= naggvars)
1365 		sortpos = 0;
1366 
1367 	/*
1368 	 * First we need to translate the specified aggregation variable IDs
1369 	 * into a linear map that will allow us to translate an aggregation
1370 	 * variable ID into its position in the specified aggvars.
1371 	 */
1372 	for (i = 0; i < naggvars; i++) {
1373 		if (aggvars[i] == DTRACE_AGGVARIDNONE || aggvars[i] < 0)
1374 			return (dt_set_errno(dtp, EDT_BADAGGVAR));
1375 
1376 		if (aggvars[i] > max)
1377 			max = aggvars[i];
1378 	}
1379 
1380 	if ((map = dt_zalloc(dtp, (max + 1) * sizeof (int))) == NULL)
1381 		return (-1);
1382 
1383 	zaggdata = dt_zalloc(dtp, naggvars * sizeof (dt_ahashent_t));
1384 
1385 	if (zaggdata == NULL)
1386 		goto out;
1387 
1388 	for (i = 0; i < naggvars; i++) {
1389 		int ndx = i + sortpos;
1390 
1391 		if (ndx >= naggvars)
1392 			ndx -= naggvars;
1393 
1394 		aggvar = aggvars[ndx];
1395 		assert(aggvar <= max);
1396 
1397 		if (map[aggvar]) {
1398 			/*
1399 			 * We have an aggregation variable that is present
1400 			 * more than once in the array of aggregation
1401 			 * variables.  While it's unclear why one might want
1402 			 * to do this, it's legal.  To support this construct,
1403 			 * we will allocate a remap that will indicate the
1404 			 * position from which this aggregation variable
1405 			 * should be pulled.  (That is, where the remap will
1406 			 * map from one position to another.)
1407 			 */
1408 			if (remap == NULL) {
1409 				remap = dt_zalloc(dtp, naggvars * sizeof (int));
1410 
1411 				if (remap == NULL)
1412 					goto out;
1413 			}
1414 
1415 			/*
1416 			 * Given that the variable is already present, assert
1417 			 * that following through the mapping and adjusting
1418 			 * for the sort position yields the same aggregation
1419 			 * variable ID.
1420 			 */
1421 			assert(aggvars[(map[aggvar] - 1 + sortpos) %
1422 			    naggvars] == aggvars[ndx]);
1423 
1424 			remap[i] = map[aggvar];
1425 			continue;
1426 		}
1427 
1428 		map[aggvar] = i + 1;
1429 	}
1430 
1431 	/*
1432 	 * We need to take two passes over the data to size our allocation, so
1433 	 * we'll use the first pass to also fill in the zero-filled data to be
1434 	 * used to properly format a zero-valued aggregation.
1435 	 */
1436 	for (h = hash->dtah_all; h != NULL; h = h->dtahe_nextall) {
1437 		dtrace_aggvarid_t id;
1438 		int ndx;
1439 
1440 		if ((id = dt_aggregate_aggvarid(h)) > max || !(ndx = map[id]))
1441 			continue;
1442 
1443 		if (zaggdata[ndx - 1].dtahe_size == 0) {
1444 			zaggdata[ndx - 1].dtahe_size = h->dtahe_size;
1445 			zaggdata[ndx - 1].dtahe_data = h->dtahe_data;
1446 		}
1447 
1448 		nentries++;
1449 	}
1450 
1451 	if (nentries == 0) {
1452 		/*
1453 		 * We couldn't find any entries; there is nothing else to do.
1454 		 */
1455 		rval = 0;
1456 		goto out;
1457 	}
1458 
1459 	/*
1460 	 * Before we sort the data, we're going to look for any holes in our
1461 	 * zero-filled data.  This will occur if an aggregation variable that
1462 	 * we are being asked to print has not yet been assigned the result of
1463 	 * any aggregating action for _any_ tuple.  The issue becomes that we
1464 	 * would like a zero value to be printed for all columns for this
1465 	 * aggregation, but without any record description, we don't know the
1466 	 * aggregating action that corresponds to the aggregation variable.  To
1467 	 * try to find a match, we're simply going to lookup aggregation IDs
1468 	 * (which are guaranteed to be contiguous and to start from 1), looking
1469 	 * for the specified aggregation variable ID.  If we find a match,
1470 	 * we'll use that.  If we iterate over all aggregation IDs and don't
1471 	 * find a match, then we must be an anonymous enabling.  (Anonymous
1472 	 * enablings can't currently derive either aggregation variable IDs or
1473 	 * aggregation variable names given only an aggregation ID.)  In this
1474 	 * obscure case (anonymous enabling, multiple aggregation printa() with
1475 	 * some aggregations not represented for any tuple), our defined
1476 	 * behavior is that the zero will be printed in the format of the first
1477 	 * aggregation variable that contains any non-zero value.
1478 	 */
1479 	for (i = 0; i < naggvars; i++) {
1480 		if (zaggdata[i].dtahe_size == 0) {
1481 			dtrace_aggvarid_t aggvar;
1482 
1483 			aggvar = aggvars[(i - sortpos + naggvars) % naggvars];
1484 			assert(zaggdata[i].dtahe_data.dtada_data == NULL);
1485 
1486 			for (j = DTRACE_AGGIDNONE + 1; ; j++) {
1487 				dtrace_aggdesc_t *agg;
1488 				dtrace_aggdata_t *aggdata;
1489 
1490 				if (dt_aggid_lookup(dtp, j, &agg) != 0)
1491 					break;
1492 
1493 				if (agg->dtagd_varid != aggvar)
1494 					continue;
1495 
1496 				/*
1497 				 * We have our description -- now we need to
1498 				 * cons up the zaggdata entry for it.
1499 				 */
1500 				aggdata = &zaggdata[i].dtahe_data;
1501 				aggdata->dtada_size = agg->dtagd_size;
1502 				aggdata->dtada_desc = agg;
1503 				aggdata->dtada_handle = dtp;
1504 				(void) dt_epid_lookup(dtp, agg->dtagd_epid,
1505 				    &aggdata->dtada_edesc,
1506 				    &aggdata->dtada_pdesc);
1507 				aggdata->dtada_normal = 1;
1508 				zaggdata[i].dtahe_hashval = 0;
1509 				zaggdata[i].dtahe_size = agg->dtagd_size;
1510 				break;
1511 			}
1512 
1513 			if (zaggdata[i].dtahe_size == 0) {
1514 				caddr_t data;
1515 
1516 				/*
1517 				 * We couldn't find this aggregation, meaning
1518 				 * that we have never seen it before for any
1519 				 * tuple _and_ this is an anonymous enabling.
1520 				 * That is, we're in the obscure case outlined
1521 				 * above.  In this case, our defined behavior
1522 				 * is to format the data in the format of the
1523 				 * first non-zero aggregation -- of which, of
1524 				 * course, we know there to be at least one
1525 				 * (or nentries would have been zero).
1526 				 */
1527 				for (j = 0; j < naggvars; j++) {
1528 					if (zaggdata[j].dtahe_size != 0)
1529 						break;
1530 				}
1531 
1532 				assert(j < naggvars);
1533 				zaggdata[i] = zaggdata[j];
1534 
1535 				data = zaggdata[i].dtahe_data.dtada_data;
1536 				assert(data != NULL);
1537 			}
1538 		}
1539 	}
1540 
1541 	/*
1542 	 * Now we need to allocate our zero-filled data for use for
1543 	 * aggregations that don't have a value corresponding to a given key.
1544 	 */
1545 	for (i = 0; i < naggvars; i++) {
1546 		dtrace_aggdata_t *aggdata = &zaggdata[i].dtahe_data;
1547 		dtrace_aggdesc_t *aggdesc = aggdata->dtada_desc;
1548 		dtrace_recdesc_t *rec;
1549 		uint64_t larg;
1550 		caddr_t zdata;
1551 
1552 		zsize = zaggdata[i].dtahe_size;
1553 		assert(zsize != 0);
1554 
1555 		if ((zdata = dt_zalloc(dtp, zsize)) == NULL) {
1556 			/*
1557 			 * If we failed to allocated some zero-filled data, we
1558 			 * need to zero out the remaining dtada_data pointers
1559 			 * to prevent the wrong data from being freed below.
1560 			 */
1561 			for (j = i; j < naggvars; j++)
1562 				zaggdata[j].dtahe_data.dtada_data = NULL;
1563 			goto out;
1564 		}
1565 
1566 		aggvar = aggvars[(i - sortpos + naggvars) % naggvars];
1567 
1568 		/*
1569 		 * First, the easy bit.  To maintain compatibility with
1570 		 * consumers that pull the compiler-generated ID out of the
1571 		 * data, we put that ID at the top of the zero-filled data.
1572 		 */
1573 		rec = &aggdesc->dtagd_rec[0];
1574 		/* LINTED - alignment */
1575 		*((dtrace_aggvarid_t *)(zdata + rec->dtrd_offset)) = aggvar;
1576 
1577 		rec = &aggdesc->dtagd_rec[aggdesc->dtagd_nrecs - 1];
1578 
1579 		/*
1580 		 * Now for the more complicated part.  If (and only if) this
1581 		 * is an lquantize() aggregating action, zero-filled data is
1582 		 * not equivalent to an empty record:  we must also get the
1583 		 * parameters for the lquantize().
1584 		 */
1585 		if (rec->dtrd_action == DTRACEAGG_LQUANTIZE) {
1586 			if (aggdata->dtada_data != NULL) {
1587 				/*
1588 				 * The easier case here is if we actually have
1589 				 * some prototype data -- in which case we
1590 				 * manually dig it out of the aggregation
1591 				 * record.
1592 				 */
1593 				/* LINTED - alignment */
1594 				larg = *((uint64_t *)(aggdata->dtada_data +
1595 				    rec->dtrd_offset));
1596 			} else {
1597 				/*
1598 				 * We don't have any prototype data.  As a
1599 				 * result, we know that we _do_ have the
1600 				 * compiler-generated information.  (If this
1601 				 * were an anonymous enabling, all of our
1602 				 * zero-filled data would have prototype data
1603 				 * -- either directly or indirectly.) So as
1604 				 * gross as it is, we'll grovel around in the
1605 				 * compiler-generated information to find the
1606 				 * lquantize() parameters.
1607 				 */
1608 				dtrace_stmtdesc_t *sdp;
1609 				dt_ident_t *aid;
1610 				dt_idsig_t *isp;
1611 
1612 				sdp = (dtrace_stmtdesc_t *)(uintptr_t)
1613 				    aggdesc->dtagd_rec[0].dtrd_uarg;
1614 				aid = sdp->dtsd_aggdata;
1615 				isp = (dt_idsig_t *)aid->di_data;
1616 				assert(isp->dis_auxinfo != 0);
1617 				larg = isp->dis_auxinfo;
1618 			}
1619 
1620 			/* LINTED - alignment */
1621 			*((uint64_t *)(zdata + rec->dtrd_offset)) = larg;
1622 		}
1623 
1624 		aggdata->dtada_data = zdata;
1625 	}
1626 
1627 	/*
1628 	 * Now that we've dealt with setting up our zero-filled data, we can
1629 	 * allocate our sorted array, and take another pass over the data to
1630 	 * fill it.
1631 	 */
1632 	sorted = dt_alloc(dtp, nentries * sizeof (dt_ahashent_t *));
1633 
1634 	if (sorted == NULL)
1635 		goto out;
1636 
1637 	for (h = hash->dtah_all, i = 0; h != NULL; h = h->dtahe_nextall) {
1638 		dtrace_aggvarid_t id;
1639 
1640 		if ((id = dt_aggregate_aggvarid(h)) > max || !map[id])
1641 			continue;
1642 
1643 		sorted[i++] = h;
1644 	}
1645 
1646 	assert(i == nentries);
1647 
1648 	/*
1649 	 * We've loaded our array; now we need to sort by value to allow us
1650 	 * to create bundles of like value.  We're going to acquire the
1651 	 * dt_qsort_lock here, and hold it across all of our subsequent
1652 	 * comparison and sorting.
1653 	 */
1654 	(void) pthread_mutex_lock(&dt_qsort_lock);
1655 
1656 	qsort(sorted, nentries, sizeof (dt_ahashent_t *),
1657 	    dt_aggregate_keyvarcmp);
1658 
1659 	/*
1660 	 * Now we need to go through and create bundles.  Because the number
1661 	 * of bundles is bounded by the size of the sorted array, we're going
1662 	 * to reuse the underlying storage.  And note that "bundle" is an
1663 	 * array of pointers to arrays of pointers to dt_ahashent_t -- making
1664 	 * its type (regrettably) "dt_ahashent_t ***".  (Regrettable because
1665 	 * '*' -- like '_' and 'X' -- should never appear in triplicate in
1666 	 * an ideal world.)
1667 	 */
1668 	bundle = (dt_ahashent_t ***)sorted;
1669 
1670 	for (i = 1, start = 0; i <= nentries; i++) {
1671 		if (i < nentries &&
1672 		    dt_aggregate_keycmp(&sorted[i], &sorted[i - 1]) == 0)
1673 			continue;
1674 
1675 		/*
1676 		 * We have a bundle boundary.  Everything from start to
1677 		 * (i - 1) belongs in one bundle.
1678 		 */
1679 		assert(i - start <= naggvars);
1680 		bundlesize = (naggvars + 2) * sizeof (dt_ahashent_t *);
1681 
1682 		if ((nbundle = dt_zalloc(dtp, bundlesize)) == NULL) {
1683 			(void) pthread_mutex_unlock(&dt_qsort_lock);
1684 			goto out;
1685 		}
1686 
1687 		for (j = start; j < i; j++) {
1688 			dtrace_aggvarid_t id = dt_aggregate_aggvarid(sorted[j]);
1689 
1690 			assert(id <= max);
1691 			assert(map[id] != 0);
1692 			assert(map[id] - 1 < naggvars);
1693 			assert(nbundle[map[id] - 1] == NULL);
1694 			nbundle[map[id] - 1] = sorted[j];
1695 
1696 			if (nbundle[naggvars] == NULL)
1697 				nbundle[naggvars] = sorted[j];
1698 		}
1699 
1700 		for (j = 0; j < naggvars; j++) {
1701 			if (nbundle[j] != NULL)
1702 				continue;
1703 
1704 			/*
1705 			 * Before we assume that this aggregation variable
1706 			 * isn't present (and fall back to using the
1707 			 * zero-filled data allocated earlier), check the
1708 			 * remap.  If we have a remapping, we'll drop it in
1709 			 * here.  Note that we might be remapping an
1710 			 * aggregation variable that isn't present for this
1711 			 * key; in this case, the aggregation data that we
1712 			 * copy will point to the zeroed data.
1713 			 */
1714 			if (remap != NULL && remap[j]) {
1715 				assert(remap[j] - 1 < j);
1716 				assert(nbundle[remap[j] - 1] != NULL);
1717 				nbundle[j] = nbundle[remap[j] - 1];
1718 			} else {
1719 				nbundle[j] = &zaggdata[j];
1720 			}
1721 		}
1722 
1723 		bundle[nbundles++] = nbundle;
1724 		start = i;
1725 	}
1726 
1727 	/*
1728 	 * Now we need to re-sort based on the first value.
1729 	 */
1730 	dt_aggregate_qsort(dtp, bundle, nbundles, sizeof (dt_ahashent_t **),
1731 	    dt_aggregate_bundlecmp);
1732 
1733 	(void) pthread_mutex_unlock(&dt_qsort_lock);
1734 
1735 	/*
1736 	 * We're done!  Now we just need to go back over the sorted bundles,
1737 	 * calling the function.
1738 	 */
1739 	data = alloca((naggvars + 1) * sizeof (dtrace_aggdata_t *));
1740 
1741 	for (i = 0; i < nbundles; i++) {
1742 		for (j = 0; j < naggvars; j++)
1743 			data[j + 1] = NULL;
1744 
1745 		for (j = 0; j < naggvars; j++) {
1746 			int ndx = j - sortpos;
1747 
1748 			if (ndx < 0)
1749 				ndx += naggvars;
1750 
1751 			assert(bundle[i][ndx] != NULL);
1752 			data[j + 1] = &bundle[i][ndx]->dtahe_data;
1753 		}
1754 
1755 		for (j = 0; j < naggvars; j++)
1756 			assert(data[j + 1] != NULL);
1757 
1758 		/*
1759 		 * The representative key is the last element in the bundle.
1760 		 * Assert that we have one, and then set it to be the first
1761 		 * element of data.
1762 		 */
1763 		assert(bundle[i][j] != NULL);
1764 		data[0] = &bundle[i][j]->dtahe_data;
1765 
1766 		if ((rval = func(data, naggvars + 1, arg)) == -1)
1767 			goto out;
1768 	}
1769 
1770 	rval = 0;
1771 out:
1772 	for (i = 0; i < nbundles; i++)
1773 		dt_free(dtp, bundle[i]);
1774 
1775 	if (zaggdata != NULL) {
1776 		for (i = 0; i < naggvars; i++)
1777 			dt_free(dtp, zaggdata[i].dtahe_data.dtada_data);
1778 	}
1779 
1780 	dt_free(dtp, zaggdata);
1781 	dt_free(dtp, sorted);
1782 	dt_free(dtp, remap);
1783 	dt_free(dtp, map);
1784 
1785 	return (rval);
1786 }
1787 
1788 int
1789 dtrace_aggregate_print(dtrace_hdl_t *dtp, FILE *fp,
1790     dtrace_aggregate_walk_f *func)
1791 {
1792 	dt_print_aggdata_t pd;
1793 
1794 	pd.dtpa_dtp = dtp;
1795 	pd.dtpa_fp = fp;
1796 	pd.dtpa_allunprint = 1;
1797 
1798 	if (func == NULL)
1799 		func = dtrace_aggregate_walk_sorted;
1800 
1801 	if ((*func)(dtp, dt_print_agg, &pd) == -1)
1802 		return (dt_set_errno(dtp, dtp->dt_errno));
1803 
1804 	return (0);
1805 }
1806 
1807 void
1808 dtrace_aggregate_clear(dtrace_hdl_t *dtp)
1809 {
1810 	dt_aggregate_t *agp = &dtp->dt_aggregate;
1811 	dt_ahash_t *hash = &agp->dtat_hash;
1812 	dt_ahashent_t *h;
1813 	dtrace_aggdata_t *data;
1814 	dtrace_aggdesc_t *aggdesc;
1815 	dtrace_recdesc_t *rec;
1816 	int i, max_cpus = agp->dtat_maxcpu;
1817 
1818 	for (h = hash->dtah_all; h != NULL; h = h->dtahe_nextall) {
1819 		aggdesc = h->dtahe_data.dtada_desc;
1820 		rec = &aggdesc->dtagd_rec[aggdesc->dtagd_nrecs - 1];
1821 		data = &h->dtahe_data;
1822 
1823 		bzero(&data->dtada_data[rec->dtrd_offset], rec->dtrd_size);
1824 
1825 		if (data->dtada_percpu == NULL)
1826 			continue;
1827 
1828 		for (i = 0; i < max_cpus; i++)
1829 			bzero(data->dtada_percpu[i], rec->dtrd_size);
1830 	}
1831 }
1832 
1833 void
1834 dt_aggregate_destroy(dtrace_hdl_t *dtp)
1835 {
1836 	dt_aggregate_t *agp = &dtp->dt_aggregate;
1837 	dt_ahash_t *hash = &agp->dtat_hash;
1838 	dt_ahashent_t *h, *next;
1839 	dtrace_aggdata_t *aggdata;
1840 	int i, max_cpus = agp->dtat_maxcpu;
1841 
1842 	if (hash->dtah_hash == NULL) {
1843 		assert(hash->dtah_all == NULL);
1844 	} else {
1845 		free(hash->dtah_hash);
1846 
1847 		for (h = hash->dtah_all; h != NULL; h = next) {
1848 			next = h->dtahe_nextall;
1849 
1850 			aggdata = &h->dtahe_data;
1851 
1852 			if (aggdata->dtada_percpu != NULL) {
1853 				for (i = 0; i < max_cpus; i++)
1854 					free(aggdata->dtada_percpu[i]);
1855 				free(aggdata->dtada_percpu);
1856 			}
1857 
1858 			free(aggdata->dtada_data);
1859 			free(h);
1860 		}
1861 
1862 		hash->dtah_hash = NULL;
1863 		hash->dtah_all = NULL;
1864 		hash->dtah_size = 0;
1865 	}
1866 
1867 	free(agp->dtat_buf.dtbd_data);
1868 	free(agp->dtat_cpus);
1869 }
1870