xref: /freebsd/cddl/contrib/opensolaris/lib/libdtrace/common/dt_aggregate.c (revision 7d8f797b725e3efc0a4256554654780df83c456c)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 /*
28  * Copyright (c) 2013, Joyent, Inc. All rights reserved.
29  * Copyright (c) 2012 by Delphix. All rights reserved.
30  */
31 
32 #include <stdlib.h>
33 #include <strings.h>
34 #include <errno.h>
35 #include <unistd.h>
36 #include <dt_impl.h>
37 #include <assert.h>
38 #ifdef illumos
39 #include <alloca.h>
40 #else
41 #include <sys/sysctl.h>
42 #include <libproc_compat.h>
43 #endif
44 #include <limits.h>
45 
46 #define	DTRACE_AHASHSIZE	32779		/* big 'ol prime */
47 
48 /*
49  * Because qsort(3C) does not allow an argument to be passed to a comparison
50  * function, the variables that affect comparison must regrettably be global;
51  * they are protected by a global static lock, dt_qsort_lock.
52  */
53 static pthread_mutex_t dt_qsort_lock = PTHREAD_MUTEX_INITIALIZER;
54 
55 static int dt_revsort;
56 static int dt_keysort;
57 static int dt_keypos;
58 
59 #define	DT_LESSTHAN	(dt_revsort == 0 ? -1 : 1)
60 #define	DT_GREATERTHAN	(dt_revsort == 0 ? 1 : -1)
61 
62 static void
63 dt_aggregate_count(int64_t *existing, int64_t *new, size_t size)
64 {
65 	uint_t i;
66 
67 	for (i = 0; i < size / sizeof (int64_t); i++)
68 		existing[i] = existing[i] + new[i];
69 }
70 
71 static int
72 dt_aggregate_countcmp(int64_t *lhs, int64_t *rhs)
73 {
74 	int64_t lvar = *lhs;
75 	int64_t rvar = *rhs;
76 
77 	if (lvar < rvar)
78 		return (DT_LESSTHAN);
79 
80 	if (lvar > rvar)
81 		return (DT_GREATERTHAN);
82 
83 	return (0);
84 }
85 
86 /*ARGSUSED*/
87 static void
88 dt_aggregate_min(int64_t *existing, int64_t *new, size_t size)
89 {
90 	if (*new < *existing)
91 		*existing = *new;
92 }
93 
94 /*ARGSUSED*/
95 static void
96 dt_aggregate_max(int64_t *existing, int64_t *new, size_t size)
97 {
98 	if (*new > *existing)
99 		*existing = *new;
100 }
101 
102 static int
103 dt_aggregate_averagecmp(int64_t *lhs, int64_t *rhs)
104 {
105 	int64_t lavg = lhs[0] ? (lhs[1] / lhs[0]) : 0;
106 	int64_t ravg = rhs[0] ? (rhs[1] / rhs[0]) : 0;
107 
108 	if (lavg < ravg)
109 		return (DT_LESSTHAN);
110 
111 	if (lavg > ravg)
112 		return (DT_GREATERTHAN);
113 
114 	return (0);
115 }
116 
117 static int
118 dt_aggregate_stddevcmp(int64_t *lhs, int64_t *rhs)
119 {
120 	uint64_t lsd = dt_stddev((uint64_t *)lhs, 1);
121 	uint64_t rsd = dt_stddev((uint64_t *)rhs, 1);
122 
123 	if (lsd < rsd)
124 		return (DT_LESSTHAN);
125 
126 	if (lsd > rsd)
127 		return (DT_GREATERTHAN);
128 
129 	return (0);
130 }
131 
132 /*ARGSUSED*/
133 static void
134 dt_aggregate_lquantize(int64_t *existing, int64_t *new, size_t size)
135 {
136 	int64_t arg = *existing++;
137 	uint16_t levels = DTRACE_LQUANTIZE_LEVELS(arg);
138 	int i;
139 
140 	for (i = 0; i <= levels + 1; i++)
141 		existing[i] = existing[i] + new[i + 1];
142 }
143 
144 static long double
145 dt_aggregate_lquantizedsum(int64_t *lquanta)
146 {
147 	int64_t arg = *lquanta++;
148 	int32_t base = DTRACE_LQUANTIZE_BASE(arg);
149 	uint16_t step = DTRACE_LQUANTIZE_STEP(arg);
150 	uint16_t levels = DTRACE_LQUANTIZE_LEVELS(arg), i;
151 	long double total = (long double)lquanta[0] * (long double)(base - 1);
152 
153 	for (i = 0; i < levels; base += step, i++)
154 		total += (long double)lquanta[i + 1] * (long double)base;
155 
156 	return (total + (long double)lquanta[levels + 1] *
157 	    (long double)(base + 1));
158 }
159 
160 static int64_t
161 dt_aggregate_lquantizedzero(int64_t *lquanta)
162 {
163 	int64_t arg = *lquanta++;
164 	int32_t base = DTRACE_LQUANTIZE_BASE(arg);
165 	uint16_t step = DTRACE_LQUANTIZE_STEP(arg);
166 	uint16_t levels = DTRACE_LQUANTIZE_LEVELS(arg), i;
167 
168 	if (base - 1 == 0)
169 		return (lquanta[0]);
170 
171 	for (i = 0; i < levels; base += step, i++) {
172 		if (base != 0)
173 			continue;
174 
175 		return (lquanta[i + 1]);
176 	}
177 
178 	if (base + 1 == 0)
179 		return (lquanta[levels + 1]);
180 
181 	return (0);
182 }
183 
184 static int
185 dt_aggregate_lquantizedcmp(int64_t *lhs, int64_t *rhs)
186 {
187 	long double lsum = dt_aggregate_lquantizedsum(lhs);
188 	long double rsum = dt_aggregate_lquantizedsum(rhs);
189 	int64_t lzero, rzero;
190 
191 	if (lsum < rsum)
192 		return (DT_LESSTHAN);
193 
194 	if (lsum > rsum)
195 		return (DT_GREATERTHAN);
196 
197 	/*
198 	 * If they're both equal, then we will compare based on the weights at
199 	 * zero.  If the weights at zero are equal (or if zero is not within
200 	 * the range of the linear quantization), then this will be judged a
201 	 * tie and will be resolved based on the key comparison.
202 	 */
203 	lzero = dt_aggregate_lquantizedzero(lhs);
204 	rzero = dt_aggregate_lquantizedzero(rhs);
205 
206 	if (lzero < rzero)
207 		return (DT_LESSTHAN);
208 
209 	if (lzero > rzero)
210 		return (DT_GREATERTHAN);
211 
212 	return (0);
213 }
214 
215 static void
216 dt_aggregate_llquantize(int64_t *existing, int64_t *new, size_t size)
217 {
218 	int i;
219 
220 	for (i = 1; i < size / sizeof (int64_t); i++)
221 		existing[i] = existing[i] + new[i];
222 }
223 
224 static long double
225 dt_aggregate_llquantizedsum(int64_t *llquanta)
226 {
227 	int64_t arg = *llquanta++;
228 	uint16_t factor = DTRACE_LLQUANTIZE_FACTOR(arg);
229 	uint16_t low = DTRACE_LLQUANTIZE_LOW(arg);
230 	uint16_t high = DTRACE_LLQUANTIZE_HIGH(arg);
231 	uint16_t nsteps = DTRACE_LLQUANTIZE_NSTEP(arg);
232 	int bin = 0, order;
233 	int64_t value = 1, next, step;
234 	long double total;
235 
236 	assert(nsteps >= factor);
237 	assert(nsteps % factor == 0);
238 
239 	for (order = 0; order < low; order++)
240 		value *= factor;
241 
242 	total = (long double)llquanta[bin++] * (long double)(value - 1);
243 
244 	next = value * factor;
245 	step = next > nsteps ? next / nsteps : 1;
246 
247 	while (order <= high) {
248 		assert(value < next);
249 		total += (long double)llquanta[bin++] * (long double)(value);
250 
251 		if ((value += step) != next)
252 			continue;
253 
254 		next = value * factor;
255 		step = next > nsteps ? next / nsteps : 1;
256 		order++;
257 	}
258 
259 	return (total + (long double)llquanta[bin] * (long double)value);
260 }
261 
262 static int
263 dt_aggregate_llquantizedcmp(int64_t *lhs, int64_t *rhs)
264 {
265 	long double lsum = dt_aggregate_llquantizedsum(lhs);
266 	long double rsum = dt_aggregate_llquantizedsum(rhs);
267 	int64_t lzero, rzero;
268 
269 	if (lsum < rsum)
270 		return (DT_LESSTHAN);
271 
272 	if (lsum > rsum)
273 		return (DT_GREATERTHAN);
274 
275 	/*
276 	 * If they're both equal, then we will compare based on the weights at
277 	 * zero.  If the weights at zero are equal, then this will be judged a
278 	 * tie and will be resolved based on the key comparison.
279 	 */
280 	lzero = lhs[1];
281 	rzero = rhs[1];
282 
283 	if (lzero < rzero)
284 		return (DT_LESSTHAN);
285 
286 	if (lzero > rzero)
287 		return (DT_GREATERTHAN);
288 
289 	return (0);
290 }
291 
292 static int
293 dt_aggregate_quantizedcmp(int64_t *lhs, int64_t *rhs)
294 {
295 	int nbuckets = DTRACE_QUANTIZE_NBUCKETS;
296 	long double ltotal = 0, rtotal = 0;
297 	int64_t lzero, rzero;
298 	uint_t i;
299 
300 	for (i = 0; i < nbuckets; i++) {
301 		int64_t bucketval = DTRACE_QUANTIZE_BUCKETVAL(i);
302 
303 		if (bucketval == 0) {
304 			lzero = lhs[i];
305 			rzero = rhs[i];
306 		}
307 
308 		ltotal += (long double)bucketval * (long double)lhs[i];
309 		rtotal += (long double)bucketval * (long double)rhs[i];
310 	}
311 
312 	if (ltotal < rtotal)
313 		return (DT_LESSTHAN);
314 
315 	if (ltotal > rtotal)
316 		return (DT_GREATERTHAN);
317 
318 	/*
319 	 * If they're both equal, then we will compare based on the weights at
320 	 * zero.  If the weights at zero are equal, then this will be judged a
321 	 * tie and will be resolved based on the key comparison.
322 	 */
323 	if (lzero < rzero)
324 		return (DT_LESSTHAN);
325 
326 	if (lzero > rzero)
327 		return (DT_GREATERTHAN);
328 
329 	return (0);
330 }
331 
332 static void
333 dt_aggregate_usym(dtrace_hdl_t *dtp, uint64_t *data)
334 {
335 	uint64_t pid = data[0];
336 	uint64_t *pc = &data[1];
337 	struct ps_prochandle *P;
338 	GElf_Sym sym;
339 
340 	if (dtp->dt_vector != NULL)
341 		return;
342 
343 	if ((P = dt_proc_grab(dtp, pid, PGRAB_RDONLY | PGRAB_FORCE, 0)) == NULL)
344 		return;
345 
346 	dt_proc_lock(dtp, P);
347 
348 	if (Plookup_by_addr(P, *pc, NULL, 0, &sym) == 0)
349 		*pc = sym.st_value;
350 
351 	dt_proc_unlock(dtp, P);
352 	dt_proc_release(dtp, P);
353 }
354 
355 static void
356 dt_aggregate_umod(dtrace_hdl_t *dtp, uint64_t *data)
357 {
358 	uint64_t pid = data[0];
359 	uint64_t *pc = &data[1];
360 	struct ps_prochandle *P;
361 	const prmap_t *map;
362 
363 	if (dtp->dt_vector != NULL)
364 		return;
365 
366 	if ((P = dt_proc_grab(dtp, pid, PGRAB_RDONLY | PGRAB_FORCE, 0)) == NULL)
367 		return;
368 
369 	dt_proc_lock(dtp, P);
370 
371 	if ((map = Paddr_to_map(P, *pc)) != NULL)
372 		*pc = map->pr_vaddr;
373 
374 	dt_proc_unlock(dtp, P);
375 	dt_proc_release(dtp, P);
376 }
377 
378 static void
379 dt_aggregate_sym(dtrace_hdl_t *dtp, uint64_t *data)
380 {
381 	GElf_Sym sym;
382 	uint64_t *pc = data;
383 
384 	if (dtrace_lookup_by_addr(dtp, *pc, &sym, NULL) == 0)
385 		*pc = sym.st_value;
386 }
387 
388 static void
389 dt_aggregate_mod(dtrace_hdl_t *dtp, uint64_t *data)
390 {
391 	uint64_t *pc = data;
392 	dt_module_t *dmp;
393 
394 	if (dtp->dt_vector != NULL) {
395 		/*
396 		 * We don't have a way of just getting the module for a
397 		 * vectored open, and it doesn't seem to be worth defining
398 		 * one.  This means that use of mod() won't get true
399 		 * aggregation in the postmortem case (some modules may
400 		 * appear more than once in aggregation output).  It seems
401 		 * unlikely that anyone will ever notice or care...
402 		 */
403 		return;
404 	}
405 
406 	for (dmp = dt_list_next(&dtp->dt_modlist); dmp != NULL;
407 	    dmp = dt_list_next(dmp)) {
408 		if (*pc - dmp->dm_text_va < dmp->dm_text_size) {
409 			*pc = dmp->dm_text_va;
410 			return;
411 		}
412 	}
413 }
414 
415 static dtrace_aggvarid_t
416 dt_aggregate_aggvarid(dt_ahashent_t *ent)
417 {
418 	dtrace_aggdesc_t *agg = ent->dtahe_data.dtada_desc;
419 	caddr_t data = ent->dtahe_data.dtada_data;
420 	dtrace_recdesc_t *rec = agg->dtagd_rec;
421 
422 	/*
423 	 * First, we'll check the variable ID in the aggdesc.  If it's valid,
424 	 * we'll return it.  If not, we'll use the compiler-generated ID
425 	 * present as the first record.
426 	 */
427 	if (agg->dtagd_varid != DTRACE_AGGVARIDNONE)
428 		return (agg->dtagd_varid);
429 
430 	agg->dtagd_varid = *((dtrace_aggvarid_t *)(uintptr_t)(data +
431 	    rec->dtrd_offset));
432 
433 	return (agg->dtagd_varid);
434 }
435 
436 
437 static int
438 dt_aggregate_snap_cpu(dtrace_hdl_t *dtp, processorid_t cpu)
439 {
440 	dtrace_epid_t id;
441 	uint64_t hashval;
442 	size_t offs, roffs, size, ndx;
443 	int i, j, rval;
444 	caddr_t addr, data;
445 	dtrace_recdesc_t *rec;
446 	dt_aggregate_t *agp = &dtp->dt_aggregate;
447 	dtrace_aggdesc_t *agg;
448 	dt_ahash_t *hash = &agp->dtat_hash;
449 	dt_ahashent_t *h;
450 	dtrace_bufdesc_t b = agp->dtat_buf, *buf = &b;
451 	dtrace_aggdata_t *aggdata;
452 	int flags = agp->dtat_flags;
453 
454 	buf->dtbd_cpu = cpu;
455 
456 #ifdef illumos
457 	if (dt_ioctl(dtp, DTRACEIOC_AGGSNAP, buf) == -1) {
458 #else
459 	if (dt_ioctl(dtp, DTRACEIOC_AGGSNAP, &buf) == -1) {
460 #endif
461 		if (errno == ENOENT) {
462 			/*
463 			 * If that failed with ENOENT, it may be because the
464 			 * CPU was unconfigured.  This is okay; we'll just
465 			 * do nothing but return success.
466 			 */
467 			return (0);
468 		}
469 
470 		return (dt_set_errno(dtp, errno));
471 	}
472 
473 	if (buf->dtbd_drops != 0) {
474 		if (dt_handle_cpudrop(dtp, cpu,
475 		    DTRACEDROP_AGGREGATION, buf->dtbd_drops) == -1)
476 			return (-1);
477 	}
478 
479 	if (buf->dtbd_size == 0)
480 		return (0);
481 
482 	if (hash->dtah_hash == NULL) {
483 		size_t size;
484 
485 		hash->dtah_size = DTRACE_AHASHSIZE;
486 		size = hash->dtah_size * sizeof (dt_ahashent_t *);
487 
488 		if ((hash->dtah_hash = malloc(size)) == NULL)
489 			return (dt_set_errno(dtp, EDT_NOMEM));
490 
491 		bzero(hash->dtah_hash, size);
492 	}
493 
494 	for (offs = 0; offs < buf->dtbd_size; ) {
495 		/*
496 		 * We're guaranteed to have an ID.
497 		 */
498 		id = *((dtrace_epid_t *)((uintptr_t)buf->dtbd_data +
499 		    (uintptr_t)offs));
500 
501 		if (id == DTRACE_AGGIDNONE) {
502 			/*
503 			 * This is filler to assure proper alignment of the
504 			 * next record; we simply ignore it.
505 			 */
506 			offs += sizeof (id);
507 			continue;
508 		}
509 
510 		if ((rval = dt_aggid_lookup(dtp, id, &agg)) != 0)
511 			return (rval);
512 
513 		addr = buf->dtbd_data + offs;
514 		size = agg->dtagd_size;
515 		hashval = 0;
516 
517 		for (j = 0; j < agg->dtagd_nrecs - 1; j++) {
518 			rec = &agg->dtagd_rec[j];
519 			roffs = rec->dtrd_offset;
520 
521 			switch (rec->dtrd_action) {
522 			case DTRACEACT_USYM:
523 				dt_aggregate_usym(dtp,
524 				    /* LINTED - alignment */
525 				    (uint64_t *)&addr[roffs]);
526 				break;
527 
528 			case DTRACEACT_UMOD:
529 				dt_aggregate_umod(dtp,
530 				    /* LINTED - alignment */
531 				    (uint64_t *)&addr[roffs]);
532 				break;
533 
534 			case DTRACEACT_SYM:
535 				/* LINTED - alignment */
536 				dt_aggregate_sym(dtp, (uint64_t *)&addr[roffs]);
537 				break;
538 
539 			case DTRACEACT_MOD:
540 				/* LINTED - alignment */
541 				dt_aggregate_mod(dtp, (uint64_t *)&addr[roffs]);
542 				break;
543 
544 			default:
545 				break;
546 			}
547 
548 			for (i = 0; i < rec->dtrd_size; i++)
549 				hashval += addr[roffs + i];
550 		}
551 
552 		ndx = hashval % hash->dtah_size;
553 
554 		for (h = hash->dtah_hash[ndx]; h != NULL; h = h->dtahe_next) {
555 			if (h->dtahe_hashval != hashval)
556 				continue;
557 
558 			if (h->dtahe_size != size)
559 				continue;
560 
561 			aggdata = &h->dtahe_data;
562 			data = aggdata->dtada_data;
563 
564 			for (j = 0; j < agg->dtagd_nrecs - 1; j++) {
565 				rec = &agg->dtagd_rec[j];
566 				roffs = rec->dtrd_offset;
567 
568 				for (i = 0; i < rec->dtrd_size; i++)
569 					if (addr[roffs + i] != data[roffs + i])
570 						goto hashnext;
571 			}
572 
573 			/*
574 			 * We found it.  Now we need to apply the aggregating
575 			 * action on the data here.
576 			 */
577 			rec = &agg->dtagd_rec[agg->dtagd_nrecs - 1];
578 			roffs = rec->dtrd_offset;
579 			/* LINTED - alignment */
580 			h->dtahe_aggregate((int64_t *)&data[roffs],
581 			    /* LINTED - alignment */
582 			    (int64_t *)&addr[roffs], rec->dtrd_size);
583 
584 			/*
585 			 * If we're keeping per CPU data, apply the aggregating
586 			 * action there as well.
587 			 */
588 			if (aggdata->dtada_percpu != NULL) {
589 				data = aggdata->dtada_percpu[cpu];
590 
591 				/* LINTED - alignment */
592 				h->dtahe_aggregate((int64_t *)data,
593 				    /* LINTED - alignment */
594 				    (int64_t *)&addr[roffs], rec->dtrd_size);
595 			}
596 
597 			goto bufnext;
598 hashnext:
599 			continue;
600 		}
601 
602 		/*
603 		 * If we're here, we couldn't find an entry for this record.
604 		 */
605 		if ((h = malloc(sizeof (dt_ahashent_t))) == NULL)
606 			return (dt_set_errno(dtp, EDT_NOMEM));
607 		bzero(h, sizeof (dt_ahashent_t));
608 		aggdata = &h->dtahe_data;
609 
610 		if ((aggdata->dtada_data = malloc(size)) == NULL) {
611 			free(h);
612 			return (dt_set_errno(dtp, EDT_NOMEM));
613 		}
614 
615 		bcopy(addr, aggdata->dtada_data, size);
616 		aggdata->dtada_size = size;
617 		aggdata->dtada_desc = agg;
618 		aggdata->dtada_handle = dtp;
619 		(void) dt_epid_lookup(dtp, agg->dtagd_epid,
620 		    &aggdata->dtada_edesc, &aggdata->dtada_pdesc);
621 		aggdata->dtada_normal = 1;
622 
623 		h->dtahe_hashval = hashval;
624 		h->dtahe_size = size;
625 		(void) dt_aggregate_aggvarid(h);
626 
627 		rec = &agg->dtagd_rec[agg->dtagd_nrecs - 1];
628 
629 		if (flags & DTRACE_A_PERCPU) {
630 			int max_cpus = agp->dtat_maxcpu;
631 			caddr_t *percpu = malloc(max_cpus * sizeof (caddr_t));
632 
633 			if (percpu == NULL) {
634 				free(aggdata->dtada_data);
635 				free(h);
636 				return (dt_set_errno(dtp, EDT_NOMEM));
637 			}
638 
639 			for (j = 0; j < max_cpus; j++) {
640 				percpu[j] = malloc(rec->dtrd_size);
641 
642 				if (percpu[j] == NULL) {
643 					while (--j >= 0)
644 						free(percpu[j]);
645 
646 					free(aggdata->dtada_data);
647 					free(h);
648 					return (dt_set_errno(dtp, EDT_NOMEM));
649 				}
650 
651 				if (j == cpu) {
652 					bcopy(&addr[rec->dtrd_offset],
653 					    percpu[j], rec->dtrd_size);
654 				} else {
655 					bzero(percpu[j], rec->dtrd_size);
656 				}
657 			}
658 
659 			aggdata->dtada_percpu = percpu;
660 		}
661 
662 		switch (rec->dtrd_action) {
663 		case DTRACEAGG_MIN:
664 			h->dtahe_aggregate = dt_aggregate_min;
665 			break;
666 
667 		case DTRACEAGG_MAX:
668 			h->dtahe_aggregate = dt_aggregate_max;
669 			break;
670 
671 		case DTRACEAGG_LQUANTIZE:
672 			h->dtahe_aggregate = dt_aggregate_lquantize;
673 			break;
674 
675 		case DTRACEAGG_LLQUANTIZE:
676 			h->dtahe_aggregate = dt_aggregate_llquantize;
677 			break;
678 
679 		case DTRACEAGG_COUNT:
680 		case DTRACEAGG_SUM:
681 		case DTRACEAGG_AVG:
682 		case DTRACEAGG_STDDEV:
683 		case DTRACEAGG_QUANTIZE:
684 			h->dtahe_aggregate = dt_aggregate_count;
685 			break;
686 
687 		default:
688 			return (dt_set_errno(dtp, EDT_BADAGG));
689 		}
690 
691 		if (hash->dtah_hash[ndx] != NULL)
692 			hash->dtah_hash[ndx]->dtahe_prev = h;
693 
694 		h->dtahe_next = hash->dtah_hash[ndx];
695 		hash->dtah_hash[ndx] = h;
696 
697 		if (hash->dtah_all != NULL)
698 			hash->dtah_all->dtahe_prevall = h;
699 
700 		h->dtahe_nextall = hash->dtah_all;
701 		hash->dtah_all = h;
702 bufnext:
703 		offs += agg->dtagd_size;
704 	}
705 
706 	return (0);
707 }
708 
709 int
710 dtrace_aggregate_snap(dtrace_hdl_t *dtp)
711 {
712 	int i, rval;
713 	dt_aggregate_t *agp = &dtp->dt_aggregate;
714 	hrtime_t now = gethrtime();
715 	dtrace_optval_t interval = dtp->dt_options[DTRACEOPT_AGGRATE];
716 
717 	if (dtp->dt_lastagg != 0) {
718 		if (now - dtp->dt_lastagg < interval)
719 			return (0);
720 
721 		dtp->dt_lastagg += interval;
722 	} else {
723 		dtp->dt_lastagg = now;
724 	}
725 
726 	if (!dtp->dt_active)
727 		return (dt_set_errno(dtp, EINVAL));
728 
729 	if (agp->dtat_buf.dtbd_size == 0)
730 		return (0);
731 
732 	for (i = 0; i < agp->dtat_ncpus; i++) {
733 		if ((rval = dt_aggregate_snap_cpu(dtp, agp->dtat_cpus[i])))
734 			return (rval);
735 	}
736 
737 	return (0);
738 }
739 
740 static int
741 dt_aggregate_hashcmp(const void *lhs, const void *rhs)
742 {
743 	dt_ahashent_t *lh = *((dt_ahashent_t **)lhs);
744 	dt_ahashent_t *rh = *((dt_ahashent_t **)rhs);
745 	dtrace_aggdesc_t *lagg = lh->dtahe_data.dtada_desc;
746 	dtrace_aggdesc_t *ragg = rh->dtahe_data.dtada_desc;
747 
748 	if (lagg->dtagd_nrecs < ragg->dtagd_nrecs)
749 		return (DT_LESSTHAN);
750 
751 	if (lagg->dtagd_nrecs > ragg->dtagd_nrecs)
752 		return (DT_GREATERTHAN);
753 
754 	return (0);
755 }
756 
757 static int
758 dt_aggregate_varcmp(const void *lhs, const void *rhs)
759 {
760 	dt_ahashent_t *lh = *((dt_ahashent_t **)lhs);
761 	dt_ahashent_t *rh = *((dt_ahashent_t **)rhs);
762 	dtrace_aggvarid_t lid, rid;
763 
764 	lid = dt_aggregate_aggvarid(lh);
765 	rid = dt_aggregate_aggvarid(rh);
766 
767 	if (lid < rid)
768 		return (DT_LESSTHAN);
769 
770 	if (lid > rid)
771 		return (DT_GREATERTHAN);
772 
773 	return (0);
774 }
775 
776 static int
777 dt_aggregate_keycmp(const void *lhs, const void *rhs)
778 {
779 	dt_ahashent_t *lh = *((dt_ahashent_t **)lhs);
780 	dt_ahashent_t *rh = *((dt_ahashent_t **)rhs);
781 	dtrace_aggdesc_t *lagg = lh->dtahe_data.dtada_desc;
782 	dtrace_aggdesc_t *ragg = rh->dtahe_data.dtada_desc;
783 	dtrace_recdesc_t *lrec, *rrec;
784 	char *ldata, *rdata;
785 	int rval, i, j, keypos, nrecs;
786 
787 	if ((rval = dt_aggregate_hashcmp(lhs, rhs)) != 0)
788 		return (rval);
789 
790 	nrecs = lagg->dtagd_nrecs - 1;
791 	assert(nrecs == ragg->dtagd_nrecs - 1);
792 
793 	keypos = dt_keypos + 1 >= nrecs ? 0 : dt_keypos;
794 
795 	for (i = 1; i < nrecs; i++) {
796 		uint64_t lval, rval;
797 		int ndx = i + keypos;
798 
799 		if (ndx >= nrecs)
800 			ndx = ndx - nrecs + 1;
801 
802 		lrec = &lagg->dtagd_rec[ndx];
803 		rrec = &ragg->dtagd_rec[ndx];
804 
805 		ldata = lh->dtahe_data.dtada_data + lrec->dtrd_offset;
806 		rdata = rh->dtahe_data.dtada_data + rrec->dtrd_offset;
807 
808 		if (lrec->dtrd_size < rrec->dtrd_size)
809 			return (DT_LESSTHAN);
810 
811 		if (lrec->dtrd_size > rrec->dtrd_size)
812 			return (DT_GREATERTHAN);
813 
814 		switch (lrec->dtrd_size) {
815 		case sizeof (uint64_t):
816 			/* LINTED - alignment */
817 			lval = *((uint64_t *)ldata);
818 			/* LINTED - alignment */
819 			rval = *((uint64_t *)rdata);
820 			break;
821 
822 		case sizeof (uint32_t):
823 			/* LINTED - alignment */
824 			lval = *((uint32_t *)ldata);
825 			/* LINTED - alignment */
826 			rval = *((uint32_t *)rdata);
827 			break;
828 
829 		case sizeof (uint16_t):
830 			/* LINTED - alignment */
831 			lval = *((uint16_t *)ldata);
832 			/* LINTED - alignment */
833 			rval = *((uint16_t *)rdata);
834 			break;
835 
836 		case sizeof (uint8_t):
837 			lval = *((uint8_t *)ldata);
838 			rval = *((uint8_t *)rdata);
839 			break;
840 
841 		default:
842 			switch (lrec->dtrd_action) {
843 			case DTRACEACT_UMOD:
844 			case DTRACEACT_UADDR:
845 			case DTRACEACT_USYM:
846 				for (j = 0; j < 2; j++) {
847 					/* LINTED - alignment */
848 					lval = ((uint64_t *)ldata)[j];
849 					/* LINTED - alignment */
850 					rval = ((uint64_t *)rdata)[j];
851 
852 					if (lval < rval)
853 						return (DT_LESSTHAN);
854 
855 					if (lval > rval)
856 						return (DT_GREATERTHAN);
857 				}
858 
859 				break;
860 
861 			default:
862 				for (j = 0; j < lrec->dtrd_size; j++) {
863 					lval = ((uint8_t *)ldata)[j];
864 					rval = ((uint8_t *)rdata)[j];
865 
866 					if (lval < rval)
867 						return (DT_LESSTHAN);
868 
869 					if (lval > rval)
870 						return (DT_GREATERTHAN);
871 				}
872 			}
873 
874 			continue;
875 		}
876 
877 		if (lval < rval)
878 			return (DT_LESSTHAN);
879 
880 		if (lval > rval)
881 			return (DT_GREATERTHAN);
882 	}
883 
884 	return (0);
885 }
886 
887 static int
888 dt_aggregate_valcmp(const void *lhs, const void *rhs)
889 {
890 	dt_ahashent_t *lh = *((dt_ahashent_t **)lhs);
891 	dt_ahashent_t *rh = *((dt_ahashent_t **)rhs);
892 	dtrace_aggdesc_t *lagg = lh->dtahe_data.dtada_desc;
893 	dtrace_aggdesc_t *ragg = rh->dtahe_data.dtada_desc;
894 	caddr_t ldata = lh->dtahe_data.dtada_data;
895 	caddr_t rdata = rh->dtahe_data.dtada_data;
896 	dtrace_recdesc_t *lrec, *rrec;
897 	int64_t *laddr, *raddr;
898 	int rval;
899 
900 	assert(lagg->dtagd_nrecs == ragg->dtagd_nrecs);
901 
902 	lrec = &lagg->dtagd_rec[lagg->dtagd_nrecs - 1];
903 	rrec = &ragg->dtagd_rec[ragg->dtagd_nrecs - 1];
904 
905 	assert(lrec->dtrd_action == rrec->dtrd_action);
906 
907 	laddr = (int64_t *)(uintptr_t)(ldata + lrec->dtrd_offset);
908 	raddr = (int64_t *)(uintptr_t)(rdata + rrec->dtrd_offset);
909 
910 	switch (lrec->dtrd_action) {
911 	case DTRACEAGG_AVG:
912 		rval = dt_aggregate_averagecmp(laddr, raddr);
913 		break;
914 
915 	case DTRACEAGG_STDDEV:
916 		rval = dt_aggregate_stddevcmp(laddr, raddr);
917 		break;
918 
919 	case DTRACEAGG_QUANTIZE:
920 		rval = dt_aggregate_quantizedcmp(laddr, raddr);
921 		break;
922 
923 	case DTRACEAGG_LQUANTIZE:
924 		rval = dt_aggregate_lquantizedcmp(laddr, raddr);
925 		break;
926 
927 	case DTRACEAGG_LLQUANTIZE:
928 		rval = dt_aggregate_llquantizedcmp(laddr, raddr);
929 		break;
930 
931 	case DTRACEAGG_COUNT:
932 	case DTRACEAGG_SUM:
933 	case DTRACEAGG_MIN:
934 	case DTRACEAGG_MAX:
935 		rval = dt_aggregate_countcmp(laddr, raddr);
936 		break;
937 
938 	default:
939 		assert(0);
940 	}
941 
942 	return (rval);
943 }
944 
945 static int
946 dt_aggregate_valkeycmp(const void *lhs, const void *rhs)
947 {
948 	int rval;
949 
950 	if ((rval = dt_aggregate_valcmp(lhs, rhs)) != 0)
951 		return (rval);
952 
953 	/*
954 	 * If we're here, the values for the two aggregation elements are
955 	 * equal.  We already know that the key layout is the same for the two
956 	 * elements; we must now compare the keys themselves as a tie-breaker.
957 	 */
958 	return (dt_aggregate_keycmp(lhs, rhs));
959 }
960 
961 static int
962 dt_aggregate_keyvarcmp(const void *lhs, const void *rhs)
963 {
964 	int rval;
965 
966 	if ((rval = dt_aggregate_keycmp(lhs, rhs)) != 0)
967 		return (rval);
968 
969 	return (dt_aggregate_varcmp(lhs, rhs));
970 }
971 
972 static int
973 dt_aggregate_varkeycmp(const void *lhs, const void *rhs)
974 {
975 	int rval;
976 
977 	if ((rval = dt_aggregate_varcmp(lhs, rhs)) != 0)
978 		return (rval);
979 
980 	return (dt_aggregate_keycmp(lhs, rhs));
981 }
982 
983 static int
984 dt_aggregate_valvarcmp(const void *lhs, const void *rhs)
985 {
986 	int rval;
987 
988 	if ((rval = dt_aggregate_valkeycmp(lhs, rhs)) != 0)
989 		return (rval);
990 
991 	return (dt_aggregate_varcmp(lhs, rhs));
992 }
993 
994 static int
995 dt_aggregate_varvalcmp(const void *lhs, const void *rhs)
996 {
997 	int rval;
998 
999 	if ((rval = dt_aggregate_varcmp(lhs, rhs)) != 0)
1000 		return (rval);
1001 
1002 	return (dt_aggregate_valkeycmp(lhs, rhs));
1003 }
1004 
1005 static int
1006 dt_aggregate_keyvarrevcmp(const void *lhs, const void *rhs)
1007 {
1008 	return (dt_aggregate_keyvarcmp(rhs, lhs));
1009 }
1010 
1011 static int
1012 dt_aggregate_varkeyrevcmp(const void *lhs, const void *rhs)
1013 {
1014 	return (dt_aggregate_varkeycmp(rhs, lhs));
1015 }
1016 
1017 static int
1018 dt_aggregate_valvarrevcmp(const void *lhs, const void *rhs)
1019 {
1020 	return (dt_aggregate_valvarcmp(rhs, lhs));
1021 }
1022 
1023 static int
1024 dt_aggregate_varvalrevcmp(const void *lhs, const void *rhs)
1025 {
1026 	return (dt_aggregate_varvalcmp(rhs, lhs));
1027 }
1028 
1029 static int
1030 dt_aggregate_bundlecmp(const void *lhs, const void *rhs)
1031 {
1032 	dt_ahashent_t **lh = *((dt_ahashent_t ***)lhs);
1033 	dt_ahashent_t **rh = *((dt_ahashent_t ***)rhs);
1034 	int i, rval;
1035 
1036 	if (dt_keysort) {
1037 		/*
1038 		 * If we're sorting on keys, we need to scan until we find the
1039 		 * last entry -- that's the representative key.  (The order of
1040 		 * the bundle is values followed by key to accommodate the
1041 		 * default behavior of sorting by value.)  If the keys are
1042 		 * equal, we'll fall into the value comparison loop, below.
1043 		 */
1044 		for (i = 0; lh[i + 1] != NULL; i++)
1045 			continue;
1046 
1047 		assert(i != 0);
1048 		assert(rh[i + 1] == NULL);
1049 
1050 		if ((rval = dt_aggregate_keycmp(&lh[i], &rh[i])) != 0)
1051 			return (rval);
1052 	}
1053 
1054 	for (i = 0; ; i++) {
1055 		if (lh[i + 1] == NULL) {
1056 			/*
1057 			 * All of the values are equal; if we're sorting on
1058 			 * keys, then we're only here because the keys were
1059 			 * found to be equal and these records are therefore
1060 			 * equal.  If we're not sorting on keys, we'll use the
1061 			 * key comparison from the representative key as the
1062 			 * tie-breaker.
1063 			 */
1064 			if (dt_keysort)
1065 				return (0);
1066 
1067 			assert(i != 0);
1068 			assert(rh[i + 1] == NULL);
1069 			return (dt_aggregate_keycmp(&lh[i], &rh[i]));
1070 		} else {
1071 			if ((rval = dt_aggregate_valcmp(&lh[i], &rh[i])) != 0)
1072 				return (rval);
1073 		}
1074 	}
1075 }
1076 
1077 int
1078 dt_aggregate_go(dtrace_hdl_t *dtp)
1079 {
1080 	dt_aggregate_t *agp = &dtp->dt_aggregate;
1081 	dtrace_optval_t size, cpu;
1082 	dtrace_bufdesc_t *buf = &agp->dtat_buf;
1083 	int rval, i;
1084 
1085 	assert(agp->dtat_maxcpu == 0);
1086 	assert(agp->dtat_ncpu == 0);
1087 	assert(agp->dtat_cpus == NULL);
1088 
1089 	agp->dtat_maxcpu = dt_sysconf(dtp, _SC_CPUID_MAX) + 1;
1090 	agp->dtat_ncpu = dt_sysconf(dtp, _SC_NPROCESSORS_MAX);
1091 	agp->dtat_cpus = malloc(agp->dtat_ncpu * sizeof (processorid_t));
1092 
1093 	if (agp->dtat_cpus == NULL)
1094 		return (dt_set_errno(dtp, EDT_NOMEM));
1095 
1096 	/*
1097 	 * Use the aggregation buffer size as reloaded from the kernel.
1098 	 */
1099 	size = dtp->dt_options[DTRACEOPT_AGGSIZE];
1100 
1101 	rval = dtrace_getopt(dtp, "aggsize", &size);
1102 	assert(rval == 0);
1103 
1104 	if (size == 0 || size == DTRACEOPT_UNSET)
1105 		return (0);
1106 
1107 	buf = &agp->dtat_buf;
1108 	buf->dtbd_size = size;
1109 
1110 	if ((buf->dtbd_data = malloc(buf->dtbd_size)) == NULL)
1111 		return (dt_set_errno(dtp, EDT_NOMEM));
1112 
1113 	/*
1114 	 * Now query for the CPUs enabled.
1115 	 */
1116 	rval = dtrace_getopt(dtp, "cpu", &cpu);
1117 	assert(rval == 0 && cpu != DTRACEOPT_UNSET);
1118 
1119 	if (cpu != DTRACE_CPUALL) {
1120 		assert(cpu < agp->dtat_ncpu);
1121 		agp->dtat_cpus[agp->dtat_ncpus++] = (processorid_t)cpu;
1122 
1123 		return (0);
1124 	}
1125 
1126 	agp->dtat_ncpus = 0;
1127 	for (i = 0; i < agp->dtat_maxcpu; i++) {
1128 		if (dt_status(dtp, i) == -1)
1129 			continue;
1130 
1131 		agp->dtat_cpus[agp->dtat_ncpus++] = i;
1132 	}
1133 
1134 	return (0);
1135 }
1136 
1137 static int
1138 dt_aggwalk_rval(dtrace_hdl_t *dtp, dt_ahashent_t *h, int rval)
1139 {
1140 	dt_aggregate_t *agp = &dtp->dt_aggregate;
1141 	dtrace_aggdata_t *data;
1142 	dtrace_aggdesc_t *aggdesc;
1143 	dtrace_recdesc_t *rec;
1144 	int i;
1145 
1146 	switch (rval) {
1147 	case DTRACE_AGGWALK_NEXT:
1148 		break;
1149 
1150 	case DTRACE_AGGWALK_CLEAR: {
1151 		uint32_t size, offs = 0;
1152 
1153 		aggdesc = h->dtahe_data.dtada_desc;
1154 		rec = &aggdesc->dtagd_rec[aggdesc->dtagd_nrecs - 1];
1155 		size = rec->dtrd_size;
1156 		data = &h->dtahe_data;
1157 
1158 		if (rec->dtrd_action == DTRACEAGG_LQUANTIZE) {
1159 			offs = sizeof (uint64_t);
1160 			size -= sizeof (uint64_t);
1161 		}
1162 
1163 		bzero(&data->dtada_data[rec->dtrd_offset] + offs, size);
1164 
1165 		if (data->dtada_percpu == NULL)
1166 			break;
1167 
1168 		for (i = 0; i < dtp->dt_aggregate.dtat_maxcpu; i++)
1169 			bzero(data->dtada_percpu[i] + offs, size);
1170 		break;
1171 	}
1172 
1173 	case DTRACE_AGGWALK_ERROR:
1174 		/*
1175 		 * We assume that errno is already set in this case.
1176 		 */
1177 		return (dt_set_errno(dtp, errno));
1178 
1179 	case DTRACE_AGGWALK_ABORT:
1180 		return (dt_set_errno(dtp, EDT_DIRABORT));
1181 
1182 	case DTRACE_AGGWALK_DENORMALIZE:
1183 		h->dtahe_data.dtada_normal = 1;
1184 		return (0);
1185 
1186 	case DTRACE_AGGWALK_NORMALIZE:
1187 		if (h->dtahe_data.dtada_normal == 0) {
1188 			h->dtahe_data.dtada_normal = 1;
1189 			return (dt_set_errno(dtp, EDT_BADRVAL));
1190 		}
1191 
1192 		return (0);
1193 
1194 	case DTRACE_AGGWALK_REMOVE: {
1195 		dtrace_aggdata_t *aggdata = &h->dtahe_data;
1196 		int max_cpus = agp->dtat_maxcpu;
1197 
1198 		/*
1199 		 * First, remove this hash entry from its hash chain.
1200 		 */
1201 		if (h->dtahe_prev != NULL) {
1202 			h->dtahe_prev->dtahe_next = h->dtahe_next;
1203 		} else {
1204 			dt_ahash_t *hash = &agp->dtat_hash;
1205 			size_t ndx = h->dtahe_hashval % hash->dtah_size;
1206 
1207 			assert(hash->dtah_hash[ndx] == h);
1208 			hash->dtah_hash[ndx] = h->dtahe_next;
1209 		}
1210 
1211 		if (h->dtahe_next != NULL)
1212 			h->dtahe_next->dtahe_prev = h->dtahe_prev;
1213 
1214 		/*
1215 		 * Now remove it from the list of all hash entries.
1216 		 */
1217 		if (h->dtahe_prevall != NULL) {
1218 			h->dtahe_prevall->dtahe_nextall = h->dtahe_nextall;
1219 		} else {
1220 			dt_ahash_t *hash = &agp->dtat_hash;
1221 
1222 			assert(hash->dtah_all == h);
1223 			hash->dtah_all = h->dtahe_nextall;
1224 		}
1225 
1226 		if (h->dtahe_nextall != NULL)
1227 			h->dtahe_nextall->dtahe_prevall = h->dtahe_prevall;
1228 
1229 		/*
1230 		 * We're unlinked.  We can safely destroy the data.
1231 		 */
1232 		if (aggdata->dtada_percpu != NULL) {
1233 			for (i = 0; i < max_cpus; i++)
1234 				free(aggdata->dtada_percpu[i]);
1235 			free(aggdata->dtada_percpu);
1236 		}
1237 
1238 		free(aggdata->dtada_data);
1239 		free(h);
1240 
1241 		return (0);
1242 	}
1243 
1244 	default:
1245 		return (dt_set_errno(dtp, EDT_BADRVAL));
1246 	}
1247 
1248 	return (0);
1249 }
1250 
1251 void
1252 dt_aggregate_qsort(dtrace_hdl_t *dtp, void *base, size_t nel, size_t width,
1253     int (*compar)(const void *, const void *))
1254 {
1255 	int rev = dt_revsort, key = dt_keysort, keypos = dt_keypos;
1256 	dtrace_optval_t keyposopt = dtp->dt_options[DTRACEOPT_AGGSORTKEYPOS];
1257 
1258 	dt_revsort = (dtp->dt_options[DTRACEOPT_AGGSORTREV] != DTRACEOPT_UNSET);
1259 	dt_keysort = (dtp->dt_options[DTRACEOPT_AGGSORTKEY] != DTRACEOPT_UNSET);
1260 
1261 	if (keyposopt != DTRACEOPT_UNSET && keyposopt <= INT_MAX) {
1262 		dt_keypos = (int)keyposopt;
1263 	} else {
1264 		dt_keypos = 0;
1265 	}
1266 
1267 	if (compar == NULL) {
1268 		if (!dt_keysort) {
1269 			compar = dt_aggregate_varvalcmp;
1270 		} else {
1271 			compar = dt_aggregate_varkeycmp;
1272 		}
1273 	}
1274 
1275 	qsort(base, nel, width, compar);
1276 
1277 	dt_revsort = rev;
1278 	dt_keysort = key;
1279 	dt_keypos = keypos;
1280 }
1281 
1282 int
1283 dtrace_aggregate_walk(dtrace_hdl_t *dtp, dtrace_aggregate_f *func, void *arg)
1284 {
1285 	dt_ahashent_t *h, *next;
1286 	dt_ahash_t *hash = &dtp->dt_aggregate.dtat_hash;
1287 
1288 	for (h = hash->dtah_all; h != NULL; h = next) {
1289 		/*
1290 		 * dt_aggwalk_rval() can potentially remove the current hash
1291 		 * entry; we need to load the next hash entry before calling
1292 		 * into it.
1293 		 */
1294 		next = h->dtahe_nextall;
1295 
1296 		if (dt_aggwalk_rval(dtp, h, func(&h->dtahe_data, arg)) == -1)
1297 			return (-1);
1298 	}
1299 
1300 	return (0);
1301 }
1302 
1303 static int
1304 dt_aggregate_total(dtrace_hdl_t *dtp, boolean_t clear)
1305 {
1306 	dt_ahashent_t *h;
1307 	dtrace_aggdata_t **total;
1308 	dtrace_aggid_t max = DTRACE_AGGVARIDNONE, id;
1309 	dt_aggregate_t *agp = &dtp->dt_aggregate;
1310 	dt_ahash_t *hash = &agp->dtat_hash;
1311 	uint32_t tflags;
1312 
1313 	tflags = DTRACE_A_TOTAL | DTRACE_A_HASNEGATIVES | DTRACE_A_HASPOSITIVES;
1314 
1315 	/*
1316 	 * If we need to deliver per-aggregation totals, we're going to take
1317 	 * three passes over the aggregate:  one to clear everything out and
1318 	 * determine our maximum aggregation ID, one to actually total
1319 	 * everything up, and a final pass to assign the totals to the
1320 	 * individual elements.
1321 	 */
1322 	for (h = hash->dtah_all; h != NULL; h = h->dtahe_nextall) {
1323 		dtrace_aggdata_t *aggdata = &h->dtahe_data;
1324 
1325 		if ((id = dt_aggregate_aggvarid(h)) > max)
1326 			max = id;
1327 
1328 		aggdata->dtada_total = 0;
1329 		aggdata->dtada_flags &= ~tflags;
1330 	}
1331 
1332 	if (clear || max == DTRACE_AGGVARIDNONE)
1333 		return (0);
1334 
1335 	total = dt_zalloc(dtp, (max + 1) * sizeof (dtrace_aggdata_t *));
1336 
1337 	if (total == NULL)
1338 		return (-1);
1339 
1340 	for (h = hash->dtah_all; h != NULL; h = h->dtahe_nextall) {
1341 		dtrace_aggdata_t *aggdata = &h->dtahe_data;
1342 		dtrace_aggdesc_t *agg = aggdata->dtada_desc;
1343 		dtrace_recdesc_t *rec;
1344 		caddr_t data;
1345 		int64_t val, *addr;
1346 
1347 		rec = &agg->dtagd_rec[agg->dtagd_nrecs - 1];
1348 		data = aggdata->dtada_data;
1349 		addr = (int64_t *)(uintptr_t)(data + rec->dtrd_offset);
1350 
1351 		switch (rec->dtrd_action) {
1352 		case DTRACEAGG_STDDEV:
1353 			val = dt_stddev((uint64_t *)addr, 1);
1354 			break;
1355 
1356 		case DTRACEAGG_SUM:
1357 		case DTRACEAGG_COUNT:
1358 			val = *addr;
1359 			break;
1360 
1361 		case DTRACEAGG_AVG:
1362 			val = addr[0] ? (addr[1] / addr[0]) : 0;
1363 			break;
1364 
1365 		default:
1366 			continue;
1367 		}
1368 
1369 		if (total[agg->dtagd_varid] == NULL) {
1370 			total[agg->dtagd_varid] = aggdata;
1371 			aggdata->dtada_flags |= DTRACE_A_TOTAL;
1372 		} else {
1373 			aggdata = total[agg->dtagd_varid];
1374 		}
1375 
1376 		if (val > 0)
1377 			aggdata->dtada_flags |= DTRACE_A_HASPOSITIVES;
1378 
1379 		if (val < 0) {
1380 			aggdata->dtada_flags |= DTRACE_A_HASNEGATIVES;
1381 			val = -val;
1382 		}
1383 
1384 		if (dtp->dt_options[DTRACEOPT_AGGZOOM] != DTRACEOPT_UNSET) {
1385 			val = (int64_t)((long double)val *
1386 			    (1 / DTRACE_AGGZOOM_MAX));
1387 
1388 			if (val > aggdata->dtada_total)
1389 				aggdata->dtada_total = val;
1390 		} else {
1391 			aggdata->dtada_total += val;
1392 		}
1393 	}
1394 
1395 	/*
1396 	 * And now one final pass to set everyone's total.
1397 	 */
1398 	for (h = hash->dtah_all; h != NULL; h = h->dtahe_nextall) {
1399 		dtrace_aggdata_t *aggdata = &h->dtahe_data, *t;
1400 		dtrace_aggdesc_t *agg = aggdata->dtada_desc;
1401 
1402 		if ((t = total[agg->dtagd_varid]) == NULL || aggdata == t)
1403 			continue;
1404 
1405 		aggdata->dtada_total = t->dtada_total;
1406 		aggdata->dtada_flags |= (t->dtada_flags & tflags);
1407 	}
1408 
1409 	dt_free(dtp, total);
1410 
1411 	return (0);
1412 }
1413 
1414 static int
1415 dt_aggregate_minmaxbin(dtrace_hdl_t *dtp, boolean_t clear)
1416 {
1417 	dt_ahashent_t *h;
1418 	dtrace_aggdata_t **minmax;
1419 	dtrace_aggid_t max = DTRACE_AGGVARIDNONE, id;
1420 	dt_aggregate_t *agp = &dtp->dt_aggregate;
1421 	dt_ahash_t *hash = &agp->dtat_hash;
1422 
1423 	for (h = hash->dtah_all; h != NULL; h = h->dtahe_nextall) {
1424 		dtrace_aggdata_t *aggdata = &h->dtahe_data;
1425 
1426 		if ((id = dt_aggregate_aggvarid(h)) > max)
1427 			max = id;
1428 
1429 		aggdata->dtada_minbin = 0;
1430 		aggdata->dtada_maxbin = 0;
1431 		aggdata->dtada_flags &= ~DTRACE_A_MINMAXBIN;
1432 	}
1433 
1434 	if (clear || max == DTRACE_AGGVARIDNONE)
1435 		return (0);
1436 
1437 	minmax = dt_zalloc(dtp, (max + 1) * sizeof (dtrace_aggdata_t *));
1438 
1439 	if (minmax == NULL)
1440 		return (-1);
1441 
1442 	for (h = hash->dtah_all; h != NULL; h = h->dtahe_nextall) {
1443 		dtrace_aggdata_t *aggdata = &h->dtahe_data;
1444 		dtrace_aggdesc_t *agg = aggdata->dtada_desc;
1445 		dtrace_recdesc_t *rec;
1446 		caddr_t data;
1447 		int64_t *addr;
1448 		int minbin = -1, maxbin = -1, i;
1449 		int start = 0, size;
1450 
1451 		rec = &agg->dtagd_rec[agg->dtagd_nrecs - 1];
1452 		size = rec->dtrd_size / sizeof (int64_t);
1453 		data = aggdata->dtada_data;
1454 		addr = (int64_t *)(uintptr_t)(data + rec->dtrd_offset);
1455 
1456 		switch (rec->dtrd_action) {
1457 		case DTRACEAGG_LQUANTIZE:
1458 			/*
1459 			 * For lquantize(), we always display the entire range
1460 			 * of the aggregation when aggpack is set.
1461 			 */
1462 			start = 1;
1463 			minbin = start;
1464 			maxbin = size - 1 - start;
1465 			break;
1466 
1467 		case DTRACEAGG_QUANTIZE:
1468 			for (i = start; i < size; i++) {
1469 				if (!addr[i])
1470 					continue;
1471 
1472 				if (minbin == -1)
1473 					minbin = i - start;
1474 
1475 				maxbin = i - start;
1476 			}
1477 
1478 			if (minbin == -1) {
1479 				/*
1480 				 * If we have no data (e.g., due to a clear()
1481 				 * or negative increments), we'll use the
1482 				 * zero bucket as both our min and max.
1483 				 */
1484 				minbin = maxbin = DTRACE_QUANTIZE_ZEROBUCKET;
1485 			}
1486 
1487 			break;
1488 
1489 		default:
1490 			continue;
1491 		}
1492 
1493 		if (minmax[agg->dtagd_varid] == NULL) {
1494 			minmax[agg->dtagd_varid] = aggdata;
1495 			aggdata->dtada_flags |= DTRACE_A_MINMAXBIN;
1496 			aggdata->dtada_minbin = minbin;
1497 			aggdata->dtada_maxbin = maxbin;
1498 			continue;
1499 		}
1500 
1501 		if (minbin < minmax[agg->dtagd_varid]->dtada_minbin)
1502 			minmax[agg->dtagd_varid]->dtada_minbin = minbin;
1503 
1504 		if (maxbin > minmax[agg->dtagd_varid]->dtada_maxbin)
1505 			minmax[agg->dtagd_varid]->dtada_maxbin = maxbin;
1506 	}
1507 
1508 	/*
1509 	 * And now one final pass to set everyone's minbin and maxbin.
1510 	 */
1511 	for (h = hash->dtah_all; h != NULL; h = h->dtahe_nextall) {
1512 		dtrace_aggdata_t *aggdata = &h->dtahe_data, *mm;
1513 		dtrace_aggdesc_t *agg = aggdata->dtada_desc;
1514 
1515 		if ((mm = minmax[agg->dtagd_varid]) == NULL || aggdata == mm)
1516 			continue;
1517 
1518 		aggdata->dtada_minbin = mm->dtada_minbin;
1519 		aggdata->dtada_maxbin = mm->dtada_maxbin;
1520 		aggdata->dtada_flags |= DTRACE_A_MINMAXBIN;
1521 	}
1522 
1523 	dt_free(dtp, minmax);
1524 
1525 	return (0);
1526 }
1527 
1528 static int
1529 dt_aggregate_walk_sorted(dtrace_hdl_t *dtp,
1530     dtrace_aggregate_f *func, void *arg,
1531     int (*sfunc)(const void *, const void *))
1532 {
1533 	dt_aggregate_t *agp = &dtp->dt_aggregate;
1534 	dt_ahashent_t *h, **sorted;
1535 	dt_ahash_t *hash = &agp->dtat_hash;
1536 	size_t i, nentries = 0;
1537 	int rval = -1;
1538 
1539 	agp->dtat_flags &= ~(DTRACE_A_TOTAL | DTRACE_A_MINMAXBIN);
1540 
1541 	if (dtp->dt_options[DTRACEOPT_AGGHIST] != DTRACEOPT_UNSET) {
1542 		agp->dtat_flags |= DTRACE_A_TOTAL;
1543 
1544 		if (dt_aggregate_total(dtp, B_FALSE) != 0)
1545 			return (-1);
1546 	}
1547 
1548 	if (dtp->dt_options[DTRACEOPT_AGGPACK] != DTRACEOPT_UNSET) {
1549 		agp->dtat_flags |= DTRACE_A_MINMAXBIN;
1550 
1551 		if (dt_aggregate_minmaxbin(dtp, B_FALSE) != 0)
1552 			return (-1);
1553 	}
1554 
1555 	for (h = hash->dtah_all; h != NULL; h = h->dtahe_nextall)
1556 		nentries++;
1557 
1558 	sorted = dt_alloc(dtp, nentries * sizeof (dt_ahashent_t *));
1559 
1560 	if (sorted == NULL)
1561 		goto out;
1562 
1563 	for (h = hash->dtah_all, i = 0; h != NULL; h = h->dtahe_nextall)
1564 		sorted[i++] = h;
1565 
1566 	(void) pthread_mutex_lock(&dt_qsort_lock);
1567 
1568 	if (sfunc == NULL) {
1569 		dt_aggregate_qsort(dtp, sorted, nentries,
1570 		    sizeof (dt_ahashent_t *), NULL);
1571 	} else {
1572 		/*
1573 		 * If we've been explicitly passed a sorting function,
1574 		 * we'll use that -- ignoring the values of the "aggsortrev",
1575 		 * "aggsortkey" and "aggsortkeypos" options.
1576 		 */
1577 		qsort(sorted, nentries, sizeof (dt_ahashent_t *), sfunc);
1578 	}
1579 
1580 	(void) pthread_mutex_unlock(&dt_qsort_lock);
1581 
1582 	for (i = 0; i < nentries; i++) {
1583 		h = sorted[i];
1584 
1585 		if (dt_aggwalk_rval(dtp, h, func(&h->dtahe_data, arg)) == -1)
1586 			goto out;
1587 	}
1588 
1589 	rval = 0;
1590 out:
1591 	if (agp->dtat_flags & DTRACE_A_TOTAL)
1592 		(void) dt_aggregate_total(dtp, B_TRUE);
1593 
1594 	if (agp->dtat_flags & DTRACE_A_MINMAXBIN)
1595 		(void) dt_aggregate_minmaxbin(dtp, B_TRUE);
1596 
1597 	dt_free(dtp, sorted);
1598 	return (rval);
1599 }
1600 
1601 int
1602 dtrace_aggregate_walk_sorted(dtrace_hdl_t *dtp,
1603     dtrace_aggregate_f *func, void *arg)
1604 {
1605 	return (dt_aggregate_walk_sorted(dtp, func, arg, NULL));
1606 }
1607 
1608 int
1609 dtrace_aggregate_walk_keysorted(dtrace_hdl_t *dtp,
1610     dtrace_aggregate_f *func, void *arg)
1611 {
1612 	return (dt_aggregate_walk_sorted(dtp, func,
1613 	    arg, dt_aggregate_varkeycmp));
1614 }
1615 
1616 int
1617 dtrace_aggregate_walk_valsorted(dtrace_hdl_t *dtp,
1618     dtrace_aggregate_f *func, void *arg)
1619 {
1620 	return (dt_aggregate_walk_sorted(dtp, func,
1621 	    arg, dt_aggregate_varvalcmp));
1622 }
1623 
1624 int
1625 dtrace_aggregate_walk_keyvarsorted(dtrace_hdl_t *dtp,
1626     dtrace_aggregate_f *func, void *arg)
1627 {
1628 	return (dt_aggregate_walk_sorted(dtp, func,
1629 	    arg, dt_aggregate_keyvarcmp));
1630 }
1631 
1632 int
1633 dtrace_aggregate_walk_valvarsorted(dtrace_hdl_t *dtp,
1634     dtrace_aggregate_f *func, void *arg)
1635 {
1636 	return (dt_aggregate_walk_sorted(dtp, func,
1637 	    arg, dt_aggregate_valvarcmp));
1638 }
1639 
1640 int
1641 dtrace_aggregate_walk_keyrevsorted(dtrace_hdl_t *dtp,
1642     dtrace_aggregate_f *func, void *arg)
1643 {
1644 	return (dt_aggregate_walk_sorted(dtp, func,
1645 	    arg, dt_aggregate_varkeyrevcmp));
1646 }
1647 
1648 int
1649 dtrace_aggregate_walk_valrevsorted(dtrace_hdl_t *dtp,
1650     dtrace_aggregate_f *func, void *arg)
1651 {
1652 	return (dt_aggregate_walk_sorted(dtp, func,
1653 	    arg, dt_aggregate_varvalrevcmp));
1654 }
1655 
1656 int
1657 dtrace_aggregate_walk_keyvarrevsorted(dtrace_hdl_t *dtp,
1658     dtrace_aggregate_f *func, void *arg)
1659 {
1660 	return (dt_aggregate_walk_sorted(dtp, func,
1661 	    arg, dt_aggregate_keyvarrevcmp));
1662 }
1663 
1664 int
1665 dtrace_aggregate_walk_valvarrevsorted(dtrace_hdl_t *dtp,
1666     dtrace_aggregate_f *func, void *arg)
1667 {
1668 	return (dt_aggregate_walk_sorted(dtp, func,
1669 	    arg, dt_aggregate_valvarrevcmp));
1670 }
1671 
1672 int
1673 dtrace_aggregate_walk_joined(dtrace_hdl_t *dtp, dtrace_aggvarid_t *aggvars,
1674     int naggvars, dtrace_aggregate_walk_joined_f *func, void *arg)
1675 {
1676 	dt_aggregate_t *agp = &dtp->dt_aggregate;
1677 	dt_ahashent_t *h, **sorted = NULL, ***bundle, **nbundle;
1678 	const dtrace_aggdata_t **data;
1679 	dt_ahashent_t *zaggdata = NULL;
1680 	dt_ahash_t *hash = &agp->dtat_hash;
1681 	size_t nentries = 0, nbundles = 0, start, zsize = 0, bundlesize;
1682 	dtrace_aggvarid_t max = 0, aggvar;
1683 	int rval = -1, *map, *remap = NULL;
1684 	int i, j;
1685 	dtrace_optval_t sortpos = dtp->dt_options[DTRACEOPT_AGGSORTPOS];
1686 
1687 	/*
1688 	 * If the sorting position is greater than the number of aggregation
1689 	 * variable IDs, we silently set it to 0.
1690 	 */
1691 	if (sortpos == DTRACEOPT_UNSET || sortpos >= naggvars)
1692 		sortpos = 0;
1693 
1694 	/*
1695 	 * First we need to translate the specified aggregation variable IDs
1696 	 * into a linear map that will allow us to translate an aggregation
1697 	 * variable ID into its position in the specified aggvars.
1698 	 */
1699 	for (i = 0; i < naggvars; i++) {
1700 		if (aggvars[i] == DTRACE_AGGVARIDNONE || aggvars[i] < 0)
1701 			return (dt_set_errno(dtp, EDT_BADAGGVAR));
1702 
1703 		if (aggvars[i] > max)
1704 			max = aggvars[i];
1705 	}
1706 
1707 	if ((map = dt_zalloc(dtp, (max + 1) * sizeof (int))) == NULL)
1708 		return (-1);
1709 
1710 	zaggdata = dt_zalloc(dtp, naggvars * sizeof (dt_ahashent_t));
1711 
1712 	if (zaggdata == NULL)
1713 		goto out;
1714 
1715 	for (i = 0; i < naggvars; i++) {
1716 		int ndx = i + sortpos;
1717 
1718 		if (ndx >= naggvars)
1719 			ndx -= naggvars;
1720 
1721 		aggvar = aggvars[ndx];
1722 		assert(aggvar <= max);
1723 
1724 		if (map[aggvar]) {
1725 			/*
1726 			 * We have an aggregation variable that is present
1727 			 * more than once in the array of aggregation
1728 			 * variables.  While it's unclear why one might want
1729 			 * to do this, it's legal.  To support this construct,
1730 			 * we will allocate a remap that will indicate the
1731 			 * position from which this aggregation variable
1732 			 * should be pulled.  (That is, where the remap will
1733 			 * map from one position to another.)
1734 			 */
1735 			if (remap == NULL) {
1736 				remap = dt_zalloc(dtp, naggvars * sizeof (int));
1737 
1738 				if (remap == NULL)
1739 					goto out;
1740 			}
1741 
1742 			/*
1743 			 * Given that the variable is already present, assert
1744 			 * that following through the mapping and adjusting
1745 			 * for the sort position yields the same aggregation
1746 			 * variable ID.
1747 			 */
1748 			assert(aggvars[(map[aggvar] - 1 + sortpos) %
1749 			    naggvars] == aggvars[ndx]);
1750 
1751 			remap[i] = map[aggvar];
1752 			continue;
1753 		}
1754 
1755 		map[aggvar] = i + 1;
1756 	}
1757 
1758 	/*
1759 	 * We need to take two passes over the data to size our allocation, so
1760 	 * we'll use the first pass to also fill in the zero-filled data to be
1761 	 * used to properly format a zero-valued aggregation.
1762 	 */
1763 	for (h = hash->dtah_all; h != NULL; h = h->dtahe_nextall) {
1764 		dtrace_aggvarid_t id;
1765 		int ndx;
1766 
1767 		if ((id = dt_aggregate_aggvarid(h)) > max || !(ndx = map[id]))
1768 			continue;
1769 
1770 		if (zaggdata[ndx - 1].dtahe_size == 0) {
1771 			zaggdata[ndx - 1].dtahe_size = h->dtahe_size;
1772 			zaggdata[ndx - 1].dtahe_data = h->dtahe_data;
1773 		}
1774 
1775 		nentries++;
1776 	}
1777 
1778 	if (nentries == 0) {
1779 		/*
1780 		 * We couldn't find any entries; there is nothing else to do.
1781 		 */
1782 		rval = 0;
1783 		goto out;
1784 	}
1785 
1786 	/*
1787 	 * Before we sort the data, we're going to look for any holes in our
1788 	 * zero-filled data.  This will occur if an aggregation variable that
1789 	 * we are being asked to print has not yet been assigned the result of
1790 	 * any aggregating action for _any_ tuple.  The issue becomes that we
1791 	 * would like a zero value to be printed for all columns for this
1792 	 * aggregation, but without any record description, we don't know the
1793 	 * aggregating action that corresponds to the aggregation variable.  To
1794 	 * try to find a match, we're simply going to lookup aggregation IDs
1795 	 * (which are guaranteed to be contiguous and to start from 1), looking
1796 	 * for the specified aggregation variable ID.  If we find a match,
1797 	 * we'll use that.  If we iterate over all aggregation IDs and don't
1798 	 * find a match, then we must be an anonymous enabling.  (Anonymous
1799 	 * enablings can't currently derive either aggregation variable IDs or
1800 	 * aggregation variable names given only an aggregation ID.)  In this
1801 	 * obscure case (anonymous enabling, multiple aggregation printa() with
1802 	 * some aggregations not represented for any tuple), our defined
1803 	 * behavior is that the zero will be printed in the format of the first
1804 	 * aggregation variable that contains any non-zero value.
1805 	 */
1806 	for (i = 0; i < naggvars; i++) {
1807 		if (zaggdata[i].dtahe_size == 0) {
1808 			dtrace_aggvarid_t aggvar;
1809 
1810 			aggvar = aggvars[(i - sortpos + naggvars) % naggvars];
1811 			assert(zaggdata[i].dtahe_data.dtada_data == NULL);
1812 
1813 			for (j = DTRACE_AGGIDNONE + 1; ; j++) {
1814 				dtrace_aggdesc_t *agg;
1815 				dtrace_aggdata_t *aggdata;
1816 
1817 				if (dt_aggid_lookup(dtp, j, &agg) != 0)
1818 					break;
1819 
1820 				if (agg->dtagd_varid != aggvar)
1821 					continue;
1822 
1823 				/*
1824 				 * We have our description -- now we need to
1825 				 * cons up the zaggdata entry for it.
1826 				 */
1827 				aggdata = &zaggdata[i].dtahe_data;
1828 				aggdata->dtada_size = agg->dtagd_size;
1829 				aggdata->dtada_desc = agg;
1830 				aggdata->dtada_handle = dtp;
1831 				(void) dt_epid_lookup(dtp, agg->dtagd_epid,
1832 				    &aggdata->dtada_edesc,
1833 				    &aggdata->dtada_pdesc);
1834 				aggdata->dtada_normal = 1;
1835 				zaggdata[i].dtahe_hashval = 0;
1836 				zaggdata[i].dtahe_size = agg->dtagd_size;
1837 				break;
1838 			}
1839 
1840 			if (zaggdata[i].dtahe_size == 0) {
1841 				caddr_t data;
1842 
1843 				/*
1844 				 * We couldn't find this aggregation, meaning
1845 				 * that we have never seen it before for any
1846 				 * tuple _and_ this is an anonymous enabling.
1847 				 * That is, we're in the obscure case outlined
1848 				 * above.  In this case, our defined behavior
1849 				 * is to format the data in the format of the
1850 				 * first non-zero aggregation -- of which, of
1851 				 * course, we know there to be at least one
1852 				 * (or nentries would have been zero).
1853 				 */
1854 				for (j = 0; j < naggvars; j++) {
1855 					if (zaggdata[j].dtahe_size != 0)
1856 						break;
1857 				}
1858 
1859 				assert(j < naggvars);
1860 				zaggdata[i] = zaggdata[j];
1861 
1862 				data = zaggdata[i].dtahe_data.dtada_data;
1863 				assert(data != NULL);
1864 			}
1865 		}
1866 	}
1867 
1868 	/*
1869 	 * Now we need to allocate our zero-filled data for use for
1870 	 * aggregations that don't have a value corresponding to a given key.
1871 	 */
1872 	for (i = 0; i < naggvars; i++) {
1873 		dtrace_aggdata_t *aggdata = &zaggdata[i].dtahe_data;
1874 		dtrace_aggdesc_t *aggdesc = aggdata->dtada_desc;
1875 		dtrace_recdesc_t *rec;
1876 		uint64_t larg;
1877 		caddr_t zdata;
1878 
1879 		zsize = zaggdata[i].dtahe_size;
1880 		assert(zsize != 0);
1881 
1882 		if ((zdata = dt_zalloc(dtp, zsize)) == NULL) {
1883 			/*
1884 			 * If we failed to allocated some zero-filled data, we
1885 			 * need to zero out the remaining dtada_data pointers
1886 			 * to prevent the wrong data from being freed below.
1887 			 */
1888 			for (j = i; j < naggvars; j++)
1889 				zaggdata[j].dtahe_data.dtada_data = NULL;
1890 			goto out;
1891 		}
1892 
1893 		aggvar = aggvars[(i - sortpos + naggvars) % naggvars];
1894 
1895 		/*
1896 		 * First, the easy bit.  To maintain compatibility with
1897 		 * consumers that pull the compiler-generated ID out of the
1898 		 * data, we put that ID at the top of the zero-filled data.
1899 		 */
1900 		rec = &aggdesc->dtagd_rec[0];
1901 		/* LINTED - alignment */
1902 		*((dtrace_aggvarid_t *)(zdata + rec->dtrd_offset)) = aggvar;
1903 
1904 		rec = &aggdesc->dtagd_rec[aggdesc->dtagd_nrecs - 1];
1905 
1906 		/*
1907 		 * Now for the more complicated part.  If (and only if) this
1908 		 * is an lquantize() aggregating action, zero-filled data is
1909 		 * not equivalent to an empty record:  we must also get the
1910 		 * parameters for the lquantize().
1911 		 */
1912 		if (rec->dtrd_action == DTRACEAGG_LQUANTIZE) {
1913 			if (aggdata->dtada_data != NULL) {
1914 				/*
1915 				 * The easier case here is if we actually have
1916 				 * some prototype data -- in which case we
1917 				 * manually dig it out of the aggregation
1918 				 * record.
1919 				 */
1920 				/* LINTED - alignment */
1921 				larg = *((uint64_t *)(aggdata->dtada_data +
1922 				    rec->dtrd_offset));
1923 			} else {
1924 				/*
1925 				 * We don't have any prototype data.  As a
1926 				 * result, we know that we _do_ have the
1927 				 * compiler-generated information.  (If this
1928 				 * were an anonymous enabling, all of our
1929 				 * zero-filled data would have prototype data
1930 				 * -- either directly or indirectly.) So as
1931 				 * gross as it is, we'll grovel around in the
1932 				 * compiler-generated information to find the
1933 				 * lquantize() parameters.
1934 				 */
1935 				dtrace_stmtdesc_t *sdp;
1936 				dt_ident_t *aid;
1937 				dt_idsig_t *isp;
1938 
1939 				sdp = (dtrace_stmtdesc_t *)(uintptr_t)
1940 				    aggdesc->dtagd_rec[0].dtrd_uarg;
1941 				aid = sdp->dtsd_aggdata;
1942 				isp = (dt_idsig_t *)aid->di_data;
1943 				assert(isp->dis_auxinfo != 0);
1944 				larg = isp->dis_auxinfo;
1945 			}
1946 
1947 			/* LINTED - alignment */
1948 			*((uint64_t *)(zdata + rec->dtrd_offset)) = larg;
1949 		}
1950 
1951 		aggdata->dtada_data = zdata;
1952 	}
1953 
1954 	/*
1955 	 * Now that we've dealt with setting up our zero-filled data, we can
1956 	 * allocate our sorted array, and take another pass over the data to
1957 	 * fill it.
1958 	 */
1959 	sorted = dt_alloc(dtp, nentries * sizeof (dt_ahashent_t *));
1960 
1961 	if (sorted == NULL)
1962 		goto out;
1963 
1964 	for (h = hash->dtah_all, i = 0; h != NULL; h = h->dtahe_nextall) {
1965 		dtrace_aggvarid_t id;
1966 
1967 		if ((id = dt_aggregate_aggvarid(h)) > max || !map[id])
1968 			continue;
1969 
1970 		sorted[i++] = h;
1971 	}
1972 
1973 	assert(i == nentries);
1974 
1975 	/*
1976 	 * We've loaded our array; now we need to sort by value to allow us
1977 	 * to create bundles of like value.  We're going to acquire the
1978 	 * dt_qsort_lock here, and hold it across all of our subsequent
1979 	 * comparison and sorting.
1980 	 */
1981 	(void) pthread_mutex_lock(&dt_qsort_lock);
1982 
1983 	qsort(sorted, nentries, sizeof (dt_ahashent_t *),
1984 	    dt_aggregate_keyvarcmp);
1985 
1986 	/*
1987 	 * Now we need to go through and create bundles.  Because the number
1988 	 * of bundles is bounded by the size of the sorted array, we're going
1989 	 * to reuse the underlying storage.  And note that "bundle" is an
1990 	 * array of pointers to arrays of pointers to dt_ahashent_t -- making
1991 	 * its type (regrettably) "dt_ahashent_t ***".  (Regrettable because
1992 	 * '*' -- like '_' and 'X' -- should never appear in triplicate in
1993 	 * an ideal world.)
1994 	 */
1995 	bundle = (dt_ahashent_t ***)sorted;
1996 
1997 	for (i = 1, start = 0; i <= nentries; i++) {
1998 		if (i < nentries &&
1999 		    dt_aggregate_keycmp(&sorted[i], &sorted[i - 1]) == 0)
2000 			continue;
2001 
2002 		/*
2003 		 * We have a bundle boundary.  Everything from start to
2004 		 * (i - 1) belongs in one bundle.
2005 		 */
2006 		assert(i - start <= naggvars);
2007 		bundlesize = (naggvars + 2) * sizeof (dt_ahashent_t *);
2008 
2009 		if ((nbundle = dt_zalloc(dtp, bundlesize)) == NULL) {
2010 			(void) pthread_mutex_unlock(&dt_qsort_lock);
2011 			goto out;
2012 		}
2013 
2014 		for (j = start; j < i; j++) {
2015 			dtrace_aggvarid_t id = dt_aggregate_aggvarid(sorted[j]);
2016 
2017 			assert(id <= max);
2018 			assert(map[id] != 0);
2019 			assert(map[id] - 1 < naggvars);
2020 			assert(nbundle[map[id] - 1] == NULL);
2021 			nbundle[map[id] - 1] = sorted[j];
2022 
2023 			if (nbundle[naggvars] == NULL)
2024 				nbundle[naggvars] = sorted[j];
2025 		}
2026 
2027 		for (j = 0; j < naggvars; j++) {
2028 			if (nbundle[j] != NULL)
2029 				continue;
2030 
2031 			/*
2032 			 * Before we assume that this aggregation variable
2033 			 * isn't present (and fall back to using the
2034 			 * zero-filled data allocated earlier), check the
2035 			 * remap.  If we have a remapping, we'll drop it in
2036 			 * here.  Note that we might be remapping an
2037 			 * aggregation variable that isn't present for this
2038 			 * key; in this case, the aggregation data that we
2039 			 * copy will point to the zeroed data.
2040 			 */
2041 			if (remap != NULL && remap[j]) {
2042 				assert(remap[j] - 1 < j);
2043 				assert(nbundle[remap[j] - 1] != NULL);
2044 				nbundle[j] = nbundle[remap[j] - 1];
2045 			} else {
2046 				nbundle[j] = &zaggdata[j];
2047 			}
2048 		}
2049 
2050 		bundle[nbundles++] = nbundle;
2051 		start = i;
2052 	}
2053 
2054 	/*
2055 	 * Now we need to re-sort based on the first value.
2056 	 */
2057 	dt_aggregate_qsort(dtp, bundle, nbundles, sizeof (dt_ahashent_t **),
2058 	    dt_aggregate_bundlecmp);
2059 
2060 	(void) pthread_mutex_unlock(&dt_qsort_lock);
2061 
2062 	/*
2063 	 * We're done!  Now we just need to go back over the sorted bundles,
2064 	 * calling the function.
2065 	 */
2066 	data = alloca((naggvars + 1) * sizeof (dtrace_aggdata_t *));
2067 
2068 	for (i = 0; i < nbundles; i++) {
2069 		for (j = 0; j < naggvars; j++)
2070 			data[j + 1] = NULL;
2071 
2072 		for (j = 0; j < naggvars; j++) {
2073 			int ndx = j - sortpos;
2074 
2075 			if (ndx < 0)
2076 				ndx += naggvars;
2077 
2078 			assert(bundle[i][ndx] != NULL);
2079 			data[j + 1] = &bundle[i][ndx]->dtahe_data;
2080 		}
2081 
2082 		for (j = 0; j < naggvars; j++)
2083 			assert(data[j + 1] != NULL);
2084 
2085 		/*
2086 		 * The representative key is the last element in the bundle.
2087 		 * Assert that we have one, and then set it to be the first
2088 		 * element of data.
2089 		 */
2090 		assert(bundle[i][j] != NULL);
2091 		data[0] = &bundle[i][j]->dtahe_data;
2092 
2093 		if ((rval = func(data, naggvars + 1, arg)) == -1)
2094 			goto out;
2095 	}
2096 
2097 	rval = 0;
2098 out:
2099 	for (i = 0; i < nbundles; i++)
2100 		dt_free(dtp, bundle[i]);
2101 
2102 	if (zaggdata != NULL) {
2103 		for (i = 0; i < naggvars; i++)
2104 			dt_free(dtp, zaggdata[i].dtahe_data.dtada_data);
2105 	}
2106 
2107 	dt_free(dtp, zaggdata);
2108 	dt_free(dtp, sorted);
2109 	dt_free(dtp, remap);
2110 	dt_free(dtp, map);
2111 
2112 	return (rval);
2113 }
2114 
2115 int
2116 dtrace_aggregate_print(dtrace_hdl_t *dtp, FILE *fp,
2117     dtrace_aggregate_walk_f *func)
2118 {
2119 	dt_print_aggdata_t pd;
2120 
2121 	bzero(&pd, sizeof (pd));
2122 
2123 	pd.dtpa_dtp = dtp;
2124 	pd.dtpa_fp = fp;
2125 	pd.dtpa_allunprint = 1;
2126 
2127 	if (func == NULL)
2128 		func = dtrace_aggregate_walk_sorted;
2129 
2130 	if ((*func)(dtp, dt_print_agg, &pd) == -1)
2131 		return (dt_set_errno(dtp, dtp->dt_errno));
2132 
2133 	return (0);
2134 }
2135 
2136 void
2137 dtrace_aggregate_clear(dtrace_hdl_t *dtp)
2138 {
2139 	dt_aggregate_t *agp = &dtp->dt_aggregate;
2140 	dt_ahash_t *hash = &agp->dtat_hash;
2141 	dt_ahashent_t *h;
2142 	dtrace_aggdata_t *data;
2143 	dtrace_aggdesc_t *aggdesc;
2144 	dtrace_recdesc_t *rec;
2145 	int i, max_cpus = agp->dtat_maxcpu;
2146 
2147 	for (h = hash->dtah_all; h != NULL; h = h->dtahe_nextall) {
2148 		aggdesc = h->dtahe_data.dtada_desc;
2149 		rec = &aggdesc->dtagd_rec[aggdesc->dtagd_nrecs - 1];
2150 		data = &h->dtahe_data;
2151 
2152 		bzero(&data->dtada_data[rec->dtrd_offset], rec->dtrd_size);
2153 
2154 		if (data->dtada_percpu == NULL)
2155 			continue;
2156 
2157 		for (i = 0; i < max_cpus; i++)
2158 			bzero(data->dtada_percpu[i], rec->dtrd_size);
2159 	}
2160 }
2161 
2162 void
2163 dt_aggregate_destroy(dtrace_hdl_t *dtp)
2164 {
2165 	dt_aggregate_t *agp = &dtp->dt_aggregate;
2166 	dt_ahash_t *hash = &agp->dtat_hash;
2167 	dt_ahashent_t *h, *next;
2168 	dtrace_aggdata_t *aggdata;
2169 	int i, max_cpus = agp->dtat_maxcpu;
2170 
2171 	if (hash->dtah_hash == NULL) {
2172 		assert(hash->dtah_all == NULL);
2173 	} else {
2174 		free(hash->dtah_hash);
2175 
2176 		for (h = hash->dtah_all; h != NULL; h = next) {
2177 			next = h->dtahe_nextall;
2178 
2179 			aggdata = &h->dtahe_data;
2180 
2181 			if (aggdata->dtada_percpu != NULL) {
2182 				for (i = 0; i < max_cpus; i++)
2183 					free(aggdata->dtada_percpu[i]);
2184 				free(aggdata->dtada_percpu);
2185 			}
2186 
2187 			free(aggdata->dtada_data);
2188 			free(h);
2189 		}
2190 
2191 		hash->dtah_hash = NULL;
2192 		hash->dtah_all = NULL;
2193 		hash->dtah_size = 0;
2194 	}
2195 
2196 	free(agp->dtat_buf.dtbd_data);
2197 	free(agp->dtat_cpus);
2198 }
2199