xref: /freebsd/cddl/contrib/opensolaris/lib/libdtrace/common/dt_cg.c (revision b1f9167f94059fd55c630891d359bcff987bd7eb)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License, Version 1.0 only
6  * (the "License").  You may not use this file except in compliance
7  * with the License.
8  *
9  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
10  * or http://www.opensolaris.org/os/licensing.
11  * See the License for the specific language governing permissions
12  * and limitations under the License.
13  *
14  * When distributing Covered Code, include this CDDL HEADER in each
15  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
16  * If applicable, add the following below this CDDL HEADER, with the
17  * fields enclosed by brackets "[]" replaced with your own identifying
18  * information: Portions Copyright [yyyy] [name of copyright owner]
19  *
20  * CDDL HEADER END
21  */
22 
23 /*
24  * Copyright 2005 Sun Microsystems, Inc.  All rights reserved.
25  * Use is subject to license terms.
26  */
27 
28 /*
29  * Copyright (c) 2012 by Delphix. All rights reserved.
30  */
31 
32 #include <sys/types.h>
33 #include <sys/sysmacros.h>
34 #include <sys/isa_defs.h>
35 
36 #include <strings.h>
37 #include <stdlib.h>
38 #include <setjmp.h>
39 #include <assert.h>
40 #include <errno.h>
41 
42 #include <dt_impl.h>
43 #include <dt_grammar.h>
44 #include <dt_parser.h>
45 #include <dt_provider.h>
46 
47 static void dt_cg_node(dt_node_t *, dt_irlist_t *, dt_regset_t *);
48 
49 static dt_irnode_t *
50 dt_cg_node_alloc(uint_t label, dif_instr_t instr)
51 {
52 	dt_irnode_t *dip = malloc(sizeof (dt_irnode_t));
53 
54 	if (dip == NULL)
55 		longjmp(yypcb->pcb_jmpbuf, EDT_NOMEM);
56 
57 	dip->di_label = label;
58 	dip->di_instr = instr;
59 	dip->di_extern = NULL;
60 	dip->di_next = NULL;
61 
62 	return (dip);
63 }
64 
65 /*
66  * Code generator wrapper function for ctf_member_info.  If we are given a
67  * reference to a forward declaration tag, search the entire type space for
68  * the actual definition and then call ctf_member_info on the result.
69  */
70 static ctf_file_t *
71 dt_cg_membinfo(ctf_file_t *fp, ctf_id_t type, const char *s, ctf_membinfo_t *mp)
72 {
73 	while (ctf_type_kind(fp, type) == CTF_K_FORWARD) {
74 		char n[DT_TYPE_NAMELEN];
75 		dtrace_typeinfo_t dtt;
76 
77 		if (ctf_type_name(fp, type, n, sizeof (n)) == NULL ||
78 		    dt_type_lookup(n, &dtt) == -1 || (
79 		    dtt.dtt_ctfp == fp && dtt.dtt_type == type))
80 			break; /* unable to improve our position */
81 
82 		fp = dtt.dtt_ctfp;
83 		type = ctf_type_resolve(fp, dtt.dtt_type);
84 	}
85 
86 	if (ctf_member_info(fp, type, s, mp) == CTF_ERR)
87 		return (NULL); /* ctf_errno is set for us */
88 
89 	return (fp);
90 }
91 
92 static void
93 dt_cg_xsetx(dt_irlist_t *dlp, dt_ident_t *idp, uint_t lbl, int reg, uint64_t x)
94 {
95 	int flag = idp != NULL ? DT_INT_PRIVATE : DT_INT_SHARED;
96 	int intoff = dt_inttab_insert(yypcb->pcb_inttab, x, flag);
97 	dif_instr_t instr = DIF_INSTR_SETX((uint_t)intoff, reg);
98 
99 	if (intoff == -1)
100 		longjmp(yypcb->pcb_jmpbuf, EDT_NOMEM);
101 
102 	if (intoff > DIF_INTOFF_MAX)
103 		longjmp(yypcb->pcb_jmpbuf, EDT_INT2BIG);
104 
105 	dt_irlist_append(dlp, dt_cg_node_alloc(lbl, instr));
106 
107 	if (idp != NULL)
108 		dlp->dl_last->di_extern = idp;
109 }
110 
111 static void
112 dt_cg_setx(dt_irlist_t *dlp, int reg, uint64_t x)
113 {
114 	dt_cg_xsetx(dlp, NULL, DT_LBL_NONE, reg, x);
115 }
116 
117 /*
118  * When loading bit-fields, we want to convert a byte count in the range
119  * 1-8 to the closest power of 2 (e.g. 3->4, 5->8, etc).  The clp2() function
120  * is a clever implementation from "Hacker's Delight" by Henry Warren, Jr.
121  */
122 static size_t
123 clp2(size_t x)
124 {
125 	x--;
126 
127 	x |= (x >> 1);
128 	x |= (x >> 2);
129 	x |= (x >> 4);
130 	x |= (x >> 8);
131 	x |= (x >> 16);
132 
133 	return (x + 1);
134 }
135 
136 /*
137  * Lookup the correct load opcode to use for the specified node and CTF type.
138  * We determine the size and convert it to a 3-bit index.  Our lookup table
139  * is constructed to use a 5-bit index, consisting of the 3-bit size 0-7, a
140  * bit for the sign, and a bit for userland address.  For example, a 4-byte
141  * signed load from userland would be at the following table index:
142  * user=1 sign=1 size=4 => binary index 11011 = decimal index 27
143  */
144 static uint_t
145 dt_cg_load(dt_node_t *dnp, ctf_file_t *ctfp, ctf_id_t type)
146 {
147 	static const uint_t ops[] = {
148 		DIF_OP_LDUB,	DIF_OP_LDUH,	0,	DIF_OP_LDUW,
149 		0,		0,		0,	DIF_OP_LDX,
150 		DIF_OP_LDSB,	DIF_OP_LDSH,	0,	DIF_OP_LDSW,
151 		0,		0,		0,	DIF_OP_LDX,
152 		DIF_OP_ULDUB,	DIF_OP_ULDUH,	0,	DIF_OP_ULDUW,
153 		0,		0,		0,	DIF_OP_ULDX,
154 		DIF_OP_ULDSB,	DIF_OP_ULDSH,	0,	DIF_OP_ULDSW,
155 		0,		0,		0,	DIF_OP_ULDX,
156 	};
157 
158 	ctf_encoding_t e;
159 	ssize_t size;
160 
161 	/*
162 	 * If we're loading a bit-field, the size of our load is found by
163 	 * rounding cte_bits up to a byte boundary and then finding the
164 	 * nearest power of two to this value (see clp2(), above).
165 	 */
166 	if ((dnp->dn_flags & DT_NF_BITFIELD) &&
167 	    ctf_type_encoding(ctfp, type, &e) != CTF_ERR)
168 		size = clp2(P2ROUNDUP(e.cte_bits, NBBY) / NBBY);
169 	else
170 		size = ctf_type_size(ctfp, type);
171 
172 	if (size < 1 || size > 8 || (size & (size - 1)) != 0) {
173 		xyerror(D_UNKNOWN, "internal error -- cg cannot load "
174 		    "size %ld when passed by value\n", (long)size);
175 	}
176 
177 	size--; /* convert size to 3-bit index */
178 
179 	if (dnp->dn_flags & DT_NF_SIGNED)
180 		size |= 0x08;
181 	if (dnp->dn_flags & DT_NF_USERLAND)
182 		size |= 0x10;
183 
184 	return (ops[size]);
185 }
186 
187 static void
188 dt_cg_ptrsize(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp,
189     uint_t op, int dreg)
190 {
191 	ctf_file_t *ctfp = dnp->dn_ctfp;
192 	ctf_arinfo_t r;
193 	dif_instr_t instr;
194 	ctf_id_t type;
195 	uint_t kind;
196 	ssize_t size;
197 	int sreg;
198 
199 	type = ctf_type_resolve(ctfp, dnp->dn_type);
200 	kind = ctf_type_kind(ctfp, type);
201 	assert(kind == CTF_K_POINTER || kind == CTF_K_ARRAY);
202 
203 	if (kind == CTF_K_ARRAY) {
204 		if (ctf_array_info(ctfp, type, &r) != 0) {
205 			yypcb->pcb_hdl->dt_ctferr = ctf_errno(ctfp);
206 			longjmp(yypcb->pcb_jmpbuf, EDT_CTF);
207 		}
208 		type = r.ctr_contents;
209 	} else
210 		type = ctf_type_reference(ctfp, type);
211 
212 	if ((size = ctf_type_size(ctfp, type)) == 1)
213 		return; /* multiply or divide by one can be omitted */
214 
215 	sreg = dt_regset_alloc(drp);
216 	dt_cg_setx(dlp, sreg, size);
217 	instr = DIF_INSTR_FMT(op, dreg, sreg, dreg);
218 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
219 	dt_regset_free(drp, sreg);
220 }
221 
222 /*
223  * If the result of a "." or "->" operation is a bit-field, we use this routine
224  * to generate an epilogue to the load instruction that extracts the value.  In
225  * the diagrams below the "ld??" is the load instruction that is generated to
226  * load the containing word that is generating prior to calling this function.
227  *
228  * Epilogue for unsigned fields:	Epilogue for signed fields:
229  *
230  * ldu?	[r1], r1			lds? [r1], r1
231  * setx	USHIFT, r2			setx 64 - SSHIFT, r2
232  * srl	r1, r2, r1			sll  r1, r2, r1
233  * setx	(1 << bits) - 1, r2		setx 64 - bits, r2
234  * and	r1, r2, r1			sra  r1, r2, r1
235  *
236  * The *SHIFT constants above changes value depending on the endian-ness of our
237  * target architecture.  Refer to the comments below for more details.
238  */
239 static void
240 dt_cg_field_get(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp,
241     ctf_file_t *fp, const ctf_membinfo_t *mp)
242 {
243 	ctf_encoding_t e;
244 	dif_instr_t instr;
245 	uint64_t shift;
246 	int r1, r2;
247 
248 	if (ctf_type_encoding(fp, mp->ctm_type, &e) != 0 || e.cte_bits > 64) {
249 		xyerror(D_UNKNOWN, "cg: bad field: off %lu type <%ld> "
250 		    "bits %u\n", mp->ctm_offset, mp->ctm_type, e.cte_bits);
251 	}
252 
253 	assert(dnp->dn_op == DT_TOK_PTR || dnp->dn_op == DT_TOK_DOT);
254 	r1 = dnp->dn_left->dn_reg;
255 	r2 = dt_regset_alloc(drp);
256 
257 	/*
258 	 * On little-endian architectures, ctm_offset counts from the right so
259 	 * ctm_offset % NBBY itself is the amount we want to shift right to
260 	 * move the value bits to the little end of the register to mask them.
261 	 * On big-endian architectures, ctm_offset counts from the left so we
262 	 * must subtract (ctm_offset % NBBY + cte_bits) from the size in bits
263 	 * we used for the load.  The size of our load in turn is found by
264 	 * rounding cte_bits up to a byte boundary and then finding the
265 	 * nearest power of two to this value (see clp2(), above).  These
266 	 * properties are used to compute shift as USHIFT or SSHIFT, below.
267 	 */
268 	if (dnp->dn_flags & DT_NF_SIGNED) {
269 #if BYTE_ORDER == _BIG_ENDIAN
270 		shift = clp2(P2ROUNDUP(e.cte_bits, NBBY) / NBBY) * NBBY -
271 		    mp->ctm_offset % NBBY;
272 #else
273 		shift = mp->ctm_offset % NBBY + e.cte_bits;
274 #endif
275 		dt_cg_setx(dlp, r2, 64 - shift);
276 		instr = DIF_INSTR_FMT(DIF_OP_SLL, r1, r2, r1);
277 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
278 
279 		dt_cg_setx(dlp, r2, 64 - e.cte_bits);
280 		instr = DIF_INSTR_FMT(DIF_OP_SRA, r1, r2, r1);
281 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
282 	} else {
283 #if BYTE_ORDER == _BIG_ENDIAN
284 		shift = clp2(P2ROUNDUP(e.cte_bits, NBBY) / NBBY) * NBBY -
285 		    (mp->ctm_offset % NBBY + e.cte_bits);
286 #else
287 		shift = mp->ctm_offset % NBBY;
288 #endif
289 		dt_cg_setx(dlp, r2, shift);
290 		instr = DIF_INSTR_FMT(DIF_OP_SRL, r1, r2, r1);
291 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
292 
293 		dt_cg_setx(dlp, r2, (1ULL << e.cte_bits) - 1);
294 		instr = DIF_INSTR_FMT(DIF_OP_AND, r1, r2, r1);
295 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
296 	}
297 
298 	dt_regset_free(drp, r2);
299 }
300 
301 /*
302  * If the destination of a store operation is a bit-field, we use this routine
303  * to generate a prologue to the store instruction that loads the surrounding
304  * bits, clears the destination field, and ORs in the new value of the field.
305  * In the diagram below the "st?" is the store instruction that is generated to
306  * store the containing word that is generating after calling this function.
307  *
308  * ld	[dst->dn_reg], r1
309  * setx	~(((1 << cte_bits) - 1) << (ctm_offset % NBBY)), r2
310  * and	r1, r2, r1
311  *
312  * setx	(1 << cte_bits) - 1, r2
313  * and	src->dn_reg, r2, r2
314  * setx ctm_offset % NBBY, r3
315  * sll	r2, r3, r2
316  *
317  * or	r1, r2, r1
318  * st?	r1, [dst->dn_reg]
319  *
320  * This routine allocates a new register to hold the value to be stored and
321  * returns it.  The caller is responsible for freeing this register later.
322  */
323 static int
324 dt_cg_field_set(dt_node_t *src, dt_irlist_t *dlp,
325     dt_regset_t *drp, dt_node_t *dst)
326 {
327 	uint64_t cmask, fmask, shift;
328 	dif_instr_t instr;
329 	int r1, r2, r3;
330 
331 	ctf_membinfo_t m;
332 	ctf_encoding_t e;
333 	ctf_file_t *fp, *ofp;
334 	ctf_id_t type;
335 
336 	assert(dst->dn_op == DT_TOK_PTR || dst->dn_op == DT_TOK_DOT);
337 	assert(dst->dn_right->dn_kind == DT_NODE_IDENT);
338 
339 	fp = dst->dn_left->dn_ctfp;
340 	type = ctf_type_resolve(fp, dst->dn_left->dn_type);
341 
342 	if (dst->dn_op == DT_TOK_PTR) {
343 		type = ctf_type_reference(fp, type);
344 		type = ctf_type_resolve(fp, type);
345 	}
346 
347 	if ((fp = dt_cg_membinfo(ofp = fp, type,
348 	    dst->dn_right->dn_string, &m)) == NULL) {
349 		yypcb->pcb_hdl->dt_ctferr = ctf_errno(ofp);
350 		longjmp(yypcb->pcb_jmpbuf, EDT_CTF);
351 	}
352 
353 	if (ctf_type_encoding(fp, m.ctm_type, &e) != 0 || e.cte_bits > 64) {
354 		xyerror(D_UNKNOWN, "cg: bad field: off %lu type <%ld> "
355 		    "bits %u\n", m.ctm_offset, m.ctm_type, e.cte_bits);
356 	}
357 
358 	r1 = dt_regset_alloc(drp);
359 	r2 = dt_regset_alloc(drp);
360 	r3 = dt_regset_alloc(drp);
361 
362 	/*
363 	 * Compute shifts and masks.  We need to compute "shift" as the amount
364 	 * we need to shift left to position our field in the containing word.
365 	 * Refer to the comments in dt_cg_field_get(), above, for more info.
366 	 * We then compute fmask as the mask that truncates the value in the
367 	 * input register to width cte_bits, and cmask as the mask used to
368 	 * pass through the containing bits and zero the field bits.
369 	 */
370 #if BYTE_ORDER == _BIG_ENDIAN
371 	shift = clp2(P2ROUNDUP(e.cte_bits, NBBY) / NBBY) * NBBY -
372 	    (m.ctm_offset % NBBY + e.cte_bits);
373 #else
374 	shift = m.ctm_offset % NBBY;
375 #endif
376 	fmask = (1ULL << e.cte_bits) - 1;
377 	cmask = ~(fmask << shift);
378 
379 	instr = DIF_INSTR_LOAD(
380 	    dt_cg_load(dst, fp, m.ctm_type), dst->dn_reg, r1);
381 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
382 
383 	dt_cg_setx(dlp, r2, cmask);
384 	instr = DIF_INSTR_FMT(DIF_OP_AND, r1, r2, r1);
385 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
386 
387 	dt_cg_setx(dlp, r2, fmask);
388 	instr = DIF_INSTR_FMT(DIF_OP_AND, src->dn_reg, r2, r2);
389 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
390 
391 	dt_cg_setx(dlp, r3, shift);
392 	instr = DIF_INSTR_FMT(DIF_OP_SLL, r2, r3, r2);
393 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
394 
395 	instr = DIF_INSTR_FMT(DIF_OP_OR, r1, r2, r1);
396 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
397 
398 	dt_regset_free(drp, r3);
399 	dt_regset_free(drp, r2);
400 
401 	return (r1);
402 }
403 
404 static void
405 dt_cg_store(dt_node_t *src, dt_irlist_t *dlp, dt_regset_t *drp, dt_node_t *dst)
406 {
407 	ctf_encoding_t e;
408 	dif_instr_t instr;
409 	size_t size;
410 	int reg;
411 
412 	/*
413 	 * If we're loading a bit-field, the size of our store is found by
414 	 * rounding dst's cte_bits up to a byte boundary and then finding the
415 	 * nearest power of two to this value (see clp2(), above).
416 	 */
417 	if ((dst->dn_flags & DT_NF_BITFIELD) &&
418 	    ctf_type_encoding(dst->dn_ctfp, dst->dn_type, &e) != CTF_ERR)
419 		size = clp2(P2ROUNDUP(e.cte_bits, NBBY) / NBBY);
420 	else
421 		size = dt_node_type_size(src);
422 
423 	if (src->dn_flags & DT_NF_REF) {
424 		reg = dt_regset_alloc(drp);
425 		dt_cg_setx(dlp, reg, size);
426 		instr = DIF_INSTR_COPYS(src->dn_reg, reg, dst->dn_reg);
427 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
428 		dt_regset_free(drp, reg);
429 	} else {
430 		if (dst->dn_flags & DT_NF_BITFIELD)
431 			reg = dt_cg_field_set(src, dlp, drp, dst);
432 		else
433 			reg = src->dn_reg;
434 
435 		switch (size) {
436 		case 1:
437 			instr = DIF_INSTR_STORE(DIF_OP_STB, reg, dst->dn_reg);
438 			break;
439 		case 2:
440 			instr = DIF_INSTR_STORE(DIF_OP_STH, reg, dst->dn_reg);
441 			break;
442 		case 4:
443 			instr = DIF_INSTR_STORE(DIF_OP_STW, reg, dst->dn_reg);
444 			break;
445 		case 8:
446 			instr = DIF_INSTR_STORE(DIF_OP_STX, reg, dst->dn_reg);
447 			break;
448 		default:
449 			xyerror(D_UNKNOWN, "internal error -- cg cannot store "
450 			    "size %lu when passed by value\n", (ulong_t)size);
451 		}
452 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
453 
454 		if (dst->dn_flags & DT_NF_BITFIELD)
455 			dt_regset_free(drp, reg);
456 	}
457 }
458 
459 /*
460  * Generate code for a typecast or for argument promotion from the type of the
461  * actual to the type of the formal.  We need to generate code for casts when
462  * a scalar type is being narrowed or changing signed-ness.  We first shift the
463  * desired bits high (losing excess bits if narrowing) and then shift them down
464  * using logical shift (unsigned result) or arithmetic shift (signed result).
465  */
466 static void
467 dt_cg_typecast(const dt_node_t *src, const dt_node_t *dst,
468     dt_irlist_t *dlp, dt_regset_t *drp)
469 {
470 	size_t srcsize = dt_node_type_size(src);
471 	size_t dstsize = dt_node_type_size(dst);
472 
473 	dif_instr_t instr;
474 	int rg;
475 
476 	if (!dt_node_is_scalar(dst))
477 		return; /* not a scalar */
478 	if (dstsize == srcsize &&
479 	    ((src->dn_flags ^ dst->dn_flags) & DT_NF_SIGNED) != 0)
480 		return; /* not narrowing or changing signed-ness */
481 	if (dstsize > srcsize && (src->dn_flags & DT_NF_SIGNED) == 0)
482 		return; /* nothing to do in this case */
483 
484 	rg = dt_regset_alloc(drp);
485 
486 	if (dstsize > srcsize) {
487 		int n = sizeof (uint64_t) * NBBY - srcsize * NBBY;
488 		int s = (dstsize - srcsize) * NBBY;
489 
490 		dt_cg_setx(dlp, rg, n);
491 
492 		instr = DIF_INSTR_FMT(DIF_OP_SLL, src->dn_reg, rg, dst->dn_reg);
493 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
494 
495 		if ((dst->dn_flags & DT_NF_SIGNED) || n == s) {
496 			instr = DIF_INSTR_FMT(DIF_OP_SRA,
497 			    dst->dn_reg, rg, dst->dn_reg);
498 			dt_irlist_append(dlp,
499 			    dt_cg_node_alloc(DT_LBL_NONE, instr));
500 		} else {
501 			dt_cg_setx(dlp, rg, s);
502 			instr = DIF_INSTR_FMT(DIF_OP_SRA,
503 			    dst->dn_reg, rg, dst->dn_reg);
504 			dt_irlist_append(dlp,
505 			    dt_cg_node_alloc(DT_LBL_NONE, instr));
506 			dt_cg_setx(dlp, rg, n - s);
507 			instr = DIF_INSTR_FMT(DIF_OP_SRL,
508 			    dst->dn_reg, rg, dst->dn_reg);
509 			dt_irlist_append(dlp,
510 			    dt_cg_node_alloc(DT_LBL_NONE, instr));
511 		}
512 	} else if (dstsize != sizeof (uint64_t)) {
513 		int n = sizeof (uint64_t) * NBBY - dstsize * NBBY;
514 
515 		dt_cg_setx(dlp, rg, n);
516 
517 		instr = DIF_INSTR_FMT(DIF_OP_SLL, src->dn_reg, rg, dst->dn_reg);
518 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
519 
520 		instr = DIF_INSTR_FMT((dst->dn_flags & DT_NF_SIGNED) ?
521 		    DIF_OP_SRA : DIF_OP_SRL, dst->dn_reg, rg, dst->dn_reg);
522 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
523 	}
524 
525 	dt_regset_free(drp, rg);
526 }
527 
528 /*
529  * Generate code to push the specified argument list on to the tuple stack.
530  * We use this routine for handling subroutine calls and associative arrays.
531  * We must first generate code for all subexpressions before loading the stack
532  * because any subexpression could itself require the use of the tuple stack.
533  * This holds a number of registers equal to the number of arguments, but this
534  * is not a huge problem because the number of arguments can't exceed the
535  * number of tuple register stack elements anyway.  At most one extra register
536  * is required (either by dt_cg_typecast() or for dtdt_size, below).  This
537  * implies that a DIF implementation should offer a number of general purpose
538  * registers at least one greater than the number of tuple registers.
539  */
540 static void
541 dt_cg_arglist(dt_ident_t *idp, dt_node_t *args,
542     dt_irlist_t *dlp, dt_regset_t *drp)
543 {
544 	const dt_idsig_t *isp = idp->di_data;
545 	dt_node_t *dnp;
546 	int i = 0;
547 
548 	for (dnp = args; dnp != NULL; dnp = dnp->dn_list)
549 		dt_cg_node(dnp, dlp, drp);
550 
551 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, DIF_INSTR_FLUSHTS));
552 
553 	for (dnp = args; dnp != NULL; dnp = dnp->dn_list, i++) {
554 		dtrace_diftype_t t;
555 		dif_instr_t instr;
556 		uint_t op;
557 		int reg;
558 
559 		dt_node_diftype(yypcb->pcb_hdl, dnp, &t);
560 
561 		isp->dis_args[i].dn_reg = dnp->dn_reg; /* re-use register */
562 		dt_cg_typecast(dnp, &isp->dis_args[i], dlp, drp);
563 		isp->dis_args[i].dn_reg = -1;
564 
565 		if (t.dtdt_flags & DIF_TF_BYREF) {
566 			op = DIF_OP_PUSHTR;
567 			if (t.dtdt_size != 0) {
568 				reg = dt_regset_alloc(drp);
569 				dt_cg_setx(dlp, reg, t.dtdt_size);
570 			} else {
571 				reg = DIF_REG_R0;
572 			}
573 		} else {
574 			op = DIF_OP_PUSHTV;
575 			reg = DIF_REG_R0;
576 		}
577 
578 		instr = DIF_INSTR_PUSHTS(op, t.dtdt_kind, reg, dnp->dn_reg);
579 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
580 		dt_regset_free(drp, dnp->dn_reg);
581 
582 		if (reg != DIF_REG_R0)
583 			dt_regset_free(drp, reg);
584 	}
585 
586 	if (i > yypcb->pcb_hdl->dt_conf.dtc_diftupregs)
587 		longjmp(yypcb->pcb_jmpbuf, EDT_NOTUPREG);
588 }
589 
590 static void
591 dt_cg_arithmetic_op(dt_node_t *dnp, dt_irlist_t *dlp,
592     dt_regset_t *drp, uint_t op)
593 {
594 	int is_ptr_op = (dnp->dn_op == DT_TOK_ADD || dnp->dn_op == DT_TOK_SUB ||
595 	    dnp->dn_op == DT_TOK_ADD_EQ || dnp->dn_op == DT_TOK_SUB_EQ);
596 
597 	int lp_is_ptr = dt_node_is_pointer(dnp->dn_left);
598 	int rp_is_ptr = dt_node_is_pointer(dnp->dn_right);
599 
600 	dif_instr_t instr;
601 
602 	if (lp_is_ptr && rp_is_ptr) {
603 		assert(dnp->dn_op == DT_TOK_SUB);
604 		is_ptr_op = 0;
605 	}
606 
607 	dt_cg_node(dnp->dn_left, dlp, drp);
608 	if (is_ptr_op && rp_is_ptr)
609 		dt_cg_ptrsize(dnp, dlp, drp, DIF_OP_MUL, dnp->dn_left->dn_reg);
610 
611 	dt_cg_node(dnp->dn_right, dlp, drp);
612 	if (is_ptr_op && lp_is_ptr)
613 		dt_cg_ptrsize(dnp, dlp, drp, DIF_OP_MUL, dnp->dn_right->dn_reg);
614 
615 	instr = DIF_INSTR_FMT(op, dnp->dn_left->dn_reg,
616 	    dnp->dn_right->dn_reg, dnp->dn_left->dn_reg);
617 
618 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
619 	dt_regset_free(drp, dnp->dn_right->dn_reg);
620 	dnp->dn_reg = dnp->dn_left->dn_reg;
621 
622 	if (lp_is_ptr && rp_is_ptr)
623 		dt_cg_ptrsize(dnp->dn_right,
624 		    dlp, drp, DIF_OP_UDIV, dnp->dn_reg);
625 }
626 
627 static uint_t
628 dt_cg_stvar(const dt_ident_t *idp)
629 {
630 	static const uint_t aops[] = { DIF_OP_STGAA, DIF_OP_STTAA, DIF_OP_NOP };
631 	static const uint_t sops[] = { DIF_OP_STGS, DIF_OP_STTS, DIF_OP_STLS };
632 
633 	uint_t i = (((idp->di_flags & DT_IDFLG_LOCAL) != 0) << 1) |
634 	    ((idp->di_flags & DT_IDFLG_TLS) != 0);
635 
636 	return (idp->di_kind == DT_IDENT_ARRAY ? aops[i] : sops[i]);
637 }
638 
639 static void
640 dt_cg_prearith_op(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp, uint_t op)
641 {
642 	ctf_file_t *ctfp = dnp->dn_ctfp;
643 	dif_instr_t instr;
644 	ctf_id_t type;
645 	ssize_t size = 1;
646 	int reg;
647 
648 	if (dt_node_is_pointer(dnp)) {
649 		type = ctf_type_resolve(ctfp, dnp->dn_type);
650 		assert(ctf_type_kind(ctfp, type) == CTF_K_POINTER);
651 		size = ctf_type_size(ctfp, ctf_type_reference(ctfp, type));
652 	}
653 
654 	dt_cg_node(dnp->dn_child, dlp, drp);
655 	dnp->dn_reg = dnp->dn_child->dn_reg;
656 
657 	reg = dt_regset_alloc(drp);
658 	dt_cg_setx(dlp, reg, size);
659 
660 	instr = DIF_INSTR_FMT(op, dnp->dn_reg, reg, dnp->dn_reg);
661 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
662 	dt_regset_free(drp, reg);
663 
664 	/*
665 	 * If we are modifying a variable, generate an stv instruction from
666 	 * the variable specified by the identifier.  If we are storing to a
667 	 * memory address, generate code again for the left-hand side using
668 	 * DT_NF_REF to get the address, and then generate a store to it.
669 	 * In both paths, we store the value in dnp->dn_reg (the new value).
670 	 */
671 	if (dnp->dn_child->dn_kind == DT_NODE_VAR) {
672 		dt_ident_t *idp = dt_ident_resolve(dnp->dn_child->dn_ident);
673 
674 		idp->di_flags |= DT_IDFLG_DIFW;
675 		instr = DIF_INSTR_STV(dt_cg_stvar(idp),
676 		    idp->di_id, dnp->dn_reg);
677 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
678 	} else {
679 		uint_t rbit = dnp->dn_child->dn_flags & DT_NF_REF;
680 
681 		assert(dnp->dn_child->dn_flags & DT_NF_WRITABLE);
682 		assert(dnp->dn_child->dn_flags & DT_NF_LVALUE);
683 
684 		dnp->dn_child->dn_flags |= DT_NF_REF; /* force pass-by-ref */
685 		dt_cg_node(dnp->dn_child, dlp, drp);
686 
687 		dt_cg_store(dnp, dlp, drp, dnp->dn_child);
688 		dt_regset_free(drp, dnp->dn_child->dn_reg);
689 
690 		dnp->dn_left->dn_flags &= ~DT_NF_REF;
691 		dnp->dn_left->dn_flags |= rbit;
692 	}
693 }
694 
695 static void
696 dt_cg_postarith_op(dt_node_t *dnp, dt_irlist_t *dlp,
697     dt_regset_t *drp, uint_t op)
698 {
699 	ctf_file_t *ctfp = dnp->dn_ctfp;
700 	dif_instr_t instr;
701 	ctf_id_t type;
702 	ssize_t size = 1;
703 	int nreg;
704 
705 	if (dt_node_is_pointer(dnp)) {
706 		type = ctf_type_resolve(ctfp, dnp->dn_type);
707 		assert(ctf_type_kind(ctfp, type) == CTF_K_POINTER);
708 		size = ctf_type_size(ctfp, ctf_type_reference(ctfp, type));
709 	}
710 
711 	dt_cg_node(dnp->dn_child, dlp, drp);
712 	dnp->dn_reg = dnp->dn_child->dn_reg;
713 
714 	nreg = dt_regset_alloc(drp);
715 	dt_cg_setx(dlp, nreg, size);
716 	instr = DIF_INSTR_FMT(op, dnp->dn_reg, nreg, nreg);
717 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
718 
719 	/*
720 	 * If we are modifying a variable, generate an stv instruction from
721 	 * the variable specified by the identifier.  If we are storing to a
722 	 * memory address, generate code again for the left-hand side using
723 	 * DT_NF_REF to get the address, and then generate a store to it.
724 	 * In both paths, we store the value from 'nreg' (the new value).
725 	 */
726 	if (dnp->dn_child->dn_kind == DT_NODE_VAR) {
727 		dt_ident_t *idp = dt_ident_resolve(dnp->dn_child->dn_ident);
728 
729 		idp->di_flags |= DT_IDFLG_DIFW;
730 		instr = DIF_INSTR_STV(dt_cg_stvar(idp), idp->di_id, nreg);
731 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
732 	} else {
733 		uint_t rbit = dnp->dn_child->dn_flags & DT_NF_REF;
734 		int oreg = dnp->dn_reg;
735 
736 		assert(dnp->dn_child->dn_flags & DT_NF_WRITABLE);
737 		assert(dnp->dn_child->dn_flags & DT_NF_LVALUE);
738 
739 		dnp->dn_child->dn_flags |= DT_NF_REF; /* force pass-by-ref */
740 		dt_cg_node(dnp->dn_child, dlp, drp);
741 
742 		dnp->dn_reg = nreg;
743 		dt_cg_store(dnp, dlp, drp, dnp->dn_child);
744 		dnp->dn_reg = oreg;
745 
746 		dt_regset_free(drp, dnp->dn_child->dn_reg);
747 		dnp->dn_left->dn_flags &= ~DT_NF_REF;
748 		dnp->dn_left->dn_flags |= rbit;
749 	}
750 
751 	dt_regset_free(drp, nreg);
752 }
753 
754 /*
755  * Determine if we should perform signed or unsigned comparison for an OP2.
756  * If both operands are of arithmetic type, perform the usual arithmetic
757  * conversions to determine the common real type for comparison [ISOC 6.5.8.3].
758  */
759 static int
760 dt_cg_compare_signed(dt_node_t *dnp)
761 {
762 	dt_node_t dn;
763 
764 	if (dt_node_is_string(dnp->dn_left) ||
765 	    dt_node_is_string(dnp->dn_right))
766 		return (1); /* strings always compare signed */
767 	else if (!dt_node_is_arith(dnp->dn_left) ||
768 	    !dt_node_is_arith(dnp->dn_right))
769 		return (0); /* non-arithmetic types always compare unsigned */
770 
771 	bzero(&dn, sizeof (dn));
772 	dt_node_promote(dnp->dn_left, dnp->dn_right, &dn);
773 	return (dn.dn_flags & DT_NF_SIGNED);
774 }
775 
776 static void
777 dt_cg_compare_op(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp, uint_t op)
778 {
779 	uint_t lbl_true = dt_irlist_label(dlp);
780 	uint_t lbl_post = dt_irlist_label(dlp);
781 
782 	dif_instr_t instr;
783 	uint_t opc;
784 
785 	dt_cg_node(dnp->dn_left, dlp, drp);
786 	dt_cg_node(dnp->dn_right, dlp, drp);
787 
788 	if (dt_node_is_string(dnp->dn_left) || dt_node_is_string(dnp->dn_right))
789 		opc = DIF_OP_SCMP;
790 	else
791 		opc = DIF_OP_CMP;
792 
793 	instr = DIF_INSTR_CMP(opc, dnp->dn_left->dn_reg, dnp->dn_right->dn_reg);
794 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
795 	dt_regset_free(drp, dnp->dn_right->dn_reg);
796 	dnp->dn_reg = dnp->dn_left->dn_reg;
797 
798 	instr = DIF_INSTR_BRANCH(op, lbl_true);
799 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
800 
801 	instr = DIF_INSTR_MOV(DIF_REG_R0, dnp->dn_reg);
802 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
803 
804 	instr = DIF_INSTR_BRANCH(DIF_OP_BA, lbl_post);
805 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
806 
807 	dt_cg_xsetx(dlp, NULL, lbl_true, dnp->dn_reg, 1);
808 	dt_irlist_append(dlp, dt_cg_node_alloc(lbl_post, DIF_INSTR_NOP));
809 }
810 
811 /*
812  * Code generation for the ternary op requires some trickery with the assembler
813  * in order to conserve registers.  We generate code for dn_expr and dn_left
814  * and free their registers so they do not have be consumed across codegen for
815  * dn_right.  We insert a dummy MOV at the end of dn_left into the destination
816  * register, which is not yet known because we haven't done dn_right yet, and
817  * save the pointer to this instruction node.  We then generate code for
818  * dn_right and use its register as our output.  Finally, we reach back and
819  * patch the instruction for dn_left to move its output into this register.
820  */
821 static void
822 dt_cg_ternary_op(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp)
823 {
824 	uint_t lbl_false = dt_irlist_label(dlp);
825 	uint_t lbl_post = dt_irlist_label(dlp);
826 
827 	dif_instr_t instr;
828 	dt_irnode_t *dip;
829 
830 	dt_cg_node(dnp->dn_expr, dlp, drp);
831 	instr = DIF_INSTR_TST(dnp->dn_expr->dn_reg);
832 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
833 	dt_regset_free(drp, dnp->dn_expr->dn_reg);
834 
835 	instr = DIF_INSTR_BRANCH(DIF_OP_BE, lbl_false);
836 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
837 
838 	dt_cg_node(dnp->dn_left, dlp, drp);
839 	instr = DIF_INSTR_MOV(dnp->dn_left->dn_reg, DIF_REG_R0);
840 	dip = dt_cg_node_alloc(DT_LBL_NONE, instr); /* save dip for below */
841 	dt_irlist_append(dlp, dip);
842 	dt_regset_free(drp, dnp->dn_left->dn_reg);
843 
844 	instr = DIF_INSTR_BRANCH(DIF_OP_BA, lbl_post);
845 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
846 
847 	dt_irlist_append(dlp, dt_cg_node_alloc(lbl_false, DIF_INSTR_NOP));
848 	dt_cg_node(dnp->dn_right, dlp, drp);
849 	dnp->dn_reg = dnp->dn_right->dn_reg;
850 
851 	/*
852 	 * Now that dn_reg is assigned, reach back and patch the correct MOV
853 	 * instruction into the tail of dn_left.  We know dn_reg was unused
854 	 * at that point because otherwise dn_right couldn't have allocated it.
855 	 */
856 	dip->di_instr = DIF_INSTR_MOV(dnp->dn_left->dn_reg, dnp->dn_reg);
857 	dt_irlist_append(dlp, dt_cg_node_alloc(lbl_post, DIF_INSTR_NOP));
858 }
859 
860 static void
861 dt_cg_logical_and(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp)
862 {
863 	uint_t lbl_false = dt_irlist_label(dlp);
864 	uint_t lbl_post = dt_irlist_label(dlp);
865 
866 	dif_instr_t instr;
867 
868 	dt_cg_node(dnp->dn_left, dlp, drp);
869 	instr = DIF_INSTR_TST(dnp->dn_left->dn_reg);
870 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
871 	dt_regset_free(drp, dnp->dn_left->dn_reg);
872 
873 	instr = DIF_INSTR_BRANCH(DIF_OP_BE, lbl_false);
874 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
875 
876 	dt_cg_node(dnp->dn_right, dlp, drp);
877 	instr = DIF_INSTR_TST(dnp->dn_right->dn_reg);
878 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
879 	dnp->dn_reg = dnp->dn_right->dn_reg;
880 
881 	instr = DIF_INSTR_BRANCH(DIF_OP_BE, lbl_false);
882 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
883 
884 	dt_cg_setx(dlp, dnp->dn_reg, 1);
885 
886 	instr = DIF_INSTR_BRANCH(DIF_OP_BA, lbl_post);
887 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
888 
889 	instr = DIF_INSTR_MOV(DIF_REG_R0, dnp->dn_reg);
890 	dt_irlist_append(dlp, dt_cg_node_alloc(lbl_false, instr));
891 
892 	dt_irlist_append(dlp, dt_cg_node_alloc(lbl_post, DIF_INSTR_NOP));
893 }
894 
895 static void
896 dt_cg_logical_xor(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp)
897 {
898 	uint_t lbl_next = dt_irlist_label(dlp);
899 	uint_t lbl_tail = dt_irlist_label(dlp);
900 
901 	dif_instr_t instr;
902 
903 	dt_cg_node(dnp->dn_left, dlp, drp);
904 	instr = DIF_INSTR_TST(dnp->dn_left->dn_reg);
905 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
906 
907 	instr = DIF_INSTR_BRANCH(DIF_OP_BE, lbl_next);
908 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
909 	dt_cg_setx(dlp, dnp->dn_left->dn_reg, 1);
910 
911 	dt_irlist_append(dlp, dt_cg_node_alloc(lbl_next, DIF_INSTR_NOP));
912 	dt_cg_node(dnp->dn_right, dlp, drp);
913 
914 	instr = DIF_INSTR_TST(dnp->dn_right->dn_reg);
915 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
916 
917 	instr = DIF_INSTR_BRANCH(DIF_OP_BE, lbl_tail);
918 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
919 	dt_cg_setx(dlp, dnp->dn_right->dn_reg, 1);
920 
921 	instr = DIF_INSTR_FMT(DIF_OP_XOR, dnp->dn_left->dn_reg,
922 	    dnp->dn_right->dn_reg, dnp->dn_left->dn_reg);
923 
924 	dt_irlist_append(dlp, dt_cg_node_alloc(lbl_tail, instr));
925 
926 	dt_regset_free(drp, dnp->dn_right->dn_reg);
927 	dnp->dn_reg = dnp->dn_left->dn_reg;
928 }
929 
930 static void
931 dt_cg_logical_or(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp)
932 {
933 	uint_t lbl_true = dt_irlist_label(dlp);
934 	uint_t lbl_false = dt_irlist_label(dlp);
935 	uint_t lbl_post = dt_irlist_label(dlp);
936 
937 	dif_instr_t instr;
938 
939 	dt_cg_node(dnp->dn_left, dlp, drp);
940 	instr = DIF_INSTR_TST(dnp->dn_left->dn_reg);
941 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
942 	dt_regset_free(drp, dnp->dn_left->dn_reg);
943 
944 	instr = DIF_INSTR_BRANCH(DIF_OP_BNE, lbl_true);
945 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
946 
947 	dt_cg_node(dnp->dn_right, dlp, drp);
948 	instr = DIF_INSTR_TST(dnp->dn_right->dn_reg);
949 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
950 	dnp->dn_reg = dnp->dn_right->dn_reg;
951 
952 	instr = DIF_INSTR_BRANCH(DIF_OP_BE, lbl_false);
953 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
954 
955 	dt_cg_xsetx(dlp, NULL, lbl_true, dnp->dn_reg, 1);
956 
957 	instr = DIF_INSTR_BRANCH(DIF_OP_BA, lbl_post);
958 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
959 
960 	instr = DIF_INSTR_MOV(DIF_REG_R0, dnp->dn_reg);
961 	dt_irlist_append(dlp, dt_cg_node_alloc(lbl_false, instr));
962 
963 	dt_irlist_append(dlp, dt_cg_node_alloc(lbl_post, DIF_INSTR_NOP));
964 }
965 
966 static void
967 dt_cg_logical_neg(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp)
968 {
969 	uint_t lbl_zero = dt_irlist_label(dlp);
970 	uint_t lbl_post = dt_irlist_label(dlp);
971 
972 	dif_instr_t instr;
973 
974 	dt_cg_node(dnp->dn_child, dlp, drp);
975 	dnp->dn_reg = dnp->dn_child->dn_reg;
976 
977 	instr = DIF_INSTR_TST(dnp->dn_reg);
978 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
979 
980 	instr = DIF_INSTR_BRANCH(DIF_OP_BE, lbl_zero);
981 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
982 
983 	instr = DIF_INSTR_MOV(DIF_REG_R0, dnp->dn_reg);
984 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
985 
986 	instr = DIF_INSTR_BRANCH(DIF_OP_BA, lbl_post);
987 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
988 
989 	dt_cg_xsetx(dlp, NULL, lbl_zero, dnp->dn_reg, 1);
990 	dt_irlist_append(dlp, dt_cg_node_alloc(lbl_post, DIF_INSTR_NOP));
991 }
992 
993 static void
994 dt_cg_asgn_op(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp)
995 {
996 	dif_instr_t instr;
997 	dt_ident_t *idp;
998 
999 	/*
1000 	 * If we are performing a structure assignment of a translated type,
1001 	 * we must instantiate all members and create a snapshot of the object
1002 	 * in scratch space.  We allocs a chunk of memory, generate code for
1003 	 * each member, and then set dnp->dn_reg to the scratch object address.
1004 	 */
1005 	if ((idp = dt_node_resolve(dnp->dn_right, DT_IDENT_XLSOU)) != NULL) {
1006 		ctf_membinfo_t ctm;
1007 		dt_xlator_t *dxp = idp->di_data;
1008 		dt_node_t *mnp, dn, mn;
1009 		int r1, r2;
1010 
1011 		/*
1012 		 * Create two fake dt_node_t's representing operator "." and a
1013 		 * right-hand identifier child node.  These will be repeatedly
1014 		 * modified according to each instantiated member so that we
1015 		 * can pass them to dt_cg_store() and effect a member store.
1016 		 */
1017 		bzero(&dn, sizeof (dt_node_t));
1018 		dn.dn_kind = DT_NODE_OP2;
1019 		dn.dn_op = DT_TOK_DOT;
1020 		dn.dn_left = dnp;
1021 		dn.dn_right = &mn;
1022 
1023 		bzero(&mn, sizeof (dt_node_t));
1024 		mn.dn_kind = DT_NODE_IDENT;
1025 		mn.dn_op = DT_TOK_IDENT;
1026 
1027 		/*
1028 		 * Allocate a register for our scratch data pointer.  First we
1029 		 * set it to the size of our data structure, and then replace
1030 		 * it with the result of an allocs of the specified size.
1031 		 */
1032 		r1 = dt_regset_alloc(drp);
1033 		dt_cg_setx(dlp, r1,
1034 		    ctf_type_size(dxp->dx_dst_ctfp, dxp->dx_dst_base));
1035 
1036 		instr = DIF_INSTR_ALLOCS(r1, r1);
1037 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1038 
1039 		/*
1040 		 * When dt_cg_asgn_op() is called, we have already generated
1041 		 * code for dnp->dn_right, which is the translator input.  We
1042 		 * now associate this register with the translator's input
1043 		 * identifier so it can be referenced during our member loop.
1044 		 */
1045 		dxp->dx_ident->di_flags |= DT_IDFLG_CGREG;
1046 		dxp->dx_ident->di_id = dnp->dn_right->dn_reg;
1047 
1048 		for (mnp = dxp->dx_members; mnp != NULL; mnp = mnp->dn_list) {
1049 			/*
1050 			 * Generate code for the translator member expression,
1051 			 * and then cast the result to the member type.
1052 			 */
1053 			dt_cg_node(mnp->dn_membexpr, dlp, drp);
1054 			mnp->dn_reg = mnp->dn_membexpr->dn_reg;
1055 			dt_cg_typecast(mnp->dn_membexpr, mnp, dlp, drp);
1056 
1057 			/*
1058 			 * Ask CTF for the offset of the member so we can store
1059 			 * to the appropriate offset.  This call has already
1060 			 * been done once by the parser, so it should succeed.
1061 			 */
1062 			if (ctf_member_info(dxp->dx_dst_ctfp, dxp->dx_dst_base,
1063 			    mnp->dn_membname, &ctm) == CTF_ERR) {
1064 				yypcb->pcb_hdl->dt_ctferr =
1065 				    ctf_errno(dxp->dx_dst_ctfp);
1066 				longjmp(yypcb->pcb_jmpbuf, EDT_CTF);
1067 			}
1068 
1069 			/*
1070 			 * If the destination member is at offset 0, store the
1071 			 * result directly to r1 (the scratch buffer address).
1072 			 * Otherwise allocate another temporary for the offset
1073 			 * and add r1 to it before storing the result.
1074 			 */
1075 			if (ctm.ctm_offset != 0) {
1076 				r2 = dt_regset_alloc(drp);
1077 
1078 				/*
1079 				 * Add the member offset rounded down to the
1080 				 * nearest byte.  If the offset was not aligned
1081 				 * on a byte boundary, this member is a bit-
1082 				 * field and dt_cg_store() will handle masking.
1083 				 */
1084 				dt_cg_setx(dlp, r2, ctm.ctm_offset / NBBY);
1085 				instr = DIF_INSTR_FMT(DIF_OP_ADD, r1, r2, r2);
1086 				dt_irlist_append(dlp,
1087 				    dt_cg_node_alloc(DT_LBL_NONE, instr));
1088 
1089 				dt_node_type_propagate(mnp, &dn);
1090 				dn.dn_right->dn_string = mnp->dn_membname;
1091 				dn.dn_reg = r2;
1092 
1093 				dt_cg_store(mnp, dlp, drp, &dn);
1094 				dt_regset_free(drp, r2);
1095 
1096 			} else {
1097 				dt_node_type_propagate(mnp, &dn);
1098 				dn.dn_right->dn_string = mnp->dn_membname;
1099 				dn.dn_reg = r1;
1100 
1101 				dt_cg_store(mnp, dlp, drp, &dn);
1102 			}
1103 
1104 			dt_regset_free(drp, mnp->dn_reg);
1105 		}
1106 
1107 		dxp->dx_ident->di_flags &= ~DT_IDFLG_CGREG;
1108 		dxp->dx_ident->di_id = 0;
1109 
1110 		if (dnp->dn_right->dn_reg != -1)
1111 			dt_regset_free(drp, dnp->dn_right->dn_reg);
1112 
1113 		assert(dnp->dn_reg == dnp->dn_right->dn_reg);
1114 		dnp->dn_reg = r1;
1115 	}
1116 
1117 	/*
1118 	 * If we are storing to a variable, generate an stv instruction from
1119 	 * the variable specified by the identifier.  If we are storing to a
1120 	 * memory address, generate code again for the left-hand side using
1121 	 * DT_NF_REF to get the address, and then generate a store to it.
1122 	 * In both paths, we assume dnp->dn_reg already has the new value.
1123 	 */
1124 	if (dnp->dn_left->dn_kind == DT_NODE_VAR) {
1125 		idp = dt_ident_resolve(dnp->dn_left->dn_ident);
1126 
1127 		if (idp->di_kind == DT_IDENT_ARRAY)
1128 			dt_cg_arglist(idp, dnp->dn_left->dn_args, dlp, drp);
1129 
1130 		idp->di_flags |= DT_IDFLG_DIFW;
1131 		instr = DIF_INSTR_STV(dt_cg_stvar(idp),
1132 		    idp->di_id, dnp->dn_reg);
1133 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1134 	} else {
1135 		uint_t rbit = dnp->dn_left->dn_flags & DT_NF_REF;
1136 
1137 		assert(dnp->dn_left->dn_flags & DT_NF_WRITABLE);
1138 		assert(dnp->dn_left->dn_flags & DT_NF_LVALUE);
1139 
1140 		dnp->dn_left->dn_flags |= DT_NF_REF; /* force pass-by-ref */
1141 
1142 		dt_cg_node(dnp->dn_left, dlp, drp);
1143 		dt_cg_store(dnp, dlp, drp, dnp->dn_left);
1144 		dt_regset_free(drp, dnp->dn_left->dn_reg);
1145 
1146 		dnp->dn_left->dn_flags &= ~DT_NF_REF;
1147 		dnp->dn_left->dn_flags |= rbit;
1148 	}
1149 }
1150 
1151 static void
1152 dt_cg_assoc_op(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp)
1153 {
1154 	dif_instr_t instr;
1155 	uint_t op;
1156 
1157 	assert(dnp->dn_kind == DT_NODE_VAR);
1158 	assert(!(dnp->dn_ident->di_flags & DT_IDFLG_LOCAL));
1159 	assert(dnp->dn_args != NULL);
1160 
1161 	dt_cg_arglist(dnp->dn_ident, dnp->dn_args, dlp, drp);
1162 
1163 	dnp->dn_reg = dt_regset_alloc(drp);
1164 
1165 	if (dnp->dn_ident->di_flags & DT_IDFLG_TLS)
1166 		op = DIF_OP_LDTAA;
1167 	else
1168 		op = DIF_OP_LDGAA;
1169 
1170 	dnp->dn_ident->di_flags |= DT_IDFLG_DIFR;
1171 	instr = DIF_INSTR_LDV(op, dnp->dn_ident->di_id, dnp->dn_reg);
1172 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1173 
1174 	/*
1175 	 * If the associative array is a pass-by-reference type, then we are
1176 	 * loading its value as a pointer to either load or store through it.
1177 	 * The array element in question may not have been faulted in yet, in
1178 	 * which case DIF_OP_LD*AA will return zero.  We append an epilogue
1179 	 * of instructions similar to the following:
1180 	 *
1181 	 *	  ld?aa	 id, %r1	! base ld?aa instruction above
1182 	 *	  tst	 %r1		! start of epilogue
1183 	 *   +--- bne	 label
1184 	 *   |    setx	 size, %r1
1185 	 *   |    allocs %r1, %r1
1186 	 *   |    st?aa	 id, %r1
1187 	 *   |    ld?aa	 id, %r1
1188 	 *   v
1189 	 * label: < rest of code >
1190 	 *
1191 	 * The idea is that we allocs a zero-filled chunk of scratch space and
1192 	 * do a DIF_OP_ST*AA to fault in and initialize the array element, and
1193 	 * then reload it to get the faulted-in address of the new variable
1194 	 * storage.  This isn't cheap, but pass-by-ref associative array values
1195 	 * are (thus far) uncommon and the allocs cost only occurs once.  If
1196 	 * this path becomes important to DTrace users, we can improve things
1197 	 * by adding a new DIF opcode to fault in associative array elements.
1198 	 */
1199 	if (dnp->dn_flags & DT_NF_REF) {
1200 		uint_t stvop = op == DIF_OP_LDTAA ? DIF_OP_STTAA : DIF_OP_STGAA;
1201 		uint_t label = dt_irlist_label(dlp);
1202 
1203 		instr = DIF_INSTR_TST(dnp->dn_reg);
1204 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1205 
1206 		instr = DIF_INSTR_BRANCH(DIF_OP_BNE, label);
1207 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1208 
1209 		dt_cg_setx(dlp, dnp->dn_reg, dt_node_type_size(dnp));
1210 		instr = DIF_INSTR_ALLOCS(dnp->dn_reg, dnp->dn_reg);
1211 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1212 
1213 		dnp->dn_ident->di_flags |= DT_IDFLG_DIFW;
1214 		instr = DIF_INSTR_STV(stvop, dnp->dn_ident->di_id, dnp->dn_reg);
1215 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1216 
1217 		instr = DIF_INSTR_LDV(op, dnp->dn_ident->di_id, dnp->dn_reg);
1218 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1219 
1220 		dt_irlist_append(dlp, dt_cg_node_alloc(label, DIF_INSTR_NOP));
1221 	}
1222 }
1223 
1224 static void
1225 dt_cg_array_op(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp)
1226 {
1227 	dt_probe_t *prp = yypcb->pcb_probe;
1228 	uintmax_t saved = dnp->dn_args->dn_value;
1229 	dt_ident_t *idp = dnp->dn_ident;
1230 
1231 	dif_instr_t instr;
1232 	uint_t op;
1233 	size_t size;
1234 	int reg, n;
1235 
1236 	assert(dnp->dn_kind == DT_NODE_VAR);
1237 	assert(!(idp->di_flags & DT_IDFLG_LOCAL));
1238 
1239 	assert(dnp->dn_args->dn_kind == DT_NODE_INT);
1240 	assert(dnp->dn_args->dn_list == NULL);
1241 
1242 	/*
1243 	 * If this is a reference in the args[] array, temporarily modify the
1244 	 * array index according to the static argument mapping (if any),
1245 	 * unless the argument reference is provided by a dynamic translator.
1246 	 * If we're using a dynamic translator for args[], then just set dn_reg
1247 	 * to an invalid reg and return: DIF_OP_XLARG will fetch the arg later.
1248 	 */
1249 	if (idp->di_id == DIF_VAR_ARGS) {
1250 		if ((idp->di_kind == DT_IDENT_XLPTR ||
1251 		    idp->di_kind == DT_IDENT_XLSOU) &&
1252 		    dt_xlator_dynamic(idp->di_data)) {
1253 			dnp->dn_reg = -1;
1254 			return;
1255 		}
1256 		dnp->dn_args->dn_value = prp->pr_mapping[saved];
1257 	}
1258 
1259 	dt_cg_node(dnp->dn_args, dlp, drp);
1260 	dnp->dn_args->dn_value = saved;
1261 
1262 	dnp->dn_reg = dnp->dn_args->dn_reg;
1263 
1264 	if (idp->di_flags & DT_IDFLG_TLS)
1265 		op = DIF_OP_LDTA;
1266 	else
1267 		op = DIF_OP_LDGA;
1268 
1269 	idp->di_flags |= DT_IDFLG_DIFR;
1270 
1271 	instr = DIF_INSTR_LDA(op, idp->di_id,
1272 	    dnp->dn_args->dn_reg, dnp->dn_reg);
1273 
1274 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1275 
1276 	/*
1277 	 * If this is a reference to the args[] array, we need to take the
1278 	 * additional step of explicitly eliminating any bits larger than the
1279 	 * type size: the DIF interpreter in the kernel will always give us
1280 	 * the raw (64-bit) argument value, and any bits larger than the type
1281 	 * size may be junk.  As a practical matter, this arises only on 64-bit
1282 	 * architectures and only when the argument index is larger than the
1283 	 * number of arguments passed directly to DTrace: if a 8-, 16- or
1284 	 * 32-bit argument must be retrieved from the stack, it is possible
1285 	 * (and it some cases, likely) that the upper bits will be garbage.
1286 	 */
1287 	if (idp->di_id != DIF_VAR_ARGS || !dt_node_is_scalar(dnp))
1288 		return;
1289 
1290 	if ((size = dt_node_type_size(dnp)) == sizeof (uint64_t))
1291 		return;
1292 
1293 	reg = dt_regset_alloc(drp);
1294 	assert(size < sizeof (uint64_t));
1295 	n = sizeof (uint64_t) * NBBY - size * NBBY;
1296 
1297 	dt_cg_setx(dlp, reg, n);
1298 
1299 	instr = DIF_INSTR_FMT(DIF_OP_SLL, dnp->dn_reg, reg, dnp->dn_reg);
1300 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1301 
1302 	instr = DIF_INSTR_FMT((dnp->dn_flags & DT_NF_SIGNED) ?
1303 	    DIF_OP_SRA : DIF_OP_SRL, dnp->dn_reg, reg, dnp->dn_reg);
1304 
1305 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1306 	dt_regset_free(drp, reg);
1307 }
1308 
1309 /*
1310  * Generate code for an inlined variable reference.  Inlines can be used to
1311  * define either scalar or associative array substitutions.  For scalars, we
1312  * simply generate code for the parse tree saved in the identifier's din_root,
1313  * and then cast the resulting expression to the inline's declaration type.
1314  * For arrays, we take the input parameter subtrees from dnp->dn_args and
1315  * temporarily store them in the din_root of each din_argv[i] identifier,
1316  * which are themselves inlines and were set up for us by the parser.  The
1317  * result is that any reference to the inlined parameter inside the top-level
1318  * din_root will turn into a recursive call to dt_cg_inline() for a scalar
1319  * inline whose din_root will refer to the subtree pointed to by the argument.
1320  */
1321 static void
1322 dt_cg_inline(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp)
1323 {
1324 	dt_ident_t *idp = dnp->dn_ident;
1325 	dt_idnode_t *inp = idp->di_iarg;
1326 
1327 	dt_idnode_t *pinp;
1328 	dt_node_t *pnp;
1329 	int i;
1330 
1331 	assert(idp->di_flags & DT_IDFLG_INLINE);
1332 	assert(idp->di_ops == &dt_idops_inline);
1333 
1334 	if (idp->di_kind == DT_IDENT_ARRAY) {
1335 		for (i = 0, pnp = dnp->dn_args;
1336 		    pnp != NULL; pnp = pnp->dn_list, i++) {
1337 			if (inp->din_argv[i] != NULL) {
1338 				pinp = inp->din_argv[i]->di_iarg;
1339 				pinp->din_root = pnp;
1340 			}
1341 		}
1342 	}
1343 
1344 	dt_cg_node(inp->din_root, dlp, drp);
1345 	dnp->dn_reg = inp->din_root->dn_reg;
1346 	dt_cg_typecast(inp->din_root, dnp, dlp, drp);
1347 
1348 	if (idp->di_kind == DT_IDENT_ARRAY) {
1349 		for (i = 0; i < inp->din_argc; i++) {
1350 			pinp = inp->din_argv[i]->di_iarg;
1351 			pinp->din_root = NULL;
1352 		}
1353 	}
1354 }
1355 
1356 static void
1357 dt_cg_func_typeref(dtrace_hdl_t *dtp, dt_node_t *dnp)
1358 {
1359 	dtrace_typeinfo_t dtt;
1360 	dt_node_t *addr = dnp->dn_args;
1361 	dt_node_t *nelm = addr->dn_list;
1362 	dt_node_t *strp = nelm->dn_list;
1363 	dt_node_t *typs = strp->dn_list;
1364 	char buf[DT_TYPE_NAMELEN];
1365 	char *p;
1366 
1367 	ctf_type_name(addr->dn_ctfp, addr->dn_type, buf, sizeof (buf));
1368 
1369 	/*
1370 	 * XXX Hack alert! XXX
1371 	 * The prototype has two dummy args that we munge to represent
1372 	 * the type string and the type size.
1373 	 *
1374 	 * Yes, I hear your grumble, but it works for now. We'll come
1375 	 * up with a more elegant implementation later. :-)
1376 	 */
1377 	free(strp->dn_string);
1378 
1379 	if ((p = strchr(buf, '*')) != NULL)
1380 		*p = '\0';
1381 
1382 	strp->dn_string = strdup(buf);
1383 
1384 	if (dtrace_lookup_by_type(dtp,  DTRACE_OBJ_EVERY, buf, &dtt) < 0)
1385 		return;
1386 
1387 	typs->dn_value = ctf_type_size(dtt.dtt_ctfp, dtt.dtt_type);
1388 }
1389 
1390 typedef struct dt_xlmemb {
1391 	dt_ident_t *dtxl_idp;		/* translated ident */
1392 	dt_irlist_t *dtxl_dlp;		/* instruction list */
1393 	dt_regset_t *dtxl_drp;		/* register set */
1394 	int dtxl_sreg;			/* location of the translation input */
1395 	int dtxl_dreg;			/* location of our allocated buffer */
1396 } dt_xlmemb_t;
1397 
1398 /*ARGSUSED*/
1399 static int
1400 dt_cg_xlate_member(const char *name, ctf_id_t type, ulong_t off, void *arg)
1401 {
1402 	dt_xlmemb_t *dx = arg;
1403 	dt_ident_t *idp = dx->dtxl_idp;
1404 	dt_irlist_t *dlp = dx->dtxl_dlp;
1405 	dt_regset_t *drp = dx->dtxl_drp;
1406 
1407 	dt_node_t *mnp;
1408 	dt_xlator_t *dxp;
1409 
1410 	int reg, treg;
1411 	uint32_t instr;
1412 	size_t size;
1413 
1414 	/* Generate code for the translation. */
1415 	dxp = idp->di_data;
1416 	mnp = dt_xlator_member(dxp, name);
1417 
1418 	/* If there's no translator for the given member, skip it. */
1419 	if (mnp == NULL)
1420 		return (0);
1421 
1422 	dxp->dx_ident->di_flags |= DT_IDFLG_CGREG;
1423 	dxp->dx_ident->di_id = dx->dtxl_sreg;
1424 
1425 	dt_cg_node(mnp->dn_membexpr, dlp, drp);
1426 
1427 	dxp->dx_ident->di_flags &= ~DT_IDFLG_CGREG;
1428 	dxp->dx_ident->di_id = 0;
1429 
1430 	treg = mnp->dn_membexpr->dn_reg;
1431 
1432 	/* Compute the offset into our buffer and store the result there. */
1433 	reg = dt_regset_alloc(drp);
1434 
1435 	dt_cg_setx(dlp, reg, off / NBBY);
1436 	instr = DIF_INSTR_FMT(DIF_OP_ADD, dx->dtxl_dreg, reg, reg);
1437 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1438 
1439 	size = ctf_type_size(mnp->dn_membexpr->dn_ctfp,
1440 	    mnp->dn_membexpr->dn_type);
1441 	if (dt_node_is_scalar(mnp->dn_membexpr)) {
1442 		/*
1443 		 * Copying scalars is simple.
1444 		 */
1445 		switch (size) {
1446 		case 1:
1447 			instr = DIF_INSTR_STORE(DIF_OP_STB, treg, reg);
1448 			break;
1449 		case 2:
1450 			instr = DIF_INSTR_STORE(DIF_OP_STH, treg, reg);
1451 			break;
1452 		case 4:
1453 			instr = DIF_INSTR_STORE(DIF_OP_STW, treg, reg);
1454 			break;
1455 		case 8:
1456 			instr = DIF_INSTR_STORE(DIF_OP_STX, treg, reg);
1457 			break;
1458 		default:
1459 			xyerror(D_UNKNOWN, "internal error -- unexpected "
1460 			    "size: %lu\n", (ulong_t)size);
1461 		}
1462 
1463 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1464 
1465 	} else if (dt_node_is_string(mnp->dn_membexpr)) {
1466 		int szreg;
1467 
1468 		/*
1469 		 * Use the copys instruction for strings.
1470 		 */
1471 		szreg = dt_regset_alloc(drp);
1472 		dt_cg_setx(dlp, szreg, size);
1473 		instr = DIF_INSTR_COPYS(treg, szreg, reg);
1474 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1475 		dt_regset_free(drp, szreg);
1476 	} else {
1477 		int szreg;
1478 
1479 		/*
1480 		 * If it's anything else then we'll just bcopy it.
1481 		 */
1482 		szreg = dt_regset_alloc(drp);
1483 		dt_cg_setx(dlp, szreg, size);
1484 		dt_irlist_append(dlp,
1485 		    dt_cg_node_alloc(DT_LBL_NONE, DIF_INSTR_FLUSHTS));
1486 		instr = DIF_INSTR_PUSHTS(DIF_OP_PUSHTV, DIF_TYPE_CTF,
1487 		    DIF_REG_R0, treg);
1488 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1489 		instr = DIF_INSTR_PUSHTS(DIF_OP_PUSHTV, DIF_TYPE_CTF,
1490 		    DIF_REG_R0, reg);
1491 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1492 		instr = DIF_INSTR_PUSHTS(DIF_OP_PUSHTV, DIF_TYPE_CTF,
1493 		    DIF_REG_R0, szreg);
1494 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1495 		instr = DIF_INSTR_CALL(DIF_SUBR_BCOPY, szreg);
1496 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1497 		dt_regset_free(drp, szreg);
1498 	}
1499 
1500 	dt_regset_free(drp, reg);
1501 	dt_regset_free(drp, treg);
1502 
1503 	return (0);
1504 }
1505 
1506 /*
1507  * If we're expanding a translated type, we create an appropriately sized
1508  * buffer with alloca() and then translate each member into it.
1509  */
1510 static int
1511 dt_cg_xlate_expand(dt_node_t *dnp, dt_ident_t *idp, dt_irlist_t *dlp,
1512     dt_regset_t *drp)
1513 {
1514 	dt_xlmemb_t dlm;
1515 	uint32_t instr;
1516 	int dreg;
1517 	size_t size;
1518 
1519 	dreg = dt_regset_alloc(drp);
1520 	size = ctf_type_size(dnp->dn_ident->di_ctfp, dnp->dn_ident->di_type);
1521 
1522 	/* Call alloca() to create the buffer. */
1523 	dt_cg_setx(dlp, dreg, size);
1524 
1525 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, DIF_INSTR_FLUSHTS));
1526 
1527 	instr = DIF_INSTR_PUSHTS(DIF_OP_PUSHTV, DIF_TYPE_CTF, DIF_REG_R0, dreg);
1528 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1529 
1530 	instr = DIF_INSTR_CALL(DIF_SUBR_ALLOCA, dreg);
1531 	dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1532 
1533 	/* Generate the translation for each member. */
1534 	dlm.dtxl_idp = idp;
1535 	dlm.dtxl_dlp = dlp;
1536 	dlm.dtxl_drp = drp;
1537 	dlm.dtxl_sreg = dnp->dn_reg;
1538 	dlm.dtxl_dreg = dreg;
1539 	(void) ctf_member_iter(dnp->dn_ident->di_ctfp,
1540 	    dnp->dn_ident->di_type, dt_cg_xlate_member,
1541 	    &dlm);
1542 
1543 	return (dreg);
1544 }
1545 
1546 static void
1547 dt_cg_node(dt_node_t *dnp, dt_irlist_t *dlp, dt_regset_t *drp)
1548 {
1549 	ctf_file_t *ctfp = dnp->dn_ctfp;
1550 	ctf_file_t *octfp;
1551 	ctf_membinfo_t m;
1552 	ctf_id_t type;
1553 
1554 	dif_instr_t instr;
1555 	dt_ident_t *idp;
1556 	ssize_t stroff;
1557 	uint_t op;
1558 
1559 	switch (dnp->dn_op) {
1560 	case DT_TOK_COMMA:
1561 		dt_cg_node(dnp->dn_left, dlp, drp);
1562 		dt_regset_free(drp, dnp->dn_left->dn_reg);
1563 		dt_cg_node(dnp->dn_right, dlp, drp);
1564 		dnp->dn_reg = dnp->dn_right->dn_reg;
1565 		break;
1566 
1567 	case DT_TOK_ASGN:
1568 		dt_cg_node(dnp->dn_right, dlp, drp);
1569 		dnp->dn_reg = dnp->dn_right->dn_reg;
1570 		dt_cg_asgn_op(dnp, dlp, drp);
1571 		break;
1572 
1573 	case DT_TOK_ADD_EQ:
1574 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_ADD);
1575 		dt_cg_asgn_op(dnp, dlp, drp);
1576 		break;
1577 
1578 	case DT_TOK_SUB_EQ:
1579 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_SUB);
1580 		dt_cg_asgn_op(dnp, dlp, drp);
1581 		break;
1582 
1583 	case DT_TOK_MUL_EQ:
1584 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_MUL);
1585 		dt_cg_asgn_op(dnp, dlp, drp);
1586 		break;
1587 
1588 	case DT_TOK_DIV_EQ:
1589 		dt_cg_arithmetic_op(dnp, dlp, drp,
1590 		    (dnp->dn_flags & DT_NF_SIGNED) ? DIF_OP_SDIV : DIF_OP_UDIV);
1591 		dt_cg_asgn_op(dnp, dlp, drp);
1592 		break;
1593 
1594 	case DT_TOK_MOD_EQ:
1595 		dt_cg_arithmetic_op(dnp, dlp, drp,
1596 		    (dnp->dn_flags & DT_NF_SIGNED) ? DIF_OP_SREM : DIF_OP_UREM);
1597 		dt_cg_asgn_op(dnp, dlp, drp);
1598 		break;
1599 
1600 	case DT_TOK_AND_EQ:
1601 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_AND);
1602 		dt_cg_asgn_op(dnp, dlp, drp);
1603 		break;
1604 
1605 	case DT_TOK_XOR_EQ:
1606 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_XOR);
1607 		dt_cg_asgn_op(dnp, dlp, drp);
1608 		break;
1609 
1610 	case DT_TOK_OR_EQ:
1611 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_OR);
1612 		dt_cg_asgn_op(dnp, dlp, drp);
1613 		break;
1614 
1615 	case DT_TOK_LSH_EQ:
1616 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_SLL);
1617 		dt_cg_asgn_op(dnp, dlp, drp);
1618 		break;
1619 
1620 	case DT_TOK_RSH_EQ:
1621 		dt_cg_arithmetic_op(dnp, dlp, drp,
1622 		    (dnp->dn_flags & DT_NF_SIGNED) ? DIF_OP_SRA : DIF_OP_SRL);
1623 		dt_cg_asgn_op(dnp, dlp, drp);
1624 		break;
1625 
1626 	case DT_TOK_QUESTION:
1627 		dt_cg_ternary_op(dnp, dlp, drp);
1628 		break;
1629 
1630 	case DT_TOK_LOR:
1631 		dt_cg_logical_or(dnp, dlp, drp);
1632 		break;
1633 
1634 	case DT_TOK_LXOR:
1635 		dt_cg_logical_xor(dnp, dlp, drp);
1636 		break;
1637 
1638 	case DT_TOK_LAND:
1639 		dt_cg_logical_and(dnp, dlp, drp);
1640 		break;
1641 
1642 	case DT_TOK_BOR:
1643 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_OR);
1644 		break;
1645 
1646 	case DT_TOK_XOR:
1647 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_XOR);
1648 		break;
1649 
1650 	case DT_TOK_BAND:
1651 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_AND);
1652 		break;
1653 
1654 	case DT_TOK_EQU:
1655 		dt_cg_compare_op(dnp, dlp, drp, DIF_OP_BE);
1656 		break;
1657 
1658 	case DT_TOK_NEQ:
1659 		dt_cg_compare_op(dnp, dlp, drp, DIF_OP_BNE);
1660 		break;
1661 
1662 	case DT_TOK_LT:
1663 		dt_cg_compare_op(dnp, dlp, drp,
1664 		    dt_cg_compare_signed(dnp) ? DIF_OP_BL : DIF_OP_BLU);
1665 		break;
1666 
1667 	case DT_TOK_LE:
1668 		dt_cg_compare_op(dnp, dlp, drp,
1669 		    dt_cg_compare_signed(dnp) ? DIF_OP_BLE : DIF_OP_BLEU);
1670 		break;
1671 
1672 	case DT_TOK_GT:
1673 		dt_cg_compare_op(dnp, dlp, drp,
1674 		    dt_cg_compare_signed(dnp) ? DIF_OP_BG : DIF_OP_BGU);
1675 		break;
1676 
1677 	case DT_TOK_GE:
1678 		dt_cg_compare_op(dnp, dlp, drp,
1679 		    dt_cg_compare_signed(dnp) ? DIF_OP_BGE : DIF_OP_BGEU);
1680 		break;
1681 
1682 	case DT_TOK_LSH:
1683 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_SLL);
1684 		break;
1685 
1686 	case DT_TOK_RSH:
1687 		dt_cg_arithmetic_op(dnp, dlp, drp,
1688 		    (dnp->dn_flags & DT_NF_SIGNED) ? DIF_OP_SRA : DIF_OP_SRL);
1689 		break;
1690 
1691 	case DT_TOK_ADD:
1692 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_ADD);
1693 		break;
1694 
1695 	case DT_TOK_SUB:
1696 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_SUB);
1697 		break;
1698 
1699 	case DT_TOK_MUL:
1700 		dt_cg_arithmetic_op(dnp, dlp, drp, DIF_OP_MUL);
1701 		break;
1702 
1703 	case DT_TOK_DIV:
1704 		dt_cg_arithmetic_op(dnp, dlp, drp,
1705 		    (dnp->dn_flags & DT_NF_SIGNED) ? DIF_OP_SDIV : DIF_OP_UDIV);
1706 		break;
1707 
1708 	case DT_TOK_MOD:
1709 		dt_cg_arithmetic_op(dnp, dlp, drp,
1710 		    (dnp->dn_flags & DT_NF_SIGNED) ? DIF_OP_SREM : DIF_OP_UREM);
1711 		break;
1712 
1713 	case DT_TOK_LNEG:
1714 		dt_cg_logical_neg(dnp, dlp, drp);
1715 		break;
1716 
1717 	case DT_TOK_BNEG:
1718 		dt_cg_node(dnp->dn_child, dlp, drp);
1719 		dnp->dn_reg = dnp->dn_child->dn_reg;
1720 		instr = DIF_INSTR_NOT(dnp->dn_reg, dnp->dn_reg);
1721 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1722 		break;
1723 
1724 	case DT_TOK_PREINC:
1725 		dt_cg_prearith_op(dnp, dlp, drp, DIF_OP_ADD);
1726 		break;
1727 
1728 	case DT_TOK_POSTINC:
1729 		dt_cg_postarith_op(dnp, dlp, drp, DIF_OP_ADD);
1730 		break;
1731 
1732 	case DT_TOK_PREDEC:
1733 		dt_cg_prearith_op(dnp, dlp, drp, DIF_OP_SUB);
1734 		break;
1735 
1736 	case DT_TOK_POSTDEC:
1737 		dt_cg_postarith_op(dnp, dlp, drp, DIF_OP_SUB);
1738 		break;
1739 
1740 	case DT_TOK_IPOS:
1741 		dt_cg_node(dnp->dn_child, dlp, drp);
1742 		dnp->dn_reg = dnp->dn_child->dn_reg;
1743 		break;
1744 
1745 	case DT_TOK_INEG:
1746 		dt_cg_node(dnp->dn_child, dlp, drp);
1747 		dnp->dn_reg = dnp->dn_child->dn_reg;
1748 
1749 		instr = DIF_INSTR_FMT(DIF_OP_SUB, DIF_REG_R0,
1750 		    dnp->dn_reg, dnp->dn_reg);
1751 
1752 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1753 		break;
1754 
1755 	case DT_TOK_DEREF:
1756 		dt_cg_node(dnp->dn_child, dlp, drp);
1757 		dnp->dn_reg = dnp->dn_child->dn_reg;
1758 
1759 		if (dt_node_is_dynamic(dnp->dn_child)) {
1760 			int reg;
1761 			idp = dt_node_resolve(dnp->dn_child, DT_IDENT_XLPTR);
1762 			assert(idp != NULL);
1763 			reg = dt_cg_xlate_expand(dnp, idp, dlp, drp);
1764 
1765 			dt_regset_free(drp, dnp->dn_child->dn_reg);
1766 			dnp->dn_reg = reg;
1767 
1768 		} else if (!(dnp->dn_flags & DT_NF_REF)) {
1769 			uint_t ubit = dnp->dn_flags & DT_NF_USERLAND;
1770 
1771 			/*
1772 			 * Save and restore DT_NF_USERLAND across dt_cg_load():
1773 			 * we need the sign bit from dnp and the user bit from
1774 			 * dnp->dn_child in order to get the proper opcode.
1775 			 */
1776 			dnp->dn_flags |=
1777 			    (dnp->dn_child->dn_flags & DT_NF_USERLAND);
1778 
1779 			instr = DIF_INSTR_LOAD(dt_cg_load(dnp, ctfp,
1780 			    dnp->dn_type), dnp->dn_reg, dnp->dn_reg);
1781 
1782 			dnp->dn_flags &= ~DT_NF_USERLAND;
1783 			dnp->dn_flags |= ubit;
1784 
1785 			dt_irlist_append(dlp,
1786 			    dt_cg_node_alloc(DT_LBL_NONE, instr));
1787 		}
1788 		break;
1789 
1790 	case DT_TOK_ADDROF: {
1791 		uint_t rbit = dnp->dn_child->dn_flags & DT_NF_REF;
1792 
1793 		dnp->dn_child->dn_flags |= DT_NF_REF; /* force pass-by-ref */
1794 		dt_cg_node(dnp->dn_child, dlp, drp);
1795 		dnp->dn_reg = dnp->dn_child->dn_reg;
1796 
1797 		dnp->dn_child->dn_flags &= ~DT_NF_REF;
1798 		dnp->dn_child->dn_flags |= rbit;
1799 		break;
1800 	}
1801 
1802 	case DT_TOK_SIZEOF: {
1803 		size_t size = dt_node_sizeof(dnp->dn_child);
1804 		dnp->dn_reg = dt_regset_alloc(drp);
1805 		assert(size != 0);
1806 		dt_cg_setx(dlp, dnp->dn_reg, size);
1807 		break;
1808 	}
1809 
1810 	case DT_TOK_STRINGOF:
1811 		dt_cg_node(dnp->dn_child, dlp, drp);
1812 		dnp->dn_reg = dnp->dn_child->dn_reg;
1813 		break;
1814 
1815 	case DT_TOK_XLATE:
1816 		/*
1817 		 * An xlate operator appears in either an XLATOR, indicating a
1818 		 * reference to a dynamic translator, or an OP2, indicating
1819 		 * use of the xlate operator in the user's program.  For the
1820 		 * dynamic case, generate an xlate opcode with a reference to
1821 		 * the corresponding member, pre-computed for us in dn_members.
1822 		 */
1823 		if (dnp->dn_kind == DT_NODE_XLATOR) {
1824 			dt_xlator_t *dxp = dnp->dn_xlator;
1825 
1826 			assert(dxp->dx_ident->di_flags & DT_IDFLG_CGREG);
1827 			assert(dxp->dx_ident->di_id != 0);
1828 
1829 			dnp->dn_reg = dt_regset_alloc(drp);
1830 
1831 			if (dxp->dx_arg == -1) {
1832 				instr = DIF_INSTR_MOV(
1833 				    dxp->dx_ident->di_id, dnp->dn_reg);
1834 				dt_irlist_append(dlp,
1835 				    dt_cg_node_alloc(DT_LBL_NONE, instr));
1836 				op = DIF_OP_XLATE;
1837 			} else
1838 				op = DIF_OP_XLARG;
1839 
1840 			instr = DIF_INSTR_XLATE(op, 0, dnp->dn_reg);
1841 			dt_irlist_append(dlp,
1842 			    dt_cg_node_alloc(DT_LBL_NONE, instr));
1843 
1844 			dlp->dl_last->di_extern = dnp->dn_xmember;
1845 			break;
1846 		}
1847 
1848 		assert(dnp->dn_kind == DT_NODE_OP2);
1849 		dt_cg_node(dnp->dn_right, dlp, drp);
1850 		dnp->dn_reg = dnp->dn_right->dn_reg;
1851 		break;
1852 
1853 	case DT_TOK_LPAR:
1854 		dt_cg_node(dnp->dn_right, dlp, drp);
1855 		dnp->dn_reg = dnp->dn_right->dn_reg;
1856 		dt_cg_typecast(dnp->dn_right, dnp, dlp, drp);
1857 		break;
1858 
1859 	case DT_TOK_PTR:
1860 	case DT_TOK_DOT:
1861 		assert(dnp->dn_right->dn_kind == DT_NODE_IDENT);
1862 		dt_cg_node(dnp->dn_left, dlp, drp);
1863 
1864 		/*
1865 		 * If the left-hand side of PTR or DOT is a dynamic variable,
1866 		 * we expect it to be the output of a D translator.   In this
1867 		 * case, we look up the parse tree corresponding to the member
1868 		 * that is being accessed and run the code generator over it.
1869 		 * We then cast the result as if by the assignment operator.
1870 		 */
1871 		if ((idp = dt_node_resolve(
1872 		    dnp->dn_left, DT_IDENT_XLSOU)) != NULL ||
1873 		    (idp = dt_node_resolve(
1874 		    dnp->dn_left, DT_IDENT_XLPTR)) != NULL) {
1875 
1876 			dt_xlator_t *dxp;
1877 			dt_node_t *mnp;
1878 
1879 			dxp = idp->di_data;
1880 			mnp = dt_xlator_member(dxp, dnp->dn_right->dn_string);
1881 			assert(mnp != NULL);
1882 
1883 			dxp->dx_ident->di_flags |= DT_IDFLG_CGREG;
1884 			dxp->dx_ident->di_id = dnp->dn_left->dn_reg;
1885 
1886 			dt_cg_node(mnp->dn_membexpr, dlp, drp);
1887 			dnp->dn_reg = mnp->dn_membexpr->dn_reg;
1888 			dt_cg_typecast(mnp->dn_membexpr, dnp, dlp, drp);
1889 
1890 			dxp->dx_ident->di_flags &= ~DT_IDFLG_CGREG;
1891 			dxp->dx_ident->di_id = 0;
1892 
1893 			if (dnp->dn_left->dn_reg != -1)
1894 				dt_regset_free(drp, dnp->dn_left->dn_reg);
1895 			break;
1896 		}
1897 
1898 		ctfp = dnp->dn_left->dn_ctfp;
1899 		type = ctf_type_resolve(ctfp, dnp->dn_left->dn_type);
1900 
1901 		if (dnp->dn_op == DT_TOK_PTR) {
1902 			type = ctf_type_reference(ctfp, type);
1903 			type = ctf_type_resolve(ctfp, type);
1904 		}
1905 
1906 		if ((ctfp = dt_cg_membinfo(octfp = ctfp, type,
1907 		    dnp->dn_right->dn_string, &m)) == NULL) {
1908 			yypcb->pcb_hdl->dt_ctferr = ctf_errno(octfp);
1909 			longjmp(yypcb->pcb_jmpbuf, EDT_CTF);
1910 		}
1911 
1912 		if (m.ctm_offset != 0) {
1913 			int reg;
1914 
1915 			reg = dt_regset_alloc(drp);
1916 
1917 			/*
1918 			 * If the offset is not aligned on a byte boundary, it
1919 			 * is a bit-field member and we will extract the value
1920 			 * bits below after we generate the appropriate load.
1921 			 */
1922 			dt_cg_setx(dlp, reg, m.ctm_offset / NBBY);
1923 
1924 			instr = DIF_INSTR_FMT(DIF_OP_ADD,
1925 			    dnp->dn_left->dn_reg, reg, dnp->dn_left->dn_reg);
1926 
1927 			dt_irlist_append(dlp,
1928 			    dt_cg_node_alloc(DT_LBL_NONE, instr));
1929 			dt_regset_free(drp, reg);
1930 		}
1931 
1932 		if (!(dnp->dn_flags & DT_NF_REF)) {
1933 			uint_t ubit = dnp->dn_flags & DT_NF_USERLAND;
1934 
1935 			/*
1936 			 * Save and restore DT_NF_USERLAND across dt_cg_load():
1937 			 * we need the sign bit from dnp and the user bit from
1938 			 * dnp->dn_left in order to get the proper opcode.
1939 			 */
1940 			dnp->dn_flags |=
1941 			    (dnp->dn_left->dn_flags & DT_NF_USERLAND);
1942 
1943 			instr = DIF_INSTR_LOAD(dt_cg_load(dnp,
1944 			    ctfp, m.ctm_type), dnp->dn_left->dn_reg,
1945 			    dnp->dn_left->dn_reg);
1946 
1947 			dnp->dn_flags &= ~DT_NF_USERLAND;
1948 			dnp->dn_flags |= ubit;
1949 
1950 			dt_irlist_append(dlp,
1951 			    dt_cg_node_alloc(DT_LBL_NONE, instr));
1952 
1953 			if (dnp->dn_flags & DT_NF_BITFIELD)
1954 				dt_cg_field_get(dnp, dlp, drp, ctfp, &m);
1955 		}
1956 
1957 		dnp->dn_reg = dnp->dn_left->dn_reg;
1958 		break;
1959 
1960 	case DT_TOK_STRING:
1961 		dnp->dn_reg = dt_regset_alloc(drp);
1962 
1963 		assert(dnp->dn_kind == DT_NODE_STRING);
1964 		stroff = dt_strtab_insert(yypcb->pcb_strtab, dnp->dn_string);
1965 
1966 		if (stroff == -1L)
1967 			longjmp(yypcb->pcb_jmpbuf, EDT_NOMEM);
1968 		if (stroff > DIF_STROFF_MAX)
1969 			longjmp(yypcb->pcb_jmpbuf, EDT_STR2BIG);
1970 
1971 		instr = DIF_INSTR_SETS((ulong_t)stroff, dnp->dn_reg);
1972 		dt_irlist_append(dlp, dt_cg_node_alloc(DT_LBL_NONE, instr));
1973 		break;
1974 
1975 	case DT_TOK_IDENT:
1976 		/*
1977 		 * If the specified identifier is a variable on which we have
1978 		 * set the code generator register flag, then this variable
1979 		 * has already had code generated for it and saved in di_id.
1980 		 * Allocate a new register and copy the existing value to it.
1981 		 */
1982 		if (dnp->dn_kind == DT_NODE_VAR &&
1983 		    (dnp->dn_ident->di_flags & DT_IDFLG_CGREG)) {
1984 			dnp->dn_reg = dt_regset_alloc(drp);
1985 			instr = DIF_INSTR_MOV(dnp->dn_ident->di_id,
1986 			    dnp->dn_reg);
1987 			dt_irlist_append(dlp,
1988 			    dt_cg_node_alloc(DT_LBL_NONE, instr));
1989 			break;
1990 		}
1991 
1992 		/*
1993 		 * Identifiers can represent function calls, variable refs, or
1994 		 * symbols.  First we check for inlined variables, and handle
1995 		 * them by generating code for the inline parse tree.
1996 		 */
1997 		if (dnp->dn_kind == DT_NODE_VAR &&
1998 		    (dnp->dn_ident->di_flags & DT_IDFLG_INLINE)) {
1999 			dt_cg_inline(dnp, dlp, drp);
2000 			break;
2001 		}
2002 
2003 		switch (dnp->dn_kind) {
2004 		case DT_NODE_FUNC: {
2005 			dtrace_hdl_t *dtp = yypcb->pcb_hdl;
2006 
2007 			if ((idp = dnp->dn_ident)->di_kind != DT_IDENT_FUNC) {
2008 				dnerror(dnp, D_CG_EXPR, "%s %s( ) may not be "
2009 				    "called from a D expression (D program "
2010 				    "context required)\n",
2011 				    dt_idkind_name(idp->di_kind), idp->di_name);
2012 			}
2013 
2014 			switch (idp->di_id) {
2015 			case DIF_SUBR_TYPEREF:
2016 				dt_cg_func_typeref(dtp, dnp);
2017 				break;
2018 
2019 			default:
2020 				break;
2021 			}
2022 
2023 			dt_cg_arglist(dnp->dn_ident, dnp->dn_args, dlp, drp);
2024 
2025 			dnp->dn_reg = dt_regset_alloc(drp);
2026 			instr = DIF_INSTR_CALL(dnp->dn_ident->di_id,
2027 			    dnp->dn_reg);
2028 
2029 			dt_irlist_append(dlp,
2030 			    dt_cg_node_alloc(DT_LBL_NONE, instr));
2031 
2032 			break;
2033 		}
2034 
2035 		case DT_NODE_VAR:
2036 			if (dnp->dn_ident->di_kind == DT_IDENT_XLSOU ||
2037 			    dnp->dn_ident->di_kind == DT_IDENT_XLPTR) {
2038 				/*
2039 				 * This can only happen if we have translated
2040 				 * args[].  See dt_idcook_args() for details.
2041 				 */
2042 				assert(dnp->dn_ident->di_id == DIF_VAR_ARGS);
2043 				dt_cg_array_op(dnp, dlp, drp);
2044 				break;
2045 			}
2046 
2047 			if (dnp->dn_ident->di_kind == DT_IDENT_ARRAY) {
2048 				if (dnp->dn_ident->di_id > DIF_VAR_ARRAY_MAX)
2049 					dt_cg_assoc_op(dnp, dlp, drp);
2050 				else
2051 					dt_cg_array_op(dnp, dlp, drp);
2052 				break;
2053 			}
2054 
2055 			dnp->dn_reg = dt_regset_alloc(drp);
2056 
2057 			if (dnp->dn_ident->di_flags & DT_IDFLG_LOCAL)
2058 				op = DIF_OP_LDLS;
2059 			else if (dnp->dn_ident->di_flags & DT_IDFLG_TLS)
2060 				op = DIF_OP_LDTS;
2061 			else
2062 				op = DIF_OP_LDGS;
2063 
2064 			dnp->dn_ident->di_flags |= DT_IDFLG_DIFR;
2065 
2066 			instr = DIF_INSTR_LDV(op,
2067 			    dnp->dn_ident->di_id, dnp->dn_reg);
2068 
2069 			dt_irlist_append(dlp,
2070 			    dt_cg_node_alloc(DT_LBL_NONE, instr));
2071 			break;
2072 
2073 		case DT_NODE_SYM: {
2074 			dtrace_hdl_t *dtp = yypcb->pcb_hdl;
2075 			dtrace_syminfo_t *sip = dnp->dn_ident->di_data;
2076 			GElf_Sym sym;
2077 
2078 			if (dtrace_lookup_by_name(dtp,
2079 			    sip->dts_object, sip->dts_name, &sym, NULL) == -1) {
2080 				xyerror(D_UNKNOWN, "cg failed for symbol %s`%s:"
2081 				    " %s\n", sip->dts_object, sip->dts_name,
2082 				    dtrace_errmsg(dtp, dtrace_errno(dtp)));
2083 			}
2084 
2085 			dnp->dn_reg = dt_regset_alloc(drp);
2086 			dt_cg_xsetx(dlp, dnp->dn_ident,
2087 			    DT_LBL_NONE, dnp->dn_reg, sym.st_value);
2088 
2089 			if (!(dnp->dn_flags & DT_NF_REF)) {
2090 				instr = DIF_INSTR_LOAD(dt_cg_load(dnp, ctfp,
2091 				    dnp->dn_type), dnp->dn_reg, dnp->dn_reg);
2092 				dt_irlist_append(dlp,
2093 				    dt_cg_node_alloc(DT_LBL_NONE, instr));
2094 			}
2095 			break;
2096 		}
2097 
2098 		default:
2099 			xyerror(D_UNKNOWN, "internal error -- node type %u is "
2100 			    "not valid for an identifier\n", dnp->dn_kind);
2101 		}
2102 		break;
2103 
2104 	case DT_TOK_INT:
2105 		dnp->dn_reg = dt_regset_alloc(drp);
2106 		dt_cg_setx(dlp, dnp->dn_reg, dnp->dn_value);
2107 		break;
2108 
2109 	default:
2110 		xyerror(D_UNKNOWN, "internal error -- token type %u is not a "
2111 		    "valid D compilation token\n", dnp->dn_op);
2112 	}
2113 }
2114 
2115 void
2116 dt_cg(dt_pcb_t *pcb, dt_node_t *dnp)
2117 {
2118 	dif_instr_t instr;
2119 	dt_xlator_t *dxp;
2120 	dt_ident_t *idp;
2121 
2122 	if (pcb->pcb_regs == NULL && (pcb->pcb_regs =
2123 	    dt_regset_create(pcb->pcb_hdl->dt_conf.dtc_difintregs)) == NULL)
2124 		longjmp(pcb->pcb_jmpbuf, EDT_NOMEM);
2125 
2126 	dt_regset_reset(pcb->pcb_regs);
2127 	(void) dt_regset_alloc(pcb->pcb_regs); /* allocate %r0 */
2128 
2129 	if (pcb->pcb_inttab != NULL)
2130 		dt_inttab_destroy(pcb->pcb_inttab);
2131 
2132 	if ((pcb->pcb_inttab = dt_inttab_create(yypcb->pcb_hdl)) == NULL)
2133 		longjmp(pcb->pcb_jmpbuf, EDT_NOMEM);
2134 
2135 	if (pcb->pcb_strtab != NULL)
2136 		dt_strtab_destroy(pcb->pcb_strtab);
2137 
2138 	if ((pcb->pcb_strtab = dt_strtab_create(BUFSIZ)) == NULL)
2139 		longjmp(pcb->pcb_jmpbuf, EDT_NOMEM);
2140 
2141 	dt_irlist_destroy(&pcb->pcb_ir);
2142 	dt_irlist_create(&pcb->pcb_ir);
2143 
2144 	assert(pcb->pcb_dret == NULL);
2145 	pcb->pcb_dret = dnp;
2146 
2147 	if (dt_node_resolve(dnp, DT_IDENT_XLPTR) != NULL) {
2148 		dnerror(dnp, D_CG_DYN, "expression cannot evaluate to result "
2149 		    "of a translated pointer\n");
2150 	}
2151 
2152 	/*
2153 	 * If we're generating code for a translator body, assign the input
2154 	 * parameter to the first available register (i.e. caller passes %r1).
2155 	 */
2156 	if (dnp->dn_kind == DT_NODE_MEMBER) {
2157 		dxp = dnp->dn_membxlator;
2158 		dnp = dnp->dn_membexpr;
2159 
2160 		dxp->dx_ident->di_flags |= DT_IDFLG_CGREG;
2161 		dxp->dx_ident->di_id = dt_regset_alloc(pcb->pcb_regs);
2162 	}
2163 
2164 	dt_cg_node(dnp, &pcb->pcb_ir, pcb->pcb_regs);
2165 
2166 	if ((idp = dt_node_resolve(dnp, DT_IDENT_XLSOU)) != NULL) {
2167 		int reg = dt_cg_xlate_expand(dnp, idp,
2168 		    &pcb->pcb_ir, pcb->pcb_regs);
2169 		dt_regset_free(pcb->pcb_regs, dnp->dn_reg);
2170 		dnp->dn_reg = reg;
2171 	}
2172 
2173 	instr = DIF_INSTR_RET(dnp->dn_reg);
2174 	dt_regset_free(pcb->pcb_regs, dnp->dn_reg);
2175 	dt_irlist_append(&pcb->pcb_ir, dt_cg_node_alloc(DT_LBL_NONE, instr));
2176 
2177 	if (dnp->dn_kind == DT_NODE_MEMBER) {
2178 		dt_regset_free(pcb->pcb_regs, dxp->dx_ident->di_id);
2179 		dxp->dx_ident->di_id = 0;
2180 		dxp->dx_ident->di_flags &= ~DT_IDFLG_CGREG;
2181 	}
2182 
2183 	dt_regset_free(pcb->pcb_regs, 0);
2184 	dt_regset_assert_free(pcb->pcb_regs);
2185 }
2186