xref: /illumos-gate/usr/src/uts/common/inet/iptun/iptun.c (revision 826ac02a0def83e0a41b29321470d299c7389aab)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved.
23  */
24 
25 /*
26  * iptun - IP Tunneling Driver
27  *
28  * This module is a GLDv3 driver that implements virtual datalinks over IP
29  * (a.k.a, IP tunneling).  The datalinks are managed through a dld ioctl
30  * interface (see iptun_ctl.c), and registered with GLDv3 using
31  * mac_register().  It implements the logic for various forms of IP (IPv4 or
32  * IPv6) encapsulation within IP (IPv4 or IPv6) by interacting with the ip
33  * module below it.  Each virtual IP tunnel datalink has a conn_t associated
34  * with it representing the "outer" IP connection.
35  *
36  * The module implements the following locking semantics:
37  *
38  * Lookups and deletions in iptun_hash are synchronized using iptun_hash_lock.
39  * See comments above iptun_hash_lock for details.
40  *
41  * No locks are ever held while calling up to GLDv3.  The general architecture
42  * of GLDv3 requires this, as the mac perimeter (essentially a lock) for a
43  * given link will be held while making downcalls (iptun_m_*() callbacks).
44  * Because we need to hold locks while handling downcalls, holding these locks
45  * while issuing upcalls results in deadlock scenarios.  See the block comment
46  * above iptun_task_cb() for details on how we safely issue upcalls without
47  * holding any locks.
48  *
49  * The contents of each iptun_t is protected by an iptun_mutex which is held
50  * in iptun_enter() (called by iptun_enter_by_linkid()), and exited in
51  * iptun_exit().
52  *
53  * See comments in iptun_delete() and iptun_free() for details on how the
54  * iptun_t is deleted safely.
55  */
56 
57 #include <sys/types.h>
58 #include <sys/kmem.h>
59 #include <sys/errno.h>
60 #include <sys/modhash.h>
61 #include <sys/list.h>
62 #include <sys/strsun.h>
63 #include <sys/file.h>
64 #include <sys/systm.h>
65 #include <sys/tihdr.h>
66 #include <sys/param.h>
67 #include <sys/mac_provider.h>
68 #include <sys/mac_ipv4.h>
69 #include <sys/mac_ipv6.h>
70 #include <sys/mac_6to4.h>
71 #include <sys/tsol/tnet.h>
72 #include <sys/sunldi.h>
73 #include <netinet/in.h>
74 #include <netinet/ip6.h>
75 #include <inet/ip.h>
76 #include <inet/ip_ire.h>
77 #include <inet/ipsec_impl.h>
78 #include <sys/tsol/label.h>
79 #include <sys/tsol/tnet.h>
80 #include <inet/iptun.h>
81 #include "iptun_impl.h"
82 
83 /* Do the tunnel type and address family match? */
84 #define	IPTUN_ADDR_MATCH(iptun_type, family)				\
85 	((iptun_type == IPTUN_TYPE_IPV4 && family == AF_INET) ||	\
86 	(iptun_type == IPTUN_TYPE_IPV6 && family == AF_INET6) ||	\
87 	(iptun_type == IPTUN_TYPE_6TO4 && family == AF_INET))
88 
89 #define	IPTUN_HASH_KEY(key)	((mod_hash_key_t)(uintptr_t)(key))
90 
91 #define	IPTUN_MIN_IPV4_MTU	576		/* ip.h still uses 68 (!) */
92 #define	IPTUN_MIN_IPV6_MTU	IPV6_MIN_MTU
93 #define	IPTUN_MAX_IPV4_MTU	(IP_MAXPACKET - sizeof (ipha_t))
94 #define	IPTUN_MAX_IPV6_MTU	(IP_MAXPACKET - sizeof (ip6_t) -	\
95 				    sizeof (iptun_encaplim_t))
96 
97 #define	IPTUN_MIN_HOPLIMIT	1
98 #define	IPTUN_MAX_HOPLIMIT	UINT8_MAX
99 
100 #define	IPTUN_MIN_ENCAPLIMIT	0
101 #define	IPTUN_MAX_ENCAPLIMIT	UINT8_MAX
102 
103 #define	IPTUN_IPSEC_REQ_MASK	(IPSEC_PREF_REQUIRED | IPSEC_PREF_NEVER)
104 
105 static iptun_encaplim_t	iptun_encaplim_init = {
106 	{ IPPROTO_NONE, 0 },
107 	IP6OPT_TUNNEL_LIMIT,
108 	1,
109 	IPTUN_DEFAULT_ENCAPLIMIT,	/* filled in with actual value later */
110 	IP6OPT_PADN,
111 	1,
112 	0
113 };
114 
115 /*
116  * Table containing per-iptun-type information.
117  * Since IPv6 can run over all of these we have the IPv6 min as the min MTU.
118  */
119 static iptun_typeinfo_t	iptun_type_table[] = {
120 	{ IPTUN_TYPE_IPV4, MAC_PLUGIN_IDENT_IPV4, IPV4_VERSION,
121 	    IPTUN_MIN_IPV6_MTU,	IPTUN_MAX_IPV4_MTU,	B_TRUE },
122 	{ IPTUN_TYPE_IPV6, MAC_PLUGIN_IDENT_IPV6, IPV6_VERSION,
123 	    IPTUN_MIN_IPV6_MTU,	IPTUN_MAX_IPV6_MTU,	B_TRUE },
124 	{ IPTUN_TYPE_6TO4, MAC_PLUGIN_IDENT_6TO4, IPV4_VERSION,
125 	    IPTUN_MIN_IPV6_MTU,	IPTUN_MAX_IPV4_MTU,	B_FALSE },
126 	{ IPTUN_TYPE_UNKNOWN, NULL, 0, 0, 0, B_FALSE }
127 };
128 
129 /*
130  * iptun_hash is an iptun_t lookup table by link ID protected by
131  * iptun_hash_lock.  While the hash table's integrity is maintained via
132  * internal locking in the mod_hash_*() functions, we need additional locking
133  * so that an iptun_t cannot be deleted after a hash lookup has returned an
134  * iptun_t and before iptun_lock has been entered.  As such, we use
135  * iptun_hash_lock when doing lookups and removals from iptun_hash.
136  */
137 mod_hash_t	*iptun_hash;
138 static kmutex_t	iptun_hash_lock;
139 
140 static uint_t	iptun_tunnelcount;	/* total for all stacks */
141 kmem_cache_t	*iptun_cache;
142 ddi_taskq_t 	*iptun_taskq;
143 
144 typedef enum {
145 	IPTUN_TASK_MTU_UPDATE,	/* tell mac about new tunnel link MTU */
146 	IPTUN_TASK_LADDR_UPDATE, /* tell mac about new local address */
147 	IPTUN_TASK_RADDR_UPDATE, /* tell mac about new remote address */
148 	IPTUN_TASK_LINK_UPDATE,	/* tell mac about new link state */
149 	IPTUN_TASK_PDATA_UPDATE	/* tell mac about updated plugin data */
150 } iptun_task_t;
151 
152 typedef struct iptun_task_data_s {
153 	iptun_task_t	itd_task;
154 	datalink_id_t	itd_linkid;
155 } iptun_task_data_t;
156 
157 static void iptun_task_dispatch(iptun_t *, iptun_task_t);
158 static int iptun_enter(iptun_t *);
159 static void iptun_exit(iptun_t *);
160 static void iptun_headergen(iptun_t *, boolean_t);
161 static void iptun_drop_pkt(mblk_t *, uint64_t *);
162 static void iptun_input(void *, mblk_t *, void *, ip_recv_attr_t *);
163 static void iptun_input_icmp(void *, mblk_t *, void *, ip_recv_attr_t *);
164 static void iptun_output(iptun_t *, mblk_t *);
165 static uint32_t iptun_get_maxmtu(iptun_t *, ip_xmit_attr_t *, uint32_t);
166 static uint32_t iptun_update_mtu(iptun_t *, ip_xmit_attr_t *, uint32_t);
167 static uint32_t iptun_get_dst_pmtu(iptun_t *, ip_xmit_attr_t *);
168 static void iptun_update_dst_pmtu(iptun_t *, ip_xmit_attr_t *);
169 static int iptun_setladdr(iptun_t *, const struct sockaddr_storage *);
170 
171 static void iptun_output_6to4(iptun_t *, mblk_t *);
172 static void iptun_output_common(iptun_t *, ip_xmit_attr_t *, mblk_t *);
173 static boolean_t iptun_verifyicmp(conn_t *, void *, icmph_t *, icmp6_t *,
174     ip_recv_attr_t *);
175 
176 static void iptun_notify(void *, ip_xmit_attr_t *, ixa_notify_type_t,
177     ixa_notify_arg_t);
178 
179 static mac_callbacks_t iptun_m_callbacks;
180 
181 static int
182 iptun_m_getstat(void *arg, uint_t stat, uint64_t *val)
183 {
184 	iptun_t	*iptun = arg;
185 	int	err = 0;
186 
187 	switch (stat) {
188 	case MAC_STAT_IERRORS:
189 		*val = iptun->iptun_ierrors;
190 		break;
191 	case MAC_STAT_OERRORS:
192 		*val = iptun->iptun_oerrors;
193 		break;
194 	case MAC_STAT_RBYTES:
195 		*val = iptun->iptun_rbytes;
196 		break;
197 	case MAC_STAT_IPACKETS:
198 		*val = iptun->iptun_ipackets;
199 		break;
200 	case MAC_STAT_OBYTES:
201 		*val = iptun->iptun_obytes;
202 		break;
203 	case MAC_STAT_OPACKETS:
204 		*val = iptun->iptun_opackets;
205 		break;
206 	case MAC_STAT_NORCVBUF:
207 		*val = iptun->iptun_norcvbuf;
208 		break;
209 	case MAC_STAT_NOXMTBUF:
210 		*val = iptun->iptun_noxmtbuf;
211 		break;
212 	default:
213 		err = ENOTSUP;
214 	}
215 
216 	return (err);
217 }
218 
219 static int
220 iptun_m_start(void *arg)
221 {
222 	iptun_t	*iptun = arg;
223 	int	err;
224 
225 	if ((err = iptun_enter(iptun)) == 0) {
226 		iptun->iptun_flags |= IPTUN_MAC_STARTED;
227 		iptun_task_dispatch(iptun, IPTUN_TASK_LINK_UPDATE);
228 		iptun_exit(iptun);
229 	}
230 	return (err);
231 }
232 
233 static void
234 iptun_m_stop(void *arg)
235 {
236 	iptun_t *iptun = arg;
237 
238 	if (iptun_enter(iptun) == 0) {
239 		iptun->iptun_flags &= ~IPTUN_MAC_STARTED;
240 		iptun_task_dispatch(iptun, IPTUN_TASK_LINK_UPDATE);
241 		iptun_exit(iptun);
242 	}
243 }
244 
245 /*
246  * iptun_m_setpromisc() does nothing and always succeeds.  This is because a
247  * tunnel data-link only ever receives packets that are destined exclusively
248  * for the local address of the tunnel.
249  */
250 /* ARGSUSED */
251 static int
252 iptun_m_setpromisc(void *arg, boolean_t on)
253 {
254 	return (0);
255 }
256 
257 /* ARGSUSED */
258 static int
259 iptun_m_multicst(void *arg, boolean_t add, const uint8_t *addrp)
260 {
261 	return (ENOTSUP);
262 }
263 
264 /*
265  * iptun_m_unicst() sets the local address.
266  */
267 /* ARGSUSED */
268 static int
269 iptun_m_unicst(void *arg, const uint8_t *addrp)
270 {
271 	iptun_t			*iptun = arg;
272 	int			err;
273 	struct sockaddr_storage	ss;
274 	struct sockaddr_in	*sin;
275 	struct sockaddr_in6	*sin6;
276 
277 	if ((err = iptun_enter(iptun)) == 0) {
278 		switch (iptun->iptun_typeinfo->iti_ipvers) {
279 		case IPV4_VERSION:
280 			sin = (struct sockaddr_in *)&ss;
281 			sin->sin_family = AF_INET;
282 			bcopy(addrp, &sin->sin_addr, sizeof (in_addr_t));
283 			break;
284 		case IPV6_VERSION:
285 			sin6 = (struct sockaddr_in6 *)&ss;
286 			sin6->sin6_family = AF_INET6;
287 			bcopy(addrp, &sin6->sin6_addr, sizeof (in6_addr_t));
288 			break;
289 		default:
290 			ASSERT(0);
291 		}
292 		err = iptun_setladdr(iptun, &ss);
293 		iptun_exit(iptun);
294 	}
295 	return (err);
296 }
297 
298 static mblk_t *
299 iptun_m_tx(void *arg, mblk_t *mpchain)
300 {
301 	mblk_t	*mp, *nmp;
302 	iptun_t	*iptun = arg;
303 
304 	if (!IS_IPTUN_RUNNING(iptun)) {
305 		iptun_drop_pkt(mpchain, &iptun->iptun_noxmtbuf);
306 		return (NULL);
307 	}
308 
309 	for (mp = mpchain; mp != NULL; mp = nmp) {
310 		nmp = mp->b_next;
311 		mp->b_next = NULL;
312 		iptun_output(iptun, mp);
313 	}
314 
315 	return (NULL);
316 }
317 
318 /* ARGSUSED */
319 static int
320 iptun_m_setprop(void *barg, const char *pr_name, mac_prop_id_t pr_num,
321     uint_t pr_valsize, const void *pr_val)
322 {
323 	iptun_t		*iptun = barg;
324 	uint32_t	value = *(uint32_t *)pr_val;
325 	int		err;
326 
327 	/*
328 	 * We need to enter this iptun_t since we'll be modifying the outer
329 	 * header.
330 	 */
331 	if ((err = iptun_enter(iptun)) != 0)
332 		return (err);
333 
334 	switch (pr_num) {
335 	case MAC_PROP_IPTUN_HOPLIMIT:
336 		if (value < IPTUN_MIN_HOPLIMIT || value > IPTUN_MAX_HOPLIMIT) {
337 			err = EINVAL;
338 			break;
339 		}
340 		if (value != iptun->iptun_hoplimit) {
341 			iptun->iptun_hoplimit = (uint8_t)value;
342 			iptun_headergen(iptun, B_TRUE);
343 		}
344 		break;
345 	case MAC_PROP_IPTUN_ENCAPLIMIT:
346 		if (iptun->iptun_typeinfo->iti_type != IPTUN_TYPE_IPV6 ||
347 		    value > IPTUN_MAX_ENCAPLIMIT) {
348 			err = EINVAL;
349 			break;
350 		}
351 		if (value != iptun->iptun_encaplimit) {
352 			iptun->iptun_encaplimit = (uint8_t)value;
353 			iptun_headergen(iptun, B_TRUE);
354 		}
355 		break;
356 	case MAC_PROP_MTU: {
357 		uint32_t maxmtu = iptun_get_maxmtu(iptun, NULL, 0);
358 
359 		if (value < iptun->iptun_typeinfo->iti_minmtu ||
360 		    value > maxmtu) {
361 			err = EINVAL;
362 			break;
363 		}
364 		iptun->iptun_flags |= IPTUN_FIXED_MTU;
365 		if (value != iptun->iptun_mtu) {
366 			iptun->iptun_mtu = value;
367 			iptun_task_dispatch(iptun, IPTUN_TASK_MTU_UPDATE);
368 		}
369 		break;
370 	}
371 	default:
372 		err = EINVAL;
373 	}
374 	iptun_exit(iptun);
375 	return (err);
376 }
377 
378 /* ARGSUSED */
379 static int
380 iptun_m_getprop(void *barg, const char *pr_name, mac_prop_id_t pr_num,
381     uint_t pr_valsize, void *pr_val)
382 {
383 	iptun_t			*iptun = barg;
384 	int			err;
385 
386 	if ((err = iptun_enter(iptun)) != 0)
387 		return (err);
388 
389 	switch (pr_num) {
390 	case MAC_PROP_IPTUN_HOPLIMIT:
391 		ASSERT(pr_valsize >= sizeof (uint32_t));
392 		*(uint32_t *)pr_val = iptun->iptun_hoplimit;
393 		break;
394 
395 	case MAC_PROP_IPTUN_ENCAPLIMIT:
396 		*(uint32_t *)pr_val = iptun->iptun_encaplimit;
397 		break;
398 	default:
399 		err = ENOTSUP;
400 	}
401 done:
402 	iptun_exit(iptun);
403 	return (err);
404 }
405 
406 /* ARGSUSED */
407 static void
408 iptun_m_propinfo(void *barg, const char *pr_name, mac_prop_id_t pr_num,
409     mac_prop_info_handle_t prh)
410 {
411 	iptun_t			*iptun = barg;
412 
413 	switch (pr_num) {
414 	case MAC_PROP_IPTUN_HOPLIMIT:
415 		mac_prop_info_set_range_uint32(prh,
416 		    IPTUN_MIN_HOPLIMIT, IPTUN_MAX_HOPLIMIT);
417 		mac_prop_info_set_default_uint32(prh, IPTUN_DEFAULT_HOPLIMIT);
418 		break;
419 
420 	case MAC_PROP_IPTUN_ENCAPLIMIT:
421 		if (iptun->iptun_typeinfo->iti_type != IPTUN_TYPE_IPV6)
422 			break;
423 		mac_prop_info_set_range_uint32(prh,
424 		    IPTUN_MIN_ENCAPLIMIT, IPTUN_MAX_ENCAPLIMIT);
425 		mac_prop_info_set_default_uint32(prh, IPTUN_DEFAULT_ENCAPLIMIT);
426 		break;
427 	case MAC_PROP_MTU:
428 		mac_prop_info_set_range_uint32(prh,
429 		    iptun->iptun_typeinfo->iti_minmtu,
430 		    iptun_get_maxmtu(iptun, NULL, 0));
431 		break;
432 	}
433 }
434 
435 uint_t
436 iptun_count(void)
437 {
438 	return (iptun_tunnelcount);
439 }
440 
441 /*
442  * Enter an iptun_t exclusively.  This is essentially just a mutex, but we
443  * don't allow iptun_enter() to succeed on a tunnel if it's in the process of
444  * being deleted.
445  */
446 static int
447 iptun_enter(iptun_t *iptun)
448 {
449 	mutex_enter(&iptun->iptun_lock);
450 	while (iptun->iptun_flags & IPTUN_DELETE_PENDING)
451 		cv_wait(&iptun->iptun_enter_cv, &iptun->iptun_lock);
452 	if (iptun->iptun_flags & IPTUN_CONDEMNED) {
453 		mutex_exit(&iptun->iptun_lock);
454 		return (ENOENT);
455 	}
456 	return (0);
457 }
458 
459 /*
460  * Exit the tunnel entered in iptun_enter().
461  */
462 static void
463 iptun_exit(iptun_t *iptun)
464 {
465 	mutex_exit(&iptun->iptun_lock);
466 }
467 
468 /*
469  * Enter the IP tunnel instance by datalink ID.
470  */
471 static int
472 iptun_enter_by_linkid(datalink_id_t linkid, iptun_t **iptun)
473 {
474 	int err;
475 
476 	mutex_enter(&iptun_hash_lock);
477 	if (mod_hash_find(iptun_hash, IPTUN_HASH_KEY(linkid),
478 	    (mod_hash_val_t *)iptun) == 0)
479 		err = iptun_enter(*iptun);
480 	else
481 		err = ENOENT;
482 	if (err != 0)
483 		*iptun = NULL;
484 	mutex_exit(&iptun_hash_lock);
485 	return (err);
486 }
487 
488 /*
489  * Handle tasks that were deferred through the iptun_taskq because they require
490  * calling up to the mac module, and we can't call up to the mac module while
491  * holding locks.
492  *
493  * This is tricky to get right without introducing race conditions and
494  * deadlocks with the mac module, as we cannot issue an upcall while in the
495  * iptun_t.  The reason is that upcalls may try and enter the mac perimeter,
496  * while iptun callbacks (such as iptun_m_setprop()) called from the mac
497  * module will already have the perimeter held, and will then try and enter
498  * the iptun_t.  You can see the lock ordering problem with this; this will
499  * deadlock.
500  *
501  * The safe way to do this is to enter the iptun_t in question and copy the
502  * information we need out of it so that we can exit it and know that the
503  * information being passed up to the upcalls won't be subject to modification
504  * by other threads.  The problem now is that we need to exit it prior to
505  * issuing the upcall, but once we do this, a thread could come along and
506  * delete the iptun_t and thus the mac handle required to issue the upcall.
507  * To prevent this, we set the IPTUN_UPCALL_PENDING flag prior to exiting the
508  * iptun_t.  This flag is the condition associated with iptun_upcall_cv, which
509  * iptun_delete() will cv_wait() on.  When the upcall completes, we clear
510  * IPTUN_UPCALL_PENDING and cv_signal() any potentially waiting
511  * iptun_delete().  We can thus still safely use iptun->iptun_mh after having
512  * exited the iptun_t.
513  */
514 static void
515 iptun_task_cb(void *arg)
516 {
517 	iptun_task_data_t	*itd = arg;
518 	iptun_task_t		task = itd->itd_task;
519 	datalink_id_t		linkid = itd->itd_linkid;
520 	iptun_t			*iptun;
521 	uint32_t		mtu;
522 	iptun_addr_t		addr;
523 	link_state_t		linkstate;
524 	size_t			header_size;
525 	iptun_header_t		header;
526 
527 	kmem_free(itd, sizeof (*itd));
528 
529 	/*
530 	 * Note that if the lookup fails, it's because the tunnel was deleted
531 	 * between the time the task was dispatched and now.  That isn't an
532 	 * error.
533 	 */
534 	if (iptun_enter_by_linkid(linkid, &iptun) != 0)
535 		return;
536 
537 	iptun->iptun_flags |= IPTUN_UPCALL_PENDING;
538 
539 	switch (task) {
540 	case IPTUN_TASK_MTU_UPDATE:
541 		mtu = iptun->iptun_mtu;
542 		break;
543 	case IPTUN_TASK_LADDR_UPDATE:
544 		addr = iptun->iptun_laddr;
545 		break;
546 	case IPTUN_TASK_RADDR_UPDATE:
547 		addr = iptun->iptun_raddr;
548 		break;
549 	case IPTUN_TASK_LINK_UPDATE:
550 		linkstate = IS_IPTUN_RUNNING(iptun) ?
551 		    LINK_STATE_UP : LINK_STATE_DOWN;
552 		break;
553 	case IPTUN_TASK_PDATA_UPDATE:
554 		header_size = iptun->iptun_header_size;
555 		header = iptun->iptun_header;
556 		break;
557 	default:
558 		ASSERT(0);
559 	}
560 
561 	iptun_exit(iptun);
562 
563 	switch (task) {
564 	case IPTUN_TASK_MTU_UPDATE:
565 		(void) mac_maxsdu_update(iptun->iptun_mh, mtu);
566 		break;
567 	case IPTUN_TASK_LADDR_UPDATE:
568 		mac_unicst_update(iptun->iptun_mh, (uint8_t *)&addr.ia_addr);
569 		break;
570 	case IPTUN_TASK_RADDR_UPDATE:
571 		mac_dst_update(iptun->iptun_mh, (uint8_t *)&addr.ia_addr);
572 		break;
573 	case IPTUN_TASK_LINK_UPDATE:
574 		mac_link_update(iptun->iptun_mh, linkstate);
575 		break;
576 	case IPTUN_TASK_PDATA_UPDATE:
577 		if (mac_pdata_update(iptun->iptun_mh,
578 		    header_size == 0 ? NULL : &header, header_size) != 0)
579 			atomic_inc_64(&iptun->iptun_taskq_fail);
580 		break;
581 	}
582 
583 	mutex_enter(&iptun->iptun_lock);
584 	iptun->iptun_flags &= ~IPTUN_UPCALL_PENDING;
585 	cv_signal(&iptun->iptun_upcall_cv);
586 	mutex_exit(&iptun->iptun_lock);
587 }
588 
589 static void
590 iptun_task_dispatch(iptun_t *iptun, iptun_task_t iptun_task)
591 {
592 	iptun_task_data_t *itd;
593 
594 	itd = kmem_alloc(sizeof (*itd), KM_NOSLEEP);
595 	if (itd == NULL) {
596 		atomic_inc_64(&iptun->iptun_taskq_fail);
597 		return;
598 	}
599 	itd->itd_task = iptun_task;
600 	itd->itd_linkid = iptun->iptun_linkid;
601 	if (ddi_taskq_dispatch(iptun_taskq, iptun_task_cb, itd, DDI_NOSLEEP)) {
602 		atomic_inc_64(&iptun->iptun_taskq_fail);
603 		kmem_free(itd, sizeof (*itd));
604 	}
605 }
606 
607 /*
608  * Convert an iptun_addr_t to sockaddr_storage.
609  */
610 static void
611 iptun_getaddr(iptun_addr_t *iptun_addr, struct sockaddr_storage *ss)
612 {
613 	struct sockaddr_in	*sin;
614 	struct sockaddr_in6	*sin6;
615 
616 	bzero(ss, sizeof (*ss));
617 	switch (iptun_addr->ia_family) {
618 	case AF_INET:
619 		sin = (struct sockaddr_in *)ss;
620 		sin->sin_addr.s_addr = iptun_addr->ia_addr.iau_addr4;
621 		break;
622 	case AF_INET6:
623 		sin6 = (struct sockaddr_in6 *)ss;
624 		sin6->sin6_addr = iptun_addr->ia_addr.iau_addr6;
625 		break;
626 	default:
627 		ASSERT(0);
628 	}
629 	ss->ss_family = iptun_addr->ia_family;
630 }
631 
632 /*
633  * General purpose function to set an IP tunnel source or destination address.
634  */
635 static int
636 iptun_setaddr(iptun_type_t iptun_type, iptun_addr_t *iptun_addr,
637     const struct sockaddr_storage *ss)
638 {
639 	if (!IPTUN_ADDR_MATCH(iptun_type, ss->ss_family))
640 		return (EINVAL);
641 
642 	switch (ss->ss_family) {
643 	case AF_INET: {
644 		struct sockaddr_in *sin = (struct sockaddr_in *)ss;
645 
646 		if ((sin->sin_addr.s_addr == INADDR_ANY) ||
647 		    (sin->sin_addr.s_addr == INADDR_BROADCAST) ||
648 		    CLASSD(sin->sin_addr.s_addr)) {
649 			return (EADDRNOTAVAIL);
650 		}
651 		iptun_addr->ia_addr.iau_addr4 = sin->sin_addr.s_addr;
652 		break;
653 	}
654 	case AF_INET6: {
655 		struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)ss;
656 
657 		if (IN6_IS_ADDR_UNSPECIFIED(&sin6->sin6_addr) ||
658 		    IN6_IS_ADDR_MULTICAST(&sin6->sin6_addr) ||
659 		    IN6_IS_ADDR_V4MAPPED(&sin6->sin6_addr)) {
660 			return (EADDRNOTAVAIL);
661 		}
662 		iptun_addr->ia_addr.iau_addr6 = sin6->sin6_addr;
663 		break;
664 	}
665 	default:
666 		return (EAFNOSUPPORT);
667 	}
668 	iptun_addr->ia_family = ss->ss_family;
669 	return (0);
670 }
671 
672 static int
673 iptun_setladdr(iptun_t *iptun, const struct sockaddr_storage *laddr)
674 {
675 	return (iptun_setaddr(iptun->iptun_typeinfo->iti_type,
676 	    &iptun->iptun_laddr, laddr));
677 }
678 
679 static int
680 iptun_setraddr(iptun_t *iptun, const struct sockaddr_storage *raddr)
681 {
682 	if (!(iptun->iptun_typeinfo->iti_hasraddr))
683 		return (EINVAL);
684 	return (iptun_setaddr(iptun->iptun_typeinfo->iti_type,
685 	    &iptun->iptun_raddr, raddr));
686 }
687 
688 static boolean_t
689 iptun_canbind(iptun_t *iptun)
690 {
691 	/*
692 	 * A tunnel may bind when its source address has been set, and if its
693 	 * tunnel type requires one, also its destination address.
694 	 */
695 	return ((iptun->iptun_flags & IPTUN_LADDR) &&
696 	    ((iptun->iptun_flags & IPTUN_RADDR) ||
697 	    !(iptun->iptun_typeinfo->iti_hasraddr)));
698 }
699 
700 /*
701  * Verify that the local address is valid, and insert in the fanout
702  */
703 static int
704 iptun_bind(iptun_t *iptun)
705 {
706 	conn_t			*connp = iptun->iptun_connp;
707 	int			error = 0;
708 	ip_xmit_attr_t		*ixa;
709 	ip_xmit_attr_t		*oldixa;
710 	iulp_t			uinfo;
711 	ip_stack_t		*ipst = connp->conn_netstack->netstack_ip;
712 
713 	/*
714 	 * Get an exclusive ixa for this thread.
715 	 * We defer updating conn_ixa until later to handle any concurrent
716 	 * conn_ixa_cleanup thread.
717 	 */
718 	ixa = conn_get_ixa(connp, B_FALSE);
719 	if (ixa == NULL)
720 		return (ENOMEM);
721 
722 	/* We create PMTU state including for 6to4 */
723 	ixa->ixa_flags |= IXAF_PMTU_DISCOVERY;
724 
725 	ASSERT(iptun_canbind(iptun));
726 
727 	mutex_enter(&connp->conn_lock);
728 	/*
729 	 * Note that conn_proto can't be set since the upper protocol
730 	 * can be both 41 and 4 when IPv6 and IPv4 are over the same tunnel.
731 	 * ipcl_iptun_classify doesn't use conn_proto.
732 	 */
733 	connp->conn_ipversion = iptun->iptun_typeinfo->iti_ipvers;
734 
735 	switch (iptun->iptun_typeinfo->iti_type) {
736 	case IPTUN_TYPE_IPV4:
737 		IN6_IPADDR_TO_V4MAPPED(iptun->iptun_laddr4,
738 		    &connp->conn_laddr_v6);
739 		IN6_IPADDR_TO_V4MAPPED(iptun->iptun_raddr4,
740 		    &connp->conn_faddr_v6);
741 		ixa->ixa_flags |= IXAF_IS_IPV4;
742 		if (ip_laddr_verify_v4(iptun->iptun_laddr4, IPCL_ZONEID(connp),
743 		    ipst, B_FALSE) != IPVL_UNICAST_UP) {
744 			mutex_exit(&connp->conn_lock);
745 			error = EADDRNOTAVAIL;
746 			goto done;
747 		}
748 		break;
749 	case IPTUN_TYPE_IPV6:
750 		connp->conn_laddr_v6 = iptun->iptun_laddr6;
751 		connp->conn_faddr_v6 = iptun->iptun_raddr6;
752 		ixa->ixa_flags &= ~IXAF_IS_IPV4;
753 		/* We use a zero scopeid for now */
754 		if (ip_laddr_verify_v6(&iptun->iptun_laddr6, IPCL_ZONEID(connp),
755 		    ipst, B_FALSE, 0) != IPVL_UNICAST_UP) {
756 			mutex_exit(&connp->conn_lock);
757 			error = EADDRNOTAVAIL;
758 			goto done;
759 		}
760 		break;
761 	case IPTUN_TYPE_6TO4:
762 		IN6_IPADDR_TO_V4MAPPED(iptun->iptun_laddr4,
763 		    &connp->conn_laddr_v6);
764 		IN6_IPADDR_TO_V4MAPPED(INADDR_ANY, &connp->conn_faddr_v6);
765 		ixa->ixa_flags |= IXAF_IS_IPV4;
766 		mutex_exit(&connp->conn_lock);
767 
768 		switch (ip_laddr_verify_v4(iptun->iptun_laddr4,
769 		    IPCL_ZONEID(connp), ipst, B_FALSE)) {
770 		case IPVL_UNICAST_UP:
771 		case IPVL_UNICAST_DOWN:
772 			break;
773 		default:
774 			error = EADDRNOTAVAIL;
775 			goto done;
776 		}
777 		goto insert;
778 	}
779 
780 	/* In case previous destination was multirt */
781 	ip_attr_newdst(ixa);
782 
783 	/*
784 	 * When we set a tunnel's destination address, we do not
785 	 * care if the destination is reachable.  Transient routing
786 	 * issues should not inhibit the creation of a tunnel
787 	 * interface, for example. Thus we pass B_FALSE here.
788 	 */
789 	connp->conn_saddr_v6 = connp->conn_laddr_v6;
790 	mutex_exit(&connp->conn_lock);
791 
792 	/* As long as the MTU is large we avoid fragmentation */
793 	ixa->ixa_flags |= IXAF_DONTFRAG | IXAF_PMTU_IPV4_DF;
794 
795 	/* We handle IPsec in iptun_output_common */
796 	error = ip_attr_connect(connp, ixa, &connp->conn_saddr_v6,
797 	    &connp->conn_faddr_v6, &connp->conn_faddr_v6, 0,
798 	    &connp->conn_saddr_v6, &uinfo, 0);
799 
800 	if (error != 0)
801 		goto done;
802 
803 	/* saddr shouldn't change since it was already set */
804 	ASSERT(IN6_ARE_ADDR_EQUAL(&connp->conn_laddr_v6,
805 	    &connp->conn_saddr_v6));
806 
807 	/* We set IXAF_VERIFY_PMTU to catch PMTU increases */
808 	ixa->ixa_flags |= IXAF_VERIFY_PMTU;
809 	ASSERT(uinfo.iulp_mtu != 0);
810 
811 	/*
812 	 * Allow setting new policies.
813 	 * The addresses/ports are already set, thus the IPsec policy calls
814 	 * can handle their passed-in conn's.
815 	 */
816 	connp->conn_policy_cached = B_FALSE;
817 
818 insert:
819 	error = ipcl_conn_insert(connp);
820 	if (error != 0)
821 		goto done;
822 
823 	/* Atomically update v6lastdst and conn_ixa */
824 	mutex_enter(&connp->conn_lock);
825 	/* Record this as the "last" send even though we haven't sent any */
826 	connp->conn_v6lastdst = connp->conn_faddr_v6;
827 
828 	iptun->iptun_flags |= IPTUN_BOUND;
829 
830 	oldixa = conn_replace_ixa(connp, ixa);
831 	/* Done with conn_t */
832 	mutex_exit(&connp->conn_lock);
833 	ixa_refrele(oldixa);
834 
835 	/*
836 	 * Now that we're bound with ip below us, this is a good
837 	 * time to initialize the destination path MTU and to
838 	 * re-calculate the tunnel's link MTU.
839 	 */
840 	(void) iptun_update_mtu(iptun, ixa, 0);
841 
842 	if (IS_IPTUN_RUNNING(iptun))
843 		iptun_task_dispatch(iptun, IPTUN_TASK_LINK_UPDATE);
844 
845 done:
846 	ixa_refrele(ixa);
847 	return (error);
848 }
849 
850 static void
851 iptun_unbind(iptun_t *iptun)
852 {
853 	ASSERT(iptun->iptun_flags & IPTUN_BOUND);
854 	ASSERT(mutex_owned(&iptun->iptun_lock) ||
855 	    (iptun->iptun_flags & IPTUN_CONDEMNED));
856 	ip_unbind(iptun->iptun_connp);
857 	iptun->iptun_flags &= ~IPTUN_BOUND;
858 	if (!(iptun->iptun_flags & IPTUN_CONDEMNED))
859 		iptun_task_dispatch(iptun, IPTUN_TASK_LINK_UPDATE);
860 }
861 
862 /*
863  * Re-generate the template data-link header for a given IP tunnel given the
864  * tunnel's current parameters.
865  */
866 static void
867 iptun_headergen(iptun_t *iptun, boolean_t update_mac)
868 {
869 	switch (iptun->iptun_typeinfo->iti_ipvers) {
870 	case IPV4_VERSION:
871 		/*
872 		 * We only need to use a custom IP header if the administrator
873 		 * has supplied a non-default hoplimit.
874 		 */
875 		if (iptun->iptun_hoplimit == IPTUN_DEFAULT_HOPLIMIT) {
876 			iptun->iptun_header_size = 0;
877 			break;
878 		}
879 		iptun->iptun_header_size = sizeof (ipha_t);
880 		iptun->iptun_header4.ipha_version_and_hdr_length =
881 		    IP_SIMPLE_HDR_VERSION;
882 		iptun->iptun_header4.ipha_fragment_offset_and_flags =
883 		    htons(IPH_DF);
884 		iptun->iptun_header4.ipha_ttl = iptun->iptun_hoplimit;
885 		break;
886 	case IPV6_VERSION: {
887 		ip6_t	*ip6hp = &iptun->iptun_header6.it6h_ip6h;
888 
889 		/*
890 		 * We only need to use a custom IPv6 header if either the
891 		 * administrator has supplied a non-default hoplimit, or we
892 		 * need to include an encapsulation limit option in the outer
893 		 * header.
894 		 */
895 		if (iptun->iptun_hoplimit == IPTUN_DEFAULT_HOPLIMIT &&
896 		    iptun->iptun_encaplimit == 0) {
897 			iptun->iptun_header_size = 0;
898 			break;
899 		}
900 
901 		(void) memset(ip6hp, 0, sizeof (*ip6hp));
902 		if (iptun->iptun_encaplimit == 0) {
903 			iptun->iptun_header_size = sizeof (ip6_t);
904 			ip6hp->ip6_nxt = IPPROTO_NONE;
905 		} else {
906 			iptun_encaplim_t	*iel;
907 
908 			iptun->iptun_header_size = sizeof (iptun_ipv6hdrs_t);
909 			/*
910 			 * The mac_ipv6 plugin requires ip6_plen to be in host
911 			 * byte order and reflect the extension headers
912 			 * present in the template.  The actual network byte
913 			 * order ip6_plen will be set on a per-packet basis on
914 			 * transmit.
915 			 */
916 			ip6hp->ip6_plen = sizeof (*iel);
917 			ip6hp->ip6_nxt = IPPROTO_DSTOPTS;
918 			iel = &iptun->iptun_header6.it6h_encaplim;
919 			*iel = iptun_encaplim_init;
920 			iel->iel_telopt.ip6ot_encap_limit =
921 			    iptun->iptun_encaplimit;
922 		}
923 
924 		ip6hp->ip6_hlim = iptun->iptun_hoplimit;
925 		break;
926 	}
927 	}
928 
929 	if (update_mac)
930 		iptun_task_dispatch(iptun, IPTUN_TASK_PDATA_UPDATE);
931 }
932 
933 /*
934  * Insert inbound and outbound IPv4 and IPv6 policy into the given policy
935  * head.
936  */
937 static boolean_t
938 iptun_insert_simple_policies(ipsec_policy_head_t *ph, ipsec_act_t *actp,
939     uint_t n, netstack_t *ns)
940 {
941 	int f = IPSEC_AF_V4;
942 
943 	if (!ipsec_polhead_insert(ph, actp, n, f, IPSEC_TYPE_INBOUND, ns) ||
944 	    !ipsec_polhead_insert(ph, actp, n, f, IPSEC_TYPE_OUTBOUND, ns))
945 		return (B_FALSE);
946 
947 	f = IPSEC_AF_V6;
948 	return (ipsec_polhead_insert(ph, actp, n, f, IPSEC_TYPE_INBOUND, ns) &&
949 	    ipsec_polhead_insert(ph, actp, n, f, IPSEC_TYPE_OUTBOUND, ns));
950 }
951 
952 /*
953  * Used to set IPsec policy when policy is set through the IPTUN_CREATE or
954  * IPTUN_MODIFY ioctls.
955  */
956 static int
957 iptun_set_sec_simple(iptun_t *iptun, const ipsec_req_t *ipsr)
958 {
959 	int		rc = 0;
960 	uint_t		nact;
961 	ipsec_act_t	*actp = NULL;
962 	boolean_t	clear_all, old_policy = B_FALSE;
963 	ipsec_tun_pol_t	*itp;
964 	char		name[MAXLINKNAMELEN];
965 	uint64_t	gen;
966 	netstack_t	*ns = iptun->iptun_ns;
967 
968 	/* Can't specify self-encap on a tunnel. */
969 	if (ipsr->ipsr_self_encap_req != 0)
970 		return (EINVAL);
971 
972 	/*
973 	 * If it's a "clear-all" entry, unset the security flags and resume
974 	 * normal cleartext (or inherit-from-global) policy.
975 	 */
976 	clear_all = ((ipsr->ipsr_ah_req & IPTUN_IPSEC_REQ_MASK) == 0 &&
977 	    (ipsr->ipsr_esp_req & IPTUN_IPSEC_REQ_MASK) == 0);
978 
979 	ASSERT(mutex_owned(&iptun->iptun_lock));
980 	itp = iptun->iptun_itp;
981 	if (itp == NULL) {
982 		if (clear_all)
983 			goto bail;
984 		if ((rc = dls_mgmt_get_linkinfo(iptun->iptun_linkid, name, NULL,
985 		    NULL, NULL)) != 0)
986 			goto bail;
987 		ASSERT(name[0] != '\0');
988 		if ((itp = create_tunnel_policy(name, &rc, &gen, ns)) == NULL)
989 			goto bail;
990 		iptun->iptun_itp = itp;
991 	}
992 
993 	/* Allocate the actvec now, before holding itp or polhead locks. */
994 	ipsec_actvec_from_req(ipsr, &actp, &nact, ns);
995 	if (actp == NULL) {
996 		rc = ENOMEM;
997 		goto bail;
998 	}
999 
1000 	/*
1001 	 * Just write on the active polhead.  Save the primary/secondary stuff
1002 	 * for spdsock operations.
1003 	 *
1004 	 * Mutex because we need to write to the polhead AND flags atomically.
1005 	 * Other threads will acquire the polhead lock as a reader if the
1006 	 * (unprotected) flag is set.
1007 	 */
1008 	mutex_enter(&itp->itp_lock);
1009 	if (itp->itp_flags & ITPF_P_TUNNEL) {
1010 		/* Oops, we lost a race.  Let's get out of here. */
1011 		rc = EBUSY;
1012 		goto mutex_bail;
1013 	}
1014 	old_policy = ((itp->itp_flags & ITPF_P_ACTIVE) != 0);
1015 
1016 	if (old_policy) {
1017 		ITPF_CLONE(itp->itp_flags);
1018 		rc = ipsec_copy_polhead(itp->itp_policy, itp->itp_inactive, ns);
1019 		if (rc != 0) {
1020 			/* inactive has already been cleared. */
1021 			itp->itp_flags &= ~ITPF_IFLAGS;
1022 			goto mutex_bail;
1023 		}
1024 		rw_enter(&itp->itp_policy->iph_lock, RW_WRITER);
1025 		ipsec_polhead_flush(itp->itp_policy, ns);
1026 	} else {
1027 		/* Else assume itp->itp_policy is already flushed. */
1028 		rw_enter(&itp->itp_policy->iph_lock, RW_WRITER);
1029 	}
1030 
1031 	if (clear_all) {
1032 		ASSERT(avl_numnodes(&itp->itp_policy->iph_rulebyid) == 0);
1033 		itp->itp_flags &= ~ITPF_PFLAGS;
1034 		rw_exit(&itp->itp_policy->iph_lock);
1035 		old_policy = B_FALSE;	/* Clear out the inactive one too. */
1036 		goto recover_bail;
1037 	}
1038 
1039 	if (iptun_insert_simple_policies(itp->itp_policy, actp, nact, ns)) {
1040 		rw_exit(&itp->itp_policy->iph_lock);
1041 		/*
1042 		 * Adjust MTU and make sure the DL side knows what's up.
1043 		 */
1044 		itp->itp_flags = ITPF_P_ACTIVE;
1045 		(void) iptun_update_mtu(iptun, NULL, 0);
1046 		old_policy = B_FALSE;	/* Blank out inactive - we succeeded */
1047 	} else {
1048 		rw_exit(&itp->itp_policy->iph_lock);
1049 		rc = ENOMEM;
1050 	}
1051 
1052 recover_bail:
1053 	if (old_policy) {
1054 		/* Recover policy in in active polhead. */
1055 		ipsec_swap_policy(itp->itp_policy, itp->itp_inactive, ns);
1056 		ITPF_SWAP(itp->itp_flags);
1057 	}
1058 
1059 	/* Clear policy in inactive polhead. */
1060 	itp->itp_flags &= ~ITPF_IFLAGS;
1061 	rw_enter(&itp->itp_inactive->iph_lock, RW_WRITER);
1062 	ipsec_polhead_flush(itp->itp_inactive, ns);
1063 	rw_exit(&itp->itp_inactive->iph_lock);
1064 
1065 mutex_bail:
1066 	mutex_exit(&itp->itp_lock);
1067 
1068 bail:
1069 	if (actp != NULL)
1070 		ipsec_actvec_free(actp, nact);
1071 
1072 	return (rc);
1073 }
1074 
1075 static iptun_typeinfo_t *
1076 iptun_gettypeinfo(iptun_type_t type)
1077 {
1078 	int i;
1079 
1080 	for (i = 0; iptun_type_table[i].iti_type != IPTUN_TYPE_UNKNOWN; i++) {
1081 		if (iptun_type_table[i].iti_type == type)
1082 			break;
1083 	}
1084 	return (&iptun_type_table[i]);
1085 }
1086 
1087 /*
1088  * Set the parameters included in ik on the tunnel iptun.  Parameters that can
1089  * only be set at creation time are set in iptun_create().
1090  */
1091 static int
1092 iptun_setparams(iptun_t *iptun, const iptun_kparams_t *ik)
1093 {
1094 	int		err = 0;
1095 	netstack_t	*ns = iptun->iptun_ns;
1096 	iptun_addr_t	orig_laddr, orig_raddr;
1097 	uint_t		orig_flags = iptun->iptun_flags;
1098 
1099 	if (ik->iptun_kparam_flags & IPTUN_KPARAM_LADDR) {
1100 		if (orig_flags & IPTUN_LADDR)
1101 			orig_laddr = iptun->iptun_laddr;
1102 		if ((err = iptun_setladdr(iptun, &ik->iptun_kparam_laddr)) != 0)
1103 			return (err);
1104 		iptun->iptun_flags |= IPTUN_LADDR;
1105 	}
1106 
1107 	if (ik->iptun_kparam_flags & IPTUN_KPARAM_RADDR) {
1108 		if (orig_flags & IPTUN_RADDR)
1109 			orig_raddr = iptun->iptun_raddr;
1110 		if ((err = iptun_setraddr(iptun, &ik->iptun_kparam_raddr)) != 0)
1111 			goto done;
1112 		iptun->iptun_flags |= IPTUN_RADDR;
1113 	}
1114 
1115 	if (ik->iptun_kparam_flags & IPTUN_KPARAM_SECINFO) {
1116 		/*
1117 		 * Set IPsec policy originating from the ifconfig(1M) command
1118 		 * line.  This is traditionally called "simple" policy because
1119 		 * the ipsec_req_t (iptun_kparam_secinfo) can only describe a
1120 		 * simple policy of "do ESP on everything" and/or "do AH on
1121 		 * everything" (as opposed to the rich policy that can be
1122 		 * defined with ipsecconf(1M)).
1123 		 */
1124 		if (iptun->iptun_typeinfo->iti_type == IPTUN_TYPE_6TO4) {
1125 			/*
1126 			 * Can't set security properties for automatic
1127 			 * tunnels.
1128 			 */
1129 			err = EINVAL;
1130 			goto done;
1131 		}
1132 
1133 		if (!ipsec_loaded(ns->netstack_ipsec)) {
1134 			/* If IPsec can be loaded, try and load it now. */
1135 			if (ipsec_failed(ns->netstack_ipsec)) {
1136 				err = EPROTONOSUPPORT;
1137 				goto done;
1138 			}
1139 			ipsec_loader_loadnow(ns->netstack_ipsec);
1140 			/*
1141 			 * ipsec_loader_loadnow() returns while IPsec is
1142 			 * loaded asynchronously.  While a method exists to
1143 			 * wait for IPsec to load (ipsec_loader_wait()), it
1144 			 * requires use of a STREAMS queue to do a qwait().
1145 			 * We're not in STREAMS context here, and so we can't
1146 			 * use it.  This is not a problem in practice because
1147 			 * in the vast majority of cases, key management and
1148 			 * global policy will have loaded before any tunnels
1149 			 * are plumbed, and so IPsec will already have been
1150 			 * loaded.
1151 			 */
1152 			err = EAGAIN;
1153 			goto done;
1154 		}
1155 
1156 		err = iptun_set_sec_simple(iptun, &ik->iptun_kparam_secinfo);
1157 		if (err == 0) {
1158 			iptun->iptun_flags |= IPTUN_SIMPLE_POLICY;
1159 			iptun->iptun_simple_policy = ik->iptun_kparam_secinfo;
1160 		}
1161 	}
1162 done:
1163 	if (err != 0) {
1164 		/* Restore original source and destination. */
1165 		if (ik->iptun_kparam_flags & IPTUN_KPARAM_LADDR &&
1166 		    (orig_flags & IPTUN_LADDR))
1167 			iptun->iptun_laddr = orig_laddr;
1168 		if ((ik->iptun_kparam_flags & IPTUN_KPARAM_RADDR) &&
1169 		    (orig_flags & IPTUN_RADDR))
1170 			iptun->iptun_raddr = orig_raddr;
1171 		iptun->iptun_flags = orig_flags;
1172 	}
1173 	return (err);
1174 }
1175 
1176 static int
1177 iptun_register(iptun_t *iptun)
1178 {
1179 	mac_register_t	*mac;
1180 	int		err;
1181 
1182 	ASSERT(!(iptun->iptun_flags & IPTUN_MAC_REGISTERED));
1183 
1184 	if ((mac = mac_alloc(MAC_VERSION)) == NULL)
1185 		return (EINVAL);
1186 
1187 	mac->m_type_ident = iptun->iptun_typeinfo->iti_ident;
1188 	mac->m_driver = iptun;
1189 	mac->m_dip = iptun_dip;
1190 	mac->m_instance = (uint_t)-1;
1191 	mac->m_src_addr = (uint8_t *)&iptun->iptun_laddr.ia_addr;
1192 	mac->m_dst_addr = iptun->iptun_typeinfo->iti_hasraddr ?
1193 	    (uint8_t *)&iptun->iptun_raddr.ia_addr : NULL;
1194 	mac->m_callbacks = &iptun_m_callbacks;
1195 	mac->m_min_sdu = iptun->iptun_typeinfo->iti_minmtu;
1196 	mac->m_max_sdu = iptun->iptun_mtu;
1197 	if (iptun->iptun_header_size != 0) {
1198 		mac->m_pdata = &iptun->iptun_header;
1199 		mac->m_pdata_size = iptun->iptun_header_size;
1200 	}
1201 	if ((err = mac_register(mac, &iptun->iptun_mh)) == 0)
1202 		iptun->iptun_flags |= IPTUN_MAC_REGISTERED;
1203 	mac_free(mac);
1204 	return (err);
1205 }
1206 
1207 static int
1208 iptun_unregister(iptun_t *iptun)
1209 {
1210 	int err;
1211 
1212 	ASSERT(iptun->iptun_flags & IPTUN_MAC_REGISTERED);
1213 	if ((err = mac_unregister(iptun->iptun_mh)) == 0)
1214 		iptun->iptun_flags &= ~IPTUN_MAC_REGISTERED;
1215 	return (err);
1216 }
1217 
1218 static conn_t *
1219 iptun_conn_create(iptun_t *iptun, netstack_t *ns, cred_t *credp)
1220 {
1221 	conn_t *connp;
1222 
1223 	if ((connp = ipcl_conn_create(IPCL_IPCCONN, KM_NOSLEEP, ns)) == NULL)
1224 		return (NULL);
1225 
1226 	connp->conn_flags |= IPCL_IPTUN;
1227 	connp->conn_iptun = iptun;
1228 	connp->conn_recv = iptun_input;
1229 	connp->conn_recvicmp = iptun_input_icmp;
1230 	connp->conn_verifyicmp = iptun_verifyicmp;
1231 
1232 	/*
1233 	 * Register iptun_notify to listen to capability changes detected by IP.
1234 	 * This upcall is made in the context of the call to conn_ip_output.
1235 	 */
1236 	connp->conn_ixa->ixa_notify = iptun_notify;
1237 	connp->conn_ixa->ixa_notify_cookie = iptun;
1238 
1239 	/*
1240 	 * For exclusive stacks we set conn_zoneid to GLOBAL_ZONEID as is done
1241 	 * for all other conn_t's.
1242 	 *
1243 	 * Note that there's an important distinction between iptun_zoneid and
1244 	 * conn_zoneid.  The conn_zoneid is set to GLOBAL_ZONEID in non-global
1245 	 * exclusive stack zones to make the ip module believe that the
1246 	 * non-global zone is actually a global zone.  Therefore, when
1247 	 * interacting with the ip module, we must always use conn_zoneid.
1248 	 */
1249 	connp->conn_zoneid = (ns->netstack_stackid == GLOBAL_NETSTACKID) ?
1250 	    crgetzoneid(credp) : GLOBAL_ZONEID;
1251 	connp->conn_cred = credp;
1252 	/* crfree() is done in ipcl_conn_destroy(), called by CONN_DEC_REF() */
1253 	crhold(connp->conn_cred);
1254 	connp->conn_cpid = NOPID;
1255 
1256 	/* conn_allzones can not be set this early, hence no IPCL_ZONEID */
1257 	connp->conn_ixa->ixa_zoneid = connp->conn_zoneid;
1258 	ASSERT(connp->conn_ref == 1);
1259 
1260 	/* Cache things in ixa without an extra refhold */
1261 	ASSERT(!(connp->conn_ixa->ixa_free_flags & IXA_FREE_CRED));
1262 	connp->conn_ixa->ixa_cred = connp->conn_cred;
1263 	connp->conn_ixa->ixa_cpid = connp->conn_cpid;
1264 	if (is_system_labeled())
1265 		connp->conn_ixa->ixa_tsl = crgetlabel(connp->conn_cred);
1266 
1267 	/*
1268 	 * Have conn_ip_output drop packets should our outer source
1269 	 * go invalid
1270 	 */
1271 	connp->conn_ixa->ixa_flags |= IXAF_VERIFY_SOURCE;
1272 
1273 	switch (iptun->iptun_typeinfo->iti_ipvers) {
1274 	case IPV4_VERSION:
1275 		connp->conn_family = AF_INET6;
1276 		break;
1277 	case IPV6_VERSION:
1278 		connp->conn_family = AF_INET;
1279 		break;
1280 	}
1281 	mutex_enter(&connp->conn_lock);
1282 	connp->conn_state_flags &= ~CONN_INCIPIENT;
1283 	mutex_exit(&connp->conn_lock);
1284 	return (connp);
1285 }
1286 
1287 static void
1288 iptun_conn_destroy(conn_t *connp)
1289 {
1290 	ip_quiesce_conn(connp);
1291 	connp->conn_iptun = NULL;
1292 	ASSERT(connp->conn_ref == 1);
1293 	CONN_DEC_REF(connp);
1294 }
1295 
1296 static iptun_t *
1297 iptun_alloc(void)
1298 {
1299 	iptun_t *iptun;
1300 
1301 	if ((iptun = kmem_cache_alloc(iptun_cache, KM_NOSLEEP)) != NULL) {
1302 		bzero(iptun, sizeof (*iptun));
1303 		atomic_inc_32(&iptun_tunnelcount);
1304 	}
1305 	return (iptun);
1306 }
1307 
1308 static void
1309 iptun_free(iptun_t *iptun)
1310 {
1311 	ASSERT(iptun->iptun_flags & IPTUN_CONDEMNED);
1312 
1313 	if (iptun->iptun_flags & IPTUN_HASH_INSERTED) {
1314 		iptun_stack_t	*iptuns = iptun->iptun_iptuns;
1315 
1316 		mutex_enter(&iptun_hash_lock);
1317 		VERIFY(mod_hash_remove(iptun_hash,
1318 		    IPTUN_HASH_KEY(iptun->iptun_linkid),
1319 		    (mod_hash_val_t *)&iptun) == 0);
1320 		mutex_exit(&iptun_hash_lock);
1321 		iptun->iptun_flags &= ~IPTUN_HASH_INSERTED;
1322 		mutex_enter(&iptuns->iptuns_lock);
1323 		list_remove(&iptuns->iptuns_iptunlist, iptun);
1324 		mutex_exit(&iptuns->iptuns_lock);
1325 	}
1326 
1327 	if (iptun->iptun_flags & IPTUN_BOUND)
1328 		iptun_unbind(iptun);
1329 
1330 	/*
1331 	 * After iptun_unregister(), there will be no threads executing a
1332 	 * downcall from the mac module, including in the tx datapath.
1333 	 */
1334 	if (iptun->iptun_flags & IPTUN_MAC_REGISTERED)
1335 		VERIFY(iptun_unregister(iptun) == 0);
1336 
1337 	if (iptun->iptun_itp != NULL) {
1338 		/*
1339 		 * Remove from the AVL tree, AND release the reference iptun_t
1340 		 * itself holds on the ITP.
1341 		 */
1342 		itp_unlink(iptun->iptun_itp, iptun->iptun_ns);
1343 		ITP_REFRELE(iptun->iptun_itp, iptun->iptun_ns);
1344 		iptun->iptun_itp = NULL;
1345 		iptun->iptun_flags &= ~IPTUN_SIMPLE_POLICY;
1346 	}
1347 
1348 	/*
1349 	 * After ipcl_conn_destroy(), there will be no threads executing an
1350 	 * upcall from ip (i.e., iptun_input()), and it is then safe to free
1351 	 * the iptun_t.
1352 	 */
1353 	if (iptun->iptun_connp != NULL) {
1354 		iptun_conn_destroy(iptun->iptun_connp);
1355 		iptun->iptun_connp = NULL;
1356 	}
1357 
1358 	kmem_cache_free(iptun_cache, iptun);
1359 	atomic_dec_32(&iptun_tunnelcount);
1360 }
1361 
1362 int
1363 iptun_create(iptun_kparams_t *ik, cred_t *credp)
1364 {
1365 	iptun_t		*iptun = NULL;
1366 	int		err = 0, mherr;
1367 	char		linkname[MAXLINKNAMELEN];
1368 	ipsec_tun_pol_t	*itp;
1369 	netstack_t	*ns = NULL;
1370 	iptun_stack_t	*iptuns;
1371 	datalink_id_t	tmpid;
1372 	zoneid_t	zoneid = crgetzoneid(credp);
1373 	boolean_t	link_created = B_FALSE;
1374 
1375 	/* The tunnel type is mandatory */
1376 	if (!(ik->iptun_kparam_flags & IPTUN_KPARAM_TYPE))
1377 		return (EINVAL);
1378 
1379 	/*
1380 	 * Is the linkid that the caller wishes to associate with this new
1381 	 * tunnel assigned to this zone?
1382 	 */
1383 	if (zone_check_datalink(&zoneid, ik->iptun_kparam_linkid) != 0) {
1384 		if (zoneid != GLOBAL_ZONEID)
1385 			return (EINVAL);
1386 	} else if (zoneid == GLOBAL_ZONEID) {
1387 		return (EINVAL);
1388 	}
1389 
1390 	/*
1391 	 * Make sure that we're not trying to create a tunnel that has already
1392 	 * been created.
1393 	 */
1394 	if (iptun_enter_by_linkid(ik->iptun_kparam_linkid, &iptun) == 0) {
1395 		iptun_exit(iptun);
1396 		iptun = NULL;
1397 		err = EEXIST;
1398 		goto done;
1399 	}
1400 
1401 	ns = netstack_find_by_cred(credp);
1402 	iptuns = ns->netstack_iptun;
1403 
1404 	if ((iptun = iptun_alloc()) == NULL) {
1405 		err = ENOMEM;
1406 		goto done;
1407 	}
1408 
1409 	iptun->iptun_linkid = ik->iptun_kparam_linkid;
1410 	iptun->iptun_zoneid = zoneid;
1411 	iptun->iptun_ns = ns;
1412 
1413 	iptun->iptun_typeinfo = iptun_gettypeinfo(ik->iptun_kparam_type);
1414 	if (iptun->iptun_typeinfo->iti_type == IPTUN_TYPE_UNKNOWN) {
1415 		err = EINVAL;
1416 		goto done;
1417 	}
1418 
1419 	if (ik->iptun_kparam_flags & IPTUN_KPARAM_IMPLICIT)
1420 		iptun->iptun_flags |= IPTUN_IMPLICIT;
1421 
1422 	if ((err = iptun_setparams(iptun, ik)) != 0)
1423 		goto done;
1424 
1425 	iptun->iptun_hoplimit = IPTUN_DEFAULT_HOPLIMIT;
1426 	if (iptun->iptun_typeinfo->iti_type == IPTUN_TYPE_IPV6)
1427 		iptun->iptun_encaplimit = IPTUN_DEFAULT_ENCAPLIMIT;
1428 
1429 	iptun_headergen(iptun, B_FALSE);
1430 
1431 	iptun->iptun_connp = iptun_conn_create(iptun, ns, credp);
1432 	if (iptun->iptun_connp == NULL) {
1433 		err = ENOMEM;
1434 		goto done;
1435 	}
1436 
1437 	iptun->iptun_mtu = iptun->iptun_typeinfo->iti_maxmtu;
1438 	iptun->iptun_dpmtu = iptun->iptun_mtu;
1439 
1440 	/*
1441 	 * Find an ITP based on linkname.  If we have parms already set via
1442 	 * the iptun_setparams() call above, it may have created an ITP for
1443 	 * us.  We always try get_tunnel_policy() for DEBUG correctness
1444 	 * checks, and we may wish to refactor this to only check when
1445 	 * iptun_itp is NULL.
1446 	 */
1447 	if ((err = dls_mgmt_get_linkinfo(iptun->iptun_linkid, linkname, NULL,
1448 	    NULL, NULL)) != 0)
1449 		goto done;
1450 	if ((itp = get_tunnel_policy(linkname, ns)) != NULL)
1451 		iptun->iptun_itp = itp;
1452 
1453 	/*
1454 	 * See if we have the necessary IP addresses assigned to this tunnel
1455 	 * to try and bind them with ip underneath us.  If we're not ready to
1456 	 * bind yet, then we'll defer the bind operation until the addresses
1457 	 * are modified.
1458 	 */
1459 	if (iptun_canbind(iptun) && ((err = iptun_bind(iptun)) != 0))
1460 		goto done;
1461 
1462 	if ((err = iptun_register(iptun)) != 0)
1463 		goto done;
1464 
1465 	err = dls_devnet_create(iptun->iptun_mh, iptun->iptun_linkid,
1466 	    iptun->iptun_zoneid);
1467 	if (err != 0)
1468 		goto done;
1469 	link_created = B_TRUE;
1470 
1471 	/*
1472 	 * We hash by link-id as that is the key used by all other iptun
1473 	 * interfaces (modify, delete, etc.).
1474 	 */
1475 	if ((mherr = mod_hash_insert(iptun_hash,
1476 	    IPTUN_HASH_KEY(iptun->iptun_linkid), (mod_hash_val_t)iptun)) == 0) {
1477 		mutex_enter(&iptuns->iptuns_lock);
1478 		list_insert_head(&iptuns->iptuns_iptunlist, iptun);
1479 		mutex_exit(&iptuns->iptuns_lock);
1480 		iptun->iptun_flags |= IPTUN_HASH_INSERTED;
1481 	} else if (mherr == MH_ERR_NOMEM) {
1482 		err = ENOMEM;
1483 	} else if (mherr == MH_ERR_DUPLICATE) {
1484 		err = EEXIST;
1485 	} else {
1486 		err = EINVAL;
1487 	}
1488 
1489 done:
1490 	if (iptun == NULL && ns != NULL)
1491 		netstack_rele(ns);
1492 	if (err != 0 && iptun != NULL) {
1493 		if (link_created) {
1494 			(void) dls_devnet_destroy(iptun->iptun_mh, &tmpid,
1495 			    B_TRUE);
1496 		}
1497 		iptun->iptun_flags |= IPTUN_CONDEMNED;
1498 		iptun_free(iptun);
1499 	}
1500 	return (err);
1501 }
1502 
1503 int
1504 iptun_delete(datalink_id_t linkid, cred_t *credp)
1505 {
1506 	int	err;
1507 	iptun_t	*iptun = NULL;
1508 
1509 	if ((err = iptun_enter_by_linkid(linkid, &iptun)) != 0)
1510 		return (err);
1511 
1512 	/* One cannot delete a tunnel that belongs to another zone. */
1513 	if (iptun->iptun_zoneid != crgetzoneid(credp)) {
1514 		iptun_exit(iptun);
1515 		return (EACCES);
1516 	}
1517 
1518 	/*
1519 	 * We need to exit iptun in order to issue calls up the stack such as
1520 	 * dls_devnet_destroy().  If we call up while still in iptun, deadlock
1521 	 * with calls coming down the stack is possible.  We prevent other
1522 	 * threads from entering this iptun after we've exited it by setting
1523 	 * the IPTUN_DELETE_PENDING flag.  This will cause callers of
1524 	 * iptun_enter() to block waiting on iptun_enter_cv.  The assumption
1525 	 * here is that the functions we're calling while IPTUN_DELETE_PENDING
1526 	 * is set dont resuult in an iptun_enter() call, as that would result
1527 	 * in deadlock.
1528 	 */
1529 	iptun->iptun_flags |= IPTUN_DELETE_PENDING;
1530 
1531 	/* Wait for any pending upcall to the mac module to complete. */
1532 	while (iptun->iptun_flags & IPTUN_UPCALL_PENDING)
1533 		cv_wait(&iptun->iptun_upcall_cv, &iptun->iptun_lock);
1534 
1535 	iptun_exit(iptun);
1536 
1537 	if ((err = dls_devnet_destroy(iptun->iptun_mh, &linkid, B_TRUE)) == 0) {
1538 		/*
1539 		 * mac_disable() will fail with EBUSY if there are references
1540 		 * to the iptun MAC.  If there are none, then mac_disable()
1541 		 * will assure that none can be acquired until the MAC is
1542 		 * unregistered.
1543 		 *
1544 		 * XXX CR 6791335 prevents us from calling mac_disable() prior
1545 		 * to dls_devnet_destroy(), so we unfortunately need to
1546 		 * attempt to re-create the devnet node if mac_disable()
1547 		 * fails.
1548 		 */
1549 		if ((err = mac_disable(iptun->iptun_mh)) != 0) {
1550 			(void) dls_devnet_create(iptun->iptun_mh, linkid,
1551 			    iptun->iptun_zoneid);
1552 		}
1553 	}
1554 
1555 	/*
1556 	 * Now that we know the fate of this iptun_t, we need to clear
1557 	 * IPTUN_DELETE_PENDING, and set IPTUN_CONDEMNED if the iptun_t is
1558 	 * slated to be freed.  Either way, we need to signal the threads
1559 	 * waiting in iptun_enter() so that they can either fail if
1560 	 * IPTUN_CONDEMNED is set, or continue if it's not.
1561 	 */
1562 	mutex_enter(&iptun->iptun_lock);
1563 	iptun->iptun_flags &= ~IPTUN_DELETE_PENDING;
1564 	if (err == 0)
1565 		iptun->iptun_flags |= IPTUN_CONDEMNED;
1566 	cv_broadcast(&iptun->iptun_enter_cv);
1567 	mutex_exit(&iptun->iptun_lock);
1568 
1569 	/*
1570 	 * Note that there is no danger in calling iptun_free() after having
1571 	 * dropped the iptun_lock since callers of iptun_enter() at this point
1572 	 * are doing so from iptun_enter_by_linkid() (mac_disable() got rid of
1573 	 * threads entering from mac callbacks which call iptun_enter()
1574 	 * directly) which holds iptun_hash_lock, and iptun_free() grabs this
1575 	 * lock in order to remove the iptun_t from the hash table.
1576 	 */
1577 	if (err == 0)
1578 		iptun_free(iptun);
1579 
1580 	return (err);
1581 }
1582 
1583 int
1584 iptun_modify(const iptun_kparams_t *ik, cred_t *credp)
1585 {
1586 	iptun_t		*iptun;
1587 	boolean_t	laddr_change = B_FALSE, raddr_change = B_FALSE;
1588 	int		err;
1589 
1590 	if ((err = iptun_enter_by_linkid(ik->iptun_kparam_linkid, &iptun)) != 0)
1591 		return (err);
1592 
1593 	/* One cannot modify a tunnel that belongs to another zone. */
1594 	if (iptun->iptun_zoneid != crgetzoneid(credp)) {
1595 		err = EACCES;
1596 		goto done;
1597 	}
1598 
1599 	/* The tunnel type cannot be changed */
1600 	if (ik->iptun_kparam_flags & IPTUN_KPARAM_TYPE) {
1601 		err = EINVAL;
1602 		goto done;
1603 	}
1604 
1605 	if ((err = iptun_setparams(iptun, ik)) != 0)
1606 		goto done;
1607 	iptun_headergen(iptun, B_FALSE);
1608 
1609 	/*
1610 	 * If any of the tunnel's addresses has been modified and the tunnel
1611 	 * has the necessary addresses assigned to it, we need to try to bind
1612 	 * with ip underneath us.  If we're not ready to bind yet, then we'll
1613 	 * try again when the addresses are modified later.
1614 	 */
1615 	laddr_change = (ik->iptun_kparam_flags & IPTUN_KPARAM_LADDR);
1616 	raddr_change = (ik->iptun_kparam_flags & IPTUN_KPARAM_RADDR);
1617 	if (laddr_change || raddr_change) {
1618 		if (iptun->iptun_flags & IPTUN_BOUND)
1619 			iptun_unbind(iptun);
1620 		if (iptun_canbind(iptun) && (err = iptun_bind(iptun)) != 0) {
1621 			if (laddr_change)
1622 				iptun->iptun_flags &= ~IPTUN_LADDR;
1623 			if (raddr_change)
1624 				iptun->iptun_flags &= ~IPTUN_RADDR;
1625 			goto done;
1626 		}
1627 	}
1628 
1629 	if (laddr_change)
1630 		iptun_task_dispatch(iptun, IPTUN_TASK_LADDR_UPDATE);
1631 	if (raddr_change)
1632 		iptun_task_dispatch(iptun, IPTUN_TASK_RADDR_UPDATE);
1633 
1634 done:
1635 	iptun_exit(iptun);
1636 	return (err);
1637 }
1638 
1639 /* Given an IP tunnel's datalink id, fill in its parameters. */
1640 int
1641 iptun_info(iptun_kparams_t *ik, cred_t *credp)
1642 {
1643 	iptun_t	*iptun;
1644 	int	err;
1645 
1646 	/* Is the tunnel link visible from the caller's zone? */
1647 	if (!dls_devnet_islinkvisible(ik->iptun_kparam_linkid,
1648 	    crgetzoneid(credp)))
1649 		return (ENOENT);
1650 
1651 	if ((err = iptun_enter_by_linkid(ik->iptun_kparam_linkid, &iptun)) != 0)
1652 		return (err);
1653 
1654 	bzero(ik, sizeof (iptun_kparams_t));
1655 
1656 	ik->iptun_kparam_linkid = iptun->iptun_linkid;
1657 	ik->iptun_kparam_type = iptun->iptun_typeinfo->iti_type;
1658 	ik->iptun_kparam_flags |= IPTUN_KPARAM_TYPE;
1659 
1660 	if (iptun->iptun_flags & IPTUN_LADDR) {
1661 		iptun_getaddr(&iptun->iptun_laddr, &ik->iptun_kparam_laddr);
1662 		ik->iptun_kparam_flags |= IPTUN_KPARAM_LADDR;
1663 	}
1664 	if (iptun->iptun_flags & IPTUN_RADDR) {
1665 		iptun_getaddr(&iptun->iptun_raddr, &ik->iptun_kparam_raddr);
1666 		ik->iptun_kparam_flags |= IPTUN_KPARAM_RADDR;
1667 	}
1668 
1669 	if (iptun->iptun_flags & IPTUN_IMPLICIT)
1670 		ik->iptun_kparam_flags |= IPTUN_KPARAM_IMPLICIT;
1671 
1672 	if (iptun->iptun_itp != NULL) {
1673 		mutex_enter(&iptun->iptun_itp->itp_lock);
1674 		if (iptun->iptun_itp->itp_flags & ITPF_P_ACTIVE) {
1675 			ik->iptun_kparam_flags |= IPTUN_KPARAM_IPSECPOL;
1676 			if (iptun->iptun_flags & IPTUN_SIMPLE_POLICY) {
1677 				ik->iptun_kparam_flags |= IPTUN_KPARAM_SECINFO;
1678 				ik->iptun_kparam_secinfo =
1679 				    iptun->iptun_simple_policy;
1680 			}
1681 		}
1682 		mutex_exit(&iptun->iptun_itp->itp_lock);
1683 	}
1684 
1685 done:
1686 	iptun_exit(iptun);
1687 	return (err);
1688 }
1689 
1690 int
1691 iptun_set_6to4relay(netstack_t *ns, ipaddr_t relay_addr)
1692 {
1693 	if (relay_addr == INADDR_BROADCAST || CLASSD(relay_addr))
1694 		return (EADDRNOTAVAIL);
1695 	ns->netstack_iptun->iptuns_relay_rtr_addr = relay_addr;
1696 	return (0);
1697 }
1698 
1699 void
1700 iptun_get_6to4relay(netstack_t *ns, ipaddr_t *relay_addr)
1701 {
1702 	*relay_addr = ns->netstack_iptun->iptuns_relay_rtr_addr;
1703 }
1704 
1705 void
1706 iptun_set_policy(datalink_id_t linkid, ipsec_tun_pol_t *itp)
1707 {
1708 	iptun_t	*iptun;
1709 
1710 	if (iptun_enter_by_linkid(linkid, &iptun) != 0)
1711 		return;
1712 	if (iptun->iptun_itp != itp) {
1713 		ASSERT(iptun->iptun_itp == NULL);
1714 		ITP_REFHOLD(itp);
1715 		iptun->iptun_itp = itp;
1716 	}
1717 	/*
1718 	 * IPsec policy means IPsec overhead, which means lower MTU.
1719 	 * Refresh the MTU for this tunnel.
1720 	 */
1721 	(void) iptun_update_mtu(iptun, NULL, 0);
1722 	iptun_exit(iptun);
1723 }
1724 
1725 /*
1726  * Obtain the path MTU to the tunnel destination.
1727  * Can return zero in some cases.
1728  */
1729 static uint32_t
1730 iptun_get_dst_pmtu(iptun_t *iptun, ip_xmit_attr_t *ixa)
1731 {
1732 	uint32_t	pmtu = 0;
1733 	conn_t		*connp = iptun->iptun_connp;
1734 	boolean_t	need_rele = B_FALSE;
1735 
1736 	/*
1737 	 * We only obtain the pmtu for tunnels that have a remote tunnel
1738 	 * address.
1739 	 */
1740 	if (!(iptun->iptun_flags & IPTUN_RADDR))
1741 		return (0);
1742 
1743 	if (ixa == NULL) {
1744 		ixa = conn_get_ixa(connp, B_FALSE);
1745 		if (ixa == NULL)
1746 			return (0);
1747 		need_rele = B_TRUE;
1748 	}
1749 	/*
1750 	 * Guard against ICMP errors before we have sent, as well as against
1751 	 * and a thread which held conn_ixa.
1752 	 */
1753 	if (ixa->ixa_ire != NULL) {
1754 		pmtu = ip_get_pmtu(ixa);
1755 
1756 		/*
1757 		 * For both IPv4 and IPv6 we can have indication that the outer
1758 		 * header needs fragmentation.
1759 		 */
1760 		if (ixa->ixa_flags & IXAF_PMTU_TOO_SMALL) {
1761 			/* Must allow fragmentation in ip_output */
1762 			ixa->ixa_flags &= ~IXAF_DONTFRAG;
1763 		} else if (iptun->iptun_typeinfo->iti_type != IPTUN_TYPE_6TO4) {
1764 			ixa->ixa_flags |= IXAF_DONTFRAG;
1765 		} else {
1766 			/* ip_get_pmtu might have set this - we don't want it */
1767 			ixa->ixa_flags &= ~IXAF_PMTU_IPV4_DF;
1768 		}
1769 	}
1770 
1771 	if (need_rele)
1772 		ixa_refrele(ixa);
1773 	return (pmtu);
1774 }
1775 
1776 /*
1777  * Update the ip_xmit_attr_t to capture the current lower path mtu as known
1778  * by ip.
1779  */
1780 static void
1781 iptun_update_dst_pmtu(iptun_t *iptun, ip_xmit_attr_t *ixa)
1782 {
1783 	uint32_t	pmtu;
1784 	conn_t		*connp = iptun->iptun_connp;
1785 	boolean_t	need_rele = B_FALSE;
1786 
1787 	/* IXAF_VERIFY_PMTU is not set if we don't have a fixed destination */
1788 	if (!(iptun->iptun_flags & IPTUN_RADDR))
1789 		return;
1790 
1791 	if (ixa == NULL) {
1792 		ixa = conn_get_ixa(connp, B_FALSE);
1793 		if (ixa == NULL)
1794 			return;
1795 		need_rele = B_TRUE;
1796 	}
1797 	/*
1798 	 * Guard against ICMP errors before we have sent, as well as against
1799 	 * and a thread which held conn_ixa.
1800 	 */
1801 	if (ixa->ixa_ire != NULL) {
1802 		pmtu = ip_get_pmtu(ixa);
1803 		/*
1804 		 * Update ixa_fragsize and ixa_pmtu.
1805 		 */
1806 		ixa->ixa_fragsize = ixa->ixa_pmtu = pmtu;
1807 
1808 		/*
1809 		 * For both IPv4 and IPv6 we can have indication that the outer
1810 		 * header needs fragmentation.
1811 		 */
1812 		if (ixa->ixa_flags & IXAF_PMTU_TOO_SMALL) {
1813 			/* Must allow fragmentation in ip_output */
1814 			ixa->ixa_flags &= ~IXAF_DONTFRAG;
1815 		} else if (iptun->iptun_typeinfo->iti_type != IPTUN_TYPE_6TO4) {
1816 			ixa->ixa_flags |= IXAF_DONTFRAG;
1817 		} else {
1818 			/* ip_get_pmtu might have set this - we don't want it */
1819 			ixa->ixa_flags &= ~IXAF_PMTU_IPV4_DF;
1820 		}
1821 	}
1822 
1823 	if (need_rele)
1824 		ixa_refrele(ixa);
1825 }
1826 
1827 /*
1828  * There is nothing that iptun can verify in addition to IP having
1829  * verified the IP addresses in the fanout.
1830  */
1831 /* ARGSUSED */
1832 static boolean_t
1833 iptun_verifyicmp(conn_t *connp, void *arg2, icmph_t *icmph, icmp6_t *icmp6,
1834     ip_recv_attr_t *ira)
1835 {
1836 	return (B_TRUE);
1837 }
1838 
1839 /*
1840  * Notify function registered with ip_xmit_attr_t.
1841  */
1842 static void
1843 iptun_notify(void *arg, ip_xmit_attr_t *ixa, ixa_notify_type_t ntype,
1844     ixa_notify_arg_t narg)
1845 {
1846 	iptun_t		*iptun = (iptun_t *)arg;
1847 
1848 	switch (ntype) {
1849 	case IXAN_PMTU:
1850 		(void) iptun_update_mtu(iptun, ixa, narg);
1851 		break;
1852 	}
1853 }
1854 
1855 /*
1856  * Returns the max of old_ovhd and the overhead associated with pol.
1857  */
1858 static uint32_t
1859 iptun_max_policy_overhead(ipsec_policy_t *pol, uint32_t old_ovhd)
1860 {
1861 	uint32_t new_ovhd = old_ovhd;
1862 
1863 	while (pol != NULL) {
1864 		new_ovhd = max(new_ovhd,
1865 		    ipsec_act_ovhd(&pol->ipsp_act->ipa_act));
1866 		pol = pol->ipsp_hash.hash_next;
1867 	}
1868 	return (new_ovhd);
1869 }
1870 
1871 static uint32_t
1872 iptun_get_ipsec_overhead(iptun_t *iptun)
1873 {
1874 	ipsec_policy_root_t	*ipr;
1875 	ipsec_policy_head_t	*iph;
1876 	ipsec_policy_t		*pol;
1877 	ipsec_selector_t	sel;
1878 	int			i;
1879 	uint32_t		ipsec_ovhd = 0;
1880 	ipsec_tun_pol_t		*itp = iptun->iptun_itp;
1881 	netstack_t		*ns = iptun->iptun_ns;
1882 
1883 	if (itp == NULL || !(itp->itp_flags & ITPF_P_ACTIVE)) {
1884 		/*
1885 		 * Consult global policy, just in case.  This will only work
1886 		 * if we have both source and destination addresses to work
1887 		 * with.
1888 		 */
1889 		if ((iptun->iptun_flags & (IPTUN_LADDR|IPTUN_RADDR)) !=
1890 		    (IPTUN_LADDR|IPTUN_RADDR))
1891 			return (0);
1892 
1893 		iph = ipsec_system_policy(ns);
1894 		bzero(&sel, sizeof (sel));
1895 		sel.ips_isv4 =
1896 		    (iptun->iptun_typeinfo->iti_ipvers == IPV4_VERSION);
1897 		switch (iptun->iptun_typeinfo->iti_ipvers) {
1898 		case IPV4_VERSION:
1899 			sel.ips_local_addr_v4 = iptun->iptun_laddr4;
1900 			sel.ips_remote_addr_v4 = iptun->iptun_raddr4;
1901 			break;
1902 		case IPV6_VERSION:
1903 			sel.ips_local_addr_v6 = iptun->iptun_laddr6;
1904 			sel.ips_remote_addr_v6 = iptun->iptun_raddr6;
1905 			break;
1906 		}
1907 		/* Check for both IPv4 and IPv6. */
1908 		sel.ips_protocol = IPPROTO_ENCAP;
1909 		pol = ipsec_find_policy_head(NULL, iph, IPSEC_TYPE_OUTBOUND,
1910 		    &sel);
1911 		if (pol != NULL) {
1912 			ipsec_ovhd = ipsec_act_ovhd(&pol->ipsp_act->ipa_act);
1913 			IPPOL_REFRELE(pol);
1914 		}
1915 		sel.ips_protocol = IPPROTO_IPV6;
1916 		pol = ipsec_find_policy_head(NULL, iph, IPSEC_TYPE_OUTBOUND,
1917 		    &sel);
1918 		if (pol != NULL) {
1919 			ipsec_ovhd = max(ipsec_ovhd,
1920 			    ipsec_act_ovhd(&pol->ipsp_act->ipa_act));
1921 			IPPOL_REFRELE(pol);
1922 		}
1923 		IPPH_REFRELE(iph, ns);
1924 	} else {
1925 		/*
1926 		 * Look through all of the possible IPsec actions for the
1927 		 * tunnel, and find the largest potential IPsec overhead.
1928 		 */
1929 		iph = itp->itp_policy;
1930 		rw_enter(&iph->iph_lock, RW_READER);
1931 		ipr = &(iph->iph_root[IPSEC_TYPE_OUTBOUND]);
1932 		ipsec_ovhd = iptun_max_policy_overhead(
1933 		    ipr->ipr_nonhash[IPSEC_AF_V4], 0);
1934 		ipsec_ovhd = iptun_max_policy_overhead(
1935 		    ipr->ipr_nonhash[IPSEC_AF_V6], ipsec_ovhd);
1936 		for (i = 0; i < ipr->ipr_nchains; i++) {
1937 			ipsec_ovhd = iptun_max_policy_overhead(
1938 			    ipr->ipr_hash[i].hash_head, ipsec_ovhd);
1939 		}
1940 		rw_exit(&iph->iph_lock);
1941 	}
1942 
1943 	return (ipsec_ovhd);
1944 }
1945 
1946 /*
1947  * Calculate and return the maximum possible upper MTU for the given tunnel.
1948  *
1949  * If new_pmtu is set then we also need to update the lower path MTU information
1950  * in the ip_xmit_attr_t. That is needed since we set IXAF_VERIFY_PMTU so that
1951  * we are notified by conn_ip_output() when the path MTU increases.
1952  */
1953 static uint32_t
1954 iptun_get_maxmtu(iptun_t *iptun, ip_xmit_attr_t *ixa, uint32_t new_pmtu)
1955 {
1956 	size_t		header_size, ipsec_overhead;
1957 	uint32_t	maxmtu, pmtu;
1958 
1959 	/*
1960 	 * Start with the path-MTU to the remote address, which is either
1961 	 * provided as the new_pmtu argument, or obtained using
1962 	 * iptun_get_dst_pmtu().
1963 	 */
1964 	if (new_pmtu != 0) {
1965 		if (iptun->iptun_flags & IPTUN_RADDR)
1966 			iptun->iptun_dpmtu = new_pmtu;
1967 		pmtu = new_pmtu;
1968 	} else if (iptun->iptun_flags & IPTUN_RADDR) {
1969 		if ((pmtu = iptun_get_dst_pmtu(iptun, ixa)) == 0) {
1970 			/*
1971 			 * We weren't able to obtain the path-MTU of the
1972 			 * destination.  Use the previous value.
1973 			 */
1974 			pmtu = iptun->iptun_dpmtu;
1975 		} else {
1976 			iptun->iptun_dpmtu = pmtu;
1977 		}
1978 	} else {
1979 		/*
1980 		 * We have no path-MTU information to go on, use the maximum
1981 		 * possible value.
1982 		 */
1983 		pmtu = iptun->iptun_typeinfo->iti_maxmtu;
1984 	}
1985 
1986 	/*
1987 	 * Now calculate tunneling overhead and subtract that from the
1988 	 * path-MTU information obtained above.
1989 	 */
1990 	if (iptun->iptun_header_size != 0) {
1991 		header_size = iptun->iptun_header_size;
1992 	} else {
1993 		switch (iptun->iptun_typeinfo->iti_ipvers) {
1994 		case IPV4_VERSION:
1995 			header_size = sizeof (ipha_t);
1996 			if (is_system_labeled())
1997 				header_size += IP_MAX_OPT_LENGTH;
1998 			break;
1999 		case IPV6_VERSION:
2000 			header_size = sizeof (iptun_ipv6hdrs_t);
2001 			break;
2002 		}
2003 	}
2004 
2005 	ipsec_overhead = iptun_get_ipsec_overhead(iptun);
2006 
2007 	maxmtu = pmtu - (header_size + ipsec_overhead);
2008 	return (max(maxmtu, iptun->iptun_typeinfo->iti_minmtu));
2009 }
2010 
2011 /*
2012  * Re-calculate the tunnel's MTU as seen from above and notify the MAC layer
2013  * of any change in MTU.  The new_pmtu argument is the new lower path MTU to
2014  * the tunnel destination to be used in the tunnel MTU calculation.  Passing
2015  * in 0 for new_pmtu causes the lower path MTU to be dynamically updated using
2016  * ip_get_pmtu().
2017  *
2018  * If the calculated tunnel MTU is different than its previous value, then we
2019  * notify the MAC layer above us of this change using mac_maxsdu_update().
2020  */
2021 static uint32_t
2022 iptun_update_mtu(iptun_t *iptun, ip_xmit_attr_t *ixa, uint32_t new_pmtu)
2023 {
2024 	uint32_t newmtu;
2025 
2026 	/* We always update the ixa since we might have set IXAF_VERIFY_PMTU */
2027 	iptun_update_dst_pmtu(iptun, ixa);
2028 
2029 	/*
2030 	 * We return the current MTU without updating it if it was pegged to a
2031 	 * static value using the MAC_PROP_MTU link property.
2032 	 */
2033 	if (iptun->iptun_flags & IPTUN_FIXED_MTU)
2034 		return (iptun->iptun_mtu);
2035 
2036 	/* If the MTU isn't fixed, then use the maximum possible value. */
2037 	newmtu = iptun_get_maxmtu(iptun, ixa, new_pmtu);
2038 	/*
2039 	 * We only dynamically adjust the tunnel MTU for tunnels with
2040 	 * destinations because dynamic MTU calculations are based on the
2041 	 * destination path-MTU.
2042 	 */
2043 	if ((iptun->iptun_flags & IPTUN_RADDR) && newmtu != iptun->iptun_mtu) {
2044 		iptun->iptun_mtu = newmtu;
2045 		if (iptun->iptun_flags & IPTUN_MAC_REGISTERED)
2046 			iptun_task_dispatch(iptun, IPTUN_TASK_MTU_UPDATE);
2047 	}
2048 
2049 	return (newmtu);
2050 }
2051 
2052 /*
2053  * Frees a packet or packet chain and bumps stat for each freed packet.
2054  */
2055 static void
2056 iptun_drop_pkt(mblk_t *mp, uint64_t *stat)
2057 {
2058 	mblk_t *pktmp;
2059 
2060 	for (pktmp = mp; pktmp != NULL; pktmp = mp) {
2061 		mp = mp->b_next;
2062 		pktmp->b_next = NULL;
2063 		if (stat != NULL)
2064 			atomic_inc_64(stat);
2065 		freemsg(pktmp);
2066 	}
2067 }
2068 
2069 /*
2070  * Allocate and return a new mblk to hold an IP and ICMP header, and chain the
2071  * original packet to its b_cont.  Returns NULL on failure.
2072  */
2073 static mblk_t *
2074 iptun_build_icmperr(size_t hdrs_size, mblk_t *orig_pkt)
2075 {
2076 	mblk_t *icmperr_mp;
2077 
2078 	if ((icmperr_mp = allocb(hdrs_size, BPRI_MED)) != NULL) {
2079 		icmperr_mp->b_wptr += hdrs_size;
2080 		/* tack on the offending packet */
2081 		icmperr_mp->b_cont = orig_pkt;
2082 	}
2083 	return (icmperr_mp);
2084 }
2085 
2086 /*
2087  * Transmit an ICMP error.  mp->b_rptr points at the packet to be included in
2088  * the ICMP error.
2089  */
2090 static void
2091 iptun_sendicmp_v4(iptun_t *iptun, icmph_t *icmp, ipha_t *orig_ipha, mblk_t *mp,
2092     ts_label_t *tsl)
2093 {
2094 	size_t	orig_pktsize, hdrs_size;
2095 	mblk_t	*icmperr_mp;
2096 	ipha_t	*new_ipha;
2097 	icmph_t	*new_icmp;
2098 	ip_xmit_attr_t	ixas;
2099 	conn_t	*connp = iptun->iptun_connp;
2100 
2101 	orig_pktsize = msgdsize(mp);
2102 	hdrs_size = sizeof (ipha_t) + sizeof (icmph_t);
2103 	if ((icmperr_mp = iptun_build_icmperr(hdrs_size, mp)) == NULL) {
2104 		iptun_drop_pkt(mp, &iptun->iptun_noxmtbuf);
2105 		return;
2106 	}
2107 
2108 	new_ipha = (ipha_t *)icmperr_mp->b_rptr;
2109 	new_icmp = (icmph_t *)(new_ipha + 1);
2110 
2111 	new_ipha->ipha_version_and_hdr_length = IP_SIMPLE_HDR_VERSION;
2112 	new_ipha->ipha_type_of_service = 0;
2113 	new_ipha->ipha_ident = 0;
2114 	new_ipha->ipha_fragment_offset_and_flags = 0;
2115 	new_ipha->ipha_ttl = orig_ipha->ipha_ttl;
2116 	new_ipha->ipha_protocol = IPPROTO_ICMP;
2117 	new_ipha->ipha_src = orig_ipha->ipha_dst;
2118 	new_ipha->ipha_dst = orig_ipha->ipha_src;
2119 	new_ipha->ipha_hdr_checksum = 0; /* will be computed by ip */
2120 	new_ipha->ipha_length = htons(hdrs_size + orig_pktsize);
2121 
2122 	*new_icmp = *icmp;
2123 	new_icmp->icmph_checksum = 0;
2124 	new_icmp->icmph_checksum = IP_CSUM(icmperr_mp, sizeof (ipha_t), 0);
2125 
2126 	bzero(&ixas, sizeof (ixas));
2127 	ixas.ixa_flags = IXAF_BASIC_SIMPLE_V4;
2128 	if (new_ipha->ipha_src == INADDR_ANY) {
2129 		ixas.ixa_flags &= ~IXAF_VERIFY_SOURCE;
2130 		ixas.ixa_flags |= IXAF_SET_SOURCE;
2131 	}
2132 
2133 	ixas.ixa_zoneid = IPCL_ZONEID(connp);
2134 	ixas.ixa_ipst = connp->conn_netstack->netstack_ip;
2135 	ixas.ixa_cred = connp->conn_cred;
2136 	ixas.ixa_cpid = NOPID;
2137 	if (is_system_labeled())
2138 		ixas.ixa_tsl = tsl;
2139 
2140 	ixas.ixa_ifindex = 0;
2141 	ixas.ixa_multicast_ttl = IP_DEFAULT_MULTICAST_TTL;
2142 
2143 	(void) ip_output_simple(icmperr_mp, &ixas);
2144 	ixa_cleanup(&ixas);
2145 }
2146 
2147 static void
2148 iptun_sendicmp_v6(iptun_t *iptun, icmp6_t *icmp6, ip6_t *orig_ip6h, mblk_t *mp,
2149     ts_label_t *tsl)
2150 {
2151 	size_t	orig_pktsize, hdrs_size;
2152 	mblk_t	*icmp6err_mp;
2153 	ip6_t	*new_ip6h;
2154 	icmp6_t	*new_icmp6;
2155 	ip_xmit_attr_t	ixas;
2156 	conn_t	*connp = iptun->iptun_connp;
2157 
2158 	orig_pktsize = msgdsize(mp);
2159 	hdrs_size = sizeof (ip6_t) + sizeof (icmp6_t);
2160 	if ((icmp6err_mp = iptun_build_icmperr(hdrs_size, mp)) == NULL) {
2161 		iptun_drop_pkt(mp, &iptun->iptun_noxmtbuf);
2162 		return;
2163 	}
2164 
2165 	new_ip6h = (ip6_t *)icmp6err_mp->b_rptr;
2166 	new_icmp6 = (icmp6_t *)(new_ip6h + 1);
2167 
2168 	new_ip6h->ip6_vcf = orig_ip6h->ip6_vcf;
2169 	new_ip6h->ip6_plen = htons(sizeof (icmp6_t) + orig_pktsize);
2170 	new_ip6h->ip6_hops = orig_ip6h->ip6_hops;
2171 	new_ip6h->ip6_nxt = IPPROTO_ICMPV6;
2172 	new_ip6h->ip6_src = orig_ip6h->ip6_dst;
2173 	new_ip6h->ip6_dst = orig_ip6h->ip6_src;
2174 
2175 	*new_icmp6 = *icmp6;
2176 	/* The checksum is calculated in ip_output_simple and friends. */
2177 	new_icmp6->icmp6_cksum = new_ip6h->ip6_plen;
2178 
2179 	bzero(&ixas, sizeof (ixas));
2180 	ixas.ixa_flags = IXAF_BASIC_SIMPLE_V6;
2181 	if (IN6_IS_ADDR_UNSPECIFIED(&new_ip6h->ip6_src)) {
2182 		ixas.ixa_flags &= ~IXAF_VERIFY_SOURCE;
2183 		ixas.ixa_flags |= IXAF_SET_SOURCE;
2184 	}
2185 
2186 	ixas.ixa_zoneid = IPCL_ZONEID(connp);
2187 	ixas.ixa_ipst = connp->conn_netstack->netstack_ip;
2188 	ixas.ixa_cred = connp->conn_cred;
2189 	ixas.ixa_cpid = NOPID;
2190 	if (is_system_labeled())
2191 		ixas.ixa_tsl = tsl;
2192 
2193 	ixas.ixa_ifindex = 0;
2194 	ixas.ixa_multicast_ttl = IP_DEFAULT_MULTICAST_TTL;
2195 
2196 	(void) ip_output_simple(icmp6err_mp, &ixas);
2197 	ixa_cleanup(&ixas);
2198 }
2199 
2200 static void
2201 iptun_icmp_error_v4(iptun_t *iptun, ipha_t *orig_ipha, mblk_t *mp,
2202     uint8_t type, uint8_t code, ts_label_t *tsl)
2203 {
2204 	icmph_t icmp;
2205 
2206 	bzero(&icmp, sizeof (icmp));
2207 	icmp.icmph_type = type;
2208 	icmp.icmph_code = code;
2209 
2210 	iptun_sendicmp_v4(iptun, &icmp, orig_ipha, mp, tsl);
2211 }
2212 
2213 static void
2214 iptun_icmp_fragneeded_v4(iptun_t *iptun, uint32_t newmtu, ipha_t *orig_ipha,
2215     mblk_t *mp, ts_label_t *tsl)
2216 {
2217 	icmph_t	icmp;
2218 
2219 	icmp.icmph_type = ICMP_DEST_UNREACHABLE;
2220 	icmp.icmph_code = ICMP_FRAGMENTATION_NEEDED;
2221 	icmp.icmph_du_zero = 0;
2222 	icmp.icmph_du_mtu = htons(newmtu);
2223 
2224 	iptun_sendicmp_v4(iptun, &icmp, orig_ipha, mp, tsl);
2225 }
2226 
2227 static void
2228 iptun_icmp_error_v6(iptun_t *iptun, ip6_t *orig_ip6h, mblk_t *mp,
2229     uint8_t type, uint8_t code, uint32_t offset, ts_label_t *tsl)
2230 {
2231 	icmp6_t icmp6;
2232 
2233 	bzero(&icmp6, sizeof (icmp6));
2234 	icmp6.icmp6_type = type;
2235 	icmp6.icmp6_code = code;
2236 	if (type == ICMP6_PARAM_PROB)
2237 		icmp6.icmp6_pptr = htonl(offset);
2238 
2239 	iptun_sendicmp_v6(iptun, &icmp6, orig_ip6h, mp, tsl);
2240 }
2241 
2242 static void
2243 iptun_icmp_toobig_v6(iptun_t *iptun, uint32_t newmtu, ip6_t *orig_ip6h,
2244     mblk_t *mp, ts_label_t *tsl)
2245 {
2246 	icmp6_t icmp6;
2247 
2248 	icmp6.icmp6_type = ICMP6_PACKET_TOO_BIG;
2249 	icmp6.icmp6_code = 0;
2250 	icmp6.icmp6_mtu = htonl(newmtu);
2251 
2252 	iptun_sendicmp_v6(iptun, &icmp6, orig_ip6h, mp, tsl);
2253 }
2254 
2255 /*
2256  * Determines if the packet pointed to by ipha or ip6h is an ICMP error.  The
2257  * mp argument is only used to do bounds checking.
2258  */
2259 static boolean_t
2260 is_icmp_error(mblk_t *mp, ipha_t *ipha, ip6_t *ip6h)
2261 {
2262 	uint16_t hlen;
2263 
2264 	if (ipha != NULL) {
2265 		icmph_t	*icmph;
2266 
2267 		ASSERT(ip6h == NULL);
2268 		if (ipha->ipha_protocol != IPPROTO_ICMP)
2269 			return (B_FALSE);
2270 
2271 		hlen = IPH_HDR_LENGTH(ipha);
2272 		icmph = (icmph_t *)((uint8_t *)ipha + hlen);
2273 		return (ICMP_IS_ERROR(icmph->icmph_type) ||
2274 		    icmph->icmph_type == ICMP_REDIRECT);
2275 	} else {
2276 		icmp6_t	*icmp6;
2277 		uint8_t	*nexthdrp;
2278 
2279 		ASSERT(ip6h != NULL);
2280 		if (!ip_hdr_length_nexthdr_v6(mp, ip6h, &hlen, &nexthdrp) ||
2281 		    *nexthdrp != IPPROTO_ICMPV6) {
2282 			return (B_FALSE);
2283 		}
2284 
2285 		icmp6 = (icmp6_t *)((uint8_t *)ip6h + hlen);
2286 		return (ICMP6_IS_ERROR(icmp6->icmp6_type) ||
2287 		    icmp6->icmp6_type == ND_REDIRECT);
2288 	}
2289 }
2290 
2291 /*
2292  * Find inner and outer IP headers from a tunneled packet as setup for calls
2293  * into ipsec_tun_{in,out}bound().
2294  * Note that we need to allow the outer header to be in a separate mblk from
2295  * the inner header.
2296  * If the caller knows the outer_hlen, the caller passes it in. Otherwise zero.
2297  */
2298 static size_t
2299 iptun_find_headers(mblk_t *mp, size_t outer_hlen, ipha_t **outer4,
2300     ipha_t **inner4, ip6_t **outer6, ip6_t **inner6)
2301 {
2302 	ipha_t	*ipha;
2303 	size_t	first_mblkl = MBLKL(mp);
2304 	mblk_t	*inner_mp;
2305 
2306 	/*
2307 	 * Don't bother handling packets that don't have a full IP header in
2308 	 * the fist mblk.  For the input path, the ip module ensures that this
2309 	 * won't happen, and on the output path, the IP tunneling MAC-type
2310 	 * plugins ensure that this also won't happen.
2311 	 */
2312 	if (first_mblkl < sizeof (ipha_t))
2313 		return (0);
2314 	ipha = (ipha_t *)(mp->b_rptr);
2315 	switch (IPH_HDR_VERSION(ipha)) {
2316 	case IPV4_VERSION:
2317 		*outer4 = ipha;
2318 		*outer6 = NULL;
2319 		if (outer_hlen == 0)
2320 			outer_hlen = IPH_HDR_LENGTH(ipha);
2321 		break;
2322 	case IPV6_VERSION:
2323 		*outer4 = NULL;
2324 		*outer6 = (ip6_t *)ipha;
2325 		if (outer_hlen == 0)
2326 			outer_hlen = ip_hdr_length_v6(mp, (ip6_t *)ipha);
2327 		break;
2328 	default:
2329 		return (0);
2330 	}
2331 
2332 	if (first_mblkl < outer_hlen ||
2333 	    (first_mblkl == outer_hlen && mp->b_cont == NULL))
2334 		return (0);
2335 
2336 	/*
2337 	 * We don't bother doing a pullup here since the outer header will
2338 	 * just get stripped off soon on input anyway.  We just want to ensure
2339 	 * that the inner* pointer points to a full header.
2340 	 */
2341 	if (first_mblkl == outer_hlen) {
2342 		inner_mp = mp->b_cont;
2343 		ipha = (ipha_t *)inner_mp->b_rptr;
2344 	} else {
2345 		inner_mp = mp;
2346 		ipha = (ipha_t *)(mp->b_rptr + outer_hlen);
2347 	}
2348 	switch (IPH_HDR_VERSION(ipha)) {
2349 	case IPV4_VERSION:
2350 		if (inner_mp->b_wptr - (uint8_t *)ipha < sizeof (ipha_t))
2351 			return (0);
2352 		*inner4 = ipha;
2353 		*inner6 = NULL;
2354 		break;
2355 	case IPV6_VERSION:
2356 		if (inner_mp->b_wptr - (uint8_t *)ipha < sizeof (ip6_t))
2357 			return (0);
2358 		*inner4 = NULL;
2359 		*inner6 = (ip6_t *)ipha;
2360 		break;
2361 	default:
2362 		return (0);
2363 	}
2364 
2365 	return (outer_hlen);
2366 }
2367 
2368 /*
2369  * Received ICMP error in response to an X over IPv4 packet that we
2370  * transmitted.
2371  *
2372  * NOTE: "outer" refers to what's inside the ICMP payload.  We will get one of
2373  * the following:
2374  *
2375  * [IPv4(0)][ICMPv4][IPv4(1)][IPv4(2)][ULP]
2376  *
2377  *	or
2378  *
2379  * [IPv4(0)][ICMPv4][IPv4(1)][IPv6][ULP]
2380  *
2381  * And "outer4" will get set to IPv4(1), and inner[46] will correspond to
2382  * whatever the very-inner packet is (IPv4(2) or IPv6).
2383  */
2384 static void
2385 iptun_input_icmp_v4(iptun_t *iptun, mblk_t *data_mp, icmph_t *icmph,
2386     ip_recv_attr_t *ira)
2387 {
2388 	uint8_t	*orig;
2389 	ipha_t	*outer4, *inner4;
2390 	ip6_t	*outer6, *inner6;
2391 	int	outer_hlen;
2392 	uint8_t	type, code;
2393 
2394 	ASSERT(data_mp->b_cont == NULL);
2395 	/*
2396 	 * Temporarily move b_rptr forward so that iptun_find_headers() can
2397 	 * find headers in the ICMP packet payload.
2398 	 */
2399 	orig = data_mp->b_rptr;
2400 	data_mp->b_rptr = (uint8_t *)(icmph + 1);
2401 	/*
2402 	 * The ip module ensures that ICMP errors contain at least the
2403 	 * original IP header (otherwise, the error would never have made it
2404 	 * here).
2405 	 */
2406 	ASSERT(MBLKL(data_mp) >= 0);
2407 	outer_hlen = iptun_find_headers(data_mp, 0, &outer4, &inner4, &outer6,
2408 	    &inner6);
2409 	ASSERT(outer6 == NULL);
2410 	data_mp->b_rptr = orig;
2411 	if (outer_hlen == 0) {
2412 		iptun_drop_pkt(data_mp, &iptun->iptun_ierrors);
2413 		return;
2414 	}
2415 
2416 	/* Only ICMP errors due to tunneled packets should reach here. */
2417 	ASSERT(outer4->ipha_protocol == IPPROTO_ENCAP ||
2418 	    outer4->ipha_protocol == IPPROTO_IPV6);
2419 
2420 	data_mp = ipsec_tun_inbound(ira, data_mp, iptun->iptun_itp,
2421 	    inner4, inner6, outer4, outer6, -outer_hlen, iptun->iptun_ns);
2422 	if (data_mp == NULL) {
2423 		/* Callee did all of the freeing. */
2424 		atomic_inc_64(&iptun->iptun_ierrors);
2425 		return;
2426 	}
2427 	/* We should never see reassembled fragment here. */
2428 	ASSERT(data_mp->b_next == NULL);
2429 
2430 	data_mp->b_rptr = (uint8_t *)outer4 + outer_hlen;
2431 
2432 	/*
2433 	 * If the original packet being transmitted was itself an ICMP error,
2434 	 * then drop this packet.  We don't want to generate an ICMP error in
2435 	 * response to an ICMP error.
2436 	 */
2437 	if (is_icmp_error(data_mp, inner4, inner6)) {
2438 		iptun_drop_pkt(data_mp, &iptun->iptun_norcvbuf);
2439 		return;
2440 	}
2441 
2442 	switch (icmph->icmph_type) {
2443 	case ICMP_DEST_UNREACHABLE:
2444 		type = (inner4 != NULL ? icmph->icmph_type : ICMP6_DST_UNREACH);
2445 		switch (icmph->icmph_code) {
2446 		case ICMP_FRAGMENTATION_NEEDED: {
2447 			uint32_t newmtu;
2448 
2449 			/*
2450 			 * We reconcile this with the fact that the tunnel may
2451 			 * also have IPsec policy by letting iptun_update_mtu
2452 			 * take care of it.
2453 			 */
2454 			newmtu = iptun_update_mtu(iptun, NULL,
2455 			    ntohs(icmph->icmph_du_mtu));
2456 
2457 			if (inner4 != NULL) {
2458 				iptun_icmp_fragneeded_v4(iptun, newmtu, inner4,
2459 				    data_mp, ira->ira_tsl);
2460 			} else {
2461 				iptun_icmp_toobig_v6(iptun, newmtu, inner6,
2462 				    data_mp, ira->ira_tsl);
2463 			}
2464 			return;
2465 		}
2466 		case ICMP_DEST_NET_UNREACH_ADMIN:
2467 		case ICMP_DEST_HOST_UNREACH_ADMIN:
2468 			code = (inner4 != NULL ? ICMP_DEST_NET_UNREACH_ADMIN :
2469 			    ICMP6_DST_UNREACH_ADMIN);
2470 			break;
2471 		default:
2472 			code = (inner4 != NULL ? ICMP_HOST_UNREACHABLE :
2473 			    ICMP6_DST_UNREACH_ADDR);
2474 			break;
2475 		}
2476 		break;
2477 	case ICMP_TIME_EXCEEDED:
2478 		if (inner6 != NULL) {
2479 			type = ICMP6_TIME_EXCEEDED;
2480 			code = 0;
2481 		} /* else we're already set. */
2482 		break;
2483 	case ICMP_PARAM_PROBLEM:
2484 		/*
2485 		 * This is a problem with the outer header we transmitted.
2486 		 * Treat this as an output error.
2487 		 */
2488 		iptun_drop_pkt(data_mp, &iptun->iptun_oerrors);
2489 		return;
2490 	default:
2491 		iptun_drop_pkt(data_mp, &iptun->iptun_norcvbuf);
2492 		return;
2493 	}
2494 
2495 	if (inner4 != NULL) {
2496 		iptun_icmp_error_v4(iptun, inner4, data_mp, type, code,
2497 		    ira->ira_tsl);
2498 	} else {
2499 		iptun_icmp_error_v6(iptun, inner6, data_mp, type, code, 0,
2500 		    ira->ira_tsl);
2501 	}
2502 }
2503 
2504 /*
2505  * Return B_TRUE if the IPv6 packet pointed to by ip6h contains a Tunnel
2506  * Encapsulation Limit destination option.  If there is one, set encaplim_ptr
2507  * to point to the option value.
2508  */
2509 static boolean_t
2510 iptun_find_encaplimit(mblk_t *mp, ip6_t *ip6h, uint8_t **encaplim_ptr)
2511 {
2512 	ip_pkt_t	pkt;
2513 	uint8_t		*endptr;
2514 	ip6_dest_t	*destp;
2515 	struct ip6_opt	*optp;
2516 
2517 	pkt.ipp_fields = 0; /* must be initialized */
2518 	(void) ip_find_hdr_v6(mp, ip6h, B_FALSE, &pkt, NULL);
2519 	if ((pkt.ipp_fields & IPPF_DSTOPTS) != 0) {
2520 		destp = pkt.ipp_dstopts;
2521 	} else if ((pkt.ipp_fields & IPPF_RTHDRDSTOPTS) != 0) {
2522 		destp = pkt.ipp_rthdrdstopts;
2523 	} else {
2524 		return (B_FALSE);
2525 	}
2526 
2527 	endptr = (uint8_t *)destp + 8 * (destp->ip6d_len + 1);
2528 	optp = (struct ip6_opt *)(destp + 1);
2529 	while (endptr - (uint8_t *)optp > sizeof (*optp)) {
2530 		if (optp->ip6o_type == IP6OPT_TUNNEL_LIMIT) {
2531 			if ((uint8_t *)(optp + 1) >= endptr)
2532 				return (B_FALSE);
2533 			*encaplim_ptr = (uint8_t *)&optp[1];
2534 			return (B_TRUE);
2535 		}
2536 		optp = (struct ip6_opt *)((uint8_t *)optp + optp->ip6o_len + 2);
2537 	}
2538 	return (B_FALSE);
2539 }
2540 
2541 /*
2542  * Received ICMPv6 error in response to an X over IPv6 packet that we
2543  * transmitted.
2544  *
2545  * NOTE: "outer" refers to what's inside the ICMP payload.  We will get one of
2546  * the following:
2547  *
2548  * [IPv6(0)][ICMPv6][IPv6(1)][IPv4][ULP]
2549  *
2550  *	or
2551  *
2552  * [IPv6(0)][ICMPv6][IPv6(1)][IPv6(2)][ULP]
2553  *
2554  * And "outer6" will get set to IPv6(1), and inner[46] will correspond to
2555  * whatever the very-inner packet is (IPv4 or IPv6(2)).
2556  */
2557 static void
2558 iptun_input_icmp_v6(iptun_t *iptun, mblk_t *data_mp, icmp6_t *icmp6h,
2559     ip_recv_attr_t *ira)
2560 {
2561 	uint8_t	*orig;
2562 	ipha_t	*outer4, *inner4;
2563 	ip6_t	*outer6, *inner6;
2564 	int	outer_hlen;
2565 	uint8_t	type, code;
2566 
2567 	ASSERT(data_mp->b_cont == NULL);
2568 
2569 	/*
2570 	 * Temporarily move b_rptr forward so that iptun_find_headers() can
2571 	 * find IP headers in the ICMP packet payload.
2572 	 */
2573 	orig = data_mp->b_rptr;
2574 	data_mp->b_rptr = (uint8_t *)(icmp6h + 1);
2575 	/*
2576 	 * The ip module ensures that ICMP errors contain at least the
2577 	 * original IP header (otherwise, the error would never have made it
2578 	 * here).
2579 	 */
2580 	ASSERT(MBLKL(data_mp) >= 0);
2581 	outer_hlen = iptun_find_headers(data_mp, 0, &outer4, &inner4, &outer6,
2582 	    &inner6);
2583 	ASSERT(outer4 == NULL);
2584 	data_mp->b_rptr = orig;	/* Restore r_ptr */
2585 	if (outer_hlen == 0) {
2586 		iptun_drop_pkt(data_mp, &iptun->iptun_ierrors);
2587 		return;
2588 	}
2589 
2590 	data_mp = ipsec_tun_inbound(ira, data_mp, iptun->iptun_itp,
2591 	    inner4, inner6, outer4, outer6, -outer_hlen, iptun->iptun_ns);
2592 	if (data_mp == NULL) {
2593 		/* Callee did all of the freeing. */
2594 		atomic_inc_64(&iptun->iptun_ierrors);
2595 		return;
2596 	}
2597 	/* We should never see reassembled fragment here. */
2598 	ASSERT(data_mp->b_next == NULL);
2599 
2600 	data_mp->b_rptr = (uint8_t *)outer6 + outer_hlen;
2601 
2602 	/*
2603 	 * If the original packet being transmitted was itself an ICMP error,
2604 	 * then drop this packet.  We don't want to generate an ICMP error in
2605 	 * response to an ICMP error.
2606 	 */
2607 	if (is_icmp_error(data_mp, inner4, inner6)) {
2608 		iptun_drop_pkt(data_mp, &iptun->iptun_norcvbuf);
2609 		return;
2610 	}
2611 
2612 	switch (icmp6h->icmp6_type) {
2613 	case ICMP6_PARAM_PROB: {
2614 		uint8_t *encaplim_ptr;
2615 
2616 		/*
2617 		 * If the ICMPv6 error points to a valid Tunnel Encapsulation
2618 		 * Limit option and the limit value is 0, then fall through
2619 		 * and send a host unreachable message.  Otherwise, treat the
2620 		 * error as an output error, as there must have been a problem
2621 		 * with a packet we sent.
2622 		 */
2623 		if (!iptun_find_encaplimit(data_mp, outer6, &encaplim_ptr) ||
2624 		    (icmp6h->icmp6_pptr !=
2625 		    ((ptrdiff_t)encaplim_ptr - (ptrdiff_t)outer6)) ||
2626 		    *encaplim_ptr != 0) {
2627 			iptun_drop_pkt(data_mp, &iptun->iptun_oerrors);
2628 			return;
2629 		}
2630 		/* FALLTHRU */
2631 	}
2632 	case ICMP6_TIME_EXCEEDED:
2633 	case ICMP6_DST_UNREACH:
2634 		type = (inner4 != NULL ? ICMP_DEST_UNREACHABLE :
2635 		    ICMP6_DST_UNREACH);
2636 		code = (inner4 != NULL ? ICMP_HOST_UNREACHABLE :
2637 		    ICMP6_DST_UNREACH_ADDR);
2638 		break;
2639 	case ICMP6_PACKET_TOO_BIG: {
2640 		uint32_t newmtu;
2641 
2642 		/*
2643 		 * We reconcile this with the fact that the tunnel may also
2644 		 * have IPsec policy by letting iptun_update_mtu take care of
2645 		 * it.
2646 		 */
2647 		newmtu = iptun_update_mtu(iptun, NULL,
2648 		    ntohl(icmp6h->icmp6_mtu));
2649 
2650 		if (inner4 != NULL) {
2651 			iptun_icmp_fragneeded_v4(iptun, newmtu, inner4,
2652 			    data_mp, ira->ira_tsl);
2653 		} else {
2654 			iptun_icmp_toobig_v6(iptun, newmtu, inner6, data_mp,
2655 			    ira->ira_tsl);
2656 		}
2657 		return;
2658 	}
2659 	default:
2660 		iptun_drop_pkt(data_mp, &iptun->iptun_norcvbuf);
2661 		return;
2662 	}
2663 
2664 	if (inner4 != NULL) {
2665 		iptun_icmp_error_v4(iptun, inner4, data_mp, type, code,
2666 		    ira->ira_tsl);
2667 	} else {
2668 		iptun_icmp_error_v6(iptun, inner6, data_mp, type, code, 0,
2669 		    ira->ira_tsl);
2670 	}
2671 }
2672 
2673 /*
2674  * Called as conn_recvicmp from IP for ICMP errors.
2675  */
2676 /* ARGSUSED2 */
2677 static void
2678 iptun_input_icmp(void *arg, mblk_t *mp, void *arg2, ip_recv_attr_t *ira)
2679 {
2680 	conn_t		*connp = arg;
2681 	iptun_t		*iptun = connp->conn_iptun;
2682 	mblk_t		*tmpmp;
2683 	size_t		hlen;
2684 
2685 	ASSERT(IPCL_IS_IPTUN(connp));
2686 
2687 	if (mp->b_cont != NULL) {
2688 		/*
2689 		 * Since ICMP error processing necessitates access to bits
2690 		 * that are within the ICMP error payload (the original packet
2691 		 * that caused the error), pull everything up into a single
2692 		 * block for convenience.
2693 		 */
2694 		if ((tmpmp = msgpullup(mp, -1)) == NULL) {
2695 			iptun_drop_pkt(mp, &iptun->iptun_norcvbuf);
2696 			return;
2697 		}
2698 		freemsg(mp);
2699 		mp = tmpmp;
2700 	}
2701 
2702 	hlen = ira->ira_ip_hdr_length;
2703 	switch (iptun->iptun_typeinfo->iti_ipvers) {
2704 	case IPV4_VERSION:
2705 		/*
2706 		 * The outer IP header coming up from IP is always ipha_t
2707 		 * alligned (otherwise, we would have crashed in ip).
2708 		 */
2709 		iptun_input_icmp_v4(iptun, mp, (icmph_t *)(mp->b_rptr + hlen),
2710 		    ira);
2711 		break;
2712 	case IPV6_VERSION:
2713 		iptun_input_icmp_v6(iptun, mp, (icmp6_t *)(mp->b_rptr + hlen),
2714 		    ira);
2715 		break;
2716 	}
2717 }
2718 
2719 static boolean_t
2720 iptun_in_6to4_ok(iptun_t *iptun, ipha_t *outer4, ip6_t *inner6)
2721 {
2722 	ipaddr_t v4addr;
2723 
2724 	/*
2725 	 * It's possible that someone sent us an IPv4-in-IPv4 packet with the
2726 	 * IPv4 address of a 6to4 tunnel as the destination.
2727 	 */
2728 	if (inner6 == NULL)
2729 		return (B_FALSE);
2730 
2731 	/*
2732 	 * Make sure that the IPv6 destination is within the site that this
2733 	 * 6to4 tunnel is routing for.  We don't want people bouncing random
2734 	 * tunneled IPv6 packets through this 6to4 router.
2735 	 */
2736 	IN6_6TO4_TO_V4ADDR(&inner6->ip6_dst, (struct in_addr *)&v4addr);
2737 	if (outer4->ipha_dst != v4addr)
2738 		return (B_FALSE);
2739 
2740 	if (IN6_IS_ADDR_6TO4(&inner6->ip6_src)) {
2741 		/*
2742 		 * Section 9 of RFC 3056 (security considerations) suggests
2743 		 * that when a packet is from a 6to4 site (i.e., it's not a
2744 		 * global address being forwarded froma relay router), make
2745 		 * sure that the packet was tunneled by that site's 6to4
2746 		 * router.
2747 		 */
2748 		IN6_6TO4_TO_V4ADDR(&inner6->ip6_src, (struct in_addr *)&v4addr);
2749 		if (outer4->ipha_src != v4addr)
2750 			return (B_FALSE);
2751 	} else {
2752 		/*
2753 		 * Only accept packets from a relay router if we've configured
2754 		 * outbound relay router functionality.
2755 		 */
2756 		if (iptun->iptun_iptuns->iptuns_relay_rtr_addr == INADDR_ANY)
2757 			return (B_FALSE);
2758 	}
2759 
2760 	return (B_TRUE);
2761 }
2762 
2763 /*
2764  * Input function for everything that comes up from the ip module below us.
2765  * This is called directly from the ip module via connp->conn_recv().
2766  *
2767  * We receive M_DATA messages with IP-in-IP tunneled packets.
2768  */
2769 /* ARGSUSED2 */
2770 static void
2771 iptun_input(void *arg, mblk_t *data_mp, void *arg2, ip_recv_attr_t *ira)
2772 {
2773 	conn_t	*connp = arg;
2774 	iptun_t	*iptun = connp->conn_iptun;
2775 	int	outer_hlen;
2776 	ipha_t	*outer4, *inner4;
2777 	ip6_t	*outer6, *inner6;
2778 
2779 	ASSERT(IPCL_IS_IPTUN(connp));
2780 	ASSERT(DB_TYPE(data_mp) == M_DATA);
2781 
2782 	outer_hlen = iptun_find_headers(data_mp, ira->ira_ip_hdr_length,
2783 	    &outer4, &inner4, &outer6, &inner6);
2784 	if (outer_hlen == 0)
2785 		goto drop;
2786 
2787 	/*
2788 	 * If the system is labeled, we call tsol_check_dest() on the packet
2789 	 * destination (our local tunnel address) to ensure that the packet as
2790 	 * labeled should be allowed to be sent to us.  We don't need to call
2791 	 * the more involved tsol_receive_local() since the tunnel link itself
2792 	 * cannot be assigned to shared-stack non-global zones.
2793 	 */
2794 	if (ira->ira_flags & IRAF_SYSTEM_LABELED) {
2795 		if (ira->ira_tsl == NULL)
2796 			goto drop;
2797 		if (tsol_check_dest(ira->ira_tsl, (outer4 != NULL ?
2798 		    (void *)&outer4->ipha_dst : (void *)&outer6->ip6_dst),
2799 		    (outer4 != NULL ? IPV4_VERSION : IPV6_VERSION),
2800 		    CONN_MAC_DEFAULT, B_FALSE, NULL) != 0)
2801 			goto drop;
2802 	}
2803 
2804 	data_mp = ipsec_tun_inbound(ira, data_mp, iptun->iptun_itp,
2805 	    inner4, inner6, outer4, outer6, outer_hlen, iptun->iptun_ns);
2806 	if (data_mp == NULL) {
2807 		/* Callee did all of the freeing. */
2808 		return;
2809 	}
2810 
2811 	if (iptun->iptun_typeinfo->iti_type == IPTUN_TYPE_6TO4 &&
2812 	    !iptun_in_6to4_ok(iptun, outer4, inner6))
2813 		goto drop;
2814 
2815 	/*
2816 	 * We need to statistically account for each packet individually, so
2817 	 * we might as well split up any b_next chains here.
2818 	 */
2819 	do {
2820 		mblk_t	*mp;
2821 
2822 		mp = data_mp->b_next;
2823 		data_mp->b_next = NULL;
2824 
2825 		atomic_inc_64(&iptun->iptun_ipackets);
2826 		atomic_add_64(&iptun->iptun_rbytes, msgdsize(data_mp));
2827 		mac_rx(iptun->iptun_mh, NULL, data_mp);
2828 
2829 		data_mp = mp;
2830 	} while (data_mp != NULL);
2831 	return;
2832 drop:
2833 	iptun_drop_pkt(data_mp, &iptun->iptun_ierrors);
2834 }
2835 
2836 /*
2837  * Do 6to4-specific header-processing on output.  Return B_TRUE if the packet
2838  * was processed without issue, or B_FALSE if the packet had issues and should
2839  * be dropped.
2840  */
2841 static boolean_t
2842 iptun_out_process_6to4(iptun_t *iptun, ipha_t *outer4, ip6_t *inner6)
2843 {
2844 	ipaddr_t v4addr;
2845 
2846 	/*
2847 	 * IPv6 source must be a 6to4 address.  This is because a conscious
2848 	 * decision was made to not allow a Solaris system to be used as a
2849 	 * relay router (for security reasons) when 6to4 was initially
2850 	 * integrated.  If this decision is ever reversed, the following check
2851 	 * can be removed.
2852 	 */
2853 	if (!IN6_IS_ADDR_6TO4(&inner6->ip6_src))
2854 		return (B_FALSE);
2855 
2856 	/*
2857 	 * RFC3056 mandates that the IPv4 source MUST be set to the IPv4
2858 	 * portion of the 6to4 IPv6 source address.  In other words, make sure
2859 	 * that we're tunneling packets from our own 6to4 site.
2860 	 */
2861 	IN6_6TO4_TO_V4ADDR(&inner6->ip6_src, (struct in_addr *)&v4addr);
2862 	if (outer4->ipha_src != v4addr)
2863 		return (B_FALSE);
2864 
2865 	/*
2866 	 * Automatically set the destination of the outer IPv4 header as
2867 	 * described in RFC3056.  There are two possibilities:
2868 	 *
2869 	 * a. If the IPv6 destination is a 6to4 address, set the IPv4 address
2870 	 *    to the IPv4 portion of the 6to4 address.
2871 	 * b. If the IPv6 destination is a native IPv6 address, set the IPv4
2872 	 *    destination to the address of a relay router.
2873 	 *
2874 	 * Design Note: b shouldn't be necessary here, and this is a flaw in
2875 	 * the design of the 6to4relay command.  Instead of setting a 6to4
2876 	 * relay address in this module via an ioctl, the 6to4relay command
2877 	 * could simply add a IPv6 route for native IPv6 addresses (such as a
2878 	 * default route) in the forwarding table that uses a 6to4 destination
2879 	 * as its next hop, and the IPv4 portion of that address could be a
2880 	 * 6to4 relay address.  In order for this to work, IP would have to
2881 	 * resolve the next hop address, which would necessitate a link-layer
2882 	 * address resolver for 6to4 links, which doesn't exist today.
2883 	 *
2884 	 * In fact, if a resolver existed for 6to4 links, then setting the
2885 	 * IPv4 destination in the outer header could be done as part of
2886 	 * link-layer address resolution and fast-path header generation, and
2887 	 * not here.
2888 	 */
2889 	if (IN6_IS_ADDR_6TO4(&inner6->ip6_dst)) {
2890 		/* destination is a 6to4 router */
2891 		IN6_6TO4_TO_V4ADDR(&inner6->ip6_dst,
2892 		    (struct in_addr *)&outer4->ipha_dst);
2893 
2894 		/* Reject attempts to send to INADDR_ANY */
2895 		if (outer4->ipha_dst == INADDR_ANY)
2896 			return (B_FALSE);
2897 	} else {
2898 		/*
2899 		 * The destination is a native IPv6 address.  If output to a
2900 		 * relay-router is enabled, use the relay-router's IPv4
2901 		 * address as the destination.
2902 		 */
2903 		if (iptun->iptun_iptuns->iptuns_relay_rtr_addr == INADDR_ANY)
2904 			return (B_FALSE);
2905 		outer4->ipha_dst = iptun->iptun_iptuns->iptuns_relay_rtr_addr;
2906 	}
2907 
2908 	/*
2909 	 * If the outer source and destination are equal, this means that the
2910 	 * 6to4 router somehow forwarded an IPv6 packet destined for its own
2911 	 * 6to4 site to its 6to4 tunnel interface, which will result in this
2912 	 * packet infinitely bouncing between ip and iptun.
2913 	 */
2914 	return (outer4->ipha_src != outer4->ipha_dst);
2915 }
2916 
2917 /*
2918  * Process output packets with outer IPv4 headers.  Frees mp and bumps stat on
2919  * error.
2920  */
2921 static mblk_t *
2922 iptun_out_process_ipv4(iptun_t *iptun, mblk_t *mp, ipha_t *outer4,
2923     ipha_t *inner4, ip6_t *inner6, ip_xmit_attr_t *ixa)
2924 {
2925 	uint8_t	*innerptr = (inner4 != NULL ?
2926 	    (uint8_t *)inner4 : (uint8_t *)inner6);
2927 	size_t	minmtu = iptun->iptun_typeinfo->iti_minmtu;
2928 
2929 	if (inner4 != NULL) {
2930 		ASSERT(outer4->ipha_protocol == IPPROTO_ENCAP);
2931 		/*
2932 		 * Copy the tos from the inner IPv4 header. We mask off ECN
2933 		 * bits (bits 6 and 7) because there is currently no
2934 		 * tunnel-tunnel communication to determine if both sides
2935 		 * support ECN.  We opt for the safe choice: don't copy the
2936 		 * ECN bits when doing encapsulation.
2937 		 */
2938 		outer4->ipha_type_of_service =
2939 		    inner4->ipha_type_of_service & ~0x03;
2940 	} else {
2941 		ASSERT(outer4->ipha_protocol == IPPROTO_IPV6 &&
2942 		    inner6 != NULL);
2943 	}
2944 	if (ixa->ixa_flags & IXAF_PMTU_IPV4_DF)
2945 		outer4->ipha_fragment_offset_and_flags |= IPH_DF_HTONS;
2946 	else
2947 		outer4->ipha_fragment_offset_and_flags &= ~IPH_DF_HTONS;
2948 
2949 	/*
2950 	 * As described in section 3.2.2 of RFC4213, if the packet payload is
2951 	 * less than or equal to the minimum MTU size, then we need to allow
2952 	 * IPv4 to fragment the packet.  The reason is that even if we end up
2953 	 * receiving an ICMP frag-needed, the interface above this tunnel
2954 	 * won't be allowed to drop its MTU as a result, since the packet was
2955 	 * already smaller than the smallest allowable MTU for that interface.
2956 	 */
2957 	if (mp->b_wptr - innerptr <= minmtu) {
2958 		outer4->ipha_fragment_offset_and_flags = 0;
2959 		ixa->ixa_flags &= ~IXAF_DONTFRAG;
2960 	} else if (!(ixa->ixa_flags & IXAF_PMTU_TOO_SMALL) &&
2961 	    (iptun->iptun_typeinfo->iti_type != IPTUN_TYPE_6TO4)) {
2962 		ixa->ixa_flags |= IXAF_DONTFRAG;
2963 	}
2964 
2965 	ixa->ixa_ip_hdr_length = IPH_HDR_LENGTH(outer4);
2966 	ixa->ixa_pktlen = msgdsize(mp);
2967 	ixa->ixa_protocol = outer4->ipha_protocol;
2968 
2969 	outer4->ipha_length = htons(ixa->ixa_pktlen);
2970 	return (mp);
2971 }
2972 
2973 /*
2974  * Insert an encapsulation limit destination option in the packet provided.
2975  * Always consumes the mp argument and returns a new mblk pointer.
2976  */
2977 static mblk_t *
2978 iptun_insert_encaplimit(iptun_t *iptun, mblk_t *mp, ip6_t *outer6,
2979     uint8_t limit)
2980 {
2981 	mblk_t			*newmp;
2982 	iptun_ipv6hdrs_t	*newouter6;
2983 
2984 	ASSERT(outer6->ip6_nxt == IPPROTO_IPV6);
2985 	ASSERT(mp->b_cont == NULL);
2986 
2987 	mp->b_rptr += sizeof (ip6_t);
2988 	newmp = allocb(sizeof (iptun_ipv6hdrs_t) + MBLKL(mp), BPRI_MED);
2989 	if (newmp == NULL) {
2990 		iptun_drop_pkt(mp, &iptun->iptun_noxmtbuf);
2991 		return (NULL);
2992 	}
2993 	newmp->b_wptr += sizeof (iptun_ipv6hdrs_t);
2994 	/* Copy the payload (Starting with the inner IPv6 header). */
2995 	bcopy(mp->b_rptr, newmp->b_wptr, MBLKL(mp));
2996 	newmp->b_wptr += MBLKL(mp);
2997 	newouter6 = (iptun_ipv6hdrs_t *)newmp->b_rptr;
2998 	/* Now copy the outer IPv6 header. */
2999 	bcopy(outer6, &newouter6->it6h_ip6h, sizeof (ip6_t));
3000 	newouter6->it6h_ip6h.ip6_nxt = IPPROTO_DSTOPTS;
3001 	newouter6->it6h_encaplim = iptun_encaplim_init;
3002 	newouter6->it6h_encaplim.iel_destopt.ip6d_nxt = outer6->ip6_nxt;
3003 	newouter6->it6h_encaplim.iel_telopt.ip6ot_encap_limit = limit;
3004 
3005 	/*
3006 	 * The payload length will be set at the end of
3007 	 * iptun_out_process_ipv6().
3008 	 */
3009 
3010 	freemsg(mp);
3011 	return (newmp);
3012 }
3013 
3014 /*
3015  * Process output packets with outer IPv6 headers.  Frees mp and bumps stats
3016  * on error.
3017  */
3018 static mblk_t *
3019 iptun_out_process_ipv6(iptun_t *iptun, mblk_t *mp, ip6_t *outer6,
3020     ipha_t *inner4, ip6_t *inner6, ip_xmit_attr_t *ixa)
3021 {
3022 	uint8_t		*innerptr = (inner4 != NULL ?
3023 	    (uint8_t *)inner4 : (uint8_t *)inner6);
3024 	size_t		minmtu = iptun->iptun_typeinfo->iti_minmtu;
3025 	uint8_t		*limit, *configlimit;
3026 	uint32_t	offset;
3027 	iptun_ipv6hdrs_t *v6hdrs;
3028 
3029 	if (inner6 != NULL && iptun_find_encaplimit(mp, inner6, &limit)) {
3030 		/*
3031 		 * The inner packet is an IPv6 packet which itself contains an
3032 		 * encapsulation limit option.  The limit variable points to
3033 		 * the value in the embedded option.  Process the
3034 		 * encapsulation limit option as specified in RFC 2473.
3035 		 *
3036 		 * If limit is 0, then we've exceeded the limit and we need to
3037 		 * send back an ICMPv6 parameter problem message.
3038 		 *
3039 		 * If limit is > 0, then we decrement it by 1 and make sure
3040 		 * that the encapsulation limit option in the outer header
3041 		 * reflects that (adding an option if one isn't already
3042 		 * there).
3043 		 */
3044 		ASSERT(limit > mp->b_rptr && limit < mp->b_wptr);
3045 		if (*limit == 0) {
3046 			mp->b_rptr = (uint8_t *)inner6;
3047 			offset = limit - mp->b_rptr;
3048 			iptun_icmp_error_v6(iptun, inner6, mp, ICMP6_PARAM_PROB,
3049 			    0, offset, ixa->ixa_tsl);
3050 			atomic_inc_64(&iptun->iptun_noxmtbuf);
3051 			return (NULL);
3052 		}
3053 
3054 		/*
3055 		 * The outer header requires an encapsulation limit option.
3056 		 * If there isn't one already, add one.
3057 		 */
3058 		if (iptun->iptun_encaplimit == 0) {
3059 			if ((mp = iptun_insert_encaplimit(iptun, mp, outer6,
3060 			    (*limit - 1))) == NULL)
3061 				return (NULL);
3062 			v6hdrs = (iptun_ipv6hdrs_t *)mp->b_rptr;
3063 		} else {
3064 			/*
3065 			 * There is an existing encapsulation limit option in
3066 			 * the outer header.  If the inner encapsulation limit
3067 			 * is less than the configured encapsulation limit,
3068 			 * update the outer encapsulation limit to reflect
3069 			 * this lesser value.
3070 			 */
3071 			v6hdrs = (iptun_ipv6hdrs_t *)mp->b_rptr;
3072 			configlimit =
3073 			    &v6hdrs->it6h_encaplim.iel_telopt.ip6ot_encap_limit;
3074 			if ((*limit - 1) < *configlimit)
3075 				*configlimit = (*limit - 1);
3076 		}
3077 		ixa->ixa_ip_hdr_length = sizeof (iptun_ipv6hdrs_t);
3078 		ixa->ixa_protocol = v6hdrs->it6h_encaplim.iel_destopt.ip6d_nxt;
3079 	} else {
3080 		ixa->ixa_ip_hdr_length = sizeof (ip6_t);
3081 		ixa->ixa_protocol = outer6->ip6_nxt;
3082 	}
3083 	/*
3084 	 * See iptun_output_process_ipv4() why we allow fragmentation for
3085 	 * small packets
3086 	 */
3087 	if (mp->b_wptr - innerptr <= minmtu)
3088 		ixa->ixa_flags &= ~IXAF_DONTFRAG;
3089 	else if (!(ixa->ixa_flags & IXAF_PMTU_TOO_SMALL))
3090 		ixa->ixa_flags |= IXAF_DONTFRAG;
3091 
3092 	ixa->ixa_pktlen = msgdsize(mp);
3093 	outer6->ip6_plen = htons(ixa->ixa_pktlen - sizeof (ip6_t));
3094 	return (mp);
3095 }
3096 
3097 /*
3098  * The IP tunneling MAC-type plugins have already done most of the header
3099  * processing and validity checks.  We are simply responsible for multiplexing
3100  * down to the ip module below us.
3101  */
3102 static void
3103 iptun_output(iptun_t *iptun, mblk_t *mp)
3104 {
3105 	conn_t	*connp = iptun->iptun_connp;
3106 	mblk_t	*newmp;
3107 	int	error;
3108 	ip_xmit_attr_t	*ixa;
3109 
3110 	ASSERT(mp->b_datap->db_type == M_DATA);
3111 
3112 	if (mp->b_cont != NULL) {
3113 		if ((newmp = msgpullup(mp, -1)) == NULL) {
3114 			iptun_drop_pkt(mp, &iptun->iptun_noxmtbuf);
3115 			return;
3116 		}
3117 		freemsg(mp);
3118 		mp = newmp;
3119 	}
3120 
3121 	if (iptun->iptun_typeinfo->iti_type == IPTUN_TYPE_6TO4) {
3122 		iptun_output_6to4(iptun, mp);
3123 		return;
3124 	}
3125 
3126 	if (is_system_labeled()) {
3127 		/*
3128 		 * Since the label can be different meaning a potentially
3129 		 * different IRE,we always use a unique ip_xmit_attr_t.
3130 		 */
3131 		ixa = conn_get_ixa_exclusive(connp);
3132 	} else {
3133 		/*
3134 		 * If no other thread is using conn_ixa this just gets a
3135 		 * reference to conn_ixa. Otherwise we get a safe copy of
3136 		 * conn_ixa.
3137 		 */
3138 		ixa = conn_get_ixa(connp, B_FALSE);
3139 	}
3140 	if (ixa == NULL) {
3141 		iptun_drop_pkt(mp, &iptun->iptun_oerrors);
3142 		return;
3143 	}
3144 
3145 	/*
3146 	 * In case we got a safe copy of conn_ixa, then we need
3147 	 * to fill in any pointers in it.
3148 	 */
3149 	if (ixa->ixa_ire == NULL) {
3150 		error = ip_attr_connect(connp, ixa, &connp->conn_saddr_v6,
3151 		    &connp->conn_faddr_v6, &connp->conn_faddr_v6, 0,
3152 		    NULL, NULL, 0);
3153 		if (error != 0) {
3154 			if (ixa->ixa_ire != NULL &&
3155 			    (error == EHOSTUNREACH || error == ENETUNREACH)) {
3156 				/*
3157 				 * Let conn_ip_output/ire_send_noroute return
3158 				 * the error and send any local ICMP error.
3159 				 */
3160 				error = 0;
3161 			} else {
3162 				ixa_refrele(ixa);
3163 				iptun_drop_pkt(mp, &iptun->iptun_oerrors);
3164 				return;
3165 			}
3166 		}
3167 	}
3168 
3169 	iptun_output_common(iptun, ixa, mp);
3170 	ixa_refrele(ixa);
3171 }
3172 
3173 /*
3174  * We use an ixa based on the last destination.
3175  */
3176 static void
3177 iptun_output_6to4(iptun_t *iptun, mblk_t *mp)
3178 {
3179 	conn_t		*connp = iptun->iptun_connp;
3180 	ipha_t		*outer4, *inner4;
3181 	ip6_t		*outer6, *inner6;
3182 	ip_xmit_attr_t	*ixa;
3183 	ip_xmit_attr_t	*oldixa;
3184 	int		error;
3185 	boolean_t	need_connect;
3186 	in6_addr_t	v6dst;
3187 
3188 	ASSERT(mp->b_cont == NULL);	/* Verified by iptun_output */
3189 
3190 	/* Make sure we set ipha_dst before we look at ipha_dst */
3191 
3192 	(void) iptun_find_headers(mp, 0, &outer4, &inner4, &outer6, &inner6);
3193 	ASSERT(outer4 != NULL);
3194 	if (!iptun_out_process_6to4(iptun, outer4, inner6)) {
3195 		iptun_drop_pkt(mp, &iptun->iptun_oerrors);
3196 		return;
3197 	}
3198 
3199 	if (is_system_labeled()) {
3200 		/*
3201 		 * Since the label can be different meaning a potentially
3202 		 * different IRE,we always use a unique ip_xmit_attr_t.
3203 		 */
3204 		ixa = conn_get_ixa_exclusive(connp);
3205 	} else {
3206 		/*
3207 		 * If no other thread is using conn_ixa this just gets a
3208 		 * reference to conn_ixa. Otherwise we get a safe copy of
3209 		 * conn_ixa.
3210 		 */
3211 		ixa = conn_get_ixa(connp, B_FALSE);
3212 	}
3213 	if (ixa == NULL) {
3214 		iptun_drop_pkt(mp, &iptun->iptun_oerrors);
3215 		return;
3216 	}
3217 
3218 	mutex_enter(&connp->conn_lock);
3219 	if (connp->conn_v4lastdst == outer4->ipha_dst) {
3220 		need_connect = (ixa->ixa_ire == NULL);
3221 	} else {
3222 		/* In case previous destination was multirt */
3223 		ip_attr_newdst(ixa);
3224 
3225 		/*
3226 		 * We later update conn_ixa when we update conn_v4lastdst
3227 		 * which enables subsequent packets to avoid redoing
3228 		 * ip_attr_connect
3229 		 */
3230 		need_connect = B_TRUE;
3231 	}
3232 	mutex_exit(&connp->conn_lock);
3233 
3234 	/*
3235 	 * In case we got a safe copy of conn_ixa, or otherwise we don't
3236 	 * have a current ixa_ire, then we need to fill in any pointers in
3237 	 * the ixa.
3238 	 */
3239 	if (need_connect) {
3240 		IN6_IPADDR_TO_V4MAPPED(outer4->ipha_dst, &v6dst);
3241 
3242 		/* We handle IPsec in iptun_output_common */
3243 		error = ip_attr_connect(connp, ixa, &connp->conn_saddr_v6,
3244 		    &v6dst, &v6dst, 0, NULL, NULL, 0);
3245 		if (error != 0) {
3246 			if (ixa->ixa_ire != NULL &&
3247 			    (error == EHOSTUNREACH || error == ENETUNREACH)) {
3248 				/*
3249 				 * Let conn_ip_output/ire_send_noroute return
3250 				 * the error and send any local ICMP error.
3251 				 */
3252 				error = 0;
3253 			} else {
3254 				ixa_refrele(ixa);
3255 				iptun_drop_pkt(mp, &iptun->iptun_oerrors);
3256 				return;
3257 			}
3258 		}
3259 	}
3260 
3261 	iptun_output_common(iptun, ixa, mp);
3262 
3263 	/* Atomically replace conn_ixa and conn_v4lastdst */
3264 	mutex_enter(&connp->conn_lock);
3265 	if (connp->conn_v4lastdst != outer4->ipha_dst) {
3266 		/* Remember the dst which corresponds to conn_ixa */
3267 		connp->conn_v6lastdst = v6dst;
3268 		oldixa = conn_replace_ixa(connp, ixa);
3269 	} else {
3270 		oldixa = NULL;
3271 	}
3272 	mutex_exit(&connp->conn_lock);
3273 	ixa_refrele(ixa);
3274 	if (oldixa != NULL)
3275 		ixa_refrele(oldixa);
3276 }
3277 
3278 /*
3279  * Check the destination/label. Modifies *mpp by adding/removing CIPSO.
3280  *
3281  * We get the label from the message in order to honor the
3282  * ULPs/IPs choice of label. This will be NULL for forwarded
3283  * packets, neighbor discovery packets and some others.
3284  */
3285 static int
3286 iptun_output_check_label(mblk_t **mpp, ip_xmit_attr_t *ixa)
3287 {
3288 	cred_t	*cr;
3289 	int	adjust;
3290 	int	iplen;
3291 	int	err;
3292 	ts_label_t *effective_tsl = NULL;
3293 
3294 
3295 	ASSERT(is_system_labeled());
3296 
3297 	cr = msg_getcred(*mpp, NULL);
3298 	if (cr == NULL)
3299 		return (0);
3300 
3301 	/*
3302 	 * We need to start with a label based on the IP/ULP above us
3303 	 */
3304 	ip_xmit_attr_restore_tsl(ixa, cr);
3305 
3306 	/*
3307 	 * Need to update packet with any CIPSO option since
3308 	 * conn_ip_output doesn't do that.
3309 	 */
3310 	if (ixa->ixa_flags & IXAF_IS_IPV4) {
3311 		ipha_t *ipha;
3312 
3313 		ipha = (ipha_t *)(*mpp)->b_rptr;
3314 		iplen = ntohs(ipha->ipha_length);
3315 		err = tsol_check_label_v4(ixa->ixa_tsl,
3316 		    ixa->ixa_zoneid, mpp, CONN_MAC_DEFAULT, B_FALSE,
3317 		    ixa->ixa_ipst, &effective_tsl);
3318 		if (err != 0)
3319 			return (err);
3320 
3321 		ipha = (ipha_t *)(*mpp)->b_rptr;
3322 		adjust = (int)ntohs(ipha->ipha_length) - iplen;
3323 	} else {
3324 		ip6_t *ip6h;
3325 
3326 		ip6h = (ip6_t *)(*mpp)->b_rptr;
3327 		iplen = ntohs(ip6h->ip6_plen);
3328 
3329 		err = tsol_check_label_v6(ixa->ixa_tsl,
3330 		    ixa->ixa_zoneid, mpp, CONN_MAC_DEFAULT, B_FALSE,
3331 		    ixa->ixa_ipst, &effective_tsl);
3332 		if (err != 0)
3333 			return (err);
3334 
3335 		ip6h = (ip6_t *)(*mpp)->b_rptr;
3336 		adjust = (int)ntohs(ip6h->ip6_plen) - iplen;
3337 	}
3338 
3339 	if (effective_tsl != NULL) {
3340 		/* Update the label */
3341 		ip_xmit_attr_replace_tsl(ixa, effective_tsl);
3342 	}
3343 	ixa->ixa_pktlen += adjust;
3344 	ixa->ixa_ip_hdr_length += adjust;
3345 	return (0);
3346 }
3347 
3348 
3349 static void
3350 iptun_output_common(iptun_t *iptun, ip_xmit_attr_t *ixa, mblk_t *mp)
3351 {
3352 	ipsec_tun_pol_t	*itp = iptun->iptun_itp;
3353 	int		outer_hlen;
3354 	mblk_t		*newmp;
3355 	ipha_t		*outer4, *inner4;
3356 	ip6_t		*outer6, *inner6;
3357 	int		error;
3358 	boolean_t	update_pktlen;
3359 
3360 	ASSERT(ixa->ixa_ire != NULL);
3361 
3362 	outer_hlen = iptun_find_headers(mp, 0, &outer4, &inner4, &outer6,
3363 	    &inner6);
3364 	if (outer_hlen == 0) {
3365 		iptun_drop_pkt(mp, &iptun->iptun_oerrors);
3366 		return;
3367 	}
3368 
3369 	/* Save IXAF_DONTFRAG value */
3370 	iaflags_t dontfrag = ixa->ixa_flags & IXAF_DONTFRAG;
3371 
3372 	/* Perform header processing. */
3373 	if (outer4 != NULL) {
3374 		mp = iptun_out_process_ipv4(iptun, mp, outer4, inner4, inner6,
3375 		    ixa);
3376 	} else {
3377 		mp = iptun_out_process_ipv6(iptun, mp, outer6, inner4, inner6,
3378 		    ixa);
3379 	}
3380 	if (mp == NULL)
3381 		return;
3382 
3383 	/*
3384 	 * Let's hope the compiler optimizes this with "branch taken".
3385 	 */
3386 	if (itp != NULL && (itp->itp_flags & ITPF_P_ACTIVE)) {
3387 		/* This updates the ip_xmit_attr_t */
3388 		mp = ipsec_tun_outbound(mp, iptun, inner4, inner6, outer4,
3389 		    outer6, outer_hlen, ixa);
3390 		if (mp == NULL) {
3391 			atomic_inc_64(&iptun->iptun_oerrors);
3392 			return;
3393 		}
3394 		if (is_system_labeled()) {
3395 			/*
3396 			 * Might change the packet by adding/removing CIPSO.
3397 			 * After this caller inner* and outer* and outer_hlen
3398 			 * might be invalid.
3399 			 */
3400 			error = iptun_output_check_label(&mp, ixa);
3401 			if (error != 0) {
3402 				ip2dbg(("label check failed (%d)\n", error));
3403 				iptun_drop_pkt(mp, &iptun->iptun_oerrors);
3404 				return;
3405 			}
3406 		}
3407 
3408 		/*
3409 		 * ipsec_tun_outbound() returns a chain of tunneled IP
3410 		 * fragments linked with b_next (or a single message if the
3411 		 * tunneled packet wasn't a fragment).
3412 		 * If fragcache returned a list then we need to update
3413 		 * ixa_pktlen for all packets in the list.
3414 		 */
3415 		update_pktlen = (mp->b_next != NULL);
3416 
3417 		/*
3418 		 * Otherwise, we're good to go.  The ixa has been updated with
3419 		 * instructions for outbound IPsec processing.
3420 		 */
3421 		for (newmp = mp; newmp != NULL; newmp = mp) {
3422 			size_t minmtu = iptun->iptun_typeinfo->iti_minmtu;
3423 
3424 			atomic_inc_64(&iptun->iptun_opackets);
3425 			atomic_add_64(&iptun->iptun_obytes, ixa->ixa_pktlen);
3426 			mp = mp->b_next;
3427 			newmp->b_next = NULL;
3428 
3429 			/*
3430 			 * The IXAF_DONTFRAG flag is global, but there is
3431 			 * a chain here.  Check if we're really already
3432 			 * smaller than the minimum allowed MTU and reset here
3433 			 * appropriately.  Otherwise one small packet can kill
3434 			 * the whole chain's path mtu discovery.
3435 			 * In addition, update the pktlen to the length of
3436 			 * the actual packet being processed.
3437 			 */
3438 			if (update_pktlen) {
3439 				ixa->ixa_pktlen = msgdsize(newmp);
3440 				if (ixa->ixa_pktlen <= minmtu)
3441 					ixa->ixa_flags &= ~IXAF_DONTFRAG;
3442 			}
3443 
3444 			atomic_inc_64(&iptun->iptun_opackets);
3445 			atomic_add_64(&iptun->iptun_obytes, ixa->ixa_pktlen);
3446 
3447 			error = conn_ip_output(newmp, ixa);
3448 
3449 			/* Restore IXAF_DONTFRAG value */
3450 			ixa->ixa_flags |= dontfrag;
3451 
3452 			if (error == EMSGSIZE) {
3453 				/* IPsec policy might have changed */
3454 				(void) iptun_update_mtu(iptun, ixa, 0);
3455 			}
3456 		}
3457 	} else {
3458 		/*
3459 		 * The ip module will potentially apply global policy to the
3460 		 * packet in its output path if there's no active tunnel
3461 		 * policy.
3462 		 */
3463 		ASSERT(ixa->ixa_ipsec_policy == NULL);
3464 		mp = ip_output_attach_policy(mp, outer4, outer6, NULL, ixa);
3465 		if (mp == NULL) {
3466 			atomic_inc_64(&iptun->iptun_oerrors);
3467 			return;
3468 		}
3469 		if (is_system_labeled()) {
3470 			/*
3471 			 * Might change the packet by adding/removing CIPSO.
3472 			 * After this caller inner* and outer* and outer_hlen
3473 			 * might be invalid.
3474 			 */
3475 			error = iptun_output_check_label(&mp, ixa);
3476 			if (error != 0) {
3477 				ip2dbg(("label check failed (%d)\n", error));
3478 				iptun_drop_pkt(mp, &iptun->iptun_oerrors);
3479 				return;
3480 			}
3481 		}
3482 
3483 		atomic_inc_64(&iptun->iptun_opackets);
3484 		atomic_add_64(&iptun->iptun_obytes, ixa->ixa_pktlen);
3485 
3486 		error = conn_ip_output(mp, ixa);
3487 		if (error == EMSGSIZE) {
3488 			/* IPsec policy might have changed */
3489 			(void) iptun_update_mtu(iptun, ixa, 0);
3490 		}
3491 	}
3492 	if (ixa->ixa_flags & IXAF_IPSEC_SECURE)
3493 		ipsec_out_release_refs(ixa);
3494 }
3495 
3496 static mac_callbacks_t iptun_m_callbacks = {
3497 	.mc_callbacks	= (MC_SETPROP | MC_GETPROP | MC_PROPINFO),
3498 	.mc_getstat	= iptun_m_getstat,
3499 	.mc_start	= iptun_m_start,
3500 	.mc_stop	= iptun_m_stop,
3501 	.mc_setpromisc	= iptun_m_setpromisc,
3502 	.mc_multicst	= iptun_m_multicst,
3503 	.mc_unicst	= iptun_m_unicst,
3504 	.mc_tx		= iptun_m_tx,
3505 	.mc_reserved	= NULL,
3506 	.mc_setprop	= iptun_m_setprop,
3507 	.mc_getprop	= iptun_m_getprop,
3508 	.mc_propinfo	= iptun_m_propinfo
3509 };
3510