xref: /titanic_50/usr/src/uts/common/inet/sctp/sctp_common.c (revision 6e375c8351497b82ffa4f33cbf61d712999b4605)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 #include <sys/types.h>
28 #include <sys/systm.h>
29 #include <sys/stream.h>
30 #include <sys/strsubr.h>
31 #include <sys/ddi.h>
32 #include <sys/sunddi.h>
33 #include <sys/kmem.h>
34 #include <sys/socket.h>
35 #include <sys/random.h>
36 #include <sys/tsol/tndb.h>
37 #include <sys/tsol/tnet.h>
38 
39 #include <netinet/in.h>
40 #include <netinet/ip6.h>
41 #include <netinet/sctp.h>
42 
43 #include <inet/common.h>
44 #include <inet/ip.h>
45 #include <inet/ip6.h>
46 #include <inet/ip_ire.h>
47 #include <inet/mib2.h>
48 #include <inet/nd.h>
49 #include <inet/optcom.h>
50 #include <inet/sctp_ip.h>
51 #include <inet/ipclassifier.h>
52 
53 #include "sctp_impl.h"
54 #include "sctp_addr.h"
55 #include "sctp_asconf.h"
56 
57 static struct kmem_cache *sctp_kmem_faddr_cache;
58 static void sctp_init_faddr(sctp_t *, sctp_faddr_t *, in6_addr_t *, mblk_t *);
59 
60 /* Set the source address.  Refer to comments in sctp_get_ire(). */
61 void
62 sctp_set_saddr(sctp_t *sctp, sctp_faddr_t *fp)
63 {
64 	boolean_t v6 = !fp->isv4;
65 	boolean_t addr_set;
66 
67 	fp->saddr = sctp_get_valid_addr(sctp, v6, &addr_set);
68 	/*
69 	 * If there is no source address avaialble, mark this peer address
70 	 * as unreachable for now.  When the heartbeat timer fires, it will
71 	 * call sctp_get_ire() to re-check if there is any source address
72 	 * available.
73 	 */
74 	if (!addr_set)
75 		fp->state = SCTP_FADDRS_UNREACH;
76 }
77 
78 /*
79  * Call this function to update the cached IRE of a peer addr fp.
80  */
81 void
82 sctp_get_ire(sctp_t *sctp, sctp_faddr_t *fp)
83 {
84 	ire_t		*ire;
85 	ipaddr_t	addr4;
86 	in6_addr_t	laddr;
87 	sctp_saddr_ipif_t *sp;
88 	int		hdrlen;
89 	ts_label_t	*tsl;
90 	sctp_stack_t	*sctps = sctp->sctp_sctps;
91 	ip_stack_t	*ipst = sctps->sctps_netstack->netstack_ip;
92 
93 	/* Remove the previous cache IRE */
94 	if ((ire = fp->ire) != NULL) {
95 		IRE_REFRELE_NOTR(ire);
96 		fp->ire = NULL;
97 	}
98 
99 	/*
100 	 * If this addr is not reachable, mark it as unconfirmed for now, the
101 	 * state will be changed back to unreachable later in this function
102 	 * if it is still the case.
103 	 */
104 	if (fp->state == SCTP_FADDRS_UNREACH) {
105 		fp->state = SCTP_FADDRS_UNCONFIRMED;
106 	}
107 
108 	tsl = crgetlabel(CONN_CRED(sctp->sctp_connp));
109 
110 	if (fp->isv4) {
111 		IN6_V4MAPPED_TO_IPADDR(&fp->faddr, addr4);
112 		ire = ire_cache_lookup(addr4, sctp->sctp_zoneid, tsl, ipst);
113 		if (ire != NULL)
114 			IN6_IPADDR_TO_V4MAPPED(ire->ire_src_addr, &laddr);
115 	} else {
116 		ire = ire_cache_lookup_v6(&fp->faddr, sctp->sctp_zoneid, tsl,
117 		    ipst);
118 		if (ire != NULL)
119 			laddr = ire->ire_src_addr_v6;
120 	}
121 
122 	if (ire == NULL) {
123 		dprint(3, ("ire2faddr: no ire for %x:%x:%x:%x\n",
124 		    SCTP_PRINTADDR(fp->faddr)));
125 		/*
126 		 * It is tempting to just leave the src addr
127 		 * unspecified and let IP figure it out, but we
128 		 * *cannot* do this, since IP may choose a src addr
129 		 * that is not part of this association... unless
130 		 * this sctp has bound to all addrs.  So if the ire
131 		 * lookup fails, try to find one in our src addr
132 		 * list, unless the sctp has bound to all addrs, in
133 		 * which case we change the src addr to unspec.
134 		 *
135 		 * Note that if this is a v6 endpoint but it does
136 		 * not have any v4 address at this point (e.g. may
137 		 * have been  deleted), sctp_get_valid_addr() will
138 		 * return mapped INADDR_ANY.  In this case, this
139 		 * address should be marked not reachable so that
140 		 * it won't be used to send data.
141 		 */
142 		sctp_set_saddr(sctp, fp);
143 		if (fp->state == SCTP_FADDRS_UNREACH)
144 			return;
145 		goto check_current;
146 	}
147 
148 	/* Make sure the laddr is part of this association */
149 	if ((sp = sctp_saddr_lookup(sctp, &ire->ire_ipif->ipif_v6lcl_addr,
150 	    0)) != NULL && !sp->saddr_ipif_dontsrc) {
151 		if (sp->saddr_ipif_unconfirmed == 1)
152 			sp->saddr_ipif_unconfirmed = 0;
153 		fp->saddr = laddr;
154 	} else {
155 		dprint(2, ("ire2faddr: src addr is not part of assc\n"));
156 
157 		/*
158 		 * Set the src to the first saddr and hope for the best.
159 		 * Note that we will still do the ire caching below.
160 		 * Otherwise, whenever we send a packet, we need to do
161 		 * the ire lookup again and still may not get the correct
162 		 * source address.  Note that this case should very seldomly
163 		 * happen.  One scenario this can happen is an app
164 		 * explicitly bind() to an address.  But that address is
165 		 * not the preferred source address to send to the peer.
166 		 */
167 		sctp_set_saddr(sctp, fp);
168 		if (fp->state == SCTP_FADDRS_UNREACH) {
169 			IRE_REFRELE(ire);
170 			return;
171 		}
172 	}
173 
174 	/*
175 	 * Note that ire_cache_lookup_*() returns an ire with the tracing
176 	 * bits enabled.  This requires the thread holding the ire also
177 	 * do the IRE_REFRELE().  Thus we need to do IRE_REFHOLD_NOTR()
178 	 * and then IRE_REFRELE() the ire here to make the tracing bits
179 	 * work.
180 	 */
181 	IRE_REFHOLD_NOTR(ire);
182 	IRE_REFRELE(ire);
183 
184 	/* Cache the IRE */
185 	fp->ire = ire;
186 	if (fp->ire->ire_type == IRE_LOOPBACK && !sctp->sctp_loopback)
187 		sctp->sctp_loopback = 1;
188 
189 	/*
190 	 * Pull out RTO information for this faddr and use it if we don't
191 	 * have any yet.
192 	 */
193 	if (fp->srtt == -1 && ire->ire_uinfo.iulp_rtt != 0) {
194 		/* The cached value is in ms. */
195 		fp->srtt = MSEC_TO_TICK(ire->ire_uinfo.iulp_rtt);
196 		fp->rttvar = MSEC_TO_TICK(ire->ire_uinfo.iulp_rtt_sd);
197 		fp->rto = 3 * fp->srtt;
198 
199 		/* Bound the RTO by configured min and max values */
200 		if (fp->rto < sctp->sctp_rto_min) {
201 			fp->rto = sctp->sctp_rto_min;
202 		}
203 		if (fp->rto > sctp->sctp_rto_max) {
204 			fp->rto = sctp->sctp_rto_max;
205 		}
206 	}
207 
208 	/*
209 	 * Record the MTU for this faddr. If the MTU for this faddr has
210 	 * changed, check if the assc MTU will also change.
211 	 */
212 	if (fp->isv4) {
213 		hdrlen = sctp->sctp_hdr_len;
214 	} else {
215 		hdrlen = sctp->sctp_hdr6_len;
216 	}
217 	if ((fp->sfa_pmss + hdrlen) != ire->ire_max_frag) {
218 		/* Make sure that sfa_pmss is a multiple of SCTP_ALIGN. */
219 		fp->sfa_pmss = (ire->ire_max_frag - hdrlen) & ~(SCTP_ALIGN - 1);
220 		if (fp->cwnd < (fp->sfa_pmss * 2)) {
221 			SET_CWND(fp, fp->sfa_pmss,
222 			    sctps->sctps_slow_start_initial);
223 		}
224 	}
225 
226 check_current:
227 	if (fp == sctp->sctp_current)
228 		sctp_set_faddr_current(sctp, fp);
229 }
230 
231 void
232 sctp_update_ire(sctp_t *sctp)
233 {
234 	ire_t		*ire;
235 	sctp_faddr_t	*fp;
236 	sctp_stack_t	*sctps = sctp->sctp_sctps;
237 
238 	for (fp = sctp->sctp_faddrs; fp != NULL; fp = fp->next) {
239 		if ((ire = fp->ire) == NULL)
240 			continue;
241 		mutex_enter(&ire->ire_lock);
242 
243 		/*
244 		 * If the cached IRE is going away, there is no point to
245 		 * update it.
246 		 */
247 		if (ire->ire_marks & IRE_MARK_CONDEMNED) {
248 			mutex_exit(&ire->ire_lock);
249 			IRE_REFRELE_NOTR(ire);
250 			fp->ire = NULL;
251 			continue;
252 		}
253 
254 		/*
255 		 * Only record the PMTU for this faddr if we actually have
256 		 * done discovery. This prevents initialized default from
257 		 * clobbering any real info that IP may have.
258 		 */
259 		if (fp->pmtu_discovered) {
260 			if (fp->isv4) {
261 				ire->ire_max_frag = fp->sfa_pmss +
262 				    sctp->sctp_hdr_len;
263 			} else {
264 				ire->ire_max_frag = fp->sfa_pmss +
265 				    sctp->sctp_hdr6_len;
266 			}
267 		}
268 
269 		if (sctps->sctps_rtt_updates != 0 &&
270 		    fp->rtt_updates >= sctps->sctps_rtt_updates) {
271 			/*
272 			 * If there is no old cached values, initialize them
273 			 * conservatively.  Set them to be (1.5 * new value).
274 			 * This code copied from ip_ire_advise().  The cached
275 			 * value is in ms.
276 			 */
277 			if (ire->ire_uinfo.iulp_rtt != 0) {
278 				ire->ire_uinfo.iulp_rtt =
279 				    (ire->ire_uinfo.iulp_rtt +
280 				    TICK_TO_MSEC(fp->srtt)) >> 1;
281 			} else {
282 				ire->ire_uinfo.iulp_rtt =
283 				    TICK_TO_MSEC(fp->srtt + (fp->srtt >> 1));
284 			}
285 			if (ire->ire_uinfo.iulp_rtt_sd != 0) {
286 				ire->ire_uinfo.iulp_rtt_sd =
287 				    (ire->ire_uinfo.iulp_rtt_sd +
288 				    TICK_TO_MSEC(fp->rttvar)) >> 1;
289 			} else {
290 				ire->ire_uinfo.iulp_rtt_sd =
291 				    TICK_TO_MSEC(fp->rttvar +
292 				    (fp->rttvar >> 1));
293 			}
294 			fp->rtt_updates = 0;
295 		}
296 		mutex_exit(&ire->ire_lock);
297 	}
298 }
299 
300 /*
301  * The sender must set the total length in the IP header.
302  * If sendto == NULL, the current will be used.
303  */
304 mblk_t *
305 sctp_make_mp(sctp_t *sctp, sctp_faddr_t *sendto, int trailer)
306 {
307 	mblk_t *mp;
308 	size_t ipsctplen;
309 	int isv4;
310 	sctp_faddr_t *fp;
311 	sctp_stack_t *sctps = sctp->sctp_sctps;
312 	boolean_t src_changed = B_FALSE;
313 
314 	ASSERT(sctp->sctp_current != NULL || sendto != NULL);
315 	if (sendto == NULL) {
316 		fp = sctp->sctp_current;
317 	} else {
318 		fp = sendto;
319 	}
320 	isv4 = fp->isv4;
321 
322 	/* Try to look for another IRE again. */
323 	if (fp->ire == NULL) {
324 		sctp_get_ire(sctp, fp);
325 		/*
326 		 * Although we still may not get an IRE, the source address
327 		 * may be changed in sctp_get_ire().  Set src_changed to
328 		 * true so that the source address is copied again.
329 		 */
330 		src_changed = B_TRUE;
331 	}
332 
333 	/* There is no suitable source address to use, return. */
334 	if (fp->state == SCTP_FADDRS_UNREACH)
335 		return (NULL);
336 	ASSERT(!IN6_IS_ADDR_V4MAPPED_ANY(&fp->saddr));
337 
338 	if (isv4) {
339 		ipsctplen = sctp->sctp_hdr_len;
340 	} else {
341 		ipsctplen = sctp->sctp_hdr6_len;
342 	}
343 
344 	mp = allocb_cred(ipsctplen + sctps->sctps_wroff_xtra + trailer,
345 	    CONN_CRED(sctp->sctp_connp));
346 	if (mp == NULL) {
347 		ip1dbg(("sctp_make_mp: error making mp..\n"));
348 		return (NULL);
349 	}
350 	mp->b_rptr += sctps->sctps_wroff_xtra;
351 	mp->b_wptr = mp->b_rptr + ipsctplen;
352 
353 	ASSERT(OK_32PTR(mp->b_wptr));
354 
355 	if (isv4) {
356 		ipha_t *iph = (ipha_t *)mp->b_rptr;
357 
358 		bcopy(sctp->sctp_iphc, mp->b_rptr, ipsctplen);
359 		if (fp != sctp->sctp_current || src_changed) {
360 			/* Fix the source and destination addresses. */
361 			IN6_V4MAPPED_TO_IPADDR(&fp->faddr, iph->ipha_dst);
362 			IN6_V4MAPPED_TO_IPADDR(&fp->saddr, iph->ipha_src);
363 		}
364 		/* set or clear the don't fragment bit */
365 		if (fp->df) {
366 			iph->ipha_fragment_offset_and_flags = htons(IPH_DF);
367 		} else {
368 			iph->ipha_fragment_offset_and_flags = 0;
369 		}
370 	} else {
371 		bcopy(sctp->sctp_iphc6, mp->b_rptr, ipsctplen);
372 		if (fp != sctp->sctp_current || src_changed) {
373 			/* Fix the source and destination addresses. */
374 			((ip6_t *)(mp->b_rptr))->ip6_dst = fp->faddr;
375 			((ip6_t *)(mp->b_rptr))->ip6_src = fp->saddr;
376 		}
377 	}
378 	ASSERT(sctp->sctp_connp != NULL);
379 
380 	/*
381 	 * IP will not free this IRE if it is condemned.  SCTP needs to
382 	 * free it.
383 	 */
384 	if ((fp->ire != NULL) && (fp->ire->ire_marks & IRE_MARK_CONDEMNED)) {
385 		IRE_REFRELE_NOTR(fp->ire);
386 		fp->ire = NULL;
387 	}
388 	/* Stash the conn and ire ptr info. for IP */
389 	SCTP_STASH_IPINFO(mp, fp->ire);
390 
391 	return (mp);
392 }
393 
394 /*
395  * Notify upper layers about preferred write offset, write size.
396  */
397 void
398 sctp_set_ulp_prop(sctp_t *sctp)
399 {
400 	int hdrlen;
401 	struct sock_proto_props sopp;
402 
403 	sctp_stack_t *sctps = sctp->sctp_sctps;
404 
405 	if (sctp->sctp_current->isv4) {
406 		hdrlen = sctp->sctp_hdr_len;
407 	} else {
408 		hdrlen = sctp->sctp_hdr6_len;
409 	}
410 	ASSERT(sctp->sctp_ulpd);
411 
412 	ASSERT(sctp->sctp_current->sfa_pmss == sctp->sctp_mss);
413 	bzero(&sopp, sizeof (sopp));
414 	sopp.sopp_flags = SOCKOPT_MAXBLK|SOCKOPT_WROFF;
415 	sopp.sopp_wroff = sctps->sctps_wroff_xtra + hdrlen +
416 	    sizeof (sctp_data_hdr_t);
417 	sopp.sopp_maxblk = sctp->sctp_mss - sizeof (sctp_data_hdr_t);
418 	sctp->sctp_ulp_prop(sctp->sctp_ulpd, &sopp);
419 }
420 
421 void
422 sctp_set_iplen(sctp_t *sctp, mblk_t *mp)
423 {
424 	uint16_t	sum = 0;
425 	ipha_t		*iph;
426 	ip6_t		*ip6h;
427 	mblk_t		*pmp = mp;
428 	boolean_t	isv4;
429 
430 	isv4 = (IPH_HDR_VERSION(mp->b_rptr) == IPV4_VERSION);
431 	for (; pmp; pmp = pmp->b_cont)
432 		sum += pmp->b_wptr - pmp->b_rptr;
433 
434 	if (isv4) {
435 		iph = (ipha_t *)mp->b_rptr;
436 		iph->ipha_length = htons(sum);
437 	} else {
438 		ip6h = (ip6_t *)mp->b_rptr;
439 		/*
440 		 * If an ip6i_t is present, the real IPv6 header
441 		 * immediately follows.
442 		 */
443 		if (ip6h->ip6_nxt == IPPROTO_RAW)
444 			ip6h = (ip6_t *)&ip6h[1];
445 		ip6h->ip6_plen = htons(sum - ((char *)&sctp->sctp_ip6h[1] -
446 		    sctp->sctp_iphc6));
447 	}
448 }
449 
450 int
451 sctp_compare_faddrsets(sctp_faddr_t *a1, sctp_faddr_t *a2)
452 {
453 	int na1 = 0;
454 	int overlap = 0;
455 	int equal = 1;
456 	int onematch;
457 	sctp_faddr_t *fp1, *fp2;
458 
459 	for (fp1 = a1; fp1; fp1 = fp1->next) {
460 		onematch = 0;
461 		for (fp2 = a2; fp2; fp2 = fp2->next) {
462 			if (IN6_ARE_ADDR_EQUAL(&fp1->faddr, &fp2->faddr)) {
463 				overlap++;
464 				onematch = 1;
465 				break;
466 			}
467 			if (!onematch) {
468 				equal = 0;
469 			}
470 		}
471 		na1++;
472 	}
473 
474 	if (equal) {
475 		return (SCTP_ADDR_EQUAL);
476 	}
477 	if (overlap == na1) {
478 		return (SCTP_ADDR_SUBSET);
479 	}
480 	if (overlap) {
481 		return (SCTP_ADDR_OVERLAP);
482 	}
483 	return (SCTP_ADDR_DISJOINT);
484 }
485 
486 /*
487  * Returns 0 on success, -1 on memory allocation failure. If sleep
488  * is true, this function should never fail.  The boolean parameter
489  * first decides whether the newly created faddr structure should be
490  * added at the beginning of the list or at the end.
491  *
492  * Note: caller must hold conn fanout lock.
493  */
494 int
495 sctp_add_faddr(sctp_t *sctp, in6_addr_t *addr, int sleep, boolean_t first)
496 {
497 	sctp_faddr_t	*faddr;
498 	mblk_t		*timer_mp;
499 
500 	if (is_system_labeled()) {
501 		ts_label_t *tsl;
502 		tsol_tpc_t *rhtp;
503 		int retv;
504 
505 		tsl = crgetlabel(CONN_CRED(sctp->sctp_connp));
506 		ASSERT(tsl != NULL);
507 
508 		/* find_tpc automatically does the right thing with IPv4 */
509 		rhtp = find_tpc(addr, IPV6_VERSION, B_FALSE);
510 		if (rhtp == NULL)
511 			return (EACCES);
512 
513 		retv = EACCES;
514 		if (tsl->tsl_doi == rhtp->tpc_tp.tp_doi) {
515 			switch (rhtp->tpc_tp.host_type) {
516 			case UNLABELED:
517 				/*
518 				 * Can talk to unlabeled hosts if any of the
519 				 * following are true:
520 				 *   1. zone's label matches the remote host's
521 				 *	default label,
522 				 *   2. mac_exempt is on and the zone dominates
523 				 *	the remote host's label, or
524 				 *   3. mac_exempt is on and the socket is from
525 				 *	the global zone.
526 				 */
527 				if (blequal(&rhtp->tpc_tp.tp_def_label,
528 				    &tsl->tsl_label) ||
529 				    (sctp->sctp_mac_exempt &&
530 				    (sctp->sctp_zoneid == GLOBAL_ZONEID ||
531 				    bldominates(&tsl->tsl_label,
532 				    &rhtp->tpc_tp.tp_def_label))))
533 					retv = 0;
534 				break;
535 			case SUN_CIPSO:
536 				if (_blinrange(&tsl->tsl_label,
537 				    &rhtp->tpc_tp.tp_sl_range_cipso) ||
538 				    blinlset(&tsl->tsl_label,
539 				    rhtp->tpc_tp.tp_sl_set_cipso))
540 					retv = 0;
541 				break;
542 			}
543 		}
544 		TPC_RELE(rhtp);
545 		if (retv != 0)
546 			return (retv);
547 	}
548 
549 	if ((faddr = kmem_cache_alloc(sctp_kmem_faddr_cache, sleep)) == NULL)
550 		return (ENOMEM);
551 	timer_mp = sctp_timer_alloc((sctp), sctp_rexmit_timer, sleep);
552 	if (timer_mp == NULL) {
553 		kmem_cache_free(sctp_kmem_faddr_cache, faddr);
554 		return (ENOMEM);
555 	}
556 	((sctpt_t *)(timer_mp->b_rptr))->sctpt_faddr = faddr;
557 
558 	sctp_init_faddr(sctp, faddr, addr, timer_mp);
559 
560 	/* Check for subnet broadcast. */
561 	if (faddr->ire != NULL && faddr->ire->ire_type & IRE_BROADCAST) {
562 		IRE_REFRELE_NOTR(faddr->ire);
563 		sctp_timer_free(timer_mp);
564 		faddr->timer_mp = NULL;
565 		kmem_cache_free(sctp_kmem_faddr_cache, faddr);
566 		return (EADDRNOTAVAIL);
567 	}
568 	ASSERT(faddr->next == NULL);
569 
570 	if (sctp->sctp_faddrs == NULL) {
571 		ASSERT(sctp->sctp_lastfaddr == NULL);
572 		/* only element on list; first and last are same */
573 		sctp->sctp_faddrs = sctp->sctp_lastfaddr = faddr;
574 	} else if (first) {
575 		ASSERT(sctp->sctp_lastfaddr != NULL);
576 		faddr->next = sctp->sctp_faddrs;
577 		sctp->sctp_faddrs = faddr;
578 	} else {
579 		sctp->sctp_lastfaddr->next = faddr;
580 		sctp->sctp_lastfaddr = faddr;
581 	}
582 	sctp->sctp_nfaddrs++;
583 
584 	return (0);
585 }
586 
587 sctp_faddr_t *
588 sctp_lookup_faddr(sctp_t *sctp, in6_addr_t *addr)
589 {
590 	sctp_faddr_t *fp;
591 
592 	for (fp = sctp->sctp_faddrs; fp != NULL; fp = fp->next) {
593 		if (IN6_ARE_ADDR_EQUAL(&fp->faddr, addr))
594 			break;
595 	}
596 
597 	return (fp);
598 }
599 
600 sctp_faddr_t *
601 sctp_lookup_faddr_nosctp(sctp_faddr_t *fp, in6_addr_t *addr)
602 {
603 	for (; fp; fp = fp->next) {
604 		if (IN6_ARE_ADDR_EQUAL(&fp->faddr, addr)) {
605 			break;
606 		}
607 	}
608 
609 	return (fp);
610 }
611 
612 /*
613  * To change the currently used peer address to the specified one.
614  */
615 void
616 sctp_set_faddr_current(sctp_t *sctp, sctp_faddr_t *fp)
617 {
618 	/* Now setup the composite header. */
619 	if (fp->isv4) {
620 		IN6_V4MAPPED_TO_IPADDR(&fp->faddr,
621 		    sctp->sctp_ipha->ipha_dst);
622 		IN6_V4MAPPED_TO_IPADDR(&fp->saddr, sctp->sctp_ipha->ipha_src);
623 		/* update don't fragment bit */
624 		if (fp->df) {
625 			sctp->sctp_ipha->ipha_fragment_offset_and_flags =
626 			    htons(IPH_DF);
627 		} else {
628 			sctp->sctp_ipha->ipha_fragment_offset_and_flags = 0;
629 		}
630 	} else {
631 		sctp->sctp_ip6h->ip6_dst = fp->faddr;
632 		sctp->sctp_ip6h->ip6_src = fp->saddr;
633 	}
634 
635 	sctp->sctp_current = fp;
636 	sctp->sctp_mss = fp->sfa_pmss;
637 
638 	/* Update the uppper layer for the change. */
639 	if (!SCTP_IS_DETACHED(sctp))
640 		sctp_set_ulp_prop(sctp);
641 }
642 
643 void
644 sctp_redo_faddr_srcs(sctp_t *sctp)
645 {
646 	sctp_faddr_t *fp;
647 
648 	for (fp = sctp->sctp_faddrs; fp != NULL; fp = fp->next) {
649 		sctp_get_ire(sctp, fp);
650 	}
651 }
652 
653 void
654 sctp_faddr_alive(sctp_t *sctp, sctp_faddr_t *fp)
655 {
656 	int64_t now = lbolt64;
657 
658 	fp->strikes = 0;
659 	sctp->sctp_strikes = 0;
660 	fp->lastactive = now;
661 	fp->hb_expiry = now + SET_HB_INTVL(fp);
662 	fp->hb_pending = B_FALSE;
663 	if (fp->state != SCTP_FADDRS_ALIVE) {
664 		fp->state = SCTP_FADDRS_ALIVE;
665 		sctp_intf_event(sctp, fp->faddr, SCTP_ADDR_AVAILABLE, 0);
666 		/* Should have a full IRE now */
667 		sctp_get_ire(sctp, fp);
668 
669 		/*
670 		 * If this is the primary, switch back to it now.  And
671 		 * we probably want to reset the source addr used to reach
672 		 * it.
673 		 */
674 		if (fp == sctp->sctp_primary) {
675 			ASSERT(fp->state != SCTP_FADDRS_UNREACH);
676 			sctp_set_faddr_current(sctp, fp);
677 			return;
678 		}
679 	}
680 }
681 
682 int
683 sctp_is_a_faddr_clean(sctp_t *sctp)
684 {
685 	sctp_faddr_t *fp;
686 
687 	for (fp = sctp->sctp_faddrs; fp; fp = fp->next) {
688 		if (fp->state == SCTP_FADDRS_ALIVE && fp->strikes == 0) {
689 			return (1);
690 		}
691 	}
692 
693 	return (0);
694 }
695 
696 /*
697  * Returns 0 if there is at leave one other active faddr, -1 if there
698  * are none. If there are none left, faddr_dead() will start killing the
699  * association.
700  * If the downed faddr was the current faddr, a new current faddr
701  * will be chosen.
702  */
703 int
704 sctp_faddr_dead(sctp_t *sctp, sctp_faddr_t *fp, int newstate)
705 {
706 	sctp_faddr_t *ofp;
707 	sctp_stack_t *sctps = sctp->sctp_sctps;
708 
709 	if (fp->state == SCTP_FADDRS_ALIVE) {
710 		sctp_intf_event(sctp, fp->faddr, SCTP_ADDR_UNREACHABLE, 0);
711 	}
712 	fp->state = newstate;
713 
714 	dprint(1, ("sctp_faddr_dead: %x:%x:%x:%x down (state=%d)\n",
715 	    SCTP_PRINTADDR(fp->faddr), newstate));
716 
717 	if (fp == sctp->sctp_current) {
718 		/* Current faddr down; need to switch it */
719 		sctp->sctp_current = NULL;
720 	}
721 
722 	/* Find next alive faddr */
723 	ofp = fp;
724 	for (fp = fp->next; fp != NULL; fp = fp->next) {
725 		if (fp->state == SCTP_FADDRS_ALIVE) {
726 			break;
727 		}
728 	}
729 
730 	if (fp == NULL) {
731 		/* Continue from beginning of list */
732 		for (fp = sctp->sctp_faddrs; fp != ofp; fp = fp->next) {
733 			if (fp->state == SCTP_FADDRS_ALIVE) {
734 				break;
735 			}
736 		}
737 	}
738 
739 	/*
740 	 * Find a new fp, so if the current faddr is dead, use the new fp
741 	 * as the current one.
742 	 */
743 	if (fp != ofp) {
744 		if (sctp->sctp_current == NULL) {
745 			dprint(1, ("sctp_faddr_dead: failover->%x:%x:%x:%x\n",
746 			    SCTP_PRINTADDR(fp->faddr)));
747 			/*
748 			 * Note that we don't need to reset the source addr
749 			 * of the new fp.
750 			 */
751 			sctp_set_faddr_current(sctp, fp);
752 		}
753 		return (0);
754 	}
755 
756 
757 	/* All faddrs are down; kill the association */
758 	dprint(1, ("sctp_faddr_dead: all faddrs down, killing assoc\n"));
759 	BUMP_MIB(&sctps->sctps_mib, sctpAborted);
760 	sctp_assoc_event(sctp, sctp->sctp_state < SCTPS_ESTABLISHED ?
761 	    SCTP_CANT_STR_ASSOC : SCTP_COMM_LOST, 0, NULL);
762 	sctp_clean_death(sctp, sctp->sctp_client_errno ?
763 	    sctp->sctp_client_errno : ETIMEDOUT);
764 
765 	return (-1);
766 }
767 
768 sctp_faddr_t *
769 sctp_rotate_faddr(sctp_t *sctp, sctp_faddr_t *ofp)
770 {
771 	sctp_faddr_t *nfp = NULL;
772 
773 	if (ofp == NULL) {
774 		ofp = sctp->sctp_current;
775 	}
776 
777 	/* Find the next live one */
778 	for (nfp = ofp->next; nfp != NULL; nfp = nfp->next) {
779 		if (nfp->state == SCTP_FADDRS_ALIVE) {
780 			break;
781 		}
782 	}
783 
784 	if (nfp == NULL) {
785 		/* Continue from beginning of list */
786 		for (nfp = sctp->sctp_faddrs; nfp != ofp; nfp = nfp->next) {
787 			if (nfp->state == SCTP_FADDRS_ALIVE) {
788 				break;
789 			}
790 		}
791 	}
792 
793 	/*
794 	 * nfp could only be NULL if all faddrs are down, and when
795 	 * this happens, faddr_dead() should have killed the
796 	 * association. Hence this assertion...
797 	 */
798 	ASSERT(nfp != NULL);
799 	return (nfp);
800 }
801 
802 void
803 sctp_unlink_faddr(sctp_t *sctp, sctp_faddr_t *fp)
804 {
805 	sctp_faddr_t *fpp;
806 
807 	if (!sctp->sctp_faddrs) {
808 		return;
809 	}
810 
811 	if (fp->timer_mp != NULL) {
812 		sctp_timer_free(fp->timer_mp);
813 		fp->timer_mp = NULL;
814 		fp->timer_running = 0;
815 	}
816 	if (fp->rc_timer_mp != NULL) {
817 		sctp_timer_free(fp->rc_timer_mp);
818 		fp->rc_timer_mp = NULL;
819 		fp->rc_timer_running = 0;
820 	}
821 	if (fp->ire != NULL) {
822 		IRE_REFRELE_NOTR(fp->ire);
823 		fp->ire = NULL;
824 	}
825 
826 	if (fp == sctp->sctp_faddrs) {
827 		goto gotit;
828 	}
829 
830 	for (fpp = sctp->sctp_faddrs; fpp->next != fp; fpp = fpp->next)
831 		;
832 
833 gotit:
834 	ASSERT(sctp->sctp_conn_tfp != NULL);
835 	mutex_enter(&sctp->sctp_conn_tfp->tf_lock);
836 	if (fp == sctp->sctp_faddrs) {
837 		sctp->sctp_faddrs = fp->next;
838 	} else {
839 		fpp->next = fp->next;
840 	}
841 	mutex_exit(&sctp->sctp_conn_tfp->tf_lock);
842 	/* XXX faddr2ire? */
843 	kmem_cache_free(sctp_kmem_faddr_cache, fp);
844 	sctp->sctp_nfaddrs--;
845 }
846 
847 void
848 sctp_zap_faddrs(sctp_t *sctp, int caller_holds_lock)
849 {
850 	sctp_faddr_t *fp, *fpn;
851 
852 	if (sctp->sctp_faddrs == NULL) {
853 		ASSERT(sctp->sctp_lastfaddr == NULL);
854 		return;
855 	}
856 
857 	ASSERT(sctp->sctp_lastfaddr != NULL);
858 	sctp->sctp_lastfaddr = NULL;
859 	sctp->sctp_current = NULL;
860 	sctp->sctp_primary = NULL;
861 
862 	sctp_free_faddr_timers(sctp);
863 
864 	if (sctp->sctp_conn_tfp != NULL && !caller_holds_lock) {
865 		/* in conn fanout; need to hold lock */
866 		mutex_enter(&sctp->sctp_conn_tfp->tf_lock);
867 	}
868 
869 	for (fp = sctp->sctp_faddrs; fp; fp = fpn) {
870 		fpn = fp->next;
871 		if (fp->ire != NULL)
872 			IRE_REFRELE_NOTR(fp->ire);
873 		kmem_cache_free(sctp_kmem_faddr_cache, fp);
874 		sctp->sctp_nfaddrs--;
875 	}
876 
877 	sctp->sctp_faddrs = NULL;
878 	ASSERT(sctp->sctp_nfaddrs == 0);
879 	if (sctp->sctp_conn_tfp != NULL && !caller_holds_lock) {
880 		mutex_exit(&sctp->sctp_conn_tfp->tf_lock);
881 	}
882 
883 }
884 
885 void
886 sctp_zap_addrs(sctp_t *sctp)
887 {
888 	sctp_zap_faddrs(sctp, 0);
889 	sctp_free_saddrs(sctp);
890 }
891 
892 /*
893  * Initialize the IPv4 header. Loses any record of any IP options.
894  */
895 int
896 sctp_header_init_ipv4(sctp_t *sctp, int sleep)
897 {
898 	sctp_hdr_t	*sctph;
899 	sctp_stack_t	*sctps = sctp->sctp_sctps;
900 
901 	/*
902 	 * This is a simple initialization. If there's
903 	 * already a template, it should never be too small,
904 	 * so reuse it.  Otherwise, allocate space for the new one.
905 	 */
906 	if (sctp->sctp_iphc != NULL) {
907 		ASSERT(sctp->sctp_iphc_len >= SCTP_MAX_COMBINED_HEADER_LENGTH);
908 		bzero(sctp->sctp_iphc, sctp->sctp_iphc_len);
909 	} else {
910 		sctp->sctp_iphc_len = SCTP_MAX_COMBINED_HEADER_LENGTH;
911 		sctp->sctp_iphc = kmem_zalloc(sctp->sctp_iphc_len, sleep);
912 		if (sctp->sctp_iphc == NULL) {
913 			sctp->sctp_iphc_len = 0;
914 			return (ENOMEM);
915 		}
916 	}
917 
918 	sctp->sctp_ipha = (ipha_t *)sctp->sctp_iphc;
919 
920 	sctp->sctp_hdr_len = sizeof (ipha_t) + sizeof (sctp_hdr_t);
921 	sctp->sctp_ip_hdr_len = sizeof (ipha_t);
922 	sctp->sctp_ipha->ipha_length = htons(sizeof (ipha_t) +
923 	    sizeof (sctp_hdr_t));
924 	sctp->sctp_ipha->ipha_version_and_hdr_length =
925 	    (IP_VERSION << 4) | IP_SIMPLE_HDR_LENGTH_IN_WORDS;
926 
927 	/*
928 	 * These two fields should be zero, and are already set above.
929 	 *
930 	 * sctp->sctp_ipha->ipha_ident,
931 	 * sctp->sctp_ipha->ipha_fragment_offset_and_flags.
932 	 */
933 
934 	sctp->sctp_ipha->ipha_ttl = sctps->sctps_ipv4_ttl;
935 	sctp->sctp_ipha->ipha_protocol = IPPROTO_SCTP;
936 
937 	sctph = (sctp_hdr_t *)(sctp->sctp_iphc + sizeof (ipha_t));
938 	sctp->sctp_sctph = sctph;
939 
940 	return (0);
941 }
942 
943 /*
944  * Update sctp_sticky_hdrs based on sctp_sticky_ipp.
945  * The headers include ip6i_t (if needed), ip6_t, any sticky extension
946  * headers, and the maximum size sctp header (to avoid reallocation
947  * on the fly for additional sctp options).
948  * Returns failure if can't allocate memory.
949  */
950 int
951 sctp_build_hdrs(sctp_t *sctp)
952 {
953 	char		*hdrs;
954 	uint_t		hdrs_len;
955 	ip6i_t		*ip6i;
956 	char		buf[SCTP_MAX_HDR_LENGTH];
957 	ip6_pkt_t	*ipp = &sctp->sctp_sticky_ipp;
958 	in6_addr_t	src;
959 	in6_addr_t	dst;
960 	sctp_stack_t	*sctps = sctp->sctp_sctps;
961 
962 	/*
963 	 * save the existing sctp header and source/dest IP addresses
964 	 */
965 	bcopy(sctp->sctp_sctph6, buf, sizeof (sctp_hdr_t));
966 	src = sctp->sctp_ip6h->ip6_src;
967 	dst = sctp->sctp_ip6h->ip6_dst;
968 	hdrs_len = ip_total_hdrs_len_v6(ipp) + SCTP_MAX_HDR_LENGTH;
969 	ASSERT(hdrs_len != 0);
970 	if (hdrs_len > sctp->sctp_iphc6_len) {
971 		/* Need to reallocate */
972 		hdrs = kmem_zalloc(hdrs_len, KM_NOSLEEP);
973 		if (hdrs == NULL)
974 			return (ENOMEM);
975 
976 		if (sctp->sctp_iphc6_len != 0)
977 			kmem_free(sctp->sctp_iphc6, sctp->sctp_iphc6_len);
978 		sctp->sctp_iphc6 = hdrs;
979 		sctp->sctp_iphc6_len = hdrs_len;
980 	}
981 	ip_build_hdrs_v6((uchar_t *)sctp->sctp_iphc6,
982 	    hdrs_len - SCTP_MAX_HDR_LENGTH, ipp, IPPROTO_SCTP);
983 
984 	/* Set header fields not in ipp */
985 	if (ipp->ipp_fields & IPPF_HAS_IP6I) {
986 		ip6i = (ip6i_t *)sctp->sctp_iphc6;
987 		sctp->sctp_ip6h = (ip6_t *)&ip6i[1];
988 	} else {
989 		sctp->sctp_ip6h = (ip6_t *)sctp->sctp_iphc6;
990 	}
991 	/*
992 	 * sctp->sctp_ip_hdr_len will include ip6i_t if there is one.
993 	 */
994 	sctp->sctp_ip_hdr6_len = hdrs_len - SCTP_MAX_HDR_LENGTH;
995 	sctp->sctp_sctph6 = (sctp_hdr_t *)(sctp->sctp_iphc6 +
996 	    sctp->sctp_ip_hdr6_len);
997 	sctp->sctp_hdr6_len = sctp->sctp_ip_hdr6_len + sizeof (sctp_hdr_t);
998 
999 	bcopy(buf, sctp->sctp_sctph6, sizeof (sctp_hdr_t));
1000 
1001 	sctp->sctp_ip6h->ip6_src = src;
1002 	sctp->sctp_ip6h->ip6_dst = dst;
1003 	/*
1004 	 * If the hoplimit was not set by ip_build_hdrs_v6(), we need to
1005 	 * set it to the default value for SCTP.
1006 	 */
1007 	if (!(ipp->ipp_fields & IPPF_UNICAST_HOPS))
1008 		sctp->sctp_ip6h->ip6_hops = sctps->sctps_ipv6_hoplimit;
1009 	/*
1010 	 * If we're setting extension headers after a connection
1011 	 * has been established, and if we have a routing header
1012 	 * among the extension headers, call ip_massage_options_v6 to
1013 	 * manipulate the routing header/ip6_dst set the checksum
1014 	 * difference in the sctp header template.
1015 	 * (This happens in sctp_connect_ipv6 if the routing header
1016 	 * is set prior to the connect.)
1017 	 */
1018 
1019 	if ((sctp->sctp_state >= SCTPS_COOKIE_WAIT) &&
1020 	    (sctp->sctp_sticky_ipp.ipp_fields & IPPF_RTHDR)) {
1021 		ip6_rthdr_t *rth;
1022 
1023 		rth = ip_find_rthdr_v6(sctp->sctp_ip6h,
1024 		    (uint8_t *)sctp->sctp_sctph6);
1025 		if (rth != NULL) {
1026 			(void) ip_massage_options_v6(sctp->sctp_ip6h, rth,
1027 			    sctps->sctps_netstack);
1028 		}
1029 	}
1030 	return (0);
1031 }
1032 
1033 /*
1034  * Initialize the IPv6 header. Loses any record of any IPv6 extension headers.
1035  */
1036 int
1037 sctp_header_init_ipv6(sctp_t *sctp, int sleep)
1038 {
1039 	sctp_hdr_t	*sctph;
1040 	sctp_stack_t	*sctps = sctp->sctp_sctps;
1041 
1042 	/*
1043 	 * This is a simple initialization. If there's
1044 	 * already a template, it should never be too small,
1045 	 * so reuse it. Otherwise, allocate space for the new one.
1046 	 * Ensure that there is enough space to "downgrade" the sctp_t
1047 	 * to an IPv4 sctp_t. This requires having space for a full load
1048 	 * of IPv4 options
1049 	 */
1050 	if (sctp->sctp_iphc6 != NULL) {
1051 		ASSERT(sctp->sctp_iphc6_len >=
1052 		    SCTP_MAX_COMBINED_HEADER_LENGTH);
1053 		bzero(sctp->sctp_iphc6, sctp->sctp_iphc6_len);
1054 	} else {
1055 		sctp->sctp_iphc6_len = SCTP_MAX_COMBINED_HEADER_LENGTH;
1056 		sctp->sctp_iphc6 = kmem_zalloc(sctp->sctp_iphc_len, sleep);
1057 		if (sctp->sctp_iphc6 == NULL) {
1058 			sctp->sctp_iphc6_len = 0;
1059 			return (ENOMEM);
1060 		}
1061 	}
1062 	sctp->sctp_hdr6_len = IPV6_HDR_LEN + sizeof (sctp_hdr_t);
1063 	sctp->sctp_ip_hdr6_len = IPV6_HDR_LEN;
1064 	sctp->sctp_ip6h = (ip6_t *)sctp->sctp_iphc6;
1065 
1066 	/* Initialize the header template */
1067 
1068 	sctp->sctp_ip6h->ip6_vcf = IPV6_DEFAULT_VERS_AND_FLOW;
1069 	sctp->sctp_ip6h->ip6_plen = ntohs(sizeof (sctp_hdr_t));
1070 	sctp->sctp_ip6h->ip6_nxt = IPPROTO_SCTP;
1071 	sctp->sctp_ip6h->ip6_hops = sctps->sctps_ipv6_hoplimit;
1072 
1073 	sctph = (sctp_hdr_t *)(sctp->sctp_iphc6 + IPV6_HDR_LEN);
1074 	sctp->sctp_sctph6 = sctph;
1075 
1076 	return (0);
1077 }
1078 
1079 static int
1080 sctp_v4_label(sctp_t *sctp)
1081 {
1082 	uchar_t optbuf[IP_MAX_OPT_LENGTH];
1083 	const cred_t *cr = CONN_CRED(sctp->sctp_connp);
1084 	int added;
1085 
1086 	if (tsol_compute_label(cr, sctp->sctp_ipha->ipha_dst, optbuf,
1087 	    sctp->sctp_mac_exempt,
1088 	    sctp->sctp_sctps->sctps_netstack->netstack_ip) != 0)
1089 		return (EACCES);
1090 
1091 	added = tsol_remove_secopt(sctp->sctp_ipha, sctp->sctp_hdr_len);
1092 	if (added == -1)
1093 		return (EACCES);
1094 	sctp->sctp_hdr_len += added;
1095 	sctp->sctp_sctph = (sctp_hdr_t *)((uchar_t *)sctp->sctp_sctph + added);
1096 	sctp->sctp_ip_hdr_len += added;
1097 	if ((sctp->sctp_v4label_len = optbuf[IPOPT_OLEN]) != 0) {
1098 		sctp->sctp_v4label_len = (sctp->sctp_v4label_len + 3) & ~3;
1099 		added = tsol_prepend_option(optbuf, sctp->sctp_ipha,
1100 		    sctp->sctp_hdr_len);
1101 		if (added == -1)
1102 			return (EACCES);
1103 		sctp->sctp_hdr_len += added;
1104 		sctp->sctp_sctph = (sctp_hdr_t *)((uchar_t *)sctp->sctp_sctph +
1105 		    added);
1106 		sctp->sctp_ip_hdr_len += added;
1107 	}
1108 	return (0);
1109 }
1110 
1111 static int
1112 sctp_v6_label(sctp_t *sctp)
1113 {
1114 	uchar_t optbuf[TSOL_MAX_IPV6_OPTION];
1115 	const cred_t *cr = CONN_CRED(sctp->sctp_connp);
1116 
1117 	if (tsol_compute_label_v6(cr, &sctp->sctp_ip6h->ip6_dst, optbuf,
1118 	    sctp->sctp_mac_exempt,
1119 	    sctp->sctp_sctps->sctps_netstack->netstack_ip) != 0)
1120 		return (EACCES);
1121 	if (tsol_update_sticky(&sctp->sctp_sticky_ipp, &sctp->sctp_v6label_len,
1122 	    optbuf) != 0)
1123 		return (EACCES);
1124 	if (sctp_build_hdrs(sctp) != 0)
1125 		return (EACCES);
1126 	return (0);
1127 }
1128 
1129 /*
1130  * XXX implement more sophisticated logic
1131  */
1132 int
1133 sctp_set_hdraddrs(sctp_t *sctp)
1134 {
1135 	sctp_faddr_t *fp;
1136 	int gotv4 = 0;
1137 	int gotv6 = 0;
1138 
1139 	ASSERT(sctp->sctp_faddrs != NULL);
1140 	ASSERT(sctp->sctp_nsaddrs > 0);
1141 
1142 	/* Set up using the primary first */
1143 	if (IN6_IS_ADDR_V4MAPPED(&sctp->sctp_primary->faddr)) {
1144 		IN6_V4MAPPED_TO_IPADDR(&sctp->sctp_primary->faddr,
1145 		    sctp->sctp_ipha->ipha_dst);
1146 		/* saddr may be unspec; make_mp() will handle this */
1147 		IN6_V4MAPPED_TO_IPADDR(&sctp->sctp_primary->saddr,
1148 		    sctp->sctp_ipha->ipha_src);
1149 		if (!is_system_labeled() || sctp_v4_label(sctp) == 0) {
1150 			gotv4 = 1;
1151 			if (sctp->sctp_ipversion == IPV4_VERSION) {
1152 				goto copyports;
1153 			}
1154 		}
1155 	} else {
1156 		sctp->sctp_ip6h->ip6_dst = sctp->sctp_primary->faddr;
1157 		/* saddr may be unspec; make_mp() will handle this */
1158 		sctp->sctp_ip6h->ip6_src = sctp->sctp_primary->saddr;
1159 		if (!is_system_labeled() || sctp_v6_label(sctp) == 0)
1160 			gotv6 = 1;
1161 	}
1162 
1163 	for (fp = sctp->sctp_faddrs; fp; fp = fp->next) {
1164 		if (!gotv4 && IN6_IS_ADDR_V4MAPPED(&fp->faddr)) {
1165 			IN6_V4MAPPED_TO_IPADDR(&fp->faddr,
1166 			    sctp->sctp_ipha->ipha_dst);
1167 			/* copy in the faddr_t's saddr */
1168 			IN6_V4MAPPED_TO_IPADDR(&fp->saddr,
1169 			    sctp->sctp_ipha->ipha_src);
1170 			if (!is_system_labeled() || sctp_v4_label(sctp) == 0) {
1171 				gotv4 = 1;
1172 				if (sctp->sctp_ipversion == IPV4_VERSION ||
1173 				    gotv6) {
1174 					break;
1175 				}
1176 			}
1177 		} else if (!gotv6 && !IN6_IS_ADDR_V4MAPPED(&fp->faddr)) {
1178 			sctp->sctp_ip6h->ip6_dst = fp->faddr;
1179 			/* copy in the faddr_t's saddr */
1180 			sctp->sctp_ip6h->ip6_src = fp->saddr;
1181 			if (!is_system_labeled() || sctp_v6_label(sctp) == 0) {
1182 				gotv6 = 1;
1183 				if (gotv4)
1184 					break;
1185 			}
1186 		}
1187 	}
1188 
1189 copyports:
1190 	if (!gotv4 && !gotv6)
1191 		return (EACCES);
1192 
1193 	/* copy in the ports for good measure */
1194 	sctp->sctp_sctph->sh_sport = sctp->sctp_lport;
1195 	sctp->sctp_sctph->sh_dport = sctp->sctp_fport;
1196 
1197 	sctp->sctp_sctph6->sh_sport = sctp->sctp_lport;
1198 	sctp->sctp_sctph6->sh_dport = sctp->sctp_fport;
1199 	return (0);
1200 }
1201 
1202 /*
1203  * got_errchunk is set B_TRUE only if called from validate_init_params(), when
1204  * an ERROR chunk is already prepended the size of which needs updating for
1205  * additional unrecognized parameters. Other callers either prepend the ERROR
1206  * chunk with the correct size after calling this function, or they are calling
1207  * to add an invalid parameter to an INIT_ACK chunk, in that case no ERROR chunk
1208  * exists, the CAUSE blocks go into the INIT_ACK directly.
1209  *
1210  * *errmp will be non-NULL both when adding an additional CAUSE block to an
1211  * existing prepended COOKIE ERROR chunk (processing params of an INIT_ACK),
1212  * and when adding unrecognized parameters after the first, to an INIT_ACK
1213  * (processing params of an INIT chunk).
1214  */
1215 void
1216 sctp_add_unrec_parm(sctp_parm_hdr_t *uph, mblk_t **errmp,
1217     boolean_t got_errchunk)
1218 {
1219 	mblk_t *mp;
1220 	sctp_parm_hdr_t *ph;
1221 	size_t len;
1222 	int pad;
1223 	sctp_chunk_hdr_t *ecp;
1224 
1225 	len = sizeof (*ph) + ntohs(uph->sph_len);
1226 	if ((pad = len % SCTP_ALIGN) != 0) {
1227 		pad = SCTP_ALIGN - pad;
1228 		len += pad;
1229 	}
1230 	mp = allocb(len, BPRI_MED);
1231 	if (mp == NULL) {
1232 		return;
1233 	}
1234 
1235 	ph = (sctp_parm_hdr_t *)(mp->b_rptr);
1236 	ph->sph_type = htons(PARM_UNRECOGNIZED);
1237 	ph->sph_len = htons(len - pad);
1238 
1239 	/* copy in the unrecognized parameter */
1240 	bcopy(uph, ph + 1, ntohs(uph->sph_len));
1241 
1242 	if (pad != 0)
1243 		bzero((mp->b_rptr + len - pad), pad);
1244 
1245 	mp->b_wptr = mp->b_rptr + len;
1246 	if (*errmp != NULL) {
1247 		/*
1248 		 * Update total length if an ERROR chunk, then link
1249 		 * this CAUSE block to the possible chain of CAUSE
1250 		 * blocks attached to the ERROR chunk or INIT_ACK
1251 		 * being created.
1252 		 */
1253 		if (got_errchunk) {
1254 			/* ERROR chunk already prepended */
1255 			ecp = (sctp_chunk_hdr_t *)((*errmp)->b_rptr);
1256 			ecp->sch_len = htons(ntohs(ecp->sch_len) + len);
1257 		}
1258 		linkb(*errmp, mp);
1259 	} else {
1260 		*errmp = mp;
1261 	}
1262 }
1263 
1264 /*
1265  * o Bounds checking
1266  * o Updates remaining
1267  * o Checks alignment
1268  */
1269 sctp_parm_hdr_t *
1270 sctp_next_parm(sctp_parm_hdr_t *current, ssize_t *remaining)
1271 {
1272 	int pad;
1273 	uint16_t len;
1274 
1275 	len = ntohs(current->sph_len);
1276 	*remaining -= len;
1277 	if (*remaining < sizeof (*current) || len < sizeof (*current)) {
1278 		return (NULL);
1279 	}
1280 	if ((pad = len & (SCTP_ALIGN - 1)) != 0) {
1281 		pad = SCTP_ALIGN - pad;
1282 		*remaining -= pad;
1283 	}
1284 	/*LINTED pointer cast may result in improper alignment*/
1285 	current = (sctp_parm_hdr_t *)((char *)current + len + pad);
1286 	return (current);
1287 }
1288 
1289 /*
1290  * Sets the address parameters given in the INIT chunk into sctp's
1291  * faddrs; if psctp is non-NULL, copies psctp's saddrs. If there are
1292  * no address parameters in the INIT chunk, a single faddr is created
1293  * from the ip hdr at the beginning of pkt.
1294  * If there already are existing addresses hanging from sctp, merge
1295  * them in, if the old info contains addresses which are not present
1296  * in this new info, get rid of them, and clean the pointers if there's
1297  * messages which have this as their target address.
1298  *
1299  * We also re-adjust the source address list here since the list may
1300  * contain more than what is actually part of the association. If
1301  * we get here from sctp_send_cookie_echo(), we are on the active
1302  * side and psctp will be NULL and ich will be the INIT-ACK chunk.
1303  * If we get here from sctp_accept_comm(), ich will be the INIT chunk
1304  * and psctp will the listening endpoint.
1305  *
1306  * INIT processing: When processing the INIT we inherit the src address
1307  * list from the listener. For a loopback or linklocal association, we
1308  * delete the list and just take the address from the IP header (since
1309  * that's how we created the INIT-ACK). Additionally, for loopback we
1310  * ignore the address params in the INIT. For determining which address
1311  * types were sent in the INIT-ACK we follow the same logic as in
1312  * creating the INIT-ACK. We delete addresses of the type that are not
1313  * supported by the peer.
1314  *
1315  * INIT-ACK processing: When processing the INIT-ACK since we had not
1316  * included addr params for loopback or linklocal addresses when creating
1317  * the INIT, we just use the address from the IP header. Further, for
1318  * loopback we ignore the addr param list. We mark addresses of the
1319  * type not supported by the peer as unconfirmed.
1320  *
1321  * In case of INIT processing we look for supported address types in the
1322  * supported address param, if present. In both cases the address type in
1323  * the IP header is supported as well as types for addresses in the param
1324  * list, if any.
1325  *
1326  * Once we have the supported address types sctp_check_saddr() runs through
1327  * the source address list and deletes or marks as unconfirmed address of
1328  * types not supported by the peer.
1329  *
1330  * Returns 0 on success, sys errno on failure
1331  */
1332 int
1333 sctp_get_addrparams(sctp_t *sctp, sctp_t *psctp, mblk_t *pkt,
1334     sctp_chunk_hdr_t *ich, uint_t *sctp_options)
1335 {
1336 	sctp_init_chunk_t	*init;
1337 	ipha_t			*iph;
1338 	ip6_t			*ip6h;
1339 	in6_addr_t		hdrsaddr[1];
1340 	in6_addr_t		hdrdaddr[1];
1341 	sctp_parm_hdr_t		*ph;
1342 	ssize_t			remaining;
1343 	int			isv4;
1344 	int			err;
1345 	sctp_faddr_t		*fp;
1346 	int			supp_af = 0;
1347 	boolean_t		check_saddr = B_TRUE;
1348 	in6_addr_t		curaddr;
1349 	sctp_stack_t		*sctps = sctp->sctp_sctps;
1350 
1351 	if (sctp_options != NULL)
1352 		*sctp_options = 0;
1353 
1354 	/* extract the address from the IP header */
1355 	isv4 = (IPH_HDR_VERSION(pkt->b_rptr) == IPV4_VERSION);
1356 	if (isv4) {
1357 		iph = (ipha_t *)pkt->b_rptr;
1358 		IN6_IPADDR_TO_V4MAPPED(iph->ipha_src, hdrsaddr);
1359 		IN6_IPADDR_TO_V4MAPPED(iph->ipha_dst, hdrdaddr);
1360 		supp_af |= PARM_SUPP_V4;
1361 	} else {
1362 		ip6h = (ip6_t *)pkt->b_rptr;
1363 		hdrsaddr[0] = ip6h->ip6_src;
1364 		hdrdaddr[0] = ip6h->ip6_dst;
1365 		supp_af |= PARM_SUPP_V6;
1366 	}
1367 
1368 	/*
1369 	 * Unfortunately, we can't delay this because adding an faddr
1370 	 * looks for the presence of the source address (from the ire
1371 	 * for the faddr) in the source address list. We could have
1372 	 * delayed this if, say, this was a loopback/linklocal connection.
1373 	 * Now, we just end up nuking this list and taking the addr from
1374 	 * the IP header for loopback/linklocal.
1375 	 */
1376 	if (psctp != NULL && psctp->sctp_nsaddrs > 0) {
1377 		ASSERT(sctp->sctp_nsaddrs == 0);
1378 
1379 		err = sctp_dup_saddrs(psctp, sctp, KM_NOSLEEP);
1380 		if (err != 0)
1381 			return (err);
1382 	}
1383 	/*
1384 	 * We will add the faddr before parsing the address list as this
1385 	 * might be a loopback connection and we would not have to
1386 	 * go through the list.
1387 	 *
1388 	 * Make sure the header's addr is in the list
1389 	 */
1390 	fp = sctp_lookup_faddr(sctp, hdrsaddr);
1391 	if (fp == NULL) {
1392 		/* not included; add it now */
1393 		err = sctp_add_faddr(sctp, hdrsaddr, KM_NOSLEEP, B_TRUE);
1394 		if (err != 0)
1395 			return (err);
1396 
1397 		/* sctp_faddrs will be the hdr addr */
1398 		fp = sctp->sctp_faddrs;
1399 	}
1400 	/* make the header addr the primary */
1401 
1402 	if (cl_sctp_assoc_change != NULL && psctp == NULL)
1403 		curaddr = sctp->sctp_current->faddr;
1404 
1405 	sctp->sctp_primary = fp;
1406 	sctp->sctp_current = fp;
1407 	sctp->sctp_mss = fp->sfa_pmss;
1408 
1409 	/* For loopback connections & linklocal get address from the header */
1410 	if (sctp->sctp_loopback || sctp->sctp_linklocal) {
1411 		if (sctp->sctp_nsaddrs != 0)
1412 			sctp_free_saddrs(sctp);
1413 		if ((err = sctp_saddr_add_addr(sctp, hdrdaddr, 0)) != 0)
1414 			return (err);
1415 		/* For loopback ignore address list */
1416 		if (sctp->sctp_loopback)
1417 			return (0);
1418 		check_saddr = B_FALSE;
1419 	}
1420 
1421 	/* Walk the params in the INIT [ACK], pulling out addr params */
1422 	remaining = ntohs(ich->sch_len) - sizeof (*ich) -
1423 	    sizeof (sctp_init_chunk_t);
1424 	if (remaining < sizeof (*ph)) {
1425 		if (check_saddr) {
1426 			sctp_check_saddr(sctp, supp_af, psctp == NULL ?
1427 			    B_FALSE : B_TRUE, hdrdaddr);
1428 		}
1429 		ASSERT(sctp_saddr_lookup(sctp, hdrdaddr, 0) != NULL);
1430 		return (0);
1431 	}
1432 
1433 	init = (sctp_init_chunk_t *)(ich + 1);
1434 	ph = (sctp_parm_hdr_t *)(init + 1);
1435 
1436 	/* params will have already been byteordered when validating */
1437 	while (ph != NULL) {
1438 		if (ph->sph_type == htons(PARM_SUPP_ADDRS)) {
1439 			int		plen;
1440 			uint16_t	*p;
1441 			uint16_t	addrtype;
1442 
1443 			ASSERT(psctp != NULL);
1444 			plen = ntohs(ph->sph_len);
1445 			p = (uint16_t *)(ph + 1);
1446 			while (plen > 0) {
1447 				addrtype = ntohs(*p);
1448 				switch (addrtype) {
1449 					case PARM_ADDR6:
1450 						supp_af |= PARM_SUPP_V6;
1451 						break;
1452 					case PARM_ADDR4:
1453 						supp_af |= PARM_SUPP_V4;
1454 						break;
1455 					default:
1456 						break;
1457 				}
1458 				p++;
1459 				plen -= sizeof (*p);
1460 			}
1461 		} else if (ph->sph_type == htons(PARM_ADDR4)) {
1462 			if (remaining >= PARM_ADDR4_LEN) {
1463 				in6_addr_t addr;
1464 				ipaddr_t ta;
1465 
1466 				supp_af |= PARM_SUPP_V4;
1467 				/*
1468 				 * Screen out broad/multicasts & loopback.
1469 				 * If the endpoint only accepts v6 address,
1470 				 * go to the next one.
1471 				 *
1472 				 * Subnet broadcast check is done in
1473 				 * sctp_add_faddr().  If the address is
1474 				 * a broadcast address, it won't be added.
1475 				 */
1476 				bcopy(ph + 1, &ta, sizeof (ta));
1477 				if (ta == 0 ||
1478 				    ta == INADDR_BROADCAST ||
1479 				    ta == htonl(INADDR_LOOPBACK) ||
1480 				    CLASSD(ta) ||
1481 				    sctp->sctp_connp->conn_ipv6_v6only) {
1482 					goto next;
1483 				}
1484 				IN6_INADDR_TO_V4MAPPED((struct in_addr *)
1485 				    (ph + 1), &addr);
1486 
1487 				/* Check for duplicate. */
1488 				if (sctp_lookup_faddr(sctp, &addr) != NULL)
1489 					goto next;
1490 
1491 				/* OK, add it to the faddr set */
1492 				err = sctp_add_faddr(sctp, &addr, KM_NOSLEEP,
1493 				    B_FALSE);
1494 				/* Something is wrong...  Try the next one. */
1495 				if (err != 0)
1496 					goto next;
1497 			}
1498 		} else if (ph->sph_type == htons(PARM_ADDR6) &&
1499 		    sctp->sctp_family == AF_INET6) {
1500 			/* An v4 socket should not take v6 addresses. */
1501 			if (remaining >= PARM_ADDR6_LEN) {
1502 				in6_addr_t *addr6;
1503 
1504 				supp_af |= PARM_SUPP_V6;
1505 				addr6 = (in6_addr_t *)(ph + 1);
1506 				/*
1507 				 * Screen out link locals, mcast, loopback
1508 				 * and bogus v6 address.
1509 				 */
1510 				if (IN6_IS_ADDR_LINKLOCAL(addr6) ||
1511 				    IN6_IS_ADDR_MULTICAST(addr6) ||
1512 				    IN6_IS_ADDR_LOOPBACK(addr6) ||
1513 				    IN6_IS_ADDR_V4MAPPED(addr6)) {
1514 					goto next;
1515 				}
1516 				/* Check for duplicate. */
1517 				if (sctp_lookup_faddr(sctp, addr6) != NULL)
1518 					goto next;
1519 
1520 				err = sctp_add_faddr(sctp,
1521 				    (in6_addr_t *)(ph + 1), KM_NOSLEEP,
1522 				    B_FALSE);
1523 				/* Something is wrong...  Try the next one. */
1524 				if (err != 0)
1525 					goto next;
1526 			}
1527 		} else if (ph->sph_type == htons(PARM_FORWARD_TSN)) {
1528 			if (sctp_options != NULL)
1529 				*sctp_options |= SCTP_PRSCTP_OPTION;
1530 		} /* else; skip */
1531 
1532 next:
1533 		ph = sctp_next_parm(ph, &remaining);
1534 	}
1535 	if (check_saddr) {
1536 		sctp_check_saddr(sctp, supp_af, psctp == NULL ? B_FALSE :
1537 		    B_TRUE, hdrdaddr);
1538 	}
1539 	ASSERT(sctp_saddr_lookup(sctp, hdrdaddr, 0) != NULL);
1540 	/*
1541 	 * We have the right address list now, update clustering's
1542 	 * knowledge because when we sent the INIT we had just added
1543 	 * the address the INIT was sent to.
1544 	 */
1545 	if (psctp == NULL && cl_sctp_assoc_change != NULL) {
1546 		uchar_t	*alist;
1547 		size_t	asize;
1548 		uchar_t	*dlist;
1549 		size_t	dsize;
1550 
1551 		asize = sizeof (in6_addr_t) * sctp->sctp_nfaddrs;
1552 		alist = kmem_alloc(asize, KM_NOSLEEP);
1553 		if (alist == NULL) {
1554 			SCTP_KSTAT(sctps, sctp_cl_assoc_change);
1555 			return (ENOMEM);
1556 		}
1557 		/*
1558 		 * Just include the address the INIT was sent to in the
1559 		 * delete list and send the entire faddr list. We could
1560 		 * do it differently (i.e include all the addresses in the
1561 		 * add list even if it contains the original address OR
1562 		 * remove the original address from the add list etc.), but
1563 		 * this seems reasonable enough.
1564 		 */
1565 		dsize = sizeof (in6_addr_t);
1566 		dlist = kmem_alloc(dsize, KM_NOSLEEP);
1567 		if (dlist == NULL) {
1568 			kmem_free(alist, asize);
1569 			SCTP_KSTAT(sctps, sctp_cl_assoc_change);
1570 			return (ENOMEM);
1571 		}
1572 		bcopy(&curaddr, dlist, sizeof (curaddr));
1573 		sctp_get_faddr_list(sctp, alist, asize);
1574 		(*cl_sctp_assoc_change)(sctp->sctp_family, alist, asize,
1575 		    sctp->sctp_nfaddrs, dlist, dsize, 1, SCTP_CL_PADDR,
1576 		    (cl_sctp_handle_t)sctp);
1577 		/* alist and dlist will be freed by the clustering module */
1578 	}
1579 	return (0);
1580 }
1581 
1582 /*
1583  * Returns 0 if the check failed and the restart should be refused,
1584  * 1 if the check succeeded.
1585  */
1586 int
1587 sctp_secure_restart_check(mblk_t *pkt, sctp_chunk_hdr_t *ich, uint32_t ports,
1588     int sleep, sctp_stack_t *sctps)
1589 {
1590 	sctp_faddr_t *fp, *fphead = NULL;
1591 	sctp_parm_hdr_t *ph;
1592 	ssize_t remaining;
1593 	int isv4;
1594 	ipha_t *iph;
1595 	ip6_t *ip6h;
1596 	in6_addr_t hdraddr[1];
1597 	int retval = 0;
1598 	sctp_tf_t *tf;
1599 	sctp_t *sctp;
1600 	int compres;
1601 	sctp_init_chunk_t *init;
1602 	int nadded = 0;
1603 
1604 	/* extract the address from the IP header */
1605 	isv4 = (IPH_HDR_VERSION(pkt->b_rptr) == IPV4_VERSION);
1606 	if (isv4) {
1607 		iph = (ipha_t *)pkt->b_rptr;
1608 		IN6_IPADDR_TO_V4MAPPED(iph->ipha_src, hdraddr);
1609 	} else {
1610 		ip6h = (ip6_t *)pkt->b_rptr;
1611 		hdraddr[0] = ip6h->ip6_src;
1612 	}
1613 
1614 	/* Walk the params in the INIT [ACK], pulling out addr params */
1615 	remaining = ntohs(ich->sch_len) - sizeof (*ich) -
1616 	    sizeof (sctp_init_chunk_t);
1617 	if (remaining < sizeof (*ph)) {
1618 		/* no parameters; restart OK */
1619 		return (1);
1620 	}
1621 	init = (sctp_init_chunk_t *)(ich + 1);
1622 	ph = (sctp_parm_hdr_t *)(init + 1);
1623 
1624 	while (ph != NULL) {
1625 		sctp_faddr_t *fpa = NULL;
1626 
1627 		/* params will have already been byteordered when validating */
1628 		if (ph->sph_type == htons(PARM_ADDR4)) {
1629 			if (remaining >= PARM_ADDR4_LEN) {
1630 				in6_addr_t addr;
1631 				IN6_INADDR_TO_V4MAPPED((struct in_addr *)
1632 				    (ph + 1), &addr);
1633 				fpa = kmem_cache_alloc(sctp_kmem_faddr_cache,
1634 				    sleep);
1635 				if (fpa == NULL) {
1636 					goto done;
1637 				}
1638 				bzero(fpa, sizeof (*fpa));
1639 				fpa->faddr = addr;
1640 				fpa->next = NULL;
1641 			}
1642 		} else if (ph->sph_type == htons(PARM_ADDR6)) {
1643 			if (remaining >= PARM_ADDR6_LEN) {
1644 				fpa = kmem_cache_alloc(sctp_kmem_faddr_cache,
1645 				    sleep);
1646 				if (fpa == NULL) {
1647 					goto done;
1648 				}
1649 				bzero(fpa, sizeof (*fpa));
1650 				bcopy(ph + 1, &fpa->faddr,
1651 				    sizeof (fpa->faddr));
1652 				fpa->next = NULL;
1653 			}
1654 		}
1655 		/* link in the new addr, if it was an addr param */
1656 		if (fpa != NULL) {
1657 			if (fphead == NULL) {
1658 				fphead = fpa;
1659 			} else {
1660 				fpa->next = fphead;
1661 				fphead = fpa;
1662 			}
1663 		}
1664 
1665 		ph = sctp_next_parm(ph, &remaining);
1666 	}
1667 
1668 	if (fphead == NULL) {
1669 		/* no addr parameters; restart OK */
1670 		return (1);
1671 	}
1672 
1673 	/*
1674 	 * got at least one; make sure the header's addr is
1675 	 * in the list
1676 	 */
1677 	fp = sctp_lookup_faddr_nosctp(fphead, hdraddr);
1678 	if (fp == NULL) {
1679 		/* not included; add it now */
1680 		fp = kmem_cache_alloc(sctp_kmem_faddr_cache, sleep);
1681 		if (fp == NULL) {
1682 			goto done;
1683 		}
1684 		bzero(fp, sizeof (*fp));
1685 		fp->faddr = *hdraddr;
1686 		fp->next = fphead;
1687 		fphead = fp;
1688 	}
1689 
1690 	/*
1691 	 * Now, we can finally do the check: For each sctp instance
1692 	 * on the hash line for ports, compare its faddr set against
1693 	 * the new one. If the new one is a strict subset of any
1694 	 * existing sctp's faddrs, the restart is OK. However, if there
1695 	 * is an overlap, this could be an attack, so return failure.
1696 	 * If all sctp's faddrs are disjoint, this is a legitimate new
1697 	 * association.
1698 	 */
1699 	tf = &(sctps->sctps_conn_fanout[SCTP_CONN_HASH(sctps, ports)]);
1700 	mutex_enter(&tf->tf_lock);
1701 
1702 	for (sctp = tf->tf_sctp; sctp; sctp = sctp->sctp_conn_hash_next) {
1703 		if (ports != sctp->sctp_ports) {
1704 			continue;
1705 		}
1706 		compres = sctp_compare_faddrsets(fphead, sctp->sctp_faddrs);
1707 		if (compres <= SCTP_ADDR_SUBSET) {
1708 			retval = 1;
1709 			mutex_exit(&tf->tf_lock);
1710 			goto done;
1711 		}
1712 		if (compres == SCTP_ADDR_OVERLAP) {
1713 			dprint(1,
1714 			    ("new assoc from %x:%x:%x:%x overlaps with %p\n",
1715 			    SCTP_PRINTADDR(*hdraddr), (void *)sctp));
1716 			/*
1717 			 * While we still hold the lock, we need to
1718 			 * figure out which addresses have been
1719 			 * added so we can include them in the abort
1720 			 * we will send back. Since these faddrs will
1721 			 * never be used, we overload the rto field
1722 			 * here, setting it to 0 if the address was
1723 			 * not added, 1 if it was added.
1724 			 */
1725 			for (fp = fphead; fp; fp = fp->next) {
1726 				if (sctp_lookup_faddr(sctp, &fp->faddr)) {
1727 					fp->rto = 0;
1728 				} else {
1729 					fp->rto = 1;
1730 					nadded++;
1731 				}
1732 			}
1733 			mutex_exit(&tf->tf_lock);
1734 			goto done;
1735 		}
1736 	}
1737 	mutex_exit(&tf->tf_lock);
1738 
1739 	/* All faddrs are disjoint; legit new association */
1740 	retval = 1;
1741 
1742 done:
1743 	/* If are attempted adds, send back an abort listing the addrs */
1744 	if (nadded > 0) {
1745 		void *dtail;
1746 		size_t dlen;
1747 
1748 		dtail = kmem_alloc(PARM_ADDR6_LEN * nadded, KM_NOSLEEP);
1749 		if (dtail == NULL) {
1750 			goto cleanup;
1751 		}
1752 
1753 		ph = dtail;
1754 		dlen = 0;
1755 		for (fp = fphead; fp; fp = fp->next) {
1756 			if (fp->rto == 0) {
1757 				continue;
1758 			}
1759 			if (IN6_IS_ADDR_V4MAPPED(&fp->faddr)) {
1760 				ipaddr_t addr4;
1761 
1762 				ph->sph_type = htons(PARM_ADDR4);
1763 				ph->sph_len = htons(PARM_ADDR4_LEN);
1764 				IN6_V4MAPPED_TO_IPADDR(&fp->faddr, addr4);
1765 				ph++;
1766 				bcopy(&addr4, ph, sizeof (addr4));
1767 				ph = (sctp_parm_hdr_t *)
1768 				    ((char *)ph + sizeof (addr4));
1769 				dlen += PARM_ADDR4_LEN;
1770 			} else {
1771 				ph->sph_type = htons(PARM_ADDR6);
1772 				ph->sph_len = htons(PARM_ADDR6_LEN);
1773 				ph++;
1774 				bcopy(&fp->faddr, ph, sizeof (fp->faddr));
1775 				ph = (sctp_parm_hdr_t *)
1776 				    ((char *)ph + sizeof (fp->faddr));
1777 				dlen += PARM_ADDR6_LEN;
1778 			}
1779 		}
1780 
1781 		/* Send off the abort */
1782 		sctp_send_abort(sctp, sctp_init2vtag(ich),
1783 		    SCTP_ERR_RESTART_NEW_ADDRS, dtail, dlen, pkt, 0, B_TRUE);
1784 
1785 		kmem_free(dtail, PARM_ADDR6_LEN * nadded);
1786 	}
1787 
1788 cleanup:
1789 	/* Clean up */
1790 	if (fphead) {
1791 		sctp_faddr_t *fpn;
1792 		for (fp = fphead; fp; fp = fpn) {
1793 			fpn = fp->next;
1794 			kmem_cache_free(sctp_kmem_faddr_cache, fp);
1795 		}
1796 	}
1797 
1798 	return (retval);
1799 }
1800 
1801 /*
1802  * Reset any state related to transmitted chunks.
1803  */
1804 void
1805 sctp_congest_reset(sctp_t *sctp)
1806 {
1807 	sctp_faddr_t	*fp;
1808 	sctp_stack_t	*sctps = sctp->sctp_sctps;
1809 	mblk_t		*mp;
1810 
1811 	for (fp = sctp->sctp_faddrs; fp != NULL; fp = fp->next) {
1812 		fp->ssthresh = sctps->sctps_initial_mtu;
1813 		SET_CWND(fp, fp->sfa_pmss, sctps->sctps_slow_start_initial);
1814 		fp->suna = 0;
1815 		fp->pba = 0;
1816 	}
1817 	/*
1818 	 * Clean up the transmit list as well since we have reset accounting
1819 	 * on all the fps. Send event upstream, if required.
1820 	 */
1821 	while ((mp = sctp->sctp_xmit_head) != NULL) {
1822 		sctp->sctp_xmit_head = mp->b_next;
1823 		mp->b_next = NULL;
1824 		if (sctp->sctp_xmit_head != NULL)
1825 			sctp->sctp_xmit_head->b_prev = NULL;
1826 		sctp_sendfail_event(sctp, mp, 0, B_TRUE);
1827 	}
1828 	sctp->sctp_xmit_head = NULL;
1829 	sctp->sctp_xmit_tail = NULL;
1830 	sctp->sctp_xmit_unacked = NULL;
1831 
1832 	sctp->sctp_unacked = 0;
1833 	/*
1834 	 * Any control message as well. We will clean-up this list as well.
1835 	 * This contains any pending ASCONF request that we have queued/sent.
1836 	 * If we do get an ACK we will just drop it. However, given that
1837 	 * we are restarting chances are we aren't going to get any.
1838 	 */
1839 	if (sctp->sctp_cxmit_list != NULL)
1840 		sctp_asconf_free_cxmit(sctp, NULL);
1841 	sctp->sctp_cxmit_list = NULL;
1842 	sctp->sctp_cchunk_pend = 0;
1843 
1844 	sctp->sctp_rexmitting = B_FALSE;
1845 	sctp->sctp_rxt_nxttsn = 0;
1846 	sctp->sctp_rxt_maxtsn = 0;
1847 
1848 	sctp->sctp_zero_win_probe = B_FALSE;
1849 }
1850 
1851 static void
1852 sctp_init_faddr(sctp_t *sctp, sctp_faddr_t *fp, in6_addr_t *addr,
1853     mblk_t *timer_mp)
1854 {
1855 	sctp_stack_t	*sctps = sctp->sctp_sctps;
1856 
1857 	bcopy(addr, &fp->faddr, sizeof (*addr));
1858 	if (IN6_IS_ADDR_V4MAPPED(addr)) {
1859 		fp->isv4 = 1;
1860 		/* Make sure that sfa_pmss is a multiple of SCTP_ALIGN. */
1861 		fp->sfa_pmss =
1862 		    (sctps->sctps_initial_mtu - sctp->sctp_hdr_len) &
1863 		    ~(SCTP_ALIGN - 1);
1864 	} else {
1865 		fp->isv4 = 0;
1866 		fp->sfa_pmss =
1867 		    (sctps->sctps_initial_mtu - sctp->sctp_hdr6_len) &
1868 		    ~(SCTP_ALIGN - 1);
1869 	}
1870 	fp->cwnd = sctps->sctps_slow_start_initial * fp->sfa_pmss;
1871 	fp->rto = MIN(sctp->sctp_rto_initial, sctp->sctp_init_rto_max);
1872 	fp->srtt = -1;
1873 	fp->rtt_updates = 0;
1874 	fp->strikes = 0;
1875 	fp->max_retr = sctp->sctp_pp_max_rxt;
1876 	/* Mark it as not confirmed. */
1877 	fp->state = SCTP_FADDRS_UNCONFIRMED;
1878 	fp->hb_interval = sctp->sctp_hb_interval;
1879 	fp->ssthresh = sctps->sctps_initial_ssthresh;
1880 	fp->suna = 0;
1881 	fp->pba = 0;
1882 	fp->acked = 0;
1883 	fp->lastactive = lbolt64;
1884 	fp->timer_mp = timer_mp;
1885 	fp->hb_pending = B_FALSE;
1886 	fp->hb_enabled = B_TRUE;
1887 	fp->df = 1;
1888 	fp->pmtu_discovered = 0;
1889 	fp->next = NULL;
1890 	fp->ire = NULL;
1891 	fp->T3expire = 0;
1892 	(void) random_get_pseudo_bytes((uint8_t *)&fp->hb_secret,
1893 	    sizeof (fp->hb_secret));
1894 	fp->hb_expiry = lbolt64;
1895 	fp->rxt_unacked = 0;
1896 
1897 	sctp_get_ire(sctp, fp);
1898 }
1899 
1900 /*ARGSUSED*/
1901 static int
1902 faddr_constructor(void *buf, void *arg, int flags)
1903 {
1904 	sctp_faddr_t *fp = buf;
1905 
1906 	fp->timer_mp = NULL;
1907 	fp->timer_running = 0;
1908 
1909 	fp->rc_timer_mp = NULL;
1910 	fp->rc_timer_running = 0;
1911 
1912 	return (0);
1913 }
1914 
1915 /*ARGSUSED*/
1916 static void
1917 faddr_destructor(void *buf, void *arg)
1918 {
1919 	sctp_faddr_t *fp = buf;
1920 
1921 	ASSERT(fp->timer_mp == NULL);
1922 	ASSERT(fp->timer_running == 0);
1923 
1924 	ASSERT(fp->rc_timer_mp == NULL);
1925 	ASSERT(fp->rc_timer_running == 0);
1926 }
1927 
1928 void
1929 sctp_faddr_init(void)
1930 {
1931 	sctp_kmem_faddr_cache = kmem_cache_create("sctp_faddr_cache",
1932 	    sizeof (sctp_faddr_t), 0, faddr_constructor, faddr_destructor,
1933 	    NULL, NULL, NULL, 0);
1934 }
1935 
1936 void
1937 sctp_faddr_fini(void)
1938 {
1939 	kmem_cache_destroy(sctp_kmem_faddr_cache);
1940 }
1941