xref: /freebsd/sys/netinet/sctp_output.c (revision 78b9f0095b4af3aca6c931b2c7b009ddb8a05125)
1 /*-
2  * SPDX-License-Identifier: BSD-3-Clause
3  *
4  * Copyright (c) 2001-2008, by Cisco Systems, Inc. All rights reserved.
5  * Copyright (c) 2008-2012, by Randall Stewart. All rights reserved.
6  * Copyright (c) 2008-2012, by Michael Tuexen. All rights reserved.
7  *
8  * Redistribution and use in source and binary forms, with or without
9  * modification, are permitted provided that the following conditions are met:
10  *
11  * a) Redistributions of source code must retain the above copyright notice,
12  *    this list of conditions and the following disclaimer.
13  *
14  * b) Redistributions in binary form must reproduce the above copyright
15  *    notice, this list of conditions and the following disclaimer in
16  *    the documentation and/or other materials provided with the distribution.
17  *
18  * c) Neither the name of Cisco Systems, Inc. nor the names of its
19  *    contributors may be used to endorse or promote products derived
20  *    from this software without specific prior written permission.
21  *
22  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
23  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
24  * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
25  * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
26  * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
27  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
28  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
29  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
30  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
31  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
32  * THE POSSIBILITY OF SUCH DAMAGE.
33  */
34 
35 #include <sys/cdefs.h>
36 __FBSDID("$FreeBSD$");
37 
38 #include <netinet/sctp_os.h>
39 #include <sys/proc.h>
40 #include <netinet/sctp_var.h>
41 #include <netinet/sctp_sysctl.h>
42 #include <netinet/sctp_header.h>
43 #include <netinet/sctp_pcb.h>
44 #include <netinet/sctputil.h>
45 #include <netinet/sctp_output.h>
46 #include <netinet/sctp_uio.h>
47 #include <netinet/sctputil.h>
48 #include <netinet/sctp_auth.h>
49 #include <netinet/sctp_timer.h>
50 #include <netinet/sctp_asconf.h>
51 #include <netinet/sctp_indata.h>
52 #include <netinet/sctp_bsd_addr.h>
53 #include <netinet/sctp_input.h>
54 #include <netinet/sctp_crc32.h>
55 #if defined(INET) || defined(INET6)
56 #include <netinet/udp.h>
57 #endif
58 #include <netinet/udp_var.h>
59 #include <machine/in_cksum.h>
60 
61 
62 
63 #define SCTP_MAX_GAPS_INARRAY 4
64 struct sack_track {
65 	uint8_t right_edge;	/* mergable on the right edge */
66 	uint8_t left_edge;	/* mergable on the left edge */
67 	uint8_t num_entries;
68 	uint8_t spare;
69 	struct sctp_gap_ack_block gaps[SCTP_MAX_GAPS_INARRAY];
70 };
71 
72 const struct sack_track sack_array[256] = {
73 	{0, 0, 0, 0,		/* 0x00 */
74 		{{0, 0},
75 		{0, 0},
76 		{0, 0},
77 		{0, 0}
78 		}
79 	},
80 	{1, 0, 1, 0,		/* 0x01 */
81 		{{0, 0},
82 		{0, 0},
83 		{0, 0},
84 		{0, 0}
85 		}
86 	},
87 	{0, 0, 1, 0,		/* 0x02 */
88 		{{1, 1},
89 		{0, 0},
90 		{0, 0},
91 		{0, 0}
92 		}
93 	},
94 	{1, 0, 1, 0,		/* 0x03 */
95 		{{0, 1},
96 		{0, 0},
97 		{0, 0},
98 		{0, 0}
99 		}
100 	},
101 	{0, 0, 1, 0,		/* 0x04 */
102 		{{2, 2},
103 		{0, 0},
104 		{0, 0},
105 		{0, 0}
106 		}
107 	},
108 	{1, 0, 2, 0,		/* 0x05 */
109 		{{0, 0},
110 		{2, 2},
111 		{0, 0},
112 		{0, 0}
113 		}
114 	},
115 	{0, 0, 1, 0,		/* 0x06 */
116 		{{1, 2},
117 		{0, 0},
118 		{0, 0},
119 		{0, 0}
120 		}
121 	},
122 	{1, 0, 1, 0,		/* 0x07 */
123 		{{0, 2},
124 		{0, 0},
125 		{0, 0},
126 		{0, 0}
127 		}
128 	},
129 	{0, 0, 1, 0,		/* 0x08 */
130 		{{3, 3},
131 		{0, 0},
132 		{0, 0},
133 		{0, 0}
134 		}
135 	},
136 	{1, 0, 2, 0,		/* 0x09 */
137 		{{0, 0},
138 		{3, 3},
139 		{0, 0},
140 		{0, 0}
141 		}
142 	},
143 	{0, 0, 2, 0,		/* 0x0a */
144 		{{1, 1},
145 		{3, 3},
146 		{0, 0},
147 		{0, 0}
148 		}
149 	},
150 	{1, 0, 2, 0,		/* 0x0b */
151 		{{0, 1},
152 		{3, 3},
153 		{0, 0},
154 		{0, 0}
155 		}
156 	},
157 	{0, 0, 1, 0,		/* 0x0c */
158 		{{2, 3},
159 		{0, 0},
160 		{0, 0},
161 		{0, 0}
162 		}
163 	},
164 	{1, 0, 2, 0,		/* 0x0d */
165 		{{0, 0},
166 		{2, 3},
167 		{0, 0},
168 		{0, 0}
169 		}
170 	},
171 	{0, 0, 1, 0,		/* 0x0e */
172 		{{1, 3},
173 		{0, 0},
174 		{0, 0},
175 		{0, 0}
176 		}
177 	},
178 	{1, 0, 1, 0,		/* 0x0f */
179 		{{0, 3},
180 		{0, 0},
181 		{0, 0},
182 		{0, 0}
183 		}
184 	},
185 	{0, 0, 1, 0,		/* 0x10 */
186 		{{4, 4},
187 		{0, 0},
188 		{0, 0},
189 		{0, 0}
190 		}
191 	},
192 	{1, 0, 2, 0,		/* 0x11 */
193 		{{0, 0},
194 		{4, 4},
195 		{0, 0},
196 		{0, 0}
197 		}
198 	},
199 	{0, 0, 2, 0,		/* 0x12 */
200 		{{1, 1},
201 		{4, 4},
202 		{0, 0},
203 		{0, 0}
204 		}
205 	},
206 	{1, 0, 2, 0,		/* 0x13 */
207 		{{0, 1},
208 		{4, 4},
209 		{0, 0},
210 		{0, 0}
211 		}
212 	},
213 	{0, 0, 2, 0,		/* 0x14 */
214 		{{2, 2},
215 		{4, 4},
216 		{0, 0},
217 		{0, 0}
218 		}
219 	},
220 	{1, 0, 3, 0,		/* 0x15 */
221 		{{0, 0},
222 		{2, 2},
223 		{4, 4},
224 		{0, 0}
225 		}
226 	},
227 	{0, 0, 2, 0,		/* 0x16 */
228 		{{1, 2},
229 		{4, 4},
230 		{0, 0},
231 		{0, 0}
232 		}
233 	},
234 	{1, 0, 2, 0,		/* 0x17 */
235 		{{0, 2},
236 		{4, 4},
237 		{0, 0},
238 		{0, 0}
239 		}
240 	},
241 	{0, 0, 1, 0,		/* 0x18 */
242 		{{3, 4},
243 		{0, 0},
244 		{0, 0},
245 		{0, 0}
246 		}
247 	},
248 	{1, 0, 2, 0,		/* 0x19 */
249 		{{0, 0},
250 		{3, 4},
251 		{0, 0},
252 		{0, 0}
253 		}
254 	},
255 	{0, 0, 2, 0,		/* 0x1a */
256 		{{1, 1},
257 		{3, 4},
258 		{0, 0},
259 		{0, 0}
260 		}
261 	},
262 	{1, 0, 2, 0,		/* 0x1b */
263 		{{0, 1},
264 		{3, 4},
265 		{0, 0},
266 		{0, 0}
267 		}
268 	},
269 	{0, 0, 1, 0,		/* 0x1c */
270 		{{2, 4},
271 		{0, 0},
272 		{0, 0},
273 		{0, 0}
274 		}
275 	},
276 	{1, 0, 2, 0,		/* 0x1d */
277 		{{0, 0},
278 		{2, 4},
279 		{0, 0},
280 		{0, 0}
281 		}
282 	},
283 	{0, 0, 1, 0,		/* 0x1e */
284 		{{1, 4},
285 		{0, 0},
286 		{0, 0},
287 		{0, 0}
288 		}
289 	},
290 	{1, 0, 1, 0,		/* 0x1f */
291 		{{0, 4},
292 		{0, 0},
293 		{0, 0},
294 		{0, 0}
295 		}
296 	},
297 	{0, 0, 1, 0,		/* 0x20 */
298 		{{5, 5},
299 		{0, 0},
300 		{0, 0},
301 		{0, 0}
302 		}
303 	},
304 	{1, 0, 2, 0,		/* 0x21 */
305 		{{0, 0},
306 		{5, 5},
307 		{0, 0},
308 		{0, 0}
309 		}
310 	},
311 	{0, 0, 2, 0,		/* 0x22 */
312 		{{1, 1},
313 		{5, 5},
314 		{0, 0},
315 		{0, 0}
316 		}
317 	},
318 	{1, 0, 2, 0,		/* 0x23 */
319 		{{0, 1},
320 		{5, 5},
321 		{0, 0},
322 		{0, 0}
323 		}
324 	},
325 	{0, 0, 2, 0,		/* 0x24 */
326 		{{2, 2},
327 		{5, 5},
328 		{0, 0},
329 		{0, 0}
330 		}
331 	},
332 	{1, 0, 3, 0,		/* 0x25 */
333 		{{0, 0},
334 		{2, 2},
335 		{5, 5},
336 		{0, 0}
337 		}
338 	},
339 	{0, 0, 2, 0,		/* 0x26 */
340 		{{1, 2},
341 		{5, 5},
342 		{0, 0},
343 		{0, 0}
344 		}
345 	},
346 	{1, 0, 2, 0,		/* 0x27 */
347 		{{0, 2},
348 		{5, 5},
349 		{0, 0},
350 		{0, 0}
351 		}
352 	},
353 	{0, 0, 2, 0,		/* 0x28 */
354 		{{3, 3},
355 		{5, 5},
356 		{0, 0},
357 		{0, 0}
358 		}
359 	},
360 	{1, 0, 3, 0,		/* 0x29 */
361 		{{0, 0},
362 		{3, 3},
363 		{5, 5},
364 		{0, 0}
365 		}
366 	},
367 	{0, 0, 3, 0,		/* 0x2a */
368 		{{1, 1},
369 		{3, 3},
370 		{5, 5},
371 		{0, 0}
372 		}
373 	},
374 	{1, 0, 3, 0,		/* 0x2b */
375 		{{0, 1},
376 		{3, 3},
377 		{5, 5},
378 		{0, 0}
379 		}
380 	},
381 	{0, 0, 2, 0,		/* 0x2c */
382 		{{2, 3},
383 		{5, 5},
384 		{0, 0},
385 		{0, 0}
386 		}
387 	},
388 	{1, 0, 3, 0,		/* 0x2d */
389 		{{0, 0},
390 		{2, 3},
391 		{5, 5},
392 		{0, 0}
393 		}
394 	},
395 	{0, 0, 2, 0,		/* 0x2e */
396 		{{1, 3},
397 		{5, 5},
398 		{0, 0},
399 		{0, 0}
400 		}
401 	},
402 	{1, 0, 2, 0,		/* 0x2f */
403 		{{0, 3},
404 		{5, 5},
405 		{0, 0},
406 		{0, 0}
407 		}
408 	},
409 	{0, 0, 1, 0,		/* 0x30 */
410 		{{4, 5},
411 		{0, 0},
412 		{0, 0},
413 		{0, 0}
414 		}
415 	},
416 	{1, 0, 2, 0,		/* 0x31 */
417 		{{0, 0},
418 		{4, 5},
419 		{0, 0},
420 		{0, 0}
421 		}
422 	},
423 	{0, 0, 2, 0,		/* 0x32 */
424 		{{1, 1},
425 		{4, 5},
426 		{0, 0},
427 		{0, 0}
428 		}
429 	},
430 	{1, 0, 2, 0,		/* 0x33 */
431 		{{0, 1},
432 		{4, 5},
433 		{0, 0},
434 		{0, 0}
435 		}
436 	},
437 	{0, 0, 2, 0,		/* 0x34 */
438 		{{2, 2},
439 		{4, 5},
440 		{0, 0},
441 		{0, 0}
442 		}
443 	},
444 	{1, 0, 3, 0,		/* 0x35 */
445 		{{0, 0},
446 		{2, 2},
447 		{4, 5},
448 		{0, 0}
449 		}
450 	},
451 	{0, 0, 2, 0,		/* 0x36 */
452 		{{1, 2},
453 		{4, 5},
454 		{0, 0},
455 		{0, 0}
456 		}
457 	},
458 	{1, 0, 2, 0,		/* 0x37 */
459 		{{0, 2},
460 		{4, 5},
461 		{0, 0},
462 		{0, 0}
463 		}
464 	},
465 	{0, 0, 1, 0,		/* 0x38 */
466 		{{3, 5},
467 		{0, 0},
468 		{0, 0},
469 		{0, 0}
470 		}
471 	},
472 	{1, 0, 2, 0,		/* 0x39 */
473 		{{0, 0},
474 		{3, 5},
475 		{0, 0},
476 		{0, 0}
477 		}
478 	},
479 	{0, 0, 2, 0,		/* 0x3a */
480 		{{1, 1},
481 		{3, 5},
482 		{0, 0},
483 		{0, 0}
484 		}
485 	},
486 	{1, 0, 2, 0,		/* 0x3b */
487 		{{0, 1},
488 		{3, 5},
489 		{0, 0},
490 		{0, 0}
491 		}
492 	},
493 	{0, 0, 1, 0,		/* 0x3c */
494 		{{2, 5},
495 		{0, 0},
496 		{0, 0},
497 		{0, 0}
498 		}
499 	},
500 	{1, 0, 2, 0,		/* 0x3d */
501 		{{0, 0},
502 		{2, 5},
503 		{0, 0},
504 		{0, 0}
505 		}
506 	},
507 	{0, 0, 1, 0,		/* 0x3e */
508 		{{1, 5},
509 		{0, 0},
510 		{0, 0},
511 		{0, 0}
512 		}
513 	},
514 	{1, 0, 1, 0,		/* 0x3f */
515 		{{0, 5},
516 		{0, 0},
517 		{0, 0},
518 		{0, 0}
519 		}
520 	},
521 	{0, 0, 1, 0,		/* 0x40 */
522 		{{6, 6},
523 		{0, 0},
524 		{0, 0},
525 		{0, 0}
526 		}
527 	},
528 	{1, 0, 2, 0,		/* 0x41 */
529 		{{0, 0},
530 		{6, 6},
531 		{0, 0},
532 		{0, 0}
533 		}
534 	},
535 	{0, 0, 2, 0,		/* 0x42 */
536 		{{1, 1},
537 		{6, 6},
538 		{0, 0},
539 		{0, 0}
540 		}
541 	},
542 	{1, 0, 2, 0,		/* 0x43 */
543 		{{0, 1},
544 		{6, 6},
545 		{0, 0},
546 		{0, 0}
547 		}
548 	},
549 	{0, 0, 2, 0,		/* 0x44 */
550 		{{2, 2},
551 		{6, 6},
552 		{0, 0},
553 		{0, 0}
554 		}
555 	},
556 	{1, 0, 3, 0,		/* 0x45 */
557 		{{0, 0},
558 		{2, 2},
559 		{6, 6},
560 		{0, 0}
561 		}
562 	},
563 	{0, 0, 2, 0,		/* 0x46 */
564 		{{1, 2},
565 		{6, 6},
566 		{0, 0},
567 		{0, 0}
568 		}
569 	},
570 	{1, 0, 2, 0,		/* 0x47 */
571 		{{0, 2},
572 		{6, 6},
573 		{0, 0},
574 		{0, 0}
575 		}
576 	},
577 	{0, 0, 2, 0,		/* 0x48 */
578 		{{3, 3},
579 		{6, 6},
580 		{0, 0},
581 		{0, 0}
582 		}
583 	},
584 	{1, 0, 3, 0,		/* 0x49 */
585 		{{0, 0},
586 		{3, 3},
587 		{6, 6},
588 		{0, 0}
589 		}
590 	},
591 	{0, 0, 3, 0,		/* 0x4a */
592 		{{1, 1},
593 		{3, 3},
594 		{6, 6},
595 		{0, 0}
596 		}
597 	},
598 	{1, 0, 3, 0,		/* 0x4b */
599 		{{0, 1},
600 		{3, 3},
601 		{6, 6},
602 		{0, 0}
603 		}
604 	},
605 	{0, 0, 2, 0,		/* 0x4c */
606 		{{2, 3},
607 		{6, 6},
608 		{0, 0},
609 		{0, 0}
610 		}
611 	},
612 	{1, 0, 3, 0,		/* 0x4d */
613 		{{0, 0},
614 		{2, 3},
615 		{6, 6},
616 		{0, 0}
617 		}
618 	},
619 	{0, 0, 2, 0,		/* 0x4e */
620 		{{1, 3},
621 		{6, 6},
622 		{0, 0},
623 		{0, 0}
624 		}
625 	},
626 	{1, 0, 2, 0,		/* 0x4f */
627 		{{0, 3},
628 		{6, 6},
629 		{0, 0},
630 		{0, 0}
631 		}
632 	},
633 	{0, 0, 2, 0,		/* 0x50 */
634 		{{4, 4},
635 		{6, 6},
636 		{0, 0},
637 		{0, 0}
638 		}
639 	},
640 	{1, 0, 3, 0,		/* 0x51 */
641 		{{0, 0},
642 		{4, 4},
643 		{6, 6},
644 		{0, 0}
645 		}
646 	},
647 	{0, 0, 3, 0,		/* 0x52 */
648 		{{1, 1},
649 		{4, 4},
650 		{6, 6},
651 		{0, 0}
652 		}
653 	},
654 	{1, 0, 3, 0,		/* 0x53 */
655 		{{0, 1},
656 		{4, 4},
657 		{6, 6},
658 		{0, 0}
659 		}
660 	},
661 	{0, 0, 3, 0,		/* 0x54 */
662 		{{2, 2},
663 		{4, 4},
664 		{6, 6},
665 		{0, 0}
666 		}
667 	},
668 	{1, 0, 4, 0,		/* 0x55 */
669 		{{0, 0},
670 		{2, 2},
671 		{4, 4},
672 		{6, 6}
673 		}
674 	},
675 	{0, 0, 3, 0,		/* 0x56 */
676 		{{1, 2},
677 		{4, 4},
678 		{6, 6},
679 		{0, 0}
680 		}
681 	},
682 	{1, 0, 3, 0,		/* 0x57 */
683 		{{0, 2},
684 		{4, 4},
685 		{6, 6},
686 		{0, 0}
687 		}
688 	},
689 	{0, 0, 2, 0,		/* 0x58 */
690 		{{3, 4},
691 		{6, 6},
692 		{0, 0},
693 		{0, 0}
694 		}
695 	},
696 	{1, 0, 3, 0,		/* 0x59 */
697 		{{0, 0},
698 		{3, 4},
699 		{6, 6},
700 		{0, 0}
701 		}
702 	},
703 	{0, 0, 3, 0,		/* 0x5a */
704 		{{1, 1},
705 		{3, 4},
706 		{6, 6},
707 		{0, 0}
708 		}
709 	},
710 	{1, 0, 3, 0,		/* 0x5b */
711 		{{0, 1},
712 		{3, 4},
713 		{6, 6},
714 		{0, 0}
715 		}
716 	},
717 	{0, 0, 2, 0,		/* 0x5c */
718 		{{2, 4},
719 		{6, 6},
720 		{0, 0},
721 		{0, 0}
722 		}
723 	},
724 	{1, 0, 3, 0,		/* 0x5d */
725 		{{0, 0},
726 		{2, 4},
727 		{6, 6},
728 		{0, 0}
729 		}
730 	},
731 	{0, 0, 2, 0,		/* 0x5e */
732 		{{1, 4},
733 		{6, 6},
734 		{0, 0},
735 		{0, 0}
736 		}
737 	},
738 	{1, 0, 2, 0,		/* 0x5f */
739 		{{0, 4},
740 		{6, 6},
741 		{0, 0},
742 		{0, 0}
743 		}
744 	},
745 	{0, 0, 1, 0,		/* 0x60 */
746 		{{5, 6},
747 		{0, 0},
748 		{0, 0},
749 		{0, 0}
750 		}
751 	},
752 	{1, 0, 2, 0,		/* 0x61 */
753 		{{0, 0},
754 		{5, 6},
755 		{0, 0},
756 		{0, 0}
757 		}
758 	},
759 	{0, 0, 2, 0,		/* 0x62 */
760 		{{1, 1},
761 		{5, 6},
762 		{0, 0},
763 		{0, 0}
764 		}
765 	},
766 	{1, 0, 2, 0,		/* 0x63 */
767 		{{0, 1},
768 		{5, 6},
769 		{0, 0},
770 		{0, 0}
771 		}
772 	},
773 	{0, 0, 2, 0,		/* 0x64 */
774 		{{2, 2},
775 		{5, 6},
776 		{0, 0},
777 		{0, 0}
778 		}
779 	},
780 	{1, 0, 3, 0,		/* 0x65 */
781 		{{0, 0},
782 		{2, 2},
783 		{5, 6},
784 		{0, 0}
785 		}
786 	},
787 	{0, 0, 2, 0,		/* 0x66 */
788 		{{1, 2},
789 		{5, 6},
790 		{0, 0},
791 		{0, 0}
792 		}
793 	},
794 	{1, 0, 2, 0,		/* 0x67 */
795 		{{0, 2},
796 		{5, 6},
797 		{0, 0},
798 		{0, 0}
799 		}
800 	},
801 	{0, 0, 2, 0,		/* 0x68 */
802 		{{3, 3},
803 		{5, 6},
804 		{0, 0},
805 		{0, 0}
806 		}
807 	},
808 	{1, 0, 3, 0,		/* 0x69 */
809 		{{0, 0},
810 		{3, 3},
811 		{5, 6},
812 		{0, 0}
813 		}
814 	},
815 	{0, 0, 3, 0,		/* 0x6a */
816 		{{1, 1},
817 		{3, 3},
818 		{5, 6},
819 		{0, 0}
820 		}
821 	},
822 	{1, 0, 3, 0,		/* 0x6b */
823 		{{0, 1},
824 		{3, 3},
825 		{5, 6},
826 		{0, 0}
827 		}
828 	},
829 	{0, 0, 2, 0,		/* 0x6c */
830 		{{2, 3},
831 		{5, 6},
832 		{0, 0},
833 		{0, 0}
834 		}
835 	},
836 	{1, 0, 3, 0,		/* 0x6d */
837 		{{0, 0},
838 		{2, 3},
839 		{5, 6},
840 		{0, 0}
841 		}
842 	},
843 	{0, 0, 2, 0,		/* 0x6e */
844 		{{1, 3},
845 		{5, 6},
846 		{0, 0},
847 		{0, 0}
848 		}
849 	},
850 	{1, 0, 2, 0,		/* 0x6f */
851 		{{0, 3},
852 		{5, 6},
853 		{0, 0},
854 		{0, 0}
855 		}
856 	},
857 	{0, 0, 1, 0,		/* 0x70 */
858 		{{4, 6},
859 		{0, 0},
860 		{0, 0},
861 		{0, 0}
862 		}
863 	},
864 	{1, 0, 2, 0,		/* 0x71 */
865 		{{0, 0},
866 		{4, 6},
867 		{0, 0},
868 		{0, 0}
869 		}
870 	},
871 	{0, 0, 2, 0,		/* 0x72 */
872 		{{1, 1},
873 		{4, 6},
874 		{0, 0},
875 		{0, 0}
876 		}
877 	},
878 	{1, 0, 2, 0,		/* 0x73 */
879 		{{0, 1},
880 		{4, 6},
881 		{0, 0},
882 		{0, 0}
883 		}
884 	},
885 	{0, 0, 2, 0,		/* 0x74 */
886 		{{2, 2},
887 		{4, 6},
888 		{0, 0},
889 		{0, 0}
890 		}
891 	},
892 	{1, 0, 3, 0,		/* 0x75 */
893 		{{0, 0},
894 		{2, 2},
895 		{4, 6},
896 		{0, 0}
897 		}
898 	},
899 	{0, 0, 2, 0,		/* 0x76 */
900 		{{1, 2},
901 		{4, 6},
902 		{0, 0},
903 		{0, 0}
904 		}
905 	},
906 	{1, 0, 2, 0,		/* 0x77 */
907 		{{0, 2},
908 		{4, 6},
909 		{0, 0},
910 		{0, 0}
911 		}
912 	},
913 	{0, 0, 1, 0,		/* 0x78 */
914 		{{3, 6},
915 		{0, 0},
916 		{0, 0},
917 		{0, 0}
918 		}
919 	},
920 	{1, 0, 2, 0,		/* 0x79 */
921 		{{0, 0},
922 		{3, 6},
923 		{0, 0},
924 		{0, 0}
925 		}
926 	},
927 	{0, 0, 2, 0,		/* 0x7a */
928 		{{1, 1},
929 		{3, 6},
930 		{0, 0},
931 		{0, 0}
932 		}
933 	},
934 	{1, 0, 2, 0,		/* 0x7b */
935 		{{0, 1},
936 		{3, 6},
937 		{0, 0},
938 		{0, 0}
939 		}
940 	},
941 	{0, 0, 1, 0,		/* 0x7c */
942 		{{2, 6},
943 		{0, 0},
944 		{0, 0},
945 		{0, 0}
946 		}
947 	},
948 	{1, 0, 2, 0,		/* 0x7d */
949 		{{0, 0},
950 		{2, 6},
951 		{0, 0},
952 		{0, 0}
953 		}
954 	},
955 	{0, 0, 1, 0,		/* 0x7e */
956 		{{1, 6},
957 		{0, 0},
958 		{0, 0},
959 		{0, 0}
960 		}
961 	},
962 	{1, 0, 1, 0,		/* 0x7f */
963 		{{0, 6},
964 		{0, 0},
965 		{0, 0},
966 		{0, 0}
967 		}
968 	},
969 	{0, 1, 1, 0,		/* 0x80 */
970 		{{7, 7},
971 		{0, 0},
972 		{0, 0},
973 		{0, 0}
974 		}
975 	},
976 	{1, 1, 2, 0,		/* 0x81 */
977 		{{0, 0},
978 		{7, 7},
979 		{0, 0},
980 		{0, 0}
981 		}
982 	},
983 	{0, 1, 2, 0,		/* 0x82 */
984 		{{1, 1},
985 		{7, 7},
986 		{0, 0},
987 		{0, 0}
988 		}
989 	},
990 	{1, 1, 2, 0,		/* 0x83 */
991 		{{0, 1},
992 		{7, 7},
993 		{0, 0},
994 		{0, 0}
995 		}
996 	},
997 	{0, 1, 2, 0,		/* 0x84 */
998 		{{2, 2},
999 		{7, 7},
1000 		{0, 0},
1001 		{0, 0}
1002 		}
1003 	},
1004 	{1, 1, 3, 0,		/* 0x85 */
1005 		{{0, 0},
1006 		{2, 2},
1007 		{7, 7},
1008 		{0, 0}
1009 		}
1010 	},
1011 	{0, 1, 2, 0,		/* 0x86 */
1012 		{{1, 2},
1013 		{7, 7},
1014 		{0, 0},
1015 		{0, 0}
1016 		}
1017 	},
1018 	{1, 1, 2, 0,		/* 0x87 */
1019 		{{0, 2},
1020 		{7, 7},
1021 		{0, 0},
1022 		{0, 0}
1023 		}
1024 	},
1025 	{0, 1, 2, 0,		/* 0x88 */
1026 		{{3, 3},
1027 		{7, 7},
1028 		{0, 0},
1029 		{0, 0}
1030 		}
1031 	},
1032 	{1, 1, 3, 0,		/* 0x89 */
1033 		{{0, 0},
1034 		{3, 3},
1035 		{7, 7},
1036 		{0, 0}
1037 		}
1038 	},
1039 	{0, 1, 3, 0,		/* 0x8a */
1040 		{{1, 1},
1041 		{3, 3},
1042 		{7, 7},
1043 		{0, 0}
1044 		}
1045 	},
1046 	{1, 1, 3, 0,		/* 0x8b */
1047 		{{0, 1},
1048 		{3, 3},
1049 		{7, 7},
1050 		{0, 0}
1051 		}
1052 	},
1053 	{0, 1, 2, 0,		/* 0x8c */
1054 		{{2, 3},
1055 		{7, 7},
1056 		{0, 0},
1057 		{0, 0}
1058 		}
1059 	},
1060 	{1, 1, 3, 0,		/* 0x8d */
1061 		{{0, 0},
1062 		{2, 3},
1063 		{7, 7},
1064 		{0, 0}
1065 		}
1066 	},
1067 	{0, 1, 2, 0,		/* 0x8e */
1068 		{{1, 3},
1069 		{7, 7},
1070 		{0, 0},
1071 		{0, 0}
1072 		}
1073 	},
1074 	{1, 1, 2, 0,		/* 0x8f */
1075 		{{0, 3},
1076 		{7, 7},
1077 		{0, 0},
1078 		{0, 0}
1079 		}
1080 	},
1081 	{0, 1, 2, 0,		/* 0x90 */
1082 		{{4, 4},
1083 		{7, 7},
1084 		{0, 0},
1085 		{0, 0}
1086 		}
1087 	},
1088 	{1, 1, 3, 0,		/* 0x91 */
1089 		{{0, 0},
1090 		{4, 4},
1091 		{7, 7},
1092 		{0, 0}
1093 		}
1094 	},
1095 	{0, 1, 3, 0,		/* 0x92 */
1096 		{{1, 1},
1097 		{4, 4},
1098 		{7, 7},
1099 		{0, 0}
1100 		}
1101 	},
1102 	{1, 1, 3, 0,		/* 0x93 */
1103 		{{0, 1},
1104 		{4, 4},
1105 		{7, 7},
1106 		{0, 0}
1107 		}
1108 	},
1109 	{0, 1, 3, 0,		/* 0x94 */
1110 		{{2, 2},
1111 		{4, 4},
1112 		{7, 7},
1113 		{0, 0}
1114 		}
1115 	},
1116 	{1, 1, 4, 0,		/* 0x95 */
1117 		{{0, 0},
1118 		{2, 2},
1119 		{4, 4},
1120 		{7, 7}
1121 		}
1122 	},
1123 	{0, 1, 3, 0,		/* 0x96 */
1124 		{{1, 2},
1125 		{4, 4},
1126 		{7, 7},
1127 		{0, 0}
1128 		}
1129 	},
1130 	{1, 1, 3, 0,		/* 0x97 */
1131 		{{0, 2},
1132 		{4, 4},
1133 		{7, 7},
1134 		{0, 0}
1135 		}
1136 	},
1137 	{0, 1, 2, 0,		/* 0x98 */
1138 		{{3, 4},
1139 		{7, 7},
1140 		{0, 0},
1141 		{0, 0}
1142 		}
1143 	},
1144 	{1, 1, 3, 0,		/* 0x99 */
1145 		{{0, 0},
1146 		{3, 4},
1147 		{7, 7},
1148 		{0, 0}
1149 		}
1150 	},
1151 	{0, 1, 3, 0,		/* 0x9a */
1152 		{{1, 1},
1153 		{3, 4},
1154 		{7, 7},
1155 		{0, 0}
1156 		}
1157 	},
1158 	{1, 1, 3, 0,		/* 0x9b */
1159 		{{0, 1},
1160 		{3, 4},
1161 		{7, 7},
1162 		{0, 0}
1163 		}
1164 	},
1165 	{0, 1, 2, 0,		/* 0x9c */
1166 		{{2, 4},
1167 		{7, 7},
1168 		{0, 0},
1169 		{0, 0}
1170 		}
1171 	},
1172 	{1, 1, 3, 0,		/* 0x9d */
1173 		{{0, 0},
1174 		{2, 4},
1175 		{7, 7},
1176 		{0, 0}
1177 		}
1178 	},
1179 	{0, 1, 2, 0,		/* 0x9e */
1180 		{{1, 4},
1181 		{7, 7},
1182 		{0, 0},
1183 		{0, 0}
1184 		}
1185 	},
1186 	{1, 1, 2, 0,		/* 0x9f */
1187 		{{0, 4},
1188 		{7, 7},
1189 		{0, 0},
1190 		{0, 0}
1191 		}
1192 	},
1193 	{0, 1, 2, 0,		/* 0xa0 */
1194 		{{5, 5},
1195 		{7, 7},
1196 		{0, 0},
1197 		{0, 0}
1198 		}
1199 	},
1200 	{1, 1, 3, 0,		/* 0xa1 */
1201 		{{0, 0},
1202 		{5, 5},
1203 		{7, 7},
1204 		{0, 0}
1205 		}
1206 	},
1207 	{0, 1, 3, 0,		/* 0xa2 */
1208 		{{1, 1},
1209 		{5, 5},
1210 		{7, 7},
1211 		{0, 0}
1212 		}
1213 	},
1214 	{1, 1, 3, 0,		/* 0xa3 */
1215 		{{0, 1},
1216 		{5, 5},
1217 		{7, 7},
1218 		{0, 0}
1219 		}
1220 	},
1221 	{0, 1, 3, 0,		/* 0xa4 */
1222 		{{2, 2},
1223 		{5, 5},
1224 		{7, 7},
1225 		{0, 0}
1226 		}
1227 	},
1228 	{1, 1, 4, 0,		/* 0xa5 */
1229 		{{0, 0},
1230 		{2, 2},
1231 		{5, 5},
1232 		{7, 7}
1233 		}
1234 	},
1235 	{0, 1, 3, 0,		/* 0xa6 */
1236 		{{1, 2},
1237 		{5, 5},
1238 		{7, 7},
1239 		{0, 0}
1240 		}
1241 	},
1242 	{1, 1, 3, 0,		/* 0xa7 */
1243 		{{0, 2},
1244 		{5, 5},
1245 		{7, 7},
1246 		{0, 0}
1247 		}
1248 	},
1249 	{0, 1, 3, 0,		/* 0xa8 */
1250 		{{3, 3},
1251 		{5, 5},
1252 		{7, 7},
1253 		{0, 0}
1254 		}
1255 	},
1256 	{1, 1, 4, 0,		/* 0xa9 */
1257 		{{0, 0},
1258 		{3, 3},
1259 		{5, 5},
1260 		{7, 7}
1261 		}
1262 	},
1263 	{0, 1, 4, 0,		/* 0xaa */
1264 		{{1, 1},
1265 		{3, 3},
1266 		{5, 5},
1267 		{7, 7}
1268 		}
1269 	},
1270 	{1, 1, 4, 0,		/* 0xab */
1271 		{{0, 1},
1272 		{3, 3},
1273 		{5, 5},
1274 		{7, 7}
1275 		}
1276 	},
1277 	{0, 1, 3, 0,		/* 0xac */
1278 		{{2, 3},
1279 		{5, 5},
1280 		{7, 7},
1281 		{0, 0}
1282 		}
1283 	},
1284 	{1, 1, 4, 0,		/* 0xad */
1285 		{{0, 0},
1286 		{2, 3},
1287 		{5, 5},
1288 		{7, 7}
1289 		}
1290 	},
1291 	{0, 1, 3, 0,		/* 0xae */
1292 		{{1, 3},
1293 		{5, 5},
1294 		{7, 7},
1295 		{0, 0}
1296 		}
1297 	},
1298 	{1, 1, 3, 0,		/* 0xaf */
1299 		{{0, 3},
1300 		{5, 5},
1301 		{7, 7},
1302 		{0, 0}
1303 		}
1304 	},
1305 	{0, 1, 2, 0,		/* 0xb0 */
1306 		{{4, 5},
1307 		{7, 7},
1308 		{0, 0},
1309 		{0, 0}
1310 		}
1311 	},
1312 	{1, 1, 3, 0,		/* 0xb1 */
1313 		{{0, 0},
1314 		{4, 5},
1315 		{7, 7},
1316 		{0, 0}
1317 		}
1318 	},
1319 	{0, 1, 3, 0,		/* 0xb2 */
1320 		{{1, 1},
1321 		{4, 5},
1322 		{7, 7},
1323 		{0, 0}
1324 		}
1325 	},
1326 	{1, 1, 3, 0,		/* 0xb3 */
1327 		{{0, 1},
1328 		{4, 5},
1329 		{7, 7},
1330 		{0, 0}
1331 		}
1332 	},
1333 	{0, 1, 3, 0,		/* 0xb4 */
1334 		{{2, 2},
1335 		{4, 5},
1336 		{7, 7},
1337 		{0, 0}
1338 		}
1339 	},
1340 	{1, 1, 4, 0,		/* 0xb5 */
1341 		{{0, 0},
1342 		{2, 2},
1343 		{4, 5},
1344 		{7, 7}
1345 		}
1346 	},
1347 	{0, 1, 3, 0,		/* 0xb6 */
1348 		{{1, 2},
1349 		{4, 5},
1350 		{7, 7},
1351 		{0, 0}
1352 		}
1353 	},
1354 	{1, 1, 3, 0,		/* 0xb7 */
1355 		{{0, 2},
1356 		{4, 5},
1357 		{7, 7},
1358 		{0, 0}
1359 		}
1360 	},
1361 	{0, 1, 2, 0,		/* 0xb8 */
1362 		{{3, 5},
1363 		{7, 7},
1364 		{0, 0},
1365 		{0, 0}
1366 		}
1367 	},
1368 	{1, 1, 3, 0,		/* 0xb9 */
1369 		{{0, 0},
1370 		{3, 5},
1371 		{7, 7},
1372 		{0, 0}
1373 		}
1374 	},
1375 	{0, 1, 3, 0,		/* 0xba */
1376 		{{1, 1},
1377 		{3, 5},
1378 		{7, 7},
1379 		{0, 0}
1380 		}
1381 	},
1382 	{1, 1, 3, 0,		/* 0xbb */
1383 		{{0, 1},
1384 		{3, 5},
1385 		{7, 7},
1386 		{0, 0}
1387 		}
1388 	},
1389 	{0, 1, 2, 0,		/* 0xbc */
1390 		{{2, 5},
1391 		{7, 7},
1392 		{0, 0},
1393 		{0, 0}
1394 		}
1395 	},
1396 	{1, 1, 3, 0,		/* 0xbd */
1397 		{{0, 0},
1398 		{2, 5},
1399 		{7, 7},
1400 		{0, 0}
1401 		}
1402 	},
1403 	{0, 1, 2, 0,		/* 0xbe */
1404 		{{1, 5},
1405 		{7, 7},
1406 		{0, 0},
1407 		{0, 0}
1408 		}
1409 	},
1410 	{1, 1, 2, 0,		/* 0xbf */
1411 		{{0, 5},
1412 		{7, 7},
1413 		{0, 0},
1414 		{0, 0}
1415 		}
1416 	},
1417 	{0, 1, 1, 0,		/* 0xc0 */
1418 		{{6, 7},
1419 		{0, 0},
1420 		{0, 0},
1421 		{0, 0}
1422 		}
1423 	},
1424 	{1, 1, 2, 0,		/* 0xc1 */
1425 		{{0, 0},
1426 		{6, 7},
1427 		{0, 0},
1428 		{0, 0}
1429 		}
1430 	},
1431 	{0, 1, 2, 0,		/* 0xc2 */
1432 		{{1, 1},
1433 		{6, 7},
1434 		{0, 0},
1435 		{0, 0}
1436 		}
1437 	},
1438 	{1, 1, 2, 0,		/* 0xc3 */
1439 		{{0, 1},
1440 		{6, 7},
1441 		{0, 0},
1442 		{0, 0}
1443 		}
1444 	},
1445 	{0, 1, 2, 0,		/* 0xc4 */
1446 		{{2, 2},
1447 		{6, 7},
1448 		{0, 0},
1449 		{0, 0}
1450 		}
1451 	},
1452 	{1, 1, 3, 0,		/* 0xc5 */
1453 		{{0, 0},
1454 		{2, 2},
1455 		{6, 7},
1456 		{0, 0}
1457 		}
1458 	},
1459 	{0, 1, 2, 0,		/* 0xc6 */
1460 		{{1, 2},
1461 		{6, 7},
1462 		{0, 0},
1463 		{0, 0}
1464 		}
1465 	},
1466 	{1, 1, 2, 0,		/* 0xc7 */
1467 		{{0, 2},
1468 		{6, 7},
1469 		{0, 0},
1470 		{0, 0}
1471 		}
1472 	},
1473 	{0, 1, 2, 0,		/* 0xc8 */
1474 		{{3, 3},
1475 		{6, 7},
1476 		{0, 0},
1477 		{0, 0}
1478 		}
1479 	},
1480 	{1, 1, 3, 0,		/* 0xc9 */
1481 		{{0, 0},
1482 		{3, 3},
1483 		{6, 7},
1484 		{0, 0}
1485 		}
1486 	},
1487 	{0, 1, 3, 0,		/* 0xca */
1488 		{{1, 1},
1489 		{3, 3},
1490 		{6, 7},
1491 		{0, 0}
1492 		}
1493 	},
1494 	{1, 1, 3, 0,		/* 0xcb */
1495 		{{0, 1},
1496 		{3, 3},
1497 		{6, 7},
1498 		{0, 0}
1499 		}
1500 	},
1501 	{0, 1, 2, 0,		/* 0xcc */
1502 		{{2, 3},
1503 		{6, 7},
1504 		{0, 0},
1505 		{0, 0}
1506 		}
1507 	},
1508 	{1, 1, 3, 0,		/* 0xcd */
1509 		{{0, 0},
1510 		{2, 3},
1511 		{6, 7},
1512 		{0, 0}
1513 		}
1514 	},
1515 	{0, 1, 2, 0,		/* 0xce */
1516 		{{1, 3},
1517 		{6, 7},
1518 		{0, 0},
1519 		{0, 0}
1520 		}
1521 	},
1522 	{1, 1, 2, 0,		/* 0xcf */
1523 		{{0, 3},
1524 		{6, 7},
1525 		{0, 0},
1526 		{0, 0}
1527 		}
1528 	},
1529 	{0, 1, 2, 0,		/* 0xd0 */
1530 		{{4, 4},
1531 		{6, 7},
1532 		{0, 0},
1533 		{0, 0}
1534 		}
1535 	},
1536 	{1, 1, 3, 0,		/* 0xd1 */
1537 		{{0, 0},
1538 		{4, 4},
1539 		{6, 7},
1540 		{0, 0}
1541 		}
1542 	},
1543 	{0, 1, 3, 0,		/* 0xd2 */
1544 		{{1, 1},
1545 		{4, 4},
1546 		{6, 7},
1547 		{0, 0}
1548 		}
1549 	},
1550 	{1, 1, 3, 0,		/* 0xd3 */
1551 		{{0, 1},
1552 		{4, 4},
1553 		{6, 7},
1554 		{0, 0}
1555 		}
1556 	},
1557 	{0, 1, 3, 0,		/* 0xd4 */
1558 		{{2, 2},
1559 		{4, 4},
1560 		{6, 7},
1561 		{0, 0}
1562 		}
1563 	},
1564 	{1, 1, 4, 0,		/* 0xd5 */
1565 		{{0, 0},
1566 		{2, 2},
1567 		{4, 4},
1568 		{6, 7}
1569 		}
1570 	},
1571 	{0, 1, 3, 0,		/* 0xd6 */
1572 		{{1, 2},
1573 		{4, 4},
1574 		{6, 7},
1575 		{0, 0}
1576 		}
1577 	},
1578 	{1, 1, 3, 0,		/* 0xd7 */
1579 		{{0, 2},
1580 		{4, 4},
1581 		{6, 7},
1582 		{0, 0}
1583 		}
1584 	},
1585 	{0, 1, 2, 0,		/* 0xd8 */
1586 		{{3, 4},
1587 		{6, 7},
1588 		{0, 0},
1589 		{0, 0}
1590 		}
1591 	},
1592 	{1, 1, 3, 0,		/* 0xd9 */
1593 		{{0, 0},
1594 		{3, 4},
1595 		{6, 7},
1596 		{0, 0}
1597 		}
1598 	},
1599 	{0, 1, 3, 0,		/* 0xda */
1600 		{{1, 1},
1601 		{3, 4},
1602 		{6, 7},
1603 		{0, 0}
1604 		}
1605 	},
1606 	{1, 1, 3, 0,		/* 0xdb */
1607 		{{0, 1},
1608 		{3, 4},
1609 		{6, 7},
1610 		{0, 0}
1611 		}
1612 	},
1613 	{0, 1, 2, 0,		/* 0xdc */
1614 		{{2, 4},
1615 		{6, 7},
1616 		{0, 0},
1617 		{0, 0}
1618 		}
1619 	},
1620 	{1, 1, 3, 0,		/* 0xdd */
1621 		{{0, 0},
1622 		{2, 4},
1623 		{6, 7},
1624 		{0, 0}
1625 		}
1626 	},
1627 	{0, 1, 2, 0,		/* 0xde */
1628 		{{1, 4},
1629 		{6, 7},
1630 		{0, 0},
1631 		{0, 0}
1632 		}
1633 	},
1634 	{1, 1, 2, 0,		/* 0xdf */
1635 		{{0, 4},
1636 		{6, 7},
1637 		{0, 0},
1638 		{0, 0}
1639 		}
1640 	},
1641 	{0, 1, 1, 0,		/* 0xe0 */
1642 		{{5, 7},
1643 		{0, 0},
1644 		{0, 0},
1645 		{0, 0}
1646 		}
1647 	},
1648 	{1, 1, 2, 0,		/* 0xe1 */
1649 		{{0, 0},
1650 		{5, 7},
1651 		{0, 0},
1652 		{0, 0}
1653 		}
1654 	},
1655 	{0, 1, 2, 0,		/* 0xe2 */
1656 		{{1, 1},
1657 		{5, 7},
1658 		{0, 0},
1659 		{0, 0}
1660 		}
1661 	},
1662 	{1, 1, 2, 0,		/* 0xe3 */
1663 		{{0, 1},
1664 		{5, 7},
1665 		{0, 0},
1666 		{0, 0}
1667 		}
1668 	},
1669 	{0, 1, 2, 0,		/* 0xe4 */
1670 		{{2, 2},
1671 		{5, 7},
1672 		{0, 0},
1673 		{0, 0}
1674 		}
1675 	},
1676 	{1, 1, 3, 0,		/* 0xe5 */
1677 		{{0, 0},
1678 		{2, 2},
1679 		{5, 7},
1680 		{0, 0}
1681 		}
1682 	},
1683 	{0, 1, 2, 0,		/* 0xe6 */
1684 		{{1, 2},
1685 		{5, 7},
1686 		{0, 0},
1687 		{0, 0}
1688 		}
1689 	},
1690 	{1, 1, 2, 0,		/* 0xe7 */
1691 		{{0, 2},
1692 		{5, 7},
1693 		{0, 0},
1694 		{0, 0}
1695 		}
1696 	},
1697 	{0, 1, 2, 0,		/* 0xe8 */
1698 		{{3, 3},
1699 		{5, 7},
1700 		{0, 0},
1701 		{0, 0}
1702 		}
1703 	},
1704 	{1, 1, 3, 0,		/* 0xe9 */
1705 		{{0, 0},
1706 		{3, 3},
1707 		{5, 7},
1708 		{0, 0}
1709 		}
1710 	},
1711 	{0, 1, 3, 0,		/* 0xea */
1712 		{{1, 1},
1713 		{3, 3},
1714 		{5, 7},
1715 		{0, 0}
1716 		}
1717 	},
1718 	{1, 1, 3, 0,		/* 0xeb */
1719 		{{0, 1},
1720 		{3, 3},
1721 		{5, 7},
1722 		{0, 0}
1723 		}
1724 	},
1725 	{0, 1, 2, 0,		/* 0xec */
1726 		{{2, 3},
1727 		{5, 7},
1728 		{0, 0},
1729 		{0, 0}
1730 		}
1731 	},
1732 	{1, 1, 3, 0,		/* 0xed */
1733 		{{0, 0},
1734 		{2, 3},
1735 		{5, 7},
1736 		{0, 0}
1737 		}
1738 	},
1739 	{0, 1, 2, 0,		/* 0xee */
1740 		{{1, 3},
1741 		{5, 7},
1742 		{0, 0},
1743 		{0, 0}
1744 		}
1745 	},
1746 	{1, 1, 2, 0,		/* 0xef */
1747 		{{0, 3},
1748 		{5, 7},
1749 		{0, 0},
1750 		{0, 0}
1751 		}
1752 	},
1753 	{0, 1, 1, 0,		/* 0xf0 */
1754 		{{4, 7},
1755 		{0, 0},
1756 		{0, 0},
1757 		{0, 0}
1758 		}
1759 	},
1760 	{1, 1, 2, 0,		/* 0xf1 */
1761 		{{0, 0},
1762 		{4, 7},
1763 		{0, 0},
1764 		{0, 0}
1765 		}
1766 	},
1767 	{0, 1, 2, 0,		/* 0xf2 */
1768 		{{1, 1},
1769 		{4, 7},
1770 		{0, 0},
1771 		{0, 0}
1772 		}
1773 	},
1774 	{1, 1, 2, 0,		/* 0xf3 */
1775 		{{0, 1},
1776 		{4, 7},
1777 		{0, 0},
1778 		{0, 0}
1779 		}
1780 	},
1781 	{0, 1, 2, 0,		/* 0xf4 */
1782 		{{2, 2},
1783 		{4, 7},
1784 		{0, 0},
1785 		{0, 0}
1786 		}
1787 	},
1788 	{1, 1, 3, 0,		/* 0xf5 */
1789 		{{0, 0},
1790 		{2, 2},
1791 		{4, 7},
1792 		{0, 0}
1793 		}
1794 	},
1795 	{0, 1, 2, 0,		/* 0xf6 */
1796 		{{1, 2},
1797 		{4, 7},
1798 		{0, 0},
1799 		{0, 0}
1800 		}
1801 	},
1802 	{1, 1, 2, 0,		/* 0xf7 */
1803 		{{0, 2},
1804 		{4, 7},
1805 		{0, 0},
1806 		{0, 0}
1807 		}
1808 	},
1809 	{0, 1, 1, 0,		/* 0xf8 */
1810 		{{3, 7},
1811 		{0, 0},
1812 		{0, 0},
1813 		{0, 0}
1814 		}
1815 	},
1816 	{1, 1, 2, 0,		/* 0xf9 */
1817 		{{0, 0},
1818 		{3, 7},
1819 		{0, 0},
1820 		{0, 0}
1821 		}
1822 	},
1823 	{0, 1, 2, 0,		/* 0xfa */
1824 		{{1, 1},
1825 		{3, 7},
1826 		{0, 0},
1827 		{0, 0}
1828 		}
1829 	},
1830 	{1, 1, 2, 0,		/* 0xfb */
1831 		{{0, 1},
1832 		{3, 7},
1833 		{0, 0},
1834 		{0, 0}
1835 		}
1836 	},
1837 	{0, 1, 1, 0,		/* 0xfc */
1838 		{{2, 7},
1839 		{0, 0},
1840 		{0, 0},
1841 		{0, 0}
1842 		}
1843 	},
1844 	{1, 1, 2, 0,		/* 0xfd */
1845 		{{0, 0},
1846 		{2, 7},
1847 		{0, 0},
1848 		{0, 0}
1849 		}
1850 	},
1851 	{0, 1, 1, 0,		/* 0xfe */
1852 		{{1, 7},
1853 		{0, 0},
1854 		{0, 0},
1855 		{0, 0}
1856 		}
1857 	},
1858 	{1, 1, 1, 0,		/* 0xff */
1859 		{{0, 7},
1860 		{0, 0},
1861 		{0, 0},
1862 		{0, 0}
1863 		}
1864 	}
1865 };
1866 
1867 
1868 int
1869 sctp_is_address_in_scope(struct sctp_ifa *ifa,
1870     struct sctp_scoping *scope,
1871     int do_update)
1872 {
1873 	if ((scope->loopback_scope == 0) &&
1874 	    (ifa->ifn_p) && SCTP_IFN_IS_IFT_LOOP(ifa->ifn_p)) {
1875 		/*
1876 		 * skip loopback if not in scope *
1877 		 */
1878 		return (0);
1879 	}
1880 	switch (ifa->address.sa.sa_family) {
1881 #ifdef INET
1882 	case AF_INET:
1883 		if (scope->ipv4_addr_legal) {
1884 			struct sockaddr_in *sin;
1885 
1886 			sin = &ifa->address.sin;
1887 			if (sin->sin_addr.s_addr == 0) {
1888 				/* not in scope , unspecified */
1889 				return (0);
1890 			}
1891 			if ((scope->ipv4_local_scope == 0) &&
1892 			    (IN4_ISPRIVATE_ADDRESS(&sin->sin_addr))) {
1893 				/* private address not in scope */
1894 				return (0);
1895 			}
1896 		} else {
1897 			return (0);
1898 		}
1899 		break;
1900 #endif
1901 #ifdef INET6
1902 	case AF_INET6:
1903 		if (scope->ipv6_addr_legal) {
1904 			struct sockaddr_in6 *sin6;
1905 
1906 			/*
1907 			 * Must update the flags,  bummer, which means any
1908 			 * IFA locks must now be applied HERE <->
1909 			 */
1910 			if (do_update) {
1911 				sctp_gather_internal_ifa_flags(ifa);
1912 			}
1913 			if (ifa->localifa_flags & SCTP_ADDR_IFA_UNUSEABLE) {
1914 				return (0);
1915 			}
1916 			/* ok to use deprecated addresses? */
1917 			sin6 = &ifa->address.sin6;
1918 			if (IN6_IS_ADDR_UNSPECIFIED(&sin6->sin6_addr)) {
1919 				/* skip unspecifed addresses */
1920 				return (0);
1921 			}
1922 			if (	/* (local_scope == 0) && */
1923 			    (IN6_IS_ADDR_LINKLOCAL(&sin6->sin6_addr))) {
1924 				return (0);
1925 			}
1926 			if ((scope->site_scope == 0) &&
1927 			    (IN6_IS_ADDR_SITELOCAL(&sin6->sin6_addr))) {
1928 				return (0);
1929 			}
1930 		} else {
1931 			return (0);
1932 		}
1933 		break;
1934 #endif
1935 	default:
1936 		return (0);
1937 	}
1938 	return (1);
1939 }
1940 
1941 static struct mbuf *
1942 sctp_add_addr_to_mbuf(struct mbuf *m, struct sctp_ifa *ifa, uint16_t *len)
1943 {
1944 #if defined(INET) || defined(INET6)
1945 	struct sctp_paramhdr *paramh;
1946 	struct mbuf *mret;
1947 	uint16_t plen;
1948 #endif
1949 
1950 	switch (ifa->address.sa.sa_family) {
1951 #ifdef INET
1952 	case AF_INET:
1953 		plen = (uint16_t)sizeof(struct sctp_ipv4addr_param);
1954 		break;
1955 #endif
1956 #ifdef INET6
1957 	case AF_INET6:
1958 		plen = (uint16_t)sizeof(struct sctp_ipv6addr_param);
1959 		break;
1960 #endif
1961 	default:
1962 		return (m);
1963 	}
1964 #if defined(INET) || defined(INET6)
1965 	if (M_TRAILINGSPACE(m) >= plen) {
1966 		/* easy side we just drop it on the end */
1967 		paramh = (struct sctp_paramhdr *)(SCTP_BUF_AT(m, SCTP_BUF_LEN(m)));
1968 		mret = m;
1969 	} else {
1970 		/* Need more space */
1971 		mret = m;
1972 		while (SCTP_BUF_NEXT(mret) != NULL) {
1973 			mret = SCTP_BUF_NEXT(mret);
1974 		}
1975 		SCTP_BUF_NEXT(mret) = sctp_get_mbuf_for_msg(plen, 0, M_NOWAIT, 1, MT_DATA);
1976 		if (SCTP_BUF_NEXT(mret) == NULL) {
1977 			/* We are hosed, can't add more addresses */
1978 			return (m);
1979 		}
1980 		mret = SCTP_BUF_NEXT(mret);
1981 		paramh = mtod(mret, struct sctp_paramhdr *);
1982 	}
1983 	/* now add the parameter */
1984 	switch (ifa->address.sa.sa_family) {
1985 #ifdef INET
1986 	case AF_INET:
1987 		{
1988 			struct sctp_ipv4addr_param *ipv4p;
1989 			struct sockaddr_in *sin;
1990 
1991 			sin = &ifa->address.sin;
1992 			ipv4p = (struct sctp_ipv4addr_param *)paramh;
1993 			paramh->param_type = htons(SCTP_IPV4_ADDRESS);
1994 			paramh->param_length = htons(plen);
1995 			ipv4p->addr = sin->sin_addr.s_addr;
1996 			SCTP_BUF_LEN(mret) += plen;
1997 			break;
1998 		}
1999 #endif
2000 #ifdef INET6
2001 	case AF_INET6:
2002 		{
2003 			struct sctp_ipv6addr_param *ipv6p;
2004 			struct sockaddr_in6 *sin6;
2005 
2006 			sin6 = &ifa->address.sin6;
2007 			ipv6p = (struct sctp_ipv6addr_param *)paramh;
2008 			paramh->param_type = htons(SCTP_IPV6_ADDRESS);
2009 			paramh->param_length = htons(plen);
2010 			memcpy(ipv6p->addr, &sin6->sin6_addr,
2011 			    sizeof(ipv6p->addr));
2012 			/* clear embedded scope in the address */
2013 			in6_clearscope((struct in6_addr *)ipv6p->addr);
2014 			SCTP_BUF_LEN(mret) += plen;
2015 			break;
2016 		}
2017 #endif
2018 	default:
2019 		return (m);
2020 	}
2021 	if (len != NULL) {
2022 		*len += plen;
2023 	}
2024 	return (mret);
2025 #endif
2026 }
2027 
2028 
2029 struct mbuf *
2030 sctp_add_addresses_to_i_ia(struct sctp_inpcb *inp, struct sctp_tcb *stcb,
2031     struct sctp_scoping *scope,
2032     struct mbuf *m_at, int cnt_inits_to,
2033     uint16_t *padding_len, uint16_t *chunk_len)
2034 {
2035 	struct sctp_vrf *vrf = NULL;
2036 	int cnt, limit_out = 0, total_count;
2037 	uint32_t vrf_id;
2038 
2039 	vrf_id = inp->def_vrf_id;
2040 	SCTP_IPI_ADDR_RLOCK();
2041 	vrf = sctp_find_vrf(vrf_id);
2042 	if (vrf == NULL) {
2043 		SCTP_IPI_ADDR_RUNLOCK();
2044 		return (m_at);
2045 	}
2046 	if (inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) {
2047 		struct sctp_ifa *sctp_ifap;
2048 		struct sctp_ifn *sctp_ifnp;
2049 
2050 		cnt = cnt_inits_to;
2051 		if (vrf->total_ifa_count > SCTP_COUNT_LIMIT) {
2052 			limit_out = 1;
2053 			cnt = SCTP_ADDRESS_LIMIT;
2054 			goto skip_count;
2055 		}
2056 		LIST_FOREACH(sctp_ifnp, &vrf->ifnlist, next_ifn) {
2057 			if ((scope->loopback_scope == 0) &&
2058 			    SCTP_IFN_IS_IFT_LOOP(sctp_ifnp)) {
2059 				/*
2060 				 * Skip loopback devices if loopback_scope
2061 				 * not set
2062 				 */
2063 				continue;
2064 			}
2065 			LIST_FOREACH(sctp_ifap, &sctp_ifnp->ifalist, next_ifa) {
2066 #ifdef INET
2067 				if ((sctp_ifap->address.sa.sa_family == AF_INET) &&
2068 				    (prison_check_ip4(inp->ip_inp.inp.inp_cred,
2069 				    &sctp_ifap->address.sin.sin_addr) != 0)) {
2070 					continue;
2071 				}
2072 #endif
2073 #ifdef INET6
2074 				if ((sctp_ifap->address.sa.sa_family == AF_INET6) &&
2075 				    (prison_check_ip6(inp->ip_inp.inp.inp_cred,
2076 				    &sctp_ifap->address.sin6.sin6_addr) != 0)) {
2077 					continue;
2078 				}
2079 #endif
2080 				if (sctp_is_addr_restricted(stcb, sctp_ifap)) {
2081 					continue;
2082 				}
2083 				if (sctp_is_address_in_scope(sctp_ifap, scope, 1) == 0) {
2084 					continue;
2085 				}
2086 				cnt++;
2087 				if (cnt > SCTP_ADDRESS_LIMIT) {
2088 					break;
2089 				}
2090 			}
2091 			if (cnt > SCTP_ADDRESS_LIMIT) {
2092 				break;
2093 			}
2094 		}
2095 skip_count:
2096 		if (cnt > 1) {
2097 			total_count = 0;
2098 			LIST_FOREACH(sctp_ifnp, &vrf->ifnlist, next_ifn) {
2099 				cnt = 0;
2100 				if ((scope->loopback_scope == 0) &&
2101 				    SCTP_IFN_IS_IFT_LOOP(sctp_ifnp)) {
2102 					/*
2103 					 * Skip loopback devices if
2104 					 * loopback_scope not set
2105 					 */
2106 					continue;
2107 				}
2108 				LIST_FOREACH(sctp_ifap, &sctp_ifnp->ifalist, next_ifa) {
2109 #ifdef INET
2110 					if ((sctp_ifap->address.sa.sa_family == AF_INET) &&
2111 					    (prison_check_ip4(inp->ip_inp.inp.inp_cred,
2112 					    &sctp_ifap->address.sin.sin_addr) != 0)) {
2113 						continue;
2114 					}
2115 #endif
2116 #ifdef INET6
2117 					if ((sctp_ifap->address.sa.sa_family == AF_INET6) &&
2118 					    (prison_check_ip6(inp->ip_inp.inp.inp_cred,
2119 					    &sctp_ifap->address.sin6.sin6_addr) != 0)) {
2120 						continue;
2121 					}
2122 #endif
2123 					if (sctp_is_addr_restricted(stcb, sctp_ifap)) {
2124 						continue;
2125 					}
2126 					if (sctp_is_address_in_scope(sctp_ifap,
2127 					    scope, 0) == 0) {
2128 						continue;
2129 					}
2130 					if ((chunk_len != NULL) &&
2131 					    (padding_len != NULL) &&
2132 					    (*padding_len > 0)) {
2133 						memset(mtod(m_at, caddr_t)+*chunk_len, 0, *padding_len);
2134 						SCTP_BUF_LEN(m_at) += *padding_len;
2135 						*chunk_len += *padding_len;
2136 						*padding_len = 0;
2137 					}
2138 					m_at = sctp_add_addr_to_mbuf(m_at, sctp_ifap, chunk_len);
2139 					if (limit_out) {
2140 						cnt++;
2141 						total_count++;
2142 						if (cnt >= 2) {
2143 							/*
2144 							 * two from each
2145 							 * address
2146 							 */
2147 							break;
2148 						}
2149 						if (total_count > SCTP_ADDRESS_LIMIT) {
2150 							/* No more addresses */
2151 							break;
2152 						}
2153 					}
2154 				}
2155 			}
2156 		}
2157 	} else {
2158 		struct sctp_laddr *laddr;
2159 
2160 		cnt = cnt_inits_to;
2161 		/* First, how many ? */
2162 		LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
2163 			if (laddr->ifa == NULL) {
2164 				continue;
2165 			}
2166 			if (laddr->ifa->localifa_flags & SCTP_BEING_DELETED)
2167 				/*
2168 				 * Address being deleted by the system, dont
2169 				 * list.
2170 				 */
2171 				continue;
2172 			if (laddr->action == SCTP_DEL_IP_ADDRESS) {
2173 				/*
2174 				 * Address being deleted on this ep don't
2175 				 * list.
2176 				 */
2177 				continue;
2178 			}
2179 			if (sctp_is_address_in_scope(laddr->ifa,
2180 			    scope, 1) == 0) {
2181 				continue;
2182 			}
2183 			cnt++;
2184 		}
2185 		/*
2186 		 * To get through a NAT we only list addresses if we have
2187 		 * more than one. That way if you just bind a single address
2188 		 * we let the source of the init dictate our address.
2189 		 */
2190 		if (cnt > 1) {
2191 			cnt = cnt_inits_to;
2192 			LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
2193 				if (laddr->ifa == NULL) {
2194 					continue;
2195 				}
2196 				if (laddr->ifa->localifa_flags & SCTP_BEING_DELETED) {
2197 					continue;
2198 				}
2199 				if (sctp_is_address_in_scope(laddr->ifa,
2200 				    scope, 0) == 0) {
2201 					continue;
2202 				}
2203 				if ((chunk_len != NULL) &&
2204 				    (padding_len != NULL) &&
2205 				    (*padding_len > 0)) {
2206 					memset(mtod(m_at, caddr_t)+*chunk_len, 0, *padding_len);
2207 					SCTP_BUF_LEN(m_at) += *padding_len;
2208 					*chunk_len += *padding_len;
2209 					*padding_len = 0;
2210 				}
2211 				m_at = sctp_add_addr_to_mbuf(m_at, laddr->ifa, chunk_len);
2212 				cnt++;
2213 				if (cnt >= SCTP_ADDRESS_LIMIT) {
2214 					break;
2215 				}
2216 			}
2217 		}
2218 	}
2219 	SCTP_IPI_ADDR_RUNLOCK();
2220 	return (m_at);
2221 }
2222 
2223 static struct sctp_ifa *
2224 sctp_is_ifa_addr_preferred(struct sctp_ifa *ifa,
2225     uint8_t dest_is_loop,
2226     uint8_t dest_is_priv,
2227     sa_family_t fam)
2228 {
2229 	uint8_t dest_is_global = 0;
2230 
2231 	/* dest_is_priv is true if destination is a private address */
2232 	/* dest_is_loop is true if destination is a loopback addresses */
2233 
2234 	/**
2235 	 * Here we determine if its a preferred address. A preferred address
2236 	 * means it is the same scope or higher scope then the destination.
2237 	 * L = loopback, P = private, G = global
2238 	 * -----------------------------------------
2239 	 *    src    |  dest | result
2240 	 *  ----------------------------------------
2241 	 *     L     |    L  |    yes
2242 	 *  -----------------------------------------
2243 	 *     P     |    L  |    yes-v4 no-v6
2244 	 *  -----------------------------------------
2245 	 *     G     |    L  |    yes-v4 no-v6
2246 	 *  -----------------------------------------
2247 	 *     L     |    P  |    no
2248 	 *  -----------------------------------------
2249 	 *     P     |    P  |    yes
2250 	 *  -----------------------------------------
2251 	 *     G     |    P  |    no
2252 	 *   -----------------------------------------
2253 	 *     L     |    G  |    no
2254 	 *   -----------------------------------------
2255 	 *     P     |    G  |    no
2256 	 *    -----------------------------------------
2257 	 *     G     |    G  |    yes
2258 	 *    -----------------------------------------
2259 	 */
2260 
2261 	if (ifa->address.sa.sa_family != fam) {
2262 		/* forget mis-matched family */
2263 		return (NULL);
2264 	}
2265 	if ((dest_is_priv == 0) && (dest_is_loop == 0)) {
2266 		dest_is_global = 1;
2267 	}
2268 	SCTPDBG(SCTP_DEBUG_OUTPUT2, "Is destination preferred:");
2269 	SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT2, &ifa->address.sa);
2270 	/* Ok the address may be ok */
2271 #ifdef INET6
2272 	if (fam == AF_INET6) {
2273 		/* ok to use deprecated addresses? no lets not! */
2274 		if (ifa->localifa_flags & SCTP_ADDR_IFA_UNUSEABLE) {
2275 			SCTPDBG(SCTP_DEBUG_OUTPUT3, "NO:1\n");
2276 			return (NULL);
2277 		}
2278 		if (ifa->src_is_priv && !ifa->src_is_loop) {
2279 			if (dest_is_loop) {
2280 				SCTPDBG(SCTP_DEBUG_OUTPUT3, "NO:2\n");
2281 				return (NULL);
2282 			}
2283 		}
2284 		if (ifa->src_is_glob) {
2285 			if (dest_is_loop) {
2286 				SCTPDBG(SCTP_DEBUG_OUTPUT3, "NO:3\n");
2287 				return (NULL);
2288 			}
2289 		}
2290 	}
2291 #endif
2292 	/*
2293 	 * Now that we know what is what, implement or table this could in
2294 	 * theory be done slicker (it used to be), but this is
2295 	 * straightforward and easier to validate :-)
2296 	 */
2297 	SCTPDBG(SCTP_DEBUG_OUTPUT3, "src_loop:%d src_priv:%d src_glob:%d\n",
2298 	    ifa->src_is_loop, ifa->src_is_priv, ifa->src_is_glob);
2299 	SCTPDBG(SCTP_DEBUG_OUTPUT3, "dest_loop:%d dest_priv:%d dest_glob:%d\n",
2300 	    dest_is_loop, dest_is_priv, dest_is_global);
2301 
2302 	if ((ifa->src_is_loop) && (dest_is_priv)) {
2303 		SCTPDBG(SCTP_DEBUG_OUTPUT3, "NO:4\n");
2304 		return (NULL);
2305 	}
2306 	if ((ifa->src_is_glob) && (dest_is_priv)) {
2307 		SCTPDBG(SCTP_DEBUG_OUTPUT3, "NO:5\n");
2308 		return (NULL);
2309 	}
2310 	if ((ifa->src_is_loop) && (dest_is_global)) {
2311 		SCTPDBG(SCTP_DEBUG_OUTPUT3, "NO:6\n");
2312 		return (NULL);
2313 	}
2314 	if ((ifa->src_is_priv) && (dest_is_global)) {
2315 		SCTPDBG(SCTP_DEBUG_OUTPUT3, "NO:7\n");
2316 		return (NULL);
2317 	}
2318 	SCTPDBG(SCTP_DEBUG_OUTPUT3, "YES\n");
2319 	/* its a preferred address */
2320 	return (ifa);
2321 }
2322 
2323 static struct sctp_ifa *
2324 sctp_is_ifa_addr_acceptable(struct sctp_ifa *ifa,
2325     uint8_t dest_is_loop,
2326     uint8_t dest_is_priv,
2327     sa_family_t fam)
2328 {
2329 	uint8_t dest_is_global = 0;
2330 
2331 	/**
2332 	 * Here we determine if its a acceptable address. A acceptable
2333 	 * address means it is the same scope or higher scope but we can
2334 	 * allow for NAT which means its ok to have a global dest and a
2335 	 * private src.
2336 	 *
2337 	 * L = loopback, P = private, G = global
2338 	 * -----------------------------------------
2339 	 *  src    |  dest | result
2340 	 * -----------------------------------------
2341 	 *   L     |   L   |    yes
2342 	 *  -----------------------------------------
2343 	 *   P     |   L   |    yes-v4 no-v6
2344 	 *  -----------------------------------------
2345 	 *   G     |   L   |    yes
2346 	 * -----------------------------------------
2347 	 *   L     |   P   |    no
2348 	 * -----------------------------------------
2349 	 *   P     |   P   |    yes
2350 	 * -----------------------------------------
2351 	 *   G     |   P   |    yes - May not work
2352 	 * -----------------------------------------
2353 	 *   L     |   G   |    no
2354 	 * -----------------------------------------
2355 	 *   P     |   G   |    yes - May not work
2356 	 * -----------------------------------------
2357 	 *   G     |   G   |    yes
2358 	 * -----------------------------------------
2359 	 */
2360 
2361 	if (ifa->address.sa.sa_family != fam) {
2362 		/* forget non matching family */
2363 		SCTPDBG(SCTP_DEBUG_OUTPUT3, "ifa_fam:%d fam:%d\n",
2364 		    ifa->address.sa.sa_family, fam);
2365 		return (NULL);
2366 	}
2367 	/* Ok the address may be ok */
2368 	SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT3, &ifa->address.sa);
2369 	SCTPDBG(SCTP_DEBUG_OUTPUT3, "dst_is_loop:%d dest_is_priv:%d\n",
2370 	    dest_is_loop, dest_is_priv);
2371 	if ((dest_is_loop == 0) && (dest_is_priv == 0)) {
2372 		dest_is_global = 1;
2373 	}
2374 #ifdef INET6
2375 	if (fam == AF_INET6) {
2376 		/* ok to use deprecated addresses? */
2377 		if (ifa->localifa_flags & SCTP_ADDR_IFA_UNUSEABLE) {
2378 			return (NULL);
2379 		}
2380 		if (ifa->src_is_priv) {
2381 			/* Special case, linklocal to loop */
2382 			if (dest_is_loop)
2383 				return (NULL);
2384 		}
2385 	}
2386 #endif
2387 	/*
2388 	 * Now that we know what is what, implement our table. This could in
2389 	 * theory be done slicker (it used to be), but this is
2390 	 * straightforward and easier to validate :-)
2391 	 */
2392 	SCTPDBG(SCTP_DEBUG_OUTPUT3, "ifa->src_is_loop:%d dest_is_priv:%d\n",
2393 	    ifa->src_is_loop,
2394 	    dest_is_priv);
2395 	if ((ifa->src_is_loop == 1) && (dest_is_priv)) {
2396 		return (NULL);
2397 	}
2398 	SCTPDBG(SCTP_DEBUG_OUTPUT3, "ifa->src_is_loop:%d dest_is_glob:%d\n",
2399 	    ifa->src_is_loop,
2400 	    dest_is_global);
2401 	if ((ifa->src_is_loop == 1) && (dest_is_global)) {
2402 		return (NULL);
2403 	}
2404 	SCTPDBG(SCTP_DEBUG_OUTPUT3, "address is acceptable\n");
2405 	/* its an acceptable address */
2406 	return (ifa);
2407 }
2408 
2409 int
2410 sctp_is_addr_restricted(struct sctp_tcb *stcb, struct sctp_ifa *ifa)
2411 {
2412 	struct sctp_laddr *laddr;
2413 
2414 	if (stcb == NULL) {
2415 		/* There are no restrictions, no TCB :-) */
2416 		return (0);
2417 	}
2418 	LIST_FOREACH(laddr, &stcb->asoc.sctp_restricted_addrs, sctp_nxt_addr) {
2419 		if (laddr->ifa == NULL) {
2420 			SCTPDBG(SCTP_DEBUG_OUTPUT1, "%s: NULL ifa\n",
2421 			    __func__);
2422 			continue;
2423 		}
2424 		if (laddr->ifa == ifa) {
2425 			/* Yes it is on the list */
2426 			return (1);
2427 		}
2428 	}
2429 	return (0);
2430 }
2431 
2432 
2433 int
2434 sctp_is_addr_in_ep(struct sctp_inpcb *inp, struct sctp_ifa *ifa)
2435 {
2436 	struct sctp_laddr *laddr;
2437 
2438 	if (ifa == NULL)
2439 		return (0);
2440 	LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
2441 		if (laddr->ifa == NULL) {
2442 			SCTPDBG(SCTP_DEBUG_OUTPUT1, "%s: NULL ifa\n",
2443 			    __func__);
2444 			continue;
2445 		}
2446 		if ((laddr->ifa == ifa) && laddr->action == 0)
2447 			/* same pointer */
2448 			return (1);
2449 	}
2450 	return (0);
2451 }
2452 
2453 
2454 
2455 static struct sctp_ifa *
2456 sctp_choose_boundspecific_inp(struct sctp_inpcb *inp,
2457     sctp_route_t *ro,
2458     uint32_t vrf_id,
2459     int non_asoc_addr_ok,
2460     uint8_t dest_is_priv,
2461     uint8_t dest_is_loop,
2462     sa_family_t fam)
2463 {
2464 	struct sctp_laddr *laddr, *starting_point;
2465 	void *ifn;
2466 	int resettotop = 0;
2467 	struct sctp_ifn *sctp_ifn;
2468 	struct sctp_ifa *sctp_ifa, *sifa;
2469 	struct sctp_vrf *vrf;
2470 	uint32_t ifn_index;
2471 
2472 	vrf = sctp_find_vrf(vrf_id);
2473 	if (vrf == NULL)
2474 		return (NULL);
2475 
2476 	ifn = SCTP_GET_IFN_VOID_FROM_ROUTE(ro);
2477 	ifn_index = SCTP_GET_IF_INDEX_FROM_ROUTE(ro);
2478 	sctp_ifn = sctp_find_ifn(ifn, ifn_index);
2479 	/*
2480 	 * first question, is the ifn we will emit on in our list, if so, we
2481 	 * want such an address. Note that we first looked for a preferred
2482 	 * address.
2483 	 */
2484 	if (sctp_ifn) {
2485 		/* is a preferred one on the interface we route out? */
2486 		LIST_FOREACH(sctp_ifa, &sctp_ifn->ifalist, next_ifa) {
2487 #ifdef INET
2488 			if ((sctp_ifa->address.sa.sa_family == AF_INET) &&
2489 			    (prison_check_ip4(inp->ip_inp.inp.inp_cred,
2490 			    &sctp_ifa->address.sin.sin_addr) != 0)) {
2491 				continue;
2492 			}
2493 #endif
2494 #ifdef INET6
2495 			if ((sctp_ifa->address.sa.sa_family == AF_INET6) &&
2496 			    (prison_check_ip6(inp->ip_inp.inp.inp_cred,
2497 			    &sctp_ifa->address.sin6.sin6_addr) != 0)) {
2498 				continue;
2499 			}
2500 #endif
2501 			if ((sctp_ifa->localifa_flags & SCTP_ADDR_DEFER_USE) &&
2502 			    (non_asoc_addr_ok == 0))
2503 				continue;
2504 			sifa = sctp_is_ifa_addr_preferred(sctp_ifa,
2505 			    dest_is_loop,
2506 			    dest_is_priv, fam);
2507 			if (sifa == NULL)
2508 				continue;
2509 			if (sctp_is_addr_in_ep(inp, sifa)) {
2510 				atomic_add_int(&sifa->refcount, 1);
2511 				return (sifa);
2512 			}
2513 		}
2514 	}
2515 	/*
2516 	 * ok, now we now need to find one on the list of the addresses. We
2517 	 * can't get one on the emitting interface so let's find first a
2518 	 * preferred one. If not that an acceptable one otherwise... we
2519 	 * return NULL.
2520 	 */
2521 	starting_point = inp->next_addr_touse;
2522 once_again:
2523 	if (inp->next_addr_touse == NULL) {
2524 		inp->next_addr_touse = LIST_FIRST(&inp->sctp_addr_list);
2525 		resettotop = 1;
2526 	}
2527 	for (laddr = inp->next_addr_touse; laddr;
2528 	    laddr = LIST_NEXT(laddr, sctp_nxt_addr)) {
2529 		if (laddr->ifa == NULL) {
2530 			/* address has been removed */
2531 			continue;
2532 		}
2533 		if (laddr->action == SCTP_DEL_IP_ADDRESS) {
2534 			/* address is being deleted */
2535 			continue;
2536 		}
2537 		sifa = sctp_is_ifa_addr_preferred(laddr->ifa, dest_is_loop,
2538 		    dest_is_priv, fam);
2539 		if (sifa == NULL)
2540 			continue;
2541 		atomic_add_int(&sifa->refcount, 1);
2542 		return (sifa);
2543 	}
2544 	if (resettotop == 0) {
2545 		inp->next_addr_touse = NULL;
2546 		goto once_again;
2547 	}
2548 
2549 	inp->next_addr_touse = starting_point;
2550 	resettotop = 0;
2551 once_again_too:
2552 	if (inp->next_addr_touse == NULL) {
2553 		inp->next_addr_touse = LIST_FIRST(&inp->sctp_addr_list);
2554 		resettotop = 1;
2555 	}
2556 
2557 	/* ok, what about an acceptable address in the inp */
2558 	for (laddr = inp->next_addr_touse; laddr;
2559 	    laddr = LIST_NEXT(laddr, sctp_nxt_addr)) {
2560 		if (laddr->ifa == NULL) {
2561 			/* address has been removed */
2562 			continue;
2563 		}
2564 		if (laddr->action == SCTP_DEL_IP_ADDRESS) {
2565 			/* address is being deleted */
2566 			continue;
2567 		}
2568 		sifa = sctp_is_ifa_addr_acceptable(laddr->ifa, dest_is_loop,
2569 		    dest_is_priv, fam);
2570 		if (sifa == NULL)
2571 			continue;
2572 		atomic_add_int(&sifa->refcount, 1);
2573 		return (sifa);
2574 	}
2575 	if (resettotop == 0) {
2576 		inp->next_addr_touse = NULL;
2577 		goto once_again_too;
2578 	}
2579 
2580 	/*
2581 	 * no address bound can be a source for the destination we are in
2582 	 * trouble
2583 	 */
2584 	return (NULL);
2585 }
2586 
2587 
2588 
2589 static struct sctp_ifa *
2590 sctp_choose_boundspecific_stcb(struct sctp_inpcb *inp,
2591     struct sctp_tcb *stcb,
2592     sctp_route_t *ro,
2593     uint32_t vrf_id,
2594     uint8_t dest_is_priv,
2595     uint8_t dest_is_loop,
2596     int non_asoc_addr_ok,
2597     sa_family_t fam)
2598 {
2599 	struct sctp_laddr *laddr, *starting_point;
2600 	void *ifn;
2601 	struct sctp_ifn *sctp_ifn;
2602 	struct sctp_ifa *sctp_ifa, *sifa;
2603 	uint8_t start_at_beginning = 0;
2604 	struct sctp_vrf *vrf;
2605 	uint32_t ifn_index;
2606 
2607 	/*
2608 	 * first question, is the ifn we will emit on in our list, if so, we
2609 	 * want that one.
2610 	 */
2611 	vrf = sctp_find_vrf(vrf_id);
2612 	if (vrf == NULL)
2613 		return (NULL);
2614 
2615 	ifn = SCTP_GET_IFN_VOID_FROM_ROUTE(ro);
2616 	ifn_index = SCTP_GET_IF_INDEX_FROM_ROUTE(ro);
2617 	sctp_ifn = sctp_find_ifn(ifn, ifn_index);
2618 
2619 	/*
2620 	 * first question, is the ifn we will emit on in our list?  If so,
2621 	 * we want that one. First we look for a preferred. Second, we go
2622 	 * for an acceptable.
2623 	 */
2624 	if (sctp_ifn) {
2625 		/* first try for a preferred address on the ep */
2626 		LIST_FOREACH(sctp_ifa, &sctp_ifn->ifalist, next_ifa) {
2627 #ifdef INET
2628 			if ((sctp_ifa->address.sa.sa_family == AF_INET) &&
2629 			    (prison_check_ip4(inp->ip_inp.inp.inp_cred,
2630 			    &sctp_ifa->address.sin.sin_addr) != 0)) {
2631 				continue;
2632 			}
2633 #endif
2634 #ifdef INET6
2635 			if ((sctp_ifa->address.sa.sa_family == AF_INET6) &&
2636 			    (prison_check_ip6(inp->ip_inp.inp.inp_cred,
2637 			    &sctp_ifa->address.sin6.sin6_addr) != 0)) {
2638 				continue;
2639 			}
2640 #endif
2641 			if ((sctp_ifa->localifa_flags & SCTP_ADDR_DEFER_USE) && (non_asoc_addr_ok == 0))
2642 				continue;
2643 			if (sctp_is_addr_in_ep(inp, sctp_ifa)) {
2644 				sifa = sctp_is_ifa_addr_preferred(sctp_ifa, dest_is_loop, dest_is_priv, fam);
2645 				if (sifa == NULL)
2646 					continue;
2647 				if (((non_asoc_addr_ok == 0) &&
2648 				    (sctp_is_addr_restricted(stcb, sifa))) ||
2649 				    (non_asoc_addr_ok &&
2650 				    (sctp_is_addr_restricted(stcb, sifa)) &&
2651 				    (!sctp_is_addr_pending(stcb, sifa)))) {
2652 					/* on the no-no list */
2653 					continue;
2654 				}
2655 				atomic_add_int(&sifa->refcount, 1);
2656 				return (sifa);
2657 			}
2658 		}
2659 		/* next try for an acceptable address on the ep */
2660 		LIST_FOREACH(sctp_ifa, &sctp_ifn->ifalist, next_ifa) {
2661 #ifdef INET
2662 			if ((sctp_ifa->address.sa.sa_family == AF_INET) &&
2663 			    (prison_check_ip4(inp->ip_inp.inp.inp_cred,
2664 			    &sctp_ifa->address.sin.sin_addr) != 0)) {
2665 				continue;
2666 			}
2667 #endif
2668 #ifdef INET6
2669 			if ((sctp_ifa->address.sa.sa_family == AF_INET6) &&
2670 			    (prison_check_ip6(inp->ip_inp.inp.inp_cred,
2671 			    &sctp_ifa->address.sin6.sin6_addr) != 0)) {
2672 				continue;
2673 			}
2674 #endif
2675 			if ((sctp_ifa->localifa_flags & SCTP_ADDR_DEFER_USE) && (non_asoc_addr_ok == 0))
2676 				continue;
2677 			if (sctp_is_addr_in_ep(inp, sctp_ifa)) {
2678 				sifa = sctp_is_ifa_addr_acceptable(sctp_ifa, dest_is_loop, dest_is_priv, fam);
2679 				if (sifa == NULL)
2680 					continue;
2681 				if (((non_asoc_addr_ok == 0) &&
2682 				    (sctp_is_addr_restricted(stcb, sifa))) ||
2683 				    (non_asoc_addr_ok &&
2684 				    (sctp_is_addr_restricted(stcb, sifa)) &&
2685 				    (!sctp_is_addr_pending(stcb, sifa)))) {
2686 					/* on the no-no list */
2687 					continue;
2688 				}
2689 				atomic_add_int(&sifa->refcount, 1);
2690 				return (sifa);
2691 			}
2692 		}
2693 
2694 	}
2695 	/*
2696 	 * if we can't find one like that then we must look at all addresses
2697 	 * bound to pick one at first preferable then secondly acceptable.
2698 	 */
2699 	starting_point = stcb->asoc.last_used_address;
2700 sctp_from_the_top:
2701 	if (stcb->asoc.last_used_address == NULL) {
2702 		start_at_beginning = 1;
2703 		stcb->asoc.last_used_address = LIST_FIRST(&inp->sctp_addr_list);
2704 	}
2705 	/* search beginning with the last used address */
2706 	for (laddr = stcb->asoc.last_used_address; laddr;
2707 	    laddr = LIST_NEXT(laddr, sctp_nxt_addr)) {
2708 		if (laddr->ifa == NULL) {
2709 			/* address has been removed */
2710 			continue;
2711 		}
2712 		if (laddr->action == SCTP_DEL_IP_ADDRESS) {
2713 			/* address is being deleted */
2714 			continue;
2715 		}
2716 		sifa = sctp_is_ifa_addr_preferred(laddr->ifa, dest_is_loop, dest_is_priv, fam);
2717 		if (sifa == NULL)
2718 			continue;
2719 		if (((non_asoc_addr_ok == 0) &&
2720 		    (sctp_is_addr_restricted(stcb, sifa))) ||
2721 		    (non_asoc_addr_ok &&
2722 		    (sctp_is_addr_restricted(stcb, sifa)) &&
2723 		    (!sctp_is_addr_pending(stcb, sifa)))) {
2724 			/* on the no-no list */
2725 			continue;
2726 		}
2727 		stcb->asoc.last_used_address = laddr;
2728 		atomic_add_int(&sifa->refcount, 1);
2729 		return (sifa);
2730 	}
2731 	if (start_at_beginning == 0) {
2732 		stcb->asoc.last_used_address = NULL;
2733 		goto sctp_from_the_top;
2734 	}
2735 	/* now try for any higher scope than the destination */
2736 	stcb->asoc.last_used_address = starting_point;
2737 	start_at_beginning = 0;
2738 sctp_from_the_top2:
2739 	if (stcb->asoc.last_used_address == NULL) {
2740 		start_at_beginning = 1;
2741 		stcb->asoc.last_used_address = LIST_FIRST(&inp->sctp_addr_list);
2742 	}
2743 	/* search beginning with the last used address */
2744 	for (laddr = stcb->asoc.last_used_address; laddr;
2745 	    laddr = LIST_NEXT(laddr, sctp_nxt_addr)) {
2746 		if (laddr->ifa == NULL) {
2747 			/* address has been removed */
2748 			continue;
2749 		}
2750 		if (laddr->action == SCTP_DEL_IP_ADDRESS) {
2751 			/* address is being deleted */
2752 			continue;
2753 		}
2754 		sifa = sctp_is_ifa_addr_acceptable(laddr->ifa, dest_is_loop,
2755 		    dest_is_priv, fam);
2756 		if (sifa == NULL)
2757 			continue;
2758 		if (((non_asoc_addr_ok == 0) &&
2759 		    (sctp_is_addr_restricted(stcb, sifa))) ||
2760 		    (non_asoc_addr_ok &&
2761 		    (sctp_is_addr_restricted(stcb, sifa)) &&
2762 		    (!sctp_is_addr_pending(stcb, sifa)))) {
2763 			/* on the no-no list */
2764 			continue;
2765 		}
2766 		stcb->asoc.last_used_address = laddr;
2767 		atomic_add_int(&sifa->refcount, 1);
2768 		return (sifa);
2769 	}
2770 	if (start_at_beginning == 0) {
2771 		stcb->asoc.last_used_address = NULL;
2772 		goto sctp_from_the_top2;
2773 	}
2774 	return (NULL);
2775 }
2776 
2777 static struct sctp_ifa *
2778 sctp_select_nth_preferred_addr_from_ifn_boundall(struct sctp_ifn *ifn,
2779     struct sctp_inpcb *inp,
2780     struct sctp_tcb *stcb,
2781     int non_asoc_addr_ok,
2782     uint8_t dest_is_loop,
2783     uint8_t dest_is_priv,
2784     int addr_wanted,
2785     sa_family_t fam,
2786     sctp_route_t *ro
2787 )
2788 {
2789 	struct sctp_ifa *ifa, *sifa;
2790 	int num_eligible_addr = 0;
2791 #ifdef INET6
2792 	struct sockaddr_in6 sin6, lsa6;
2793 
2794 	if (fam == AF_INET6) {
2795 		memcpy(&sin6, &ro->ro_dst, sizeof(struct sockaddr_in6));
2796 		(void)sa6_recoverscope(&sin6);
2797 	}
2798 #endif				/* INET6 */
2799 	LIST_FOREACH(ifa, &ifn->ifalist, next_ifa) {
2800 #ifdef INET
2801 		if ((ifa->address.sa.sa_family == AF_INET) &&
2802 		    (prison_check_ip4(inp->ip_inp.inp.inp_cred,
2803 		    &ifa->address.sin.sin_addr) != 0)) {
2804 			continue;
2805 		}
2806 #endif
2807 #ifdef INET6
2808 		if ((ifa->address.sa.sa_family == AF_INET6) &&
2809 		    (prison_check_ip6(inp->ip_inp.inp.inp_cred,
2810 		    &ifa->address.sin6.sin6_addr) != 0)) {
2811 			continue;
2812 		}
2813 #endif
2814 		if ((ifa->localifa_flags & SCTP_ADDR_DEFER_USE) &&
2815 		    (non_asoc_addr_ok == 0))
2816 			continue;
2817 		sifa = sctp_is_ifa_addr_preferred(ifa, dest_is_loop,
2818 		    dest_is_priv, fam);
2819 		if (sifa == NULL)
2820 			continue;
2821 #ifdef INET6
2822 		if (fam == AF_INET6 &&
2823 		    dest_is_loop &&
2824 		    sifa->src_is_loop && sifa->src_is_priv) {
2825 			/*
2826 			 * don't allow fe80::1 to be a src on loop ::1, we
2827 			 * don't list it to the peer so we will get an
2828 			 * abort.
2829 			 */
2830 			continue;
2831 		}
2832 		if (fam == AF_INET6 &&
2833 		    IN6_IS_ADDR_LINKLOCAL(&sifa->address.sin6.sin6_addr) &&
2834 		    IN6_IS_ADDR_LINKLOCAL(&sin6.sin6_addr)) {
2835 			/*
2836 			 * link-local <-> link-local must belong to the same
2837 			 * scope.
2838 			 */
2839 			memcpy(&lsa6, &sifa->address.sin6, sizeof(struct sockaddr_in6));
2840 			(void)sa6_recoverscope(&lsa6);
2841 			if (sin6.sin6_scope_id != lsa6.sin6_scope_id) {
2842 				continue;
2843 			}
2844 		}
2845 #endif				/* INET6 */
2846 
2847 		/*
2848 		 * Check if the IPv6 address matches to next-hop. In the
2849 		 * mobile case, old IPv6 address may be not deleted from the
2850 		 * interface. Then, the interface has previous and new
2851 		 * addresses.  We should use one corresponding to the
2852 		 * next-hop.  (by micchie)
2853 		 */
2854 #ifdef INET6
2855 		if (stcb && fam == AF_INET6 &&
2856 		    sctp_is_mobility_feature_on(stcb->sctp_ep, SCTP_MOBILITY_BASE)) {
2857 			if (sctp_v6src_match_nexthop(&sifa->address.sin6, ro)
2858 			    == 0) {
2859 				continue;
2860 			}
2861 		}
2862 #endif
2863 #ifdef INET
2864 		/* Avoid topologically incorrect IPv4 address */
2865 		if (stcb && fam == AF_INET &&
2866 		    sctp_is_mobility_feature_on(stcb->sctp_ep, SCTP_MOBILITY_BASE)) {
2867 			if (sctp_v4src_match_nexthop(sifa, ro) == 0) {
2868 				continue;
2869 			}
2870 		}
2871 #endif
2872 		if (stcb) {
2873 			if (sctp_is_address_in_scope(ifa, &stcb->asoc.scope, 0) == 0) {
2874 				continue;
2875 			}
2876 			if (((non_asoc_addr_ok == 0) &&
2877 			    (sctp_is_addr_restricted(stcb, sifa))) ||
2878 			    (non_asoc_addr_ok &&
2879 			    (sctp_is_addr_restricted(stcb, sifa)) &&
2880 			    (!sctp_is_addr_pending(stcb, sifa)))) {
2881 				/*
2882 				 * It is restricted for some reason..
2883 				 * probably not yet added.
2884 				 */
2885 				continue;
2886 			}
2887 		}
2888 		if (num_eligible_addr >= addr_wanted) {
2889 			return (sifa);
2890 		}
2891 		num_eligible_addr++;
2892 	}
2893 	return (NULL);
2894 }
2895 
2896 
2897 static int
2898 sctp_count_num_preferred_boundall(struct sctp_ifn *ifn,
2899     struct sctp_inpcb *inp,
2900     struct sctp_tcb *stcb,
2901     int non_asoc_addr_ok,
2902     uint8_t dest_is_loop,
2903     uint8_t dest_is_priv,
2904     sa_family_t fam)
2905 {
2906 	struct sctp_ifa *ifa, *sifa;
2907 	int num_eligible_addr = 0;
2908 
2909 	LIST_FOREACH(ifa, &ifn->ifalist, next_ifa) {
2910 #ifdef INET
2911 		if ((ifa->address.sa.sa_family == AF_INET) &&
2912 		    (prison_check_ip4(inp->ip_inp.inp.inp_cred,
2913 		    &ifa->address.sin.sin_addr) != 0)) {
2914 			continue;
2915 		}
2916 #endif
2917 #ifdef INET6
2918 		if ((ifa->address.sa.sa_family == AF_INET6) &&
2919 		    (stcb != NULL) &&
2920 		    (prison_check_ip6(inp->ip_inp.inp.inp_cred,
2921 		    &ifa->address.sin6.sin6_addr) != 0)) {
2922 			continue;
2923 		}
2924 #endif
2925 		if ((ifa->localifa_flags & SCTP_ADDR_DEFER_USE) &&
2926 		    (non_asoc_addr_ok == 0)) {
2927 			continue;
2928 		}
2929 		sifa = sctp_is_ifa_addr_preferred(ifa, dest_is_loop,
2930 		    dest_is_priv, fam);
2931 		if (sifa == NULL) {
2932 			continue;
2933 		}
2934 		if (stcb) {
2935 			if (sctp_is_address_in_scope(ifa, &stcb->asoc.scope, 0) == 0) {
2936 				continue;
2937 			}
2938 			if (((non_asoc_addr_ok == 0) &&
2939 			    (sctp_is_addr_restricted(stcb, sifa))) ||
2940 			    (non_asoc_addr_ok &&
2941 			    (sctp_is_addr_restricted(stcb, sifa)) &&
2942 			    (!sctp_is_addr_pending(stcb, sifa)))) {
2943 				/*
2944 				 * It is restricted for some reason..
2945 				 * probably not yet added.
2946 				 */
2947 				continue;
2948 			}
2949 		}
2950 		num_eligible_addr++;
2951 	}
2952 	return (num_eligible_addr);
2953 }
2954 
2955 static struct sctp_ifa *
2956 sctp_choose_boundall(struct sctp_inpcb *inp,
2957     struct sctp_tcb *stcb,
2958     struct sctp_nets *net,
2959     sctp_route_t *ro,
2960     uint32_t vrf_id,
2961     uint8_t dest_is_priv,
2962     uint8_t dest_is_loop,
2963     int non_asoc_addr_ok,
2964     sa_family_t fam)
2965 {
2966 	int cur_addr_num = 0, num_preferred = 0;
2967 	void *ifn;
2968 	struct sctp_ifn *sctp_ifn, *looked_at = NULL, *emit_ifn;
2969 	struct sctp_ifa *sctp_ifa, *sifa;
2970 	uint32_t ifn_index;
2971 	struct sctp_vrf *vrf;
2972 #ifdef INET
2973 	int retried = 0;
2974 #endif
2975 
2976 	/*-
2977 	 * For boundall we can use any address in the association.
2978 	 * If non_asoc_addr_ok is set we can use any address (at least in
2979 	 * theory). So we look for preferred addresses first. If we find one,
2980 	 * we use it. Otherwise we next try to get an address on the
2981 	 * interface, which we should be able to do (unless non_asoc_addr_ok
2982 	 * is false and we are routed out that way). In these cases where we
2983 	 * can't use the address of the interface we go through all the
2984 	 * ifn's looking for an address we can use and fill that in. Punting
2985 	 * means we send back address 0, which will probably cause problems
2986 	 * actually since then IP will fill in the address of the route ifn,
2987 	 * which means we probably already rejected it.. i.e. here comes an
2988 	 * abort :-<.
2989 	 */
2990 	vrf = sctp_find_vrf(vrf_id);
2991 	if (vrf == NULL)
2992 		return (NULL);
2993 
2994 	ifn = SCTP_GET_IFN_VOID_FROM_ROUTE(ro);
2995 	ifn_index = SCTP_GET_IF_INDEX_FROM_ROUTE(ro);
2996 	SCTPDBG(SCTP_DEBUG_OUTPUT2, "ifn from route:%p ifn_index:%d\n", ifn, ifn_index);
2997 	emit_ifn = looked_at = sctp_ifn = sctp_find_ifn(ifn, ifn_index);
2998 	if (sctp_ifn == NULL) {
2999 		/* ?? We don't have this guy ?? */
3000 		SCTPDBG(SCTP_DEBUG_OUTPUT2, "No ifn emit interface?\n");
3001 		goto bound_all_plan_b;
3002 	}
3003 	SCTPDBG(SCTP_DEBUG_OUTPUT2, "ifn_index:%d name:%s is emit interface\n",
3004 	    ifn_index, sctp_ifn->ifn_name);
3005 
3006 	if (net) {
3007 		cur_addr_num = net->indx_of_eligible_next_to_use;
3008 	}
3009 	num_preferred = sctp_count_num_preferred_boundall(sctp_ifn,
3010 	    inp, stcb,
3011 	    non_asoc_addr_ok,
3012 	    dest_is_loop,
3013 	    dest_is_priv, fam);
3014 	SCTPDBG(SCTP_DEBUG_OUTPUT2, "Found %d preferred source addresses for intf:%s\n",
3015 	    num_preferred, sctp_ifn->ifn_name);
3016 	if (num_preferred == 0) {
3017 		/*
3018 		 * no eligible addresses, we must use some other interface
3019 		 * address if we can find one.
3020 		 */
3021 		goto bound_all_plan_b;
3022 	}
3023 	/*
3024 	 * Ok we have num_eligible_addr set with how many we can use, this
3025 	 * may vary from call to call due to addresses being deprecated
3026 	 * etc..
3027 	 */
3028 	if (cur_addr_num >= num_preferred) {
3029 		cur_addr_num = 0;
3030 	}
3031 	/*
3032 	 * select the nth address from the list (where cur_addr_num is the
3033 	 * nth) and 0 is the first one, 1 is the second one etc...
3034 	 */
3035 	SCTPDBG(SCTP_DEBUG_OUTPUT2, "cur_addr_num:%d\n", cur_addr_num);
3036 
3037 	sctp_ifa = sctp_select_nth_preferred_addr_from_ifn_boundall(sctp_ifn, inp, stcb, non_asoc_addr_ok, dest_is_loop,
3038 	    dest_is_priv, cur_addr_num, fam, ro);
3039 
3040 	/* if sctp_ifa is NULL something changed??, fall to plan b. */
3041 	if (sctp_ifa) {
3042 		atomic_add_int(&sctp_ifa->refcount, 1);
3043 		if (net) {
3044 			/* save off where the next one we will want */
3045 			net->indx_of_eligible_next_to_use = cur_addr_num + 1;
3046 		}
3047 		return (sctp_ifa);
3048 	}
3049 	/*
3050 	 * plan_b: Look at all interfaces and find a preferred address. If
3051 	 * no preferred fall through to plan_c.
3052 	 */
3053 bound_all_plan_b:
3054 	SCTPDBG(SCTP_DEBUG_OUTPUT2, "Trying Plan B\n");
3055 	LIST_FOREACH(sctp_ifn, &vrf->ifnlist, next_ifn) {
3056 		SCTPDBG(SCTP_DEBUG_OUTPUT2, "Examine interface %s\n",
3057 		    sctp_ifn->ifn_name);
3058 		if (dest_is_loop == 0 && SCTP_IFN_IS_IFT_LOOP(sctp_ifn)) {
3059 			/* wrong base scope */
3060 			SCTPDBG(SCTP_DEBUG_OUTPUT2, "skip\n");
3061 			continue;
3062 		}
3063 		if ((sctp_ifn == looked_at) && looked_at) {
3064 			/* already looked at this guy */
3065 			SCTPDBG(SCTP_DEBUG_OUTPUT2, "already seen\n");
3066 			continue;
3067 		}
3068 		num_preferred = sctp_count_num_preferred_boundall(sctp_ifn, inp, stcb, non_asoc_addr_ok,
3069 		    dest_is_loop, dest_is_priv, fam);
3070 		SCTPDBG(SCTP_DEBUG_OUTPUT2,
3071 		    "Found ifn:%p %d preferred source addresses\n",
3072 		    ifn, num_preferred);
3073 		if (num_preferred == 0) {
3074 			/* None on this interface. */
3075 			SCTPDBG(SCTP_DEBUG_OUTPUT2, "No preferred -- skipping to next\n");
3076 			continue;
3077 		}
3078 		SCTPDBG(SCTP_DEBUG_OUTPUT2,
3079 		    "num preferred:%d on interface:%p cur_addr_num:%d\n",
3080 		    num_preferred, (void *)sctp_ifn, cur_addr_num);
3081 
3082 		/*
3083 		 * Ok we have num_eligible_addr set with how many we can
3084 		 * use, this may vary from call to call due to addresses
3085 		 * being deprecated etc..
3086 		 */
3087 		if (cur_addr_num >= num_preferred) {
3088 			cur_addr_num = 0;
3089 		}
3090 		sifa = sctp_select_nth_preferred_addr_from_ifn_boundall(sctp_ifn, inp, stcb, non_asoc_addr_ok, dest_is_loop,
3091 		    dest_is_priv, cur_addr_num, fam, ro);
3092 		if (sifa == NULL)
3093 			continue;
3094 		if (net) {
3095 			net->indx_of_eligible_next_to_use = cur_addr_num + 1;
3096 			SCTPDBG(SCTP_DEBUG_OUTPUT2, "we selected %d\n",
3097 			    cur_addr_num);
3098 			SCTPDBG(SCTP_DEBUG_OUTPUT2, "Source:");
3099 			SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT2, &sifa->address.sa);
3100 			SCTPDBG(SCTP_DEBUG_OUTPUT2, "Dest:");
3101 			SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT2, &net->ro._l_addr.sa);
3102 		}
3103 		atomic_add_int(&sifa->refcount, 1);
3104 		return (sifa);
3105 	}
3106 #ifdef INET
3107 again_with_private_addresses_allowed:
3108 #endif
3109 	/* plan_c: do we have an acceptable address on the emit interface */
3110 	sifa = NULL;
3111 	SCTPDBG(SCTP_DEBUG_OUTPUT2, "Trying Plan C: find acceptable on interface\n");
3112 	if (emit_ifn == NULL) {
3113 		SCTPDBG(SCTP_DEBUG_OUTPUT2, "Jump to Plan D - no emit_ifn\n");
3114 		goto plan_d;
3115 	}
3116 	LIST_FOREACH(sctp_ifa, &emit_ifn->ifalist, next_ifa) {
3117 		SCTPDBG(SCTP_DEBUG_OUTPUT2, "ifa:%p\n", (void *)sctp_ifa);
3118 #ifdef INET
3119 		if ((sctp_ifa->address.sa.sa_family == AF_INET) &&
3120 		    (prison_check_ip4(inp->ip_inp.inp.inp_cred,
3121 		    &sctp_ifa->address.sin.sin_addr) != 0)) {
3122 			SCTPDBG(SCTP_DEBUG_OUTPUT2, "Jailed\n");
3123 			continue;
3124 		}
3125 #endif
3126 #ifdef INET6
3127 		if ((sctp_ifa->address.sa.sa_family == AF_INET6) &&
3128 		    (prison_check_ip6(inp->ip_inp.inp.inp_cred,
3129 		    &sctp_ifa->address.sin6.sin6_addr) != 0)) {
3130 			SCTPDBG(SCTP_DEBUG_OUTPUT2, "Jailed\n");
3131 			continue;
3132 		}
3133 #endif
3134 		if ((sctp_ifa->localifa_flags & SCTP_ADDR_DEFER_USE) &&
3135 		    (non_asoc_addr_ok == 0)) {
3136 			SCTPDBG(SCTP_DEBUG_OUTPUT2, "Defer\n");
3137 			continue;
3138 		}
3139 		sifa = sctp_is_ifa_addr_acceptable(sctp_ifa, dest_is_loop,
3140 		    dest_is_priv, fam);
3141 		if (sifa == NULL) {
3142 			SCTPDBG(SCTP_DEBUG_OUTPUT2, "IFA not acceptable\n");
3143 			continue;
3144 		}
3145 		if (stcb) {
3146 			if (sctp_is_address_in_scope(sifa, &stcb->asoc.scope, 0) == 0) {
3147 				SCTPDBG(SCTP_DEBUG_OUTPUT2, "NOT in scope\n");
3148 				sifa = NULL;
3149 				continue;
3150 			}
3151 			if (((non_asoc_addr_ok == 0) &&
3152 			    (sctp_is_addr_restricted(stcb, sifa))) ||
3153 			    (non_asoc_addr_ok &&
3154 			    (sctp_is_addr_restricted(stcb, sifa)) &&
3155 			    (!sctp_is_addr_pending(stcb, sifa)))) {
3156 				/*
3157 				 * It is restricted for some reason..
3158 				 * probably not yet added.
3159 				 */
3160 				SCTPDBG(SCTP_DEBUG_OUTPUT2, "Its restricted\n");
3161 				sifa = NULL;
3162 				continue;
3163 			}
3164 		}
3165 		atomic_add_int(&sifa->refcount, 1);
3166 		goto out;
3167 	}
3168 plan_d:
3169 	/*
3170 	 * plan_d: We are in trouble. No preferred address on the emit
3171 	 * interface. And not even a preferred address on all interfaces. Go
3172 	 * out and see if we can find an acceptable address somewhere
3173 	 * amongst all interfaces.
3174 	 */
3175 	SCTPDBG(SCTP_DEBUG_OUTPUT2, "Trying Plan D looked_at is %p\n", (void *)looked_at);
3176 	LIST_FOREACH(sctp_ifn, &vrf->ifnlist, next_ifn) {
3177 		if (dest_is_loop == 0 && SCTP_IFN_IS_IFT_LOOP(sctp_ifn)) {
3178 			/* wrong base scope */
3179 			continue;
3180 		}
3181 		LIST_FOREACH(sctp_ifa, &sctp_ifn->ifalist, next_ifa) {
3182 #ifdef INET
3183 			if ((sctp_ifa->address.sa.sa_family == AF_INET) &&
3184 			    (prison_check_ip4(inp->ip_inp.inp.inp_cred,
3185 			    &sctp_ifa->address.sin.sin_addr) != 0)) {
3186 				continue;
3187 			}
3188 #endif
3189 #ifdef INET6
3190 			if ((sctp_ifa->address.sa.sa_family == AF_INET6) &&
3191 			    (prison_check_ip6(inp->ip_inp.inp.inp_cred,
3192 			    &sctp_ifa->address.sin6.sin6_addr) != 0)) {
3193 				continue;
3194 			}
3195 #endif
3196 			if ((sctp_ifa->localifa_flags & SCTP_ADDR_DEFER_USE) &&
3197 			    (non_asoc_addr_ok == 0))
3198 				continue;
3199 			sifa = sctp_is_ifa_addr_acceptable(sctp_ifa,
3200 			    dest_is_loop,
3201 			    dest_is_priv, fam);
3202 			if (sifa == NULL)
3203 				continue;
3204 			if (stcb) {
3205 				if (sctp_is_address_in_scope(sifa, &stcb->asoc.scope, 0) == 0) {
3206 					sifa = NULL;
3207 					continue;
3208 				}
3209 				if (((non_asoc_addr_ok == 0) &&
3210 				    (sctp_is_addr_restricted(stcb, sifa))) ||
3211 				    (non_asoc_addr_ok &&
3212 				    (sctp_is_addr_restricted(stcb, sifa)) &&
3213 				    (!sctp_is_addr_pending(stcb, sifa)))) {
3214 					/*
3215 					 * It is restricted for some
3216 					 * reason.. probably not yet added.
3217 					 */
3218 					sifa = NULL;
3219 					continue;
3220 				}
3221 			}
3222 			goto out;
3223 		}
3224 	}
3225 #ifdef INET
3226 	if (stcb) {
3227 		if ((retried == 0) && (stcb->asoc.scope.ipv4_local_scope == 0)) {
3228 			stcb->asoc.scope.ipv4_local_scope = 1;
3229 			retried = 1;
3230 			goto again_with_private_addresses_allowed;
3231 		} else if (retried == 1) {
3232 			stcb->asoc.scope.ipv4_local_scope = 0;
3233 		}
3234 	}
3235 #endif
3236 out:
3237 #ifdef INET
3238 	if (sifa) {
3239 		if (retried == 1) {
3240 			LIST_FOREACH(sctp_ifn, &vrf->ifnlist, next_ifn) {
3241 				if (dest_is_loop == 0 && SCTP_IFN_IS_IFT_LOOP(sctp_ifn)) {
3242 					/* wrong base scope */
3243 					continue;
3244 				}
3245 				LIST_FOREACH(sctp_ifa, &sctp_ifn->ifalist, next_ifa) {
3246 					struct sctp_ifa *tmp_sifa;
3247 
3248 #ifdef INET
3249 					if ((sctp_ifa->address.sa.sa_family == AF_INET) &&
3250 					    (prison_check_ip4(inp->ip_inp.inp.inp_cred,
3251 					    &sctp_ifa->address.sin.sin_addr) != 0)) {
3252 						continue;
3253 					}
3254 #endif
3255 #ifdef INET6
3256 					if ((sctp_ifa->address.sa.sa_family == AF_INET6) &&
3257 					    (prison_check_ip6(inp->ip_inp.inp.inp_cred,
3258 					    &sctp_ifa->address.sin6.sin6_addr) != 0)) {
3259 						continue;
3260 					}
3261 #endif
3262 					if ((sctp_ifa->localifa_flags & SCTP_ADDR_DEFER_USE) &&
3263 					    (non_asoc_addr_ok == 0))
3264 						continue;
3265 					tmp_sifa = sctp_is_ifa_addr_acceptable(sctp_ifa,
3266 					    dest_is_loop,
3267 					    dest_is_priv, fam);
3268 					if (tmp_sifa == NULL) {
3269 						continue;
3270 					}
3271 					if (tmp_sifa == sifa) {
3272 						continue;
3273 					}
3274 					if (stcb) {
3275 						if (sctp_is_address_in_scope(tmp_sifa,
3276 						    &stcb->asoc.scope, 0) == 0) {
3277 							continue;
3278 						}
3279 						if (((non_asoc_addr_ok == 0) &&
3280 						    (sctp_is_addr_restricted(stcb, tmp_sifa))) ||
3281 						    (non_asoc_addr_ok &&
3282 						    (sctp_is_addr_restricted(stcb, tmp_sifa)) &&
3283 						    (!sctp_is_addr_pending(stcb, tmp_sifa)))) {
3284 							/*
3285 							 * It is restricted
3286 							 * for some reason..
3287 							 * probably not yet
3288 							 * added.
3289 							 */
3290 							continue;
3291 						}
3292 					}
3293 					if ((tmp_sifa->address.sin.sin_family == AF_INET) &&
3294 					    (IN4_ISPRIVATE_ADDRESS(&(tmp_sifa->address.sin.sin_addr)))) {
3295 						sctp_add_local_addr_restricted(stcb, tmp_sifa);
3296 					}
3297 				}
3298 			}
3299 		}
3300 		atomic_add_int(&sifa->refcount, 1);
3301 	}
3302 #endif
3303 	return (sifa);
3304 }
3305 
3306 
3307 
3308 /* tcb may be NULL */
3309 struct sctp_ifa *
3310 sctp_source_address_selection(struct sctp_inpcb *inp,
3311     struct sctp_tcb *stcb,
3312     sctp_route_t *ro,
3313     struct sctp_nets *net,
3314     int non_asoc_addr_ok, uint32_t vrf_id)
3315 {
3316 	struct sctp_ifa *answer;
3317 	uint8_t dest_is_priv, dest_is_loop;
3318 	sa_family_t fam;
3319 #ifdef INET
3320 	struct sockaddr_in *to = (struct sockaddr_in *)&ro->ro_dst;
3321 #endif
3322 #ifdef INET6
3323 	struct sockaddr_in6 *to6 = (struct sockaddr_in6 *)&ro->ro_dst;
3324 #endif
3325 
3326 	/**
3327 	 * Rules:
3328 	 * - Find the route if needed, cache if I can.
3329 	 * - Look at interface address in route, Is it in the bound list. If so we
3330 	 *   have the best source.
3331 	 * - If not we must rotate amongst the addresses.
3332 	 *
3333 	 * Cavets and issues
3334 	 *
3335 	 * Do we need to pay attention to scope. We can have a private address
3336 	 * or a global address we are sourcing or sending to. So if we draw
3337 	 * it out
3338 	 * zzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzz
3339 	 * For V4
3340 	 * ------------------------------------------
3341 	 *      source     *      dest  *  result
3342 	 * -----------------------------------------
3343 	 * <a>  Private    *    Global  *  NAT
3344 	 * -----------------------------------------
3345 	 * <b>  Private    *    Private *  No problem
3346 	 * -----------------------------------------
3347 	 * <c>  Global     *    Private *  Huh, How will this work?
3348 	 * -----------------------------------------
3349 	 * <d>  Global     *    Global  *  No Problem
3350 	 *------------------------------------------
3351 	 * zzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzz
3352 	 * For V6
3353 	 *------------------------------------------
3354 	 *      source     *      dest  *  result
3355 	 * -----------------------------------------
3356 	 * <a>  Linklocal  *    Global  *
3357 	 * -----------------------------------------
3358 	 * <b>  Linklocal  * Linklocal  *  No problem
3359 	 * -----------------------------------------
3360 	 * <c>  Global     * Linklocal  *  Huh, How will this work?
3361 	 * -----------------------------------------
3362 	 * <d>  Global     *    Global  *  No Problem
3363 	 *------------------------------------------
3364 	 * zzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzz
3365 	 *
3366 	 * And then we add to that what happens if there are multiple addresses
3367 	 * assigned to an interface. Remember the ifa on a ifn is a linked
3368 	 * list of addresses. So one interface can have more than one IP
3369 	 * address. What happens if we have both a private and a global
3370 	 * address? Do we then use context of destination to sort out which
3371 	 * one is best? And what about NAT's sending P->G may get you a NAT
3372 	 * translation, or should you select the G thats on the interface in
3373 	 * preference.
3374 	 *
3375 	 * Decisions:
3376 	 *
3377 	 * - count the number of addresses on the interface.
3378 	 * - if it is one, no problem except case <c>.
3379 	 *   For <a> we will assume a NAT out there.
3380 	 * - if there are more than one, then we need to worry about scope P
3381 	 *   or G. We should prefer G -> G and P -> P if possible.
3382 	 *   Then as a secondary fall back to mixed types G->P being a last
3383 	 *   ditch one.
3384 	 * - The above all works for bound all, but bound specific we need to
3385 	 *   use the same concept but instead only consider the bound
3386 	 *   addresses. If the bound set is NOT assigned to the interface then
3387 	 *   we must use rotation amongst the bound addresses..
3388 	 */
3389 	if (ro->ro_rt == NULL) {
3390 		/*
3391 		 * Need a route to cache.
3392 		 */
3393 		SCTP_RTALLOC(ro, vrf_id, inp->fibnum);
3394 	}
3395 	if (ro->ro_rt == NULL) {
3396 		return (NULL);
3397 	}
3398 	fam = ro->ro_dst.sa_family;
3399 	dest_is_priv = dest_is_loop = 0;
3400 	/* Setup our scopes for the destination */
3401 	switch (fam) {
3402 #ifdef INET
3403 	case AF_INET:
3404 		/* Scope based on outbound address */
3405 		if (IN4_ISLOOPBACK_ADDRESS(&to->sin_addr)) {
3406 			dest_is_loop = 1;
3407 			if (net != NULL) {
3408 				/* mark it as local */
3409 				net->addr_is_local = 1;
3410 			}
3411 		} else if ((IN4_ISPRIVATE_ADDRESS(&to->sin_addr))) {
3412 			dest_is_priv = 1;
3413 		}
3414 		break;
3415 #endif
3416 #ifdef INET6
3417 	case AF_INET6:
3418 		/* Scope based on outbound address */
3419 		if (IN6_IS_ADDR_LOOPBACK(&to6->sin6_addr) ||
3420 		    SCTP_ROUTE_IS_REAL_LOOP(ro)) {
3421 			/*
3422 			 * If the address is a loopback address, which
3423 			 * consists of "::1" OR "fe80::1%lo0", we are
3424 			 * loopback scope. But we don't use dest_is_priv
3425 			 * (link local addresses).
3426 			 */
3427 			dest_is_loop = 1;
3428 			if (net != NULL) {
3429 				/* mark it as local */
3430 				net->addr_is_local = 1;
3431 			}
3432 		} else if (IN6_IS_ADDR_LINKLOCAL(&to6->sin6_addr)) {
3433 			dest_is_priv = 1;
3434 		}
3435 		break;
3436 #endif
3437 	}
3438 	SCTPDBG(SCTP_DEBUG_OUTPUT2, "Select source addr for:");
3439 	SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT2, (struct sockaddr *)&ro->ro_dst);
3440 	SCTP_IPI_ADDR_RLOCK();
3441 	if (inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) {
3442 		/*
3443 		 * Bound all case
3444 		 */
3445 		answer = sctp_choose_boundall(inp, stcb, net, ro, vrf_id,
3446 		    dest_is_priv, dest_is_loop,
3447 		    non_asoc_addr_ok, fam);
3448 		SCTP_IPI_ADDR_RUNLOCK();
3449 		return (answer);
3450 	}
3451 	/*
3452 	 * Subset bound case
3453 	 */
3454 	if (stcb) {
3455 		answer = sctp_choose_boundspecific_stcb(inp, stcb, ro,
3456 		    vrf_id, dest_is_priv,
3457 		    dest_is_loop,
3458 		    non_asoc_addr_ok, fam);
3459 	} else {
3460 		answer = sctp_choose_boundspecific_inp(inp, ro, vrf_id,
3461 		    non_asoc_addr_ok,
3462 		    dest_is_priv,
3463 		    dest_is_loop, fam);
3464 	}
3465 	SCTP_IPI_ADDR_RUNLOCK();
3466 	return (answer);
3467 }
3468 
3469 static int
3470 sctp_find_cmsg(int c_type, void *data, struct mbuf *control, size_t cpsize)
3471 {
3472 	struct cmsghdr cmh;
3473 	struct sctp_sndinfo sndinfo;
3474 	struct sctp_prinfo prinfo;
3475 	struct sctp_authinfo authinfo;
3476 	int tot_len, rem_len, cmsg_data_len, cmsg_data_off, off;
3477 	int found;
3478 
3479 	/*
3480 	 * Independent of how many mbufs, find the c_type inside the control
3481 	 * structure and copy out the data.
3482 	 */
3483 	found = 0;
3484 	tot_len = SCTP_BUF_LEN(control);
3485 	for (off = 0; off < tot_len; off += CMSG_ALIGN(cmh.cmsg_len)) {
3486 		rem_len = tot_len - off;
3487 		if (rem_len < (int)CMSG_ALIGN(sizeof(cmh))) {
3488 			/* There is not enough room for one more. */
3489 			return (found);
3490 		}
3491 		m_copydata(control, off, sizeof(cmh), (caddr_t)&cmh);
3492 		if (cmh.cmsg_len < CMSG_ALIGN(sizeof(cmh))) {
3493 			/* We dont't have a complete CMSG header. */
3494 			return (found);
3495 		}
3496 		if ((cmh.cmsg_len > INT_MAX) || ((int)cmh.cmsg_len > rem_len)) {
3497 			/* We don't have the complete CMSG. */
3498 			return (found);
3499 		}
3500 		cmsg_data_len = (int)cmh.cmsg_len - CMSG_ALIGN(sizeof(cmh));
3501 		cmsg_data_off = off + CMSG_ALIGN(sizeof(cmh));
3502 		if ((cmh.cmsg_level == IPPROTO_SCTP) &&
3503 		    ((c_type == cmh.cmsg_type) ||
3504 		    ((c_type == SCTP_SNDRCV) &&
3505 		    ((cmh.cmsg_type == SCTP_SNDINFO) ||
3506 		    (cmh.cmsg_type == SCTP_PRINFO) ||
3507 		    (cmh.cmsg_type == SCTP_AUTHINFO))))) {
3508 			if (c_type == cmh.cmsg_type) {
3509 				if (cpsize > INT_MAX) {
3510 					return (found);
3511 				}
3512 				if (cmsg_data_len < (int)cpsize) {
3513 					return (found);
3514 				}
3515 				/* It is exactly what we want. Copy it out. */
3516 				m_copydata(control, cmsg_data_off, (int)cpsize, (caddr_t)data);
3517 				return (1);
3518 			} else {
3519 				struct sctp_sndrcvinfo *sndrcvinfo;
3520 
3521 				sndrcvinfo = (struct sctp_sndrcvinfo *)data;
3522 				if (found == 0) {
3523 					if (cpsize < sizeof(struct sctp_sndrcvinfo)) {
3524 						return (found);
3525 					}
3526 					memset(sndrcvinfo, 0, sizeof(struct sctp_sndrcvinfo));
3527 				}
3528 				switch (cmh.cmsg_type) {
3529 				case SCTP_SNDINFO:
3530 					if (cmsg_data_len < (int)sizeof(struct sctp_sndinfo)) {
3531 						return (found);
3532 					}
3533 					m_copydata(control, cmsg_data_off, sizeof(struct sctp_sndinfo), (caddr_t)&sndinfo);
3534 					sndrcvinfo->sinfo_stream = sndinfo.snd_sid;
3535 					sndrcvinfo->sinfo_flags = sndinfo.snd_flags;
3536 					sndrcvinfo->sinfo_ppid = sndinfo.snd_ppid;
3537 					sndrcvinfo->sinfo_context = sndinfo.snd_context;
3538 					sndrcvinfo->sinfo_assoc_id = sndinfo.snd_assoc_id;
3539 					break;
3540 				case SCTP_PRINFO:
3541 					if (cmsg_data_len < (int)sizeof(struct sctp_prinfo)) {
3542 						return (found);
3543 					}
3544 					m_copydata(control, cmsg_data_off, sizeof(struct sctp_prinfo), (caddr_t)&prinfo);
3545 					if (prinfo.pr_policy != SCTP_PR_SCTP_NONE) {
3546 						sndrcvinfo->sinfo_timetolive = prinfo.pr_value;
3547 					} else {
3548 						sndrcvinfo->sinfo_timetolive = 0;
3549 					}
3550 					sndrcvinfo->sinfo_flags |= prinfo.pr_policy;
3551 					break;
3552 				case SCTP_AUTHINFO:
3553 					if (cmsg_data_len < (int)sizeof(struct sctp_authinfo)) {
3554 						return (found);
3555 					}
3556 					m_copydata(control, cmsg_data_off, sizeof(struct sctp_authinfo), (caddr_t)&authinfo);
3557 					sndrcvinfo->sinfo_keynumber_valid = 1;
3558 					sndrcvinfo->sinfo_keynumber = authinfo.auth_keynumber;
3559 					break;
3560 				default:
3561 					return (found);
3562 				}
3563 				found = 1;
3564 			}
3565 		}
3566 	}
3567 	return (found);
3568 }
3569 
3570 static int
3571 sctp_process_cmsgs_for_init(struct sctp_tcb *stcb, struct mbuf *control, int *error)
3572 {
3573 	struct cmsghdr cmh;
3574 	int tlen, at;
3575 	struct sctp_initmsg initmsg;
3576 #ifdef INET
3577 	struct sockaddr_in sin;
3578 #endif
3579 #ifdef INET6
3580 	struct sockaddr_in6 sin6;
3581 #endif
3582 
3583 	tlen = SCTP_BUF_LEN(control);
3584 	at = 0;
3585 	while (at < tlen) {
3586 		if ((tlen - at) < (int)CMSG_ALIGN(sizeof(cmh))) {
3587 			/* There is not enough room for one more. */
3588 			*error = EINVAL;
3589 			return (1);
3590 		}
3591 		m_copydata(control, at, sizeof(cmh), (caddr_t)&cmh);
3592 		if (cmh.cmsg_len < CMSG_ALIGN(sizeof(cmh))) {
3593 			/* We dont't have a complete CMSG header. */
3594 			*error = EINVAL;
3595 			return (1);
3596 		}
3597 		if (((int)cmh.cmsg_len + at) > tlen) {
3598 			/* We don't have the complete CMSG. */
3599 			*error = EINVAL;
3600 			return (1);
3601 		}
3602 		if (cmh.cmsg_level == IPPROTO_SCTP) {
3603 			switch (cmh.cmsg_type) {
3604 			case SCTP_INIT:
3605 				if ((size_t)(cmh.cmsg_len - CMSG_ALIGN(sizeof(cmh))) < sizeof(struct sctp_initmsg)) {
3606 					*error = EINVAL;
3607 					return (1);
3608 				}
3609 				m_copydata(control, at + CMSG_ALIGN(sizeof(cmh)), sizeof(struct sctp_initmsg), (caddr_t)&initmsg);
3610 				if (initmsg.sinit_max_attempts)
3611 					stcb->asoc.max_init_times = initmsg.sinit_max_attempts;
3612 				if (initmsg.sinit_num_ostreams)
3613 					stcb->asoc.pre_open_streams = initmsg.sinit_num_ostreams;
3614 				if (initmsg.sinit_max_instreams)
3615 					stcb->asoc.max_inbound_streams = initmsg.sinit_max_instreams;
3616 				if (initmsg.sinit_max_init_timeo)
3617 					stcb->asoc.initial_init_rto_max = initmsg.sinit_max_init_timeo;
3618 				if (stcb->asoc.streamoutcnt < stcb->asoc.pre_open_streams) {
3619 					struct sctp_stream_out *tmp_str;
3620 					unsigned int i;
3621 #if defined(SCTP_DETAILED_STR_STATS)
3622 					int j;
3623 #endif
3624 
3625 					/* Default is NOT correct */
3626 					SCTPDBG(SCTP_DEBUG_OUTPUT1, "Ok, default:%d pre_open:%d\n",
3627 					    stcb->asoc.streamoutcnt, stcb->asoc.pre_open_streams);
3628 					SCTP_TCB_UNLOCK(stcb);
3629 					SCTP_MALLOC(tmp_str,
3630 					    struct sctp_stream_out *,
3631 					    (stcb->asoc.pre_open_streams * sizeof(struct sctp_stream_out)),
3632 					    SCTP_M_STRMO);
3633 					SCTP_TCB_LOCK(stcb);
3634 					if (tmp_str != NULL) {
3635 						SCTP_FREE(stcb->asoc.strmout, SCTP_M_STRMO);
3636 						stcb->asoc.strmout = tmp_str;
3637 						stcb->asoc.strm_realoutsize = stcb->asoc.streamoutcnt = stcb->asoc.pre_open_streams;
3638 					} else {
3639 						stcb->asoc.pre_open_streams = stcb->asoc.streamoutcnt;
3640 					}
3641 					for (i = 0; i < stcb->asoc.streamoutcnt; i++) {
3642 						TAILQ_INIT(&stcb->asoc.strmout[i].outqueue);
3643 						stcb->asoc.strmout[i].chunks_on_queues = 0;
3644 						stcb->asoc.strmout[i].next_mid_ordered = 0;
3645 						stcb->asoc.strmout[i].next_mid_unordered = 0;
3646 #if defined(SCTP_DETAILED_STR_STATS)
3647 						for (j = 0; j < SCTP_PR_SCTP_MAX + 1; j++) {
3648 							stcb->asoc.strmout[i].abandoned_sent[j] = 0;
3649 							stcb->asoc.strmout[i].abandoned_unsent[j] = 0;
3650 						}
3651 #else
3652 						stcb->asoc.strmout[i].abandoned_sent[0] = 0;
3653 						stcb->asoc.strmout[i].abandoned_unsent[0] = 0;
3654 #endif
3655 						stcb->asoc.strmout[i].sid = i;
3656 						stcb->asoc.strmout[i].last_msg_incomplete = 0;
3657 						stcb->asoc.strmout[i].state = SCTP_STREAM_OPENING;
3658 						stcb->asoc.ss_functions.sctp_ss_init_stream(stcb, &stcb->asoc.strmout[i], NULL);
3659 					}
3660 				}
3661 				break;
3662 #ifdef INET
3663 			case SCTP_DSTADDRV4:
3664 				if ((size_t)(cmh.cmsg_len - CMSG_ALIGN(sizeof(cmh))) < sizeof(struct in_addr)) {
3665 					*error = EINVAL;
3666 					return (1);
3667 				}
3668 				memset(&sin, 0, sizeof(struct sockaddr_in));
3669 				sin.sin_family = AF_INET;
3670 				sin.sin_len = sizeof(struct sockaddr_in);
3671 				sin.sin_port = stcb->rport;
3672 				m_copydata(control, at + CMSG_ALIGN(sizeof(cmh)), sizeof(struct in_addr), (caddr_t)&sin.sin_addr);
3673 				if ((sin.sin_addr.s_addr == INADDR_ANY) ||
3674 				    (sin.sin_addr.s_addr == INADDR_BROADCAST) ||
3675 				    IN_MULTICAST(ntohl(sin.sin_addr.s_addr))) {
3676 					*error = EINVAL;
3677 					return (1);
3678 				}
3679 				if (sctp_add_remote_addr(stcb, (struct sockaddr *)&sin, NULL, stcb->asoc.port,
3680 				    SCTP_DONOT_SETSCOPE, SCTP_ADDR_IS_CONFIRMED)) {
3681 					*error = ENOBUFS;
3682 					return (1);
3683 				}
3684 				break;
3685 #endif
3686 #ifdef INET6
3687 			case SCTP_DSTADDRV6:
3688 				if ((size_t)(cmh.cmsg_len - CMSG_ALIGN(sizeof(cmh))) < sizeof(struct in6_addr)) {
3689 					*error = EINVAL;
3690 					return (1);
3691 				}
3692 				memset(&sin6, 0, sizeof(struct sockaddr_in6));
3693 				sin6.sin6_family = AF_INET6;
3694 				sin6.sin6_len = sizeof(struct sockaddr_in6);
3695 				sin6.sin6_port = stcb->rport;
3696 				m_copydata(control, at + CMSG_ALIGN(sizeof(cmh)), sizeof(struct in6_addr), (caddr_t)&sin6.sin6_addr);
3697 				if (IN6_IS_ADDR_UNSPECIFIED(&sin6.sin6_addr) ||
3698 				    IN6_IS_ADDR_MULTICAST(&sin6.sin6_addr)) {
3699 					*error = EINVAL;
3700 					return (1);
3701 				}
3702 #ifdef INET
3703 				if (IN6_IS_ADDR_V4MAPPED(&sin6.sin6_addr)) {
3704 					in6_sin6_2_sin(&sin, &sin6);
3705 					if ((sin.sin_addr.s_addr == INADDR_ANY) ||
3706 					    (sin.sin_addr.s_addr == INADDR_BROADCAST) ||
3707 					    IN_MULTICAST(ntohl(sin.sin_addr.s_addr))) {
3708 						*error = EINVAL;
3709 						return (1);
3710 					}
3711 					if (sctp_add_remote_addr(stcb, (struct sockaddr *)&sin, NULL, stcb->asoc.port,
3712 					    SCTP_DONOT_SETSCOPE, SCTP_ADDR_IS_CONFIRMED)) {
3713 						*error = ENOBUFS;
3714 						return (1);
3715 					}
3716 				} else
3717 #endif
3718 					if (sctp_add_remote_addr(stcb, (struct sockaddr *)&sin6, NULL, stcb->asoc.port,
3719 				    SCTP_DONOT_SETSCOPE, SCTP_ADDR_IS_CONFIRMED)) {
3720 					*error = ENOBUFS;
3721 					return (1);
3722 				}
3723 				break;
3724 #endif
3725 			default:
3726 				break;
3727 			}
3728 		}
3729 		at += CMSG_ALIGN(cmh.cmsg_len);
3730 	}
3731 	return (0);
3732 }
3733 
3734 static struct sctp_tcb *
3735 sctp_findassociation_cmsgs(struct sctp_inpcb **inp_p,
3736     uint16_t port,
3737     struct mbuf *control,
3738     struct sctp_nets **net_p,
3739     int *error)
3740 {
3741 	struct cmsghdr cmh;
3742 	int tlen, at;
3743 	struct sctp_tcb *stcb;
3744 	struct sockaddr *addr;
3745 #ifdef INET
3746 	struct sockaddr_in sin;
3747 #endif
3748 #ifdef INET6
3749 	struct sockaddr_in6 sin6;
3750 #endif
3751 
3752 	tlen = SCTP_BUF_LEN(control);
3753 	at = 0;
3754 	while (at < tlen) {
3755 		if ((tlen - at) < (int)CMSG_ALIGN(sizeof(cmh))) {
3756 			/* There is not enough room for one more. */
3757 			*error = EINVAL;
3758 			return (NULL);
3759 		}
3760 		m_copydata(control, at, sizeof(cmh), (caddr_t)&cmh);
3761 		if (cmh.cmsg_len < CMSG_ALIGN(sizeof(cmh))) {
3762 			/* We dont't have a complete CMSG header. */
3763 			*error = EINVAL;
3764 			return (NULL);
3765 		}
3766 		if (((int)cmh.cmsg_len + at) > tlen) {
3767 			/* We don't have the complete CMSG. */
3768 			*error = EINVAL;
3769 			return (NULL);
3770 		}
3771 		if (cmh.cmsg_level == IPPROTO_SCTP) {
3772 			switch (cmh.cmsg_type) {
3773 #ifdef INET
3774 			case SCTP_DSTADDRV4:
3775 				if ((size_t)(cmh.cmsg_len - CMSG_ALIGN(sizeof(cmh))) < sizeof(struct in_addr)) {
3776 					*error = EINVAL;
3777 					return (NULL);
3778 				}
3779 				memset(&sin, 0, sizeof(struct sockaddr_in));
3780 				sin.sin_family = AF_INET;
3781 				sin.sin_len = sizeof(struct sockaddr_in);
3782 				sin.sin_port = port;
3783 				m_copydata(control, at + CMSG_ALIGN(sizeof(cmh)), sizeof(struct in_addr), (caddr_t)&sin.sin_addr);
3784 				addr = (struct sockaddr *)&sin;
3785 				break;
3786 #endif
3787 #ifdef INET6
3788 			case SCTP_DSTADDRV6:
3789 				if ((size_t)(cmh.cmsg_len - CMSG_ALIGN(sizeof(cmh))) < sizeof(struct in6_addr)) {
3790 					*error = EINVAL;
3791 					return (NULL);
3792 				}
3793 				memset(&sin6, 0, sizeof(struct sockaddr_in6));
3794 				sin6.sin6_family = AF_INET6;
3795 				sin6.sin6_len = sizeof(struct sockaddr_in6);
3796 				sin6.sin6_port = port;
3797 				m_copydata(control, at + CMSG_ALIGN(sizeof(cmh)), sizeof(struct in6_addr), (caddr_t)&sin6.sin6_addr);
3798 #ifdef INET
3799 				if (IN6_IS_ADDR_V4MAPPED(&sin6.sin6_addr)) {
3800 					in6_sin6_2_sin(&sin, &sin6);
3801 					addr = (struct sockaddr *)&sin;
3802 				} else
3803 #endif
3804 					addr = (struct sockaddr *)&sin6;
3805 				break;
3806 #endif
3807 			default:
3808 				addr = NULL;
3809 				break;
3810 			}
3811 			if (addr) {
3812 				stcb = sctp_findassociation_ep_addr(inp_p, addr, net_p, NULL, NULL);
3813 				if (stcb != NULL) {
3814 					return (stcb);
3815 				}
3816 			}
3817 		}
3818 		at += CMSG_ALIGN(cmh.cmsg_len);
3819 	}
3820 	return (NULL);
3821 }
3822 
3823 static struct mbuf *
3824 sctp_add_cookie(struct mbuf *init, int init_offset,
3825     struct mbuf *initack, int initack_offset, struct sctp_state_cookie *stc_in, uint8_t **signature)
3826 {
3827 	struct mbuf *copy_init, *copy_initack, *m_at, *sig, *mret;
3828 	struct sctp_state_cookie *stc;
3829 	struct sctp_paramhdr *ph;
3830 	uint8_t *foo;
3831 	int sig_offset;
3832 	uint16_t cookie_sz;
3833 
3834 	mret = sctp_get_mbuf_for_msg((sizeof(struct sctp_state_cookie) +
3835 	    sizeof(struct sctp_paramhdr)), 0,
3836 	    M_NOWAIT, 1, MT_DATA);
3837 	if (mret == NULL) {
3838 		return (NULL);
3839 	}
3840 	copy_init = SCTP_M_COPYM(init, init_offset, M_COPYALL, M_NOWAIT);
3841 	if (copy_init == NULL) {
3842 		sctp_m_freem(mret);
3843 		return (NULL);
3844 	}
3845 #ifdef SCTP_MBUF_LOGGING
3846 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_MBUF_LOGGING_ENABLE) {
3847 		sctp_log_mbc(copy_init, SCTP_MBUF_ICOPY);
3848 	}
3849 #endif
3850 	copy_initack = SCTP_M_COPYM(initack, initack_offset, M_COPYALL,
3851 	    M_NOWAIT);
3852 	if (copy_initack == NULL) {
3853 		sctp_m_freem(mret);
3854 		sctp_m_freem(copy_init);
3855 		return (NULL);
3856 	}
3857 #ifdef SCTP_MBUF_LOGGING
3858 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_MBUF_LOGGING_ENABLE) {
3859 		sctp_log_mbc(copy_initack, SCTP_MBUF_ICOPY);
3860 	}
3861 #endif
3862 	/* easy side we just drop it on the end */
3863 	ph = mtod(mret, struct sctp_paramhdr *);
3864 	SCTP_BUF_LEN(mret) = sizeof(struct sctp_state_cookie) +
3865 	    sizeof(struct sctp_paramhdr);
3866 	stc = (struct sctp_state_cookie *)((caddr_t)ph +
3867 	    sizeof(struct sctp_paramhdr));
3868 	ph->param_type = htons(SCTP_STATE_COOKIE);
3869 	ph->param_length = 0;	/* fill in at the end */
3870 	/* Fill in the stc cookie data */
3871 	memcpy(stc, stc_in, sizeof(struct sctp_state_cookie));
3872 
3873 	/* tack the INIT and then the INIT-ACK onto the chain */
3874 	cookie_sz = 0;
3875 	for (m_at = mret; m_at; m_at = SCTP_BUF_NEXT(m_at)) {
3876 		cookie_sz += SCTP_BUF_LEN(m_at);
3877 		if (SCTP_BUF_NEXT(m_at) == NULL) {
3878 			SCTP_BUF_NEXT(m_at) = copy_init;
3879 			break;
3880 		}
3881 	}
3882 	for (m_at = copy_init; m_at; m_at = SCTP_BUF_NEXT(m_at)) {
3883 		cookie_sz += SCTP_BUF_LEN(m_at);
3884 		if (SCTP_BUF_NEXT(m_at) == NULL) {
3885 			SCTP_BUF_NEXT(m_at) = copy_initack;
3886 			break;
3887 		}
3888 	}
3889 	for (m_at = copy_initack; m_at; m_at = SCTP_BUF_NEXT(m_at)) {
3890 		cookie_sz += SCTP_BUF_LEN(m_at);
3891 		if (SCTP_BUF_NEXT(m_at) == NULL) {
3892 			break;
3893 		}
3894 	}
3895 	sig = sctp_get_mbuf_for_msg(SCTP_SECRET_SIZE, 0, M_NOWAIT, 1, MT_DATA);
3896 	if (sig == NULL) {
3897 		/* no space, so free the entire chain */
3898 		sctp_m_freem(mret);
3899 		return (NULL);
3900 	}
3901 	SCTP_BUF_LEN(sig) = 0;
3902 	SCTP_BUF_NEXT(m_at) = sig;
3903 	sig_offset = 0;
3904 	foo = (uint8_t *)(mtod(sig, caddr_t)+sig_offset);
3905 	memset(foo, 0, SCTP_SIGNATURE_SIZE);
3906 	*signature = foo;
3907 	SCTP_BUF_LEN(sig) += SCTP_SIGNATURE_SIZE;
3908 	cookie_sz += SCTP_SIGNATURE_SIZE;
3909 	ph->param_length = htons(cookie_sz);
3910 	return (mret);
3911 }
3912 
3913 
3914 static uint8_t
3915 sctp_get_ect(struct sctp_tcb *stcb)
3916 {
3917 	if ((stcb != NULL) && (stcb->asoc.ecn_supported == 1)) {
3918 		return (SCTP_ECT0_BIT);
3919 	} else {
3920 		return (0);
3921 	}
3922 }
3923 
3924 #if defined(INET) || defined(INET6)
3925 static void
3926 sctp_handle_no_route(struct sctp_tcb *stcb,
3927     struct sctp_nets *net,
3928     int so_locked)
3929 {
3930 	SCTPDBG(SCTP_DEBUG_OUTPUT1, "dropped packet - no valid source addr\n");
3931 
3932 	if (net) {
3933 		SCTPDBG(SCTP_DEBUG_OUTPUT1, "Destination was ");
3934 		SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT1, &net->ro._l_addr.sa);
3935 		if (net->dest_state & SCTP_ADDR_CONFIRMED) {
3936 			if ((net->dest_state & SCTP_ADDR_REACHABLE) && stcb) {
3937 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "no route takes interface %p down\n", (void *)net);
3938 				sctp_ulp_notify(SCTP_NOTIFY_INTERFACE_DOWN,
3939 				    stcb, 0,
3940 				    (void *)net,
3941 				    so_locked);
3942 				net->dest_state &= ~SCTP_ADDR_REACHABLE;
3943 				net->dest_state &= ~SCTP_ADDR_PF;
3944 			}
3945 		}
3946 		if (stcb) {
3947 			if (net == stcb->asoc.primary_destination) {
3948 				/* need a new primary */
3949 				struct sctp_nets *alt;
3950 
3951 				alt = sctp_find_alternate_net(stcb, net, 0);
3952 				if (alt != net) {
3953 					if (stcb->asoc.alternate) {
3954 						sctp_free_remote_addr(stcb->asoc.alternate);
3955 					}
3956 					stcb->asoc.alternate = alt;
3957 					atomic_add_int(&stcb->asoc.alternate->ref_count, 1);
3958 					if (net->ro._s_addr) {
3959 						sctp_free_ifa(net->ro._s_addr);
3960 						net->ro._s_addr = NULL;
3961 					}
3962 					net->src_addr_selected = 0;
3963 				}
3964 			}
3965 		}
3966 	}
3967 }
3968 #endif
3969 
3970 static int
3971 sctp_lowlevel_chunk_output(struct sctp_inpcb *inp,
3972     struct sctp_tcb *stcb,	/* may be NULL */
3973     struct sctp_nets *net,
3974     struct sockaddr *to,
3975     struct mbuf *m,
3976     uint32_t auth_offset,
3977     struct sctp_auth_chunk *auth,
3978     uint16_t auth_keyid,
3979     int nofragment_flag,
3980     int ecn_ok,
3981     int out_of_asoc_ok,
3982     uint16_t src_port,
3983     uint16_t dest_port,
3984     uint32_t v_tag,
3985     uint16_t port,
3986     union sctp_sockstore *over_addr,
3987     uint8_t mflowtype, uint32_t mflowid,
3988 #if !defined(__APPLE__) && !defined(SCTP_SO_LOCK_TESTING)
3989     int so_locked SCTP_UNUSED
3990 #else
3991     int so_locked
3992 #endif
3993 )
3994 {
3995 /* nofragment_flag to tell if IP_DF should be set (IPv4 only) */
3996 	/**
3997 	 * Given a mbuf chain (via SCTP_BUF_NEXT()) that holds a packet header
3998 	 * WITH an SCTPHDR but no IP header, endpoint inp and sa structure:
3999 	 * - fill in the HMAC digest of any AUTH chunk in the packet.
4000 	 * - calculate and fill in the SCTP checksum.
4001 	 * - prepend an IP address header.
4002 	 * - if boundall use INADDR_ANY.
4003 	 * - if boundspecific do source address selection.
4004 	 * - set fragmentation option for ipV4.
4005 	 * - On return from IP output, check/adjust mtu size of output
4006 	 *   interface and smallest_mtu size as well.
4007 	 */
4008 	/* Will need ifdefs around this */
4009 	struct mbuf *newm;
4010 	struct sctphdr *sctphdr;
4011 	int packet_length;
4012 	int ret;
4013 #if defined(INET) || defined(INET6)
4014 	uint32_t vrf_id;
4015 #endif
4016 #if defined(INET) || defined(INET6)
4017 	struct mbuf *o_pak;
4018 	sctp_route_t *ro = NULL;
4019 	struct udphdr *udp = NULL;
4020 #endif
4021 	uint8_t tos_value;
4022 #if defined(__APPLE__) || defined(SCTP_SO_LOCK_TESTING)
4023 	struct socket *so = NULL;
4024 #endif
4025 
4026 	if ((net) && (net->dest_state & SCTP_ADDR_OUT_OF_SCOPE)) {
4027 		SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EFAULT);
4028 		sctp_m_freem(m);
4029 		return (EFAULT);
4030 	}
4031 #if defined(INET) || defined(INET6)
4032 	if (stcb) {
4033 		vrf_id = stcb->asoc.vrf_id;
4034 	} else {
4035 		vrf_id = inp->def_vrf_id;
4036 	}
4037 #endif
4038 	/* fill in the HMAC digest for any AUTH chunk in the packet */
4039 	if ((auth != NULL) && (stcb != NULL)) {
4040 		sctp_fill_hmac_digest_m(m, auth_offset, auth, stcb, auth_keyid);
4041 	}
4042 
4043 	if (net) {
4044 		tos_value = net->dscp;
4045 	} else if (stcb) {
4046 		tos_value = stcb->asoc.default_dscp;
4047 	} else {
4048 		tos_value = inp->sctp_ep.default_dscp;
4049 	}
4050 
4051 	switch (to->sa_family) {
4052 #ifdef INET
4053 	case AF_INET:
4054 		{
4055 			struct ip *ip = NULL;
4056 			sctp_route_t iproute;
4057 			int len;
4058 
4059 			len = SCTP_MIN_V4_OVERHEAD;
4060 			if (port) {
4061 				len += sizeof(struct udphdr);
4062 			}
4063 			newm = sctp_get_mbuf_for_msg(len, 1, M_NOWAIT, 1, MT_DATA);
4064 			if (newm == NULL) {
4065 				sctp_m_freem(m);
4066 				SCTP_LTRACE_ERR_RET(inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
4067 				return (ENOMEM);
4068 			}
4069 			SCTP_ALIGN_TO_END(newm, len);
4070 			SCTP_BUF_LEN(newm) = len;
4071 			SCTP_BUF_NEXT(newm) = m;
4072 			m = newm;
4073 			if (net != NULL) {
4074 				m->m_pkthdr.flowid = net->flowid;
4075 				M_HASHTYPE_SET(m, net->flowtype);
4076 			} else {
4077 				m->m_pkthdr.flowid = mflowid;
4078 				M_HASHTYPE_SET(m, mflowtype);
4079 			}
4080 			packet_length = sctp_calculate_len(m);
4081 			ip = mtod(m, struct ip *);
4082 			ip->ip_v = IPVERSION;
4083 			ip->ip_hl = (sizeof(struct ip) >> 2);
4084 			if (tos_value == 0) {
4085 				/*
4086 				 * This means especially, that it is not set
4087 				 * at the SCTP layer. So use the value from
4088 				 * the IP layer.
4089 				 */
4090 				tos_value = inp->ip_inp.inp.inp_ip_tos;
4091 			}
4092 			tos_value &= 0xfc;
4093 			if (ecn_ok) {
4094 				tos_value |= sctp_get_ect(stcb);
4095 			}
4096 			if ((nofragment_flag) && (port == 0)) {
4097 				ip->ip_off = htons(IP_DF);
4098 			} else {
4099 				ip->ip_off = htons(0);
4100 			}
4101 			/* FreeBSD has a function for ip_id's */
4102 			ip_fillid(ip);
4103 
4104 			ip->ip_ttl = inp->ip_inp.inp.inp_ip_ttl;
4105 			ip->ip_len = htons(packet_length);
4106 			ip->ip_tos = tos_value;
4107 			if (port) {
4108 				ip->ip_p = IPPROTO_UDP;
4109 			} else {
4110 				ip->ip_p = IPPROTO_SCTP;
4111 			}
4112 			ip->ip_sum = 0;
4113 			if (net == NULL) {
4114 				ro = &iproute;
4115 				memset(&iproute, 0, sizeof(iproute));
4116 				memcpy(&ro->ro_dst, to, to->sa_len);
4117 			} else {
4118 				ro = (sctp_route_t *)&net->ro;
4119 			}
4120 			/* Now the address selection part */
4121 			ip->ip_dst.s_addr = ((struct sockaddr_in *)to)->sin_addr.s_addr;
4122 
4123 			/* call the routine to select the src address */
4124 			if (net && out_of_asoc_ok == 0) {
4125 				if (net->ro._s_addr && (net->ro._s_addr->localifa_flags & (SCTP_BEING_DELETED | SCTP_ADDR_IFA_UNUSEABLE))) {
4126 					sctp_free_ifa(net->ro._s_addr);
4127 					net->ro._s_addr = NULL;
4128 					net->src_addr_selected = 0;
4129 					if (ro->ro_rt) {
4130 						RTFREE(ro->ro_rt);
4131 						ro->ro_rt = NULL;
4132 					}
4133 				}
4134 				if (net->src_addr_selected == 0) {
4135 					/* Cache the source address */
4136 					net->ro._s_addr = sctp_source_address_selection(inp, stcb,
4137 					    ro, net, 0,
4138 					    vrf_id);
4139 					net->src_addr_selected = 1;
4140 				}
4141 				if (net->ro._s_addr == NULL) {
4142 					/* No route to host */
4143 					net->src_addr_selected = 0;
4144 					sctp_handle_no_route(stcb, net, so_locked);
4145 					SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, EHOSTUNREACH);
4146 					sctp_m_freem(m);
4147 					return (EHOSTUNREACH);
4148 				}
4149 				ip->ip_src = net->ro._s_addr->address.sin.sin_addr;
4150 			} else {
4151 				if (over_addr == NULL) {
4152 					struct sctp_ifa *_lsrc;
4153 
4154 					_lsrc = sctp_source_address_selection(inp, stcb, ro,
4155 					    net,
4156 					    out_of_asoc_ok,
4157 					    vrf_id);
4158 					if (_lsrc == NULL) {
4159 						sctp_handle_no_route(stcb, net, so_locked);
4160 						SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, EHOSTUNREACH);
4161 						sctp_m_freem(m);
4162 						return (EHOSTUNREACH);
4163 					}
4164 					ip->ip_src = _lsrc->address.sin.sin_addr;
4165 					sctp_free_ifa(_lsrc);
4166 				} else {
4167 					ip->ip_src = over_addr->sin.sin_addr;
4168 					SCTP_RTALLOC(ro, vrf_id, inp->fibnum);
4169 				}
4170 			}
4171 			if (port) {
4172 				if (htons(SCTP_BASE_SYSCTL(sctp_udp_tunneling_port)) == 0) {
4173 					sctp_handle_no_route(stcb, net, so_locked);
4174 					SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, EHOSTUNREACH);
4175 					sctp_m_freem(m);
4176 					return (EHOSTUNREACH);
4177 				}
4178 				udp = (struct udphdr *)((caddr_t)ip + sizeof(struct ip));
4179 				udp->uh_sport = htons(SCTP_BASE_SYSCTL(sctp_udp_tunneling_port));
4180 				udp->uh_dport = port;
4181 				udp->uh_ulen = htons((uint16_t)(packet_length - sizeof(struct ip)));
4182 				if (V_udp_cksum) {
4183 					udp->uh_sum = in_pseudo(ip->ip_src.s_addr, ip->ip_dst.s_addr, udp->uh_ulen + htons(IPPROTO_UDP));
4184 				} else {
4185 					udp->uh_sum = 0;
4186 				}
4187 				sctphdr = (struct sctphdr *)((caddr_t)udp + sizeof(struct udphdr));
4188 			} else {
4189 				sctphdr = (struct sctphdr *)((caddr_t)ip + sizeof(struct ip));
4190 			}
4191 
4192 			sctphdr->src_port = src_port;
4193 			sctphdr->dest_port = dest_port;
4194 			sctphdr->v_tag = v_tag;
4195 			sctphdr->checksum = 0;
4196 
4197 			/*
4198 			 * If source address selection fails and we find no
4199 			 * route then the ip_output should fail as well with
4200 			 * a NO_ROUTE_TO_HOST type error. We probably should
4201 			 * catch that somewhere and abort the association
4202 			 * right away (assuming this is an INIT being sent).
4203 			 */
4204 			if (ro->ro_rt == NULL) {
4205 				/*
4206 				 * src addr selection failed to find a route
4207 				 * (or valid source addr), so we can't get
4208 				 * there from here (yet)!
4209 				 */
4210 				sctp_handle_no_route(stcb, net, so_locked);
4211 				SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, EHOSTUNREACH);
4212 				sctp_m_freem(m);
4213 				return (EHOSTUNREACH);
4214 			}
4215 			if (ro != &iproute) {
4216 				memcpy(&iproute, ro, sizeof(*ro));
4217 			}
4218 			SCTPDBG(SCTP_DEBUG_OUTPUT3, "Calling ipv4 output routine from low level src addr:%x\n",
4219 			    (uint32_t)(ntohl(ip->ip_src.s_addr)));
4220 			SCTPDBG(SCTP_DEBUG_OUTPUT3, "Destination is %x\n",
4221 			    (uint32_t)(ntohl(ip->ip_dst.s_addr)));
4222 			SCTPDBG(SCTP_DEBUG_OUTPUT3, "RTP route is %p through\n",
4223 			    (void *)ro->ro_rt);
4224 
4225 			if (SCTP_GET_HEADER_FOR_OUTPUT(o_pak)) {
4226 				/* failed to prepend data, give up */
4227 				SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
4228 				sctp_m_freem(m);
4229 				return (ENOMEM);
4230 			}
4231 			SCTP_ATTACH_CHAIN(o_pak, m, packet_length);
4232 			if (port) {
4233 				sctphdr->checksum = sctp_calculate_cksum(m, sizeof(struct ip) + sizeof(struct udphdr));
4234 				SCTP_STAT_INCR(sctps_sendswcrc);
4235 				if (V_udp_cksum) {
4236 					SCTP_ENABLE_UDP_CSUM(o_pak);
4237 				}
4238 			} else {
4239 				m->m_pkthdr.csum_flags = CSUM_SCTP;
4240 				m->m_pkthdr.csum_data = offsetof(struct sctphdr, checksum);
4241 				SCTP_STAT_INCR(sctps_sendhwcrc);
4242 			}
4243 #ifdef SCTP_PACKET_LOGGING
4244 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LAST_PACKET_TRACING)
4245 				sctp_packet_log(o_pak);
4246 #endif
4247 			/* send it out.  table id is taken from stcb */
4248 #if defined(__APPLE__) || defined(SCTP_SO_LOCK_TESTING)
4249 			if ((SCTP_BASE_SYSCTL(sctp_output_unlocked)) && (so_locked)) {
4250 				so = SCTP_INP_SO(inp);
4251 				SCTP_SOCKET_UNLOCK(so, 0);
4252 			}
4253 #endif
4254 			SCTP_IP_OUTPUT(ret, o_pak, ro, stcb, vrf_id);
4255 #if defined(__APPLE__) || defined(SCTP_SO_LOCK_TESTING)
4256 			if ((SCTP_BASE_SYSCTL(sctp_output_unlocked)) && (so_locked)) {
4257 				atomic_add_int(&stcb->asoc.refcnt, 1);
4258 				SCTP_TCB_UNLOCK(stcb);
4259 				SCTP_SOCKET_LOCK(so, 0);
4260 				SCTP_TCB_LOCK(stcb);
4261 				atomic_subtract_int(&stcb->asoc.refcnt, 1);
4262 			}
4263 #endif
4264 			SCTP_STAT_INCR(sctps_sendpackets);
4265 			SCTP_STAT_INCR_COUNTER64(sctps_outpackets);
4266 			if (ret)
4267 				SCTP_STAT_INCR(sctps_senderrors);
4268 
4269 			SCTPDBG(SCTP_DEBUG_OUTPUT3, "IP output returns %d\n", ret);
4270 			if (net == NULL) {
4271 				/* free tempy routes */
4272 				RO_RTFREE(ro);
4273 			} else {
4274 				if ((ro->ro_rt != NULL) && (net->ro._s_addr) &&
4275 				    ((net->dest_state & SCTP_ADDR_NO_PMTUD) == 0)) {
4276 					uint32_t mtu;
4277 
4278 					mtu = SCTP_GATHER_MTU_FROM_ROUTE(net->ro._s_addr, &net->ro._l_addr.sa, ro->ro_rt);
4279 					if (mtu > 0) {
4280 						if (net->port) {
4281 							mtu -= sizeof(struct udphdr);
4282 						}
4283 						if ((stcb != NULL) && (stcb->asoc.smallest_mtu > mtu)) {
4284 							sctp_mtu_size_reset(inp, &stcb->asoc, mtu);
4285 						}
4286 						net->mtu = mtu;
4287 					}
4288 				} else if (ro->ro_rt == NULL) {
4289 					/* route was freed */
4290 					if (net->ro._s_addr &&
4291 					    net->src_addr_selected) {
4292 						sctp_free_ifa(net->ro._s_addr);
4293 						net->ro._s_addr = NULL;
4294 					}
4295 					net->src_addr_selected = 0;
4296 				}
4297 			}
4298 			return (ret);
4299 		}
4300 #endif
4301 #ifdef INET6
4302 	case AF_INET6:
4303 		{
4304 			uint32_t flowlabel, flowinfo;
4305 			struct ip6_hdr *ip6h;
4306 			struct route_in6 ip6route;
4307 			struct ifnet *ifp;
4308 			struct sockaddr_in6 *sin6, tmp, *lsa6, lsa6_tmp;
4309 			int prev_scope = 0;
4310 			struct sockaddr_in6 lsa6_storage;
4311 			int error;
4312 			u_short prev_port = 0;
4313 			int len;
4314 
4315 			if (net) {
4316 				flowlabel = net->flowlabel;
4317 			} else if (stcb) {
4318 				flowlabel = stcb->asoc.default_flowlabel;
4319 			} else {
4320 				flowlabel = inp->sctp_ep.default_flowlabel;
4321 			}
4322 			if (flowlabel == 0) {
4323 				/*
4324 				 * This means especially, that it is not set
4325 				 * at the SCTP layer. So use the value from
4326 				 * the IP layer.
4327 				 */
4328 				flowlabel = ntohl(((struct in6pcb *)inp)->in6p_flowinfo);
4329 			}
4330 			flowlabel &= 0x000fffff;
4331 			len = SCTP_MIN_OVERHEAD;
4332 			if (port) {
4333 				len += sizeof(struct udphdr);
4334 			}
4335 			newm = sctp_get_mbuf_for_msg(len, 1, M_NOWAIT, 1, MT_DATA);
4336 			if (newm == NULL) {
4337 				sctp_m_freem(m);
4338 				SCTP_LTRACE_ERR_RET(inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
4339 				return (ENOMEM);
4340 			}
4341 			SCTP_ALIGN_TO_END(newm, len);
4342 			SCTP_BUF_LEN(newm) = len;
4343 			SCTP_BUF_NEXT(newm) = m;
4344 			m = newm;
4345 			if (net != NULL) {
4346 				m->m_pkthdr.flowid = net->flowid;
4347 				M_HASHTYPE_SET(m, net->flowtype);
4348 			} else {
4349 				m->m_pkthdr.flowid = mflowid;
4350 				M_HASHTYPE_SET(m, mflowtype);
4351 			}
4352 			packet_length = sctp_calculate_len(m);
4353 
4354 			ip6h = mtod(m, struct ip6_hdr *);
4355 			/* protect *sin6 from overwrite */
4356 			sin6 = (struct sockaddr_in6 *)to;
4357 			tmp = *sin6;
4358 			sin6 = &tmp;
4359 
4360 			/* KAME hack: embed scopeid */
4361 			if (sa6_embedscope(sin6, MODULE_GLOBAL(ip6_use_defzone)) != 0) {
4362 				SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
4363 				return (EINVAL);
4364 			}
4365 			if (net == NULL) {
4366 				memset(&ip6route, 0, sizeof(ip6route));
4367 				ro = (sctp_route_t *)&ip6route;
4368 				memcpy(&ro->ro_dst, sin6, sin6->sin6_len);
4369 			} else {
4370 				ro = (sctp_route_t *)&net->ro;
4371 			}
4372 			/*
4373 			 * We assume here that inp_flow is in host byte
4374 			 * order within the TCB!
4375 			 */
4376 			if (tos_value == 0) {
4377 				/*
4378 				 * This means especially, that it is not set
4379 				 * at the SCTP layer. So use the value from
4380 				 * the IP layer.
4381 				 */
4382 				tos_value = (ntohl(((struct in6pcb *)inp)->in6p_flowinfo) >> 20) & 0xff;
4383 			}
4384 			tos_value &= 0xfc;
4385 			if (ecn_ok) {
4386 				tos_value |= sctp_get_ect(stcb);
4387 			}
4388 			flowinfo = 0x06;
4389 			flowinfo <<= 8;
4390 			flowinfo |= tos_value;
4391 			flowinfo <<= 20;
4392 			flowinfo |= flowlabel;
4393 			ip6h->ip6_flow = htonl(flowinfo);
4394 			if (port) {
4395 				ip6h->ip6_nxt = IPPROTO_UDP;
4396 			} else {
4397 				ip6h->ip6_nxt = IPPROTO_SCTP;
4398 			}
4399 			ip6h->ip6_plen = htons((uint16_t)(packet_length - sizeof(struct ip6_hdr)));
4400 			ip6h->ip6_dst = sin6->sin6_addr;
4401 
4402 			/*
4403 			 * Add SRC address selection here: we can only reuse
4404 			 * to a limited degree the kame src-addr-sel, since
4405 			 * we can try their selection but it may not be
4406 			 * bound.
4407 			 */
4408 			memset(&lsa6_tmp, 0, sizeof(lsa6_tmp));
4409 			lsa6_tmp.sin6_family = AF_INET6;
4410 			lsa6_tmp.sin6_len = sizeof(lsa6_tmp);
4411 			lsa6 = &lsa6_tmp;
4412 			if (net && out_of_asoc_ok == 0) {
4413 				if (net->ro._s_addr && (net->ro._s_addr->localifa_flags & (SCTP_BEING_DELETED | SCTP_ADDR_IFA_UNUSEABLE))) {
4414 					sctp_free_ifa(net->ro._s_addr);
4415 					net->ro._s_addr = NULL;
4416 					net->src_addr_selected = 0;
4417 					if (ro->ro_rt) {
4418 						RTFREE(ro->ro_rt);
4419 						ro->ro_rt = NULL;
4420 					}
4421 				}
4422 				if (net->src_addr_selected == 0) {
4423 					sin6 = (struct sockaddr_in6 *)&net->ro._l_addr;
4424 					/* KAME hack: embed scopeid */
4425 					if (sa6_embedscope(sin6, MODULE_GLOBAL(ip6_use_defzone)) != 0) {
4426 						SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
4427 						return (EINVAL);
4428 					}
4429 					/* Cache the source address */
4430 					net->ro._s_addr = sctp_source_address_selection(inp,
4431 					    stcb,
4432 					    ro,
4433 					    net,
4434 					    0,
4435 					    vrf_id);
4436 					(void)sa6_recoverscope(sin6);
4437 					net->src_addr_selected = 1;
4438 				}
4439 				if (net->ro._s_addr == NULL) {
4440 					SCTPDBG(SCTP_DEBUG_OUTPUT3, "V6:No route to host\n");
4441 					net->src_addr_selected = 0;
4442 					sctp_handle_no_route(stcb, net, so_locked);
4443 					SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, EHOSTUNREACH);
4444 					sctp_m_freem(m);
4445 					return (EHOSTUNREACH);
4446 				}
4447 				lsa6->sin6_addr = net->ro._s_addr->address.sin6.sin6_addr;
4448 			} else {
4449 				sin6 = (struct sockaddr_in6 *)&ro->ro_dst;
4450 				/* KAME hack: embed scopeid */
4451 				if (sa6_embedscope(sin6, MODULE_GLOBAL(ip6_use_defzone)) != 0) {
4452 					SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
4453 					return (EINVAL);
4454 				}
4455 				if (over_addr == NULL) {
4456 					struct sctp_ifa *_lsrc;
4457 
4458 					_lsrc = sctp_source_address_selection(inp, stcb, ro,
4459 					    net,
4460 					    out_of_asoc_ok,
4461 					    vrf_id);
4462 					if (_lsrc == NULL) {
4463 						sctp_handle_no_route(stcb, net, so_locked);
4464 						SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, EHOSTUNREACH);
4465 						sctp_m_freem(m);
4466 						return (EHOSTUNREACH);
4467 					}
4468 					lsa6->sin6_addr = _lsrc->address.sin6.sin6_addr;
4469 					sctp_free_ifa(_lsrc);
4470 				} else {
4471 					lsa6->sin6_addr = over_addr->sin6.sin6_addr;
4472 					SCTP_RTALLOC(ro, vrf_id, inp->fibnum);
4473 				}
4474 				(void)sa6_recoverscope(sin6);
4475 			}
4476 			lsa6->sin6_port = inp->sctp_lport;
4477 
4478 			if (ro->ro_rt == NULL) {
4479 				/*
4480 				 * src addr selection failed to find a route
4481 				 * (or valid source addr), so we can't get
4482 				 * there from here!
4483 				 */
4484 				sctp_handle_no_route(stcb, net, so_locked);
4485 				SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, EHOSTUNREACH);
4486 				sctp_m_freem(m);
4487 				return (EHOSTUNREACH);
4488 			}
4489 			/*
4490 			 * XXX: sa6 may not have a valid sin6_scope_id in
4491 			 * the non-SCOPEDROUTING case.
4492 			 */
4493 			memset(&lsa6_storage, 0, sizeof(lsa6_storage));
4494 			lsa6_storage.sin6_family = AF_INET6;
4495 			lsa6_storage.sin6_len = sizeof(lsa6_storage);
4496 			lsa6_storage.sin6_addr = lsa6->sin6_addr;
4497 			if ((error = sa6_recoverscope(&lsa6_storage)) != 0) {
4498 				SCTPDBG(SCTP_DEBUG_OUTPUT3, "recover scope fails error %d\n", error);
4499 				sctp_m_freem(m);
4500 				return (error);
4501 			}
4502 			/* XXX */
4503 			lsa6_storage.sin6_addr = lsa6->sin6_addr;
4504 			lsa6_storage.sin6_port = inp->sctp_lport;
4505 			lsa6 = &lsa6_storage;
4506 			ip6h->ip6_src = lsa6->sin6_addr;
4507 
4508 			if (port) {
4509 				if (htons(SCTP_BASE_SYSCTL(sctp_udp_tunneling_port)) == 0) {
4510 					sctp_handle_no_route(stcb, net, so_locked);
4511 					SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, EHOSTUNREACH);
4512 					sctp_m_freem(m);
4513 					return (EHOSTUNREACH);
4514 				}
4515 				udp = (struct udphdr *)((caddr_t)ip6h + sizeof(struct ip6_hdr));
4516 				udp->uh_sport = htons(SCTP_BASE_SYSCTL(sctp_udp_tunneling_port));
4517 				udp->uh_dport = port;
4518 				udp->uh_ulen = htons((uint16_t)(packet_length - sizeof(struct ip6_hdr)));
4519 				udp->uh_sum = 0;
4520 				sctphdr = (struct sctphdr *)((caddr_t)udp + sizeof(struct udphdr));
4521 			} else {
4522 				sctphdr = (struct sctphdr *)((caddr_t)ip6h + sizeof(struct ip6_hdr));
4523 			}
4524 
4525 			sctphdr->src_port = src_port;
4526 			sctphdr->dest_port = dest_port;
4527 			sctphdr->v_tag = v_tag;
4528 			sctphdr->checksum = 0;
4529 
4530 			/*
4531 			 * We set the hop limit now since there is a good
4532 			 * chance that our ro pointer is now filled
4533 			 */
4534 			ip6h->ip6_hlim = SCTP_GET_HLIM(inp, ro);
4535 			ifp = SCTP_GET_IFN_VOID_FROM_ROUTE(ro);
4536 
4537 #ifdef SCTP_DEBUG
4538 			/* Copy to be sure something bad is not happening */
4539 			sin6->sin6_addr = ip6h->ip6_dst;
4540 			lsa6->sin6_addr = ip6h->ip6_src;
4541 #endif
4542 
4543 			SCTPDBG(SCTP_DEBUG_OUTPUT3, "Calling ipv6 output routine from low level\n");
4544 			SCTPDBG(SCTP_DEBUG_OUTPUT3, "src: ");
4545 			SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT3, (struct sockaddr *)lsa6);
4546 			SCTPDBG(SCTP_DEBUG_OUTPUT3, "dst: ");
4547 			SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT3, (struct sockaddr *)sin6);
4548 			if (net) {
4549 				sin6 = (struct sockaddr_in6 *)&net->ro._l_addr;
4550 				/*
4551 				 * preserve the port and scope for link
4552 				 * local send
4553 				 */
4554 				prev_scope = sin6->sin6_scope_id;
4555 				prev_port = sin6->sin6_port;
4556 			}
4557 
4558 			if (SCTP_GET_HEADER_FOR_OUTPUT(o_pak)) {
4559 				/* failed to prepend data, give up */
4560 				sctp_m_freem(m);
4561 				SCTP_LTRACE_ERR_RET(inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
4562 				return (ENOMEM);
4563 			}
4564 			SCTP_ATTACH_CHAIN(o_pak, m, packet_length);
4565 			if (port) {
4566 				sctphdr->checksum = sctp_calculate_cksum(m, sizeof(struct ip6_hdr) + sizeof(struct udphdr));
4567 				SCTP_STAT_INCR(sctps_sendswcrc);
4568 				if ((udp->uh_sum = in6_cksum(o_pak, IPPROTO_UDP, sizeof(struct ip6_hdr), packet_length - sizeof(struct ip6_hdr))) == 0) {
4569 					udp->uh_sum = 0xffff;
4570 				}
4571 			} else {
4572 				m->m_pkthdr.csum_flags = CSUM_SCTP_IPV6;
4573 				m->m_pkthdr.csum_data = offsetof(struct sctphdr, checksum);
4574 				SCTP_STAT_INCR(sctps_sendhwcrc);
4575 			}
4576 			/* send it out. table id is taken from stcb */
4577 #if defined(__APPLE__) || defined(SCTP_SO_LOCK_TESTING)
4578 			if ((SCTP_BASE_SYSCTL(sctp_output_unlocked)) && (so_locked)) {
4579 				so = SCTP_INP_SO(inp);
4580 				SCTP_SOCKET_UNLOCK(so, 0);
4581 			}
4582 #endif
4583 #ifdef SCTP_PACKET_LOGGING
4584 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LAST_PACKET_TRACING)
4585 				sctp_packet_log(o_pak);
4586 #endif
4587 			SCTP_IP6_OUTPUT(ret, o_pak, (struct route_in6 *)ro, &ifp, stcb, vrf_id);
4588 #if defined(__APPLE__) || defined(SCTP_SO_LOCK_TESTING)
4589 			if ((SCTP_BASE_SYSCTL(sctp_output_unlocked)) && (so_locked)) {
4590 				atomic_add_int(&stcb->asoc.refcnt, 1);
4591 				SCTP_TCB_UNLOCK(stcb);
4592 				SCTP_SOCKET_LOCK(so, 0);
4593 				SCTP_TCB_LOCK(stcb);
4594 				atomic_subtract_int(&stcb->asoc.refcnt, 1);
4595 			}
4596 #endif
4597 			if (net) {
4598 				/* for link local this must be done */
4599 				sin6->sin6_scope_id = prev_scope;
4600 				sin6->sin6_port = prev_port;
4601 			}
4602 			SCTPDBG(SCTP_DEBUG_OUTPUT3, "return from send is %d\n", ret);
4603 			SCTP_STAT_INCR(sctps_sendpackets);
4604 			SCTP_STAT_INCR_COUNTER64(sctps_outpackets);
4605 			if (ret) {
4606 				SCTP_STAT_INCR(sctps_senderrors);
4607 			}
4608 			if (net == NULL) {
4609 				/* Now if we had a temp route free it */
4610 				RO_RTFREE(ro);
4611 			} else {
4612 				/*
4613 				 * PMTU check versus smallest asoc MTU goes
4614 				 * here
4615 				 */
4616 				if (ro->ro_rt == NULL) {
4617 					/* Route was freed */
4618 					if (net->ro._s_addr &&
4619 					    net->src_addr_selected) {
4620 						sctp_free_ifa(net->ro._s_addr);
4621 						net->ro._s_addr = NULL;
4622 					}
4623 					net->src_addr_selected = 0;
4624 				}
4625 				if ((ro->ro_rt != NULL) && (net->ro._s_addr) &&
4626 				    ((net->dest_state & SCTP_ADDR_NO_PMTUD) == 0)) {
4627 					uint32_t mtu;
4628 
4629 					mtu = SCTP_GATHER_MTU_FROM_ROUTE(net->ro._s_addr, &net->ro._l_addr.sa, ro->ro_rt);
4630 					if (mtu > 0) {
4631 						if (net->port) {
4632 							mtu -= sizeof(struct udphdr);
4633 						}
4634 						if ((stcb != NULL) && (stcb->asoc.smallest_mtu > mtu)) {
4635 							sctp_mtu_size_reset(inp, &stcb->asoc, mtu);
4636 						}
4637 						net->mtu = mtu;
4638 					}
4639 				} else if (ifp) {
4640 					if (ND_IFINFO(ifp)->linkmtu &&
4641 					    (stcb->asoc.smallest_mtu > ND_IFINFO(ifp)->linkmtu)) {
4642 						sctp_mtu_size_reset(inp,
4643 						    &stcb->asoc,
4644 						    ND_IFINFO(ifp)->linkmtu);
4645 					}
4646 				}
4647 			}
4648 			return (ret);
4649 		}
4650 #endif
4651 	default:
4652 		SCTPDBG(SCTP_DEBUG_OUTPUT1, "Unknown protocol (TSNH) type %d\n",
4653 		    ((struct sockaddr *)to)->sa_family);
4654 		sctp_m_freem(m);
4655 		SCTP_LTRACE_ERR_RET_PKT(m, inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EFAULT);
4656 		return (EFAULT);
4657 	}
4658 }
4659 
4660 
4661 void
4662 sctp_send_initiate(struct sctp_inpcb *inp, struct sctp_tcb *stcb, int so_locked
4663 #if !defined(__APPLE__) && !defined(SCTP_SO_LOCK_TESTING)
4664     SCTP_UNUSED
4665 #endif
4666 )
4667 {
4668 	struct mbuf *m, *m_last;
4669 	struct sctp_nets *net;
4670 	struct sctp_init_chunk *init;
4671 	struct sctp_supported_addr_param *sup_addr;
4672 	struct sctp_adaptation_layer_indication *ali;
4673 	struct sctp_supported_chunk_types_param *pr_supported;
4674 	struct sctp_paramhdr *ph;
4675 	int cnt_inits_to = 0;
4676 	int error;
4677 	uint16_t num_ext, chunk_len, padding_len, parameter_len;
4678 
4679 	/* INIT's always go to the primary (and usually ONLY address) */
4680 	net = stcb->asoc.primary_destination;
4681 	if (net == NULL) {
4682 		net = TAILQ_FIRST(&stcb->asoc.nets);
4683 		if (net == NULL) {
4684 			/* TSNH */
4685 			return;
4686 		}
4687 		/* we confirm any address we send an INIT to */
4688 		net->dest_state &= ~SCTP_ADDR_UNCONFIRMED;
4689 		(void)sctp_set_primary_addr(stcb, NULL, net);
4690 	} else {
4691 		/* we confirm any address we send an INIT to */
4692 		net->dest_state &= ~SCTP_ADDR_UNCONFIRMED;
4693 	}
4694 	SCTPDBG(SCTP_DEBUG_OUTPUT4, "Sending INIT\n");
4695 #ifdef INET6
4696 	if (net->ro._l_addr.sa.sa_family == AF_INET6) {
4697 		/*
4698 		 * special hook, if we are sending to link local it will not
4699 		 * show up in our private address count.
4700 		 */
4701 		if (IN6_IS_ADDR_LINKLOCAL(&net->ro._l_addr.sin6.sin6_addr))
4702 			cnt_inits_to = 1;
4703 	}
4704 #endif
4705 	if (SCTP_OS_TIMER_PENDING(&net->rxt_timer.timer)) {
4706 		/* This case should not happen */
4707 		SCTPDBG(SCTP_DEBUG_OUTPUT4, "Sending INIT - failed timer?\n");
4708 		return;
4709 	}
4710 	/* start the INIT timer */
4711 	sctp_timer_start(SCTP_TIMER_TYPE_INIT, inp, stcb, net);
4712 
4713 	m = sctp_get_mbuf_for_msg(MCLBYTES, 1, M_NOWAIT, 1, MT_DATA);
4714 	if (m == NULL) {
4715 		/* No memory, INIT timer will re-attempt. */
4716 		SCTPDBG(SCTP_DEBUG_OUTPUT4, "Sending INIT - mbuf?\n");
4717 		return;
4718 	}
4719 	chunk_len = (uint16_t)sizeof(struct sctp_init_chunk);
4720 	padding_len = 0;
4721 	/* Now lets put the chunk header in place */
4722 	init = mtod(m, struct sctp_init_chunk *);
4723 	/* now the chunk header */
4724 	init->ch.chunk_type = SCTP_INITIATION;
4725 	init->ch.chunk_flags = 0;
4726 	/* fill in later from mbuf we build */
4727 	init->ch.chunk_length = 0;
4728 	/* place in my tag */
4729 	init->init.initiate_tag = htonl(stcb->asoc.my_vtag);
4730 	/* set up some of the credits. */
4731 	init->init.a_rwnd = htonl(max(inp->sctp_socket ? SCTP_SB_LIMIT_RCV(inp->sctp_socket) : 0,
4732 	    SCTP_MINIMAL_RWND));
4733 	init->init.num_outbound_streams = htons(stcb->asoc.pre_open_streams);
4734 	init->init.num_inbound_streams = htons(stcb->asoc.max_inbound_streams);
4735 	init->init.initial_tsn = htonl(stcb->asoc.init_seq_number);
4736 
4737 	/* Adaptation layer indication parameter */
4738 	if (inp->sctp_ep.adaptation_layer_indicator_provided) {
4739 		parameter_len = (uint16_t)sizeof(struct sctp_adaptation_layer_indication);
4740 		ali = (struct sctp_adaptation_layer_indication *)(mtod(m, caddr_t)+chunk_len);
4741 		ali->ph.param_type = htons(SCTP_ULP_ADAPTATION);
4742 		ali->ph.param_length = htons(parameter_len);
4743 		ali->indication = htonl(inp->sctp_ep.adaptation_layer_indicator);
4744 		chunk_len += parameter_len;
4745 	}
4746 
4747 	/* ECN parameter */
4748 	if (stcb->asoc.ecn_supported == 1) {
4749 		parameter_len = (uint16_t)sizeof(struct sctp_paramhdr);
4750 		ph = (struct sctp_paramhdr *)(mtod(m, caddr_t)+chunk_len);
4751 		ph->param_type = htons(SCTP_ECN_CAPABLE);
4752 		ph->param_length = htons(parameter_len);
4753 		chunk_len += parameter_len;
4754 	}
4755 
4756 	/* PR-SCTP supported parameter */
4757 	if (stcb->asoc.prsctp_supported == 1) {
4758 		parameter_len = (uint16_t)sizeof(struct sctp_paramhdr);
4759 		ph = (struct sctp_paramhdr *)(mtod(m, caddr_t)+chunk_len);
4760 		ph->param_type = htons(SCTP_PRSCTP_SUPPORTED);
4761 		ph->param_length = htons(parameter_len);
4762 		chunk_len += parameter_len;
4763 	}
4764 
4765 	/* Add NAT friendly parameter. */
4766 	if (SCTP_BASE_SYSCTL(sctp_inits_include_nat_friendly)) {
4767 		parameter_len = (uint16_t)sizeof(struct sctp_paramhdr);
4768 		ph = (struct sctp_paramhdr *)(mtod(m, caddr_t)+chunk_len);
4769 		ph->param_type = htons(SCTP_HAS_NAT_SUPPORT);
4770 		ph->param_length = htons(parameter_len);
4771 		chunk_len += parameter_len;
4772 	}
4773 
4774 	/* And now tell the peer which extensions we support */
4775 	num_ext = 0;
4776 	pr_supported = (struct sctp_supported_chunk_types_param *)(mtod(m, caddr_t)+chunk_len);
4777 	if (stcb->asoc.prsctp_supported == 1) {
4778 		pr_supported->chunk_types[num_ext++] = SCTP_FORWARD_CUM_TSN;
4779 		if (stcb->asoc.idata_supported) {
4780 			pr_supported->chunk_types[num_ext++] = SCTP_IFORWARD_CUM_TSN;
4781 		}
4782 	}
4783 	if (stcb->asoc.auth_supported == 1) {
4784 		pr_supported->chunk_types[num_ext++] = SCTP_AUTHENTICATION;
4785 	}
4786 	if (stcb->asoc.asconf_supported == 1) {
4787 		pr_supported->chunk_types[num_ext++] = SCTP_ASCONF;
4788 		pr_supported->chunk_types[num_ext++] = SCTP_ASCONF_ACK;
4789 	}
4790 	if (stcb->asoc.reconfig_supported == 1) {
4791 		pr_supported->chunk_types[num_ext++] = SCTP_STREAM_RESET;
4792 	}
4793 	if (stcb->asoc.idata_supported) {
4794 		pr_supported->chunk_types[num_ext++] = SCTP_IDATA;
4795 	}
4796 	if (stcb->asoc.nrsack_supported == 1) {
4797 		pr_supported->chunk_types[num_ext++] = SCTP_NR_SELECTIVE_ACK;
4798 	}
4799 	if (stcb->asoc.pktdrop_supported == 1) {
4800 		pr_supported->chunk_types[num_ext++] = SCTP_PACKET_DROPPED;
4801 	}
4802 	if (num_ext > 0) {
4803 		parameter_len = (uint16_t)sizeof(struct sctp_supported_chunk_types_param) + num_ext;
4804 		pr_supported->ph.param_type = htons(SCTP_SUPPORTED_CHUNK_EXT);
4805 		pr_supported->ph.param_length = htons(parameter_len);
4806 		padding_len = SCTP_SIZE32(parameter_len) - parameter_len;
4807 		chunk_len += parameter_len;
4808 	}
4809 	/* add authentication parameters */
4810 	if (stcb->asoc.auth_supported) {
4811 		/* attach RANDOM parameter, if available */
4812 		if (stcb->asoc.authinfo.random != NULL) {
4813 			struct sctp_auth_random *randp;
4814 
4815 			if (padding_len > 0) {
4816 				memset(mtod(m, caddr_t)+chunk_len, 0, padding_len);
4817 				chunk_len += padding_len;
4818 				padding_len = 0;
4819 			}
4820 			randp = (struct sctp_auth_random *)(mtod(m, caddr_t)+chunk_len);
4821 			parameter_len = (uint16_t)sizeof(struct sctp_auth_random) + stcb->asoc.authinfo.random_len;
4822 			/* random key already contains the header */
4823 			memcpy(randp, stcb->asoc.authinfo.random->key, parameter_len);
4824 			padding_len = SCTP_SIZE32(parameter_len) - parameter_len;
4825 			chunk_len += parameter_len;
4826 		}
4827 		/* add HMAC_ALGO parameter */
4828 		if (stcb->asoc.local_hmacs != NULL) {
4829 			struct sctp_auth_hmac_algo *hmacs;
4830 
4831 			if (padding_len > 0) {
4832 				memset(mtod(m, caddr_t)+chunk_len, 0, padding_len);
4833 				chunk_len += padding_len;
4834 				padding_len = 0;
4835 			}
4836 			hmacs = (struct sctp_auth_hmac_algo *)(mtod(m, caddr_t)+chunk_len);
4837 			parameter_len = (uint16_t)(sizeof(struct sctp_auth_hmac_algo) +
4838 			    stcb->asoc.local_hmacs->num_algo * sizeof(uint16_t));
4839 			hmacs->ph.param_type = htons(SCTP_HMAC_LIST);
4840 			hmacs->ph.param_length = htons(parameter_len);
4841 			sctp_serialize_hmaclist(stcb->asoc.local_hmacs, (uint8_t *)hmacs->hmac_ids);
4842 			padding_len = SCTP_SIZE32(parameter_len) - parameter_len;
4843 			chunk_len += parameter_len;
4844 		}
4845 		/* add CHUNKS parameter */
4846 		if (stcb->asoc.local_auth_chunks != NULL) {
4847 			struct sctp_auth_chunk_list *chunks;
4848 
4849 			if (padding_len > 0) {
4850 				memset(mtod(m, caddr_t)+chunk_len, 0, padding_len);
4851 				chunk_len += padding_len;
4852 				padding_len = 0;
4853 			}
4854 			chunks = (struct sctp_auth_chunk_list *)(mtod(m, caddr_t)+chunk_len);
4855 			parameter_len = (uint16_t)(sizeof(struct sctp_auth_chunk_list) +
4856 			    sctp_auth_get_chklist_size(stcb->asoc.local_auth_chunks));
4857 			chunks->ph.param_type = htons(SCTP_CHUNK_LIST);
4858 			chunks->ph.param_length = htons(parameter_len);
4859 			sctp_serialize_auth_chunks(stcb->asoc.local_auth_chunks, chunks->chunk_types);
4860 			padding_len = SCTP_SIZE32(parameter_len) - parameter_len;
4861 			chunk_len += parameter_len;
4862 		}
4863 	}
4864 
4865 	/* now any cookie time extensions */
4866 	if (stcb->asoc.cookie_preserve_req) {
4867 		struct sctp_cookie_perserve_param *cookie_preserve;
4868 
4869 		if (padding_len > 0) {
4870 			memset(mtod(m, caddr_t)+chunk_len, 0, padding_len);
4871 			chunk_len += padding_len;
4872 			padding_len = 0;
4873 		}
4874 		parameter_len = (uint16_t)sizeof(struct sctp_cookie_perserve_param);
4875 		cookie_preserve = (struct sctp_cookie_perserve_param *)(mtod(m, caddr_t)+chunk_len);
4876 		cookie_preserve->ph.param_type = htons(SCTP_COOKIE_PRESERVE);
4877 		cookie_preserve->ph.param_length = htons(parameter_len);
4878 		cookie_preserve->time = htonl(stcb->asoc.cookie_preserve_req);
4879 		stcb->asoc.cookie_preserve_req = 0;
4880 		chunk_len += parameter_len;
4881 	}
4882 
4883 	if (stcb->asoc.scope.ipv4_addr_legal || stcb->asoc.scope.ipv6_addr_legal) {
4884 		uint8_t i;
4885 
4886 		if (padding_len > 0) {
4887 			memset(mtod(m, caddr_t)+chunk_len, 0, padding_len);
4888 			chunk_len += padding_len;
4889 			padding_len = 0;
4890 		}
4891 		parameter_len = (uint16_t)sizeof(struct sctp_paramhdr);
4892 		if (stcb->asoc.scope.ipv4_addr_legal) {
4893 			parameter_len += (uint16_t)sizeof(uint16_t);
4894 		}
4895 		if (stcb->asoc.scope.ipv6_addr_legal) {
4896 			parameter_len += (uint16_t)sizeof(uint16_t);
4897 		}
4898 		sup_addr = (struct sctp_supported_addr_param *)(mtod(m, caddr_t)+chunk_len);
4899 		sup_addr->ph.param_type = htons(SCTP_SUPPORTED_ADDRTYPE);
4900 		sup_addr->ph.param_length = htons(parameter_len);
4901 		i = 0;
4902 		if (stcb->asoc.scope.ipv4_addr_legal) {
4903 			sup_addr->addr_type[i++] = htons(SCTP_IPV4_ADDRESS);
4904 		}
4905 		if (stcb->asoc.scope.ipv6_addr_legal) {
4906 			sup_addr->addr_type[i++] = htons(SCTP_IPV6_ADDRESS);
4907 		}
4908 		padding_len = 4 - 2 * i;
4909 		chunk_len += parameter_len;
4910 	}
4911 
4912 	SCTP_BUF_LEN(m) = chunk_len;
4913 	/* now the addresses */
4914 	/*
4915 	 * To optimize this we could put the scoping stuff into a structure
4916 	 * and remove the individual uint8's from the assoc structure. Then
4917 	 * we could just sifa in the address within the stcb. But for now
4918 	 * this is a quick hack to get the address stuff teased apart.
4919 	 */
4920 	m_last = sctp_add_addresses_to_i_ia(inp, stcb, &stcb->asoc.scope,
4921 	    m, cnt_inits_to,
4922 	    &padding_len, &chunk_len);
4923 
4924 	init->ch.chunk_length = htons(chunk_len);
4925 	if (padding_len > 0) {
4926 		if (sctp_add_pad_tombuf(m_last, padding_len) == NULL) {
4927 			sctp_m_freem(m);
4928 			return;
4929 		}
4930 	}
4931 	SCTPDBG(SCTP_DEBUG_OUTPUT4, "Sending INIT - calls lowlevel_output\n");
4932 	if ((error = sctp_lowlevel_chunk_output(inp, stcb, net,
4933 	    (struct sockaddr *)&net->ro._l_addr,
4934 	    m, 0, NULL, 0, 0, 0, 0,
4935 	    inp->sctp_lport, stcb->rport, htonl(0),
4936 	    net->port, NULL,
4937 	    0, 0,
4938 	    so_locked))) {
4939 		SCTPDBG(SCTP_DEBUG_OUTPUT4, "Gak send error %d\n", error);
4940 		if (error == ENOBUFS) {
4941 			stcb->asoc.ifp_had_enobuf = 1;
4942 			SCTP_STAT_INCR(sctps_lowlevelerr);
4943 		}
4944 	} else {
4945 		stcb->asoc.ifp_had_enobuf = 0;
4946 	}
4947 	SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
4948 	(void)SCTP_GETTIME_TIMEVAL(&net->last_sent_time);
4949 }
4950 
4951 struct mbuf *
4952 sctp_arethere_unrecognized_parameters(struct mbuf *in_initpkt,
4953     int param_offset, int *abort_processing, struct sctp_chunkhdr *cp, int *nat_friendly)
4954 {
4955 	/*
4956 	 * Given a mbuf containing an INIT or INIT-ACK with the param_offset
4957 	 * being equal to the beginning of the params i.e. (iphlen +
4958 	 * sizeof(struct sctp_init_msg) parse through the parameters to the
4959 	 * end of the mbuf verifying that all parameters are known.
4960 	 *
4961 	 * For unknown parameters build and return a mbuf with
4962 	 * UNRECOGNIZED_PARAMETER errors. If the flags indicate to stop
4963 	 * processing this chunk stop, and set *abort_processing to 1.
4964 	 *
4965 	 * By having param_offset be pre-set to where parameters begin it is
4966 	 * hoped that this routine may be reused in the future by new
4967 	 * features.
4968 	 */
4969 	struct sctp_paramhdr *phdr, params;
4970 
4971 	struct mbuf *mat, *op_err;
4972 	char tempbuf[SCTP_PARAM_BUFFER_SIZE];
4973 	int at, limit, pad_needed;
4974 	uint16_t ptype, plen, padded_size;
4975 	int err_at;
4976 
4977 	*abort_processing = 0;
4978 	mat = in_initpkt;
4979 	err_at = 0;
4980 	limit = ntohs(cp->chunk_length) - sizeof(struct sctp_init_chunk);
4981 	at = param_offset;
4982 	op_err = NULL;
4983 	SCTPDBG(SCTP_DEBUG_OUTPUT1, "Check for unrecognized param's\n");
4984 	phdr = sctp_get_next_param(mat, at, &params, sizeof(params));
4985 	while ((phdr != NULL) && ((size_t)limit >= sizeof(struct sctp_paramhdr))) {
4986 		ptype = ntohs(phdr->param_type);
4987 		plen = ntohs(phdr->param_length);
4988 		if ((plen > limit) || (plen < sizeof(struct sctp_paramhdr))) {
4989 			/* wacked parameter */
4990 			SCTPDBG(SCTP_DEBUG_OUTPUT1, "Invalid size - error %d\n", plen);
4991 			goto invalid_size;
4992 		}
4993 		limit -= SCTP_SIZE32(plen);
4994 		/*-
4995 		 * All parameters for all chunks that we know/understand are
4996 		 * listed here. We process them other places and make
4997 		 * appropriate stop actions per the upper bits. However this
4998 		 * is the generic routine processor's can call to get back
4999 		 * an operr.. to either incorporate (init-ack) or send.
5000 		 */
5001 		padded_size = SCTP_SIZE32(plen);
5002 		switch (ptype) {
5003 			/* Param's with variable size */
5004 		case SCTP_HEARTBEAT_INFO:
5005 		case SCTP_STATE_COOKIE:
5006 		case SCTP_UNRECOG_PARAM:
5007 		case SCTP_ERROR_CAUSE_IND:
5008 			/* ok skip fwd */
5009 			at += padded_size;
5010 			break;
5011 			/* Param's with variable size within a range */
5012 		case SCTP_CHUNK_LIST:
5013 		case SCTP_SUPPORTED_CHUNK_EXT:
5014 			if (padded_size > (sizeof(struct sctp_supported_chunk_types_param) + (sizeof(uint8_t) * SCTP_MAX_SUPPORTED_EXT))) {
5015 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "Invalid size - error chklist %d\n", plen);
5016 				goto invalid_size;
5017 			}
5018 			at += padded_size;
5019 			break;
5020 		case SCTP_SUPPORTED_ADDRTYPE:
5021 			if (padded_size > SCTP_MAX_ADDR_PARAMS_SIZE) {
5022 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "Invalid size - error supaddrtype %d\n", plen);
5023 				goto invalid_size;
5024 			}
5025 			at += padded_size;
5026 			break;
5027 		case SCTP_RANDOM:
5028 			if (padded_size > (sizeof(struct sctp_auth_random) + SCTP_RANDOM_MAX_SIZE)) {
5029 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "Invalid size - error random %d\n", plen);
5030 				goto invalid_size;
5031 			}
5032 			at += padded_size;
5033 			break;
5034 		case SCTP_SET_PRIM_ADDR:
5035 		case SCTP_DEL_IP_ADDRESS:
5036 		case SCTP_ADD_IP_ADDRESS:
5037 			if ((padded_size != sizeof(struct sctp_asconf_addrv4_param)) &&
5038 			    (padded_size != sizeof(struct sctp_asconf_addr_param))) {
5039 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "Invalid size - error setprim %d\n", plen);
5040 				goto invalid_size;
5041 			}
5042 			at += padded_size;
5043 			break;
5044 			/* Param's with a fixed size */
5045 		case SCTP_IPV4_ADDRESS:
5046 			if (padded_size != sizeof(struct sctp_ipv4addr_param)) {
5047 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "Invalid size - error ipv4 addr %d\n", plen);
5048 				goto invalid_size;
5049 			}
5050 			at += padded_size;
5051 			break;
5052 		case SCTP_IPV6_ADDRESS:
5053 			if (padded_size != sizeof(struct sctp_ipv6addr_param)) {
5054 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "Invalid size - error ipv6 addr %d\n", plen);
5055 				goto invalid_size;
5056 			}
5057 			at += padded_size;
5058 			break;
5059 		case SCTP_COOKIE_PRESERVE:
5060 			if (padded_size != sizeof(struct sctp_cookie_perserve_param)) {
5061 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "Invalid size - error cookie-preserve %d\n", plen);
5062 				goto invalid_size;
5063 			}
5064 			at += padded_size;
5065 			break;
5066 		case SCTP_HAS_NAT_SUPPORT:
5067 			*nat_friendly = 1;
5068 			/* fall through */
5069 		case SCTP_PRSCTP_SUPPORTED:
5070 			if (padded_size != sizeof(struct sctp_paramhdr)) {
5071 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "Invalid size - error prsctp/nat support %d\n", plen);
5072 				goto invalid_size;
5073 			}
5074 			at += padded_size;
5075 			break;
5076 		case SCTP_ECN_CAPABLE:
5077 			if (padded_size != sizeof(struct sctp_paramhdr)) {
5078 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "Invalid size - error ecn %d\n", plen);
5079 				goto invalid_size;
5080 			}
5081 			at += padded_size;
5082 			break;
5083 		case SCTP_ULP_ADAPTATION:
5084 			if (padded_size != sizeof(struct sctp_adaptation_layer_indication)) {
5085 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "Invalid size - error adapatation %d\n", plen);
5086 				goto invalid_size;
5087 			}
5088 			at += padded_size;
5089 			break;
5090 		case SCTP_SUCCESS_REPORT:
5091 			if (padded_size != sizeof(struct sctp_asconf_paramhdr)) {
5092 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "Invalid size - error success %d\n", plen);
5093 				goto invalid_size;
5094 			}
5095 			at += padded_size;
5096 			break;
5097 		case SCTP_HOSTNAME_ADDRESS:
5098 			{
5099 				/* We can NOT handle HOST NAME addresses!! */
5100 				int l_len;
5101 
5102 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "Can't handle hostname addresses.. abort processing\n");
5103 				*abort_processing = 1;
5104 				if (op_err == NULL) {
5105 					/* Ok need to try to get a mbuf */
5106 #ifdef INET6
5107 					l_len = SCTP_MIN_OVERHEAD;
5108 #else
5109 					l_len = SCTP_MIN_V4_OVERHEAD;
5110 #endif
5111 					l_len += sizeof(struct sctp_chunkhdr);
5112 					l_len += plen;
5113 					l_len += sizeof(struct sctp_paramhdr);
5114 					op_err = sctp_get_mbuf_for_msg(l_len, 0, M_NOWAIT, 1, MT_DATA);
5115 					if (op_err) {
5116 						SCTP_BUF_LEN(op_err) = 0;
5117 						/*
5118 						 * pre-reserve space for ip
5119 						 * and sctp header  and
5120 						 * chunk hdr
5121 						 */
5122 #ifdef INET6
5123 						SCTP_BUF_RESV_UF(op_err, sizeof(struct ip6_hdr));
5124 #else
5125 						SCTP_BUF_RESV_UF(op_err, sizeof(struct ip));
5126 #endif
5127 						SCTP_BUF_RESV_UF(op_err, sizeof(struct sctphdr));
5128 						SCTP_BUF_RESV_UF(op_err, sizeof(struct sctp_chunkhdr));
5129 					}
5130 				}
5131 				if (op_err) {
5132 					/* If we have space */
5133 					struct sctp_paramhdr s;
5134 
5135 					if (err_at % 4) {
5136 						uint32_t cpthis = 0;
5137 
5138 						pad_needed = 4 - (err_at % 4);
5139 						m_copyback(op_err, err_at, pad_needed, (caddr_t)&cpthis);
5140 						err_at += pad_needed;
5141 					}
5142 					s.param_type = htons(SCTP_CAUSE_UNRESOLVABLE_ADDR);
5143 					s.param_length = htons(sizeof(s) + plen);
5144 					m_copyback(op_err, err_at, sizeof(s), (caddr_t)&s);
5145 					err_at += sizeof(s);
5146 					if (plen > sizeof(tempbuf)) {
5147 						plen = sizeof(tempbuf);
5148 					}
5149 					phdr = sctp_get_next_param(mat, at, (struct sctp_paramhdr *)tempbuf, plen);
5150 					if (phdr == NULL) {
5151 						sctp_m_freem(op_err);
5152 						/*
5153 						 * we are out of memory but
5154 						 * we still need to have a
5155 						 * look at what to do (the
5156 						 * system is in trouble
5157 						 * though).
5158 						 */
5159 						return (NULL);
5160 					}
5161 					m_copyback(op_err, err_at, plen, (caddr_t)phdr);
5162 				}
5163 				return (op_err);
5164 				break;
5165 			}
5166 		default:
5167 			/*
5168 			 * we do not recognize the parameter figure out what
5169 			 * we do.
5170 			 */
5171 			SCTPDBG(SCTP_DEBUG_OUTPUT1, "Hit default param %x\n", ptype);
5172 			if ((ptype & 0x4000) == 0x4000) {
5173 				/* Report bit is set?? */
5174 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "report op err\n");
5175 				if (op_err == NULL) {
5176 					int l_len;
5177 
5178 					/* Ok need to try to get an mbuf */
5179 #ifdef INET6
5180 					l_len = SCTP_MIN_OVERHEAD;
5181 #else
5182 					l_len = SCTP_MIN_V4_OVERHEAD;
5183 #endif
5184 					l_len += sizeof(struct sctp_chunkhdr);
5185 					l_len += plen;
5186 					l_len += sizeof(struct sctp_paramhdr);
5187 					op_err = sctp_get_mbuf_for_msg(l_len, 0, M_NOWAIT, 1, MT_DATA);
5188 					if (op_err) {
5189 						SCTP_BUF_LEN(op_err) = 0;
5190 #ifdef INET6
5191 						SCTP_BUF_RESV_UF(op_err, sizeof(struct ip6_hdr));
5192 #else
5193 						SCTP_BUF_RESV_UF(op_err, sizeof(struct ip));
5194 #endif
5195 						SCTP_BUF_RESV_UF(op_err, sizeof(struct sctphdr));
5196 						SCTP_BUF_RESV_UF(op_err, sizeof(struct sctp_chunkhdr));
5197 					}
5198 				}
5199 				if (op_err) {
5200 					/* If we have space */
5201 					struct sctp_paramhdr s;
5202 
5203 					if (err_at % 4) {
5204 						uint32_t cpthis = 0;
5205 
5206 						pad_needed = 4 - (err_at % 4);
5207 						m_copyback(op_err, err_at, pad_needed, (caddr_t)&cpthis);
5208 						err_at += pad_needed;
5209 					}
5210 					s.param_type = htons(SCTP_UNRECOG_PARAM);
5211 					s.param_length = htons(sizeof(s) + plen);
5212 					m_copyback(op_err, err_at, sizeof(s), (caddr_t)&s);
5213 					err_at += sizeof(s);
5214 					if (plen > sizeof(tempbuf)) {
5215 						plen = sizeof(tempbuf);
5216 					}
5217 					phdr = sctp_get_next_param(mat, at, (struct sctp_paramhdr *)tempbuf, plen);
5218 					if (phdr == NULL) {
5219 						sctp_m_freem(op_err);
5220 						/*
5221 						 * we are out of memory but
5222 						 * we still need to have a
5223 						 * look at what to do (the
5224 						 * system is in trouble
5225 						 * though).
5226 						 */
5227 						op_err = NULL;
5228 						goto more_processing;
5229 					}
5230 					m_copyback(op_err, err_at, plen, (caddr_t)phdr);
5231 					err_at += plen;
5232 				}
5233 			}
5234 	more_processing:
5235 			if ((ptype & 0x8000) == 0x0000) {
5236 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "stop proc\n");
5237 				return (op_err);
5238 			} else {
5239 				/* skip this chunk and continue processing */
5240 				SCTPDBG(SCTP_DEBUG_OUTPUT1, "move on\n");
5241 				at += SCTP_SIZE32(plen);
5242 			}
5243 			break;
5244 
5245 		}
5246 		phdr = sctp_get_next_param(mat, at, &params, sizeof(params));
5247 	}
5248 	return (op_err);
5249 invalid_size:
5250 	SCTPDBG(SCTP_DEBUG_OUTPUT1, "abort flag set\n");
5251 	*abort_processing = 1;
5252 	if ((op_err == NULL) && phdr) {
5253 		int l_len;
5254 #ifdef INET6
5255 		l_len = SCTP_MIN_OVERHEAD;
5256 #else
5257 		l_len = SCTP_MIN_V4_OVERHEAD;
5258 #endif
5259 		l_len += sizeof(struct sctp_chunkhdr);
5260 		l_len += (2 * sizeof(struct sctp_paramhdr));
5261 		op_err = sctp_get_mbuf_for_msg(l_len, 0, M_NOWAIT, 1, MT_DATA);
5262 		if (op_err) {
5263 			SCTP_BUF_LEN(op_err) = 0;
5264 #ifdef INET6
5265 			SCTP_BUF_RESV_UF(op_err, sizeof(struct ip6_hdr));
5266 #else
5267 			SCTP_BUF_RESV_UF(op_err, sizeof(struct ip));
5268 #endif
5269 			SCTP_BUF_RESV_UF(op_err, sizeof(struct sctphdr));
5270 			SCTP_BUF_RESV_UF(op_err, sizeof(struct sctp_chunkhdr));
5271 		}
5272 	}
5273 	if ((op_err) && phdr) {
5274 		struct sctp_paramhdr s;
5275 
5276 		if (err_at % 4) {
5277 			uint32_t cpthis = 0;
5278 
5279 			pad_needed = 4 - (err_at % 4);
5280 			m_copyback(op_err, err_at, pad_needed, (caddr_t)&cpthis);
5281 			err_at += pad_needed;
5282 		}
5283 		s.param_type = htons(SCTP_CAUSE_PROTOCOL_VIOLATION);
5284 		s.param_length = htons(sizeof(s) + sizeof(struct sctp_paramhdr));
5285 		m_copyback(op_err, err_at, sizeof(s), (caddr_t)&s);
5286 		err_at += sizeof(s);
5287 		/* Only copy back the p-hdr that caused the issue */
5288 		m_copyback(op_err, err_at, sizeof(struct sctp_paramhdr), (caddr_t)phdr);
5289 	}
5290 	return (op_err);
5291 }
5292 
5293 static int
5294 sctp_are_there_new_addresses(struct sctp_association *asoc,
5295     struct mbuf *in_initpkt, int offset, struct sockaddr *src)
5296 {
5297 	/*
5298 	 * Given a INIT packet, look through the packet to verify that there
5299 	 * are NO new addresses. As we go through the parameters add reports
5300 	 * of any un-understood parameters that require an error.  Also we
5301 	 * must return (1) to drop the packet if we see a un-understood
5302 	 * parameter that tells us to drop the chunk.
5303 	 */
5304 	struct sockaddr *sa_touse;
5305 	struct sockaddr *sa;
5306 	struct sctp_paramhdr *phdr, params;
5307 	uint16_t ptype, plen;
5308 	uint8_t fnd;
5309 	struct sctp_nets *net;
5310 	int check_src;
5311 #ifdef INET
5312 	struct sockaddr_in sin4, *sa4;
5313 #endif
5314 #ifdef INET6
5315 	struct sockaddr_in6 sin6, *sa6;
5316 #endif
5317 
5318 #ifdef INET
5319 	memset(&sin4, 0, sizeof(sin4));
5320 	sin4.sin_family = AF_INET;
5321 	sin4.sin_len = sizeof(sin4);
5322 #endif
5323 #ifdef INET6
5324 	memset(&sin6, 0, sizeof(sin6));
5325 	sin6.sin6_family = AF_INET6;
5326 	sin6.sin6_len = sizeof(sin6);
5327 #endif
5328 	/* First what about the src address of the pkt ? */
5329 	check_src = 0;
5330 	switch (src->sa_family) {
5331 #ifdef INET
5332 	case AF_INET:
5333 		if (asoc->scope.ipv4_addr_legal) {
5334 			check_src = 1;
5335 		}
5336 		break;
5337 #endif
5338 #ifdef INET6
5339 	case AF_INET6:
5340 		if (asoc->scope.ipv6_addr_legal) {
5341 			check_src = 1;
5342 		}
5343 		break;
5344 #endif
5345 	default:
5346 		/* TSNH */
5347 		break;
5348 	}
5349 	if (check_src) {
5350 		fnd = 0;
5351 		TAILQ_FOREACH(net, &asoc->nets, sctp_next) {
5352 			sa = (struct sockaddr *)&net->ro._l_addr;
5353 			if (sa->sa_family == src->sa_family) {
5354 #ifdef INET
5355 				if (sa->sa_family == AF_INET) {
5356 					struct sockaddr_in *src4;
5357 
5358 					sa4 = (struct sockaddr_in *)sa;
5359 					src4 = (struct sockaddr_in *)src;
5360 					if (sa4->sin_addr.s_addr == src4->sin_addr.s_addr) {
5361 						fnd = 1;
5362 						break;
5363 					}
5364 				}
5365 #endif
5366 #ifdef INET6
5367 				if (sa->sa_family == AF_INET6) {
5368 					struct sockaddr_in6 *src6;
5369 
5370 					sa6 = (struct sockaddr_in6 *)sa;
5371 					src6 = (struct sockaddr_in6 *)src;
5372 					if (SCTP6_ARE_ADDR_EQUAL(sa6, src6)) {
5373 						fnd = 1;
5374 						break;
5375 					}
5376 				}
5377 #endif
5378 			}
5379 		}
5380 		if (fnd == 0) {
5381 			/* New address added! no need to look further. */
5382 			return (1);
5383 		}
5384 	}
5385 	/* Ok so far lets munge through the rest of the packet */
5386 	offset += sizeof(struct sctp_init_chunk);
5387 	phdr = sctp_get_next_param(in_initpkt, offset, &params, sizeof(params));
5388 	while (phdr) {
5389 		sa_touse = NULL;
5390 		ptype = ntohs(phdr->param_type);
5391 		plen = ntohs(phdr->param_length);
5392 		switch (ptype) {
5393 #ifdef INET
5394 		case SCTP_IPV4_ADDRESS:
5395 			{
5396 				struct sctp_ipv4addr_param *p4, p4_buf;
5397 
5398 				if (plen != sizeof(struct sctp_ipv4addr_param)) {
5399 					return (1);
5400 				}
5401 				phdr = sctp_get_next_param(in_initpkt, offset,
5402 				    (struct sctp_paramhdr *)&p4_buf, sizeof(p4_buf));
5403 				if (phdr == NULL) {
5404 					return (1);
5405 				}
5406 				if (asoc->scope.ipv4_addr_legal) {
5407 					p4 = (struct sctp_ipv4addr_param *)phdr;
5408 					sin4.sin_addr.s_addr = p4->addr;
5409 					sa_touse = (struct sockaddr *)&sin4;
5410 				}
5411 				break;
5412 			}
5413 #endif
5414 #ifdef INET6
5415 		case SCTP_IPV6_ADDRESS:
5416 			{
5417 				struct sctp_ipv6addr_param *p6, p6_buf;
5418 
5419 				if (plen != sizeof(struct sctp_ipv6addr_param)) {
5420 					return (1);
5421 				}
5422 				phdr = sctp_get_next_param(in_initpkt, offset,
5423 				    (struct sctp_paramhdr *)&p6_buf, sizeof(p6_buf));
5424 				if (phdr == NULL) {
5425 					return (1);
5426 				}
5427 				if (asoc->scope.ipv6_addr_legal) {
5428 					p6 = (struct sctp_ipv6addr_param *)phdr;
5429 					memcpy((caddr_t)&sin6.sin6_addr, p6->addr,
5430 					    sizeof(p6->addr));
5431 					sa_touse = (struct sockaddr *)&sin6;
5432 				}
5433 				break;
5434 			}
5435 #endif
5436 		default:
5437 			sa_touse = NULL;
5438 			break;
5439 		}
5440 		if (sa_touse) {
5441 			/* ok, sa_touse points to one to check */
5442 			fnd = 0;
5443 			TAILQ_FOREACH(net, &asoc->nets, sctp_next) {
5444 				sa = (struct sockaddr *)&net->ro._l_addr;
5445 				if (sa->sa_family != sa_touse->sa_family) {
5446 					continue;
5447 				}
5448 #ifdef INET
5449 				if (sa->sa_family == AF_INET) {
5450 					sa4 = (struct sockaddr_in *)sa;
5451 					if (sa4->sin_addr.s_addr ==
5452 					    sin4.sin_addr.s_addr) {
5453 						fnd = 1;
5454 						break;
5455 					}
5456 				}
5457 #endif
5458 #ifdef INET6
5459 				if (sa->sa_family == AF_INET6) {
5460 					sa6 = (struct sockaddr_in6 *)sa;
5461 					if (SCTP6_ARE_ADDR_EQUAL(
5462 					    sa6, &sin6)) {
5463 						fnd = 1;
5464 						break;
5465 					}
5466 				}
5467 #endif
5468 			}
5469 			if (!fnd) {
5470 				/* New addr added! no need to look further */
5471 				return (1);
5472 			}
5473 		}
5474 		offset += SCTP_SIZE32(plen);
5475 		phdr = sctp_get_next_param(in_initpkt, offset, &params, sizeof(params));
5476 	}
5477 	return (0);
5478 }
5479 
5480 /*
5481  * Given a MBUF chain that was sent into us containing an INIT. Build a
5482  * INIT-ACK with COOKIE and send back. We assume that the in_initpkt has done
5483  * a pullup to include IPv6/4header, SCTP header and initial part of INIT
5484  * message (i.e. the struct sctp_init_msg).
5485  */
5486 void
5487 sctp_send_initiate_ack(struct sctp_inpcb *inp, struct sctp_tcb *stcb,
5488     struct sctp_nets *src_net, struct mbuf *init_pkt,
5489     int iphlen, int offset,
5490     struct sockaddr *src, struct sockaddr *dst,
5491     struct sctphdr *sh, struct sctp_init_chunk *init_chk,
5492     uint8_t mflowtype, uint32_t mflowid,
5493     uint32_t vrf_id, uint16_t port)
5494 {
5495 	struct sctp_association *asoc;
5496 	struct mbuf *m, *m_tmp, *m_last, *m_cookie, *op_err;
5497 	struct sctp_init_ack_chunk *initack;
5498 	struct sctp_adaptation_layer_indication *ali;
5499 	struct sctp_supported_chunk_types_param *pr_supported;
5500 	struct sctp_paramhdr *ph;
5501 	union sctp_sockstore *over_addr;
5502 	struct sctp_scoping scp;
5503 	struct timeval now;
5504 #ifdef INET
5505 	struct sockaddr_in *dst4 = (struct sockaddr_in *)dst;
5506 	struct sockaddr_in *src4 = (struct sockaddr_in *)src;
5507 	struct sockaddr_in *sin;
5508 #endif
5509 #ifdef INET6
5510 	struct sockaddr_in6 *dst6 = (struct sockaddr_in6 *)dst;
5511 	struct sockaddr_in6 *src6 = (struct sockaddr_in6 *)src;
5512 	struct sockaddr_in6 *sin6;
5513 #endif
5514 	struct sockaddr *to;
5515 	struct sctp_state_cookie stc;
5516 	struct sctp_nets *net = NULL;
5517 	uint8_t *signature = NULL;
5518 	int cnt_inits_to = 0;
5519 	uint16_t his_limit, i_want;
5520 	int abort_flag;
5521 	int nat_friendly = 0;
5522 	int error;
5523 	struct socket *so;
5524 	uint16_t num_ext, chunk_len, padding_len, parameter_len;
5525 
5526 	if (stcb) {
5527 		asoc = &stcb->asoc;
5528 	} else {
5529 		asoc = NULL;
5530 	}
5531 	if ((asoc != NULL) &&
5532 	    (SCTP_GET_STATE(stcb) != SCTP_STATE_COOKIE_WAIT)) {
5533 		if (sctp_are_there_new_addresses(asoc, init_pkt, offset, src)) {
5534 			/*
5535 			 * new addresses, out of here in non-cookie-wait
5536 			 * states
5537 			 *
5538 			 * Send an ABORT, without the new address error
5539 			 * cause. This looks no different than if no
5540 			 * listener was present.
5541 			 */
5542 			op_err = sctp_generate_cause(SCTP_BASE_SYSCTL(sctp_diag_info_code),
5543 			    "Address added");
5544 			sctp_send_abort(init_pkt, iphlen, src, dst, sh, 0, op_err,
5545 			    mflowtype, mflowid, inp->fibnum,
5546 			    vrf_id, port);
5547 			return;
5548 		}
5549 		if (src_net != NULL && (src_net->port != port)) {
5550 			/*
5551 			 * change of remote encapsulation port, out of here
5552 			 * in non-cookie-wait states
5553 			 *
5554 			 * Send an ABORT, without an specific error cause.
5555 			 * This looks no different than if no listener was
5556 			 * present.
5557 			 */
5558 			op_err = sctp_generate_cause(SCTP_BASE_SYSCTL(sctp_diag_info_code),
5559 			    "Remote encapsulation port changed");
5560 			sctp_send_abort(init_pkt, iphlen, src, dst, sh, 0, op_err,
5561 			    mflowtype, mflowid, inp->fibnum,
5562 			    vrf_id, port);
5563 			return;
5564 		}
5565 	}
5566 	abort_flag = 0;
5567 	op_err = sctp_arethere_unrecognized_parameters(init_pkt,
5568 	    (offset + sizeof(struct sctp_init_chunk)),
5569 	    &abort_flag, (struct sctp_chunkhdr *)init_chk, &nat_friendly);
5570 	if (abort_flag) {
5571 do_a_abort:
5572 		if (op_err == NULL) {
5573 			char msg[SCTP_DIAG_INFO_LEN];
5574 
5575 			snprintf(msg, sizeof(msg), "%s:%d at %s", __FILE__, __LINE__, __func__);
5576 			op_err = sctp_generate_cause(SCTP_BASE_SYSCTL(sctp_diag_info_code),
5577 			    msg);
5578 		}
5579 		sctp_send_abort(init_pkt, iphlen, src, dst, sh,
5580 		    init_chk->init.initiate_tag, op_err,
5581 		    mflowtype, mflowid, inp->fibnum,
5582 		    vrf_id, port);
5583 		return;
5584 	}
5585 	m = sctp_get_mbuf_for_msg(MCLBYTES, 0, M_NOWAIT, 1, MT_DATA);
5586 	if (m == NULL) {
5587 		/* No memory, INIT timer will re-attempt. */
5588 		if (op_err)
5589 			sctp_m_freem(op_err);
5590 		return;
5591 	}
5592 	chunk_len = (uint16_t)sizeof(struct sctp_init_ack_chunk);
5593 	padding_len = 0;
5594 
5595 	/*
5596 	 * We might not overwrite the identification[] completely and on
5597 	 * some platforms time_entered will contain some padding. Therefore
5598 	 * zero out the cookie to avoid putting uninitialized memory on the
5599 	 * wire.
5600 	 */
5601 	memset(&stc, 0, sizeof(struct sctp_state_cookie));
5602 
5603 	/* the time I built cookie */
5604 	(void)SCTP_GETTIME_TIMEVAL(&now);
5605 	stc.time_entered.tv_sec = now.tv_sec;
5606 	stc.time_entered.tv_usec = now.tv_usec;
5607 
5608 	/* populate any tie tags */
5609 	if (asoc != NULL) {
5610 		/* unlock before tag selections */
5611 		stc.tie_tag_my_vtag = asoc->my_vtag_nonce;
5612 		stc.tie_tag_peer_vtag = asoc->peer_vtag_nonce;
5613 		stc.cookie_life = asoc->cookie_life;
5614 		net = asoc->primary_destination;
5615 	} else {
5616 		stc.tie_tag_my_vtag = 0;
5617 		stc.tie_tag_peer_vtag = 0;
5618 		/* life I will award this cookie */
5619 		stc.cookie_life = inp->sctp_ep.def_cookie_life;
5620 	}
5621 
5622 	/* copy in the ports for later check */
5623 	stc.myport = sh->dest_port;
5624 	stc.peerport = sh->src_port;
5625 
5626 	/*
5627 	 * If we wanted to honor cookie life extensions, we would add to
5628 	 * stc.cookie_life. For now we should NOT honor any extension
5629 	 */
5630 	stc.site_scope = stc.local_scope = stc.loopback_scope = 0;
5631 	if (inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) {
5632 		stc.ipv6_addr_legal = 1;
5633 		if (SCTP_IPV6_V6ONLY(inp)) {
5634 			stc.ipv4_addr_legal = 0;
5635 		} else {
5636 			stc.ipv4_addr_legal = 1;
5637 		}
5638 	} else {
5639 		stc.ipv6_addr_legal = 0;
5640 		stc.ipv4_addr_legal = 1;
5641 	}
5642 	stc.ipv4_scope = 0;
5643 	if (net == NULL) {
5644 		to = src;
5645 		switch (dst->sa_family) {
5646 #ifdef INET
5647 		case AF_INET:
5648 			{
5649 				/* lookup address */
5650 				stc.address[0] = src4->sin_addr.s_addr;
5651 				stc.address[1] = 0;
5652 				stc.address[2] = 0;
5653 				stc.address[3] = 0;
5654 				stc.addr_type = SCTP_IPV4_ADDRESS;
5655 				/* local from address */
5656 				stc.laddress[0] = dst4->sin_addr.s_addr;
5657 				stc.laddress[1] = 0;
5658 				stc.laddress[2] = 0;
5659 				stc.laddress[3] = 0;
5660 				stc.laddr_type = SCTP_IPV4_ADDRESS;
5661 				/* scope_id is only for v6 */
5662 				stc.scope_id = 0;
5663 				if ((IN4_ISPRIVATE_ADDRESS(&src4->sin_addr)) ||
5664 				    (IN4_ISPRIVATE_ADDRESS(&dst4->sin_addr))) {
5665 					stc.ipv4_scope = 1;
5666 				}
5667 				/* Must use the address in this case */
5668 				if (sctp_is_address_on_local_host(src, vrf_id)) {
5669 					stc.loopback_scope = 1;
5670 					stc.ipv4_scope = 1;
5671 					stc.site_scope = 1;
5672 					stc.local_scope = 0;
5673 				}
5674 				break;
5675 			}
5676 #endif
5677 #ifdef INET6
5678 		case AF_INET6:
5679 			{
5680 				stc.addr_type = SCTP_IPV6_ADDRESS;
5681 				memcpy(&stc.address, &src6->sin6_addr, sizeof(struct in6_addr));
5682 				stc.scope_id = ntohs(in6_getscope(&src6->sin6_addr));
5683 				if (sctp_is_address_on_local_host(src, vrf_id)) {
5684 					stc.loopback_scope = 1;
5685 					stc.local_scope = 0;
5686 					stc.site_scope = 1;
5687 					stc.ipv4_scope = 1;
5688 				} else if (IN6_IS_ADDR_LINKLOCAL(&src6->sin6_addr) ||
5689 				    IN6_IS_ADDR_LINKLOCAL(&dst6->sin6_addr)) {
5690 					/*
5691 					 * If the new destination or source
5692 					 * is a LINK_LOCAL we must have
5693 					 * common both site and local scope.
5694 					 * Don't set local scope though
5695 					 * since we must depend on the
5696 					 * source to be added implicitly. We
5697 					 * cannot assure just because we
5698 					 * share one link that all links are
5699 					 * common.
5700 					 */
5701 					stc.local_scope = 0;
5702 					stc.site_scope = 1;
5703 					stc.ipv4_scope = 1;
5704 					/*
5705 					 * we start counting for the private
5706 					 * address stuff at 1. since the
5707 					 * link local we source from won't
5708 					 * show up in our scoped count.
5709 					 */
5710 					cnt_inits_to = 1;
5711 					/*
5712 					 * pull out the scope_id from
5713 					 * incoming pkt
5714 					 */
5715 				} else if (IN6_IS_ADDR_SITELOCAL(&src6->sin6_addr) ||
5716 				    IN6_IS_ADDR_SITELOCAL(&dst6->sin6_addr)) {
5717 					/*
5718 					 * If the new destination or source
5719 					 * is SITE_LOCAL then we must have
5720 					 * site scope in common.
5721 					 */
5722 					stc.site_scope = 1;
5723 				}
5724 				memcpy(&stc.laddress, &dst6->sin6_addr, sizeof(struct in6_addr));
5725 				stc.laddr_type = SCTP_IPV6_ADDRESS;
5726 				break;
5727 			}
5728 #endif
5729 		default:
5730 			/* TSNH */
5731 			goto do_a_abort;
5732 			break;
5733 		}
5734 	} else {
5735 		/* set the scope per the existing tcb */
5736 
5737 #ifdef INET6
5738 		struct sctp_nets *lnet;
5739 #endif
5740 
5741 		stc.loopback_scope = asoc->scope.loopback_scope;
5742 		stc.ipv4_scope = asoc->scope.ipv4_local_scope;
5743 		stc.site_scope = asoc->scope.site_scope;
5744 		stc.local_scope = asoc->scope.local_scope;
5745 #ifdef INET6
5746 		/* Why do we not consider IPv4 LL addresses? */
5747 		TAILQ_FOREACH(lnet, &asoc->nets, sctp_next) {
5748 			if (lnet->ro._l_addr.sin6.sin6_family == AF_INET6) {
5749 				if (IN6_IS_ADDR_LINKLOCAL(&lnet->ro._l_addr.sin6.sin6_addr)) {
5750 					/*
5751 					 * if we have a LL address, start
5752 					 * counting at 1.
5753 					 */
5754 					cnt_inits_to = 1;
5755 				}
5756 			}
5757 		}
5758 #endif
5759 		/* use the net pointer */
5760 		to = (struct sockaddr *)&net->ro._l_addr;
5761 		switch (to->sa_family) {
5762 #ifdef INET
5763 		case AF_INET:
5764 			sin = (struct sockaddr_in *)to;
5765 			stc.address[0] = sin->sin_addr.s_addr;
5766 			stc.address[1] = 0;
5767 			stc.address[2] = 0;
5768 			stc.address[3] = 0;
5769 			stc.addr_type = SCTP_IPV4_ADDRESS;
5770 			if (net->src_addr_selected == 0) {
5771 				/*
5772 				 * strange case here, the INIT should have
5773 				 * did the selection.
5774 				 */
5775 				net->ro._s_addr = sctp_source_address_selection(inp,
5776 				    stcb, (sctp_route_t *)&net->ro,
5777 				    net, 0, vrf_id);
5778 				if (net->ro._s_addr == NULL)
5779 					return;
5780 
5781 				net->src_addr_selected = 1;
5782 
5783 			}
5784 			stc.laddress[0] = net->ro._s_addr->address.sin.sin_addr.s_addr;
5785 			stc.laddress[1] = 0;
5786 			stc.laddress[2] = 0;
5787 			stc.laddress[3] = 0;
5788 			stc.laddr_type = SCTP_IPV4_ADDRESS;
5789 			/* scope_id is only for v6 */
5790 			stc.scope_id = 0;
5791 			break;
5792 #endif
5793 #ifdef INET6
5794 		case AF_INET6:
5795 			sin6 = (struct sockaddr_in6 *)to;
5796 			memcpy(&stc.address, &sin6->sin6_addr,
5797 			    sizeof(struct in6_addr));
5798 			stc.addr_type = SCTP_IPV6_ADDRESS;
5799 			stc.scope_id = sin6->sin6_scope_id;
5800 			if (net->src_addr_selected == 0) {
5801 				/*
5802 				 * strange case here, the INIT should have
5803 				 * done the selection.
5804 				 */
5805 				net->ro._s_addr = sctp_source_address_selection(inp,
5806 				    stcb, (sctp_route_t *)&net->ro,
5807 				    net, 0, vrf_id);
5808 				if (net->ro._s_addr == NULL)
5809 					return;
5810 
5811 				net->src_addr_selected = 1;
5812 			}
5813 			memcpy(&stc.laddress, &net->ro._s_addr->address.sin6.sin6_addr,
5814 			    sizeof(struct in6_addr));
5815 			stc.laddr_type = SCTP_IPV6_ADDRESS;
5816 			break;
5817 #endif
5818 		}
5819 	}
5820 	/* Now lets put the SCTP header in place */
5821 	initack = mtod(m, struct sctp_init_ack_chunk *);
5822 	/* Save it off for quick ref */
5823 	stc.peers_vtag = ntohl(init_chk->init.initiate_tag);
5824 	/* who are we */
5825 	memcpy(stc.identification, SCTP_VERSION_STRING,
5826 	    min(strlen(SCTP_VERSION_STRING), sizeof(stc.identification)));
5827 	memset(stc.reserved, 0, SCTP_RESERVE_SPACE);
5828 	/* now the chunk header */
5829 	initack->ch.chunk_type = SCTP_INITIATION_ACK;
5830 	initack->ch.chunk_flags = 0;
5831 	/* fill in later from mbuf we build */
5832 	initack->ch.chunk_length = 0;
5833 	/* place in my tag */
5834 	if ((asoc != NULL) &&
5835 	    ((SCTP_GET_STATE(stcb) == SCTP_STATE_COOKIE_WAIT) ||
5836 	    (SCTP_GET_STATE(stcb) == SCTP_STATE_INUSE) ||
5837 	    (SCTP_GET_STATE(stcb) == SCTP_STATE_COOKIE_ECHOED))) {
5838 		/* re-use the v-tags and init-seq here */
5839 		initack->init.initiate_tag = htonl(asoc->my_vtag);
5840 		initack->init.initial_tsn = htonl(asoc->init_seq_number);
5841 	} else {
5842 		uint32_t vtag, itsn;
5843 
5844 		if (asoc) {
5845 			atomic_add_int(&asoc->refcnt, 1);
5846 			SCTP_TCB_UNLOCK(stcb);
5847 	new_tag:
5848 			vtag = sctp_select_a_tag(inp, inp->sctp_lport, sh->src_port, 1);
5849 			if ((asoc->peer_supports_nat) && (vtag == asoc->my_vtag)) {
5850 				/*
5851 				 * Got a duplicate vtag on some guy behind a
5852 				 * nat make sure we don't use it.
5853 				 */
5854 				goto new_tag;
5855 			}
5856 			initack->init.initiate_tag = htonl(vtag);
5857 			/* get a TSN to use too */
5858 			itsn = sctp_select_initial_TSN(&inp->sctp_ep);
5859 			initack->init.initial_tsn = htonl(itsn);
5860 			SCTP_TCB_LOCK(stcb);
5861 			atomic_add_int(&asoc->refcnt, -1);
5862 		} else {
5863 			SCTP_INP_INCR_REF(inp);
5864 			SCTP_INP_RUNLOCK(inp);
5865 			vtag = sctp_select_a_tag(inp, inp->sctp_lport, sh->src_port, 1);
5866 			initack->init.initiate_tag = htonl(vtag);
5867 			/* get a TSN to use too */
5868 			initack->init.initial_tsn = htonl(sctp_select_initial_TSN(&inp->sctp_ep));
5869 			SCTP_INP_RLOCK(inp);
5870 			SCTP_INP_DECR_REF(inp);
5871 		}
5872 	}
5873 	/* save away my tag to */
5874 	stc.my_vtag = initack->init.initiate_tag;
5875 
5876 	/* set up some of the credits. */
5877 	so = inp->sctp_socket;
5878 	if (so == NULL) {
5879 		/* memory problem */
5880 		sctp_m_freem(m);
5881 		return;
5882 	} else {
5883 		initack->init.a_rwnd = htonl(max(SCTP_SB_LIMIT_RCV(so), SCTP_MINIMAL_RWND));
5884 	}
5885 	/* set what I want */
5886 	his_limit = ntohs(init_chk->init.num_inbound_streams);
5887 	/* choose what I want */
5888 	if (asoc != NULL) {
5889 		if (asoc->streamoutcnt > asoc->pre_open_streams) {
5890 			i_want = asoc->streamoutcnt;
5891 		} else {
5892 			i_want = asoc->pre_open_streams;
5893 		}
5894 	} else {
5895 		i_want = inp->sctp_ep.pre_open_stream_count;
5896 	}
5897 	if (his_limit < i_want) {
5898 		/* I Want more :< */
5899 		initack->init.num_outbound_streams = init_chk->init.num_inbound_streams;
5900 	} else {
5901 		/* I can have what I want :> */
5902 		initack->init.num_outbound_streams = htons(i_want);
5903 	}
5904 	/* tell him his limit. */
5905 	initack->init.num_inbound_streams =
5906 	    htons(inp->sctp_ep.max_open_streams_intome);
5907 
5908 	/* adaptation layer indication parameter */
5909 	if (inp->sctp_ep.adaptation_layer_indicator_provided) {
5910 		parameter_len = (uint16_t)sizeof(struct sctp_adaptation_layer_indication);
5911 		ali = (struct sctp_adaptation_layer_indication *)(mtod(m, caddr_t)+chunk_len);
5912 		ali->ph.param_type = htons(SCTP_ULP_ADAPTATION);
5913 		ali->ph.param_length = htons(parameter_len);
5914 		ali->indication = htonl(inp->sctp_ep.adaptation_layer_indicator);
5915 		chunk_len += parameter_len;
5916 	}
5917 
5918 	/* ECN parameter */
5919 	if (((asoc != NULL) && (asoc->ecn_supported == 1)) ||
5920 	    ((asoc == NULL) && (inp->ecn_supported == 1))) {
5921 		parameter_len = (uint16_t)sizeof(struct sctp_paramhdr);
5922 		ph = (struct sctp_paramhdr *)(mtod(m, caddr_t)+chunk_len);
5923 		ph->param_type = htons(SCTP_ECN_CAPABLE);
5924 		ph->param_length = htons(parameter_len);
5925 		chunk_len += parameter_len;
5926 	}
5927 
5928 	/* PR-SCTP supported parameter */
5929 	if (((asoc != NULL) && (asoc->prsctp_supported == 1)) ||
5930 	    ((asoc == NULL) && (inp->prsctp_supported == 1))) {
5931 		parameter_len = (uint16_t)sizeof(struct sctp_paramhdr);
5932 		ph = (struct sctp_paramhdr *)(mtod(m, caddr_t)+chunk_len);
5933 		ph->param_type = htons(SCTP_PRSCTP_SUPPORTED);
5934 		ph->param_length = htons(parameter_len);
5935 		chunk_len += parameter_len;
5936 	}
5937 
5938 	/* Add NAT friendly parameter */
5939 	if (nat_friendly) {
5940 		parameter_len = (uint16_t)sizeof(struct sctp_paramhdr);
5941 		ph = (struct sctp_paramhdr *)(mtod(m, caddr_t)+chunk_len);
5942 		ph->param_type = htons(SCTP_HAS_NAT_SUPPORT);
5943 		ph->param_length = htons(parameter_len);
5944 		chunk_len += parameter_len;
5945 	}
5946 
5947 	/* And now tell the peer which extensions we support */
5948 	num_ext = 0;
5949 	pr_supported = (struct sctp_supported_chunk_types_param *)(mtod(m, caddr_t)+chunk_len);
5950 	if (((asoc != NULL) && (asoc->prsctp_supported == 1)) ||
5951 	    ((asoc == NULL) && (inp->prsctp_supported == 1))) {
5952 		pr_supported->chunk_types[num_ext++] = SCTP_FORWARD_CUM_TSN;
5953 		if (((asoc != NULL) && (asoc->idata_supported == 1)) ||
5954 		    ((asoc == NULL) && (inp->idata_supported == 1))) {
5955 			pr_supported->chunk_types[num_ext++] = SCTP_IFORWARD_CUM_TSN;
5956 		}
5957 	}
5958 	if (((asoc != NULL) && (asoc->auth_supported == 1)) ||
5959 	    ((asoc == NULL) && (inp->auth_supported == 1))) {
5960 		pr_supported->chunk_types[num_ext++] = SCTP_AUTHENTICATION;
5961 	}
5962 	if (((asoc != NULL) && (asoc->asconf_supported == 1)) ||
5963 	    ((asoc == NULL) && (inp->asconf_supported == 1))) {
5964 		pr_supported->chunk_types[num_ext++] = SCTP_ASCONF;
5965 		pr_supported->chunk_types[num_ext++] = SCTP_ASCONF_ACK;
5966 	}
5967 	if (((asoc != NULL) && (asoc->reconfig_supported == 1)) ||
5968 	    ((asoc == NULL) && (inp->reconfig_supported == 1))) {
5969 		pr_supported->chunk_types[num_ext++] = SCTP_STREAM_RESET;
5970 	}
5971 	if (((asoc != NULL) && (asoc->idata_supported == 1)) ||
5972 	    ((asoc == NULL) && (inp->idata_supported == 1))) {
5973 		pr_supported->chunk_types[num_ext++] = SCTP_IDATA;
5974 	}
5975 	if (((asoc != NULL) && (asoc->nrsack_supported == 1)) ||
5976 	    ((asoc == NULL) && (inp->nrsack_supported == 1))) {
5977 		pr_supported->chunk_types[num_ext++] = SCTP_NR_SELECTIVE_ACK;
5978 	}
5979 	if (((asoc != NULL) && (asoc->pktdrop_supported == 1)) ||
5980 	    ((asoc == NULL) && (inp->pktdrop_supported == 1))) {
5981 		pr_supported->chunk_types[num_ext++] = SCTP_PACKET_DROPPED;
5982 	}
5983 	if (num_ext > 0) {
5984 		parameter_len = (uint16_t)sizeof(struct sctp_supported_chunk_types_param) + num_ext;
5985 		pr_supported->ph.param_type = htons(SCTP_SUPPORTED_CHUNK_EXT);
5986 		pr_supported->ph.param_length = htons(parameter_len);
5987 		padding_len = SCTP_SIZE32(parameter_len) - parameter_len;
5988 		chunk_len += parameter_len;
5989 	}
5990 
5991 	/* add authentication parameters */
5992 	if (((asoc != NULL) && (asoc->auth_supported == 1)) ||
5993 	    ((asoc == NULL) && (inp->auth_supported == 1))) {
5994 		struct sctp_auth_random *randp;
5995 		struct sctp_auth_hmac_algo *hmacs;
5996 		struct sctp_auth_chunk_list *chunks;
5997 
5998 		if (padding_len > 0) {
5999 			memset(mtod(m, caddr_t)+chunk_len, 0, padding_len);
6000 			chunk_len += padding_len;
6001 			padding_len = 0;
6002 		}
6003 		/* generate and add RANDOM parameter */
6004 		randp = (struct sctp_auth_random *)(mtod(m, caddr_t)+chunk_len);
6005 		parameter_len = (uint16_t)sizeof(struct sctp_auth_random) +
6006 		    SCTP_AUTH_RANDOM_SIZE_DEFAULT;
6007 		randp->ph.param_type = htons(SCTP_RANDOM);
6008 		randp->ph.param_length = htons(parameter_len);
6009 		SCTP_READ_RANDOM(randp->random_data, SCTP_AUTH_RANDOM_SIZE_DEFAULT);
6010 		padding_len = SCTP_SIZE32(parameter_len) - parameter_len;
6011 		chunk_len += parameter_len;
6012 
6013 		if (padding_len > 0) {
6014 			memset(mtod(m, caddr_t)+chunk_len, 0, padding_len);
6015 			chunk_len += padding_len;
6016 			padding_len = 0;
6017 		}
6018 		/* add HMAC_ALGO parameter */
6019 		hmacs = (struct sctp_auth_hmac_algo *)(mtod(m, caddr_t)+chunk_len);
6020 		parameter_len = (uint16_t)sizeof(struct sctp_auth_hmac_algo) +
6021 		    sctp_serialize_hmaclist(inp->sctp_ep.local_hmacs,
6022 		    (uint8_t *)hmacs->hmac_ids);
6023 		hmacs->ph.param_type = htons(SCTP_HMAC_LIST);
6024 		hmacs->ph.param_length = htons(parameter_len);
6025 		padding_len = SCTP_SIZE32(parameter_len) - parameter_len;
6026 		chunk_len += parameter_len;
6027 
6028 		if (padding_len > 0) {
6029 			memset(mtod(m, caddr_t)+chunk_len, 0, padding_len);
6030 			chunk_len += padding_len;
6031 			padding_len = 0;
6032 		}
6033 		/* add CHUNKS parameter */
6034 		chunks = (struct sctp_auth_chunk_list *)(mtod(m, caddr_t)+chunk_len);
6035 		parameter_len = (uint16_t)sizeof(struct sctp_auth_chunk_list) +
6036 		    sctp_serialize_auth_chunks(inp->sctp_ep.local_auth_chunks,
6037 		    chunks->chunk_types);
6038 		chunks->ph.param_type = htons(SCTP_CHUNK_LIST);
6039 		chunks->ph.param_length = htons(parameter_len);
6040 		padding_len = SCTP_SIZE32(parameter_len) - parameter_len;
6041 		chunk_len += parameter_len;
6042 	}
6043 	SCTP_BUF_LEN(m) = chunk_len;
6044 	m_last = m;
6045 	/* now the addresses */
6046 	/*
6047 	 * To optimize this we could put the scoping stuff into a structure
6048 	 * and remove the individual uint8's from the stc structure. Then we
6049 	 * could just sifa in the address within the stc.. but for now this
6050 	 * is a quick hack to get the address stuff teased apart.
6051 	 */
6052 	scp.ipv4_addr_legal = stc.ipv4_addr_legal;
6053 	scp.ipv6_addr_legal = stc.ipv6_addr_legal;
6054 	scp.loopback_scope = stc.loopback_scope;
6055 	scp.ipv4_local_scope = stc.ipv4_scope;
6056 	scp.local_scope = stc.local_scope;
6057 	scp.site_scope = stc.site_scope;
6058 	m_last = sctp_add_addresses_to_i_ia(inp, stcb, &scp, m_last,
6059 	    cnt_inits_to,
6060 	    &padding_len, &chunk_len);
6061 	/* padding_len can only be positive, if no addresses have been added */
6062 	if (padding_len > 0) {
6063 		memset(mtod(m, caddr_t)+chunk_len, 0, padding_len);
6064 		chunk_len += padding_len;
6065 		SCTP_BUF_LEN(m) += padding_len;
6066 		padding_len = 0;
6067 	}
6068 
6069 	/* tack on the operational error if present */
6070 	if (op_err) {
6071 		parameter_len = 0;
6072 		for (m_tmp = op_err; m_tmp != NULL; m_tmp = SCTP_BUF_NEXT(m_tmp)) {
6073 			parameter_len += SCTP_BUF_LEN(m_tmp);
6074 		}
6075 		padding_len = SCTP_SIZE32(parameter_len) - parameter_len;
6076 		SCTP_BUF_NEXT(m_last) = op_err;
6077 		while (SCTP_BUF_NEXT(m_last) != NULL) {
6078 			m_last = SCTP_BUF_NEXT(m_last);
6079 		}
6080 		chunk_len += parameter_len;
6081 	}
6082 	if (padding_len > 0) {
6083 		m_last = sctp_add_pad_tombuf(m_last, padding_len);
6084 		if (m_last == NULL) {
6085 			/* Houston we have a problem, no space */
6086 			sctp_m_freem(m);
6087 			return;
6088 		}
6089 		chunk_len += padding_len;
6090 		padding_len = 0;
6091 	}
6092 	/* Now we must build a cookie */
6093 	m_cookie = sctp_add_cookie(init_pkt, offset, m, 0, &stc, &signature);
6094 	if (m_cookie == NULL) {
6095 		/* memory problem */
6096 		sctp_m_freem(m);
6097 		return;
6098 	}
6099 	/* Now append the cookie to the end and update the space/size */
6100 	SCTP_BUF_NEXT(m_last) = m_cookie;
6101 	parameter_len = 0;
6102 	for (m_tmp = m_cookie; m_tmp != NULL; m_tmp = SCTP_BUF_NEXT(m_tmp)) {
6103 		parameter_len += SCTP_BUF_LEN(m_tmp);
6104 		if (SCTP_BUF_NEXT(m_tmp) == NULL) {
6105 			m_last = m_tmp;
6106 		}
6107 	}
6108 	padding_len = SCTP_SIZE32(parameter_len) - parameter_len;
6109 	chunk_len += parameter_len;
6110 
6111 	/*
6112 	 * Place in the size, but we don't include the last pad (if any) in
6113 	 * the INIT-ACK.
6114 	 */
6115 	initack->ch.chunk_length = htons(chunk_len);
6116 
6117 	/*
6118 	 * Time to sign the cookie, we don't sign over the cookie signature
6119 	 * though thus we set trailer.
6120 	 */
6121 	(void)sctp_hmac_m(SCTP_HMAC,
6122 	    (uint8_t *)inp->sctp_ep.secret_key[(int)(inp->sctp_ep.current_secret_number)],
6123 	    SCTP_SECRET_SIZE, m_cookie, sizeof(struct sctp_paramhdr),
6124 	    (uint8_t *)signature, SCTP_SIGNATURE_SIZE);
6125 	/*
6126 	 * We sifa 0 here to NOT set IP_DF if its IPv4, we ignore the return
6127 	 * here since the timer will drive a retranmission.
6128 	 */
6129 	if (padding_len > 0) {
6130 		if (sctp_add_pad_tombuf(m_last, padding_len) == NULL) {
6131 			sctp_m_freem(m);
6132 			return;
6133 		}
6134 	}
6135 	if (stc.loopback_scope) {
6136 		over_addr = (union sctp_sockstore *)dst;
6137 	} else {
6138 		over_addr = NULL;
6139 	}
6140 
6141 	if ((error = sctp_lowlevel_chunk_output(inp, NULL, NULL, to, m, 0, NULL, 0, 0,
6142 	    0, 0,
6143 	    inp->sctp_lport, sh->src_port, init_chk->init.initiate_tag,
6144 	    port, over_addr,
6145 	    mflowtype, mflowid,
6146 	    SCTP_SO_NOT_LOCKED))) {
6147 		SCTPDBG(SCTP_DEBUG_OUTPUT4, "Gak send error %d\n", error);
6148 		if (error == ENOBUFS) {
6149 			if (asoc != NULL) {
6150 				asoc->ifp_had_enobuf = 1;
6151 			}
6152 			SCTP_STAT_INCR(sctps_lowlevelerr);
6153 		}
6154 	} else {
6155 		if (asoc != NULL) {
6156 			asoc->ifp_had_enobuf = 0;
6157 		}
6158 	}
6159 	SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
6160 }
6161 
6162 
6163 static void
6164 sctp_prune_prsctp(struct sctp_tcb *stcb,
6165     struct sctp_association *asoc,
6166     struct sctp_sndrcvinfo *srcv,
6167     int dataout)
6168 {
6169 	int freed_spc = 0;
6170 	struct sctp_tmit_chunk *chk, *nchk;
6171 
6172 	SCTP_TCB_LOCK_ASSERT(stcb);
6173 	if ((asoc->prsctp_supported) &&
6174 	    (asoc->sent_queue_cnt_removeable > 0)) {
6175 		TAILQ_FOREACH(chk, &asoc->sent_queue, sctp_next) {
6176 			/*
6177 			 * Look for chunks marked with the PR_SCTP flag AND
6178 			 * the buffer space flag. If the one being sent is
6179 			 * equal or greater priority then purge the old one
6180 			 * and free some space.
6181 			 */
6182 			if (PR_SCTP_BUF_ENABLED(chk->flags)) {
6183 				/*
6184 				 * This one is PR-SCTP AND buffer space
6185 				 * limited type
6186 				 */
6187 				if (chk->rec.data.timetodrop.tv_sec >= (long)srcv->sinfo_timetolive) {
6188 					/*
6189 					 * Lower numbers equates to higher
6190 					 * priority so if the one we are
6191 					 * looking at has a larger or equal
6192 					 * priority we want to drop the data
6193 					 * and NOT retransmit it.
6194 					 */
6195 					if (chk->data) {
6196 						/*
6197 						 * We release the book_size
6198 						 * if the mbuf is here
6199 						 */
6200 						int ret_spc;
6201 						uint8_t sent;
6202 
6203 						if (chk->sent > SCTP_DATAGRAM_UNSENT)
6204 							sent = 1;
6205 						else
6206 							sent = 0;
6207 						ret_spc = sctp_release_pr_sctp_chunk(stcb, chk,
6208 						    sent,
6209 						    SCTP_SO_LOCKED);
6210 						freed_spc += ret_spc;
6211 						if (freed_spc >= dataout) {
6212 							return;
6213 						}
6214 					}	/* if chunk was present */
6215 				}	/* if of sufficient priority */
6216 			}	/* if chunk has enabled */
6217 		}		/* tailqforeach */
6218 
6219 		TAILQ_FOREACH_SAFE(chk, &asoc->send_queue, sctp_next, nchk) {
6220 			/* Here we must move to the sent queue and mark */
6221 			if (PR_SCTP_BUF_ENABLED(chk->flags)) {
6222 				if (chk->rec.data.timetodrop.tv_sec >= (long)srcv->sinfo_timetolive) {
6223 					if (chk->data) {
6224 						/*
6225 						 * We release the book_size
6226 						 * if the mbuf is here
6227 						 */
6228 						int ret_spc;
6229 
6230 						ret_spc = sctp_release_pr_sctp_chunk(stcb, chk,
6231 						    0, SCTP_SO_LOCKED);
6232 
6233 						freed_spc += ret_spc;
6234 						if (freed_spc >= dataout) {
6235 							return;
6236 						}
6237 					}	/* end if chk->data */
6238 				}	/* end if right class */
6239 			}	/* end if chk pr-sctp */
6240 		}		/* tailqforeachsafe (chk) */
6241 	}			/* if enabled in asoc */
6242 }
6243 
6244 int
6245 sctp_get_frag_point(struct sctp_tcb *stcb,
6246     struct sctp_association *asoc)
6247 {
6248 	int siz, ovh;
6249 
6250 	/*
6251 	 * For endpoints that have both v6 and v4 addresses we must reserve
6252 	 * room for the ipv6 header, for those that are only dealing with V4
6253 	 * we use a larger frag point.
6254 	 */
6255 	if (stcb->sctp_ep->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) {
6256 		ovh = SCTP_MIN_OVERHEAD;
6257 	} else {
6258 		ovh = SCTP_MIN_V4_OVERHEAD;
6259 	}
6260 	ovh += SCTP_DATA_CHUNK_OVERHEAD(stcb);
6261 	if (stcb->asoc.sctp_frag_point > asoc->smallest_mtu)
6262 		siz = asoc->smallest_mtu - ovh;
6263 	else
6264 		siz = (stcb->asoc.sctp_frag_point - ovh);
6265 	/*
6266 	 * if (siz > (MCLBYTES-sizeof(struct sctp_data_chunk))) {
6267 	 */
6268 	/* A data chunk MUST fit in a cluster */
6269 	/* siz = (MCLBYTES - sizeof(struct sctp_data_chunk)); */
6270 	/* } */
6271 
6272 	/* adjust for an AUTH chunk if DATA requires auth */
6273 	if (sctp_auth_is_required_chunk(SCTP_DATA, stcb->asoc.peer_auth_chunks))
6274 		siz -= sctp_get_auth_chunk_len(stcb->asoc.peer_hmac_id);
6275 
6276 	if (siz % 4) {
6277 		/* make it an even word boundary please */
6278 		siz -= (siz % 4);
6279 	}
6280 	return (siz);
6281 }
6282 
6283 static void
6284 sctp_set_prsctp_policy(struct sctp_stream_queue_pending *sp)
6285 {
6286 	/*
6287 	 * We assume that the user wants PR_SCTP_TTL if the user provides a
6288 	 * positive lifetime but does not specify any PR_SCTP policy.
6289 	 */
6290 	if (PR_SCTP_ENABLED(sp->sinfo_flags)) {
6291 		sp->act_flags |= PR_SCTP_POLICY(sp->sinfo_flags);
6292 	} else if (sp->timetolive > 0) {
6293 		sp->sinfo_flags |= SCTP_PR_SCTP_TTL;
6294 		sp->act_flags |= PR_SCTP_POLICY(sp->sinfo_flags);
6295 	} else {
6296 		return;
6297 	}
6298 	switch (PR_SCTP_POLICY(sp->sinfo_flags)) {
6299 	case CHUNK_FLAGS_PR_SCTP_BUF:
6300 		/*
6301 		 * Time to live is a priority stored in tv_sec when doing
6302 		 * the buffer drop thing.
6303 		 */
6304 		sp->ts.tv_sec = sp->timetolive;
6305 		sp->ts.tv_usec = 0;
6306 		break;
6307 	case CHUNK_FLAGS_PR_SCTP_TTL:
6308 		{
6309 			struct timeval tv;
6310 
6311 			(void)SCTP_GETTIME_TIMEVAL(&sp->ts);
6312 			tv.tv_sec = sp->timetolive / 1000;
6313 			tv.tv_usec = (sp->timetolive * 1000) % 1000000;
6314 			/*
6315 			 * TODO sctp_constants.h needs alternative time
6316 			 * macros when _KERNEL is undefined.
6317 			 */
6318 			timevaladd(&sp->ts, &tv);
6319 		}
6320 		break;
6321 	case CHUNK_FLAGS_PR_SCTP_RTX:
6322 		/*
6323 		 * Time to live is a the number or retransmissions stored in
6324 		 * tv_sec.
6325 		 */
6326 		sp->ts.tv_sec = sp->timetolive;
6327 		sp->ts.tv_usec = 0;
6328 		break;
6329 	default:
6330 		SCTPDBG(SCTP_DEBUG_USRREQ1,
6331 		    "Unknown PR_SCTP policy %u.\n",
6332 		    PR_SCTP_POLICY(sp->sinfo_flags));
6333 		break;
6334 	}
6335 }
6336 
6337 static int
6338 sctp_msg_append(struct sctp_tcb *stcb,
6339     struct sctp_nets *net,
6340     struct mbuf *m,
6341     struct sctp_sndrcvinfo *srcv, int hold_stcb_lock)
6342 {
6343 	int error = 0;
6344 	struct mbuf *at;
6345 	struct sctp_stream_queue_pending *sp = NULL;
6346 	struct sctp_stream_out *strm;
6347 
6348 	/*
6349 	 * Given an mbuf chain, put it into the association send queue and
6350 	 * place it on the wheel
6351 	 */
6352 	if (srcv->sinfo_stream >= stcb->asoc.streamoutcnt) {
6353 		/* Invalid stream number */
6354 		SCTP_LTRACE_ERR_RET_PKT(m, NULL, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
6355 		error = EINVAL;
6356 		goto out_now;
6357 	}
6358 	if ((stcb->asoc.stream_locked) &&
6359 	    (stcb->asoc.stream_locked_on != srcv->sinfo_stream)) {
6360 		SCTP_LTRACE_ERR_RET_PKT(m, NULL, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
6361 		error = EINVAL;
6362 		goto out_now;
6363 	}
6364 	strm = &stcb->asoc.strmout[srcv->sinfo_stream];
6365 	/* Now can we send this? */
6366 	if ((SCTP_GET_STATE(stcb) == SCTP_STATE_SHUTDOWN_SENT) ||
6367 	    (SCTP_GET_STATE(stcb) == SCTP_STATE_SHUTDOWN_ACK_SENT) ||
6368 	    (SCTP_GET_STATE(stcb) == SCTP_STATE_SHUTDOWN_RECEIVED) ||
6369 	    (stcb->asoc.state & SCTP_STATE_SHUTDOWN_PENDING)) {
6370 		/* got data while shutting down */
6371 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ECONNRESET);
6372 		error = ECONNRESET;
6373 		goto out_now;
6374 	}
6375 	sctp_alloc_a_strmoq(stcb, sp);
6376 	if (sp == NULL) {
6377 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
6378 		error = ENOMEM;
6379 		goto out_now;
6380 	}
6381 	sp->sinfo_flags = srcv->sinfo_flags;
6382 	sp->timetolive = srcv->sinfo_timetolive;
6383 	sp->ppid = srcv->sinfo_ppid;
6384 	sp->context = srcv->sinfo_context;
6385 	sp->fsn = 0;
6386 	if (sp->sinfo_flags & SCTP_ADDR_OVER) {
6387 		sp->net = net;
6388 		atomic_add_int(&sp->net->ref_count, 1);
6389 	} else {
6390 		sp->net = NULL;
6391 	}
6392 	(void)SCTP_GETTIME_TIMEVAL(&sp->ts);
6393 	sp->sid = srcv->sinfo_stream;
6394 	sp->msg_is_complete = 1;
6395 	sp->sender_all_done = 1;
6396 	sp->some_taken = 0;
6397 	sp->data = m;
6398 	sp->tail_mbuf = NULL;
6399 	sctp_set_prsctp_policy(sp);
6400 	/*
6401 	 * We could in theory (for sendall) sifa the length in, but we would
6402 	 * still have to hunt through the chain since we need to setup the
6403 	 * tail_mbuf
6404 	 */
6405 	sp->length = 0;
6406 	for (at = m; at; at = SCTP_BUF_NEXT(at)) {
6407 		if (SCTP_BUF_NEXT(at) == NULL)
6408 			sp->tail_mbuf = at;
6409 		sp->length += SCTP_BUF_LEN(at);
6410 	}
6411 	if (srcv->sinfo_keynumber_valid) {
6412 		sp->auth_keyid = srcv->sinfo_keynumber;
6413 	} else {
6414 		sp->auth_keyid = stcb->asoc.authinfo.active_keyid;
6415 	}
6416 	if (sctp_auth_is_required_chunk(SCTP_DATA, stcb->asoc.peer_auth_chunks)) {
6417 		sctp_auth_key_acquire(stcb, sp->auth_keyid);
6418 		sp->holds_key_ref = 1;
6419 	}
6420 	if (hold_stcb_lock == 0) {
6421 		SCTP_TCB_SEND_LOCK(stcb);
6422 	}
6423 	sctp_snd_sb_alloc(stcb, sp->length);
6424 	atomic_add_int(&stcb->asoc.stream_queue_cnt, 1);
6425 	TAILQ_INSERT_TAIL(&strm->outqueue, sp, next);
6426 	stcb->asoc.ss_functions.sctp_ss_add_to_stream(stcb, &stcb->asoc, strm, sp, 1);
6427 	m = NULL;
6428 	if (hold_stcb_lock == 0) {
6429 		SCTP_TCB_SEND_UNLOCK(stcb);
6430 	}
6431 out_now:
6432 	if (m) {
6433 		sctp_m_freem(m);
6434 	}
6435 	return (error);
6436 }
6437 
6438 
6439 static struct mbuf *
6440 sctp_copy_mbufchain(struct mbuf *clonechain,
6441     struct mbuf *outchain,
6442     struct mbuf **endofchain,
6443     int can_take_mbuf,
6444     int sizeofcpy,
6445     uint8_t copy_by_ref)
6446 {
6447 	struct mbuf *m;
6448 	struct mbuf *appendchain;
6449 	caddr_t cp;
6450 	int len;
6451 
6452 	if (endofchain == NULL) {
6453 		/* error */
6454 error_out:
6455 		if (outchain)
6456 			sctp_m_freem(outchain);
6457 		return (NULL);
6458 	}
6459 	if (can_take_mbuf) {
6460 		appendchain = clonechain;
6461 	} else {
6462 		if (!copy_by_ref &&
6463 		    (sizeofcpy <= (int)((((SCTP_BASE_SYSCTL(sctp_mbuf_threshold_count) - 1) * MLEN) + MHLEN)))
6464 		    ) {
6465 			/* Its not in a cluster */
6466 			if (*endofchain == NULL) {
6467 				/* lets get a mbuf cluster */
6468 				if (outchain == NULL) {
6469 					/* This is the general case */
6470 			new_mbuf:
6471 					outchain = sctp_get_mbuf_for_msg(MCLBYTES, 0, M_NOWAIT, 1, MT_HEADER);
6472 					if (outchain == NULL) {
6473 						goto error_out;
6474 					}
6475 					SCTP_BUF_LEN(outchain) = 0;
6476 					*endofchain = outchain;
6477 					/* get the prepend space */
6478 					SCTP_BUF_RESV_UF(outchain, (SCTP_FIRST_MBUF_RESV + 4));
6479 				} else {
6480 					/*
6481 					 * We really should not get a NULL
6482 					 * in endofchain
6483 					 */
6484 					/* find end */
6485 					m = outchain;
6486 					while (m) {
6487 						if (SCTP_BUF_NEXT(m) == NULL) {
6488 							*endofchain = m;
6489 							break;
6490 						}
6491 						m = SCTP_BUF_NEXT(m);
6492 					}
6493 					/* sanity */
6494 					if (*endofchain == NULL) {
6495 						/*
6496 						 * huh, TSNH XXX maybe we
6497 						 * should panic
6498 						 */
6499 						sctp_m_freem(outchain);
6500 						goto new_mbuf;
6501 					}
6502 				}
6503 				/* get the new end of length */
6504 				len = (int)M_TRAILINGSPACE(*endofchain);
6505 			} else {
6506 				/* how much is left at the end? */
6507 				len = (int)M_TRAILINGSPACE(*endofchain);
6508 			}
6509 			/* Find the end of the data, for appending */
6510 			cp = (mtod((*endofchain), caddr_t)+SCTP_BUF_LEN((*endofchain)));
6511 
6512 			/* Now lets copy it out */
6513 			if (len >= sizeofcpy) {
6514 				/* It all fits, copy it in */
6515 				m_copydata(clonechain, 0, sizeofcpy, cp);
6516 				SCTP_BUF_LEN((*endofchain)) += sizeofcpy;
6517 			} else {
6518 				/* fill up the end of the chain */
6519 				if (len > 0) {
6520 					m_copydata(clonechain, 0, len, cp);
6521 					SCTP_BUF_LEN((*endofchain)) += len;
6522 					/* now we need another one */
6523 					sizeofcpy -= len;
6524 				}
6525 				m = sctp_get_mbuf_for_msg(MCLBYTES, 0, M_NOWAIT, 1, MT_HEADER);
6526 				if (m == NULL) {
6527 					/* We failed */
6528 					goto error_out;
6529 				}
6530 				SCTP_BUF_NEXT((*endofchain)) = m;
6531 				*endofchain = m;
6532 				cp = mtod((*endofchain), caddr_t);
6533 				m_copydata(clonechain, len, sizeofcpy, cp);
6534 				SCTP_BUF_LEN((*endofchain)) += sizeofcpy;
6535 			}
6536 			return (outchain);
6537 		} else {
6538 			/* copy the old fashion way */
6539 			appendchain = SCTP_M_COPYM(clonechain, 0, M_COPYALL, M_NOWAIT);
6540 #ifdef SCTP_MBUF_LOGGING
6541 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_MBUF_LOGGING_ENABLE) {
6542 				sctp_log_mbc(appendchain, SCTP_MBUF_ICOPY);
6543 			}
6544 #endif
6545 		}
6546 	}
6547 	if (appendchain == NULL) {
6548 		/* error */
6549 		if (outchain)
6550 			sctp_m_freem(outchain);
6551 		return (NULL);
6552 	}
6553 	if (outchain) {
6554 		/* tack on to the end */
6555 		if (*endofchain != NULL) {
6556 			SCTP_BUF_NEXT(((*endofchain))) = appendchain;
6557 		} else {
6558 			m = outchain;
6559 			while (m) {
6560 				if (SCTP_BUF_NEXT(m) == NULL) {
6561 					SCTP_BUF_NEXT(m) = appendchain;
6562 					break;
6563 				}
6564 				m = SCTP_BUF_NEXT(m);
6565 			}
6566 		}
6567 		/*
6568 		 * save off the end and update the end-chain position
6569 		 */
6570 		m = appendchain;
6571 		while (m) {
6572 			if (SCTP_BUF_NEXT(m) == NULL) {
6573 				*endofchain = m;
6574 				break;
6575 			}
6576 			m = SCTP_BUF_NEXT(m);
6577 		}
6578 		return (outchain);
6579 	} else {
6580 		/* save off the end and update the end-chain position */
6581 		m = appendchain;
6582 		while (m) {
6583 			if (SCTP_BUF_NEXT(m) == NULL) {
6584 				*endofchain = m;
6585 				break;
6586 			}
6587 			m = SCTP_BUF_NEXT(m);
6588 		}
6589 		return (appendchain);
6590 	}
6591 }
6592 
6593 static int
6594 sctp_med_chunk_output(struct sctp_inpcb *inp,
6595     struct sctp_tcb *stcb,
6596     struct sctp_association *asoc,
6597     int *num_out,
6598     int *reason_code,
6599     int control_only, int from_where,
6600     struct timeval *now, int *now_filled, int frag_point, int so_locked
6601 #if !defined(__APPLE__) && !defined(SCTP_SO_LOCK_TESTING)
6602     SCTP_UNUSED
6603 #endif
6604 );
6605 
6606 static void
6607 sctp_sendall_iterator(struct sctp_inpcb *inp, struct sctp_tcb *stcb, void *ptr,
6608     uint32_t val SCTP_UNUSED)
6609 {
6610 	struct sctp_copy_all *ca;
6611 	struct mbuf *m;
6612 	int ret = 0;
6613 	int added_control = 0;
6614 	int un_sent, do_chunk_output = 1;
6615 	struct sctp_association *asoc;
6616 	struct sctp_nets *net;
6617 
6618 	ca = (struct sctp_copy_all *)ptr;
6619 	if (ca->m == NULL) {
6620 		return;
6621 	}
6622 	if (ca->inp != inp) {
6623 		/* TSNH */
6624 		return;
6625 	}
6626 	if (ca->sndlen > 0) {
6627 		m = SCTP_M_COPYM(ca->m, 0, M_COPYALL, M_NOWAIT);
6628 		if (m == NULL) {
6629 			/* can't copy so we are done */
6630 			ca->cnt_failed++;
6631 			return;
6632 		}
6633 #ifdef SCTP_MBUF_LOGGING
6634 		if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_MBUF_LOGGING_ENABLE) {
6635 			sctp_log_mbc(m, SCTP_MBUF_ICOPY);
6636 		}
6637 #endif
6638 	} else {
6639 		m = NULL;
6640 	}
6641 	SCTP_TCB_LOCK_ASSERT(stcb);
6642 	if (stcb->asoc.alternate) {
6643 		net = stcb->asoc.alternate;
6644 	} else {
6645 		net = stcb->asoc.primary_destination;
6646 	}
6647 	if (ca->sndrcv.sinfo_flags & SCTP_ABORT) {
6648 		/* Abort this assoc with m as the user defined reason */
6649 		if (m != NULL) {
6650 			SCTP_BUF_PREPEND(m, sizeof(struct sctp_paramhdr), M_NOWAIT);
6651 		} else {
6652 			m = sctp_get_mbuf_for_msg(sizeof(struct sctp_paramhdr),
6653 			    0, M_NOWAIT, 1, MT_DATA);
6654 			SCTP_BUF_LEN(m) = sizeof(struct sctp_paramhdr);
6655 		}
6656 		if (m != NULL) {
6657 			struct sctp_paramhdr *ph;
6658 
6659 			ph = mtod(m, struct sctp_paramhdr *);
6660 			ph->param_type = htons(SCTP_CAUSE_USER_INITIATED_ABT);
6661 			ph->param_length = htons((uint16_t)(sizeof(struct sctp_paramhdr) + ca->sndlen));
6662 		}
6663 		/*
6664 		 * We add one here to keep the assoc from dis-appearing on
6665 		 * us.
6666 		 */
6667 		atomic_add_int(&stcb->asoc.refcnt, 1);
6668 		sctp_abort_an_association(inp, stcb, m, SCTP_SO_NOT_LOCKED);
6669 		/*
6670 		 * sctp_abort_an_association calls sctp_free_asoc() free
6671 		 * association will NOT free it since we incremented the
6672 		 * refcnt .. we do this to prevent it being freed and things
6673 		 * getting tricky since we could end up (from free_asoc)
6674 		 * calling inpcb_free which would get a recursive lock call
6675 		 * to the iterator lock.. But as a consequence of that the
6676 		 * stcb will return to us un-locked.. since free_asoc
6677 		 * returns with either no TCB or the TCB unlocked, we must
6678 		 * relock.. to unlock in the iterator timer :-0
6679 		 */
6680 		SCTP_TCB_LOCK(stcb);
6681 		atomic_add_int(&stcb->asoc.refcnt, -1);
6682 		goto no_chunk_output;
6683 	} else {
6684 		if (m) {
6685 			ret = sctp_msg_append(stcb, net, m,
6686 			    &ca->sndrcv, 1);
6687 		}
6688 		asoc = &stcb->asoc;
6689 		if (ca->sndrcv.sinfo_flags & SCTP_EOF) {
6690 			/* shutdown this assoc */
6691 			if (TAILQ_EMPTY(&asoc->send_queue) &&
6692 			    TAILQ_EMPTY(&asoc->sent_queue) &&
6693 			    sctp_is_there_unsent_data(stcb, SCTP_SO_NOT_LOCKED) == 0) {
6694 				if ((*asoc->ss_functions.sctp_ss_is_user_msgs_incomplete) (stcb, asoc)) {
6695 					goto abort_anyway;
6696 				}
6697 				/*
6698 				 * there is nothing queued to send, so I'm
6699 				 * done...
6700 				 */
6701 				if ((SCTP_GET_STATE(stcb) != SCTP_STATE_SHUTDOWN_SENT) &&
6702 				    (SCTP_GET_STATE(stcb) != SCTP_STATE_SHUTDOWN_RECEIVED) &&
6703 				    (SCTP_GET_STATE(stcb) != SCTP_STATE_SHUTDOWN_ACK_SENT)) {
6704 					/*
6705 					 * only send SHUTDOWN the first time
6706 					 * through
6707 					 */
6708 					if (SCTP_GET_STATE(stcb) == SCTP_STATE_OPEN) {
6709 						SCTP_STAT_DECR_GAUGE32(sctps_currestab);
6710 					}
6711 					SCTP_SET_STATE(stcb, SCTP_STATE_SHUTDOWN_SENT);
6712 					SCTP_CLEAR_SUBSTATE(stcb, SCTP_STATE_SHUTDOWN_PENDING);
6713 					sctp_stop_timers_for_shutdown(stcb);
6714 					sctp_send_shutdown(stcb, net);
6715 					sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWN, stcb->sctp_ep, stcb,
6716 					    net);
6717 					sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWNGUARD, stcb->sctp_ep, stcb,
6718 					    asoc->primary_destination);
6719 					added_control = 1;
6720 					do_chunk_output = 0;
6721 				}
6722 			} else {
6723 				/*
6724 				 * we still got (or just got) data to send,
6725 				 * so set SHUTDOWN_PENDING
6726 				 */
6727 				/*
6728 				 * XXX sockets draft says that SCTP_EOF
6729 				 * should be sent with no data.  currently,
6730 				 * we will allow user data to be sent first
6731 				 * and move to SHUTDOWN-PENDING
6732 				 */
6733 				if ((SCTP_GET_STATE(stcb) != SCTP_STATE_SHUTDOWN_SENT) &&
6734 				    (SCTP_GET_STATE(stcb) != SCTP_STATE_SHUTDOWN_RECEIVED) &&
6735 				    (SCTP_GET_STATE(stcb) != SCTP_STATE_SHUTDOWN_ACK_SENT)) {
6736 					if ((*asoc->ss_functions.sctp_ss_is_user_msgs_incomplete) (stcb, asoc)) {
6737 						SCTP_ADD_SUBSTATE(stcb, SCTP_STATE_PARTIAL_MSG_LEFT);
6738 					}
6739 					SCTP_ADD_SUBSTATE(stcb, SCTP_STATE_SHUTDOWN_PENDING);
6740 					if (TAILQ_EMPTY(&asoc->send_queue) &&
6741 					    TAILQ_EMPTY(&asoc->sent_queue) &&
6742 					    (asoc->state & SCTP_STATE_PARTIAL_MSG_LEFT)) {
6743 						struct mbuf *op_err;
6744 						char msg[SCTP_DIAG_INFO_LEN];
6745 
6746 				abort_anyway:
6747 						snprintf(msg, sizeof(msg),
6748 						    "%s:%d at %s", __FILE__, __LINE__, __func__);
6749 						op_err = sctp_generate_cause(SCTP_BASE_SYSCTL(sctp_diag_info_code),
6750 						    msg);
6751 						atomic_add_int(&stcb->asoc.refcnt, 1);
6752 						sctp_abort_an_association(stcb->sctp_ep, stcb,
6753 						    op_err, SCTP_SO_NOT_LOCKED);
6754 						atomic_add_int(&stcb->asoc.refcnt, -1);
6755 						goto no_chunk_output;
6756 					}
6757 					sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWNGUARD, stcb->sctp_ep, stcb,
6758 					    asoc->primary_destination);
6759 				}
6760 			}
6761 
6762 		}
6763 	}
6764 	un_sent = ((stcb->asoc.total_output_queue_size - stcb->asoc.total_flight) +
6765 	    (stcb->asoc.stream_queue_cnt * SCTP_DATA_CHUNK_OVERHEAD(stcb)));
6766 
6767 	if ((sctp_is_feature_off(inp, SCTP_PCB_FLAGS_NODELAY)) &&
6768 	    (stcb->asoc.total_flight > 0) &&
6769 	    (un_sent < (int)(stcb->asoc.smallest_mtu - SCTP_MIN_OVERHEAD))) {
6770 		do_chunk_output = 0;
6771 	}
6772 	if (do_chunk_output)
6773 		sctp_chunk_output(inp, stcb, SCTP_OUTPUT_FROM_USR_SEND, SCTP_SO_NOT_LOCKED);
6774 	else if (added_control) {
6775 		int num_out, reason, now_filled = 0;
6776 		struct timeval now;
6777 		int frag_point;
6778 
6779 		frag_point = sctp_get_frag_point(stcb, &stcb->asoc);
6780 		(void)sctp_med_chunk_output(inp, stcb, &stcb->asoc, &num_out,
6781 		    &reason, 1, 1, &now, &now_filled, frag_point, SCTP_SO_NOT_LOCKED);
6782 	}
6783 no_chunk_output:
6784 	if (ret) {
6785 		ca->cnt_failed++;
6786 	} else {
6787 		ca->cnt_sent++;
6788 	}
6789 }
6790 
6791 static void
6792 sctp_sendall_completes(void *ptr, uint32_t val SCTP_UNUSED)
6793 {
6794 	struct sctp_copy_all *ca;
6795 
6796 	ca = (struct sctp_copy_all *)ptr;
6797 	/*
6798 	 * Do a notify here? Kacheong suggests that the notify be done at
6799 	 * the send time.. so you would push up a notification if any send
6800 	 * failed. Don't know if this is feasible since the only failures we
6801 	 * have is "memory" related and if you cannot get an mbuf to send
6802 	 * the data you surely can't get an mbuf to send up to notify the
6803 	 * user you can't send the data :->
6804 	 */
6805 
6806 	/* now free everything */
6807 	sctp_m_freem(ca->m);
6808 	SCTP_FREE(ca, SCTP_M_COPYAL);
6809 }
6810 
6811 static struct mbuf *
6812 sctp_copy_out_all(struct uio *uio, int len)
6813 {
6814 	struct mbuf *ret, *at;
6815 	int left, willcpy, cancpy, error;
6816 
6817 	ret = sctp_get_mbuf_for_msg(MCLBYTES, 0, M_WAITOK, 1, MT_DATA);
6818 	if (ret == NULL) {
6819 		/* TSNH */
6820 		return (NULL);
6821 	}
6822 	left = len;
6823 	SCTP_BUF_LEN(ret) = 0;
6824 	/* save space for the data chunk header */
6825 	cancpy = (int)M_TRAILINGSPACE(ret);
6826 	willcpy = min(cancpy, left);
6827 	at = ret;
6828 	while (left > 0) {
6829 		/* Align data to the end */
6830 		error = uiomove(mtod(at, caddr_t), willcpy, uio);
6831 		if (error) {
6832 	err_out_now:
6833 			sctp_m_freem(at);
6834 			return (NULL);
6835 		}
6836 		SCTP_BUF_LEN(at) = willcpy;
6837 		SCTP_BUF_NEXT_PKT(at) = SCTP_BUF_NEXT(at) = 0;
6838 		left -= willcpy;
6839 		if (left > 0) {
6840 			SCTP_BUF_NEXT(at) = sctp_get_mbuf_for_msg(left, 0, M_WAITOK, 1, MT_DATA);
6841 			if (SCTP_BUF_NEXT(at) == NULL) {
6842 				goto err_out_now;
6843 			}
6844 			at = SCTP_BUF_NEXT(at);
6845 			SCTP_BUF_LEN(at) = 0;
6846 			cancpy = (int)M_TRAILINGSPACE(at);
6847 			willcpy = min(cancpy, left);
6848 		}
6849 	}
6850 	return (ret);
6851 }
6852 
6853 static int
6854 sctp_sendall(struct sctp_inpcb *inp, struct uio *uio, struct mbuf *m,
6855     struct sctp_sndrcvinfo *srcv)
6856 {
6857 	int ret;
6858 	struct sctp_copy_all *ca;
6859 
6860 	SCTP_MALLOC(ca, struct sctp_copy_all *, sizeof(struct sctp_copy_all),
6861 	    SCTP_M_COPYAL);
6862 	if (ca == NULL) {
6863 		sctp_m_freem(m);
6864 		SCTP_LTRACE_ERR_RET(inp, NULL, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
6865 		return (ENOMEM);
6866 	}
6867 	memset(ca, 0, sizeof(struct sctp_copy_all));
6868 
6869 	ca->inp = inp;
6870 	if (srcv) {
6871 		memcpy(&ca->sndrcv, srcv, sizeof(struct sctp_nonpad_sndrcvinfo));
6872 	}
6873 	/*
6874 	 * take off the sendall flag, it would be bad if we failed to do
6875 	 * this :-0
6876 	 */
6877 	ca->sndrcv.sinfo_flags &= ~SCTP_SENDALL;
6878 	/* get length and mbuf chain */
6879 	if (uio) {
6880 		ca->sndlen = (int)uio->uio_resid;
6881 		ca->m = sctp_copy_out_all(uio, ca->sndlen);
6882 		if (ca->m == NULL) {
6883 			SCTP_FREE(ca, SCTP_M_COPYAL);
6884 			SCTP_LTRACE_ERR_RET(inp, NULL, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
6885 			return (ENOMEM);
6886 		}
6887 	} else {
6888 		/* Gather the length of the send */
6889 		struct mbuf *mat;
6890 
6891 		ca->sndlen = 0;
6892 		for (mat = m; mat; mat = SCTP_BUF_NEXT(mat)) {
6893 			ca->sndlen += SCTP_BUF_LEN(mat);
6894 		}
6895 	}
6896 	ret = sctp_initiate_iterator(NULL, sctp_sendall_iterator, NULL,
6897 	    SCTP_PCB_ANY_FLAGS, SCTP_PCB_ANY_FEATURES,
6898 	    SCTP_ASOC_ANY_STATE,
6899 	    (void *)ca, 0,
6900 	    sctp_sendall_completes, inp, 1);
6901 	if (ret) {
6902 		SCTP_PRINTF("Failed to initiate iterator for sendall\n");
6903 		SCTP_FREE(ca, SCTP_M_COPYAL);
6904 		SCTP_LTRACE_ERR_RET_PKT(m, inp, NULL, NULL, SCTP_FROM_SCTP_OUTPUT, EFAULT);
6905 		return (EFAULT);
6906 	}
6907 	return (0);
6908 }
6909 
6910 
6911 void
6912 sctp_toss_old_cookies(struct sctp_tcb *stcb, struct sctp_association *asoc)
6913 {
6914 	struct sctp_tmit_chunk *chk, *nchk;
6915 
6916 	TAILQ_FOREACH_SAFE(chk, &asoc->control_send_queue, sctp_next, nchk) {
6917 		if (chk->rec.chunk_id.id == SCTP_COOKIE_ECHO) {
6918 			TAILQ_REMOVE(&asoc->control_send_queue, chk, sctp_next);
6919 			asoc->ctrl_queue_cnt--;
6920 			if (chk->data) {
6921 				sctp_m_freem(chk->data);
6922 				chk->data = NULL;
6923 			}
6924 			sctp_free_a_chunk(stcb, chk, SCTP_SO_NOT_LOCKED);
6925 		}
6926 	}
6927 }
6928 
6929 void
6930 sctp_toss_old_asconf(struct sctp_tcb *stcb)
6931 {
6932 	struct sctp_association *asoc;
6933 	struct sctp_tmit_chunk *chk, *nchk;
6934 	struct sctp_asconf_chunk *acp;
6935 
6936 	asoc = &stcb->asoc;
6937 	TAILQ_FOREACH_SAFE(chk, &asoc->asconf_send_queue, sctp_next, nchk) {
6938 		/* find SCTP_ASCONF chunk in queue */
6939 		if (chk->rec.chunk_id.id == SCTP_ASCONF) {
6940 			if (chk->data) {
6941 				acp = mtod(chk->data, struct sctp_asconf_chunk *);
6942 				if (SCTP_TSN_GT(ntohl(acp->serial_number), asoc->asconf_seq_out_acked)) {
6943 					/* Not Acked yet */
6944 					break;
6945 				}
6946 			}
6947 			TAILQ_REMOVE(&asoc->asconf_send_queue, chk, sctp_next);
6948 			asoc->ctrl_queue_cnt--;
6949 			if (chk->data) {
6950 				sctp_m_freem(chk->data);
6951 				chk->data = NULL;
6952 			}
6953 			sctp_free_a_chunk(stcb, chk, SCTP_SO_NOT_LOCKED);
6954 		}
6955 	}
6956 }
6957 
6958 
6959 static void
6960 sctp_clean_up_datalist(struct sctp_tcb *stcb,
6961     struct sctp_association *asoc,
6962     struct sctp_tmit_chunk **data_list,
6963     int bundle_at,
6964     struct sctp_nets *net)
6965 {
6966 	int i;
6967 	struct sctp_tmit_chunk *tp1;
6968 
6969 	for (i = 0; i < bundle_at; i++) {
6970 		/* off of the send queue */
6971 		TAILQ_REMOVE(&asoc->send_queue, data_list[i], sctp_next);
6972 		asoc->send_queue_cnt--;
6973 		if (i > 0) {
6974 			/*
6975 			 * Any chunk NOT 0 you zap the time chunk 0 gets
6976 			 * zapped or set based on if a RTO measurment is
6977 			 * needed.
6978 			 */
6979 			data_list[i]->do_rtt = 0;
6980 		}
6981 		/* record time */
6982 		data_list[i]->sent_rcv_time = net->last_sent_time;
6983 		data_list[i]->rec.data.cwnd_at_send = net->cwnd;
6984 		data_list[i]->rec.data.fast_retran_tsn = data_list[i]->rec.data.tsn;
6985 		if (data_list[i]->whoTo == NULL) {
6986 			data_list[i]->whoTo = net;
6987 			atomic_add_int(&net->ref_count, 1);
6988 		}
6989 		/* on to the sent queue */
6990 		tp1 = TAILQ_LAST(&asoc->sent_queue, sctpchunk_listhead);
6991 		if ((tp1) && SCTP_TSN_GT(tp1->rec.data.tsn, data_list[i]->rec.data.tsn)) {
6992 			struct sctp_tmit_chunk *tpp;
6993 
6994 			/* need to move back */
6995 	back_up_more:
6996 			tpp = TAILQ_PREV(tp1, sctpchunk_listhead, sctp_next);
6997 			if (tpp == NULL) {
6998 				TAILQ_INSERT_BEFORE(tp1, data_list[i], sctp_next);
6999 				goto all_done;
7000 			}
7001 			tp1 = tpp;
7002 			if (SCTP_TSN_GT(tp1->rec.data.tsn, data_list[i]->rec.data.tsn)) {
7003 				goto back_up_more;
7004 			}
7005 			TAILQ_INSERT_AFTER(&asoc->sent_queue, tp1, data_list[i], sctp_next);
7006 		} else {
7007 			TAILQ_INSERT_TAIL(&asoc->sent_queue,
7008 			    data_list[i],
7009 			    sctp_next);
7010 		}
7011 all_done:
7012 		/* This does not lower until the cum-ack passes it */
7013 		asoc->sent_queue_cnt++;
7014 		if ((asoc->peers_rwnd <= 0) &&
7015 		    (asoc->total_flight == 0) &&
7016 		    (bundle_at == 1)) {
7017 			/* Mark the chunk as being a window probe */
7018 			SCTP_STAT_INCR(sctps_windowprobed);
7019 		}
7020 #ifdef SCTP_AUDITING_ENABLED
7021 		sctp_audit_log(0xC2, 3);
7022 #endif
7023 		data_list[i]->sent = SCTP_DATAGRAM_SENT;
7024 		data_list[i]->snd_count = 1;
7025 		data_list[i]->rec.data.chunk_was_revoked = 0;
7026 		if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_FLIGHT_LOGGING_ENABLE) {
7027 			sctp_misc_ints(SCTP_FLIGHT_LOG_UP,
7028 			    data_list[i]->whoTo->flight_size,
7029 			    data_list[i]->book_size,
7030 			    (uint32_t)(uintptr_t)data_list[i]->whoTo,
7031 			    data_list[i]->rec.data.tsn);
7032 		}
7033 		sctp_flight_size_increase(data_list[i]);
7034 		sctp_total_flight_increase(stcb, data_list[i]);
7035 		if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOG_RWND_ENABLE) {
7036 			sctp_log_rwnd(SCTP_DECREASE_PEER_RWND,
7037 			    asoc->peers_rwnd, data_list[i]->send_size, SCTP_BASE_SYSCTL(sctp_peer_chunk_oh));
7038 		}
7039 		asoc->peers_rwnd = sctp_sbspace_sub(asoc->peers_rwnd,
7040 		    (uint32_t)(data_list[i]->send_size + SCTP_BASE_SYSCTL(sctp_peer_chunk_oh)));
7041 		if (asoc->peers_rwnd < stcb->sctp_ep->sctp_ep.sctp_sws_sender) {
7042 			/* SWS sender side engages */
7043 			asoc->peers_rwnd = 0;
7044 		}
7045 	}
7046 	if (asoc->cc_functions.sctp_cwnd_update_packet_transmitted) {
7047 		(*asoc->cc_functions.sctp_cwnd_update_packet_transmitted) (stcb, net);
7048 	}
7049 }
7050 
7051 static void
7052 sctp_clean_up_ctl(struct sctp_tcb *stcb, struct sctp_association *asoc, int so_locked
7053 #if !defined(__APPLE__) && !defined(SCTP_SO_LOCK_TESTING)
7054     SCTP_UNUSED
7055 #endif
7056 )
7057 {
7058 	struct sctp_tmit_chunk *chk, *nchk;
7059 
7060 	TAILQ_FOREACH_SAFE(chk, &asoc->control_send_queue, sctp_next, nchk) {
7061 		if ((chk->rec.chunk_id.id == SCTP_SELECTIVE_ACK) ||
7062 		    (chk->rec.chunk_id.id == SCTP_NR_SELECTIVE_ACK) ||	/* EY */
7063 		    (chk->rec.chunk_id.id == SCTP_HEARTBEAT_REQUEST) ||
7064 		    (chk->rec.chunk_id.id == SCTP_HEARTBEAT_ACK) ||
7065 		    (chk->rec.chunk_id.id == SCTP_FORWARD_CUM_TSN) ||
7066 		    (chk->rec.chunk_id.id == SCTP_SHUTDOWN) ||
7067 		    (chk->rec.chunk_id.id == SCTP_SHUTDOWN_ACK) ||
7068 		    (chk->rec.chunk_id.id == SCTP_OPERATION_ERROR) ||
7069 		    (chk->rec.chunk_id.id == SCTP_PACKET_DROPPED) ||
7070 		    (chk->rec.chunk_id.id == SCTP_COOKIE_ACK) ||
7071 		    (chk->rec.chunk_id.id == SCTP_ECN_CWR) ||
7072 		    (chk->rec.chunk_id.id == SCTP_ASCONF_ACK)) {
7073 			/* Stray chunks must be cleaned up */
7074 	clean_up_anyway:
7075 			TAILQ_REMOVE(&asoc->control_send_queue, chk, sctp_next);
7076 			asoc->ctrl_queue_cnt--;
7077 			if (chk->data) {
7078 				sctp_m_freem(chk->data);
7079 				chk->data = NULL;
7080 			}
7081 			if (chk->rec.chunk_id.id == SCTP_FORWARD_CUM_TSN) {
7082 				asoc->fwd_tsn_cnt--;
7083 			}
7084 			sctp_free_a_chunk(stcb, chk, so_locked);
7085 		} else if (chk->rec.chunk_id.id == SCTP_STREAM_RESET) {
7086 			/* special handling, we must look into the param */
7087 			if (chk != asoc->str_reset) {
7088 				goto clean_up_anyway;
7089 			}
7090 		}
7091 	}
7092 }
7093 
7094 static uint32_t
7095 sctp_can_we_split_this(struct sctp_tcb *stcb, uint32_t length,
7096     uint32_t space_left, uint32_t frag_point, int eeor_on)
7097 {
7098 	/*
7099 	 * Make a decision on if I should split a msg into multiple parts.
7100 	 * This is only asked of incomplete messages.
7101 	 */
7102 	if (eeor_on) {
7103 		/*
7104 		 * If we are doing EEOR we need to always send it if its the
7105 		 * entire thing, since it might be all the guy is putting in
7106 		 * the hopper.
7107 		 */
7108 		if (space_left >= length) {
7109 			/*-
7110 			 * If we have data outstanding,
7111 			 * we get another chance when the sack
7112 			 * arrives to transmit - wait for more data
7113 			 */
7114 			if (stcb->asoc.total_flight == 0) {
7115 				/*
7116 				 * If nothing is in flight, we zero the
7117 				 * packet counter.
7118 				 */
7119 				return (length);
7120 			}
7121 			return (0);
7122 
7123 		} else {
7124 			/* You can fill the rest */
7125 			return (space_left);
7126 		}
7127 	}
7128 	/*-
7129 	 * For those strange folk that make the send buffer
7130 	 * smaller than our fragmentation point, we can't
7131 	 * get a full msg in so we have to allow splitting.
7132 	 */
7133 	if (SCTP_SB_LIMIT_SND(stcb->sctp_socket) < frag_point) {
7134 		return (length);
7135 	}
7136 	if ((length <= space_left) ||
7137 	    ((length - space_left) < SCTP_BASE_SYSCTL(sctp_min_residual))) {
7138 		/* Sub-optimial residual don't split in non-eeor mode. */
7139 		return (0);
7140 	}
7141 	/*
7142 	 * If we reach here length is larger than the space_left. Do we wish
7143 	 * to split it for the sake of packet putting together?
7144 	 */
7145 	if (space_left >= min(SCTP_BASE_SYSCTL(sctp_min_split_point), frag_point)) {
7146 		/* Its ok to split it */
7147 		return (min(space_left, frag_point));
7148 	}
7149 	/* Nope, can't split */
7150 	return (0);
7151 }
7152 
7153 static uint32_t
7154 sctp_move_to_outqueue(struct sctp_tcb *stcb,
7155     struct sctp_stream_out *strq,
7156     uint32_t space_left,
7157     uint32_t frag_point,
7158     int *giveup,
7159     int eeor_mode,
7160     int *bail,
7161     int so_locked
7162 #if !defined(__APPLE__) && !defined(SCTP_SO_LOCK_TESTING)
7163     SCTP_UNUSED
7164 #endif
7165 )
7166 {
7167 	/* Move from the stream to the send_queue keeping track of the total */
7168 	struct sctp_association *asoc;
7169 	struct sctp_stream_queue_pending *sp;
7170 	struct sctp_tmit_chunk *chk;
7171 	struct sctp_data_chunk *dchkh = NULL;
7172 	struct sctp_idata_chunk *ndchkh = NULL;
7173 	uint32_t to_move, length;
7174 	int leading;
7175 	uint8_t rcv_flags = 0;
7176 	uint8_t some_taken;
7177 	uint8_t send_lock_up = 0;
7178 
7179 	SCTP_TCB_LOCK_ASSERT(stcb);
7180 	asoc = &stcb->asoc;
7181 one_more_time:
7182 	/* sa_ignore FREED_MEMORY */
7183 	sp = TAILQ_FIRST(&strq->outqueue);
7184 	if (sp == NULL) {
7185 		if (send_lock_up == 0) {
7186 			SCTP_TCB_SEND_LOCK(stcb);
7187 			send_lock_up = 1;
7188 		}
7189 		sp = TAILQ_FIRST(&strq->outqueue);
7190 		if (sp) {
7191 			goto one_more_time;
7192 		}
7193 		if ((sctp_is_feature_on(stcb->sctp_ep, SCTP_PCB_FLAGS_EXPLICIT_EOR) == 0) &&
7194 		    (stcb->asoc.idata_supported == 0) &&
7195 		    (strq->last_msg_incomplete)) {
7196 			SCTP_PRINTF("Huh? Stream:%d lm_in_c=%d but queue is NULL\n",
7197 			    strq->sid,
7198 			    strq->last_msg_incomplete);
7199 			strq->last_msg_incomplete = 0;
7200 		}
7201 		to_move = 0;
7202 		if (send_lock_up) {
7203 			SCTP_TCB_SEND_UNLOCK(stcb);
7204 			send_lock_up = 0;
7205 		}
7206 		goto out_of;
7207 	}
7208 	if ((sp->msg_is_complete) && (sp->length == 0)) {
7209 		if (sp->sender_all_done) {
7210 			/*
7211 			 * We are doing differed cleanup. Last time through
7212 			 * when we took all the data the sender_all_done was
7213 			 * not set.
7214 			 */
7215 			if ((sp->put_last_out == 0) && (sp->discard_rest == 0)) {
7216 				SCTP_PRINTF("Gak, put out entire msg with NO end!-1\n");
7217 				SCTP_PRINTF("sender_done:%d len:%d msg_comp:%d put_last_out:%d send_lock:%d\n",
7218 				    sp->sender_all_done,
7219 				    sp->length,
7220 				    sp->msg_is_complete,
7221 				    sp->put_last_out,
7222 				    send_lock_up);
7223 			}
7224 			if ((TAILQ_NEXT(sp, next) == NULL) && (send_lock_up == 0)) {
7225 				SCTP_TCB_SEND_LOCK(stcb);
7226 				send_lock_up = 1;
7227 			}
7228 			atomic_subtract_int(&asoc->stream_queue_cnt, 1);
7229 			TAILQ_REMOVE(&strq->outqueue, sp, next);
7230 			stcb->asoc.ss_functions.sctp_ss_remove_from_stream(stcb, asoc, strq, sp, send_lock_up);
7231 			if ((strq->state == SCTP_STREAM_RESET_PENDING) &&
7232 			    (strq->chunks_on_queues == 0) &&
7233 			    TAILQ_EMPTY(&strq->outqueue)) {
7234 				stcb->asoc.trigger_reset = 1;
7235 			}
7236 			if (sp->net) {
7237 				sctp_free_remote_addr(sp->net);
7238 				sp->net = NULL;
7239 			}
7240 			if (sp->data) {
7241 				sctp_m_freem(sp->data);
7242 				sp->data = NULL;
7243 			}
7244 			sctp_free_a_strmoq(stcb, sp, so_locked);
7245 			/* we can't be locked to it */
7246 			if (send_lock_up) {
7247 				SCTP_TCB_SEND_UNLOCK(stcb);
7248 				send_lock_up = 0;
7249 			}
7250 			/* back to get the next msg */
7251 			goto one_more_time;
7252 		} else {
7253 			/*
7254 			 * sender just finished this but still holds a
7255 			 * reference
7256 			 */
7257 			*giveup = 1;
7258 			to_move = 0;
7259 			goto out_of;
7260 		}
7261 	} else {
7262 		/* is there some to get */
7263 		if (sp->length == 0) {
7264 			/* no */
7265 			*giveup = 1;
7266 			to_move = 0;
7267 			goto out_of;
7268 		} else if (sp->discard_rest) {
7269 			if (send_lock_up == 0) {
7270 				SCTP_TCB_SEND_LOCK(stcb);
7271 				send_lock_up = 1;
7272 			}
7273 			/* Whack down the size */
7274 			atomic_subtract_int(&stcb->asoc.total_output_queue_size, sp->length);
7275 			if ((stcb->sctp_socket != NULL) &&
7276 			    ((stcb->sctp_ep->sctp_flags & SCTP_PCB_FLAGS_TCPTYPE) ||
7277 			    (stcb->sctp_ep->sctp_flags & SCTP_PCB_FLAGS_IN_TCPPOOL))) {
7278 				atomic_subtract_int(&stcb->sctp_socket->so_snd.sb_cc, sp->length);
7279 			}
7280 			if (sp->data) {
7281 				sctp_m_freem(sp->data);
7282 				sp->data = NULL;
7283 				sp->tail_mbuf = NULL;
7284 			}
7285 			sp->length = 0;
7286 			sp->some_taken = 1;
7287 			*giveup = 1;
7288 			to_move = 0;
7289 			goto out_of;
7290 		}
7291 	}
7292 	some_taken = sp->some_taken;
7293 re_look:
7294 	length = sp->length;
7295 	if (sp->msg_is_complete) {
7296 		/* The message is complete */
7297 		to_move = min(length, frag_point);
7298 		if (to_move == length) {
7299 			/* All of it fits in the MTU */
7300 			if (sp->some_taken) {
7301 				rcv_flags |= SCTP_DATA_LAST_FRAG;
7302 			} else {
7303 				rcv_flags |= SCTP_DATA_NOT_FRAG;
7304 			}
7305 			sp->put_last_out = 1;
7306 			if (sp->sinfo_flags & SCTP_SACK_IMMEDIATELY) {
7307 				rcv_flags |= SCTP_DATA_SACK_IMMEDIATELY;
7308 			}
7309 		} else {
7310 			/* Not all of it fits, we fragment */
7311 			if (sp->some_taken == 0) {
7312 				rcv_flags |= SCTP_DATA_FIRST_FRAG;
7313 			}
7314 			sp->some_taken = 1;
7315 		}
7316 	} else {
7317 		to_move = sctp_can_we_split_this(stcb, length, space_left, frag_point, eeor_mode);
7318 		if (to_move) {
7319 			/*-
7320 			 * We use a snapshot of length in case it
7321 			 * is expanding during the compare.
7322 			 */
7323 			uint32_t llen;
7324 
7325 			llen = length;
7326 			if (to_move >= llen) {
7327 				to_move = llen;
7328 				if (send_lock_up == 0) {
7329 					/*-
7330 					 * We are taking all of an incomplete msg
7331 					 * thus we need a send lock.
7332 					 */
7333 					SCTP_TCB_SEND_LOCK(stcb);
7334 					send_lock_up = 1;
7335 					if (sp->msg_is_complete) {
7336 						/*
7337 						 * the sender finished the
7338 						 * msg
7339 						 */
7340 						goto re_look;
7341 					}
7342 				}
7343 			}
7344 			if (sp->some_taken == 0) {
7345 				rcv_flags |= SCTP_DATA_FIRST_FRAG;
7346 				sp->some_taken = 1;
7347 			}
7348 		} else {
7349 			/* Nothing to take. */
7350 			*giveup = 1;
7351 			to_move = 0;
7352 			goto out_of;
7353 		}
7354 	}
7355 
7356 	/* If we reach here, we can copy out a chunk */
7357 	sctp_alloc_a_chunk(stcb, chk);
7358 	if (chk == NULL) {
7359 		/* No chunk memory */
7360 		*giveup = 1;
7361 		to_move = 0;
7362 		goto out_of;
7363 	}
7364 	/*
7365 	 * Setup for unordered if needed by looking at the user sent info
7366 	 * flags.
7367 	 */
7368 	if (sp->sinfo_flags & SCTP_UNORDERED) {
7369 		rcv_flags |= SCTP_DATA_UNORDERED;
7370 	}
7371 	if (SCTP_BASE_SYSCTL(sctp_enable_sack_immediately) &&
7372 	    (sp->sinfo_flags & SCTP_EOF) == SCTP_EOF) {
7373 		rcv_flags |= SCTP_DATA_SACK_IMMEDIATELY;
7374 	}
7375 	/* clear out the chunk before setting up */
7376 	memset(chk, 0, sizeof(*chk));
7377 	chk->rec.data.rcv_flags = rcv_flags;
7378 
7379 	if (to_move >= length) {
7380 		/* we think we can steal the whole thing */
7381 		if ((sp->sender_all_done == 0) && (send_lock_up == 0)) {
7382 			SCTP_TCB_SEND_LOCK(stcb);
7383 			send_lock_up = 1;
7384 		}
7385 		if (to_move < sp->length) {
7386 			/* bail, it changed */
7387 			goto dont_do_it;
7388 		}
7389 		chk->data = sp->data;
7390 		chk->last_mbuf = sp->tail_mbuf;
7391 		/* register the stealing */
7392 		sp->data = sp->tail_mbuf = NULL;
7393 	} else {
7394 		struct mbuf *m;
7395 
7396 dont_do_it:
7397 		chk->data = SCTP_M_COPYM(sp->data, 0, to_move, M_NOWAIT);
7398 		chk->last_mbuf = NULL;
7399 		if (chk->data == NULL) {
7400 			sp->some_taken = some_taken;
7401 			sctp_free_a_chunk(stcb, chk, so_locked);
7402 			*bail = 1;
7403 			to_move = 0;
7404 			goto out_of;
7405 		}
7406 #ifdef SCTP_MBUF_LOGGING
7407 		if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_MBUF_LOGGING_ENABLE) {
7408 			sctp_log_mbc(chk->data, SCTP_MBUF_ICOPY);
7409 		}
7410 #endif
7411 		/* Pull off the data */
7412 		m_adj(sp->data, to_move);
7413 		/* Now lets work our way down and compact it */
7414 		m = sp->data;
7415 		while (m && (SCTP_BUF_LEN(m) == 0)) {
7416 			sp->data = SCTP_BUF_NEXT(m);
7417 			SCTP_BUF_NEXT(m) = NULL;
7418 			if (sp->tail_mbuf == m) {
7419 				/*-
7420 				 * Freeing tail? TSNH since
7421 				 * we supposedly were taking less
7422 				 * than the sp->length.
7423 				 */
7424 #ifdef INVARIANTS
7425 				panic("Huh, freing tail? - TSNH");
7426 #else
7427 				SCTP_PRINTF("Huh, freeing tail? - TSNH\n");
7428 				sp->tail_mbuf = sp->data = NULL;
7429 				sp->length = 0;
7430 #endif
7431 
7432 			}
7433 			sctp_m_free(m);
7434 			m = sp->data;
7435 		}
7436 	}
7437 	if (SCTP_BUF_IS_EXTENDED(chk->data)) {
7438 		chk->copy_by_ref = 1;
7439 	} else {
7440 		chk->copy_by_ref = 0;
7441 	}
7442 	/*
7443 	 * get last_mbuf and counts of mb usage This is ugly but hopefully
7444 	 * its only one mbuf.
7445 	 */
7446 	if (chk->last_mbuf == NULL) {
7447 		chk->last_mbuf = chk->data;
7448 		while (SCTP_BUF_NEXT(chk->last_mbuf) != NULL) {
7449 			chk->last_mbuf = SCTP_BUF_NEXT(chk->last_mbuf);
7450 		}
7451 	}
7452 
7453 	if (to_move > length) {
7454 		/*- This should not happen either
7455 		 * since we always lower to_move to the size
7456 		 * of sp->length if its larger.
7457 		 */
7458 #ifdef INVARIANTS
7459 		panic("Huh, how can to_move be larger?");
7460 #else
7461 		SCTP_PRINTF("Huh, how can to_move be larger?\n");
7462 		sp->length = 0;
7463 #endif
7464 	} else {
7465 		atomic_subtract_int(&sp->length, to_move);
7466 	}
7467 	leading = SCTP_DATA_CHUNK_OVERHEAD(stcb);
7468 	if (M_LEADINGSPACE(chk->data) < leading) {
7469 		/* Not enough room for a chunk header, get some */
7470 		struct mbuf *m;
7471 
7472 		m = sctp_get_mbuf_for_msg(1, 0, M_NOWAIT, 1, MT_DATA);
7473 		if (m == NULL) {
7474 			/*
7475 			 * we're in trouble here. _PREPEND below will free
7476 			 * all the data if there is no leading space, so we
7477 			 * must put the data back and restore.
7478 			 */
7479 			if (send_lock_up == 0) {
7480 				SCTP_TCB_SEND_LOCK(stcb);
7481 				send_lock_up = 1;
7482 			}
7483 			if (sp->data == NULL) {
7484 				/* unsteal the data */
7485 				sp->data = chk->data;
7486 				sp->tail_mbuf = chk->last_mbuf;
7487 			} else {
7488 				struct mbuf *m_tmp;
7489 
7490 				/* reassemble the data */
7491 				m_tmp = sp->data;
7492 				sp->data = chk->data;
7493 				SCTP_BUF_NEXT(chk->last_mbuf) = m_tmp;
7494 			}
7495 			sp->some_taken = some_taken;
7496 			atomic_add_int(&sp->length, to_move);
7497 			chk->data = NULL;
7498 			*bail = 1;
7499 			sctp_free_a_chunk(stcb, chk, so_locked);
7500 			to_move = 0;
7501 			goto out_of;
7502 		} else {
7503 			SCTP_BUF_LEN(m) = 0;
7504 			SCTP_BUF_NEXT(m) = chk->data;
7505 			chk->data = m;
7506 			M_ALIGN(chk->data, 4);
7507 		}
7508 	}
7509 	SCTP_BUF_PREPEND(chk->data, SCTP_DATA_CHUNK_OVERHEAD(stcb), M_NOWAIT);
7510 	if (chk->data == NULL) {
7511 		/* HELP, TSNH since we assured it would not above? */
7512 #ifdef INVARIANTS
7513 		panic("prepend failes HELP?");
7514 #else
7515 		SCTP_PRINTF("prepend fails HELP?\n");
7516 		sctp_free_a_chunk(stcb, chk, so_locked);
7517 #endif
7518 		*bail = 1;
7519 		to_move = 0;
7520 		goto out_of;
7521 	}
7522 	sctp_snd_sb_alloc(stcb, SCTP_DATA_CHUNK_OVERHEAD(stcb));
7523 	chk->book_size = chk->send_size = (uint16_t)(to_move + SCTP_DATA_CHUNK_OVERHEAD(stcb));
7524 	chk->book_size_scale = 0;
7525 	chk->sent = SCTP_DATAGRAM_UNSENT;
7526 
7527 	chk->flags = 0;
7528 	chk->asoc = &stcb->asoc;
7529 	chk->pad_inplace = 0;
7530 	chk->no_fr_allowed = 0;
7531 	if (stcb->asoc.idata_supported == 0) {
7532 		if (rcv_flags & SCTP_DATA_UNORDERED) {
7533 			/* Just use 0. The receiver ignores the values. */
7534 			chk->rec.data.mid = 0;
7535 		} else {
7536 			chk->rec.data.mid = strq->next_mid_ordered;
7537 			if (rcv_flags & SCTP_DATA_LAST_FRAG) {
7538 				strq->next_mid_ordered++;
7539 			}
7540 		}
7541 	} else {
7542 		if (rcv_flags & SCTP_DATA_UNORDERED) {
7543 			chk->rec.data.mid = strq->next_mid_unordered;
7544 			if (rcv_flags & SCTP_DATA_LAST_FRAG) {
7545 				strq->next_mid_unordered++;
7546 			}
7547 		} else {
7548 			chk->rec.data.mid = strq->next_mid_ordered;
7549 			if (rcv_flags & SCTP_DATA_LAST_FRAG) {
7550 				strq->next_mid_ordered++;
7551 			}
7552 		}
7553 	}
7554 	chk->rec.data.sid = sp->sid;
7555 	chk->rec.data.ppid = sp->ppid;
7556 	chk->rec.data.context = sp->context;
7557 	chk->rec.data.doing_fast_retransmit = 0;
7558 
7559 	chk->rec.data.timetodrop = sp->ts;
7560 	chk->flags = sp->act_flags;
7561 
7562 	if (sp->net) {
7563 		chk->whoTo = sp->net;
7564 		atomic_add_int(&chk->whoTo->ref_count, 1);
7565 	} else
7566 		chk->whoTo = NULL;
7567 
7568 	if (sp->holds_key_ref) {
7569 		chk->auth_keyid = sp->auth_keyid;
7570 		sctp_auth_key_acquire(stcb, chk->auth_keyid);
7571 		chk->holds_key_ref = 1;
7572 	}
7573 	chk->rec.data.tsn = atomic_fetchadd_int(&asoc->sending_seq, 1);
7574 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOG_AT_SEND_2_OUTQ) {
7575 		sctp_misc_ints(SCTP_STRMOUT_LOG_SEND,
7576 		    (uint32_t)(uintptr_t)stcb, sp->length,
7577 		    (uint32_t)((chk->rec.data.sid << 16) | (0x0000ffff & chk->rec.data.mid)),
7578 		    chk->rec.data.tsn);
7579 	}
7580 	if (stcb->asoc.idata_supported == 0) {
7581 		dchkh = mtod(chk->data, struct sctp_data_chunk *);
7582 	} else {
7583 		ndchkh = mtod(chk->data, struct sctp_idata_chunk *);
7584 	}
7585 	/*
7586 	 * Put the rest of the things in place now. Size was done earlier in
7587 	 * previous loop prior to padding.
7588 	 */
7589 
7590 #ifdef SCTP_ASOCLOG_OF_TSNS
7591 	SCTP_TCB_LOCK_ASSERT(stcb);
7592 	if (asoc->tsn_out_at >= SCTP_TSN_LOG_SIZE) {
7593 		asoc->tsn_out_at = 0;
7594 		asoc->tsn_out_wrapped = 1;
7595 	}
7596 	asoc->out_tsnlog[asoc->tsn_out_at].tsn = chk->rec.data.tsn;
7597 	asoc->out_tsnlog[asoc->tsn_out_at].strm = chk->rec.data.sid;
7598 	asoc->out_tsnlog[asoc->tsn_out_at].seq = chk->rec.data.mid;
7599 	asoc->out_tsnlog[asoc->tsn_out_at].sz = chk->send_size;
7600 	asoc->out_tsnlog[asoc->tsn_out_at].flgs = chk->rec.data.rcv_flags;
7601 	asoc->out_tsnlog[asoc->tsn_out_at].stcb = (void *)stcb;
7602 	asoc->out_tsnlog[asoc->tsn_out_at].in_pos = asoc->tsn_out_at;
7603 	asoc->out_tsnlog[asoc->tsn_out_at].in_out = 2;
7604 	asoc->tsn_out_at++;
7605 #endif
7606 	if (stcb->asoc.idata_supported == 0) {
7607 		dchkh->ch.chunk_type = SCTP_DATA;
7608 		dchkh->ch.chunk_flags = chk->rec.data.rcv_flags;
7609 		dchkh->dp.tsn = htonl(chk->rec.data.tsn);
7610 		dchkh->dp.sid = htons(strq->sid);
7611 		dchkh->dp.ssn = htons((uint16_t)chk->rec.data.mid);
7612 		dchkh->dp.ppid = chk->rec.data.ppid;
7613 		dchkh->ch.chunk_length = htons(chk->send_size);
7614 	} else {
7615 		ndchkh->ch.chunk_type = SCTP_IDATA;
7616 		ndchkh->ch.chunk_flags = chk->rec.data.rcv_flags;
7617 		ndchkh->dp.tsn = htonl(chk->rec.data.tsn);
7618 		ndchkh->dp.sid = htons(strq->sid);
7619 		ndchkh->dp.reserved = htons(0);
7620 		ndchkh->dp.mid = htonl(chk->rec.data.mid);
7621 		if (sp->fsn == 0)
7622 			ndchkh->dp.ppid_fsn.ppid = chk->rec.data.ppid;
7623 		else
7624 			ndchkh->dp.ppid_fsn.fsn = htonl(sp->fsn);
7625 		sp->fsn++;
7626 		ndchkh->ch.chunk_length = htons(chk->send_size);
7627 	}
7628 	/* Now advance the chk->send_size by the actual pad needed. */
7629 	if (chk->send_size < SCTP_SIZE32(chk->book_size)) {
7630 		/* need a pad */
7631 		struct mbuf *lm;
7632 		int pads;
7633 
7634 		pads = SCTP_SIZE32(chk->book_size) - chk->send_size;
7635 		lm = sctp_pad_lastmbuf(chk->data, pads, chk->last_mbuf);
7636 		if (lm != NULL) {
7637 			chk->last_mbuf = lm;
7638 			chk->pad_inplace = 1;
7639 		}
7640 		chk->send_size += pads;
7641 	}
7642 	if (PR_SCTP_ENABLED(chk->flags)) {
7643 		asoc->pr_sctp_cnt++;
7644 	}
7645 	if (sp->msg_is_complete && (sp->length == 0) && (sp->sender_all_done)) {
7646 		/* All done pull and kill the message */
7647 		if (sp->put_last_out == 0) {
7648 			SCTP_PRINTF("Gak, put out entire msg with NO end!-2\n");
7649 			SCTP_PRINTF("sender_done:%d len:%d msg_comp:%d put_last_out:%d send_lock:%d\n",
7650 			    sp->sender_all_done,
7651 			    sp->length,
7652 			    sp->msg_is_complete,
7653 			    sp->put_last_out,
7654 			    send_lock_up);
7655 		}
7656 		if ((send_lock_up == 0) && (TAILQ_NEXT(sp, next) == NULL)) {
7657 			SCTP_TCB_SEND_LOCK(stcb);
7658 			send_lock_up = 1;
7659 		}
7660 		atomic_subtract_int(&asoc->stream_queue_cnt, 1);
7661 		TAILQ_REMOVE(&strq->outqueue, sp, next);
7662 		stcb->asoc.ss_functions.sctp_ss_remove_from_stream(stcb, asoc, strq, sp, send_lock_up);
7663 		if ((strq->state == SCTP_STREAM_RESET_PENDING) &&
7664 		    (strq->chunks_on_queues == 0) &&
7665 		    TAILQ_EMPTY(&strq->outqueue)) {
7666 			stcb->asoc.trigger_reset = 1;
7667 		}
7668 		if (sp->net) {
7669 			sctp_free_remote_addr(sp->net);
7670 			sp->net = NULL;
7671 		}
7672 		if (sp->data) {
7673 			sctp_m_freem(sp->data);
7674 			sp->data = NULL;
7675 		}
7676 		sctp_free_a_strmoq(stcb, sp, so_locked);
7677 	}
7678 	asoc->chunks_on_out_queue++;
7679 	strq->chunks_on_queues++;
7680 	TAILQ_INSERT_TAIL(&asoc->send_queue, chk, sctp_next);
7681 	asoc->send_queue_cnt++;
7682 out_of:
7683 	if (send_lock_up) {
7684 		SCTP_TCB_SEND_UNLOCK(stcb);
7685 	}
7686 	return (to_move);
7687 }
7688 
7689 
7690 static void
7691 sctp_fill_outqueue(struct sctp_tcb *stcb,
7692     struct sctp_nets *net, int frag_point, int eeor_mode, int *quit_now, int so_locked
7693 #if !defined(__APPLE__) && !defined(SCTP_SO_LOCK_TESTING)
7694     SCTP_UNUSED
7695 #endif
7696 )
7697 {
7698 	struct sctp_association *asoc;
7699 	struct sctp_stream_out *strq;
7700 	uint32_t space_left, moved, total_moved;
7701 	int bail, giveup;
7702 
7703 	SCTP_TCB_LOCK_ASSERT(stcb);
7704 	asoc = &stcb->asoc;
7705 	total_moved = 0;
7706 	switch (net->ro._l_addr.sa.sa_family) {
7707 #ifdef INET
7708 	case AF_INET:
7709 		space_left = net->mtu - SCTP_MIN_V4_OVERHEAD;
7710 		break;
7711 #endif
7712 #ifdef INET6
7713 	case AF_INET6:
7714 		space_left = net->mtu - SCTP_MIN_OVERHEAD;
7715 		break;
7716 #endif
7717 	default:
7718 		/* TSNH */
7719 		space_left = net->mtu;
7720 		break;
7721 	}
7722 	/* Need an allowance for the data chunk header too */
7723 	space_left -= SCTP_DATA_CHUNK_OVERHEAD(stcb);
7724 
7725 	/* must make even word boundary */
7726 	space_left &= 0xfffffffc;
7727 	strq = stcb->asoc.ss_functions.sctp_ss_select_stream(stcb, net, asoc);
7728 	giveup = 0;
7729 	bail = 0;
7730 	while ((space_left > 0) && (strq != NULL)) {
7731 		moved = sctp_move_to_outqueue(stcb, strq, space_left, frag_point,
7732 		    &giveup, eeor_mode, &bail, so_locked);
7733 		stcb->asoc.ss_functions.sctp_ss_scheduled(stcb, net, asoc, strq, moved);
7734 		if ((giveup != 0) || (bail != 0)) {
7735 			break;
7736 		}
7737 		strq = stcb->asoc.ss_functions.sctp_ss_select_stream(stcb, net, asoc);
7738 		total_moved += moved;
7739 		space_left -= moved;
7740 		if (space_left >= SCTP_DATA_CHUNK_OVERHEAD(stcb)) {
7741 			space_left -= SCTP_DATA_CHUNK_OVERHEAD(stcb);
7742 		} else {
7743 			space_left = 0;
7744 		}
7745 		space_left &= 0xfffffffc;
7746 	}
7747 	if (bail != 0)
7748 		*quit_now = 1;
7749 
7750 	stcb->asoc.ss_functions.sctp_ss_packet_done(stcb, net, asoc);
7751 
7752 	if (total_moved == 0) {
7753 		if ((stcb->asoc.sctp_cmt_on_off == 0) &&
7754 		    (net == stcb->asoc.primary_destination)) {
7755 			/* ran dry for primary network net */
7756 			SCTP_STAT_INCR(sctps_primary_randry);
7757 		} else if (stcb->asoc.sctp_cmt_on_off > 0) {
7758 			/* ran dry with CMT on */
7759 			SCTP_STAT_INCR(sctps_cmt_randry);
7760 		}
7761 	}
7762 }
7763 
7764 void
7765 sctp_fix_ecn_echo(struct sctp_association *asoc)
7766 {
7767 	struct sctp_tmit_chunk *chk;
7768 
7769 	TAILQ_FOREACH(chk, &asoc->control_send_queue, sctp_next) {
7770 		if (chk->rec.chunk_id.id == SCTP_ECN_ECHO) {
7771 			chk->sent = SCTP_DATAGRAM_UNSENT;
7772 		}
7773 	}
7774 }
7775 
7776 void
7777 sctp_move_chunks_from_net(struct sctp_tcb *stcb, struct sctp_nets *net)
7778 {
7779 	struct sctp_association *asoc;
7780 	struct sctp_tmit_chunk *chk;
7781 	struct sctp_stream_queue_pending *sp;
7782 	unsigned int i;
7783 
7784 	if (net == NULL) {
7785 		return;
7786 	}
7787 	asoc = &stcb->asoc;
7788 	for (i = 0; i < stcb->asoc.streamoutcnt; i++) {
7789 		TAILQ_FOREACH(sp, &stcb->asoc.strmout[i].outqueue, next) {
7790 			if (sp->net == net) {
7791 				sctp_free_remote_addr(sp->net);
7792 				sp->net = NULL;
7793 			}
7794 		}
7795 	}
7796 	TAILQ_FOREACH(chk, &asoc->send_queue, sctp_next) {
7797 		if (chk->whoTo == net) {
7798 			sctp_free_remote_addr(chk->whoTo);
7799 			chk->whoTo = NULL;
7800 		}
7801 	}
7802 }
7803 
7804 int
7805 sctp_med_chunk_output(struct sctp_inpcb *inp,
7806     struct sctp_tcb *stcb,
7807     struct sctp_association *asoc,
7808     int *num_out,
7809     int *reason_code,
7810     int control_only, int from_where,
7811     struct timeval *now, int *now_filled, int frag_point, int so_locked
7812 #if !defined(__APPLE__) && !defined(SCTP_SO_LOCK_TESTING)
7813     SCTP_UNUSED
7814 #endif
7815 )
7816 {
7817 	/**
7818 	 * Ok this is the generic chunk service queue. we must do the
7819 	 * following:
7820 	 * - Service the stream queue that is next, moving any
7821 	 *   message (note I must get a complete message i.e. FIRST/MIDDLE and
7822 	 *   LAST to the out queue in one pass) and assigning TSN's. This
7823 	 *   only applys though if the peer does not support NDATA. For NDATA
7824 	 *   chunks its ok to not send the entire message ;-)
7825 	 * - Check to see if the cwnd/rwnd allows any output, if so we go ahead and
7826 	 *   fomulate and send the low level chunks. Making sure to combine
7827 	 *   any control in the control chunk queue also.
7828 	 */
7829 	struct sctp_nets *net, *start_at, *sack_goes_to = NULL, *old_start_at = NULL;
7830 	struct mbuf *outchain, *endoutchain;
7831 	struct sctp_tmit_chunk *chk, *nchk;
7832 
7833 	/* temp arrays for unlinking */
7834 	struct sctp_tmit_chunk *data_list[SCTP_MAX_DATA_BUNDLING];
7835 	int no_fragmentflg, error;
7836 	unsigned int max_rwnd_per_dest, max_send_per_dest;
7837 	int one_chunk, hbflag, skip_data_for_this_net;
7838 	int asconf, cookie, no_out_cnt;
7839 	int bundle_at, ctl_cnt, no_data_chunks, eeor_mode;
7840 	unsigned int mtu, r_mtu, omtu, mx_mtu, to_out;
7841 	int tsns_sent = 0;
7842 	uint32_t auth_offset = 0;
7843 	struct sctp_auth_chunk *auth = NULL;
7844 	uint16_t auth_keyid;
7845 	int override_ok = 1;
7846 	int skip_fill_up = 0;
7847 	int data_auth_reqd = 0;
7848 
7849 	/*
7850 	 * JRS 5/14/07 - Add flag for whether a heartbeat is sent to the
7851 	 * destination.
7852 	 */
7853 	int quit_now = 0;
7854 
7855 	*num_out = 0;
7856 	*reason_code = 0;
7857 	auth_keyid = stcb->asoc.authinfo.active_keyid;
7858 	if ((asoc->state & SCTP_STATE_SHUTDOWN_PENDING) ||
7859 	    (SCTP_GET_STATE(stcb) == SCTP_STATE_SHUTDOWN_RECEIVED) ||
7860 	    (sctp_is_feature_on(inp, SCTP_PCB_FLAGS_EXPLICIT_EOR))) {
7861 		eeor_mode = 1;
7862 	} else {
7863 		eeor_mode = 0;
7864 	}
7865 	ctl_cnt = no_out_cnt = asconf = cookie = 0;
7866 	/*
7867 	 * First lets prime the pump. For each destination, if there is room
7868 	 * in the flight size, attempt to pull an MTU's worth out of the
7869 	 * stream queues into the general send_queue
7870 	 */
7871 #ifdef SCTP_AUDITING_ENABLED
7872 	sctp_audit_log(0xC2, 2);
7873 #endif
7874 	SCTP_TCB_LOCK_ASSERT(stcb);
7875 	hbflag = 0;
7876 	if (control_only)
7877 		no_data_chunks = 1;
7878 	else
7879 		no_data_chunks = 0;
7880 
7881 	/* Nothing to possible to send? */
7882 	if ((TAILQ_EMPTY(&asoc->control_send_queue) ||
7883 	    (asoc->ctrl_queue_cnt == stcb->asoc.ecn_echo_cnt_onq)) &&
7884 	    TAILQ_EMPTY(&asoc->asconf_send_queue) &&
7885 	    TAILQ_EMPTY(&asoc->send_queue) &&
7886 	    sctp_is_there_unsent_data(stcb, so_locked) == 0) {
7887 nothing_to_send:
7888 		*reason_code = 9;
7889 		return (0);
7890 	}
7891 	if (asoc->peers_rwnd == 0) {
7892 		/* No room in peers rwnd */
7893 		*reason_code = 1;
7894 		if (asoc->total_flight > 0) {
7895 			/* we are allowed one chunk in flight */
7896 			no_data_chunks = 1;
7897 		}
7898 	}
7899 	if (stcb->asoc.ecn_echo_cnt_onq) {
7900 		/* Record where a sack goes, if any */
7901 		if (no_data_chunks &&
7902 		    (asoc->ctrl_queue_cnt == stcb->asoc.ecn_echo_cnt_onq)) {
7903 			/* Nothing but ECNe to send - we don't do that */
7904 			goto nothing_to_send;
7905 		}
7906 		TAILQ_FOREACH(chk, &asoc->control_send_queue, sctp_next) {
7907 			if ((chk->rec.chunk_id.id == SCTP_SELECTIVE_ACK) ||
7908 			    (chk->rec.chunk_id.id == SCTP_NR_SELECTIVE_ACK)) {
7909 				sack_goes_to = chk->whoTo;
7910 				break;
7911 			}
7912 		}
7913 	}
7914 	max_rwnd_per_dest = ((asoc->peers_rwnd + asoc->total_flight) / asoc->numnets);
7915 	if (stcb->sctp_socket)
7916 		max_send_per_dest = SCTP_SB_LIMIT_SND(stcb->sctp_socket) / asoc->numnets;
7917 	else
7918 		max_send_per_dest = 0;
7919 	if (no_data_chunks == 0) {
7920 		/* How many non-directed chunks are there? */
7921 		TAILQ_FOREACH(chk, &asoc->send_queue, sctp_next) {
7922 			if (chk->whoTo == NULL) {
7923 				/*
7924 				 * We already have non-directed chunks on
7925 				 * the queue, no need to do a fill-up.
7926 				 */
7927 				skip_fill_up = 1;
7928 				break;
7929 			}
7930 		}
7931 
7932 	}
7933 	if ((no_data_chunks == 0) &&
7934 	    (skip_fill_up == 0) &&
7935 	    (!stcb->asoc.ss_functions.sctp_ss_is_empty(stcb, asoc))) {
7936 		TAILQ_FOREACH(net, &asoc->nets, sctp_next) {
7937 			/*
7938 			 * This for loop we are in takes in each net, if
7939 			 * its's got space in cwnd and has data sent to it
7940 			 * (when CMT is off) then it calls
7941 			 * sctp_fill_outqueue for the net. This gets data on
7942 			 * the send queue for that network.
7943 			 *
7944 			 * In sctp_fill_outqueue TSN's are assigned and data
7945 			 * is copied out of the stream buffers. Note mostly
7946 			 * copy by reference (we hope).
7947 			 */
7948 			net->window_probe = 0;
7949 			if ((net != stcb->asoc.alternate) &&
7950 			    ((net->dest_state & SCTP_ADDR_PF) ||
7951 			    (!(net->dest_state & SCTP_ADDR_REACHABLE)) ||
7952 			    (net->dest_state & SCTP_ADDR_UNCONFIRMED))) {
7953 				if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_CWND_LOGGING_ENABLE) {
7954 					sctp_log_cwnd(stcb, net, 1,
7955 					    SCTP_CWND_LOG_FILL_OUTQ_CALLED);
7956 				}
7957 				continue;
7958 			}
7959 			if ((stcb->asoc.cc_functions.sctp_cwnd_new_transmission_begins) &&
7960 			    (net->flight_size == 0)) {
7961 				(*stcb->asoc.cc_functions.sctp_cwnd_new_transmission_begins) (stcb, net);
7962 			}
7963 			if (net->flight_size >= net->cwnd) {
7964 				/* skip this network, no room - can't fill */
7965 				if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_CWND_LOGGING_ENABLE) {
7966 					sctp_log_cwnd(stcb, net, 3,
7967 					    SCTP_CWND_LOG_FILL_OUTQ_CALLED);
7968 				}
7969 				continue;
7970 			}
7971 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_CWND_LOGGING_ENABLE) {
7972 				sctp_log_cwnd(stcb, net, 4, SCTP_CWND_LOG_FILL_OUTQ_CALLED);
7973 			}
7974 			sctp_fill_outqueue(stcb, net, frag_point, eeor_mode, &quit_now, so_locked);
7975 			if (quit_now) {
7976 				/* memory alloc failure */
7977 				no_data_chunks = 1;
7978 				break;
7979 			}
7980 		}
7981 	}
7982 	/* now service each destination and send out what we can for it */
7983 	/* Nothing to send? */
7984 	if (TAILQ_EMPTY(&asoc->control_send_queue) &&
7985 	    TAILQ_EMPTY(&asoc->asconf_send_queue) &&
7986 	    TAILQ_EMPTY(&asoc->send_queue)) {
7987 		*reason_code = 8;
7988 		return (0);
7989 	}
7990 
7991 	if (asoc->sctp_cmt_on_off > 0) {
7992 		/* get the last start point */
7993 		start_at = asoc->last_net_cmt_send_started;
7994 		if (start_at == NULL) {
7995 			/* null so to beginning */
7996 			start_at = TAILQ_FIRST(&asoc->nets);
7997 		} else {
7998 			start_at = TAILQ_NEXT(asoc->last_net_cmt_send_started, sctp_next);
7999 			if (start_at == NULL) {
8000 				start_at = TAILQ_FIRST(&asoc->nets);
8001 			}
8002 		}
8003 		asoc->last_net_cmt_send_started = start_at;
8004 	} else {
8005 		start_at = TAILQ_FIRST(&asoc->nets);
8006 	}
8007 	TAILQ_FOREACH(chk, &asoc->control_send_queue, sctp_next) {
8008 		if (chk->whoTo == NULL) {
8009 			if (asoc->alternate) {
8010 				chk->whoTo = asoc->alternate;
8011 			} else {
8012 				chk->whoTo = asoc->primary_destination;
8013 			}
8014 			atomic_add_int(&chk->whoTo->ref_count, 1);
8015 		}
8016 	}
8017 	old_start_at = NULL;
8018 again_one_more_time:
8019 	for (net = start_at; net != NULL; net = TAILQ_NEXT(net, sctp_next)) {
8020 		/* how much can we send? */
8021 		/* SCTPDBG("Examine for sending net:%x\n", (uint32_t)net); */
8022 		if (old_start_at && (old_start_at == net)) {
8023 			/* through list ocmpletely. */
8024 			break;
8025 		}
8026 		tsns_sent = 0xa;
8027 		if (TAILQ_EMPTY(&asoc->control_send_queue) &&
8028 		    TAILQ_EMPTY(&asoc->asconf_send_queue) &&
8029 		    (net->flight_size >= net->cwnd)) {
8030 			/*
8031 			 * Nothing on control or asconf and flight is full,
8032 			 * we can skip even in the CMT case.
8033 			 */
8034 			continue;
8035 		}
8036 		bundle_at = 0;
8037 		endoutchain = outchain = NULL;
8038 		no_fragmentflg = 1;
8039 		one_chunk = 0;
8040 		if (net->dest_state & SCTP_ADDR_UNCONFIRMED) {
8041 			skip_data_for_this_net = 1;
8042 		} else {
8043 			skip_data_for_this_net = 0;
8044 		}
8045 		switch (((struct sockaddr *)&net->ro._l_addr)->sa_family) {
8046 #ifdef INET
8047 		case AF_INET:
8048 			mtu = net->mtu - SCTP_MIN_V4_OVERHEAD;
8049 			break;
8050 #endif
8051 #ifdef INET6
8052 		case AF_INET6:
8053 			mtu = net->mtu - SCTP_MIN_OVERHEAD;
8054 			break;
8055 #endif
8056 		default:
8057 			/* TSNH */
8058 			mtu = net->mtu;
8059 			break;
8060 		}
8061 		mx_mtu = mtu;
8062 		to_out = 0;
8063 		if (mtu > asoc->peers_rwnd) {
8064 			if (asoc->total_flight > 0) {
8065 				/* We have a packet in flight somewhere */
8066 				r_mtu = asoc->peers_rwnd;
8067 			} else {
8068 				/* We are always allowed to send one MTU out */
8069 				one_chunk = 1;
8070 				r_mtu = mtu;
8071 			}
8072 		} else {
8073 			r_mtu = mtu;
8074 		}
8075 		error = 0;
8076 		/************************/
8077 		/* ASCONF transmission */
8078 		/************************/
8079 		/* Now first lets go through the asconf queue */
8080 		TAILQ_FOREACH_SAFE(chk, &asoc->asconf_send_queue, sctp_next, nchk) {
8081 			if (chk->rec.chunk_id.id != SCTP_ASCONF) {
8082 				continue;
8083 			}
8084 			if (chk->whoTo == NULL) {
8085 				if (asoc->alternate == NULL) {
8086 					if (asoc->primary_destination != net) {
8087 						break;
8088 					}
8089 				} else {
8090 					if (asoc->alternate != net) {
8091 						break;
8092 					}
8093 				}
8094 			} else {
8095 				if (chk->whoTo != net) {
8096 					break;
8097 				}
8098 			}
8099 			if (chk->data == NULL) {
8100 				break;
8101 			}
8102 			if (chk->sent != SCTP_DATAGRAM_UNSENT &&
8103 			    chk->sent != SCTP_DATAGRAM_RESEND) {
8104 				break;
8105 			}
8106 			/*
8107 			 * if no AUTH is yet included and this chunk
8108 			 * requires it, make sure to account for it.  We
8109 			 * don't apply the size until the AUTH chunk is
8110 			 * actually added below in case there is no room for
8111 			 * this chunk. NOTE: we overload the use of "omtu"
8112 			 * here
8113 			 */
8114 			if ((auth == NULL) &&
8115 			    sctp_auth_is_required_chunk(chk->rec.chunk_id.id,
8116 			    stcb->asoc.peer_auth_chunks)) {
8117 				omtu = sctp_get_auth_chunk_len(stcb->asoc.peer_hmac_id);
8118 			} else
8119 				omtu = 0;
8120 			/* Here we do NOT factor the r_mtu */
8121 			if ((chk->send_size < (int)(mtu - omtu)) ||
8122 			    (chk->flags & CHUNK_FLAGS_FRAGMENT_OK)) {
8123 				/*
8124 				 * We probably should glom the mbuf chain
8125 				 * from the chk->data for control but the
8126 				 * problem is it becomes yet one more level
8127 				 * of tracking to do if for some reason
8128 				 * output fails. Then I have got to
8129 				 * reconstruct the merged control chain.. el
8130 				 * yucko.. for now we take the easy way and
8131 				 * do the copy
8132 				 */
8133 				/*
8134 				 * Add an AUTH chunk, if chunk requires it
8135 				 * save the offset into the chain for AUTH
8136 				 */
8137 				if ((auth == NULL) &&
8138 				    (sctp_auth_is_required_chunk(chk->rec.chunk_id.id,
8139 				    stcb->asoc.peer_auth_chunks))) {
8140 					outchain = sctp_add_auth_chunk(outchain,
8141 					    &endoutchain,
8142 					    &auth,
8143 					    &auth_offset,
8144 					    stcb,
8145 					    chk->rec.chunk_id.id);
8146 					SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
8147 				}
8148 				outchain = sctp_copy_mbufchain(chk->data, outchain, &endoutchain,
8149 				    (int)chk->rec.chunk_id.can_take_data,
8150 				    chk->send_size, chk->copy_by_ref);
8151 				if (outchain == NULL) {
8152 					*reason_code = 8;
8153 					SCTP_LTRACE_ERR_RET(inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
8154 					return (ENOMEM);
8155 				}
8156 				SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
8157 				/* update our MTU size */
8158 				if (mtu > (chk->send_size + omtu))
8159 					mtu -= (chk->send_size + omtu);
8160 				else
8161 					mtu = 0;
8162 				to_out += (chk->send_size + omtu);
8163 				/* Do clear IP_DF ? */
8164 				if (chk->flags & CHUNK_FLAGS_FRAGMENT_OK) {
8165 					no_fragmentflg = 0;
8166 				}
8167 				if (chk->rec.chunk_id.can_take_data)
8168 					chk->data = NULL;
8169 				/*
8170 				 * set hb flag since we can use these for
8171 				 * RTO
8172 				 */
8173 				hbflag = 1;
8174 				asconf = 1;
8175 				/*
8176 				 * should sysctl this: don't bundle data
8177 				 * with ASCONF since it requires AUTH
8178 				 */
8179 				no_data_chunks = 1;
8180 				chk->sent = SCTP_DATAGRAM_SENT;
8181 				if (chk->whoTo == NULL) {
8182 					chk->whoTo = net;
8183 					atomic_add_int(&net->ref_count, 1);
8184 				}
8185 				chk->snd_count++;
8186 				if (mtu == 0) {
8187 					/*
8188 					 * Ok we are out of room but we can
8189 					 * output without effecting the
8190 					 * flight size since this little guy
8191 					 * is a control only packet.
8192 					 */
8193 					sctp_timer_start(SCTP_TIMER_TYPE_ASCONF, inp, stcb, net);
8194 					/*
8195 					 * do NOT clear the asconf flag as
8196 					 * it is used to do appropriate
8197 					 * source address selection.
8198 					 */
8199 					if (*now_filled == 0) {
8200 						(void)SCTP_GETTIME_TIMEVAL(now);
8201 						*now_filled = 1;
8202 					}
8203 					net->last_sent_time = *now;
8204 					hbflag = 0;
8205 					if ((error = sctp_lowlevel_chunk_output(inp, stcb, net,
8206 					    (struct sockaddr *)&net->ro._l_addr,
8207 					    outchain, auth_offset, auth,
8208 					    stcb->asoc.authinfo.active_keyid,
8209 					    no_fragmentflg, 0, asconf,
8210 					    inp->sctp_lport, stcb->rport,
8211 					    htonl(stcb->asoc.peer_vtag),
8212 					    net->port, NULL,
8213 					    0, 0,
8214 					    so_locked))) {
8215 						/*
8216 						 * error, we could not
8217 						 * output
8218 						 */
8219 						SCTPDBG(SCTP_DEBUG_OUTPUT3, "Gak send error %d\n", error);
8220 						if (from_where == 0) {
8221 							SCTP_STAT_INCR(sctps_lowlevelerrusr);
8222 						}
8223 						if (error == ENOBUFS) {
8224 							asoc->ifp_had_enobuf = 1;
8225 							SCTP_STAT_INCR(sctps_lowlevelerr);
8226 						}
8227 						/* error, could not output */
8228 						if (error == EHOSTUNREACH) {
8229 							/*
8230 							 * Destination went
8231 							 * unreachable
8232 							 * during this send
8233 							 */
8234 							sctp_move_chunks_from_net(stcb, net);
8235 						}
8236 						*reason_code = 7;
8237 						break;
8238 					} else {
8239 						asoc->ifp_had_enobuf = 0;
8240 					}
8241 					/*
8242 					 * increase the number we sent, if a
8243 					 * cookie is sent we don't tell them
8244 					 * any was sent out.
8245 					 */
8246 					outchain = endoutchain = NULL;
8247 					auth = NULL;
8248 					auth_offset = 0;
8249 					if (!no_out_cnt)
8250 						*num_out += ctl_cnt;
8251 					/* recalc a clean slate and setup */
8252 					switch (net->ro._l_addr.sa.sa_family) {
8253 #ifdef INET
8254 					case AF_INET:
8255 						mtu = net->mtu - SCTP_MIN_V4_OVERHEAD;
8256 						break;
8257 #endif
8258 #ifdef INET6
8259 					case AF_INET6:
8260 						mtu = net->mtu - SCTP_MIN_OVERHEAD;
8261 						break;
8262 #endif
8263 					default:
8264 						/* TSNH */
8265 						mtu = net->mtu;
8266 						break;
8267 					}
8268 					to_out = 0;
8269 					no_fragmentflg = 1;
8270 				}
8271 			}
8272 		}
8273 		if (error != 0) {
8274 			/* try next net */
8275 			continue;
8276 		}
8277 		/************************/
8278 		/* Control transmission */
8279 		/************************/
8280 		/* Now first lets go through the control queue */
8281 		TAILQ_FOREACH_SAFE(chk, &asoc->control_send_queue, sctp_next, nchk) {
8282 			if ((sack_goes_to) &&
8283 			    (chk->rec.chunk_id.id == SCTP_ECN_ECHO) &&
8284 			    (chk->whoTo != sack_goes_to)) {
8285 				/*
8286 				 * if we have a sack in queue, and we are
8287 				 * looking at an ecn echo that is NOT queued
8288 				 * to where the sack is going..
8289 				 */
8290 				if (chk->whoTo == net) {
8291 					/*
8292 					 * Don't transmit it to where its
8293 					 * going (current net)
8294 					 */
8295 					continue;
8296 				} else if (sack_goes_to == net) {
8297 					/*
8298 					 * But do transmit it to this
8299 					 * address
8300 					 */
8301 					goto skip_net_check;
8302 				}
8303 			}
8304 			if (chk->whoTo == NULL) {
8305 				if (asoc->alternate == NULL) {
8306 					if (asoc->primary_destination != net) {
8307 						continue;
8308 					}
8309 				} else {
8310 					if (asoc->alternate != net) {
8311 						continue;
8312 					}
8313 				}
8314 			} else {
8315 				if (chk->whoTo != net) {
8316 					continue;
8317 				}
8318 			}
8319 	skip_net_check:
8320 			if (chk->data == NULL) {
8321 				continue;
8322 			}
8323 			if (chk->sent != SCTP_DATAGRAM_UNSENT) {
8324 				/*
8325 				 * It must be unsent. Cookies and ASCONF's
8326 				 * hang around but there timers will force
8327 				 * when marked for resend.
8328 				 */
8329 				continue;
8330 			}
8331 			/*
8332 			 * if no AUTH is yet included and this chunk
8333 			 * requires it, make sure to account for it.  We
8334 			 * don't apply the size until the AUTH chunk is
8335 			 * actually added below in case there is no room for
8336 			 * this chunk. NOTE: we overload the use of "omtu"
8337 			 * here
8338 			 */
8339 			if ((auth == NULL) &&
8340 			    sctp_auth_is_required_chunk(chk->rec.chunk_id.id,
8341 			    stcb->asoc.peer_auth_chunks)) {
8342 				omtu = sctp_get_auth_chunk_len(stcb->asoc.peer_hmac_id);
8343 			} else
8344 				omtu = 0;
8345 			/* Here we do NOT factor the r_mtu */
8346 			if ((chk->send_size <= (int)(mtu - omtu)) ||
8347 			    (chk->flags & CHUNK_FLAGS_FRAGMENT_OK)) {
8348 				/*
8349 				 * We probably should glom the mbuf chain
8350 				 * from the chk->data for control but the
8351 				 * problem is it becomes yet one more level
8352 				 * of tracking to do if for some reason
8353 				 * output fails. Then I have got to
8354 				 * reconstruct the merged control chain.. el
8355 				 * yucko.. for now we take the easy way and
8356 				 * do the copy
8357 				 */
8358 				/*
8359 				 * Add an AUTH chunk, if chunk requires it
8360 				 * save the offset into the chain for AUTH
8361 				 */
8362 				if ((auth == NULL) &&
8363 				    (sctp_auth_is_required_chunk(chk->rec.chunk_id.id,
8364 				    stcb->asoc.peer_auth_chunks))) {
8365 					outchain = sctp_add_auth_chunk(outchain,
8366 					    &endoutchain,
8367 					    &auth,
8368 					    &auth_offset,
8369 					    stcb,
8370 					    chk->rec.chunk_id.id);
8371 					SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
8372 				}
8373 				outchain = sctp_copy_mbufchain(chk->data, outchain, &endoutchain,
8374 				    (int)chk->rec.chunk_id.can_take_data,
8375 				    chk->send_size, chk->copy_by_ref);
8376 				if (outchain == NULL) {
8377 					*reason_code = 8;
8378 					SCTP_LTRACE_ERR_RET(inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
8379 					return (ENOMEM);
8380 				}
8381 				SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
8382 				/* update our MTU size */
8383 				if (mtu > (chk->send_size + omtu))
8384 					mtu -= (chk->send_size + omtu);
8385 				else
8386 					mtu = 0;
8387 				to_out += (chk->send_size + omtu);
8388 				/* Do clear IP_DF ? */
8389 				if (chk->flags & CHUNK_FLAGS_FRAGMENT_OK) {
8390 					no_fragmentflg = 0;
8391 				}
8392 				if (chk->rec.chunk_id.can_take_data)
8393 					chk->data = NULL;
8394 				/* Mark things to be removed, if needed */
8395 				if ((chk->rec.chunk_id.id == SCTP_SELECTIVE_ACK) ||
8396 				    (chk->rec.chunk_id.id == SCTP_NR_SELECTIVE_ACK) ||	/* EY */
8397 				    (chk->rec.chunk_id.id == SCTP_HEARTBEAT_REQUEST) ||
8398 				    (chk->rec.chunk_id.id == SCTP_HEARTBEAT_ACK) ||
8399 				    (chk->rec.chunk_id.id == SCTP_SHUTDOWN) ||
8400 				    (chk->rec.chunk_id.id == SCTP_SHUTDOWN_ACK) ||
8401 				    (chk->rec.chunk_id.id == SCTP_OPERATION_ERROR) ||
8402 				    (chk->rec.chunk_id.id == SCTP_COOKIE_ACK) ||
8403 				    (chk->rec.chunk_id.id == SCTP_ECN_CWR) ||
8404 				    (chk->rec.chunk_id.id == SCTP_PACKET_DROPPED) ||
8405 				    (chk->rec.chunk_id.id == SCTP_ASCONF_ACK)) {
8406 					if (chk->rec.chunk_id.id == SCTP_HEARTBEAT_REQUEST) {
8407 						hbflag = 1;
8408 					}
8409 					/* remove these chunks at the end */
8410 					if ((chk->rec.chunk_id.id == SCTP_SELECTIVE_ACK) ||
8411 					    (chk->rec.chunk_id.id == SCTP_NR_SELECTIVE_ACK)) {
8412 						/* turn off the timer */
8413 						if (SCTP_OS_TIMER_PENDING(&stcb->asoc.dack_timer.timer)) {
8414 							sctp_timer_stop(SCTP_TIMER_TYPE_RECV,
8415 							    inp, stcb, net,
8416 							    SCTP_FROM_SCTP_OUTPUT + SCTP_LOC_1);
8417 						}
8418 					}
8419 					ctl_cnt++;
8420 				} else {
8421 					/*
8422 					 * Other chunks, since they have
8423 					 * timers running (i.e. COOKIE) we
8424 					 * just "trust" that it gets sent or
8425 					 * retransmitted.
8426 					 */
8427 					ctl_cnt++;
8428 					if (chk->rec.chunk_id.id == SCTP_COOKIE_ECHO) {
8429 						cookie = 1;
8430 						no_out_cnt = 1;
8431 					} else if (chk->rec.chunk_id.id == SCTP_ECN_ECHO) {
8432 						/*
8433 						 * Increment ecne send count
8434 						 * here this means we may be
8435 						 * over-zealous in our
8436 						 * counting if the send
8437 						 * fails, but its the best
8438 						 * place to do it (we used
8439 						 * to do it in the queue of
8440 						 * the chunk, but that did
8441 						 * not tell how many times
8442 						 * it was sent.
8443 						 */
8444 						SCTP_STAT_INCR(sctps_sendecne);
8445 					}
8446 					chk->sent = SCTP_DATAGRAM_SENT;
8447 					if (chk->whoTo == NULL) {
8448 						chk->whoTo = net;
8449 						atomic_add_int(&net->ref_count, 1);
8450 					}
8451 					chk->snd_count++;
8452 				}
8453 				if (mtu == 0) {
8454 					/*
8455 					 * Ok we are out of room but we can
8456 					 * output without effecting the
8457 					 * flight size since this little guy
8458 					 * is a control only packet.
8459 					 */
8460 					if (asconf) {
8461 						sctp_timer_start(SCTP_TIMER_TYPE_ASCONF, inp, stcb, net);
8462 						/*
8463 						 * do NOT clear the asconf
8464 						 * flag as it is used to do
8465 						 * appropriate source
8466 						 * address selection.
8467 						 */
8468 					}
8469 					if (cookie) {
8470 						sctp_timer_start(SCTP_TIMER_TYPE_COOKIE, inp, stcb, net);
8471 						cookie = 0;
8472 					}
8473 					/* Only HB or ASCONF advances time */
8474 					if (hbflag) {
8475 						if (*now_filled == 0) {
8476 							(void)SCTP_GETTIME_TIMEVAL(now);
8477 							*now_filled = 1;
8478 						}
8479 						net->last_sent_time = *now;
8480 						hbflag = 0;
8481 					}
8482 					if ((error = sctp_lowlevel_chunk_output(inp, stcb, net,
8483 					    (struct sockaddr *)&net->ro._l_addr,
8484 					    outchain,
8485 					    auth_offset, auth,
8486 					    stcb->asoc.authinfo.active_keyid,
8487 					    no_fragmentflg, 0, asconf,
8488 					    inp->sctp_lport, stcb->rport,
8489 					    htonl(stcb->asoc.peer_vtag),
8490 					    net->port, NULL,
8491 					    0, 0,
8492 					    so_locked))) {
8493 						/*
8494 						 * error, we could not
8495 						 * output
8496 						 */
8497 						SCTPDBG(SCTP_DEBUG_OUTPUT3, "Gak send error %d\n", error);
8498 						if (from_where == 0) {
8499 							SCTP_STAT_INCR(sctps_lowlevelerrusr);
8500 						}
8501 						if (error == ENOBUFS) {
8502 							asoc->ifp_had_enobuf = 1;
8503 							SCTP_STAT_INCR(sctps_lowlevelerr);
8504 						}
8505 						if (error == EHOSTUNREACH) {
8506 							/*
8507 							 * Destination went
8508 							 * unreachable
8509 							 * during this send
8510 							 */
8511 							sctp_move_chunks_from_net(stcb, net);
8512 						}
8513 						*reason_code = 7;
8514 						break;
8515 					} else {
8516 						asoc->ifp_had_enobuf = 0;
8517 					}
8518 					/*
8519 					 * increase the number we sent, if a
8520 					 * cookie is sent we don't tell them
8521 					 * any was sent out.
8522 					 */
8523 					outchain = endoutchain = NULL;
8524 					auth = NULL;
8525 					auth_offset = 0;
8526 					if (!no_out_cnt)
8527 						*num_out += ctl_cnt;
8528 					/* recalc a clean slate and setup */
8529 					switch (net->ro._l_addr.sa.sa_family) {
8530 #ifdef INET
8531 					case AF_INET:
8532 						mtu = net->mtu - SCTP_MIN_V4_OVERHEAD;
8533 						break;
8534 #endif
8535 #ifdef INET6
8536 					case AF_INET6:
8537 						mtu = net->mtu - SCTP_MIN_OVERHEAD;
8538 						break;
8539 #endif
8540 					default:
8541 						/* TSNH */
8542 						mtu = net->mtu;
8543 						break;
8544 					}
8545 					to_out = 0;
8546 					no_fragmentflg = 1;
8547 				}
8548 			}
8549 		}
8550 		if (error != 0) {
8551 			/* try next net */
8552 			continue;
8553 		}
8554 		/* JRI: if dest is in PF state, do not send data to it */
8555 		if ((asoc->sctp_cmt_on_off > 0) &&
8556 		    (net != stcb->asoc.alternate) &&
8557 		    (net->dest_state & SCTP_ADDR_PF)) {
8558 			goto no_data_fill;
8559 		}
8560 		if (net->flight_size >= net->cwnd) {
8561 			goto no_data_fill;
8562 		}
8563 		if ((asoc->sctp_cmt_on_off > 0) &&
8564 		    (SCTP_BASE_SYSCTL(sctp_buffer_splitting) & SCTP_RECV_BUFFER_SPLITTING) &&
8565 		    (net->flight_size > max_rwnd_per_dest)) {
8566 			goto no_data_fill;
8567 		}
8568 		/*
8569 		 * We need a specific accounting for the usage of the send
8570 		 * buffer. We also need to check the number of messages per
8571 		 * net. For now, this is better than nothing and it disabled
8572 		 * by default...
8573 		 */
8574 		if ((asoc->sctp_cmt_on_off > 0) &&
8575 		    (SCTP_BASE_SYSCTL(sctp_buffer_splitting) & SCTP_SEND_BUFFER_SPLITTING) &&
8576 		    (max_send_per_dest > 0) &&
8577 		    (net->flight_size > max_send_per_dest)) {
8578 			goto no_data_fill;
8579 		}
8580 		/*********************/
8581 		/* Data transmission */
8582 		/*********************/
8583 		/*
8584 		 * if AUTH for DATA is required and no AUTH has been added
8585 		 * yet, account for this in the mtu now... if no data can be
8586 		 * bundled, this adjustment won't matter anyways since the
8587 		 * packet will be going out...
8588 		 */
8589 		data_auth_reqd = sctp_auth_is_required_chunk(SCTP_DATA,
8590 		    stcb->asoc.peer_auth_chunks);
8591 		if (data_auth_reqd && (auth == NULL)) {
8592 			mtu -= sctp_get_auth_chunk_len(stcb->asoc.peer_hmac_id);
8593 		}
8594 		/* now lets add any data within the MTU constraints */
8595 		switch (((struct sockaddr *)&net->ro._l_addr)->sa_family) {
8596 #ifdef INET
8597 		case AF_INET:
8598 			if (net->mtu > SCTP_MIN_V4_OVERHEAD)
8599 				omtu = net->mtu - SCTP_MIN_V4_OVERHEAD;
8600 			else
8601 				omtu = 0;
8602 			break;
8603 #endif
8604 #ifdef INET6
8605 		case AF_INET6:
8606 			if (net->mtu > SCTP_MIN_OVERHEAD)
8607 				omtu = net->mtu - SCTP_MIN_OVERHEAD;
8608 			else
8609 				omtu = 0;
8610 			break;
8611 #endif
8612 		default:
8613 			/* TSNH */
8614 			omtu = 0;
8615 			break;
8616 		}
8617 		if ((((SCTP_GET_STATE(stcb) == SCTP_STATE_OPEN) ||
8618 		    (SCTP_GET_STATE(stcb) == SCTP_STATE_SHUTDOWN_RECEIVED)) &&
8619 		    (skip_data_for_this_net == 0)) ||
8620 		    (cookie)) {
8621 			TAILQ_FOREACH_SAFE(chk, &asoc->send_queue, sctp_next, nchk) {
8622 				if (no_data_chunks) {
8623 					/* let only control go out */
8624 					*reason_code = 1;
8625 					break;
8626 				}
8627 				if (net->flight_size >= net->cwnd) {
8628 					/* skip this net, no room for data */
8629 					*reason_code = 2;
8630 					break;
8631 				}
8632 				if ((chk->whoTo != NULL) &&
8633 				    (chk->whoTo != net)) {
8634 					/* Don't send the chunk on this net */
8635 					continue;
8636 				}
8637 
8638 				if (asoc->sctp_cmt_on_off == 0) {
8639 					if ((asoc->alternate) &&
8640 					    (asoc->alternate != net) &&
8641 					    (chk->whoTo == NULL)) {
8642 						continue;
8643 					} else if ((net != asoc->primary_destination) &&
8644 						    (asoc->alternate == NULL) &&
8645 					    (chk->whoTo == NULL)) {
8646 						continue;
8647 					}
8648 				}
8649 				if ((chk->send_size > omtu) && ((chk->flags & CHUNK_FLAGS_FRAGMENT_OK) == 0)) {
8650 					/*-
8651 					 * strange, we have a chunk that is
8652 					 * to big for its destination and
8653 					 * yet no fragment ok flag.
8654 					 * Something went wrong when the
8655 					 * PMTU changed...we did not mark
8656 					 * this chunk for some reason?? I
8657 					 * will fix it here by letting IP
8658 					 * fragment it for now and printing
8659 					 * a warning. This really should not
8660 					 * happen ...
8661 					 */
8662 					SCTP_PRINTF("Warning chunk of %d bytes > mtu:%d and yet PMTU disc missed\n",
8663 					    chk->send_size, mtu);
8664 					chk->flags |= CHUNK_FLAGS_FRAGMENT_OK;
8665 				}
8666 				if (SCTP_BASE_SYSCTL(sctp_enable_sack_immediately) &&
8667 				    (asoc->state & SCTP_STATE_SHUTDOWN_PENDING)) {
8668 					struct sctp_data_chunk *dchkh;
8669 
8670 					dchkh = mtod(chk->data, struct sctp_data_chunk *);
8671 					dchkh->ch.chunk_flags |= SCTP_DATA_SACK_IMMEDIATELY;
8672 				}
8673 				if (((chk->send_size <= mtu) && (chk->send_size <= r_mtu)) ||
8674 				    ((chk->flags & CHUNK_FLAGS_FRAGMENT_OK) && (chk->send_size <= asoc->peers_rwnd))) {
8675 					/* ok we will add this one */
8676 
8677 					/*
8678 					 * Add an AUTH chunk, if chunk
8679 					 * requires it, save the offset into
8680 					 * the chain for AUTH
8681 					 */
8682 					if (data_auth_reqd) {
8683 						if (auth == NULL) {
8684 							outchain = sctp_add_auth_chunk(outchain,
8685 							    &endoutchain,
8686 							    &auth,
8687 							    &auth_offset,
8688 							    stcb,
8689 							    SCTP_DATA);
8690 							auth_keyid = chk->auth_keyid;
8691 							override_ok = 0;
8692 							SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
8693 						} else if (override_ok) {
8694 							/*
8695 							 * use this data's
8696 							 * keyid
8697 							 */
8698 							auth_keyid = chk->auth_keyid;
8699 							override_ok = 0;
8700 						} else if (auth_keyid != chk->auth_keyid) {
8701 							/*
8702 							 * different keyid,
8703 							 * so done bundling
8704 							 */
8705 							break;
8706 						}
8707 					}
8708 					outchain = sctp_copy_mbufchain(chk->data, outchain, &endoutchain, 0,
8709 					    chk->send_size, chk->copy_by_ref);
8710 					if (outchain == NULL) {
8711 						SCTPDBG(SCTP_DEBUG_OUTPUT3, "No memory?\n");
8712 						if (!SCTP_OS_TIMER_PENDING(&net->rxt_timer.timer)) {
8713 							sctp_timer_start(SCTP_TIMER_TYPE_SEND, inp, stcb, net);
8714 						}
8715 						*reason_code = 3;
8716 						SCTP_LTRACE_ERR_RET(inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
8717 						return (ENOMEM);
8718 					}
8719 					/* upate our MTU size */
8720 					/* Do clear IP_DF ? */
8721 					if (chk->flags & CHUNK_FLAGS_FRAGMENT_OK) {
8722 						no_fragmentflg = 0;
8723 					}
8724 					/* unsigned subtraction of mtu */
8725 					if (mtu > chk->send_size)
8726 						mtu -= chk->send_size;
8727 					else
8728 						mtu = 0;
8729 					/* unsigned subtraction of r_mtu */
8730 					if (r_mtu > chk->send_size)
8731 						r_mtu -= chk->send_size;
8732 					else
8733 						r_mtu = 0;
8734 
8735 					to_out += chk->send_size;
8736 					if ((to_out > mx_mtu) && no_fragmentflg) {
8737 #ifdef INVARIANTS
8738 						panic("Exceeding mtu of %d out size is %d", mx_mtu, to_out);
8739 #else
8740 						SCTP_PRINTF("Exceeding mtu of %d out size is %d\n",
8741 						    mx_mtu, to_out);
8742 #endif
8743 					}
8744 					chk->window_probe = 0;
8745 					data_list[bundle_at++] = chk;
8746 					if (bundle_at >= SCTP_MAX_DATA_BUNDLING) {
8747 						break;
8748 					}
8749 					if (chk->sent == SCTP_DATAGRAM_UNSENT) {
8750 						if ((chk->rec.data.rcv_flags & SCTP_DATA_UNORDERED) == 0) {
8751 							SCTP_STAT_INCR_COUNTER64(sctps_outorderchunks);
8752 						} else {
8753 							SCTP_STAT_INCR_COUNTER64(sctps_outunorderchunks);
8754 						}
8755 						if (((chk->rec.data.rcv_flags & SCTP_DATA_LAST_FRAG) == SCTP_DATA_LAST_FRAG) &&
8756 						    ((chk->rec.data.rcv_flags & SCTP_DATA_FIRST_FRAG) == 0))
8757 							/*
8758 							 * Count number of
8759 							 * user msg's that
8760 							 * were fragmented
8761 							 * we do this by
8762 							 * counting when we
8763 							 * see a LAST
8764 							 * fragment only.
8765 							 */
8766 							SCTP_STAT_INCR_COUNTER64(sctps_fragusrmsgs);
8767 					}
8768 					if ((mtu == 0) || (r_mtu == 0) || (one_chunk)) {
8769 						if ((one_chunk) && (stcb->asoc.total_flight == 0)) {
8770 							data_list[0]->window_probe = 1;
8771 							net->window_probe = 1;
8772 						}
8773 						break;
8774 					}
8775 				} else {
8776 					/*
8777 					 * Must be sent in order of the
8778 					 * TSN's (on a network)
8779 					 */
8780 					break;
8781 				}
8782 			}	/* for (chunk gather loop for this net) */
8783 		}		/* if asoc.state OPEN */
8784 no_data_fill:
8785 		/* Is there something to send for this destination? */
8786 		if (outchain) {
8787 			/* We may need to start a control timer or two */
8788 			if (asconf) {
8789 				sctp_timer_start(SCTP_TIMER_TYPE_ASCONF, inp,
8790 				    stcb, net);
8791 				/*
8792 				 * do NOT clear the asconf flag as it is
8793 				 * used to do appropriate source address
8794 				 * selection.
8795 				 */
8796 			}
8797 			if (cookie) {
8798 				sctp_timer_start(SCTP_TIMER_TYPE_COOKIE, inp, stcb, net);
8799 				cookie = 0;
8800 			}
8801 			/* must start a send timer if data is being sent */
8802 			if (bundle_at && (!SCTP_OS_TIMER_PENDING(&net->rxt_timer.timer))) {
8803 				/*
8804 				 * no timer running on this destination
8805 				 * restart it.
8806 				 */
8807 				sctp_timer_start(SCTP_TIMER_TYPE_SEND, inp, stcb, net);
8808 			}
8809 			if (bundle_at || hbflag) {
8810 				/* For data/asconf and hb set time */
8811 				if (*now_filled == 0) {
8812 					(void)SCTP_GETTIME_TIMEVAL(now);
8813 					*now_filled = 1;
8814 				}
8815 				net->last_sent_time = *now;
8816 			}
8817 			/* Now send it, if there is anything to send :> */
8818 			if ((error = sctp_lowlevel_chunk_output(inp,
8819 			    stcb,
8820 			    net,
8821 			    (struct sockaddr *)&net->ro._l_addr,
8822 			    outchain,
8823 			    auth_offset,
8824 			    auth,
8825 			    auth_keyid,
8826 			    no_fragmentflg,
8827 			    bundle_at,
8828 			    asconf,
8829 			    inp->sctp_lport, stcb->rport,
8830 			    htonl(stcb->asoc.peer_vtag),
8831 			    net->port, NULL,
8832 			    0, 0,
8833 			    so_locked))) {
8834 				/* error, we could not output */
8835 				SCTPDBG(SCTP_DEBUG_OUTPUT3, "Gak send error %d\n", error);
8836 				if (from_where == 0) {
8837 					SCTP_STAT_INCR(sctps_lowlevelerrusr);
8838 				}
8839 				if (error == ENOBUFS) {
8840 					asoc->ifp_had_enobuf = 1;
8841 					SCTP_STAT_INCR(sctps_lowlevelerr);
8842 				}
8843 				if (error == EHOSTUNREACH) {
8844 					/*
8845 					 * Destination went unreachable
8846 					 * during this send
8847 					 */
8848 					sctp_move_chunks_from_net(stcb, net);
8849 				}
8850 				*reason_code = 6;
8851 				/*-
8852 				 * I add this line to be paranoid. As far as
8853 				 * I can tell the continue, takes us back to
8854 				 * the top of the for, but just to make sure
8855 				 * I will reset these again here.
8856 				 */
8857 				ctl_cnt = bundle_at = 0;
8858 				continue;	/* This takes us back to the
8859 						 * for() for the nets. */
8860 			} else {
8861 				asoc->ifp_had_enobuf = 0;
8862 			}
8863 			endoutchain = NULL;
8864 			auth = NULL;
8865 			auth_offset = 0;
8866 			if (!no_out_cnt) {
8867 				*num_out += (ctl_cnt + bundle_at);
8868 			}
8869 			if (bundle_at) {
8870 				/* setup for a RTO measurement */
8871 				tsns_sent = data_list[0]->rec.data.tsn;
8872 				/* fill time if not already filled */
8873 				if (*now_filled == 0) {
8874 					(void)SCTP_GETTIME_TIMEVAL(&asoc->time_last_sent);
8875 					*now_filled = 1;
8876 					*now = asoc->time_last_sent;
8877 				} else {
8878 					asoc->time_last_sent = *now;
8879 				}
8880 				if (net->rto_needed) {
8881 					data_list[0]->do_rtt = 1;
8882 					net->rto_needed = 0;
8883 				}
8884 				SCTP_STAT_INCR_BY(sctps_senddata, bundle_at);
8885 				sctp_clean_up_datalist(stcb, asoc, data_list, bundle_at, net);
8886 			}
8887 			if (one_chunk) {
8888 				break;
8889 			}
8890 		}
8891 		if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_CWND_LOGGING_ENABLE) {
8892 			sctp_log_cwnd(stcb, net, tsns_sent, SCTP_CWND_LOG_FROM_SEND);
8893 		}
8894 	}
8895 	if (old_start_at == NULL) {
8896 		old_start_at = start_at;
8897 		start_at = TAILQ_FIRST(&asoc->nets);
8898 		if (old_start_at)
8899 			goto again_one_more_time;
8900 	}
8901 
8902 	/*
8903 	 * At the end there should be no NON timed chunks hanging on this
8904 	 * queue.
8905 	 */
8906 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_CWND_LOGGING_ENABLE) {
8907 		sctp_log_cwnd(stcb, net, *num_out, SCTP_CWND_LOG_FROM_SEND);
8908 	}
8909 	if ((*num_out == 0) && (*reason_code == 0)) {
8910 		*reason_code = 4;
8911 	} else {
8912 		*reason_code = 5;
8913 	}
8914 	sctp_clean_up_ctl(stcb, asoc, so_locked);
8915 	return (0);
8916 }
8917 
8918 void
8919 sctp_queue_op_err(struct sctp_tcb *stcb, struct mbuf *op_err)
8920 {
8921 	/*-
8922 	 * Prepend a OPERATIONAL_ERROR chunk header and put on the end of
8923 	 * the control chunk queue.
8924 	 */
8925 	struct sctp_chunkhdr *hdr;
8926 	struct sctp_tmit_chunk *chk;
8927 	struct mbuf *mat, *last_mbuf;
8928 	uint32_t chunk_length;
8929 	uint16_t padding_length;
8930 
8931 	SCTP_TCB_LOCK_ASSERT(stcb);
8932 	SCTP_BUF_PREPEND(op_err, sizeof(struct sctp_chunkhdr), M_NOWAIT);
8933 	if (op_err == NULL) {
8934 		return;
8935 	}
8936 	last_mbuf = NULL;
8937 	chunk_length = 0;
8938 	for (mat = op_err; mat != NULL; mat = SCTP_BUF_NEXT(mat)) {
8939 		chunk_length += SCTP_BUF_LEN(mat);
8940 		if (SCTP_BUF_NEXT(mat) == NULL) {
8941 			last_mbuf = mat;
8942 		}
8943 	}
8944 	if (chunk_length > SCTP_MAX_CHUNK_LENGTH) {
8945 		sctp_m_freem(op_err);
8946 		return;
8947 	}
8948 	padding_length = chunk_length % 4;
8949 	if (padding_length != 0) {
8950 		padding_length = 4 - padding_length;
8951 	}
8952 	if (padding_length != 0) {
8953 		if (sctp_add_pad_tombuf(last_mbuf, padding_length) == NULL) {
8954 			sctp_m_freem(op_err);
8955 			return;
8956 		}
8957 	}
8958 	sctp_alloc_a_chunk(stcb, chk);
8959 	if (chk == NULL) {
8960 		/* no memory */
8961 		sctp_m_freem(op_err);
8962 		return;
8963 	}
8964 	chk->copy_by_ref = 0;
8965 	chk->send_size = (uint16_t)chunk_length;
8966 	chk->sent = SCTP_DATAGRAM_UNSENT;
8967 	chk->snd_count = 0;
8968 	chk->asoc = &stcb->asoc;
8969 	chk->data = op_err;
8970 	chk->whoTo = NULL;
8971 	chk->rec.chunk_id.id = SCTP_OPERATION_ERROR;
8972 	chk->rec.chunk_id.can_take_data = 0;
8973 	hdr = mtod(op_err, struct sctp_chunkhdr *);
8974 	hdr->chunk_type = SCTP_OPERATION_ERROR;
8975 	hdr->chunk_flags = 0;
8976 	hdr->chunk_length = htons(chk->send_size);
8977 	TAILQ_INSERT_TAIL(&chk->asoc->control_send_queue, chk, sctp_next);
8978 	chk->asoc->ctrl_queue_cnt++;
8979 }
8980 
8981 int
8982 sctp_send_cookie_echo(struct mbuf *m,
8983     int offset,
8984     struct sctp_tcb *stcb,
8985     struct sctp_nets *net)
8986 {
8987 	/*-
8988 	 * pull out the cookie and put it at the front of the control chunk
8989 	 * queue.
8990 	 */
8991 	int at;
8992 	struct mbuf *cookie;
8993 	struct sctp_paramhdr param, *phdr;
8994 	struct sctp_chunkhdr *hdr;
8995 	struct sctp_tmit_chunk *chk;
8996 	uint16_t ptype, plen;
8997 
8998 	SCTP_TCB_LOCK_ASSERT(stcb);
8999 	/* First find the cookie in the param area */
9000 	cookie = NULL;
9001 	at = offset + sizeof(struct sctp_init_chunk);
9002 	for (;;) {
9003 		phdr = sctp_get_next_param(m, at, &param, sizeof(param));
9004 		if (phdr == NULL) {
9005 			return (-3);
9006 		}
9007 		ptype = ntohs(phdr->param_type);
9008 		plen = ntohs(phdr->param_length);
9009 		if (ptype == SCTP_STATE_COOKIE) {
9010 			int pad;
9011 
9012 			/* found the cookie */
9013 			if ((pad = (plen % 4))) {
9014 				plen += 4 - pad;
9015 			}
9016 			cookie = SCTP_M_COPYM(m, at, plen, M_NOWAIT);
9017 			if (cookie == NULL) {
9018 				/* No memory */
9019 				return (-2);
9020 			}
9021 #ifdef SCTP_MBUF_LOGGING
9022 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_MBUF_LOGGING_ENABLE) {
9023 				sctp_log_mbc(cookie, SCTP_MBUF_ICOPY);
9024 			}
9025 #endif
9026 			break;
9027 		}
9028 		at += SCTP_SIZE32(plen);
9029 	}
9030 	/* ok, we got the cookie lets change it into a cookie echo chunk */
9031 	/* first the change from param to cookie */
9032 	hdr = mtod(cookie, struct sctp_chunkhdr *);
9033 	hdr->chunk_type = SCTP_COOKIE_ECHO;
9034 	hdr->chunk_flags = 0;
9035 	/* get the chunk stuff now and place it in the FRONT of the queue */
9036 	sctp_alloc_a_chunk(stcb, chk);
9037 	if (chk == NULL) {
9038 		/* no memory */
9039 		sctp_m_freem(cookie);
9040 		return (-5);
9041 	}
9042 	chk->copy_by_ref = 0;
9043 	chk->rec.chunk_id.id = SCTP_COOKIE_ECHO;
9044 	chk->rec.chunk_id.can_take_data = 0;
9045 	chk->flags = CHUNK_FLAGS_FRAGMENT_OK;
9046 	chk->send_size = plen;
9047 	chk->sent = SCTP_DATAGRAM_UNSENT;
9048 	chk->snd_count = 0;
9049 	chk->asoc = &stcb->asoc;
9050 	chk->data = cookie;
9051 	chk->whoTo = net;
9052 	atomic_add_int(&chk->whoTo->ref_count, 1);
9053 	TAILQ_INSERT_HEAD(&chk->asoc->control_send_queue, chk, sctp_next);
9054 	chk->asoc->ctrl_queue_cnt++;
9055 	return (0);
9056 }
9057 
9058 void
9059 sctp_send_heartbeat_ack(struct sctp_tcb *stcb,
9060     struct mbuf *m,
9061     int offset,
9062     int chk_length,
9063     struct sctp_nets *net)
9064 {
9065 	/*
9066 	 * take a HB request and make it into a HB ack and send it.
9067 	 */
9068 	struct mbuf *outchain;
9069 	struct sctp_chunkhdr *chdr;
9070 	struct sctp_tmit_chunk *chk;
9071 
9072 
9073 	if (net == NULL)
9074 		/* must have a net pointer */
9075 		return;
9076 
9077 	outchain = SCTP_M_COPYM(m, offset, chk_length, M_NOWAIT);
9078 	if (outchain == NULL) {
9079 		/* gak out of memory */
9080 		return;
9081 	}
9082 #ifdef SCTP_MBUF_LOGGING
9083 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_MBUF_LOGGING_ENABLE) {
9084 		sctp_log_mbc(outchain, SCTP_MBUF_ICOPY);
9085 	}
9086 #endif
9087 	chdr = mtod(outchain, struct sctp_chunkhdr *);
9088 	chdr->chunk_type = SCTP_HEARTBEAT_ACK;
9089 	chdr->chunk_flags = 0;
9090 	if (chk_length % 4) {
9091 		/* need pad */
9092 		uint32_t cpthis = 0;
9093 		int padlen;
9094 
9095 		padlen = 4 - (chk_length % 4);
9096 		m_copyback(outchain, chk_length, padlen, (caddr_t)&cpthis);
9097 	}
9098 	sctp_alloc_a_chunk(stcb, chk);
9099 	if (chk == NULL) {
9100 		/* no memory */
9101 		sctp_m_freem(outchain);
9102 		return;
9103 	}
9104 	chk->copy_by_ref = 0;
9105 	chk->rec.chunk_id.id = SCTP_HEARTBEAT_ACK;
9106 	chk->rec.chunk_id.can_take_data = 1;
9107 	chk->flags = 0;
9108 	chk->send_size = chk_length;
9109 	chk->sent = SCTP_DATAGRAM_UNSENT;
9110 	chk->snd_count = 0;
9111 	chk->asoc = &stcb->asoc;
9112 	chk->data = outchain;
9113 	chk->whoTo = net;
9114 	atomic_add_int(&chk->whoTo->ref_count, 1);
9115 	TAILQ_INSERT_TAIL(&chk->asoc->control_send_queue, chk, sctp_next);
9116 	chk->asoc->ctrl_queue_cnt++;
9117 }
9118 
9119 void
9120 sctp_send_cookie_ack(struct sctp_tcb *stcb)
9121 {
9122 	/* formulate and queue a cookie-ack back to sender */
9123 	struct mbuf *cookie_ack;
9124 	struct sctp_chunkhdr *hdr;
9125 	struct sctp_tmit_chunk *chk;
9126 
9127 	SCTP_TCB_LOCK_ASSERT(stcb);
9128 
9129 	cookie_ack = sctp_get_mbuf_for_msg(sizeof(struct sctp_chunkhdr), 0, M_NOWAIT, 1, MT_HEADER);
9130 	if (cookie_ack == NULL) {
9131 		/* no mbuf's */
9132 		return;
9133 	}
9134 	SCTP_BUF_RESV_UF(cookie_ack, SCTP_MIN_OVERHEAD);
9135 	sctp_alloc_a_chunk(stcb, chk);
9136 	if (chk == NULL) {
9137 		/* no memory */
9138 		sctp_m_freem(cookie_ack);
9139 		return;
9140 	}
9141 	chk->copy_by_ref = 0;
9142 	chk->rec.chunk_id.id = SCTP_COOKIE_ACK;
9143 	chk->rec.chunk_id.can_take_data = 1;
9144 	chk->flags = 0;
9145 	chk->send_size = sizeof(struct sctp_chunkhdr);
9146 	chk->sent = SCTP_DATAGRAM_UNSENT;
9147 	chk->snd_count = 0;
9148 	chk->asoc = &stcb->asoc;
9149 	chk->data = cookie_ack;
9150 	if (chk->asoc->last_control_chunk_from != NULL) {
9151 		chk->whoTo = chk->asoc->last_control_chunk_from;
9152 		atomic_add_int(&chk->whoTo->ref_count, 1);
9153 	} else {
9154 		chk->whoTo = NULL;
9155 	}
9156 	hdr = mtod(cookie_ack, struct sctp_chunkhdr *);
9157 	hdr->chunk_type = SCTP_COOKIE_ACK;
9158 	hdr->chunk_flags = 0;
9159 	hdr->chunk_length = htons(chk->send_size);
9160 	SCTP_BUF_LEN(cookie_ack) = chk->send_size;
9161 	TAILQ_INSERT_TAIL(&chk->asoc->control_send_queue, chk, sctp_next);
9162 	chk->asoc->ctrl_queue_cnt++;
9163 	return;
9164 }
9165 
9166 
9167 void
9168 sctp_send_shutdown_ack(struct sctp_tcb *stcb, struct sctp_nets *net)
9169 {
9170 	/* formulate and queue a SHUTDOWN-ACK back to the sender */
9171 	struct mbuf *m_shutdown_ack;
9172 	struct sctp_shutdown_ack_chunk *ack_cp;
9173 	struct sctp_tmit_chunk *chk;
9174 
9175 	m_shutdown_ack = sctp_get_mbuf_for_msg(sizeof(struct sctp_shutdown_ack_chunk), 0, M_NOWAIT, 1, MT_HEADER);
9176 	if (m_shutdown_ack == NULL) {
9177 		/* no mbuf's */
9178 		return;
9179 	}
9180 	SCTP_BUF_RESV_UF(m_shutdown_ack, SCTP_MIN_OVERHEAD);
9181 	sctp_alloc_a_chunk(stcb, chk);
9182 	if (chk == NULL) {
9183 		/* no memory */
9184 		sctp_m_freem(m_shutdown_ack);
9185 		return;
9186 	}
9187 	chk->copy_by_ref = 0;
9188 	chk->rec.chunk_id.id = SCTP_SHUTDOWN_ACK;
9189 	chk->rec.chunk_id.can_take_data = 1;
9190 	chk->flags = 0;
9191 	chk->send_size = sizeof(struct sctp_chunkhdr);
9192 	chk->sent = SCTP_DATAGRAM_UNSENT;
9193 	chk->snd_count = 0;
9194 	chk->flags = 0;
9195 	chk->asoc = &stcb->asoc;
9196 	chk->data = m_shutdown_ack;
9197 	chk->whoTo = net;
9198 	if (chk->whoTo) {
9199 		atomic_add_int(&chk->whoTo->ref_count, 1);
9200 	}
9201 	ack_cp = mtod(m_shutdown_ack, struct sctp_shutdown_ack_chunk *);
9202 	ack_cp->ch.chunk_type = SCTP_SHUTDOWN_ACK;
9203 	ack_cp->ch.chunk_flags = 0;
9204 	ack_cp->ch.chunk_length = htons(chk->send_size);
9205 	SCTP_BUF_LEN(m_shutdown_ack) = chk->send_size;
9206 	TAILQ_INSERT_TAIL(&chk->asoc->control_send_queue, chk, sctp_next);
9207 	chk->asoc->ctrl_queue_cnt++;
9208 	return;
9209 }
9210 
9211 void
9212 sctp_send_shutdown(struct sctp_tcb *stcb, struct sctp_nets *net)
9213 {
9214 	/* formulate and queue a SHUTDOWN to the sender */
9215 	struct mbuf *m_shutdown;
9216 	struct sctp_shutdown_chunk *shutdown_cp;
9217 	struct sctp_tmit_chunk *chk;
9218 
9219 	TAILQ_FOREACH(chk, &stcb->asoc.control_send_queue, sctp_next) {
9220 		if (chk->rec.chunk_id.id == SCTP_SHUTDOWN) {
9221 			/* We already have a SHUTDOWN queued. Reuse it. */
9222 			if (chk->whoTo) {
9223 				sctp_free_remote_addr(chk->whoTo);
9224 				chk->whoTo = NULL;
9225 			}
9226 			break;
9227 		}
9228 	}
9229 	if (chk == NULL) {
9230 		m_shutdown = sctp_get_mbuf_for_msg(sizeof(struct sctp_shutdown_chunk), 0, M_NOWAIT, 1, MT_HEADER);
9231 		if (m_shutdown == NULL) {
9232 			/* no mbuf's */
9233 			return;
9234 		}
9235 		SCTP_BUF_RESV_UF(m_shutdown, SCTP_MIN_OVERHEAD);
9236 		sctp_alloc_a_chunk(stcb, chk);
9237 		if (chk == NULL) {
9238 			/* no memory */
9239 			sctp_m_freem(m_shutdown);
9240 			return;
9241 		}
9242 		chk->copy_by_ref = 0;
9243 		chk->rec.chunk_id.id = SCTP_SHUTDOWN;
9244 		chk->rec.chunk_id.can_take_data = 1;
9245 		chk->flags = 0;
9246 		chk->send_size = sizeof(struct sctp_shutdown_chunk);
9247 		chk->sent = SCTP_DATAGRAM_UNSENT;
9248 		chk->snd_count = 0;
9249 		chk->flags = 0;
9250 		chk->asoc = &stcb->asoc;
9251 		chk->data = m_shutdown;
9252 		chk->whoTo = net;
9253 		if (chk->whoTo) {
9254 			atomic_add_int(&chk->whoTo->ref_count, 1);
9255 		}
9256 		shutdown_cp = mtod(m_shutdown, struct sctp_shutdown_chunk *);
9257 		shutdown_cp->ch.chunk_type = SCTP_SHUTDOWN;
9258 		shutdown_cp->ch.chunk_flags = 0;
9259 		shutdown_cp->ch.chunk_length = htons(chk->send_size);
9260 		shutdown_cp->cumulative_tsn_ack = htonl(stcb->asoc.cumulative_tsn);
9261 		SCTP_BUF_LEN(m_shutdown) = chk->send_size;
9262 		TAILQ_INSERT_TAIL(&chk->asoc->control_send_queue, chk, sctp_next);
9263 		chk->asoc->ctrl_queue_cnt++;
9264 	} else {
9265 		TAILQ_REMOVE(&stcb->asoc.control_send_queue, chk, sctp_next);
9266 		chk->whoTo = net;
9267 		if (chk->whoTo) {
9268 			atomic_add_int(&chk->whoTo->ref_count, 1);
9269 		}
9270 		shutdown_cp = mtod(chk->data, struct sctp_shutdown_chunk *);
9271 		shutdown_cp->cumulative_tsn_ack = htonl(stcb->asoc.cumulative_tsn);
9272 		TAILQ_INSERT_TAIL(&stcb->asoc.control_send_queue, chk, sctp_next);
9273 	}
9274 	return;
9275 }
9276 
9277 void
9278 sctp_send_asconf(struct sctp_tcb *stcb, struct sctp_nets *net, int addr_locked)
9279 {
9280 	/*
9281 	 * formulate and queue an ASCONF to the peer. ASCONF parameters
9282 	 * should be queued on the assoc queue.
9283 	 */
9284 	struct sctp_tmit_chunk *chk;
9285 	struct mbuf *m_asconf;
9286 	int len;
9287 
9288 	SCTP_TCB_LOCK_ASSERT(stcb);
9289 
9290 	if ((!TAILQ_EMPTY(&stcb->asoc.asconf_send_queue)) &&
9291 	    (!sctp_is_feature_on(stcb->sctp_ep, SCTP_PCB_FLAGS_MULTIPLE_ASCONFS))) {
9292 		/* can't send a new one if there is one in flight already */
9293 		return;
9294 	}
9295 
9296 	/* compose an ASCONF chunk, maximum length is PMTU */
9297 	m_asconf = sctp_compose_asconf(stcb, &len, addr_locked);
9298 	if (m_asconf == NULL) {
9299 		return;
9300 	}
9301 
9302 	sctp_alloc_a_chunk(stcb, chk);
9303 	if (chk == NULL) {
9304 		/* no memory */
9305 		sctp_m_freem(m_asconf);
9306 		return;
9307 	}
9308 
9309 	chk->copy_by_ref = 0;
9310 	chk->rec.chunk_id.id = SCTP_ASCONF;
9311 	chk->rec.chunk_id.can_take_data = 0;
9312 	chk->flags = CHUNK_FLAGS_FRAGMENT_OK;
9313 	chk->data = m_asconf;
9314 	chk->send_size = len;
9315 	chk->sent = SCTP_DATAGRAM_UNSENT;
9316 	chk->snd_count = 0;
9317 	chk->asoc = &stcb->asoc;
9318 	chk->whoTo = net;
9319 	if (chk->whoTo) {
9320 		atomic_add_int(&chk->whoTo->ref_count, 1);
9321 	}
9322 	TAILQ_INSERT_TAIL(&chk->asoc->asconf_send_queue, chk, sctp_next);
9323 	chk->asoc->ctrl_queue_cnt++;
9324 	return;
9325 }
9326 
9327 void
9328 sctp_send_asconf_ack(struct sctp_tcb *stcb)
9329 {
9330 	/*
9331 	 * formulate and queue a asconf-ack back to sender. the asconf-ack
9332 	 * must be stored in the tcb.
9333 	 */
9334 	struct sctp_tmit_chunk *chk;
9335 	struct sctp_asconf_ack *ack, *latest_ack;
9336 	struct mbuf *m_ack;
9337 	struct sctp_nets *net = NULL;
9338 
9339 	SCTP_TCB_LOCK_ASSERT(stcb);
9340 	/* Get the latest ASCONF-ACK */
9341 	latest_ack = TAILQ_LAST(&stcb->asoc.asconf_ack_sent, sctp_asconf_ackhead);
9342 	if (latest_ack == NULL) {
9343 		return;
9344 	}
9345 	if (latest_ack->last_sent_to != NULL &&
9346 	    latest_ack->last_sent_to == stcb->asoc.last_control_chunk_from) {
9347 		/* we're doing a retransmission */
9348 		net = sctp_find_alternate_net(stcb, stcb->asoc.last_control_chunk_from, 0);
9349 		if (net == NULL) {
9350 			/* no alternate */
9351 			if (stcb->asoc.last_control_chunk_from == NULL) {
9352 				if (stcb->asoc.alternate) {
9353 					net = stcb->asoc.alternate;
9354 				} else {
9355 					net = stcb->asoc.primary_destination;
9356 				}
9357 			} else {
9358 				net = stcb->asoc.last_control_chunk_from;
9359 			}
9360 		}
9361 	} else {
9362 		/* normal case */
9363 		if (stcb->asoc.last_control_chunk_from == NULL) {
9364 			if (stcb->asoc.alternate) {
9365 				net = stcb->asoc.alternate;
9366 			} else {
9367 				net = stcb->asoc.primary_destination;
9368 			}
9369 		} else {
9370 			net = stcb->asoc.last_control_chunk_from;
9371 		}
9372 	}
9373 	latest_ack->last_sent_to = net;
9374 
9375 	TAILQ_FOREACH(ack, &stcb->asoc.asconf_ack_sent, next) {
9376 		if (ack->data == NULL) {
9377 			continue;
9378 		}
9379 
9380 		/* copy the asconf_ack */
9381 		m_ack = SCTP_M_COPYM(ack->data, 0, M_COPYALL, M_NOWAIT);
9382 		if (m_ack == NULL) {
9383 			/* couldn't copy it */
9384 			return;
9385 		}
9386 #ifdef SCTP_MBUF_LOGGING
9387 		if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_MBUF_LOGGING_ENABLE) {
9388 			sctp_log_mbc(m_ack, SCTP_MBUF_ICOPY);
9389 		}
9390 #endif
9391 
9392 		sctp_alloc_a_chunk(stcb, chk);
9393 		if (chk == NULL) {
9394 			/* no memory */
9395 			if (m_ack)
9396 				sctp_m_freem(m_ack);
9397 			return;
9398 		}
9399 		chk->copy_by_ref = 0;
9400 		chk->rec.chunk_id.id = SCTP_ASCONF_ACK;
9401 		chk->rec.chunk_id.can_take_data = 1;
9402 		chk->flags = CHUNK_FLAGS_FRAGMENT_OK;
9403 		chk->whoTo = net;
9404 		if (chk->whoTo) {
9405 			atomic_add_int(&chk->whoTo->ref_count, 1);
9406 		}
9407 		chk->data = m_ack;
9408 		chk->send_size = ack->len;
9409 		chk->sent = SCTP_DATAGRAM_UNSENT;
9410 		chk->snd_count = 0;
9411 		chk->asoc = &stcb->asoc;
9412 
9413 		TAILQ_INSERT_TAIL(&chk->asoc->control_send_queue, chk, sctp_next);
9414 		chk->asoc->ctrl_queue_cnt++;
9415 	}
9416 	return;
9417 }
9418 
9419 
9420 static int
9421 sctp_chunk_retransmission(struct sctp_inpcb *inp,
9422     struct sctp_tcb *stcb,
9423     struct sctp_association *asoc,
9424     int *cnt_out, struct timeval *now, int *now_filled, int *fr_done, int so_locked
9425 #if !defined(__APPLE__) && !defined(SCTP_SO_LOCK_TESTING)
9426     SCTP_UNUSED
9427 #endif
9428 )
9429 {
9430 	/*-
9431 	 * send out one MTU of retransmission. If fast_retransmit is
9432 	 * happening we ignore the cwnd. Otherwise we obey the cwnd and
9433 	 * rwnd. For a Cookie or Asconf in the control chunk queue we
9434 	 * retransmit them by themselves.
9435 	 *
9436 	 * For data chunks we will pick out the lowest TSN's in the sent_queue
9437 	 * marked for resend and bundle them all together (up to a MTU of
9438 	 * destination). The address to send to should have been
9439 	 * selected/changed where the retransmission was marked (i.e. in FR
9440 	 * or t3-timeout routines).
9441 	 */
9442 	struct sctp_tmit_chunk *data_list[SCTP_MAX_DATA_BUNDLING];
9443 	struct sctp_tmit_chunk *chk, *fwd;
9444 	struct mbuf *m, *endofchain;
9445 	struct sctp_nets *net = NULL;
9446 	uint32_t tsns_sent = 0;
9447 	int no_fragmentflg, bundle_at, cnt_thru;
9448 	unsigned int mtu;
9449 	int error, i, one_chunk, fwd_tsn, ctl_cnt, tmr_started;
9450 	struct sctp_auth_chunk *auth = NULL;
9451 	uint32_t auth_offset = 0;
9452 	uint16_t auth_keyid;
9453 	int override_ok = 1;
9454 	int data_auth_reqd = 0;
9455 	uint32_t dmtu = 0;
9456 
9457 	SCTP_TCB_LOCK_ASSERT(stcb);
9458 	tmr_started = ctl_cnt = bundle_at = error = 0;
9459 	no_fragmentflg = 1;
9460 	fwd_tsn = 0;
9461 	*cnt_out = 0;
9462 	fwd = NULL;
9463 	endofchain = m = NULL;
9464 	auth_keyid = stcb->asoc.authinfo.active_keyid;
9465 #ifdef SCTP_AUDITING_ENABLED
9466 	sctp_audit_log(0xC3, 1);
9467 #endif
9468 	if ((TAILQ_EMPTY(&asoc->sent_queue)) &&
9469 	    (TAILQ_EMPTY(&asoc->control_send_queue))) {
9470 		SCTPDBG(SCTP_DEBUG_OUTPUT1, "SCTP hits empty queue with cnt set to %d?\n",
9471 		    asoc->sent_queue_retran_cnt);
9472 		asoc->sent_queue_cnt = 0;
9473 		asoc->sent_queue_cnt_removeable = 0;
9474 		/* send back 0/0 so we enter normal transmission */
9475 		*cnt_out = 0;
9476 		return (0);
9477 	}
9478 	TAILQ_FOREACH(chk, &asoc->control_send_queue, sctp_next) {
9479 		if ((chk->rec.chunk_id.id == SCTP_COOKIE_ECHO) ||
9480 		    (chk->rec.chunk_id.id == SCTP_STREAM_RESET) ||
9481 		    (chk->rec.chunk_id.id == SCTP_FORWARD_CUM_TSN)) {
9482 			if (chk->sent != SCTP_DATAGRAM_RESEND) {
9483 				continue;
9484 			}
9485 			if (chk->rec.chunk_id.id == SCTP_STREAM_RESET) {
9486 				if (chk != asoc->str_reset) {
9487 					/*
9488 					 * not eligible for retran if its
9489 					 * not ours
9490 					 */
9491 					continue;
9492 				}
9493 			}
9494 			ctl_cnt++;
9495 			if (chk->rec.chunk_id.id == SCTP_FORWARD_CUM_TSN) {
9496 				fwd_tsn = 1;
9497 			}
9498 			/*
9499 			 * Add an AUTH chunk, if chunk requires it save the
9500 			 * offset into the chain for AUTH
9501 			 */
9502 			if ((auth == NULL) &&
9503 			    (sctp_auth_is_required_chunk(chk->rec.chunk_id.id,
9504 			    stcb->asoc.peer_auth_chunks))) {
9505 				m = sctp_add_auth_chunk(m, &endofchain,
9506 				    &auth, &auth_offset,
9507 				    stcb,
9508 				    chk->rec.chunk_id.id);
9509 				SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
9510 			}
9511 			m = sctp_copy_mbufchain(chk->data, m, &endofchain, 0, chk->send_size, chk->copy_by_ref);
9512 			break;
9513 		}
9514 	}
9515 	one_chunk = 0;
9516 	cnt_thru = 0;
9517 	/* do we have control chunks to retransmit? */
9518 	if (m != NULL) {
9519 		/* Start a timer no matter if we succeed or fail */
9520 		if (chk->rec.chunk_id.id == SCTP_COOKIE_ECHO) {
9521 			sctp_timer_start(SCTP_TIMER_TYPE_COOKIE, inp, stcb, chk->whoTo);
9522 		} else if (chk->rec.chunk_id.id == SCTP_ASCONF)
9523 			sctp_timer_start(SCTP_TIMER_TYPE_ASCONF, inp, stcb, chk->whoTo);
9524 		chk->snd_count++;	/* update our count */
9525 		if ((error = sctp_lowlevel_chunk_output(inp, stcb, chk->whoTo,
9526 		    (struct sockaddr *)&chk->whoTo->ro._l_addr, m,
9527 		    auth_offset, auth, stcb->asoc.authinfo.active_keyid,
9528 		    no_fragmentflg, 0, 0,
9529 		    inp->sctp_lport, stcb->rport, htonl(stcb->asoc.peer_vtag),
9530 		    chk->whoTo->port, NULL,
9531 		    0, 0,
9532 		    so_locked))) {
9533 			SCTPDBG(SCTP_DEBUG_OUTPUT3, "Gak send error %d\n", error);
9534 			if (error == ENOBUFS) {
9535 				asoc->ifp_had_enobuf = 1;
9536 				SCTP_STAT_INCR(sctps_lowlevelerr);
9537 			}
9538 			return (error);
9539 		} else {
9540 			asoc->ifp_had_enobuf = 0;
9541 		}
9542 		endofchain = NULL;
9543 		auth = NULL;
9544 		auth_offset = 0;
9545 		/*
9546 		 * We don't want to mark the net->sent time here since this
9547 		 * we use this for HB and retrans cannot measure RTT
9548 		 */
9549 		/* (void)SCTP_GETTIME_TIMEVAL(&chk->whoTo->last_sent_time); */
9550 		*cnt_out += 1;
9551 		chk->sent = SCTP_DATAGRAM_SENT;
9552 		sctp_ucount_decr(stcb->asoc.sent_queue_retran_cnt);
9553 		if (fwd_tsn == 0) {
9554 			return (0);
9555 		} else {
9556 			/* Clean up the fwd-tsn list */
9557 			sctp_clean_up_ctl(stcb, asoc, so_locked);
9558 			return (0);
9559 		}
9560 	}
9561 	/*
9562 	 * Ok, it is just data retransmission we need to do or that and a
9563 	 * fwd-tsn with it all.
9564 	 */
9565 	if (TAILQ_EMPTY(&asoc->sent_queue)) {
9566 		return (SCTP_RETRAN_DONE);
9567 	}
9568 	if ((SCTP_GET_STATE(stcb) == SCTP_STATE_COOKIE_ECHOED) ||
9569 	    (SCTP_GET_STATE(stcb) == SCTP_STATE_COOKIE_WAIT)) {
9570 		/* not yet open, resend the cookie and that is it */
9571 		return (1);
9572 	}
9573 #ifdef SCTP_AUDITING_ENABLED
9574 	sctp_auditing(20, inp, stcb, NULL);
9575 #endif
9576 	data_auth_reqd = sctp_auth_is_required_chunk(SCTP_DATA, stcb->asoc.peer_auth_chunks);
9577 	TAILQ_FOREACH(chk, &asoc->sent_queue, sctp_next) {
9578 		if (chk->sent != SCTP_DATAGRAM_RESEND) {
9579 			/* No, not sent to this net or not ready for rtx */
9580 			continue;
9581 		}
9582 		if (chk->data == NULL) {
9583 			SCTP_PRINTF("TSN:%x chk->snd_count:%d chk->sent:%d can't retran - no data\n",
9584 			    chk->rec.data.tsn, chk->snd_count, chk->sent);
9585 			continue;
9586 		}
9587 		if ((SCTP_BASE_SYSCTL(sctp_max_retran_chunk)) &&
9588 		    (chk->snd_count >= SCTP_BASE_SYSCTL(sctp_max_retran_chunk))) {
9589 			struct mbuf *op_err;
9590 			char msg[SCTP_DIAG_INFO_LEN];
9591 
9592 			snprintf(msg, sizeof(msg), "TSN %8.8x retransmitted %d times, giving up",
9593 			    chk->rec.data.tsn, chk->snd_count);
9594 			op_err = sctp_generate_cause(SCTP_BASE_SYSCTL(sctp_diag_info_code),
9595 			    msg);
9596 			atomic_add_int(&stcb->asoc.refcnt, 1);
9597 			sctp_abort_an_association(stcb->sctp_ep, stcb, op_err,
9598 			    so_locked);
9599 			SCTP_TCB_LOCK(stcb);
9600 			atomic_subtract_int(&stcb->asoc.refcnt, 1);
9601 			return (SCTP_RETRAN_EXIT);
9602 		}
9603 		/* pick up the net */
9604 		net = chk->whoTo;
9605 		switch (net->ro._l_addr.sa.sa_family) {
9606 #ifdef INET
9607 		case AF_INET:
9608 			mtu = net->mtu - SCTP_MIN_V4_OVERHEAD;
9609 			break;
9610 #endif
9611 #ifdef INET6
9612 		case AF_INET6:
9613 			mtu = net->mtu - SCTP_MIN_OVERHEAD;
9614 			break;
9615 #endif
9616 		default:
9617 			/* TSNH */
9618 			mtu = net->mtu;
9619 			break;
9620 		}
9621 
9622 		if ((asoc->peers_rwnd < mtu) && (asoc->total_flight > 0)) {
9623 			/* No room in peers rwnd */
9624 			uint32_t tsn;
9625 
9626 			tsn = asoc->last_acked_seq + 1;
9627 			if (tsn == chk->rec.data.tsn) {
9628 				/*
9629 				 * we make a special exception for this
9630 				 * case. The peer has no rwnd but is missing
9631 				 * the lowest chunk.. which is probably what
9632 				 * is holding up the rwnd.
9633 				 */
9634 				goto one_chunk_around;
9635 			}
9636 			return (1);
9637 		}
9638 one_chunk_around:
9639 		if (asoc->peers_rwnd < mtu) {
9640 			one_chunk = 1;
9641 			if ((asoc->peers_rwnd == 0) &&
9642 			    (asoc->total_flight == 0)) {
9643 				chk->window_probe = 1;
9644 				chk->whoTo->window_probe = 1;
9645 			}
9646 		}
9647 #ifdef SCTP_AUDITING_ENABLED
9648 		sctp_audit_log(0xC3, 2);
9649 #endif
9650 		bundle_at = 0;
9651 		m = NULL;
9652 		net->fast_retran_ip = 0;
9653 		if (chk->rec.data.doing_fast_retransmit == 0) {
9654 			/*
9655 			 * if no FR in progress skip destination that have
9656 			 * flight_size > cwnd.
9657 			 */
9658 			if (net->flight_size >= net->cwnd) {
9659 				continue;
9660 			}
9661 		} else {
9662 			/*
9663 			 * Mark the destination net to have FR recovery
9664 			 * limits put on it.
9665 			 */
9666 			*fr_done = 1;
9667 			net->fast_retran_ip = 1;
9668 		}
9669 
9670 		/*
9671 		 * if no AUTH is yet included and this chunk requires it,
9672 		 * make sure to account for it.  We don't apply the size
9673 		 * until the AUTH chunk is actually added below in case
9674 		 * there is no room for this chunk.
9675 		 */
9676 		if (data_auth_reqd && (auth == NULL)) {
9677 			dmtu = sctp_get_auth_chunk_len(stcb->asoc.peer_hmac_id);
9678 		} else
9679 			dmtu = 0;
9680 
9681 		if ((chk->send_size <= (mtu - dmtu)) ||
9682 		    (chk->flags & CHUNK_FLAGS_FRAGMENT_OK)) {
9683 			/* ok we will add this one */
9684 			if (data_auth_reqd) {
9685 				if (auth == NULL) {
9686 					m = sctp_add_auth_chunk(m,
9687 					    &endofchain,
9688 					    &auth,
9689 					    &auth_offset,
9690 					    stcb,
9691 					    SCTP_DATA);
9692 					auth_keyid = chk->auth_keyid;
9693 					override_ok = 0;
9694 					SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
9695 				} else if (override_ok) {
9696 					auth_keyid = chk->auth_keyid;
9697 					override_ok = 0;
9698 				} else if (chk->auth_keyid != auth_keyid) {
9699 					/* different keyid, so done bundling */
9700 					break;
9701 				}
9702 			}
9703 			m = sctp_copy_mbufchain(chk->data, m, &endofchain, 0, chk->send_size, chk->copy_by_ref);
9704 			if (m == NULL) {
9705 				SCTP_LTRACE_ERR_RET(inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
9706 				return (ENOMEM);
9707 			}
9708 			/* Do clear IP_DF ? */
9709 			if (chk->flags & CHUNK_FLAGS_FRAGMENT_OK) {
9710 				no_fragmentflg = 0;
9711 			}
9712 			/* upate our MTU size */
9713 			if (mtu > (chk->send_size + dmtu))
9714 				mtu -= (chk->send_size + dmtu);
9715 			else
9716 				mtu = 0;
9717 			data_list[bundle_at++] = chk;
9718 			if (one_chunk && (asoc->total_flight <= 0)) {
9719 				SCTP_STAT_INCR(sctps_windowprobed);
9720 			}
9721 		}
9722 		if (one_chunk == 0) {
9723 			/*
9724 			 * now are there anymore forward from chk to pick
9725 			 * up?
9726 			 */
9727 			for (fwd = TAILQ_NEXT(chk, sctp_next); fwd != NULL; fwd = TAILQ_NEXT(fwd, sctp_next)) {
9728 				if (fwd->sent != SCTP_DATAGRAM_RESEND) {
9729 					/* Nope, not for retran */
9730 					continue;
9731 				}
9732 				if (fwd->whoTo != net) {
9733 					/* Nope, not the net in question */
9734 					continue;
9735 				}
9736 				if (data_auth_reqd && (auth == NULL)) {
9737 					dmtu = sctp_get_auth_chunk_len(stcb->asoc.peer_hmac_id);
9738 				} else
9739 					dmtu = 0;
9740 				if (fwd->send_size <= (mtu - dmtu)) {
9741 					if (data_auth_reqd) {
9742 						if (auth == NULL) {
9743 							m = sctp_add_auth_chunk(m,
9744 							    &endofchain,
9745 							    &auth,
9746 							    &auth_offset,
9747 							    stcb,
9748 							    SCTP_DATA);
9749 							auth_keyid = fwd->auth_keyid;
9750 							override_ok = 0;
9751 							SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
9752 						} else if (override_ok) {
9753 							auth_keyid = fwd->auth_keyid;
9754 							override_ok = 0;
9755 						} else if (fwd->auth_keyid != auth_keyid) {
9756 							/*
9757 							 * different keyid,
9758 							 * so done bundling
9759 							 */
9760 							break;
9761 						}
9762 					}
9763 					m = sctp_copy_mbufchain(fwd->data, m, &endofchain, 0, fwd->send_size, fwd->copy_by_ref);
9764 					if (m == NULL) {
9765 						SCTP_LTRACE_ERR_RET(inp, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
9766 						return (ENOMEM);
9767 					}
9768 					/* Do clear IP_DF ? */
9769 					if (fwd->flags & CHUNK_FLAGS_FRAGMENT_OK) {
9770 						no_fragmentflg = 0;
9771 					}
9772 					/* upate our MTU size */
9773 					if (mtu > (fwd->send_size + dmtu))
9774 						mtu -= (fwd->send_size + dmtu);
9775 					else
9776 						mtu = 0;
9777 					data_list[bundle_at++] = fwd;
9778 					if (bundle_at >= SCTP_MAX_DATA_BUNDLING) {
9779 						break;
9780 					}
9781 				} else {
9782 					/* can't fit so we are done */
9783 					break;
9784 				}
9785 			}
9786 		}
9787 		/* Is there something to send for this destination? */
9788 		if (m) {
9789 			/*
9790 			 * No matter if we fail/or succeed we should start a
9791 			 * timer. A failure is like a lost IP packet :-)
9792 			 */
9793 			if (!SCTP_OS_TIMER_PENDING(&net->rxt_timer.timer)) {
9794 				/*
9795 				 * no timer running on this destination
9796 				 * restart it.
9797 				 */
9798 				sctp_timer_start(SCTP_TIMER_TYPE_SEND, inp, stcb, net);
9799 				tmr_started = 1;
9800 			}
9801 			/* Now lets send it, if there is anything to send :> */
9802 			if ((error = sctp_lowlevel_chunk_output(inp, stcb, net,
9803 			    (struct sockaddr *)&net->ro._l_addr, m,
9804 			    auth_offset, auth, auth_keyid,
9805 			    no_fragmentflg, 0, 0,
9806 			    inp->sctp_lport, stcb->rport, htonl(stcb->asoc.peer_vtag),
9807 			    net->port, NULL,
9808 			    0, 0,
9809 			    so_locked))) {
9810 				/* error, we could not output */
9811 				SCTPDBG(SCTP_DEBUG_OUTPUT3, "Gak send error %d\n", error);
9812 				if (error == ENOBUFS) {
9813 					asoc->ifp_had_enobuf = 1;
9814 					SCTP_STAT_INCR(sctps_lowlevelerr);
9815 				}
9816 				return (error);
9817 			} else {
9818 				asoc->ifp_had_enobuf = 0;
9819 			}
9820 			endofchain = NULL;
9821 			auth = NULL;
9822 			auth_offset = 0;
9823 			/* For HB's */
9824 			/*
9825 			 * We don't want to mark the net->sent time here
9826 			 * since this we use this for HB and retrans cannot
9827 			 * measure RTT
9828 			 */
9829 			/* (void)SCTP_GETTIME_TIMEVAL(&net->last_sent_time); */
9830 
9831 			/* For auto-close */
9832 			cnt_thru++;
9833 			if (*now_filled == 0) {
9834 				(void)SCTP_GETTIME_TIMEVAL(&asoc->time_last_sent);
9835 				*now = asoc->time_last_sent;
9836 				*now_filled = 1;
9837 			} else {
9838 				asoc->time_last_sent = *now;
9839 			}
9840 			*cnt_out += bundle_at;
9841 #ifdef SCTP_AUDITING_ENABLED
9842 			sctp_audit_log(0xC4, bundle_at);
9843 #endif
9844 			if (bundle_at) {
9845 				tsns_sent = data_list[0]->rec.data.tsn;
9846 			}
9847 			for (i = 0; i < bundle_at; i++) {
9848 				SCTP_STAT_INCR(sctps_sendretransdata);
9849 				data_list[i]->sent = SCTP_DATAGRAM_SENT;
9850 				/*
9851 				 * When we have a revoked data, and we
9852 				 * retransmit it, then we clear the revoked
9853 				 * flag since this flag dictates if we
9854 				 * subtracted from the fs
9855 				 */
9856 				if (data_list[i]->rec.data.chunk_was_revoked) {
9857 					/* Deflate the cwnd */
9858 					data_list[i]->whoTo->cwnd -= data_list[i]->book_size;
9859 					data_list[i]->rec.data.chunk_was_revoked = 0;
9860 				}
9861 				data_list[i]->snd_count++;
9862 				sctp_ucount_decr(asoc->sent_queue_retran_cnt);
9863 				/* record the time */
9864 				data_list[i]->sent_rcv_time = asoc->time_last_sent;
9865 				if (data_list[i]->book_size_scale) {
9866 					/*
9867 					 * need to double the book size on
9868 					 * this one
9869 					 */
9870 					data_list[i]->book_size_scale = 0;
9871 					/*
9872 					 * Since we double the booksize, we
9873 					 * must also double the output queue
9874 					 * size, since this get shrunk when
9875 					 * we free by this amount.
9876 					 */
9877 					atomic_add_int(&((asoc)->total_output_queue_size), data_list[i]->book_size);
9878 					data_list[i]->book_size *= 2;
9879 
9880 
9881 				} else {
9882 					if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOG_RWND_ENABLE) {
9883 						sctp_log_rwnd(SCTP_DECREASE_PEER_RWND,
9884 						    asoc->peers_rwnd, data_list[i]->send_size, SCTP_BASE_SYSCTL(sctp_peer_chunk_oh));
9885 					}
9886 					asoc->peers_rwnd = sctp_sbspace_sub(asoc->peers_rwnd,
9887 					    (uint32_t)(data_list[i]->send_size +
9888 					    SCTP_BASE_SYSCTL(sctp_peer_chunk_oh)));
9889 				}
9890 				if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_FLIGHT_LOGGING_ENABLE) {
9891 					sctp_misc_ints(SCTP_FLIGHT_LOG_UP_RSND,
9892 					    data_list[i]->whoTo->flight_size,
9893 					    data_list[i]->book_size,
9894 					    (uint32_t)(uintptr_t)data_list[i]->whoTo,
9895 					    data_list[i]->rec.data.tsn);
9896 				}
9897 				sctp_flight_size_increase(data_list[i]);
9898 				sctp_total_flight_increase(stcb, data_list[i]);
9899 				if (asoc->peers_rwnd < stcb->sctp_ep->sctp_ep.sctp_sws_sender) {
9900 					/* SWS sender side engages */
9901 					asoc->peers_rwnd = 0;
9902 				}
9903 				if ((i == 0) &&
9904 				    (data_list[i]->rec.data.doing_fast_retransmit)) {
9905 					SCTP_STAT_INCR(sctps_sendfastretrans);
9906 					if ((data_list[i] == TAILQ_FIRST(&asoc->sent_queue)) &&
9907 					    (tmr_started == 0)) {
9908 						/*-
9909 						 * ok we just fast-retrans'd
9910 						 * the lowest TSN, i.e the
9911 						 * first on the list. In
9912 						 * this case we want to give
9913 						 * some more time to get a
9914 						 * SACK back without a
9915 						 * t3-expiring.
9916 						 */
9917 						sctp_timer_stop(SCTP_TIMER_TYPE_SEND, inp, stcb, net,
9918 						    SCTP_FROM_SCTP_OUTPUT + SCTP_LOC_2);
9919 						sctp_timer_start(SCTP_TIMER_TYPE_SEND, inp, stcb, net);
9920 					}
9921 				}
9922 			}
9923 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_CWND_LOGGING_ENABLE) {
9924 				sctp_log_cwnd(stcb, net, tsns_sent, SCTP_CWND_LOG_FROM_RESEND);
9925 			}
9926 #ifdef SCTP_AUDITING_ENABLED
9927 			sctp_auditing(21, inp, stcb, NULL);
9928 #endif
9929 		} else {
9930 			/* None will fit */
9931 			return (1);
9932 		}
9933 		if (asoc->sent_queue_retran_cnt <= 0) {
9934 			/* all done we have no more to retran */
9935 			asoc->sent_queue_retran_cnt = 0;
9936 			break;
9937 		}
9938 		if (one_chunk) {
9939 			/* No more room in rwnd */
9940 			return (1);
9941 		}
9942 		/* stop the for loop here. we sent out a packet */
9943 		break;
9944 	}
9945 	return (0);
9946 }
9947 
9948 static void
9949 sctp_timer_validation(struct sctp_inpcb *inp,
9950     struct sctp_tcb *stcb,
9951     struct sctp_association *asoc)
9952 {
9953 	struct sctp_nets *net;
9954 
9955 	/* Validate that a timer is running somewhere */
9956 	TAILQ_FOREACH(net, &asoc->nets, sctp_next) {
9957 		if (SCTP_OS_TIMER_PENDING(&net->rxt_timer.timer)) {
9958 			/* Here is a timer */
9959 			return;
9960 		}
9961 	}
9962 	SCTP_TCB_LOCK_ASSERT(stcb);
9963 	/* Gak, we did not have a timer somewhere */
9964 	SCTPDBG(SCTP_DEBUG_OUTPUT3, "Deadlock avoided starting timer on a dest at retran\n");
9965 	if (asoc->alternate) {
9966 		sctp_timer_start(SCTP_TIMER_TYPE_SEND, inp, stcb, asoc->alternate);
9967 	} else {
9968 		sctp_timer_start(SCTP_TIMER_TYPE_SEND, inp, stcb, asoc->primary_destination);
9969 	}
9970 	return;
9971 }
9972 
9973 void
9974 sctp_chunk_output(struct sctp_inpcb *inp,
9975     struct sctp_tcb *stcb,
9976     int from_where,
9977     int so_locked
9978 #if !defined(__APPLE__) && !defined(SCTP_SO_LOCK_TESTING)
9979     SCTP_UNUSED
9980 #endif
9981 )
9982 {
9983 	/*-
9984 	 * Ok this is the generic chunk service queue. we must do the
9985 	 * following:
9986 	 * - See if there are retransmits pending, if so we must
9987 	 *   do these first.
9988 	 * - Service the stream queue that is next, moving any
9989 	 *   message (note I must get a complete message i.e.
9990 	 *   FIRST/MIDDLE and LAST to the out queue in one pass) and assigning
9991 	 *   TSN's
9992 	 * - Check to see if the cwnd/rwnd allows any output, if so we
9993 	 *   go ahead and fomulate and send the low level chunks. Making sure
9994 	 *   to combine any control in the control chunk queue also.
9995 	 */
9996 	struct sctp_association *asoc;
9997 	struct sctp_nets *net;
9998 	int error = 0, num_out, tot_out = 0, ret = 0, reason_code;
9999 	unsigned int burst_cnt = 0;
10000 	struct timeval now;
10001 	int now_filled = 0;
10002 	int nagle_on;
10003 	int frag_point = sctp_get_frag_point(stcb, &stcb->asoc);
10004 	int un_sent = 0;
10005 	int fr_done;
10006 	unsigned int tot_frs = 0;
10007 
10008 	asoc = &stcb->asoc;
10009 do_it_again:
10010 	/* The Nagle algorithm is only applied when handling a send call. */
10011 	if (from_where == SCTP_OUTPUT_FROM_USR_SEND) {
10012 		if (sctp_is_feature_on(inp, SCTP_PCB_FLAGS_NODELAY)) {
10013 			nagle_on = 0;
10014 		} else {
10015 			nagle_on = 1;
10016 		}
10017 	} else {
10018 		nagle_on = 0;
10019 	}
10020 	SCTP_TCB_LOCK_ASSERT(stcb);
10021 
10022 	un_sent = (stcb->asoc.total_output_queue_size - stcb->asoc.total_flight);
10023 
10024 	if ((un_sent <= 0) &&
10025 	    (TAILQ_EMPTY(&asoc->control_send_queue)) &&
10026 	    (TAILQ_EMPTY(&asoc->asconf_send_queue)) &&
10027 	    (asoc->sent_queue_retran_cnt == 0) &&
10028 	    (asoc->trigger_reset == 0)) {
10029 		/* Nothing to do unless there is something to be sent left */
10030 		return;
10031 	}
10032 	/*
10033 	 * Do we have something to send, data or control AND a sack timer
10034 	 * running, if so piggy-back the sack.
10035 	 */
10036 	if (SCTP_OS_TIMER_PENDING(&stcb->asoc.dack_timer.timer)) {
10037 		sctp_send_sack(stcb, so_locked);
10038 		(void)SCTP_OS_TIMER_STOP(&stcb->asoc.dack_timer.timer);
10039 	}
10040 	while (asoc->sent_queue_retran_cnt) {
10041 		/*-
10042 		 * Ok, it is retransmission time only, we send out only ONE
10043 		 * packet with a single call off to the retran code.
10044 		 */
10045 		if (from_where == SCTP_OUTPUT_FROM_COOKIE_ACK) {
10046 			/*-
10047 			 * Special hook for handling cookiess discarded
10048 			 * by peer that carried data. Send cookie-ack only
10049 			 * and then the next call with get the retran's.
10050 			 */
10051 			(void)sctp_med_chunk_output(inp, stcb, asoc, &num_out, &reason_code, 1,
10052 			    from_where,
10053 			    &now, &now_filled, frag_point, so_locked);
10054 			return;
10055 		} else if (from_where != SCTP_OUTPUT_FROM_HB_TMR) {
10056 			/* if its not from a HB then do it */
10057 			fr_done = 0;
10058 			ret = sctp_chunk_retransmission(inp, stcb, asoc, &num_out, &now, &now_filled, &fr_done, so_locked);
10059 			if (fr_done) {
10060 				tot_frs++;
10061 			}
10062 		} else {
10063 			/*
10064 			 * its from any other place, we don't allow retran
10065 			 * output (only control)
10066 			 */
10067 			ret = 1;
10068 		}
10069 		if (ret > 0) {
10070 			/* Can't send anymore */
10071 			/*-
10072 			 * now lets push out control by calling med-level
10073 			 * output once. this assures that we WILL send HB's
10074 			 * if queued too.
10075 			 */
10076 			(void)sctp_med_chunk_output(inp, stcb, asoc, &num_out, &reason_code, 1,
10077 			    from_where,
10078 			    &now, &now_filled, frag_point, so_locked);
10079 #ifdef SCTP_AUDITING_ENABLED
10080 			sctp_auditing(8, inp, stcb, NULL);
10081 #endif
10082 			sctp_timer_validation(inp, stcb, asoc);
10083 			return;
10084 		}
10085 		if (ret < 0) {
10086 			/*-
10087 			 * The count was off.. retran is not happening so do
10088 			 * the normal retransmission.
10089 			 */
10090 #ifdef SCTP_AUDITING_ENABLED
10091 			sctp_auditing(9, inp, stcb, NULL);
10092 #endif
10093 			if (ret == SCTP_RETRAN_EXIT) {
10094 				return;
10095 			}
10096 			break;
10097 		}
10098 		if (from_where == SCTP_OUTPUT_FROM_T3) {
10099 			/* Only one transmission allowed out of a timeout */
10100 #ifdef SCTP_AUDITING_ENABLED
10101 			sctp_auditing(10, inp, stcb, NULL);
10102 #endif
10103 			/* Push out any control */
10104 			(void)sctp_med_chunk_output(inp, stcb, asoc, &num_out, &reason_code, 1, from_where,
10105 			    &now, &now_filled, frag_point, so_locked);
10106 			return;
10107 		}
10108 		if ((asoc->fr_max_burst > 0) && (tot_frs >= asoc->fr_max_burst)) {
10109 			/* Hit FR burst limit */
10110 			return;
10111 		}
10112 		if ((num_out == 0) && (ret == 0)) {
10113 			/* No more retrans to send */
10114 			break;
10115 		}
10116 	}
10117 #ifdef SCTP_AUDITING_ENABLED
10118 	sctp_auditing(12, inp, stcb, NULL);
10119 #endif
10120 	/* Check for bad destinations, if they exist move chunks around. */
10121 	TAILQ_FOREACH(net, &asoc->nets, sctp_next) {
10122 		if (!(net->dest_state & SCTP_ADDR_REACHABLE)) {
10123 			/*-
10124 			 * if possible move things off of this address we
10125 			 * still may send below due to the dormant state but
10126 			 * we try to find an alternate address to send to
10127 			 * and if we have one we move all queued data on the
10128 			 * out wheel to this alternate address.
10129 			 */
10130 			if (net->ref_count > 1)
10131 				sctp_move_chunks_from_net(stcb, net);
10132 		} else {
10133 			/*-
10134 			 * if ((asoc->sat_network) || (net->addr_is_local))
10135 			 * { burst_limit = asoc->max_burst *
10136 			 * SCTP_SAT_NETWORK_BURST_INCR; }
10137 			 */
10138 			if (asoc->max_burst > 0) {
10139 				if (SCTP_BASE_SYSCTL(sctp_use_cwnd_based_maxburst)) {
10140 					if ((net->flight_size + (asoc->max_burst * net->mtu)) < net->cwnd) {
10141 						/*
10142 						 * JRS - Use the congestion
10143 						 * control given in the
10144 						 * congestion control module
10145 						 */
10146 						asoc->cc_functions.sctp_cwnd_update_after_output(stcb, net, asoc->max_burst);
10147 						if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOG_MAXBURST_ENABLE) {
10148 							sctp_log_maxburst(stcb, net, 0, asoc->max_burst, SCTP_MAX_BURST_APPLIED);
10149 						}
10150 						SCTP_STAT_INCR(sctps_maxburstqueued);
10151 					}
10152 					net->fast_retran_ip = 0;
10153 				} else {
10154 					if (net->flight_size == 0) {
10155 						/*
10156 						 * Should be decaying the
10157 						 * cwnd here
10158 						 */
10159 						;
10160 					}
10161 				}
10162 			}
10163 		}
10164 
10165 	}
10166 	burst_cnt = 0;
10167 	do {
10168 		error = sctp_med_chunk_output(inp, stcb, asoc, &num_out,
10169 		    &reason_code, 0, from_where,
10170 		    &now, &now_filled, frag_point, so_locked);
10171 		if (error) {
10172 			SCTPDBG(SCTP_DEBUG_OUTPUT1, "Error %d was returned from med-c-op\n", error);
10173 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOG_MAXBURST_ENABLE) {
10174 				sctp_log_maxburst(stcb, asoc->primary_destination, error, burst_cnt, SCTP_MAX_BURST_ERROR_STOP);
10175 			}
10176 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_CWND_LOGGING_ENABLE) {
10177 				sctp_log_cwnd(stcb, NULL, error, SCTP_SEND_NOW_COMPLETES);
10178 				sctp_log_cwnd(stcb, NULL, 0xdeadbeef, SCTP_SEND_NOW_COMPLETES);
10179 			}
10180 			break;
10181 		}
10182 		SCTPDBG(SCTP_DEBUG_OUTPUT3, "m-c-o put out %d\n", num_out);
10183 
10184 		tot_out += num_out;
10185 		burst_cnt++;
10186 		if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_CWND_LOGGING_ENABLE) {
10187 			sctp_log_cwnd(stcb, NULL, num_out, SCTP_SEND_NOW_COMPLETES);
10188 			if (num_out == 0) {
10189 				sctp_log_cwnd(stcb, NULL, reason_code, SCTP_SEND_NOW_COMPLETES);
10190 			}
10191 		}
10192 		if (nagle_on) {
10193 			/*
10194 			 * When the Nagle algorithm is used, look at how
10195 			 * much is unsent, then if its smaller than an MTU
10196 			 * and we have data in flight we stop, except if we
10197 			 * are handling a fragmented user message.
10198 			 */
10199 			un_sent = stcb->asoc.total_output_queue_size - stcb->asoc.total_flight;
10200 			if ((un_sent < (int)(stcb->asoc.smallest_mtu - SCTP_MIN_OVERHEAD)) &&
10201 			    (stcb->asoc.total_flight > 0)) {
10202 /*	&&		     sctp_is_feature_on(inp, SCTP_PCB_FLAGS_EXPLICIT_EOR))) {*/
10203 				break;
10204 			}
10205 		}
10206 		if (TAILQ_EMPTY(&asoc->control_send_queue) &&
10207 		    TAILQ_EMPTY(&asoc->send_queue) &&
10208 		    sctp_is_there_unsent_data(stcb, so_locked) == 0) {
10209 			/* Nothing left to send */
10210 			break;
10211 		}
10212 		if ((stcb->asoc.total_output_queue_size - stcb->asoc.total_flight) <= 0) {
10213 			/* Nothing left to send */
10214 			break;
10215 		}
10216 	} while (num_out &&
10217 	    ((asoc->max_burst == 0) ||
10218 	    SCTP_BASE_SYSCTL(sctp_use_cwnd_based_maxburst) ||
10219 	    (burst_cnt < asoc->max_burst)));
10220 
10221 	if (SCTP_BASE_SYSCTL(sctp_use_cwnd_based_maxburst) == 0) {
10222 		if ((asoc->max_burst > 0) && (burst_cnt >= asoc->max_burst)) {
10223 			SCTP_STAT_INCR(sctps_maxburstqueued);
10224 			asoc->burst_limit_applied = 1;
10225 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOG_MAXBURST_ENABLE) {
10226 				sctp_log_maxburst(stcb, asoc->primary_destination, 0, burst_cnt, SCTP_MAX_BURST_APPLIED);
10227 			}
10228 		} else {
10229 			asoc->burst_limit_applied = 0;
10230 		}
10231 	}
10232 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_CWND_LOGGING_ENABLE) {
10233 		sctp_log_cwnd(stcb, NULL, tot_out, SCTP_SEND_NOW_COMPLETES);
10234 	}
10235 	SCTPDBG(SCTP_DEBUG_OUTPUT1, "Ok, we have put out %d chunks\n",
10236 	    tot_out);
10237 
10238 	/*-
10239 	 * Now we need to clean up the control chunk chain if a ECNE is on
10240 	 * it. It must be marked as UNSENT again so next call will continue
10241 	 * to send it until such time that we get a CWR, to remove it.
10242 	 */
10243 	if (stcb->asoc.ecn_echo_cnt_onq)
10244 		sctp_fix_ecn_echo(asoc);
10245 
10246 	if (stcb->asoc.trigger_reset) {
10247 		if (sctp_send_stream_reset_out_if_possible(stcb, so_locked) == 0) {
10248 			goto do_it_again;
10249 		}
10250 	}
10251 	return;
10252 }
10253 
10254 
10255 int
10256 sctp_output(
10257     struct sctp_inpcb *inp,
10258     struct mbuf *m,
10259     struct sockaddr *addr,
10260     struct mbuf *control,
10261     struct thread *p,
10262     int flags)
10263 {
10264 	if (inp == NULL) {
10265 		SCTP_LTRACE_ERR_RET_PKT(m, inp, NULL, NULL, SCTP_FROM_SCTP_OUTPUT, EINVAL);
10266 		return (EINVAL);
10267 	}
10268 
10269 	if (inp->sctp_socket == NULL) {
10270 		SCTP_LTRACE_ERR_RET_PKT(m, inp, NULL, NULL, SCTP_FROM_SCTP_OUTPUT, EINVAL);
10271 		return (EINVAL);
10272 	}
10273 	return (sctp_sosend(inp->sctp_socket,
10274 	    addr,
10275 	    (struct uio *)NULL,
10276 	    m,
10277 	    control,
10278 	    flags, p
10279 	    ));
10280 }
10281 
10282 void
10283 send_forward_tsn(struct sctp_tcb *stcb,
10284     struct sctp_association *asoc)
10285 {
10286 	struct sctp_tmit_chunk *chk, *at, *tp1, *last;
10287 	struct sctp_forward_tsn_chunk *fwdtsn;
10288 	struct sctp_strseq *strseq;
10289 	struct sctp_strseq_mid *strseq_m;
10290 	uint32_t advance_peer_ack_point;
10291 	unsigned int cnt_of_space, i, ovh;
10292 	unsigned int space_needed;
10293 	unsigned int cnt_of_skipped = 0;
10294 
10295 	SCTP_TCB_LOCK_ASSERT(stcb);
10296 	TAILQ_FOREACH(chk, &asoc->control_send_queue, sctp_next) {
10297 		if (chk->rec.chunk_id.id == SCTP_FORWARD_CUM_TSN) {
10298 			/* mark it to unsent */
10299 			chk->sent = SCTP_DATAGRAM_UNSENT;
10300 			chk->snd_count = 0;
10301 			/* Do we correct its output location? */
10302 			if (chk->whoTo) {
10303 				sctp_free_remote_addr(chk->whoTo);
10304 				chk->whoTo = NULL;
10305 			}
10306 			goto sctp_fill_in_rest;
10307 		}
10308 	}
10309 	/* Ok if we reach here we must build one */
10310 	sctp_alloc_a_chunk(stcb, chk);
10311 	if (chk == NULL) {
10312 		return;
10313 	}
10314 	asoc->fwd_tsn_cnt++;
10315 	chk->copy_by_ref = 0;
10316 	/*
10317 	 * We don't do the old thing here since this is used not for on-wire
10318 	 * but to tell if we are sending a fwd-tsn by the stack during
10319 	 * output. And if its a IFORWARD or a FORWARD it is a fwd-tsn.
10320 	 */
10321 	chk->rec.chunk_id.id = SCTP_FORWARD_CUM_TSN;
10322 	chk->rec.chunk_id.can_take_data = 0;
10323 	chk->flags = 0;
10324 	chk->asoc = asoc;
10325 	chk->whoTo = NULL;
10326 	chk->data = sctp_get_mbuf_for_msg(MCLBYTES, 0, M_NOWAIT, 1, MT_DATA);
10327 	if (chk->data == NULL) {
10328 		sctp_free_a_chunk(stcb, chk, SCTP_SO_NOT_LOCKED);
10329 		return;
10330 	}
10331 	SCTP_BUF_RESV_UF(chk->data, SCTP_MIN_OVERHEAD);
10332 	chk->sent = SCTP_DATAGRAM_UNSENT;
10333 	chk->snd_count = 0;
10334 	TAILQ_INSERT_TAIL(&asoc->control_send_queue, chk, sctp_next);
10335 	asoc->ctrl_queue_cnt++;
10336 sctp_fill_in_rest:
10337 	/*-
10338 	 * Here we go through and fill out the part that deals with
10339 	 * stream/seq of the ones we skip.
10340 	 */
10341 	SCTP_BUF_LEN(chk->data) = 0;
10342 	TAILQ_FOREACH(at, &asoc->sent_queue, sctp_next) {
10343 		if ((at->sent != SCTP_FORWARD_TSN_SKIP) &&
10344 		    (at->sent != SCTP_DATAGRAM_NR_ACKED)) {
10345 			/* no more to look at */
10346 			break;
10347 		}
10348 		if (!asoc->idata_supported && (at->rec.data.rcv_flags & SCTP_DATA_UNORDERED)) {
10349 			/* We don't report these */
10350 			continue;
10351 		}
10352 		cnt_of_skipped++;
10353 	}
10354 	if (asoc->idata_supported) {
10355 		space_needed = (sizeof(struct sctp_forward_tsn_chunk) +
10356 		    (cnt_of_skipped * sizeof(struct sctp_strseq_mid)));
10357 	} else {
10358 		space_needed = (sizeof(struct sctp_forward_tsn_chunk) +
10359 		    (cnt_of_skipped * sizeof(struct sctp_strseq)));
10360 	}
10361 	cnt_of_space = (unsigned int)M_TRAILINGSPACE(chk->data);
10362 
10363 	if (stcb->sctp_ep->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) {
10364 		ovh = SCTP_MIN_OVERHEAD;
10365 	} else {
10366 		ovh = SCTP_MIN_V4_OVERHEAD;
10367 	}
10368 	if (cnt_of_space > (asoc->smallest_mtu - ovh)) {
10369 		/* trim to a mtu size */
10370 		cnt_of_space = asoc->smallest_mtu - ovh;
10371 	}
10372 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOG_TRY_ADVANCE) {
10373 		sctp_misc_ints(SCTP_FWD_TSN_CHECK,
10374 		    0xff, 0, cnt_of_skipped,
10375 		    asoc->advanced_peer_ack_point);
10376 	}
10377 	advance_peer_ack_point = asoc->advanced_peer_ack_point;
10378 	if (cnt_of_space < space_needed) {
10379 		/*-
10380 		 * ok we must trim down the chunk by lowering the
10381 		 * advance peer ack point.
10382 		 */
10383 		if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOG_TRY_ADVANCE) {
10384 			sctp_misc_ints(SCTP_FWD_TSN_CHECK,
10385 			    0xff, 0xff, cnt_of_space,
10386 			    space_needed);
10387 		}
10388 		cnt_of_skipped = cnt_of_space - sizeof(struct sctp_forward_tsn_chunk);
10389 		if (asoc->idata_supported) {
10390 			cnt_of_skipped /= sizeof(struct sctp_strseq_mid);
10391 		} else {
10392 			cnt_of_skipped /= sizeof(struct sctp_strseq);
10393 		}
10394 		/*-
10395 		 * Go through and find the TSN that will be the one
10396 		 * we report.
10397 		 */
10398 		at = TAILQ_FIRST(&asoc->sent_queue);
10399 		if (at != NULL) {
10400 			for (i = 0; i < cnt_of_skipped; i++) {
10401 				tp1 = TAILQ_NEXT(at, sctp_next);
10402 				if (tp1 == NULL) {
10403 					break;
10404 				}
10405 				at = tp1;
10406 			}
10407 		}
10408 		if (at && SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LOG_TRY_ADVANCE) {
10409 			sctp_misc_ints(SCTP_FWD_TSN_CHECK,
10410 			    0xff, cnt_of_skipped, at->rec.data.tsn,
10411 			    asoc->advanced_peer_ack_point);
10412 		}
10413 		last = at;
10414 		/*-
10415 		 * last now points to last one I can report, update
10416 		 * peer ack point
10417 		 */
10418 		if (last) {
10419 			advance_peer_ack_point = last->rec.data.tsn;
10420 		}
10421 		if (asoc->idata_supported) {
10422 			space_needed = sizeof(struct sctp_forward_tsn_chunk) +
10423 			    cnt_of_skipped * sizeof(struct sctp_strseq_mid);
10424 		} else {
10425 			space_needed = sizeof(struct sctp_forward_tsn_chunk) +
10426 			    cnt_of_skipped * sizeof(struct sctp_strseq);
10427 		}
10428 	}
10429 	chk->send_size = space_needed;
10430 	/* Setup the chunk */
10431 	fwdtsn = mtod(chk->data, struct sctp_forward_tsn_chunk *);
10432 	fwdtsn->ch.chunk_length = htons(chk->send_size);
10433 	fwdtsn->ch.chunk_flags = 0;
10434 	if (asoc->idata_supported) {
10435 		fwdtsn->ch.chunk_type = SCTP_IFORWARD_CUM_TSN;
10436 	} else {
10437 		fwdtsn->ch.chunk_type = SCTP_FORWARD_CUM_TSN;
10438 	}
10439 	fwdtsn->new_cumulative_tsn = htonl(advance_peer_ack_point);
10440 	SCTP_BUF_LEN(chk->data) = chk->send_size;
10441 	fwdtsn++;
10442 	/*-
10443 	 * Move pointer to after the fwdtsn and transfer to the
10444 	 * strseq pointer.
10445 	 */
10446 	if (asoc->idata_supported) {
10447 		strseq_m = (struct sctp_strseq_mid *)fwdtsn;
10448 		strseq = NULL;
10449 	} else {
10450 		strseq = (struct sctp_strseq *)fwdtsn;
10451 		strseq_m = NULL;
10452 	}
10453 	/*-
10454 	 * Now populate the strseq list. This is done blindly
10455 	 * without pulling out duplicate stream info. This is
10456 	 * inefficent but won't harm the process since the peer will
10457 	 * look at these in sequence and will thus release anything.
10458 	 * It could mean we exceed the PMTU and chop off some that
10459 	 * we could have included.. but this is unlikely (aka 1432/4
10460 	 * would mean 300+ stream seq's would have to be reported in
10461 	 * one FWD-TSN. With a bit of work we can later FIX this to
10462 	 * optimize and pull out duplicates.. but it does add more
10463 	 * overhead. So for now... not!
10464 	 */
10465 	i = 0;
10466 	TAILQ_FOREACH(at, &asoc->sent_queue, sctp_next) {
10467 		if (i >= cnt_of_skipped) {
10468 			break;
10469 		}
10470 		if (!asoc->idata_supported && (at->rec.data.rcv_flags & SCTP_DATA_UNORDERED)) {
10471 			/* We don't report these */
10472 			continue;
10473 		}
10474 		if (at->rec.data.tsn == advance_peer_ack_point) {
10475 			at->rec.data.fwd_tsn_cnt = 0;
10476 		}
10477 		if (asoc->idata_supported) {
10478 			strseq_m->sid = htons(at->rec.data.sid);
10479 			if (at->rec.data.rcv_flags & SCTP_DATA_UNORDERED) {
10480 				strseq_m->flags = htons(PR_SCTP_UNORDERED_FLAG);
10481 			} else {
10482 				strseq_m->flags = 0;
10483 			}
10484 			strseq_m->mid = htonl(at->rec.data.mid);
10485 			strseq_m++;
10486 		} else {
10487 			strseq->sid = htons(at->rec.data.sid);
10488 			strseq->ssn = htons((uint16_t)at->rec.data.mid);
10489 			strseq++;
10490 		}
10491 		i++;
10492 	}
10493 	return;
10494 }
10495 
10496 void
10497 sctp_send_sack(struct sctp_tcb *stcb, int so_locked
10498 #if !defined(__APPLE__) && !defined(SCTP_SO_LOCK_TESTING)
10499     SCTP_UNUSED
10500 #endif
10501 )
10502 {
10503 	/*-
10504 	 * Queue up a SACK or NR-SACK in the control queue.
10505 	 * We must first check to see if a SACK or NR-SACK is
10506 	 * somehow on the control queue.
10507 	 * If so, we will take and and remove the old one.
10508 	 */
10509 	struct sctp_association *asoc;
10510 	struct sctp_tmit_chunk *chk, *a_chk;
10511 	struct sctp_sack_chunk *sack;
10512 	struct sctp_nr_sack_chunk *nr_sack;
10513 	struct sctp_gap_ack_block *gap_descriptor;
10514 	const struct sack_track *selector;
10515 	int mergeable = 0;
10516 	int offset;
10517 	caddr_t limit;
10518 	uint32_t *dup;
10519 	int limit_reached = 0;
10520 	unsigned int i, siz, j;
10521 	unsigned int num_gap_blocks = 0, num_nr_gap_blocks = 0, space;
10522 	int num_dups = 0;
10523 	int space_req;
10524 	uint32_t highest_tsn;
10525 	uint8_t flags;
10526 	uint8_t type;
10527 	uint8_t tsn_map;
10528 
10529 	if (stcb->asoc.nrsack_supported == 1) {
10530 		type = SCTP_NR_SELECTIVE_ACK;
10531 	} else {
10532 		type = SCTP_SELECTIVE_ACK;
10533 	}
10534 	a_chk = NULL;
10535 	asoc = &stcb->asoc;
10536 	SCTP_TCB_LOCK_ASSERT(stcb);
10537 	if (asoc->last_data_chunk_from == NULL) {
10538 		/* Hmm we never received anything */
10539 		return;
10540 	}
10541 	sctp_slide_mapping_arrays(stcb);
10542 	sctp_set_rwnd(stcb, asoc);
10543 	TAILQ_FOREACH(chk, &asoc->control_send_queue, sctp_next) {
10544 		if (chk->rec.chunk_id.id == type) {
10545 			/* Hmm, found a sack already on queue, remove it */
10546 			TAILQ_REMOVE(&asoc->control_send_queue, chk, sctp_next);
10547 			asoc->ctrl_queue_cnt--;
10548 			a_chk = chk;
10549 			if (a_chk->data) {
10550 				sctp_m_freem(a_chk->data);
10551 				a_chk->data = NULL;
10552 			}
10553 			if (a_chk->whoTo) {
10554 				sctp_free_remote_addr(a_chk->whoTo);
10555 				a_chk->whoTo = NULL;
10556 			}
10557 			break;
10558 		}
10559 	}
10560 	if (a_chk == NULL) {
10561 		sctp_alloc_a_chunk(stcb, a_chk);
10562 		if (a_chk == NULL) {
10563 			/* No memory so we drop the idea, and set a timer */
10564 			if (stcb->asoc.delayed_ack) {
10565 				sctp_timer_stop(SCTP_TIMER_TYPE_RECV,
10566 				    stcb->sctp_ep, stcb, NULL,
10567 				    SCTP_FROM_SCTP_OUTPUT + SCTP_LOC_3);
10568 				sctp_timer_start(SCTP_TIMER_TYPE_RECV,
10569 				    stcb->sctp_ep, stcb, NULL);
10570 			} else {
10571 				stcb->asoc.send_sack = 1;
10572 			}
10573 			return;
10574 		}
10575 		a_chk->copy_by_ref = 0;
10576 		a_chk->rec.chunk_id.id = type;
10577 		a_chk->rec.chunk_id.can_take_data = 1;
10578 	}
10579 	/* Clear our pkt counts */
10580 	asoc->data_pkts_seen = 0;
10581 
10582 	a_chk->flags = 0;
10583 	a_chk->asoc = asoc;
10584 	a_chk->snd_count = 0;
10585 	a_chk->send_size = 0;	/* fill in later */
10586 	a_chk->sent = SCTP_DATAGRAM_UNSENT;
10587 	a_chk->whoTo = NULL;
10588 
10589 	if (!(asoc->last_data_chunk_from->dest_state & SCTP_ADDR_REACHABLE)) {
10590 		/*-
10591 		 * Ok, the destination for the SACK is unreachable, lets see if
10592 		 * we can select an alternate to asoc->last_data_chunk_from
10593 		 */
10594 		a_chk->whoTo = sctp_find_alternate_net(stcb, asoc->last_data_chunk_from, 0);
10595 		if (a_chk->whoTo == NULL) {
10596 			/* Nope, no alternate */
10597 			a_chk->whoTo = asoc->last_data_chunk_from;
10598 		}
10599 	} else {
10600 		a_chk->whoTo = asoc->last_data_chunk_from;
10601 	}
10602 	if (a_chk->whoTo) {
10603 		atomic_add_int(&a_chk->whoTo->ref_count, 1);
10604 	}
10605 	if (SCTP_TSN_GT(asoc->highest_tsn_inside_map, asoc->highest_tsn_inside_nr_map)) {
10606 		highest_tsn = asoc->highest_tsn_inside_map;
10607 	} else {
10608 		highest_tsn = asoc->highest_tsn_inside_nr_map;
10609 	}
10610 	if (highest_tsn == asoc->cumulative_tsn) {
10611 		/* no gaps */
10612 		if (type == SCTP_SELECTIVE_ACK) {
10613 			space_req = sizeof(struct sctp_sack_chunk);
10614 		} else {
10615 			space_req = sizeof(struct sctp_nr_sack_chunk);
10616 		}
10617 	} else {
10618 		/* gaps get a cluster */
10619 		space_req = MCLBYTES;
10620 	}
10621 	/* Ok now lets formulate a MBUF with our sack */
10622 	a_chk->data = sctp_get_mbuf_for_msg(space_req, 0, M_NOWAIT, 1, MT_DATA);
10623 	if ((a_chk->data == NULL) ||
10624 	    (a_chk->whoTo == NULL)) {
10625 		/* rats, no mbuf memory */
10626 		if (a_chk->data) {
10627 			/* was a problem with the destination */
10628 			sctp_m_freem(a_chk->data);
10629 			a_chk->data = NULL;
10630 		}
10631 		sctp_free_a_chunk(stcb, a_chk, so_locked);
10632 		/* sa_ignore NO_NULL_CHK */
10633 		if (stcb->asoc.delayed_ack) {
10634 			sctp_timer_stop(SCTP_TIMER_TYPE_RECV,
10635 			    stcb->sctp_ep, stcb, NULL,
10636 			    SCTP_FROM_SCTP_OUTPUT + SCTP_LOC_4);
10637 			sctp_timer_start(SCTP_TIMER_TYPE_RECV,
10638 			    stcb->sctp_ep, stcb, NULL);
10639 		} else {
10640 			stcb->asoc.send_sack = 1;
10641 		}
10642 		return;
10643 	}
10644 	/* ok, lets go through and fill it in */
10645 	SCTP_BUF_RESV_UF(a_chk->data, SCTP_MIN_OVERHEAD);
10646 	space = (unsigned int)M_TRAILINGSPACE(a_chk->data);
10647 	if (space > (a_chk->whoTo->mtu - SCTP_MIN_OVERHEAD)) {
10648 		space = (a_chk->whoTo->mtu - SCTP_MIN_OVERHEAD);
10649 	}
10650 	limit = mtod(a_chk->data, caddr_t);
10651 	limit += space;
10652 
10653 	flags = 0;
10654 
10655 	if ((asoc->sctp_cmt_on_off > 0) &&
10656 	    SCTP_BASE_SYSCTL(sctp_cmt_use_dac)) {
10657 		/*-
10658 		 * CMT DAC algorithm: If 2 (i.e., 0x10) packets have been
10659 		 * received, then set high bit to 1, else 0. Reset
10660 		 * pkts_rcvd.
10661 		 */
10662 		flags |= (asoc->cmt_dac_pkts_rcvd << 6);
10663 		asoc->cmt_dac_pkts_rcvd = 0;
10664 	}
10665 #ifdef SCTP_ASOCLOG_OF_TSNS
10666 	stcb->asoc.cumack_logsnt[stcb->asoc.cumack_log_atsnt] = asoc->cumulative_tsn;
10667 	stcb->asoc.cumack_log_atsnt++;
10668 	if (stcb->asoc.cumack_log_atsnt >= SCTP_TSN_LOG_SIZE) {
10669 		stcb->asoc.cumack_log_atsnt = 0;
10670 	}
10671 #endif
10672 	/* reset the readers interpretation */
10673 	stcb->freed_by_sorcv_sincelast = 0;
10674 
10675 	if (type == SCTP_SELECTIVE_ACK) {
10676 		sack = mtod(a_chk->data, struct sctp_sack_chunk *);
10677 		nr_sack = NULL;
10678 		gap_descriptor = (struct sctp_gap_ack_block *)((caddr_t)sack + sizeof(struct sctp_sack_chunk));
10679 		if (highest_tsn > asoc->mapping_array_base_tsn) {
10680 			siz = (((highest_tsn - asoc->mapping_array_base_tsn) + 1) + 7) / 8;
10681 		} else {
10682 			siz = (((MAX_TSN - highest_tsn) + 1) + highest_tsn + 7) / 8;
10683 		}
10684 	} else {
10685 		sack = NULL;
10686 		nr_sack = mtod(a_chk->data, struct sctp_nr_sack_chunk *);
10687 		gap_descriptor = (struct sctp_gap_ack_block *)((caddr_t)nr_sack + sizeof(struct sctp_nr_sack_chunk));
10688 		if (asoc->highest_tsn_inside_map > asoc->mapping_array_base_tsn) {
10689 			siz = (((asoc->highest_tsn_inside_map - asoc->mapping_array_base_tsn) + 1) + 7) / 8;
10690 		} else {
10691 			siz = (((MAX_TSN - asoc->mapping_array_base_tsn) + 1) + asoc->highest_tsn_inside_map + 7) / 8;
10692 		}
10693 	}
10694 
10695 	if (SCTP_TSN_GT(asoc->mapping_array_base_tsn, asoc->cumulative_tsn)) {
10696 		offset = 1;
10697 	} else {
10698 		offset = asoc->mapping_array_base_tsn - asoc->cumulative_tsn;
10699 	}
10700 	if (((type == SCTP_SELECTIVE_ACK) &&
10701 	    SCTP_TSN_GT(highest_tsn, asoc->cumulative_tsn)) ||
10702 	    ((type == SCTP_NR_SELECTIVE_ACK) &&
10703 	    SCTP_TSN_GT(asoc->highest_tsn_inside_map, asoc->cumulative_tsn))) {
10704 		/* we have a gap .. maybe */
10705 		for (i = 0; i < siz; i++) {
10706 			tsn_map = asoc->mapping_array[i];
10707 			if (type == SCTP_SELECTIVE_ACK) {
10708 				tsn_map |= asoc->nr_mapping_array[i];
10709 			}
10710 			if (i == 0) {
10711 				/*
10712 				 * Clear all bits corresponding to TSNs
10713 				 * smaller or equal to the cumulative TSN.
10714 				 */
10715 				tsn_map &= (~0U << (1 - offset));
10716 			}
10717 			selector = &sack_array[tsn_map];
10718 			if (mergeable && selector->right_edge) {
10719 				/*
10720 				 * Backup, left and right edges were ok to
10721 				 * merge.
10722 				 */
10723 				num_gap_blocks--;
10724 				gap_descriptor--;
10725 			}
10726 			if (selector->num_entries == 0)
10727 				mergeable = 0;
10728 			else {
10729 				for (j = 0; j < selector->num_entries; j++) {
10730 					if (mergeable && selector->right_edge) {
10731 						/*
10732 						 * do a merge by NOT setting
10733 						 * the left side
10734 						 */
10735 						mergeable = 0;
10736 					} else {
10737 						/*
10738 						 * no merge, set the left
10739 						 * side
10740 						 */
10741 						mergeable = 0;
10742 						gap_descriptor->start = htons((selector->gaps[j].start + offset));
10743 					}
10744 					gap_descriptor->end = htons((selector->gaps[j].end + offset));
10745 					num_gap_blocks++;
10746 					gap_descriptor++;
10747 					if (((caddr_t)gap_descriptor + sizeof(struct sctp_gap_ack_block)) > limit) {
10748 						/* no more room */
10749 						limit_reached = 1;
10750 						break;
10751 					}
10752 				}
10753 				if (selector->left_edge) {
10754 					mergeable = 1;
10755 				}
10756 			}
10757 			if (limit_reached) {
10758 				/* Reached the limit stop */
10759 				break;
10760 			}
10761 			offset += 8;
10762 		}
10763 	}
10764 	if ((type == SCTP_NR_SELECTIVE_ACK) &&
10765 	    (limit_reached == 0)) {
10766 
10767 		mergeable = 0;
10768 
10769 		if (asoc->highest_tsn_inside_nr_map > asoc->mapping_array_base_tsn) {
10770 			siz = (((asoc->highest_tsn_inside_nr_map - asoc->mapping_array_base_tsn) + 1) + 7) / 8;
10771 		} else {
10772 			siz = (((MAX_TSN - asoc->mapping_array_base_tsn) + 1) + asoc->highest_tsn_inside_nr_map + 7) / 8;
10773 		}
10774 
10775 		if (SCTP_TSN_GT(asoc->mapping_array_base_tsn, asoc->cumulative_tsn)) {
10776 			offset = 1;
10777 		} else {
10778 			offset = asoc->mapping_array_base_tsn - asoc->cumulative_tsn;
10779 		}
10780 		if (SCTP_TSN_GT(asoc->highest_tsn_inside_nr_map, asoc->cumulative_tsn)) {
10781 			/* we have a gap .. maybe */
10782 			for (i = 0; i < siz; i++) {
10783 				tsn_map = asoc->nr_mapping_array[i];
10784 				if (i == 0) {
10785 					/*
10786 					 * Clear all bits corresponding to
10787 					 * TSNs smaller or equal to the
10788 					 * cumulative TSN.
10789 					 */
10790 					tsn_map &= (~0U << (1 - offset));
10791 				}
10792 				selector = &sack_array[tsn_map];
10793 				if (mergeable && selector->right_edge) {
10794 					/*
10795 					 * Backup, left and right edges were
10796 					 * ok to merge.
10797 					 */
10798 					num_nr_gap_blocks--;
10799 					gap_descriptor--;
10800 				}
10801 				if (selector->num_entries == 0)
10802 					mergeable = 0;
10803 				else {
10804 					for (j = 0; j < selector->num_entries; j++) {
10805 						if (mergeable && selector->right_edge) {
10806 							/*
10807 							 * do a merge by NOT
10808 							 * setting the left
10809 							 * side
10810 							 */
10811 							mergeable = 0;
10812 						} else {
10813 							/*
10814 							 * no merge, set the
10815 							 * left side
10816 							 */
10817 							mergeable = 0;
10818 							gap_descriptor->start = htons((selector->gaps[j].start + offset));
10819 						}
10820 						gap_descriptor->end = htons((selector->gaps[j].end + offset));
10821 						num_nr_gap_blocks++;
10822 						gap_descriptor++;
10823 						if (((caddr_t)gap_descriptor + sizeof(struct sctp_gap_ack_block)) > limit) {
10824 							/* no more room */
10825 							limit_reached = 1;
10826 							break;
10827 						}
10828 					}
10829 					if (selector->left_edge) {
10830 						mergeable = 1;
10831 					}
10832 				}
10833 				if (limit_reached) {
10834 					/* Reached the limit stop */
10835 					break;
10836 				}
10837 				offset += 8;
10838 			}
10839 		}
10840 	}
10841 	/* now we must add any dups we are going to report. */
10842 	if ((limit_reached == 0) && (asoc->numduptsns)) {
10843 		dup = (uint32_t *)gap_descriptor;
10844 		for (i = 0; i < asoc->numduptsns; i++) {
10845 			*dup = htonl(asoc->dup_tsns[i]);
10846 			dup++;
10847 			num_dups++;
10848 			if (((caddr_t)dup + sizeof(uint32_t)) > limit) {
10849 				/* no more room */
10850 				break;
10851 			}
10852 		}
10853 		asoc->numduptsns = 0;
10854 	}
10855 	/*
10856 	 * now that the chunk is prepared queue it to the control chunk
10857 	 * queue.
10858 	 */
10859 	if (type == SCTP_SELECTIVE_ACK) {
10860 		a_chk->send_size = (uint16_t)(sizeof(struct sctp_sack_chunk) +
10861 		    (num_gap_blocks + num_nr_gap_blocks) * sizeof(struct sctp_gap_ack_block) +
10862 		    num_dups * sizeof(int32_t));
10863 		SCTP_BUF_LEN(a_chk->data) = a_chk->send_size;
10864 		sack->sack.cum_tsn_ack = htonl(asoc->cumulative_tsn);
10865 		sack->sack.a_rwnd = htonl(asoc->my_rwnd);
10866 		sack->sack.num_gap_ack_blks = htons(num_gap_blocks);
10867 		sack->sack.num_dup_tsns = htons(num_dups);
10868 		sack->ch.chunk_type = type;
10869 		sack->ch.chunk_flags = flags;
10870 		sack->ch.chunk_length = htons(a_chk->send_size);
10871 	} else {
10872 		a_chk->send_size = (uint16_t)(sizeof(struct sctp_nr_sack_chunk) +
10873 		    (num_gap_blocks + num_nr_gap_blocks) * sizeof(struct sctp_gap_ack_block) +
10874 		    num_dups * sizeof(int32_t));
10875 		SCTP_BUF_LEN(a_chk->data) = a_chk->send_size;
10876 		nr_sack->nr_sack.cum_tsn_ack = htonl(asoc->cumulative_tsn);
10877 		nr_sack->nr_sack.a_rwnd = htonl(asoc->my_rwnd);
10878 		nr_sack->nr_sack.num_gap_ack_blks = htons(num_gap_blocks);
10879 		nr_sack->nr_sack.num_nr_gap_ack_blks = htons(num_nr_gap_blocks);
10880 		nr_sack->nr_sack.num_dup_tsns = htons(num_dups);
10881 		nr_sack->nr_sack.reserved = 0;
10882 		nr_sack->ch.chunk_type = type;
10883 		nr_sack->ch.chunk_flags = flags;
10884 		nr_sack->ch.chunk_length = htons(a_chk->send_size);
10885 	}
10886 	TAILQ_INSERT_TAIL(&asoc->control_send_queue, a_chk, sctp_next);
10887 	asoc->my_last_reported_rwnd = asoc->my_rwnd;
10888 	asoc->ctrl_queue_cnt++;
10889 	asoc->send_sack = 0;
10890 	SCTP_STAT_INCR(sctps_sendsacks);
10891 	return;
10892 }
10893 
10894 void
10895 sctp_send_abort_tcb(struct sctp_tcb *stcb, struct mbuf *operr, int so_locked
10896 #if !defined(__APPLE__) && !defined(SCTP_SO_LOCK_TESTING)
10897     SCTP_UNUSED
10898 #endif
10899 )
10900 {
10901 	struct mbuf *m_abort, *m, *m_last;
10902 	struct mbuf *m_out, *m_end = NULL;
10903 	struct sctp_abort_chunk *abort;
10904 	struct sctp_auth_chunk *auth = NULL;
10905 	struct sctp_nets *net;
10906 	uint32_t vtag;
10907 	uint32_t auth_offset = 0;
10908 	int error;
10909 	uint16_t cause_len, chunk_len, padding_len;
10910 
10911 	SCTP_TCB_LOCK_ASSERT(stcb);
10912 	/*-
10913 	 * Add an AUTH chunk, if chunk requires it and save the offset into
10914 	 * the chain for AUTH
10915 	 */
10916 	if (sctp_auth_is_required_chunk(SCTP_ABORT_ASSOCIATION,
10917 	    stcb->asoc.peer_auth_chunks)) {
10918 		m_out = sctp_add_auth_chunk(NULL, &m_end, &auth, &auth_offset,
10919 		    stcb, SCTP_ABORT_ASSOCIATION);
10920 		SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
10921 	} else {
10922 		m_out = NULL;
10923 	}
10924 	m_abort = sctp_get_mbuf_for_msg(sizeof(struct sctp_abort_chunk), 0, M_NOWAIT, 1, MT_HEADER);
10925 	if (m_abort == NULL) {
10926 		if (m_out) {
10927 			sctp_m_freem(m_out);
10928 		}
10929 		if (operr) {
10930 			sctp_m_freem(operr);
10931 		}
10932 		return;
10933 	}
10934 	/* link in any error */
10935 	SCTP_BUF_NEXT(m_abort) = operr;
10936 	cause_len = 0;
10937 	m_last = NULL;
10938 	for (m = operr; m; m = SCTP_BUF_NEXT(m)) {
10939 		cause_len += (uint16_t)SCTP_BUF_LEN(m);
10940 		if (SCTP_BUF_NEXT(m) == NULL) {
10941 			m_last = m;
10942 		}
10943 	}
10944 	SCTP_BUF_LEN(m_abort) = sizeof(struct sctp_abort_chunk);
10945 	chunk_len = (uint16_t)sizeof(struct sctp_abort_chunk) + cause_len;
10946 	padding_len = SCTP_SIZE32(chunk_len) - chunk_len;
10947 	if (m_out == NULL) {
10948 		/* NO Auth chunk prepended, so reserve space in front */
10949 		SCTP_BUF_RESV_UF(m_abort, SCTP_MIN_OVERHEAD);
10950 		m_out = m_abort;
10951 	} else {
10952 		/* Put AUTH chunk at the front of the chain */
10953 		SCTP_BUF_NEXT(m_end) = m_abort;
10954 	}
10955 	if (stcb->asoc.alternate) {
10956 		net = stcb->asoc.alternate;
10957 	} else {
10958 		net = stcb->asoc.primary_destination;
10959 	}
10960 	/* Fill in the ABORT chunk header. */
10961 	abort = mtod(m_abort, struct sctp_abort_chunk *);
10962 	abort->ch.chunk_type = SCTP_ABORT_ASSOCIATION;
10963 	if (stcb->asoc.peer_vtag == 0) {
10964 		/* This happens iff the assoc is in COOKIE-WAIT state. */
10965 		vtag = stcb->asoc.my_vtag;
10966 		abort->ch.chunk_flags = SCTP_HAD_NO_TCB;
10967 	} else {
10968 		vtag = stcb->asoc.peer_vtag;
10969 		abort->ch.chunk_flags = 0;
10970 	}
10971 	abort->ch.chunk_length = htons(chunk_len);
10972 	/* Add padding, if necessary. */
10973 	if (padding_len > 0) {
10974 		if ((m_last == NULL) ||
10975 		    (sctp_add_pad_tombuf(m_last, padding_len) == NULL)) {
10976 			sctp_m_freem(m_out);
10977 			return;
10978 		}
10979 	}
10980 	if ((error = sctp_lowlevel_chunk_output(stcb->sctp_ep, stcb, net,
10981 	    (struct sockaddr *)&net->ro._l_addr,
10982 	    m_out, auth_offset, auth, stcb->asoc.authinfo.active_keyid, 1, 0, 0,
10983 	    stcb->sctp_ep->sctp_lport, stcb->rport, htonl(vtag),
10984 	    stcb->asoc.primary_destination->port, NULL,
10985 	    0, 0,
10986 	    so_locked))) {
10987 		SCTPDBG(SCTP_DEBUG_OUTPUT3, "Gak send error %d\n", error);
10988 		if (error == ENOBUFS) {
10989 			stcb->asoc.ifp_had_enobuf = 1;
10990 			SCTP_STAT_INCR(sctps_lowlevelerr);
10991 		}
10992 	} else {
10993 		stcb->asoc.ifp_had_enobuf = 0;
10994 	}
10995 	SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
10996 }
10997 
10998 void
10999 sctp_send_shutdown_complete(struct sctp_tcb *stcb,
11000     struct sctp_nets *net,
11001     int reflect_vtag)
11002 {
11003 	/* formulate and SEND a SHUTDOWN-COMPLETE */
11004 	struct mbuf *m_shutdown_comp;
11005 	struct sctp_shutdown_complete_chunk *shutdown_complete;
11006 	uint32_t vtag;
11007 	int error;
11008 	uint8_t flags;
11009 
11010 	m_shutdown_comp = sctp_get_mbuf_for_msg(sizeof(struct sctp_chunkhdr), 0, M_NOWAIT, 1, MT_HEADER);
11011 	if (m_shutdown_comp == NULL) {
11012 		/* no mbuf's */
11013 		return;
11014 	}
11015 	if (reflect_vtag) {
11016 		flags = SCTP_HAD_NO_TCB;
11017 		vtag = stcb->asoc.my_vtag;
11018 	} else {
11019 		flags = 0;
11020 		vtag = stcb->asoc.peer_vtag;
11021 	}
11022 	shutdown_complete = mtod(m_shutdown_comp, struct sctp_shutdown_complete_chunk *);
11023 	shutdown_complete->ch.chunk_type = SCTP_SHUTDOWN_COMPLETE;
11024 	shutdown_complete->ch.chunk_flags = flags;
11025 	shutdown_complete->ch.chunk_length = htons(sizeof(struct sctp_shutdown_complete_chunk));
11026 	SCTP_BUF_LEN(m_shutdown_comp) = sizeof(struct sctp_shutdown_complete_chunk);
11027 	if ((error = sctp_lowlevel_chunk_output(stcb->sctp_ep, stcb, net,
11028 	    (struct sockaddr *)&net->ro._l_addr,
11029 	    m_shutdown_comp, 0, NULL, 0, 1, 0, 0,
11030 	    stcb->sctp_ep->sctp_lport, stcb->rport,
11031 	    htonl(vtag),
11032 	    net->port, NULL,
11033 	    0, 0,
11034 	    SCTP_SO_NOT_LOCKED))) {
11035 		SCTPDBG(SCTP_DEBUG_OUTPUT3, "Gak send error %d\n", error);
11036 		if (error == ENOBUFS) {
11037 			stcb->asoc.ifp_had_enobuf = 1;
11038 			SCTP_STAT_INCR(sctps_lowlevelerr);
11039 		}
11040 	} else {
11041 		stcb->asoc.ifp_had_enobuf = 0;
11042 	}
11043 	SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
11044 	return;
11045 }
11046 
11047 static void
11048 sctp_send_resp_msg(struct sockaddr *src, struct sockaddr *dst,
11049     struct sctphdr *sh, uint32_t vtag,
11050     uint8_t type, struct mbuf *cause,
11051     uint8_t mflowtype, uint32_t mflowid, uint16_t fibnum,
11052     uint32_t vrf_id, uint16_t port)
11053 {
11054 	struct mbuf *o_pak;
11055 	struct mbuf *mout;
11056 	struct sctphdr *shout;
11057 	struct sctp_chunkhdr *ch;
11058 #if defined(INET) || defined(INET6)
11059 	struct udphdr *udp;
11060 #endif
11061 	int ret, len, cause_len, padding_len;
11062 #ifdef INET
11063 	struct sockaddr_in *src_sin, *dst_sin;
11064 	struct ip *ip;
11065 #endif
11066 #ifdef INET6
11067 	struct sockaddr_in6 *src_sin6, *dst_sin6;
11068 	struct ip6_hdr *ip6;
11069 #endif
11070 
11071 	/* Compute the length of the cause and add final padding. */
11072 	cause_len = 0;
11073 	if (cause != NULL) {
11074 		struct mbuf *m_at, *m_last = NULL;
11075 
11076 		for (m_at = cause; m_at; m_at = SCTP_BUF_NEXT(m_at)) {
11077 			if (SCTP_BUF_NEXT(m_at) == NULL)
11078 				m_last = m_at;
11079 			cause_len += SCTP_BUF_LEN(m_at);
11080 		}
11081 		padding_len = cause_len % 4;
11082 		if (padding_len != 0) {
11083 			padding_len = 4 - padding_len;
11084 		}
11085 		if (padding_len != 0) {
11086 			if (sctp_add_pad_tombuf(m_last, padding_len) == NULL) {
11087 				sctp_m_freem(cause);
11088 				return;
11089 			}
11090 		}
11091 	} else {
11092 		padding_len = 0;
11093 	}
11094 	/* Get an mbuf for the header. */
11095 	len = sizeof(struct sctphdr) + sizeof(struct sctp_chunkhdr);
11096 	switch (dst->sa_family) {
11097 #ifdef INET
11098 	case AF_INET:
11099 		len += sizeof(struct ip);
11100 		break;
11101 #endif
11102 #ifdef INET6
11103 	case AF_INET6:
11104 		len += sizeof(struct ip6_hdr);
11105 		break;
11106 #endif
11107 	default:
11108 		break;
11109 	}
11110 #if defined(INET) || defined(INET6)
11111 	if (port) {
11112 		len += sizeof(struct udphdr);
11113 	}
11114 #endif
11115 	mout = sctp_get_mbuf_for_msg(len + max_linkhdr, 1, M_NOWAIT, 1, MT_DATA);
11116 	if (mout == NULL) {
11117 		if (cause) {
11118 			sctp_m_freem(cause);
11119 		}
11120 		return;
11121 	}
11122 	SCTP_BUF_RESV_UF(mout, max_linkhdr);
11123 	SCTP_BUF_LEN(mout) = len;
11124 	SCTP_BUF_NEXT(mout) = cause;
11125 	M_SETFIB(mout, fibnum);
11126 	mout->m_pkthdr.flowid = mflowid;
11127 	M_HASHTYPE_SET(mout, mflowtype);
11128 #ifdef INET
11129 	ip = NULL;
11130 #endif
11131 #ifdef INET6
11132 	ip6 = NULL;
11133 #endif
11134 	switch (dst->sa_family) {
11135 #ifdef INET
11136 	case AF_INET:
11137 		src_sin = (struct sockaddr_in *)src;
11138 		dst_sin = (struct sockaddr_in *)dst;
11139 		ip = mtod(mout, struct ip *);
11140 		ip->ip_v = IPVERSION;
11141 		ip->ip_hl = (sizeof(struct ip) >> 2);
11142 		ip->ip_tos = 0;
11143 		ip->ip_off = htons(IP_DF);
11144 		ip_fillid(ip);
11145 		ip->ip_ttl = MODULE_GLOBAL(ip_defttl);
11146 		if (port) {
11147 			ip->ip_p = IPPROTO_UDP;
11148 		} else {
11149 			ip->ip_p = IPPROTO_SCTP;
11150 		}
11151 		ip->ip_src.s_addr = dst_sin->sin_addr.s_addr;
11152 		ip->ip_dst.s_addr = src_sin->sin_addr.s_addr;
11153 		ip->ip_sum = 0;
11154 		len = sizeof(struct ip);
11155 		shout = (struct sctphdr *)((caddr_t)ip + len);
11156 		break;
11157 #endif
11158 #ifdef INET6
11159 	case AF_INET6:
11160 		src_sin6 = (struct sockaddr_in6 *)src;
11161 		dst_sin6 = (struct sockaddr_in6 *)dst;
11162 		ip6 = mtod(mout, struct ip6_hdr *);
11163 		ip6->ip6_flow = htonl(0x60000000);
11164 		if (V_ip6_auto_flowlabel) {
11165 			ip6->ip6_flow |= (htonl(ip6_randomflowlabel()) & IPV6_FLOWLABEL_MASK);
11166 		}
11167 		ip6->ip6_hlim = MODULE_GLOBAL(ip6_defhlim);
11168 		if (port) {
11169 			ip6->ip6_nxt = IPPROTO_UDP;
11170 		} else {
11171 			ip6->ip6_nxt = IPPROTO_SCTP;
11172 		}
11173 		ip6->ip6_src = dst_sin6->sin6_addr;
11174 		ip6->ip6_dst = src_sin6->sin6_addr;
11175 		len = sizeof(struct ip6_hdr);
11176 		shout = (struct sctphdr *)((caddr_t)ip6 + len);
11177 		break;
11178 #endif
11179 	default:
11180 		len = 0;
11181 		shout = mtod(mout, struct sctphdr *);
11182 		break;
11183 	}
11184 #if defined(INET) || defined(INET6)
11185 	if (port) {
11186 		if (htons(SCTP_BASE_SYSCTL(sctp_udp_tunneling_port)) == 0) {
11187 			sctp_m_freem(mout);
11188 			return;
11189 		}
11190 		udp = (struct udphdr *)shout;
11191 		udp->uh_sport = htons(SCTP_BASE_SYSCTL(sctp_udp_tunneling_port));
11192 		udp->uh_dport = port;
11193 		udp->uh_sum = 0;
11194 		udp->uh_ulen = htons((uint16_t)(sizeof(struct udphdr) +
11195 		    sizeof(struct sctphdr) +
11196 		    sizeof(struct sctp_chunkhdr) +
11197 		    cause_len + padding_len));
11198 		len += sizeof(struct udphdr);
11199 		shout = (struct sctphdr *)((caddr_t)shout + sizeof(struct udphdr));
11200 	} else {
11201 		udp = NULL;
11202 	}
11203 #endif
11204 	shout->src_port = sh->dest_port;
11205 	shout->dest_port = sh->src_port;
11206 	shout->checksum = 0;
11207 	if (vtag) {
11208 		shout->v_tag = htonl(vtag);
11209 	} else {
11210 		shout->v_tag = sh->v_tag;
11211 	}
11212 	len += sizeof(struct sctphdr);
11213 	ch = (struct sctp_chunkhdr *)((caddr_t)shout + sizeof(struct sctphdr));
11214 	ch->chunk_type = type;
11215 	if (vtag) {
11216 		ch->chunk_flags = 0;
11217 	} else {
11218 		ch->chunk_flags = SCTP_HAD_NO_TCB;
11219 	}
11220 	ch->chunk_length = htons((uint16_t)(sizeof(struct sctp_chunkhdr) + cause_len));
11221 	len += sizeof(struct sctp_chunkhdr);
11222 	len += cause_len + padding_len;
11223 
11224 	if (SCTP_GET_HEADER_FOR_OUTPUT(o_pak)) {
11225 		sctp_m_freem(mout);
11226 		return;
11227 	}
11228 	SCTP_ATTACH_CHAIN(o_pak, mout, len);
11229 	switch (dst->sa_family) {
11230 #ifdef INET
11231 	case AF_INET:
11232 		if (port) {
11233 			if (V_udp_cksum) {
11234 				udp->uh_sum = in_pseudo(ip->ip_src.s_addr, ip->ip_dst.s_addr, udp->uh_ulen + htons(IPPROTO_UDP));
11235 			} else {
11236 				udp->uh_sum = 0;
11237 			}
11238 		}
11239 		ip->ip_len = htons(len);
11240 		if (port) {
11241 			shout->checksum = sctp_calculate_cksum(mout, sizeof(struct ip) + sizeof(struct udphdr));
11242 			SCTP_STAT_INCR(sctps_sendswcrc);
11243 			if (V_udp_cksum) {
11244 				SCTP_ENABLE_UDP_CSUM(o_pak);
11245 			}
11246 		} else {
11247 			mout->m_pkthdr.csum_flags = CSUM_SCTP;
11248 			mout->m_pkthdr.csum_data = offsetof(struct sctphdr, checksum);
11249 			SCTP_STAT_INCR(sctps_sendhwcrc);
11250 		}
11251 #ifdef SCTP_PACKET_LOGGING
11252 		if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LAST_PACKET_TRACING) {
11253 			sctp_packet_log(o_pak);
11254 		}
11255 #endif
11256 		SCTP_IP_OUTPUT(ret, o_pak, NULL, NULL, vrf_id);
11257 		break;
11258 #endif
11259 #ifdef INET6
11260 	case AF_INET6:
11261 		ip6->ip6_plen = htons((uint16_t)(len - sizeof(struct ip6_hdr)));
11262 		if (port) {
11263 			shout->checksum = sctp_calculate_cksum(mout, sizeof(struct ip6_hdr) + sizeof(struct udphdr));
11264 			SCTP_STAT_INCR(sctps_sendswcrc);
11265 			if ((udp->uh_sum = in6_cksum(o_pak, IPPROTO_UDP, sizeof(struct ip6_hdr), len - sizeof(struct ip6_hdr))) == 0) {
11266 				udp->uh_sum = 0xffff;
11267 			}
11268 		} else {
11269 			mout->m_pkthdr.csum_flags = CSUM_SCTP_IPV6;
11270 			mout->m_pkthdr.csum_data = offsetof(struct sctphdr, checksum);
11271 			SCTP_STAT_INCR(sctps_sendhwcrc);
11272 		}
11273 #ifdef SCTP_PACKET_LOGGING
11274 		if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_LAST_PACKET_TRACING) {
11275 			sctp_packet_log(o_pak);
11276 		}
11277 #endif
11278 		SCTP_IP6_OUTPUT(ret, o_pak, NULL, NULL, NULL, vrf_id);
11279 		break;
11280 #endif
11281 	default:
11282 		SCTPDBG(SCTP_DEBUG_OUTPUT1, "Unknown protocol (TSNH) type %d\n",
11283 		    dst->sa_family);
11284 		sctp_m_freem(mout);
11285 		SCTP_LTRACE_ERR_RET_PKT(mout, NULL, NULL, NULL, SCTP_FROM_SCTP_OUTPUT, EFAULT);
11286 		return;
11287 	}
11288 	SCTPDBG(SCTP_DEBUG_OUTPUT3, "return from send is %d\n", ret);
11289 	SCTP_STAT_INCR(sctps_sendpackets);
11290 	SCTP_STAT_INCR_COUNTER64(sctps_outpackets);
11291 	SCTP_STAT_INCR_COUNTER64(sctps_outcontrolchunks);
11292 	if (ret) {
11293 		SCTP_STAT_INCR(sctps_senderrors);
11294 	}
11295 	return;
11296 }
11297 
11298 void
11299 sctp_send_shutdown_complete2(struct sockaddr *src, struct sockaddr *dst,
11300     struct sctphdr *sh,
11301     uint8_t mflowtype, uint32_t mflowid, uint16_t fibnum,
11302     uint32_t vrf_id, uint16_t port)
11303 {
11304 	sctp_send_resp_msg(src, dst, sh, 0, SCTP_SHUTDOWN_COMPLETE, NULL,
11305 	    mflowtype, mflowid, fibnum,
11306 	    vrf_id, port);
11307 }
11308 
11309 void
11310 sctp_send_hb(struct sctp_tcb *stcb, struct sctp_nets *net, int so_locked
11311 #if !defined(__APPLE__) && !defined(SCTP_SO_LOCK_TESTING)
11312     SCTP_UNUSED
11313 #endif
11314 )
11315 {
11316 	struct sctp_tmit_chunk *chk;
11317 	struct sctp_heartbeat_chunk *hb;
11318 	struct timeval now;
11319 
11320 	SCTP_TCB_LOCK_ASSERT(stcb);
11321 	if (net == NULL) {
11322 		return;
11323 	}
11324 	(void)SCTP_GETTIME_TIMEVAL(&now);
11325 	switch (net->ro._l_addr.sa.sa_family) {
11326 #ifdef INET
11327 	case AF_INET:
11328 		break;
11329 #endif
11330 #ifdef INET6
11331 	case AF_INET6:
11332 		break;
11333 #endif
11334 	default:
11335 		return;
11336 	}
11337 	sctp_alloc_a_chunk(stcb, chk);
11338 	if (chk == NULL) {
11339 		SCTPDBG(SCTP_DEBUG_OUTPUT4, "Gak, can't get a chunk for hb\n");
11340 		return;
11341 	}
11342 
11343 	chk->copy_by_ref = 0;
11344 	chk->rec.chunk_id.id = SCTP_HEARTBEAT_REQUEST;
11345 	chk->rec.chunk_id.can_take_data = 1;
11346 	chk->flags = 0;
11347 	chk->asoc = &stcb->asoc;
11348 	chk->send_size = sizeof(struct sctp_heartbeat_chunk);
11349 
11350 	chk->data = sctp_get_mbuf_for_msg(chk->send_size, 0, M_NOWAIT, 1, MT_HEADER);
11351 	if (chk->data == NULL) {
11352 		sctp_free_a_chunk(stcb, chk, so_locked);
11353 		return;
11354 	}
11355 	SCTP_BUF_RESV_UF(chk->data, SCTP_MIN_OVERHEAD);
11356 	SCTP_BUF_LEN(chk->data) = chk->send_size;
11357 	chk->sent = SCTP_DATAGRAM_UNSENT;
11358 	chk->snd_count = 0;
11359 	chk->whoTo = net;
11360 	atomic_add_int(&chk->whoTo->ref_count, 1);
11361 	/* Now we have a mbuf that we can fill in with the details */
11362 	hb = mtod(chk->data, struct sctp_heartbeat_chunk *);
11363 	memset(hb, 0, sizeof(struct sctp_heartbeat_chunk));
11364 	/* fill out chunk header */
11365 	hb->ch.chunk_type = SCTP_HEARTBEAT_REQUEST;
11366 	hb->ch.chunk_flags = 0;
11367 	hb->ch.chunk_length = htons(chk->send_size);
11368 	/* Fill out hb parameter */
11369 	hb->heartbeat.hb_info.ph.param_type = htons(SCTP_HEARTBEAT_INFO);
11370 	hb->heartbeat.hb_info.ph.param_length = htons(sizeof(struct sctp_heartbeat_info_param));
11371 	hb->heartbeat.hb_info.time_value_1 = now.tv_sec;
11372 	hb->heartbeat.hb_info.time_value_2 = now.tv_usec;
11373 	/* Did our user request this one, put it in */
11374 	hb->heartbeat.hb_info.addr_family = (uint8_t)net->ro._l_addr.sa.sa_family;
11375 	hb->heartbeat.hb_info.addr_len = net->ro._l_addr.sa.sa_len;
11376 	if (net->dest_state & SCTP_ADDR_UNCONFIRMED) {
11377 		/*
11378 		 * we only take from the entropy pool if the address is not
11379 		 * confirmed.
11380 		 */
11381 		net->heartbeat_random1 = hb->heartbeat.hb_info.random_value1 = sctp_select_initial_TSN(&stcb->sctp_ep->sctp_ep);
11382 		net->heartbeat_random2 = hb->heartbeat.hb_info.random_value2 = sctp_select_initial_TSN(&stcb->sctp_ep->sctp_ep);
11383 	} else {
11384 		net->heartbeat_random1 = hb->heartbeat.hb_info.random_value1 = 0;
11385 		net->heartbeat_random2 = hb->heartbeat.hb_info.random_value2 = 0;
11386 	}
11387 	switch (net->ro._l_addr.sa.sa_family) {
11388 #ifdef INET
11389 	case AF_INET:
11390 		memcpy(hb->heartbeat.hb_info.address,
11391 		    &net->ro._l_addr.sin.sin_addr,
11392 		    sizeof(net->ro._l_addr.sin.sin_addr));
11393 		break;
11394 #endif
11395 #ifdef INET6
11396 	case AF_INET6:
11397 		memcpy(hb->heartbeat.hb_info.address,
11398 		    &net->ro._l_addr.sin6.sin6_addr,
11399 		    sizeof(net->ro._l_addr.sin6.sin6_addr));
11400 		break;
11401 #endif
11402 	default:
11403 		if (chk->data) {
11404 			sctp_m_freem(chk->data);
11405 			chk->data = NULL;
11406 		}
11407 		sctp_free_a_chunk(stcb, chk, so_locked);
11408 		return;
11409 		break;
11410 	}
11411 	net->hb_responded = 0;
11412 	TAILQ_INSERT_TAIL(&stcb->asoc.control_send_queue, chk, sctp_next);
11413 	stcb->asoc.ctrl_queue_cnt++;
11414 	SCTP_STAT_INCR(sctps_sendheartbeat);
11415 	return;
11416 }
11417 
11418 void
11419 sctp_send_ecn_echo(struct sctp_tcb *stcb, struct sctp_nets *net,
11420     uint32_t high_tsn)
11421 {
11422 	struct sctp_association *asoc;
11423 	struct sctp_ecne_chunk *ecne;
11424 	struct sctp_tmit_chunk *chk;
11425 
11426 	if (net == NULL) {
11427 		return;
11428 	}
11429 	asoc = &stcb->asoc;
11430 	SCTP_TCB_LOCK_ASSERT(stcb);
11431 	TAILQ_FOREACH(chk, &asoc->control_send_queue, sctp_next) {
11432 		if ((chk->rec.chunk_id.id == SCTP_ECN_ECHO) && (net == chk->whoTo)) {
11433 			/* found a previous ECN_ECHO update it if needed */
11434 			uint32_t cnt, ctsn;
11435 
11436 			ecne = mtod(chk->data, struct sctp_ecne_chunk *);
11437 			ctsn = ntohl(ecne->tsn);
11438 			if (SCTP_TSN_GT(high_tsn, ctsn)) {
11439 				ecne->tsn = htonl(high_tsn);
11440 				SCTP_STAT_INCR(sctps_queue_upd_ecne);
11441 			}
11442 			cnt = ntohl(ecne->num_pkts_since_cwr);
11443 			cnt++;
11444 			ecne->num_pkts_since_cwr = htonl(cnt);
11445 			return;
11446 		}
11447 	}
11448 	/* nope could not find one to update so we must build one */
11449 	sctp_alloc_a_chunk(stcb, chk);
11450 	if (chk == NULL) {
11451 		return;
11452 	}
11453 	SCTP_STAT_INCR(sctps_queue_upd_ecne);
11454 	chk->copy_by_ref = 0;
11455 	chk->rec.chunk_id.id = SCTP_ECN_ECHO;
11456 	chk->rec.chunk_id.can_take_data = 0;
11457 	chk->flags = 0;
11458 	chk->asoc = &stcb->asoc;
11459 	chk->send_size = sizeof(struct sctp_ecne_chunk);
11460 	chk->data = sctp_get_mbuf_for_msg(chk->send_size, 0, M_NOWAIT, 1, MT_HEADER);
11461 	if (chk->data == NULL) {
11462 		sctp_free_a_chunk(stcb, chk, SCTP_SO_NOT_LOCKED);
11463 		return;
11464 	}
11465 	SCTP_BUF_RESV_UF(chk->data, SCTP_MIN_OVERHEAD);
11466 	SCTP_BUF_LEN(chk->data) = chk->send_size;
11467 	chk->sent = SCTP_DATAGRAM_UNSENT;
11468 	chk->snd_count = 0;
11469 	chk->whoTo = net;
11470 	atomic_add_int(&chk->whoTo->ref_count, 1);
11471 
11472 	stcb->asoc.ecn_echo_cnt_onq++;
11473 	ecne = mtod(chk->data, struct sctp_ecne_chunk *);
11474 	ecne->ch.chunk_type = SCTP_ECN_ECHO;
11475 	ecne->ch.chunk_flags = 0;
11476 	ecne->ch.chunk_length = htons(sizeof(struct sctp_ecne_chunk));
11477 	ecne->tsn = htonl(high_tsn);
11478 	ecne->num_pkts_since_cwr = htonl(1);
11479 	TAILQ_INSERT_HEAD(&stcb->asoc.control_send_queue, chk, sctp_next);
11480 	asoc->ctrl_queue_cnt++;
11481 }
11482 
11483 void
11484 sctp_send_packet_dropped(struct sctp_tcb *stcb, struct sctp_nets *net,
11485     struct mbuf *m, int len, int iphlen, int bad_crc)
11486 {
11487 	struct sctp_association *asoc;
11488 	struct sctp_pktdrop_chunk *drp;
11489 	struct sctp_tmit_chunk *chk;
11490 	uint8_t *datap;
11491 	int was_trunc = 0;
11492 	int fullsz = 0;
11493 	long spc;
11494 	int offset;
11495 	struct sctp_chunkhdr *ch, chunk_buf;
11496 	unsigned int chk_length;
11497 
11498 	if (!stcb) {
11499 		return;
11500 	}
11501 	asoc = &stcb->asoc;
11502 	SCTP_TCB_LOCK_ASSERT(stcb);
11503 	if (asoc->pktdrop_supported == 0) {
11504 		/*-
11505 		 * peer must declare support before I send one.
11506 		 */
11507 		return;
11508 	}
11509 	if (stcb->sctp_socket == NULL) {
11510 		return;
11511 	}
11512 	sctp_alloc_a_chunk(stcb, chk);
11513 	if (chk == NULL) {
11514 		return;
11515 	}
11516 	chk->copy_by_ref = 0;
11517 	chk->rec.chunk_id.id = SCTP_PACKET_DROPPED;
11518 	chk->rec.chunk_id.can_take_data = 1;
11519 	chk->flags = 0;
11520 	len -= iphlen;
11521 	chk->send_size = len;
11522 	/* Validate that we do not have an ABORT in here. */
11523 	offset = iphlen + sizeof(struct sctphdr);
11524 	ch = (struct sctp_chunkhdr *)sctp_m_getptr(m, offset,
11525 	    sizeof(*ch), (uint8_t *)&chunk_buf);
11526 	while (ch != NULL) {
11527 		chk_length = ntohs(ch->chunk_length);
11528 		if (chk_length < sizeof(*ch)) {
11529 			/* break to abort land */
11530 			break;
11531 		}
11532 		switch (ch->chunk_type) {
11533 		case SCTP_PACKET_DROPPED:
11534 		case SCTP_ABORT_ASSOCIATION:
11535 		case SCTP_INITIATION_ACK:
11536 			/**
11537 			 * We don't respond with an PKT-DROP to an ABORT
11538 			 * or PKT-DROP. We also do not respond to an
11539 			 * INIT-ACK, because we can't know if the initiation
11540 			 * tag is correct or not.
11541 			 */
11542 			sctp_free_a_chunk(stcb, chk, SCTP_SO_NOT_LOCKED);
11543 			return;
11544 		default:
11545 			break;
11546 		}
11547 		offset += SCTP_SIZE32(chk_length);
11548 		ch = (struct sctp_chunkhdr *)sctp_m_getptr(m, offset,
11549 		    sizeof(*ch), (uint8_t *)&chunk_buf);
11550 	}
11551 
11552 	if ((len + SCTP_MAX_OVERHEAD + sizeof(struct sctp_pktdrop_chunk)) >
11553 	    min(stcb->asoc.smallest_mtu, MCLBYTES)) {
11554 		/*
11555 		 * only send 1 mtu worth, trim off the excess on the end.
11556 		 */
11557 		fullsz = len;
11558 		len = min(stcb->asoc.smallest_mtu, MCLBYTES) - SCTP_MAX_OVERHEAD;
11559 		was_trunc = 1;
11560 	}
11561 	chk->asoc = &stcb->asoc;
11562 	chk->data = sctp_get_mbuf_for_msg(MCLBYTES, 0, M_NOWAIT, 1, MT_DATA);
11563 	if (chk->data == NULL) {
11564 jump_out:
11565 		sctp_free_a_chunk(stcb, chk, SCTP_SO_NOT_LOCKED);
11566 		return;
11567 	}
11568 	SCTP_BUF_RESV_UF(chk->data, SCTP_MIN_OVERHEAD);
11569 	drp = mtod(chk->data, struct sctp_pktdrop_chunk *);
11570 	if (drp == NULL) {
11571 		sctp_m_freem(chk->data);
11572 		chk->data = NULL;
11573 		goto jump_out;
11574 	}
11575 	chk->book_size = SCTP_SIZE32((chk->send_size + sizeof(struct sctp_pktdrop_chunk) +
11576 	    sizeof(struct sctphdr) + SCTP_MED_OVERHEAD));
11577 	chk->book_size_scale = 0;
11578 	if (was_trunc) {
11579 		drp->ch.chunk_flags = SCTP_PACKET_TRUNCATED;
11580 		drp->trunc_len = htons(fullsz);
11581 		/*
11582 		 * Len is already adjusted to size minus overhead above take
11583 		 * out the pkt_drop chunk itself from it.
11584 		 */
11585 		chk->send_size = (uint16_t)(len - sizeof(struct sctp_pktdrop_chunk));
11586 		len = chk->send_size;
11587 	} else {
11588 		/* no truncation needed */
11589 		drp->ch.chunk_flags = 0;
11590 		drp->trunc_len = htons(0);
11591 	}
11592 	if (bad_crc) {
11593 		drp->ch.chunk_flags |= SCTP_BADCRC;
11594 	}
11595 	chk->send_size += sizeof(struct sctp_pktdrop_chunk);
11596 	SCTP_BUF_LEN(chk->data) = chk->send_size;
11597 	chk->sent = SCTP_DATAGRAM_UNSENT;
11598 	chk->snd_count = 0;
11599 	if (net) {
11600 		/* we should hit here */
11601 		chk->whoTo = net;
11602 		atomic_add_int(&chk->whoTo->ref_count, 1);
11603 	} else {
11604 		chk->whoTo = NULL;
11605 	}
11606 	drp->ch.chunk_type = SCTP_PACKET_DROPPED;
11607 	drp->ch.chunk_length = htons(chk->send_size);
11608 	spc = SCTP_SB_LIMIT_RCV(stcb->sctp_socket);
11609 	if (spc < 0) {
11610 		spc = 0;
11611 	}
11612 	drp->bottle_bw = htonl(spc);
11613 	if (asoc->my_rwnd) {
11614 		drp->current_onq = htonl(asoc->size_on_reasm_queue +
11615 		    asoc->size_on_all_streams +
11616 		    asoc->my_rwnd_control_len +
11617 		    stcb->sctp_socket->so_rcv.sb_cc);
11618 	} else {
11619 		/*-
11620 		 * If my rwnd is 0, possibly from mbuf depletion as well as
11621 		 * space used, tell the peer there is NO space aka onq == bw
11622 		 */
11623 		drp->current_onq = htonl(spc);
11624 	}
11625 	drp->reserved = 0;
11626 	datap = drp->data;
11627 	m_copydata(m, iphlen, len, (caddr_t)datap);
11628 	TAILQ_INSERT_TAIL(&stcb->asoc.control_send_queue, chk, sctp_next);
11629 	asoc->ctrl_queue_cnt++;
11630 }
11631 
11632 void
11633 sctp_send_cwr(struct sctp_tcb *stcb, struct sctp_nets *net, uint32_t high_tsn, uint8_t override)
11634 {
11635 	struct sctp_association *asoc;
11636 	struct sctp_cwr_chunk *cwr;
11637 	struct sctp_tmit_chunk *chk;
11638 
11639 	SCTP_TCB_LOCK_ASSERT(stcb);
11640 	if (net == NULL) {
11641 		return;
11642 	}
11643 	asoc = &stcb->asoc;
11644 	TAILQ_FOREACH(chk, &asoc->control_send_queue, sctp_next) {
11645 		if ((chk->rec.chunk_id.id == SCTP_ECN_CWR) && (net == chk->whoTo)) {
11646 			/*
11647 			 * found a previous CWR queued to same destination
11648 			 * update it if needed
11649 			 */
11650 			uint32_t ctsn;
11651 
11652 			cwr = mtod(chk->data, struct sctp_cwr_chunk *);
11653 			ctsn = ntohl(cwr->tsn);
11654 			if (SCTP_TSN_GT(high_tsn, ctsn)) {
11655 				cwr->tsn = htonl(high_tsn);
11656 			}
11657 			if (override & SCTP_CWR_REDUCE_OVERRIDE) {
11658 				/* Make sure override is carried */
11659 				cwr->ch.chunk_flags |= SCTP_CWR_REDUCE_OVERRIDE;
11660 			}
11661 			return;
11662 		}
11663 	}
11664 	sctp_alloc_a_chunk(stcb, chk);
11665 	if (chk == NULL) {
11666 		return;
11667 	}
11668 	chk->copy_by_ref = 0;
11669 	chk->rec.chunk_id.id = SCTP_ECN_CWR;
11670 	chk->rec.chunk_id.can_take_data = 1;
11671 	chk->flags = 0;
11672 	chk->asoc = &stcb->asoc;
11673 	chk->send_size = sizeof(struct sctp_cwr_chunk);
11674 	chk->data = sctp_get_mbuf_for_msg(chk->send_size, 0, M_NOWAIT, 1, MT_HEADER);
11675 	if (chk->data == NULL) {
11676 		sctp_free_a_chunk(stcb, chk, SCTP_SO_NOT_LOCKED);
11677 		return;
11678 	}
11679 	SCTP_BUF_RESV_UF(chk->data, SCTP_MIN_OVERHEAD);
11680 	SCTP_BUF_LEN(chk->data) = chk->send_size;
11681 	chk->sent = SCTP_DATAGRAM_UNSENT;
11682 	chk->snd_count = 0;
11683 	chk->whoTo = net;
11684 	atomic_add_int(&chk->whoTo->ref_count, 1);
11685 	cwr = mtod(chk->data, struct sctp_cwr_chunk *);
11686 	cwr->ch.chunk_type = SCTP_ECN_CWR;
11687 	cwr->ch.chunk_flags = override;
11688 	cwr->ch.chunk_length = htons(sizeof(struct sctp_cwr_chunk));
11689 	cwr->tsn = htonl(high_tsn);
11690 	TAILQ_INSERT_TAIL(&stcb->asoc.control_send_queue, chk, sctp_next);
11691 	asoc->ctrl_queue_cnt++;
11692 }
11693 
11694 static int
11695 sctp_add_stream_reset_out(struct sctp_tcb *stcb, struct sctp_tmit_chunk *chk,
11696     uint32_t seq, uint32_t resp_seq, uint32_t last_sent)
11697 {
11698 	uint16_t len, old_len, i;
11699 	struct sctp_stream_reset_out_request *req_out;
11700 	struct sctp_chunkhdr *ch;
11701 	int at;
11702 	int number_entries = 0;
11703 
11704 	ch = mtod(chk->data, struct sctp_chunkhdr *);
11705 	old_len = len = SCTP_SIZE32(ntohs(ch->chunk_length));
11706 	/* get to new offset for the param. */
11707 	req_out = (struct sctp_stream_reset_out_request *)((caddr_t)ch + len);
11708 	/* now how long will this param be? */
11709 	for (i = 0; i < stcb->asoc.streamoutcnt; i++) {
11710 		if ((stcb->asoc.strmout[i].state == SCTP_STREAM_RESET_PENDING) &&
11711 		    (stcb->asoc.strmout[i].chunks_on_queues == 0) &&
11712 		    TAILQ_EMPTY(&stcb->asoc.strmout[i].outqueue)) {
11713 			number_entries++;
11714 		}
11715 	}
11716 	if (number_entries == 0) {
11717 		return (0);
11718 	}
11719 	if (number_entries == stcb->asoc.streamoutcnt) {
11720 		number_entries = 0;
11721 	}
11722 	if (number_entries > SCTP_MAX_STREAMS_AT_ONCE_RESET) {
11723 		number_entries = SCTP_MAX_STREAMS_AT_ONCE_RESET;
11724 	}
11725 	len = (uint16_t)(sizeof(struct sctp_stream_reset_out_request) + (sizeof(uint16_t) * number_entries));
11726 	req_out->ph.param_type = htons(SCTP_STR_RESET_OUT_REQUEST);
11727 	req_out->ph.param_length = htons(len);
11728 	req_out->request_seq = htonl(seq);
11729 	req_out->response_seq = htonl(resp_seq);
11730 	req_out->send_reset_at_tsn = htonl(last_sent);
11731 	at = 0;
11732 	if (number_entries) {
11733 		for (i = 0; i < stcb->asoc.streamoutcnt; i++) {
11734 			if ((stcb->asoc.strmout[i].state == SCTP_STREAM_RESET_PENDING) &&
11735 			    (stcb->asoc.strmout[i].chunks_on_queues == 0) &&
11736 			    TAILQ_EMPTY(&stcb->asoc.strmout[i].outqueue)) {
11737 				req_out->list_of_streams[at] = htons(i);
11738 				at++;
11739 				stcb->asoc.strmout[i].state = SCTP_STREAM_RESET_IN_FLIGHT;
11740 				if (at >= number_entries) {
11741 					break;
11742 				}
11743 			}
11744 		}
11745 	} else {
11746 		for (i = 0; i < stcb->asoc.streamoutcnt; i++) {
11747 			stcb->asoc.strmout[i].state = SCTP_STREAM_RESET_IN_FLIGHT;
11748 		}
11749 	}
11750 	if (SCTP_SIZE32(len) > len) {
11751 		/*-
11752 		 * Need to worry about the pad we may end up adding to the
11753 		 * end. This is easy since the struct is either aligned to 4
11754 		 * bytes or 2 bytes off.
11755 		 */
11756 		req_out->list_of_streams[number_entries] = 0;
11757 	}
11758 	/* now fix the chunk length */
11759 	ch->chunk_length = htons(len + old_len);
11760 	chk->book_size = len + old_len;
11761 	chk->book_size_scale = 0;
11762 	chk->send_size = SCTP_SIZE32(chk->book_size);
11763 	SCTP_BUF_LEN(chk->data) = chk->send_size;
11764 	return (1);
11765 }
11766 
11767 static void
11768 sctp_add_stream_reset_in(struct sctp_tmit_chunk *chk,
11769     int number_entries, uint16_t *list,
11770     uint32_t seq)
11771 {
11772 	uint16_t len, old_len, i;
11773 	struct sctp_stream_reset_in_request *req_in;
11774 	struct sctp_chunkhdr *ch;
11775 
11776 	ch = mtod(chk->data, struct sctp_chunkhdr *);
11777 	old_len = len = SCTP_SIZE32(ntohs(ch->chunk_length));
11778 
11779 	/* get to new offset for the param. */
11780 	req_in = (struct sctp_stream_reset_in_request *)((caddr_t)ch + len);
11781 	/* now how long will this param be? */
11782 	len = (uint16_t)(sizeof(struct sctp_stream_reset_in_request) + (sizeof(uint16_t) * number_entries));
11783 	req_in->ph.param_type = htons(SCTP_STR_RESET_IN_REQUEST);
11784 	req_in->ph.param_length = htons(len);
11785 	req_in->request_seq = htonl(seq);
11786 	if (number_entries) {
11787 		for (i = 0; i < number_entries; i++) {
11788 			req_in->list_of_streams[i] = htons(list[i]);
11789 		}
11790 	}
11791 	if (SCTP_SIZE32(len) > len) {
11792 		/*-
11793 		 * Need to worry about the pad we may end up adding to the
11794 		 * end. This is easy since the struct is either aligned to 4
11795 		 * bytes or 2 bytes off.
11796 		 */
11797 		req_in->list_of_streams[number_entries] = 0;
11798 	}
11799 	/* now fix the chunk length */
11800 	ch->chunk_length = htons(len + old_len);
11801 	chk->book_size = len + old_len;
11802 	chk->book_size_scale = 0;
11803 	chk->send_size = SCTP_SIZE32(chk->book_size);
11804 	SCTP_BUF_LEN(chk->data) = chk->send_size;
11805 	return;
11806 }
11807 
11808 static void
11809 sctp_add_stream_reset_tsn(struct sctp_tmit_chunk *chk,
11810     uint32_t seq)
11811 {
11812 	uint16_t len, old_len;
11813 	struct sctp_stream_reset_tsn_request *req_tsn;
11814 	struct sctp_chunkhdr *ch;
11815 
11816 	ch = mtod(chk->data, struct sctp_chunkhdr *);
11817 	old_len = len = SCTP_SIZE32(ntohs(ch->chunk_length));
11818 
11819 	/* get to new offset for the param. */
11820 	req_tsn = (struct sctp_stream_reset_tsn_request *)((caddr_t)ch + len);
11821 	/* now how long will this param be? */
11822 	len = sizeof(struct sctp_stream_reset_tsn_request);
11823 	req_tsn->ph.param_type = htons(SCTP_STR_RESET_TSN_REQUEST);
11824 	req_tsn->ph.param_length = htons(len);
11825 	req_tsn->request_seq = htonl(seq);
11826 
11827 	/* now fix the chunk length */
11828 	ch->chunk_length = htons(len + old_len);
11829 	chk->send_size = len + old_len;
11830 	chk->book_size = SCTP_SIZE32(chk->send_size);
11831 	chk->book_size_scale = 0;
11832 	SCTP_BUF_LEN(chk->data) = SCTP_SIZE32(chk->send_size);
11833 	return;
11834 }
11835 
11836 void
11837 sctp_add_stream_reset_result(struct sctp_tmit_chunk *chk,
11838     uint32_t resp_seq, uint32_t result)
11839 {
11840 	uint16_t len, old_len;
11841 	struct sctp_stream_reset_response *resp;
11842 	struct sctp_chunkhdr *ch;
11843 
11844 	ch = mtod(chk->data, struct sctp_chunkhdr *);
11845 	old_len = len = SCTP_SIZE32(ntohs(ch->chunk_length));
11846 
11847 	/* get to new offset for the param. */
11848 	resp = (struct sctp_stream_reset_response *)((caddr_t)ch + len);
11849 	/* now how long will this param be? */
11850 	len = sizeof(struct sctp_stream_reset_response);
11851 	resp->ph.param_type = htons(SCTP_STR_RESET_RESPONSE);
11852 	resp->ph.param_length = htons(len);
11853 	resp->response_seq = htonl(resp_seq);
11854 	resp->result = ntohl(result);
11855 
11856 	/* now fix the chunk length */
11857 	ch->chunk_length = htons(len + old_len);
11858 	chk->book_size = len + old_len;
11859 	chk->book_size_scale = 0;
11860 	chk->send_size = SCTP_SIZE32(chk->book_size);
11861 	SCTP_BUF_LEN(chk->data) = chk->send_size;
11862 	return;
11863 }
11864 
11865 void
11866 sctp_send_deferred_reset_response(struct sctp_tcb *stcb,
11867     struct sctp_stream_reset_list *ent,
11868     int response)
11869 {
11870 	struct sctp_association *asoc;
11871 	struct sctp_tmit_chunk *chk;
11872 	struct sctp_chunkhdr *ch;
11873 
11874 	asoc = &stcb->asoc;
11875 
11876 	/*
11877 	 * Reset our last reset action to the new one IP -> response
11878 	 * (PERFORMED probably). This assures that if we fail to send, a
11879 	 * retran from the peer will get the new response.
11880 	 */
11881 	asoc->last_reset_action[0] = response;
11882 	if (asoc->stream_reset_outstanding) {
11883 		return;
11884 	}
11885 	sctp_alloc_a_chunk(stcb, chk);
11886 	if (chk == NULL) {
11887 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
11888 		return;
11889 	}
11890 	chk->copy_by_ref = 0;
11891 	chk->rec.chunk_id.id = SCTP_STREAM_RESET;
11892 	chk->rec.chunk_id.can_take_data = 0;
11893 	chk->flags = 0;
11894 	chk->asoc = &stcb->asoc;
11895 	chk->book_size = sizeof(struct sctp_chunkhdr);
11896 	chk->send_size = SCTP_SIZE32(chk->book_size);
11897 	chk->book_size_scale = 0;
11898 	chk->data = sctp_get_mbuf_for_msg(MCLBYTES, 0, M_NOWAIT, 1, MT_DATA);
11899 	if (chk->data == NULL) {
11900 		sctp_free_a_chunk(stcb, chk, SCTP_SO_LOCKED);
11901 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
11902 		return;
11903 	}
11904 	SCTP_BUF_RESV_UF(chk->data, SCTP_MIN_OVERHEAD);
11905 	/* setup chunk parameters */
11906 	chk->sent = SCTP_DATAGRAM_UNSENT;
11907 	chk->snd_count = 0;
11908 	if (stcb->asoc.alternate) {
11909 		chk->whoTo = stcb->asoc.alternate;
11910 	} else {
11911 		chk->whoTo = stcb->asoc.primary_destination;
11912 	}
11913 	ch = mtod(chk->data, struct sctp_chunkhdr *);
11914 	ch->chunk_type = SCTP_STREAM_RESET;
11915 	ch->chunk_flags = 0;
11916 	ch->chunk_length = htons(chk->book_size);
11917 	atomic_add_int(&chk->whoTo->ref_count, 1);
11918 	SCTP_BUF_LEN(chk->data) = chk->send_size;
11919 	sctp_add_stream_reset_result(chk, ent->seq, response);
11920 	/* insert the chunk for sending */
11921 	TAILQ_INSERT_TAIL(&asoc->control_send_queue,
11922 	    chk,
11923 	    sctp_next);
11924 	asoc->ctrl_queue_cnt++;
11925 }
11926 
11927 void
11928 sctp_add_stream_reset_result_tsn(struct sctp_tmit_chunk *chk,
11929     uint32_t resp_seq, uint32_t result,
11930     uint32_t send_una, uint32_t recv_next)
11931 {
11932 	uint16_t len, old_len;
11933 	struct sctp_stream_reset_response_tsn *resp;
11934 	struct sctp_chunkhdr *ch;
11935 
11936 	ch = mtod(chk->data, struct sctp_chunkhdr *);
11937 	old_len = len = SCTP_SIZE32(ntohs(ch->chunk_length));
11938 
11939 	/* get to new offset for the param. */
11940 	resp = (struct sctp_stream_reset_response_tsn *)((caddr_t)ch + len);
11941 	/* now how long will this param be? */
11942 	len = sizeof(struct sctp_stream_reset_response_tsn);
11943 	resp->ph.param_type = htons(SCTP_STR_RESET_RESPONSE);
11944 	resp->ph.param_length = htons(len);
11945 	resp->response_seq = htonl(resp_seq);
11946 	resp->result = htonl(result);
11947 	resp->senders_next_tsn = htonl(send_una);
11948 	resp->receivers_next_tsn = htonl(recv_next);
11949 
11950 	/* now fix the chunk length */
11951 	ch->chunk_length = htons(len + old_len);
11952 	chk->book_size = len + old_len;
11953 	chk->send_size = SCTP_SIZE32(chk->book_size);
11954 	chk->book_size_scale = 0;
11955 	SCTP_BUF_LEN(chk->data) = chk->send_size;
11956 	return;
11957 }
11958 
11959 static void
11960 sctp_add_an_out_stream(struct sctp_tmit_chunk *chk,
11961     uint32_t seq,
11962     uint16_t adding)
11963 {
11964 	uint16_t len, old_len;
11965 	struct sctp_chunkhdr *ch;
11966 	struct sctp_stream_reset_add_strm *addstr;
11967 
11968 	ch = mtod(chk->data, struct sctp_chunkhdr *);
11969 	old_len = len = SCTP_SIZE32(ntohs(ch->chunk_length));
11970 
11971 	/* get to new offset for the param. */
11972 	addstr = (struct sctp_stream_reset_add_strm *)((caddr_t)ch + len);
11973 	/* now how long will this param be? */
11974 	len = sizeof(struct sctp_stream_reset_add_strm);
11975 
11976 	/* Fill it out. */
11977 	addstr->ph.param_type = htons(SCTP_STR_RESET_ADD_OUT_STREAMS);
11978 	addstr->ph.param_length = htons(len);
11979 	addstr->request_seq = htonl(seq);
11980 	addstr->number_of_streams = htons(adding);
11981 	addstr->reserved = 0;
11982 
11983 	/* now fix the chunk length */
11984 	ch->chunk_length = htons(len + old_len);
11985 	chk->send_size = len + old_len;
11986 	chk->book_size = SCTP_SIZE32(chk->send_size);
11987 	chk->book_size_scale = 0;
11988 	SCTP_BUF_LEN(chk->data) = SCTP_SIZE32(chk->send_size);
11989 	return;
11990 }
11991 
11992 static void
11993 sctp_add_an_in_stream(struct sctp_tmit_chunk *chk,
11994     uint32_t seq,
11995     uint16_t adding)
11996 {
11997 	uint16_t len, old_len;
11998 	struct sctp_chunkhdr *ch;
11999 	struct sctp_stream_reset_add_strm *addstr;
12000 
12001 	ch = mtod(chk->data, struct sctp_chunkhdr *);
12002 	old_len = len = SCTP_SIZE32(ntohs(ch->chunk_length));
12003 
12004 	/* get to new offset for the param. */
12005 	addstr = (struct sctp_stream_reset_add_strm *)((caddr_t)ch + len);
12006 	/* now how long will this param be? */
12007 	len = sizeof(struct sctp_stream_reset_add_strm);
12008 	/* Fill it out. */
12009 	addstr->ph.param_type = htons(SCTP_STR_RESET_ADD_IN_STREAMS);
12010 	addstr->ph.param_length = htons(len);
12011 	addstr->request_seq = htonl(seq);
12012 	addstr->number_of_streams = htons(adding);
12013 	addstr->reserved = 0;
12014 
12015 	/* now fix the chunk length */
12016 	ch->chunk_length = htons(len + old_len);
12017 	chk->send_size = len + old_len;
12018 	chk->book_size = SCTP_SIZE32(chk->send_size);
12019 	chk->book_size_scale = 0;
12020 	SCTP_BUF_LEN(chk->data) = SCTP_SIZE32(chk->send_size);
12021 	return;
12022 }
12023 
12024 int
12025 sctp_send_stream_reset_out_if_possible(struct sctp_tcb *stcb, int so_locked)
12026 {
12027 	struct sctp_association *asoc;
12028 	struct sctp_tmit_chunk *chk;
12029 	struct sctp_chunkhdr *ch;
12030 	uint32_t seq;
12031 
12032 	asoc = &stcb->asoc;
12033 	asoc->trigger_reset = 0;
12034 	if (asoc->stream_reset_outstanding) {
12035 		return (EALREADY);
12036 	}
12037 	sctp_alloc_a_chunk(stcb, chk);
12038 	if (chk == NULL) {
12039 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
12040 		return (ENOMEM);
12041 	}
12042 	chk->copy_by_ref = 0;
12043 	chk->rec.chunk_id.id = SCTP_STREAM_RESET;
12044 	chk->rec.chunk_id.can_take_data = 0;
12045 	chk->flags = 0;
12046 	chk->asoc = &stcb->asoc;
12047 	chk->book_size = sizeof(struct sctp_chunkhdr);
12048 	chk->send_size = SCTP_SIZE32(chk->book_size);
12049 	chk->book_size_scale = 0;
12050 	chk->data = sctp_get_mbuf_for_msg(MCLBYTES, 0, M_NOWAIT, 1, MT_DATA);
12051 	if (chk->data == NULL) {
12052 		sctp_free_a_chunk(stcb, chk, so_locked);
12053 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
12054 		return (ENOMEM);
12055 	}
12056 	SCTP_BUF_RESV_UF(chk->data, SCTP_MIN_OVERHEAD);
12057 
12058 	/* setup chunk parameters */
12059 	chk->sent = SCTP_DATAGRAM_UNSENT;
12060 	chk->snd_count = 0;
12061 	if (stcb->asoc.alternate) {
12062 		chk->whoTo = stcb->asoc.alternate;
12063 	} else {
12064 		chk->whoTo = stcb->asoc.primary_destination;
12065 	}
12066 	ch = mtod(chk->data, struct sctp_chunkhdr *);
12067 	ch->chunk_type = SCTP_STREAM_RESET;
12068 	ch->chunk_flags = 0;
12069 	ch->chunk_length = htons(chk->book_size);
12070 	atomic_add_int(&chk->whoTo->ref_count, 1);
12071 	SCTP_BUF_LEN(chk->data) = chk->send_size;
12072 	seq = stcb->asoc.str_reset_seq_out;
12073 	if (sctp_add_stream_reset_out(stcb, chk, seq, (stcb->asoc.str_reset_seq_in - 1), (stcb->asoc.sending_seq - 1))) {
12074 		seq++;
12075 		asoc->stream_reset_outstanding++;
12076 	} else {
12077 		m_freem(chk->data);
12078 		chk->data = NULL;
12079 		sctp_free_a_chunk(stcb, chk, so_locked);
12080 		return (ENOENT);
12081 	}
12082 	asoc->str_reset = chk;
12083 	/* insert the chunk for sending */
12084 	TAILQ_INSERT_TAIL(&asoc->control_send_queue,
12085 	    chk,
12086 	    sctp_next);
12087 	asoc->ctrl_queue_cnt++;
12088 
12089 	if (stcb->asoc.send_sack) {
12090 		sctp_send_sack(stcb, so_locked);
12091 	}
12092 	sctp_timer_start(SCTP_TIMER_TYPE_STRRESET, stcb->sctp_ep, stcb, chk->whoTo);
12093 	return (0);
12094 }
12095 
12096 int
12097 sctp_send_str_reset_req(struct sctp_tcb *stcb,
12098     uint16_t number_entries, uint16_t *list,
12099     uint8_t send_in_req,
12100     uint8_t send_tsn_req,
12101     uint8_t add_stream,
12102     uint16_t adding_o,
12103     uint16_t adding_i, uint8_t peer_asked)
12104 {
12105 	struct sctp_association *asoc;
12106 	struct sctp_tmit_chunk *chk;
12107 	struct sctp_chunkhdr *ch;
12108 	int can_send_out_req = 0;
12109 	uint32_t seq;
12110 
12111 	asoc = &stcb->asoc;
12112 	if (asoc->stream_reset_outstanding) {
12113 		/*-
12114 		 * Already one pending, must get ACK back to clear the flag.
12115 		 */
12116 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, EBUSY);
12117 		return (EBUSY);
12118 	}
12119 	if ((send_in_req == 0) && (send_tsn_req == 0) &&
12120 	    (add_stream == 0)) {
12121 		/* nothing to do */
12122 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12123 		return (EINVAL);
12124 	}
12125 	if (send_tsn_req && send_in_req) {
12126 		/* error, can't do that */
12127 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12128 		return (EINVAL);
12129 	} else if (send_in_req) {
12130 		can_send_out_req = 1;
12131 	}
12132 	if (number_entries > (MCLBYTES -
12133 	    SCTP_MIN_OVERHEAD -
12134 	    sizeof(struct sctp_chunkhdr) -
12135 	    sizeof(struct sctp_stream_reset_out_request)) /
12136 	    sizeof(uint16_t)) {
12137 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
12138 		return (ENOMEM);
12139 	}
12140 	sctp_alloc_a_chunk(stcb, chk);
12141 	if (chk == NULL) {
12142 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
12143 		return (ENOMEM);
12144 	}
12145 	chk->copy_by_ref = 0;
12146 	chk->rec.chunk_id.id = SCTP_STREAM_RESET;
12147 	chk->rec.chunk_id.can_take_data = 0;
12148 	chk->flags = 0;
12149 	chk->asoc = &stcb->asoc;
12150 	chk->book_size = sizeof(struct sctp_chunkhdr);
12151 	chk->send_size = SCTP_SIZE32(chk->book_size);
12152 	chk->book_size_scale = 0;
12153 
12154 	chk->data = sctp_get_mbuf_for_msg(MCLBYTES, 0, M_NOWAIT, 1, MT_DATA);
12155 	if (chk->data == NULL) {
12156 		sctp_free_a_chunk(stcb, chk, SCTP_SO_LOCKED);
12157 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
12158 		return (ENOMEM);
12159 	}
12160 	SCTP_BUF_RESV_UF(chk->data, SCTP_MIN_OVERHEAD);
12161 
12162 	/* setup chunk parameters */
12163 	chk->sent = SCTP_DATAGRAM_UNSENT;
12164 	chk->snd_count = 0;
12165 	if (stcb->asoc.alternate) {
12166 		chk->whoTo = stcb->asoc.alternate;
12167 	} else {
12168 		chk->whoTo = stcb->asoc.primary_destination;
12169 	}
12170 	atomic_add_int(&chk->whoTo->ref_count, 1);
12171 	ch = mtod(chk->data, struct sctp_chunkhdr *);
12172 	ch->chunk_type = SCTP_STREAM_RESET;
12173 	ch->chunk_flags = 0;
12174 	ch->chunk_length = htons(chk->book_size);
12175 	SCTP_BUF_LEN(chk->data) = chk->send_size;
12176 
12177 	seq = stcb->asoc.str_reset_seq_out;
12178 	if (can_send_out_req) {
12179 		int ret;
12180 
12181 		ret = sctp_add_stream_reset_out(stcb, chk, seq, (stcb->asoc.str_reset_seq_in - 1), (stcb->asoc.sending_seq - 1));
12182 		if (ret) {
12183 			seq++;
12184 			asoc->stream_reset_outstanding++;
12185 		}
12186 	}
12187 	if ((add_stream & 1) &&
12188 	    ((stcb->asoc.strm_realoutsize - stcb->asoc.streamoutcnt) < adding_o)) {
12189 		/* Need to allocate more */
12190 		struct sctp_stream_out *oldstream;
12191 		struct sctp_stream_queue_pending *sp, *nsp;
12192 		int i;
12193 #if defined(SCTP_DETAILED_STR_STATS)
12194 		int j;
12195 #endif
12196 
12197 		oldstream = stcb->asoc.strmout;
12198 		/* get some more */
12199 		SCTP_MALLOC(stcb->asoc.strmout, struct sctp_stream_out *,
12200 		    (stcb->asoc.streamoutcnt + adding_o) * sizeof(struct sctp_stream_out),
12201 		    SCTP_M_STRMO);
12202 		if (stcb->asoc.strmout == NULL) {
12203 			uint8_t x;
12204 
12205 			stcb->asoc.strmout = oldstream;
12206 			/* Turn off the bit */
12207 			x = add_stream & 0xfe;
12208 			add_stream = x;
12209 			goto skip_stuff;
12210 		}
12211 		/*
12212 		 * Ok now we proceed with copying the old out stuff and
12213 		 * initializing the new stuff.
12214 		 */
12215 		SCTP_TCB_SEND_LOCK(stcb);
12216 		stcb->asoc.ss_functions.sctp_ss_clear(stcb, &stcb->asoc, 0, 1);
12217 		for (i = 0; i < stcb->asoc.streamoutcnt; i++) {
12218 			TAILQ_INIT(&stcb->asoc.strmout[i].outqueue);
12219 			stcb->asoc.strmout[i].chunks_on_queues = oldstream[i].chunks_on_queues;
12220 			stcb->asoc.strmout[i].next_mid_ordered = oldstream[i].next_mid_ordered;
12221 			stcb->asoc.strmout[i].next_mid_unordered = oldstream[i].next_mid_unordered;
12222 			stcb->asoc.strmout[i].last_msg_incomplete = oldstream[i].last_msg_incomplete;
12223 			stcb->asoc.strmout[i].sid = i;
12224 			stcb->asoc.strmout[i].state = oldstream[i].state;
12225 			/* FIX ME FIX ME */
12226 			/*
12227 			 * This should be a SS_COPY operation FIX ME STREAM
12228 			 * SCHEDULER EXPERT
12229 			 */
12230 			stcb->asoc.ss_functions.sctp_ss_init_stream(stcb, &stcb->asoc.strmout[i], &oldstream[i]);
12231 			/* now anything on those queues? */
12232 			TAILQ_FOREACH_SAFE(sp, &oldstream[i].outqueue, next, nsp) {
12233 				TAILQ_REMOVE(&oldstream[i].outqueue, sp, next);
12234 				TAILQ_INSERT_TAIL(&stcb->asoc.strmout[i].outqueue, sp, next);
12235 			}
12236 
12237 		}
12238 		/* now the new streams */
12239 		stcb->asoc.ss_functions.sctp_ss_init(stcb, &stcb->asoc, 1);
12240 		for (i = stcb->asoc.streamoutcnt; i < (stcb->asoc.streamoutcnt + adding_o); i++) {
12241 			TAILQ_INIT(&stcb->asoc.strmout[i].outqueue);
12242 			stcb->asoc.strmout[i].chunks_on_queues = 0;
12243 #if defined(SCTP_DETAILED_STR_STATS)
12244 			for (j = 0; j < SCTP_PR_SCTP_MAX + 1; j++) {
12245 				stcb->asoc.strmout[i].abandoned_sent[j] = 0;
12246 				stcb->asoc.strmout[i].abandoned_unsent[j] = 0;
12247 			}
12248 #else
12249 			stcb->asoc.strmout[i].abandoned_sent[0] = 0;
12250 			stcb->asoc.strmout[i].abandoned_unsent[0] = 0;
12251 #endif
12252 			stcb->asoc.strmout[i].next_mid_ordered = 0;
12253 			stcb->asoc.strmout[i].next_mid_unordered = 0;
12254 			stcb->asoc.strmout[i].sid = i;
12255 			stcb->asoc.strmout[i].last_msg_incomplete = 0;
12256 			stcb->asoc.ss_functions.sctp_ss_init_stream(stcb, &stcb->asoc.strmout[i], NULL);
12257 			stcb->asoc.strmout[i].state = SCTP_STREAM_CLOSED;
12258 		}
12259 		stcb->asoc.strm_realoutsize = stcb->asoc.streamoutcnt + adding_o;
12260 		SCTP_FREE(oldstream, SCTP_M_STRMO);
12261 		SCTP_TCB_SEND_UNLOCK(stcb);
12262 	}
12263 skip_stuff:
12264 	if ((add_stream & 1) && (adding_o > 0)) {
12265 		asoc->strm_pending_add_size = adding_o;
12266 		asoc->peer_req_out = peer_asked;
12267 		sctp_add_an_out_stream(chk, seq, adding_o);
12268 		seq++;
12269 		asoc->stream_reset_outstanding++;
12270 	}
12271 	if ((add_stream & 2) && (adding_i > 0)) {
12272 		sctp_add_an_in_stream(chk, seq, adding_i);
12273 		seq++;
12274 		asoc->stream_reset_outstanding++;
12275 	}
12276 	if (send_in_req) {
12277 		sctp_add_stream_reset_in(chk, number_entries, list, seq);
12278 		seq++;
12279 		asoc->stream_reset_outstanding++;
12280 	}
12281 	if (send_tsn_req) {
12282 		sctp_add_stream_reset_tsn(chk, seq);
12283 		asoc->stream_reset_outstanding++;
12284 	}
12285 	asoc->str_reset = chk;
12286 	/* insert the chunk for sending */
12287 	TAILQ_INSERT_TAIL(&asoc->control_send_queue,
12288 	    chk,
12289 	    sctp_next);
12290 	asoc->ctrl_queue_cnt++;
12291 	if (stcb->asoc.send_sack) {
12292 		sctp_send_sack(stcb, SCTP_SO_LOCKED);
12293 	}
12294 	sctp_timer_start(SCTP_TIMER_TYPE_STRRESET, stcb->sctp_ep, stcb, chk->whoTo);
12295 	return (0);
12296 }
12297 
12298 void
12299 sctp_send_abort(struct mbuf *m, int iphlen, struct sockaddr *src, struct sockaddr *dst,
12300     struct sctphdr *sh, uint32_t vtag, struct mbuf *cause,
12301     uint8_t mflowtype, uint32_t mflowid, uint16_t fibnum,
12302     uint32_t vrf_id, uint16_t port)
12303 {
12304 	/* Don't respond to an ABORT with an ABORT. */
12305 	if (sctp_is_there_an_abort_here(m, iphlen, &vtag)) {
12306 		if (cause)
12307 			sctp_m_freem(cause);
12308 		return;
12309 	}
12310 	sctp_send_resp_msg(src, dst, sh, vtag, SCTP_ABORT_ASSOCIATION, cause,
12311 	    mflowtype, mflowid, fibnum,
12312 	    vrf_id, port);
12313 	return;
12314 }
12315 
12316 void
12317 sctp_send_operr_to(struct sockaddr *src, struct sockaddr *dst,
12318     struct sctphdr *sh, uint32_t vtag, struct mbuf *cause,
12319     uint8_t mflowtype, uint32_t mflowid, uint16_t fibnum,
12320     uint32_t vrf_id, uint16_t port)
12321 {
12322 	sctp_send_resp_msg(src, dst, sh, vtag, SCTP_OPERATION_ERROR, cause,
12323 	    mflowtype, mflowid, fibnum,
12324 	    vrf_id, port);
12325 	return;
12326 }
12327 
12328 static struct mbuf *
12329 sctp_copy_resume(struct uio *uio,
12330     int max_send_len,
12331     int user_marks_eor,
12332     int *error,
12333     uint32_t *sndout,
12334     struct mbuf **new_tail)
12335 {
12336 	struct mbuf *m;
12337 
12338 	m = m_uiotombuf(uio, M_WAITOK, max_send_len, 0,
12339 	    (M_PKTHDR | (user_marks_eor ? M_EOR : 0)));
12340 	if (m == NULL) {
12341 		SCTP_LTRACE_ERR_RET(NULL, NULL, NULL, SCTP_FROM_SCTP_OUTPUT, ENOBUFS);
12342 		*error = ENOBUFS;
12343 	} else {
12344 		*sndout = m_length(m, NULL);
12345 		*new_tail = m_last(m);
12346 	}
12347 	return (m);
12348 }
12349 
12350 static int
12351 sctp_copy_one(struct sctp_stream_queue_pending *sp,
12352     struct uio *uio,
12353     int resv_upfront)
12354 {
12355 	sp->data = m_uiotombuf(uio, M_WAITOK, sp->length,
12356 	    resv_upfront, 0);
12357 	if (sp->data == NULL) {
12358 		SCTP_LTRACE_ERR_RET(NULL, NULL, NULL, SCTP_FROM_SCTP_OUTPUT, ENOBUFS);
12359 		return (ENOBUFS);
12360 	}
12361 
12362 	sp->tail_mbuf = m_last(sp->data);
12363 	return (0);
12364 }
12365 
12366 
12367 
12368 static struct sctp_stream_queue_pending *
12369 sctp_copy_it_in(struct sctp_tcb *stcb,
12370     struct sctp_association *asoc,
12371     struct sctp_sndrcvinfo *srcv,
12372     struct uio *uio,
12373     struct sctp_nets *net,
12374     int max_send_len,
12375     int user_marks_eor,
12376     int *error)
12377 {
12378 
12379 	/*-
12380 	 * This routine must be very careful in its work. Protocol
12381 	 * processing is up and running so care must be taken to spl...()
12382 	 * when you need to do something that may effect the stcb/asoc. The
12383 	 * sb is locked however. When data is copied the protocol processing
12384 	 * should be enabled since this is a slower operation...
12385 	 */
12386 	struct sctp_stream_queue_pending *sp = NULL;
12387 	int resv_in_first;
12388 
12389 	*error = 0;
12390 	/* Now can we send this? */
12391 	if ((SCTP_GET_STATE(stcb) == SCTP_STATE_SHUTDOWN_SENT) ||
12392 	    (SCTP_GET_STATE(stcb) == SCTP_STATE_SHUTDOWN_ACK_SENT) ||
12393 	    (SCTP_GET_STATE(stcb) == SCTP_STATE_SHUTDOWN_RECEIVED) ||
12394 	    (asoc->state & SCTP_STATE_SHUTDOWN_PENDING)) {
12395 		/* got data while shutting down */
12396 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ECONNRESET);
12397 		*error = ECONNRESET;
12398 		goto out_now;
12399 	}
12400 	sctp_alloc_a_strmoq(stcb, sp);
12401 	if (sp == NULL) {
12402 		SCTP_LTRACE_ERR_RET(NULL, stcb, net, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
12403 		*error = ENOMEM;
12404 		goto out_now;
12405 	}
12406 	sp->act_flags = 0;
12407 	sp->sender_all_done = 0;
12408 	sp->sinfo_flags = srcv->sinfo_flags;
12409 	sp->timetolive = srcv->sinfo_timetolive;
12410 	sp->ppid = srcv->sinfo_ppid;
12411 	sp->context = srcv->sinfo_context;
12412 	sp->fsn = 0;
12413 	(void)SCTP_GETTIME_TIMEVAL(&sp->ts);
12414 
12415 	sp->sid = srcv->sinfo_stream;
12416 	sp->length = (uint32_t)min(uio->uio_resid, max_send_len);
12417 	if ((sp->length == (uint32_t)uio->uio_resid) &&
12418 	    ((user_marks_eor == 0) ||
12419 	    (srcv->sinfo_flags & SCTP_EOF) ||
12420 	    (user_marks_eor && (srcv->sinfo_flags & SCTP_EOR)))) {
12421 		sp->msg_is_complete = 1;
12422 	} else {
12423 		sp->msg_is_complete = 0;
12424 	}
12425 	sp->sender_all_done = 0;
12426 	sp->some_taken = 0;
12427 	sp->put_last_out = 0;
12428 	resv_in_first = SCTP_DATA_CHUNK_OVERHEAD(stcb);
12429 	sp->data = sp->tail_mbuf = NULL;
12430 	if (sp->length == 0) {
12431 		goto skip_copy;
12432 	}
12433 	if (srcv->sinfo_keynumber_valid) {
12434 		sp->auth_keyid = srcv->sinfo_keynumber;
12435 	} else {
12436 		sp->auth_keyid = stcb->asoc.authinfo.active_keyid;
12437 	}
12438 	if (sctp_auth_is_required_chunk(SCTP_DATA, stcb->asoc.peer_auth_chunks)) {
12439 		sctp_auth_key_acquire(stcb, sp->auth_keyid);
12440 		sp->holds_key_ref = 1;
12441 	}
12442 	*error = sctp_copy_one(sp, uio, resv_in_first);
12443 skip_copy:
12444 	if (*error) {
12445 		sctp_free_a_strmoq(stcb, sp, SCTP_SO_LOCKED);
12446 		sp = NULL;
12447 	} else {
12448 		if (sp->sinfo_flags & SCTP_ADDR_OVER) {
12449 			sp->net = net;
12450 			atomic_add_int(&sp->net->ref_count, 1);
12451 		} else {
12452 			sp->net = NULL;
12453 		}
12454 		sctp_set_prsctp_policy(sp);
12455 	}
12456 out_now:
12457 	return (sp);
12458 }
12459 
12460 
12461 int
12462 sctp_sosend(struct socket *so,
12463     struct sockaddr *addr,
12464     struct uio *uio,
12465     struct mbuf *top,
12466     struct mbuf *control,
12467     int flags,
12468     struct thread *p
12469 )
12470 {
12471 	int error, use_sndinfo = 0;
12472 	struct sctp_sndrcvinfo sndrcvninfo;
12473 	struct sockaddr *addr_to_use;
12474 #if defined(INET) && defined(INET6)
12475 	struct sockaddr_in sin;
12476 #endif
12477 
12478 	if (control) {
12479 		/* process cmsg snd/rcv info (maybe a assoc-id) */
12480 		if (sctp_find_cmsg(SCTP_SNDRCV, (void *)&sndrcvninfo, control,
12481 		    sizeof(sndrcvninfo))) {
12482 			/* got one */
12483 			use_sndinfo = 1;
12484 		}
12485 	}
12486 	addr_to_use = addr;
12487 #if defined(INET) && defined(INET6)
12488 	if ((addr) && (addr->sa_family == AF_INET6)) {
12489 		struct sockaddr_in6 *sin6;
12490 
12491 		sin6 = (struct sockaddr_in6 *)addr;
12492 		if (IN6_IS_ADDR_V4MAPPED(&sin6->sin6_addr)) {
12493 			in6_sin6_2_sin(&sin, sin6);
12494 			addr_to_use = (struct sockaddr *)&sin;
12495 		}
12496 	}
12497 #endif
12498 	error = sctp_lower_sosend(so, addr_to_use, uio, top,
12499 	    control,
12500 	    flags,
12501 	    use_sndinfo ? &sndrcvninfo : NULL
12502 	    ,p
12503 	    );
12504 	return (error);
12505 }
12506 
12507 
12508 int
12509 sctp_lower_sosend(struct socket *so,
12510     struct sockaddr *addr,
12511     struct uio *uio,
12512     struct mbuf *i_pak,
12513     struct mbuf *control,
12514     int flags,
12515     struct sctp_sndrcvinfo *srcv
12516     ,
12517     struct thread *p
12518 )
12519 {
12520 	unsigned int sndlen = 0, max_len;
12521 	int error, len;
12522 	struct mbuf *top = NULL;
12523 	int queue_only = 0, queue_only_for_init = 0;
12524 	int free_cnt_applied = 0;
12525 	int un_sent;
12526 	int now_filled = 0;
12527 	unsigned int inqueue_bytes = 0;
12528 	struct sctp_block_entry be;
12529 	struct sctp_inpcb *inp;
12530 	struct sctp_tcb *stcb = NULL;
12531 	struct timeval now;
12532 	struct sctp_nets *net;
12533 	struct sctp_association *asoc;
12534 	struct sctp_inpcb *t_inp;
12535 	int user_marks_eor;
12536 	int create_lock_applied = 0;
12537 	int nagle_applies = 0;
12538 	int some_on_control = 0;
12539 	int got_all_of_the_send = 0;
12540 	int hold_tcblock = 0;
12541 	int non_blocking = 0;
12542 	uint32_t local_add_more, local_soresv = 0;
12543 	uint16_t port;
12544 	uint16_t sinfo_flags;
12545 	sctp_assoc_t sinfo_assoc_id;
12546 
12547 	error = 0;
12548 	net = NULL;
12549 	stcb = NULL;
12550 	asoc = NULL;
12551 
12552 	t_inp = inp = (struct sctp_inpcb *)so->so_pcb;
12553 	if (inp == NULL) {
12554 		SCTP_LTRACE_ERR_RET(NULL, NULL, NULL, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12555 		error = EINVAL;
12556 		if (i_pak) {
12557 			SCTP_RELEASE_PKT(i_pak);
12558 		}
12559 		return (error);
12560 	}
12561 	if ((uio == NULL) && (i_pak == NULL)) {
12562 		SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12563 		return (EINVAL);
12564 	}
12565 	user_marks_eor = sctp_is_feature_on(inp, SCTP_PCB_FLAGS_EXPLICIT_EOR);
12566 	atomic_add_int(&inp->total_sends, 1);
12567 	if (uio) {
12568 		if (uio->uio_resid < 0) {
12569 			SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12570 			return (EINVAL);
12571 		}
12572 		sndlen = (unsigned int)uio->uio_resid;
12573 	} else {
12574 		top = SCTP_HEADER_TO_CHAIN(i_pak);
12575 		sndlen = SCTP_HEADER_LEN(i_pak);
12576 	}
12577 	SCTPDBG(SCTP_DEBUG_OUTPUT1, "Send called addr:%p send length %d\n",
12578 	    (void *)addr,
12579 	    sndlen);
12580 	if ((inp->sctp_flags & SCTP_PCB_FLAGS_TCPTYPE) &&
12581 	    SCTP_IS_LISTENING(inp)) {
12582 		/* The listener can NOT send */
12583 		SCTP_LTRACE_ERR_RET(NULL, NULL, NULL, SCTP_FROM_SCTP_OUTPUT, ENOTCONN);
12584 		error = ENOTCONN;
12585 		goto out_unlocked;
12586 	}
12587 	/**
12588 	 * Pre-screen address, if one is given the sin-len
12589 	 * must be set correctly!
12590 	 */
12591 	if (addr) {
12592 		union sctp_sockstore *raddr = (union sctp_sockstore *)addr;
12593 
12594 		switch (raddr->sa.sa_family) {
12595 #ifdef INET
12596 		case AF_INET:
12597 			if (raddr->sin.sin_len != sizeof(struct sockaddr_in)) {
12598 				SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12599 				error = EINVAL;
12600 				goto out_unlocked;
12601 			}
12602 			port = raddr->sin.sin_port;
12603 			break;
12604 #endif
12605 #ifdef INET6
12606 		case AF_INET6:
12607 			if (raddr->sin6.sin6_len != sizeof(struct sockaddr_in6)) {
12608 				SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12609 				error = EINVAL;
12610 				goto out_unlocked;
12611 			}
12612 			port = raddr->sin6.sin6_port;
12613 			break;
12614 #endif
12615 		default:
12616 			SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EAFNOSUPPORT);
12617 			error = EAFNOSUPPORT;
12618 			goto out_unlocked;
12619 		}
12620 	} else
12621 		port = 0;
12622 
12623 	if (srcv) {
12624 		sinfo_flags = srcv->sinfo_flags;
12625 		sinfo_assoc_id = srcv->sinfo_assoc_id;
12626 		if (INVALID_SINFO_FLAG(sinfo_flags) ||
12627 		    PR_SCTP_INVALID_POLICY(sinfo_flags)) {
12628 			SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12629 			error = EINVAL;
12630 			goto out_unlocked;
12631 		}
12632 		if (srcv->sinfo_flags)
12633 			SCTP_STAT_INCR(sctps_sends_with_flags);
12634 	} else {
12635 		sinfo_flags = inp->def_send.sinfo_flags;
12636 		sinfo_assoc_id = inp->def_send.sinfo_assoc_id;
12637 	}
12638 	if (sinfo_flags & SCTP_SENDALL) {
12639 		/* its a sendall */
12640 		error = sctp_sendall(inp, uio, top, srcv);
12641 		top = NULL;
12642 		goto out_unlocked;
12643 	}
12644 	if ((sinfo_flags & SCTP_ADDR_OVER) && (addr == NULL)) {
12645 		SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12646 		error = EINVAL;
12647 		goto out_unlocked;
12648 	}
12649 	/* now we must find the assoc */
12650 	if ((inp->sctp_flags & SCTP_PCB_FLAGS_CONNECTED) ||
12651 	    (inp->sctp_flags & SCTP_PCB_FLAGS_IN_TCPPOOL)) {
12652 		SCTP_INP_RLOCK(inp);
12653 		stcb = LIST_FIRST(&inp->sctp_asoc_list);
12654 		if (stcb) {
12655 			SCTP_TCB_LOCK(stcb);
12656 			hold_tcblock = 1;
12657 		}
12658 		SCTP_INP_RUNLOCK(inp);
12659 	} else if (sinfo_assoc_id) {
12660 		stcb = sctp_findassociation_ep_asocid(inp, sinfo_assoc_id, 1);
12661 		if (stcb != NULL) {
12662 			hold_tcblock = 1;
12663 		}
12664 	} else if (addr) {
12665 		/*-
12666 		 * Since we did not use findep we must
12667 		 * increment it, and if we don't find a tcb
12668 		 * decrement it.
12669 		 */
12670 		SCTP_INP_WLOCK(inp);
12671 		SCTP_INP_INCR_REF(inp);
12672 		SCTP_INP_WUNLOCK(inp);
12673 		stcb = sctp_findassociation_ep_addr(&t_inp, addr, &net, NULL, NULL);
12674 		if (stcb == NULL) {
12675 			SCTP_INP_WLOCK(inp);
12676 			SCTP_INP_DECR_REF(inp);
12677 			SCTP_INP_WUNLOCK(inp);
12678 		} else {
12679 			hold_tcblock = 1;
12680 		}
12681 	}
12682 	if ((stcb == NULL) && (addr)) {
12683 		/* Possible implicit send? */
12684 		SCTP_ASOC_CREATE_LOCK(inp);
12685 		create_lock_applied = 1;
12686 		if ((inp->sctp_flags & SCTP_PCB_FLAGS_SOCKET_GONE) ||
12687 		    (inp->sctp_flags & SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
12688 			/* Should I really unlock ? */
12689 			SCTP_LTRACE_ERR_RET(NULL, NULL, NULL, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12690 			error = EINVAL;
12691 			goto out_unlocked;
12692 
12693 		}
12694 		if (((inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) == 0) &&
12695 		    (addr->sa_family == AF_INET6)) {
12696 			SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12697 			error = EINVAL;
12698 			goto out_unlocked;
12699 		}
12700 		SCTP_INP_WLOCK(inp);
12701 		SCTP_INP_INCR_REF(inp);
12702 		SCTP_INP_WUNLOCK(inp);
12703 		/* With the lock applied look again */
12704 		stcb = sctp_findassociation_ep_addr(&t_inp, addr, &net, NULL, NULL);
12705 		if ((stcb == NULL) && (control != NULL) && (port > 0)) {
12706 			stcb = sctp_findassociation_cmsgs(&t_inp, port, control, &net, &error);
12707 		}
12708 		if (stcb == NULL) {
12709 			SCTP_INP_WLOCK(inp);
12710 			SCTP_INP_DECR_REF(inp);
12711 			SCTP_INP_WUNLOCK(inp);
12712 		} else {
12713 			hold_tcblock = 1;
12714 		}
12715 		if (error) {
12716 			goto out_unlocked;
12717 		}
12718 		if (t_inp != inp) {
12719 			SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, ENOTCONN);
12720 			error = ENOTCONN;
12721 			goto out_unlocked;
12722 		}
12723 	}
12724 	if (stcb == NULL) {
12725 		if (addr == NULL) {
12726 			SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, ENOENT);
12727 			error = ENOENT;
12728 			goto out_unlocked;
12729 		} else {
12730 			/* We must go ahead and start the INIT process */
12731 			uint32_t vrf_id;
12732 
12733 			if ((sinfo_flags & SCTP_ABORT) ||
12734 			    ((sinfo_flags & SCTP_EOF) && (sndlen == 0))) {
12735 				/*-
12736 				 * User asks to abort a non-existant assoc,
12737 				 * or EOF a non-existant assoc with no data
12738 				 */
12739 				SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, ENOENT);
12740 				error = ENOENT;
12741 				goto out_unlocked;
12742 			}
12743 			/* get an asoc/stcb struct */
12744 			vrf_id = inp->def_vrf_id;
12745 #ifdef INVARIANTS
12746 			if (create_lock_applied == 0) {
12747 				panic("Error, should hold create lock and I don't?");
12748 			}
12749 #endif
12750 			stcb = sctp_aloc_assoc(inp, addr, &error, 0, vrf_id,
12751 			    inp->sctp_ep.pre_open_stream_count,
12752 			    inp->sctp_ep.port,
12753 			    p);
12754 			if (stcb == NULL) {
12755 				/* Error is setup for us in the call */
12756 				goto out_unlocked;
12757 			}
12758 			if (stcb->sctp_ep->sctp_flags & SCTP_PCB_FLAGS_TCPTYPE) {
12759 				stcb->sctp_ep->sctp_flags |= SCTP_PCB_FLAGS_CONNECTED;
12760 				/*
12761 				 * Set the connected flag so we can queue
12762 				 * data
12763 				 */
12764 				soisconnecting(so);
12765 			}
12766 			hold_tcblock = 1;
12767 			if (create_lock_applied) {
12768 				SCTP_ASOC_CREATE_UNLOCK(inp);
12769 				create_lock_applied = 0;
12770 			} else {
12771 				SCTP_PRINTF("Huh-3? create lock should have been on??\n");
12772 			}
12773 			/*
12774 			 * Turn on queue only flag to prevent data from
12775 			 * being sent
12776 			 */
12777 			queue_only = 1;
12778 			asoc = &stcb->asoc;
12779 			SCTP_SET_STATE(stcb, SCTP_STATE_COOKIE_WAIT);
12780 			(void)SCTP_GETTIME_TIMEVAL(&asoc->time_entered);
12781 
12782 			/* initialize authentication params for the assoc */
12783 			sctp_initialize_auth_params(inp, stcb);
12784 
12785 			if (control) {
12786 				if (sctp_process_cmsgs_for_init(stcb, control, &error)) {
12787 					sctp_free_assoc(inp, stcb, SCTP_PCBFREE_FORCE,
12788 					    SCTP_FROM_SCTP_OUTPUT + SCTP_LOC_5);
12789 					hold_tcblock = 0;
12790 					stcb = NULL;
12791 					goto out_unlocked;
12792 				}
12793 			}
12794 			/* out with the INIT */
12795 			queue_only_for_init = 1;
12796 			/*-
12797 			 * we may want to dig in after this call and adjust the MTU
12798 			 * value. It defaulted to 1500 (constant) but the ro
12799 			 * structure may now have an update and thus we may need to
12800 			 * change it BEFORE we append the message.
12801 			 */
12802 		}
12803 	} else
12804 		asoc = &stcb->asoc;
12805 	if (srcv == NULL)
12806 		srcv = (struct sctp_sndrcvinfo *)&asoc->def_send;
12807 	if (srcv->sinfo_flags & SCTP_ADDR_OVER) {
12808 		if (addr)
12809 			net = sctp_findnet(stcb, addr);
12810 		else
12811 			net = NULL;
12812 		if ((net == NULL) ||
12813 		    ((port != 0) && (port != stcb->rport))) {
12814 			SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12815 			error = EINVAL;
12816 			goto out_unlocked;
12817 		}
12818 	} else {
12819 		if (stcb->asoc.alternate) {
12820 			net = stcb->asoc.alternate;
12821 		} else {
12822 			net = stcb->asoc.primary_destination;
12823 		}
12824 	}
12825 	atomic_add_int(&stcb->total_sends, 1);
12826 	/* Keep the stcb from being freed under our feet */
12827 	atomic_add_int(&asoc->refcnt, 1);
12828 	free_cnt_applied = 1;
12829 
12830 	if (sctp_is_feature_on(inp, SCTP_PCB_FLAGS_NO_FRAGMENT)) {
12831 		if (sndlen > asoc->smallest_mtu) {
12832 			SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EMSGSIZE);
12833 			error = EMSGSIZE;
12834 			goto out_unlocked;
12835 		}
12836 	}
12837 	if (SCTP_SO_IS_NBIO(so)
12838 	    || (flags & MSG_NBIO)
12839 	    ) {
12840 		non_blocking = 1;
12841 	}
12842 	/* would we block? */
12843 	if (non_blocking) {
12844 		uint32_t amount;
12845 
12846 		if (hold_tcblock == 0) {
12847 			SCTP_TCB_LOCK(stcb);
12848 			hold_tcblock = 1;
12849 		}
12850 		inqueue_bytes = stcb->asoc.total_output_queue_size - (stcb->asoc.chunks_on_out_queue * SCTP_DATA_CHUNK_OVERHEAD(stcb));
12851 		if (user_marks_eor == 0) {
12852 			amount = sndlen;
12853 		} else {
12854 			amount = 1;
12855 		}
12856 		if ((SCTP_SB_LIMIT_SND(so) < (amount + inqueue_bytes + stcb->asoc.sb_send_resv)) ||
12857 		    (stcb->asoc.chunks_on_out_queue >= SCTP_BASE_SYSCTL(sctp_max_chunks_on_queue))) {
12858 			SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EWOULDBLOCK);
12859 			if (sndlen > SCTP_SB_LIMIT_SND(so))
12860 				error = EMSGSIZE;
12861 			else
12862 				error = EWOULDBLOCK;
12863 			goto out_unlocked;
12864 		}
12865 		stcb->asoc.sb_send_resv += sndlen;
12866 		SCTP_TCB_UNLOCK(stcb);
12867 		hold_tcblock = 0;
12868 	} else {
12869 		atomic_add_int(&stcb->asoc.sb_send_resv, sndlen);
12870 	}
12871 	local_soresv = sndlen;
12872 	if (stcb->asoc.state & SCTP_STATE_ABOUT_TO_BE_FREED) {
12873 		SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ECONNRESET);
12874 		error = ECONNRESET;
12875 		goto out_unlocked;
12876 	}
12877 	if (create_lock_applied) {
12878 		SCTP_ASOC_CREATE_UNLOCK(inp);
12879 		create_lock_applied = 0;
12880 	}
12881 	/* Is the stream no. valid? */
12882 	if (srcv->sinfo_stream >= asoc->streamoutcnt) {
12883 		/* Invalid stream number */
12884 		SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12885 		error = EINVAL;
12886 		goto out_unlocked;
12887 	}
12888 	if ((asoc->strmout[srcv->sinfo_stream].state != SCTP_STREAM_OPEN) &&
12889 	    (asoc->strmout[srcv->sinfo_stream].state != SCTP_STREAM_OPENING)) {
12890 		/*
12891 		 * Can't queue any data while stream reset is underway.
12892 		 */
12893 		if (asoc->strmout[srcv->sinfo_stream].state > SCTP_STREAM_OPEN) {
12894 			error = EAGAIN;
12895 		} else {
12896 			error = EINVAL;
12897 		}
12898 		SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, error);
12899 		goto out_unlocked;
12900 	}
12901 	if ((SCTP_GET_STATE(stcb) == SCTP_STATE_COOKIE_WAIT) ||
12902 	    (SCTP_GET_STATE(stcb) == SCTP_STATE_COOKIE_ECHOED)) {
12903 		queue_only = 1;
12904 	}
12905 	/* we are now done with all control */
12906 	if (control) {
12907 		sctp_m_freem(control);
12908 		control = NULL;
12909 	}
12910 	if ((SCTP_GET_STATE(stcb) == SCTP_STATE_SHUTDOWN_SENT) ||
12911 	    (SCTP_GET_STATE(stcb) == SCTP_STATE_SHUTDOWN_RECEIVED) ||
12912 	    (SCTP_GET_STATE(stcb) == SCTP_STATE_SHUTDOWN_ACK_SENT) ||
12913 	    (asoc->state & SCTP_STATE_SHUTDOWN_PENDING)) {
12914 		if (srcv->sinfo_flags & SCTP_ABORT) {
12915 			;
12916 		} else {
12917 			SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ECONNRESET);
12918 			error = ECONNRESET;
12919 			goto out_unlocked;
12920 		}
12921 	}
12922 	/* Ok, we will attempt a msgsnd :> */
12923 	if (p) {
12924 		p->td_ru.ru_msgsnd++;
12925 	}
12926 	/* Are we aborting? */
12927 	if (srcv->sinfo_flags & SCTP_ABORT) {
12928 		struct mbuf *mm;
12929 		int tot_demand, tot_out = 0, max_out;
12930 
12931 		SCTP_STAT_INCR(sctps_sends_with_abort);
12932 		if ((SCTP_GET_STATE(stcb) == SCTP_STATE_COOKIE_WAIT) ||
12933 		    (SCTP_GET_STATE(stcb) == SCTP_STATE_COOKIE_ECHOED)) {
12934 			/* It has to be up before we abort */
12935 			/* how big is the user initiated abort? */
12936 			SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
12937 			error = EINVAL;
12938 			goto out;
12939 		}
12940 		if (hold_tcblock) {
12941 			SCTP_TCB_UNLOCK(stcb);
12942 			hold_tcblock = 0;
12943 		}
12944 		if (top) {
12945 			struct mbuf *cntm = NULL;
12946 
12947 			mm = sctp_get_mbuf_for_msg(sizeof(struct sctp_paramhdr), 0, M_WAITOK, 1, MT_DATA);
12948 			if (sndlen != 0) {
12949 				for (cntm = top; cntm; cntm = SCTP_BUF_NEXT(cntm)) {
12950 					tot_out += SCTP_BUF_LEN(cntm);
12951 				}
12952 			}
12953 		} else {
12954 			/* Must fit in a MTU */
12955 			tot_out = sndlen;
12956 			tot_demand = (tot_out + sizeof(struct sctp_paramhdr));
12957 			if (tot_demand > SCTP_DEFAULT_ADD_MORE) {
12958 				/* To big */
12959 				SCTP_LTRACE_ERR_RET(NULL, stcb, net, SCTP_FROM_SCTP_OUTPUT, EMSGSIZE);
12960 				error = EMSGSIZE;
12961 				goto out;
12962 			}
12963 			mm = sctp_get_mbuf_for_msg(tot_demand, 0, M_WAITOK, 1, MT_DATA);
12964 		}
12965 		if (mm == NULL) {
12966 			SCTP_LTRACE_ERR_RET(NULL, stcb, net, SCTP_FROM_SCTP_OUTPUT, ENOMEM);
12967 			error = ENOMEM;
12968 			goto out;
12969 		}
12970 		max_out = asoc->smallest_mtu - sizeof(struct sctp_paramhdr);
12971 		max_out -= sizeof(struct sctp_abort_msg);
12972 		if (tot_out > max_out) {
12973 			tot_out = max_out;
12974 		}
12975 		if (mm) {
12976 			struct sctp_paramhdr *ph;
12977 
12978 			/* now move forward the data pointer */
12979 			ph = mtod(mm, struct sctp_paramhdr *);
12980 			ph->param_type = htons(SCTP_CAUSE_USER_INITIATED_ABT);
12981 			ph->param_length = htons((uint16_t)(sizeof(struct sctp_paramhdr) + tot_out));
12982 			ph++;
12983 			SCTP_BUF_LEN(mm) = tot_out + sizeof(struct sctp_paramhdr);
12984 			if (top == NULL) {
12985 				error = uiomove((caddr_t)ph, (int)tot_out, uio);
12986 				if (error) {
12987 					/*-
12988 					 * Here if we can't get his data we
12989 					 * still abort we just don't get to
12990 					 * send the users note :-0
12991 					 */
12992 					sctp_m_freem(mm);
12993 					mm = NULL;
12994 				}
12995 			} else {
12996 				if (sndlen != 0) {
12997 					SCTP_BUF_NEXT(mm) = top;
12998 				}
12999 			}
13000 		}
13001 		if (hold_tcblock == 0) {
13002 			SCTP_TCB_LOCK(stcb);
13003 		}
13004 		atomic_add_int(&stcb->asoc.refcnt, -1);
13005 		free_cnt_applied = 0;
13006 		/* release this lock, otherwise we hang on ourselves */
13007 		sctp_abort_an_association(stcb->sctp_ep, stcb, mm, SCTP_SO_LOCKED);
13008 		/* now relock the stcb so everything is sane */
13009 		hold_tcblock = 0;
13010 		stcb = NULL;
13011 		/*
13012 		 * In this case top is already chained to mm avoid double
13013 		 * free, since we free it below if top != NULL and driver
13014 		 * would free it after sending the packet out
13015 		 */
13016 		if (sndlen != 0) {
13017 			top = NULL;
13018 		}
13019 		goto out_unlocked;
13020 	}
13021 	/* Calculate the maximum we can send */
13022 	inqueue_bytes = stcb->asoc.total_output_queue_size - (stcb->asoc.chunks_on_out_queue * SCTP_DATA_CHUNK_OVERHEAD(stcb));
13023 	if (SCTP_SB_LIMIT_SND(so) > inqueue_bytes) {
13024 		if (non_blocking) {
13025 			/* we already checked for non-blocking above. */
13026 			max_len = sndlen;
13027 		} else {
13028 			max_len = SCTP_SB_LIMIT_SND(so) - inqueue_bytes;
13029 		}
13030 	} else {
13031 		max_len = 0;
13032 	}
13033 	if (hold_tcblock) {
13034 		SCTP_TCB_UNLOCK(stcb);
13035 		hold_tcblock = 0;
13036 	}
13037 	if (asoc->strmout == NULL) {
13038 		/* huh? software error */
13039 		SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EFAULT);
13040 		error = EFAULT;
13041 		goto out_unlocked;
13042 	}
13043 
13044 	/* Unless E_EOR mode is on, we must make a send FIT in one call. */
13045 	if ((user_marks_eor == 0) &&
13046 	    (sndlen > SCTP_SB_LIMIT_SND(stcb->sctp_socket))) {
13047 		/* It will NEVER fit */
13048 		SCTP_LTRACE_ERR_RET(NULL, stcb, net, SCTP_FROM_SCTP_OUTPUT, EMSGSIZE);
13049 		error = EMSGSIZE;
13050 		goto out_unlocked;
13051 	}
13052 	if ((uio == NULL) && user_marks_eor) {
13053 		/*-
13054 		 * We do not support eeor mode for
13055 		 * sending with mbuf chains (like sendfile).
13056 		 */
13057 		SCTP_LTRACE_ERR_RET(NULL, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
13058 		error = EINVAL;
13059 		goto out_unlocked;
13060 	}
13061 
13062 	if (user_marks_eor) {
13063 		local_add_more = min(SCTP_SB_LIMIT_SND(so), SCTP_BASE_SYSCTL(sctp_add_more_threshold));
13064 	} else {
13065 		/*-
13066 		 * For non-eeor the whole message must fit in
13067 		 * the socket send buffer.
13068 		 */
13069 		local_add_more = sndlen;
13070 	}
13071 	len = 0;
13072 	if (non_blocking) {
13073 		goto skip_preblock;
13074 	}
13075 	if (((max_len <= local_add_more) &&
13076 	    (SCTP_SB_LIMIT_SND(so) >= local_add_more)) ||
13077 	    (max_len == 0) ||
13078 	    ((stcb->asoc.chunks_on_out_queue + stcb->asoc.stream_queue_cnt) >= SCTP_BASE_SYSCTL(sctp_max_chunks_on_queue))) {
13079 		/* No room right now ! */
13080 		SOCKBUF_LOCK(&so->so_snd);
13081 		inqueue_bytes = stcb->asoc.total_output_queue_size - (stcb->asoc.chunks_on_out_queue * SCTP_DATA_CHUNK_OVERHEAD(stcb));
13082 		while ((SCTP_SB_LIMIT_SND(so) < (inqueue_bytes + local_add_more)) ||
13083 		    ((stcb->asoc.stream_queue_cnt + stcb->asoc.chunks_on_out_queue) >= SCTP_BASE_SYSCTL(sctp_max_chunks_on_queue))) {
13084 			SCTPDBG(SCTP_DEBUG_OUTPUT1, "pre_block limit:%u <(inq:%d + %d) || (%d+%d > %d)\n",
13085 			    (unsigned int)SCTP_SB_LIMIT_SND(so),
13086 			    inqueue_bytes,
13087 			    local_add_more,
13088 			    stcb->asoc.stream_queue_cnt,
13089 			    stcb->asoc.chunks_on_out_queue,
13090 			    SCTP_BASE_SYSCTL(sctp_max_chunks_on_queue));
13091 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_BLK_LOGGING_ENABLE) {
13092 				sctp_log_block(SCTP_BLOCK_LOG_INTO_BLKA, asoc, sndlen);
13093 			}
13094 			be.error = 0;
13095 			stcb->block_entry = &be;
13096 			error = sbwait(&so->so_snd);
13097 			stcb->block_entry = NULL;
13098 			if (error || so->so_error || be.error) {
13099 				if (error == 0) {
13100 					if (so->so_error)
13101 						error = so->so_error;
13102 					if (be.error) {
13103 						error = be.error;
13104 					}
13105 				}
13106 				SOCKBUF_UNLOCK(&so->so_snd);
13107 				goto out_unlocked;
13108 			}
13109 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_BLK_LOGGING_ENABLE) {
13110 				sctp_log_block(SCTP_BLOCK_LOG_OUTOF_BLK,
13111 				    asoc, stcb->asoc.total_output_queue_size);
13112 			}
13113 			if (stcb->asoc.state & SCTP_STATE_ABOUT_TO_BE_FREED) {
13114 				SOCKBUF_UNLOCK(&so->so_snd);
13115 				goto out_unlocked;
13116 			}
13117 			inqueue_bytes = stcb->asoc.total_output_queue_size - (stcb->asoc.chunks_on_out_queue * SCTP_DATA_CHUNK_OVERHEAD(stcb));
13118 		}
13119 		if (SCTP_SB_LIMIT_SND(so) > inqueue_bytes) {
13120 			max_len = SCTP_SB_LIMIT_SND(so) - inqueue_bytes;
13121 		} else {
13122 			max_len = 0;
13123 		}
13124 		SOCKBUF_UNLOCK(&so->so_snd);
13125 	}
13126 
13127 skip_preblock:
13128 	if (stcb->asoc.state & SCTP_STATE_ABOUT_TO_BE_FREED) {
13129 		goto out_unlocked;
13130 	}
13131 	/*
13132 	 * sndlen covers for mbuf case uio_resid covers for the non-mbuf
13133 	 * case NOTE: uio will be null when top/mbuf is passed
13134 	 */
13135 	if (sndlen == 0) {
13136 		if (srcv->sinfo_flags & SCTP_EOF) {
13137 			got_all_of_the_send = 1;
13138 			goto dataless_eof;
13139 		} else {
13140 			SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
13141 			error = EINVAL;
13142 			goto out;
13143 		}
13144 	}
13145 	if (top == NULL) {
13146 		struct sctp_stream_queue_pending *sp;
13147 		struct sctp_stream_out *strm;
13148 		uint32_t sndout;
13149 
13150 		SCTP_TCB_SEND_LOCK(stcb);
13151 		if ((asoc->stream_locked) &&
13152 		    (asoc->stream_locked_on != srcv->sinfo_stream)) {
13153 			SCTP_TCB_SEND_UNLOCK(stcb);
13154 			SCTP_LTRACE_ERR_RET(inp, stcb, net, SCTP_FROM_SCTP_OUTPUT, EINVAL);
13155 			error = EINVAL;
13156 			goto out;
13157 		}
13158 		SCTP_TCB_SEND_UNLOCK(stcb);
13159 
13160 		strm = &stcb->asoc.strmout[srcv->sinfo_stream];
13161 		if (strm->last_msg_incomplete == 0) {
13162 	do_a_copy_in:
13163 			sp = sctp_copy_it_in(stcb, asoc, srcv, uio, net, max_len, user_marks_eor, &error);
13164 			if (error) {
13165 				goto out;
13166 			}
13167 			SCTP_TCB_SEND_LOCK(stcb);
13168 			if (sp->msg_is_complete) {
13169 				strm->last_msg_incomplete = 0;
13170 				asoc->stream_locked = 0;
13171 			} else {
13172 				/*
13173 				 * Just got locked to this guy in case of an
13174 				 * interrupt.
13175 				 */
13176 				strm->last_msg_incomplete = 1;
13177 				if (stcb->asoc.idata_supported == 0) {
13178 					asoc->stream_locked = 1;
13179 					asoc->stream_locked_on = srcv->sinfo_stream;
13180 				}
13181 				sp->sender_all_done = 0;
13182 			}
13183 			sctp_snd_sb_alloc(stcb, sp->length);
13184 			atomic_add_int(&asoc->stream_queue_cnt, 1);
13185 			if (srcv->sinfo_flags & SCTP_UNORDERED) {
13186 				SCTP_STAT_INCR(sctps_sends_with_unord);
13187 			}
13188 			TAILQ_INSERT_TAIL(&strm->outqueue, sp, next);
13189 			stcb->asoc.ss_functions.sctp_ss_add_to_stream(stcb, asoc, strm, sp, 1);
13190 			SCTP_TCB_SEND_UNLOCK(stcb);
13191 		} else {
13192 			SCTP_TCB_SEND_LOCK(stcb);
13193 			sp = TAILQ_LAST(&strm->outqueue, sctp_streamhead);
13194 			SCTP_TCB_SEND_UNLOCK(stcb);
13195 			if (sp == NULL) {
13196 				/* ???? Huh ??? last msg is gone */
13197 #ifdef INVARIANTS
13198 				panic("Warning: Last msg marked incomplete, yet nothing left?");
13199 #else
13200 				SCTP_PRINTF("Warning: Last msg marked incomplete, yet nothing left?\n");
13201 				strm->last_msg_incomplete = 0;
13202 #endif
13203 				goto do_a_copy_in;
13204 
13205 			}
13206 		}
13207 		while (uio->uio_resid > 0) {
13208 			/* How much room do we have? */
13209 			struct mbuf *new_tail, *mm;
13210 
13211 			inqueue_bytes = stcb->asoc.total_output_queue_size - (stcb->asoc.chunks_on_out_queue * SCTP_DATA_CHUNK_OVERHEAD(stcb));
13212 			if (SCTP_SB_LIMIT_SND(so) > inqueue_bytes)
13213 				max_len = SCTP_SB_LIMIT_SND(so) - inqueue_bytes;
13214 			else
13215 				max_len = 0;
13216 
13217 			if ((max_len > SCTP_BASE_SYSCTL(sctp_add_more_threshold)) ||
13218 			    (max_len && (SCTP_SB_LIMIT_SND(so) < SCTP_BASE_SYSCTL(sctp_add_more_threshold))) ||
13219 			    (uio->uio_resid && (uio->uio_resid <= (int)max_len))) {
13220 				sndout = 0;
13221 				new_tail = NULL;
13222 				if (hold_tcblock) {
13223 					SCTP_TCB_UNLOCK(stcb);
13224 					hold_tcblock = 0;
13225 				}
13226 				mm = sctp_copy_resume(uio, max_len, user_marks_eor, &error, &sndout, &new_tail);
13227 				if ((mm == NULL) || error) {
13228 					if (mm) {
13229 						sctp_m_freem(mm);
13230 					}
13231 					goto out;
13232 				}
13233 				/* Update the mbuf and count */
13234 				SCTP_TCB_SEND_LOCK(stcb);
13235 				if (stcb->asoc.state & SCTP_STATE_ABOUT_TO_BE_FREED) {
13236 					/*
13237 					 * we need to get out. Peer probably
13238 					 * aborted.
13239 					 */
13240 					sctp_m_freem(mm);
13241 					if (stcb->asoc.state & SCTP_PCB_FLAGS_WAS_ABORTED) {
13242 						SCTP_LTRACE_ERR_RET(NULL, stcb, NULL, SCTP_FROM_SCTP_OUTPUT, ECONNRESET);
13243 						error = ECONNRESET;
13244 					}
13245 					SCTP_TCB_SEND_UNLOCK(stcb);
13246 					goto out;
13247 				}
13248 				if (sp->tail_mbuf) {
13249 					/* tack it to the end */
13250 					SCTP_BUF_NEXT(sp->tail_mbuf) = mm;
13251 					sp->tail_mbuf = new_tail;
13252 				} else {
13253 					/* A stolen mbuf */
13254 					sp->data = mm;
13255 					sp->tail_mbuf = new_tail;
13256 				}
13257 				sctp_snd_sb_alloc(stcb, sndout);
13258 				atomic_add_int(&sp->length, sndout);
13259 				len += sndout;
13260 				if (srcv->sinfo_flags & SCTP_SACK_IMMEDIATELY) {
13261 					sp->sinfo_flags |= SCTP_SACK_IMMEDIATELY;
13262 				}
13263 
13264 				/* Did we reach EOR? */
13265 				if ((uio->uio_resid == 0) &&
13266 				    ((user_marks_eor == 0) ||
13267 				    (srcv->sinfo_flags & SCTP_EOF) ||
13268 				    (user_marks_eor && (srcv->sinfo_flags & SCTP_EOR)))) {
13269 					sp->msg_is_complete = 1;
13270 				} else {
13271 					sp->msg_is_complete = 0;
13272 				}
13273 				SCTP_TCB_SEND_UNLOCK(stcb);
13274 			}
13275 			if (uio->uio_resid == 0) {
13276 				/* got it all? */
13277 				continue;
13278 			}
13279 			/* PR-SCTP? */
13280 			if ((asoc->prsctp_supported) && (asoc->sent_queue_cnt_removeable > 0)) {
13281 				/*
13282 				 * This is ugly but we must assure locking
13283 				 * order
13284 				 */
13285 				if (hold_tcblock == 0) {
13286 					SCTP_TCB_LOCK(stcb);
13287 					hold_tcblock = 1;
13288 				}
13289 				sctp_prune_prsctp(stcb, asoc, srcv, sndlen);
13290 				inqueue_bytes = stcb->asoc.total_output_queue_size - (stcb->asoc.chunks_on_out_queue * SCTP_DATA_CHUNK_OVERHEAD(stcb));
13291 				if (SCTP_SB_LIMIT_SND(so) > inqueue_bytes)
13292 					max_len = SCTP_SB_LIMIT_SND(so) - inqueue_bytes;
13293 				else
13294 					max_len = 0;
13295 				if (max_len > 0) {
13296 					continue;
13297 				}
13298 				SCTP_TCB_UNLOCK(stcb);
13299 				hold_tcblock = 0;
13300 			}
13301 			/* wait for space now */
13302 			if (non_blocking) {
13303 				/* Non-blocking io in place out */
13304 				goto skip_out_eof;
13305 			}
13306 			/* What about the INIT, send it maybe */
13307 			if (queue_only_for_init) {
13308 				if (hold_tcblock == 0) {
13309 					SCTP_TCB_LOCK(stcb);
13310 					hold_tcblock = 1;
13311 				}
13312 				if (SCTP_GET_STATE(stcb) == SCTP_STATE_OPEN) {
13313 					/* a collision took us forward? */
13314 					queue_only = 0;
13315 				} else {
13316 					sctp_send_initiate(inp, stcb, SCTP_SO_LOCKED);
13317 					SCTP_SET_STATE(stcb, SCTP_STATE_COOKIE_WAIT);
13318 					queue_only = 1;
13319 				}
13320 			}
13321 			if ((net->flight_size > net->cwnd) &&
13322 			    (asoc->sctp_cmt_on_off == 0)) {
13323 				SCTP_STAT_INCR(sctps_send_cwnd_avoid);
13324 				queue_only = 1;
13325 			} else if (asoc->ifp_had_enobuf) {
13326 				SCTP_STAT_INCR(sctps_ifnomemqueued);
13327 				if (net->flight_size > (2 * net->mtu)) {
13328 					queue_only = 1;
13329 				}
13330 				asoc->ifp_had_enobuf = 0;
13331 			}
13332 			un_sent = stcb->asoc.total_output_queue_size - stcb->asoc.total_flight;
13333 			if ((sctp_is_feature_off(inp, SCTP_PCB_FLAGS_NODELAY)) &&
13334 			    (stcb->asoc.total_flight > 0) &&
13335 			    (stcb->asoc.stream_queue_cnt < SCTP_MAX_DATA_BUNDLING) &&
13336 			    (un_sent < (int)(stcb->asoc.smallest_mtu - SCTP_MIN_OVERHEAD))) {
13337 
13338 				/*-
13339 				 * Ok, Nagle is set on and we have data outstanding.
13340 				 * Don't send anything and let SACKs drive out the
13341 				 * data unless we have a "full" segment to send.
13342 				 */
13343 				if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_NAGLE_LOGGING_ENABLE) {
13344 					sctp_log_nagle_event(stcb, SCTP_NAGLE_APPLIED);
13345 				}
13346 				SCTP_STAT_INCR(sctps_naglequeued);
13347 				nagle_applies = 1;
13348 			} else {
13349 				if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_NAGLE_LOGGING_ENABLE) {
13350 					if (sctp_is_feature_off(inp, SCTP_PCB_FLAGS_NODELAY))
13351 						sctp_log_nagle_event(stcb, SCTP_NAGLE_SKIPPED);
13352 				}
13353 				SCTP_STAT_INCR(sctps_naglesent);
13354 				nagle_applies = 0;
13355 			}
13356 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_BLK_LOGGING_ENABLE) {
13357 
13358 				sctp_misc_ints(SCTP_CWNDLOG_PRESEND, queue_only_for_init, queue_only,
13359 				    nagle_applies, un_sent);
13360 				sctp_misc_ints(SCTP_CWNDLOG_PRESEND, stcb->asoc.total_output_queue_size,
13361 				    stcb->asoc.total_flight,
13362 				    stcb->asoc.chunks_on_out_queue, stcb->asoc.total_flight_count);
13363 			}
13364 			if (queue_only_for_init)
13365 				queue_only_for_init = 0;
13366 			if ((queue_only == 0) && (nagle_applies == 0)) {
13367 				/*-
13368 				 * need to start chunk output
13369 				 * before blocking.. note that if
13370 				 * a lock is already applied, then
13371 				 * the input via the net is happening
13372 				 * and I don't need to start output :-D
13373 				 */
13374 				if (hold_tcblock == 0) {
13375 					if (SCTP_TCB_TRYLOCK(stcb)) {
13376 						hold_tcblock = 1;
13377 						sctp_chunk_output(inp,
13378 						    stcb,
13379 						    SCTP_OUTPUT_FROM_USR_SEND, SCTP_SO_LOCKED);
13380 					}
13381 				} else {
13382 					sctp_chunk_output(inp,
13383 					    stcb,
13384 					    SCTP_OUTPUT_FROM_USR_SEND, SCTP_SO_LOCKED);
13385 				}
13386 				if (hold_tcblock == 1) {
13387 					SCTP_TCB_UNLOCK(stcb);
13388 					hold_tcblock = 0;
13389 				}
13390 			}
13391 			SOCKBUF_LOCK(&so->so_snd);
13392 			/*-
13393 			 * This is a bit strange, but I think it will
13394 			 * work. The total_output_queue_size is locked and
13395 			 * protected by the TCB_LOCK, which we just released.
13396 			 * There is a race that can occur between releasing it
13397 			 * above, and me getting the socket lock, where sacks
13398 			 * come in but we have not put the SB_WAIT on the
13399 			 * so_snd buffer to get the wakeup. After the LOCK
13400 			 * is applied the sack_processing will also need to
13401 			 * LOCK the so->so_snd to do the actual sowwakeup(). So
13402 			 * once we have the socket buffer lock if we recheck the
13403 			 * size we KNOW we will get to sleep safely with the
13404 			 * wakeup flag in place.
13405 			 */
13406 			inqueue_bytes = stcb->asoc.total_output_queue_size - (stcb->asoc.chunks_on_out_queue * SCTP_DATA_CHUNK_OVERHEAD(stcb));
13407 			if (SCTP_SB_LIMIT_SND(so) <= (inqueue_bytes +
13408 			    min(SCTP_BASE_SYSCTL(sctp_add_more_threshold), SCTP_SB_LIMIT_SND(so)))) {
13409 				if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_BLK_LOGGING_ENABLE) {
13410 					sctp_log_block(SCTP_BLOCK_LOG_INTO_BLK,
13411 					    asoc, (size_t)uio->uio_resid);
13412 				}
13413 				be.error = 0;
13414 				stcb->block_entry = &be;
13415 				error = sbwait(&so->so_snd);
13416 				stcb->block_entry = NULL;
13417 
13418 				if (error || so->so_error || be.error) {
13419 					if (error == 0) {
13420 						if (so->so_error)
13421 							error = so->so_error;
13422 						if (be.error) {
13423 							error = be.error;
13424 						}
13425 					}
13426 					SOCKBUF_UNLOCK(&so->so_snd);
13427 					goto out_unlocked;
13428 				}
13429 
13430 				if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_BLK_LOGGING_ENABLE) {
13431 					sctp_log_block(SCTP_BLOCK_LOG_OUTOF_BLK,
13432 					    asoc, stcb->asoc.total_output_queue_size);
13433 				}
13434 			}
13435 			SOCKBUF_UNLOCK(&so->so_snd);
13436 			if (stcb->asoc.state & SCTP_STATE_ABOUT_TO_BE_FREED) {
13437 				goto out_unlocked;
13438 			}
13439 		}
13440 		SCTP_TCB_SEND_LOCK(stcb);
13441 		if (stcb->asoc.state & SCTP_STATE_ABOUT_TO_BE_FREED) {
13442 			SCTP_TCB_SEND_UNLOCK(stcb);
13443 			goto out_unlocked;
13444 		}
13445 		if (sp) {
13446 			if (sp->msg_is_complete == 0) {
13447 				strm->last_msg_incomplete = 1;
13448 				if (stcb->asoc.idata_supported == 0) {
13449 					asoc->stream_locked = 1;
13450 					asoc->stream_locked_on = srcv->sinfo_stream;
13451 				}
13452 			} else {
13453 				sp->sender_all_done = 1;
13454 				strm->last_msg_incomplete = 0;
13455 				asoc->stream_locked = 0;
13456 			}
13457 		} else {
13458 			SCTP_PRINTF("Huh no sp TSNH?\n");
13459 			strm->last_msg_incomplete = 0;
13460 			asoc->stream_locked = 0;
13461 		}
13462 		SCTP_TCB_SEND_UNLOCK(stcb);
13463 		if (uio->uio_resid == 0) {
13464 			got_all_of_the_send = 1;
13465 		}
13466 	} else {
13467 		/* We send in a 0, since we do NOT have any locks */
13468 		error = sctp_msg_append(stcb, net, top, srcv, 0);
13469 		top = NULL;
13470 		if (srcv->sinfo_flags & SCTP_EOF) {
13471 			/*
13472 			 * This should only happen for Panda for the mbuf
13473 			 * send case, which does NOT yet support EEOR mode.
13474 			 * Thus, we can just set this flag to do the proper
13475 			 * EOF handling.
13476 			 */
13477 			got_all_of_the_send = 1;
13478 		}
13479 	}
13480 	if (error) {
13481 		goto out;
13482 	}
13483 dataless_eof:
13484 	/* EOF thing ? */
13485 	if ((srcv->sinfo_flags & SCTP_EOF) &&
13486 	    (got_all_of_the_send == 1)) {
13487 		SCTP_STAT_INCR(sctps_sends_with_eof);
13488 		error = 0;
13489 		if (hold_tcblock == 0) {
13490 			SCTP_TCB_LOCK(stcb);
13491 			hold_tcblock = 1;
13492 		}
13493 		if (TAILQ_EMPTY(&asoc->send_queue) &&
13494 		    TAILQ_EMPTY(&asoc->sent_queue) &&
13495 		    sctp_is_there_unsent_data(stcb, SCTP_SO_LOCKED) == 0) {
13496 			if ((*asoc->ss_functions.sctp_ss_is_user_msgs_incomplete) (stcb, asoc)) {
13497 				goto abort_anyway;
13498 			}
13499 			/* there is nothing queued to send, so I'm done... */
13500 			if ((SCTP_GET_STATE(stcb) != SCTP_STATE_SHUTDOWN_SENT) &&
13501 			    (SCTP_GET_STATE(stcb) != SCTP_STATE_SHUTDOWN_RECEIVED) &&
13502 			    (SCTP_GET_STATE(stcb) != SCTP_STATE_SHUTDOWN_ACK_SENT)) {
13503 				struct sctp_nets *netp;
13504 
13505 				/* only send SHUTDOWN the first time through */
13506 				if (SCTP_GET_STATE(stcb) == SCTP_STATE_OPEN) {
13507 					SCTP_STAT_DECR_GAUGE32(sctps_currestab);
13508 				}
13509 				SCTP_SET_STATE(stcb, SCTP_STATE_SHUTDOWN_SENT);
13510 				SCTP_CLEAR_SUBSTATE(stcb, SCTP_STATE_SHUTDOWN_PENDING);
13511 				sctp_stop_timers_for_shutdown(stcb);
13512 				if (stcb->asoc.alternate) {
13513 					netp = stcb->asoc.alternate;
13514 				} else {
13515 					netp = stcb->asoc.primary_destination;
13516 				}
13517 				sctp_send_shutdown(stcb, netp);
13518 				sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWN, stcb->sctp_ep, stcb,
13519 				    netp);
13520 				sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWNGUARD, stcb->sctp_ep, stcb,
13521 				    asoc->primary_destination);
13522 			}
13523 		} else {
13524 			/*-
13525 			 * we still got (or just got) data to send, so set
13526 			 * SHUTDOWN_PENDING
13527 			 */
13528 			/*-
13529 			 * XXX sockets draft says that SCTP_EOF should be
13530 			 * sent with no data.  currently, we will allow user
13531 			 * data to be sent first and move to
13532 			 * SHUTDOWN-PENDING
13533 			 */
13534 			if ((SCTP_GET_STATE(stcb) != SCTP_STATE_SHUTDOWN_SENT) &&
13535 			    (SCTP_GET_STATE(stcb) != SCTP_STATE_SHUTDOWN_RECEIVED) &&
13536 			    (SCTP_GET_STATE(stcb) != SCTP_STATE_SHUTDOWN_ACK_SENT)) {
13537 				if (hold_tcblock == 0) {
13538 					SCTP_TCB_LOCK(stcb);
13539 					hold_tcblock = 1;
13540 				}
13541 				if ((*asoc->ss_functions.sctp_ss_is_user_msgs_incomplete) (stcb, asoc)) {
13542 					SCTP_ADD_SUBSTATE(stcb, SCTP_STATE_PARTIAL_MSG_LEFT);
13543 				}
13544 				SCTP_ADD_SUBSTATE(stcb, SCTP_STATE_SHUTDOWN_PENDING);
13545 				if (TAILQ_EMPTY(&asoc->send_queue) &&
13546 				    TAILQ_EMPTY(&asoc->sent_queue) &&
13547 				    (asoc->state & SCTP_STATE_PARTIAL_MSG_LEFT)) {
13548 					struct mbuf *op_err;
13549 					char msg[SCTP_DIAG_INFO_LEN];
13550 
13551 			abort_anyway:
13552 					if (free_cnt_applied) {
13553 						atomic_add_int(&stcb->asoc.refcnt, -1);
13554 						free_cnt_applied = 0;
13555 					}
13556 					snprintf(msg, sizeof(msg),
13557 					    "%s:%d at %s", __FILE__, __LINE__, __func__);
13558 					op_err = sctp_generate_cause(SCTP_BASE_SYSCTL(sctp_diag_info_code),
13559 					    msg);
13560 					sctp_abort_an_association(stcb->sctp_ep, stcb,
13561 					    op_err, SCTP_SO_LOCKED);
13562 					/*
13563 					 * now relock the stcb so everything
13564 					 * is sane
13565 					 */
13566 					hold_tcblock = 0;
13567 					stcb = NULL;
13568 					goto out;
13569 				}
13570 				sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWNGUARD, stcb->sctp_ep, stcb,
13571 				    asoc->primary_destination);
13572 				sctp_feature_off(inp, SCTP_PCB_FLAGS_NODELAY);
13573 			}
13574 		}
13575 	}
13576 skip_out_eof:
13577 	if (!TAILQ_EMPTY(&stcb->asoc.control_send_queue)) {
13578 		some_on_control = 1;
13579 	}
13580 	if (queue_only_for_init) {
13581 		if (hold_tcblock == 0) {
13582 			SCTP_TCB_LOCK(stcb);
13583 			hold_tcblock = 1;
13584 		}
13585 		if (SCTP_GET_STATE(stcb) == SCTP_STATE_OPEN) {
13586 			/* a collision took us forward? */
13587 			queue_only = 0;
13588 		} else {
13589 			sctp_send_initiate(inp, stcb, SCTP_SO_LOCKED);
13590 			SCTP_SET_STATE(stcb, SCTP_STATE_COOKIE_WAIT);
13591 			queue_only = 1;
13592 		}
13593 	}
13594 	if ((net->flight_size > net->cwnd) &&
13595 	    (stcb->asoc.sctp_cmt_on_off == 0)) {
13596 		SCTP_STAT_INCR(sctps_send_cwnd_avoid);
13597 		queue_only = 1;
13598 	} else if (asoc->ifp_had_enobuf) {
13599 		SCTP_STAT_INCR(sctps_ifnomemqueued);
13600 		if (net->flight_size > (2 * net->mtu)) {
13601 			queue_only = 1;
13602 		}
13603 		asoc->ifp_had_enobuf = 0;
13604 	}
13605 	un_sent = stcb->asoc.total_output_queue_size - stcb->asoc.total_flight;
13606 	if ((sctp_is_feature_off(inp, SCTP_PCB_FLAGS_NODELAY)) &&
13607 	    (stcb->asoc.total_flight > 0) &&
13608 	    (stcb->asoc.stream_queue_cnt < SCTP_MAX_DATA_BUNDLING) &&
13609 	    (un_sent < (int)(stcb->asoc.smallest_mtu - SCTP_MIN_OVERHEAD))) {
13610 		/*-
13611 		 * Ok, Nagle is set on and we have data outstanding.
13612 		 * Don't send anything and let SACKs drive out the
13613 		 * data unless wen have a "full" segment to send.
13614 		 */
13615 		if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_NAGLE_LOGGING_ENABLE) {
13616 			sctp_log_nagle_event(stcb, SCTP_NAGLE_APPLIED);
13617 		}
13618 		SCTP_STAT_INCR(sctps_naglequeued);
13619 		nagle_applies = 1;
13620 	} else {
13621 		if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_NAGLE_LOGGING_ENABLE) {
13622 			if (sctp_is_feature_off(inp, SCTP_PCB_FLAGS_NODELAY))
13623 				sctp_log_nagle_event(stcb, SCTP_NAGLE_SKIPPED);
13624 		}
13625 		SCTP_STAT_INCR(sctps_naglesent);
13626 		nagle_applies = 0;
13627 	}
13628 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_BLK_LOGGING_ENABLE) {
13629 		sctp_misc_ints(SCTP_CWNDLOG_PRESEND, queue_only_for_init, queue_only,
13630 		    nagle_applies, un_sent);
13631 		sctp_misc_ints(SCTP_CWNDLOG_PRESEND, stcb->asoc.total_output_queue_size,
13632 		    stcb->asoc.total_flight,
13633 		    stcb->asoc.chunks_on_out_queue, stcb->asoc.total_flight_count);
13634 	}
13635 	if ((queue_only == 0) && (nagle_applies == 0) && (stcb->asoc.peers_rwnd && un_sent)) {
13636 		/* we can attempt to send too. */
13637 		if (hold_tcblock == 0) {
13638 			/*
13639 			 * If there is activity recv'ing sacks no need to
13640 			 * send
13641 			 */
13642 			if (SCTP_TCB_TRYLOCK(stcb)) {
13643 				sctp_chunk_output(inp, stcb, SCTP_OUTPUT_FROM_USR_SEND, SCTP_SO_LOCKED);
13644 				hold_tcblock = 1;
13645 			}
13646 		} else {
13647 			sctp_chunk_output(inp, stcb, SCTP_OUTPUT_FROM_USR_SEND, SCTP_SO_LOCKED);
13648 		}
13649 	} else if ((queue_only == 0) &&
13650 		    (stcb->asoc.peers_rwnd == 0) &&
13651 	    (stcb->asoc.total_flight == 0)) {
13652 		/* We get to have a probe outstanding */
13653 		if (hold_tcblock == 0) {
13654 			hold_tcblock = 1;
13655 			SCTP_TCB_LOCK(stcb);
13656 		}
13657 		sctp_chunk_output(inp, stcb, SCTP_OUTPUT_FROM_USR_SEND, SCTP_SO_LOCKED);
13658 	} else if (some_on_control) {
13659 		int num_out, reason, frag_point;
13660 
13661 		/* Here we do control only */
13662 		if (hold_tcblock == 0) {
13663 			hold_tcblock = 1;
13664 			SCTP_TCB_LOCK(stcb);
13665 		}
13666 		frag_point = sctp_get_frag_point(stcb, &stcb->asoc);
13667 		(void)sctp_med_chunk_output(inp, stcb, &stcb->asoc, &num_out,
13668 		    &reason, 1, 1, &now, &now_filled, frag_point, SCTP_SO_LOCKED);
13669 	}
13670 	SCTPDBG(SCTP_DEBUG_OUTPUT1, "USR Send complete qo:%d prw:%d unsent:%d tf:%d cooq:%d toqs:%d err:%d\n",
13671 	    queue_only, stcb->asoc.peers_rwnd, un_sent,
13672 	    stcb->asoc.total_flight, stcb->asoc.chunks_on_out_queue,
13673 	    stcb->asoc.total_output_queue_size, error);
13674 
13675 out:
13676 out_unlocked:
13677 
13678 	if (local_soresv && stcb) {
13679 		atomic_subtract_int(&stcb->asoc.sb_send_resv, sndlen);
13680 	}
13681 	if (create_lock_applied) {
13682 		SCTP_ASOC_CREATE_UNLOCK(inp);
13683 	}
13684 	if ((stcb) && hold_tcblock) {
13685 		SCTP_TCB_UNLOCK(stcb);
13686 	}
13687 	if (stcb && free_cnt_applied) {
13688 		atomic_add_int(&stcb->asoc.refcnt, -1);
13689 	}
13690 #ifdef INVARIANTS
13691 	if (stcb) {
13692 		if (mtx_owned(&stcb->tcb_mtx)) {
13693 			panic("Leaving with tcb mtx owned?");
13694 		}
13695 		if (mtx_owned(&stcb->tcb_send_mtx)) {
13696 			panic("Leaving with tcb send mtx owned?");
13697 		}
13698 	}
13699 #endif
13700 	if (top) {
13701 		sctp_m_freem(top);
13702 	}
13703 	if (control) {
13704 		sctp_m_freem(control);
13705 	}
13706 	return (error);
13707 }
13708 
13709 
13710 /*
13711  * generate an AUTHentication chunk, if required
13712  */
13713 struct mbuf *
13714 sctp_add_auth_chunk(struct mbuf *m, struct mbuf **m_end,
13715     struct sctp_auth_chunk **auth_ret, uint32_t *offset,
13716     struct sctp_tcb *stcb, uint8_t chunk)
13717 {
13718 	struct mbuf *m_auth;
13719 	struct sctp_auth_chunk *auth;
13720 	int chunk_len;
13721 	struct mbuf *cn;
13722 
13723 	if ((m_end == NULL) || (auth_ret == NULL) || (offset == NULL) ||
13724 	    (stcb == NULL))
13725 		return (m);
13726 
13727 	if (stcb->asoc.auth_supported == 0) {
13728 		return (m);
13729 	}
13730 	/* does the requested chunk require auth? */
13731 	if (!sctp_auth_is_required_chunk(chunk, stcb->asoc.peer_auth_chunks)) {
13732 		return (m);
13733 	}
13734 	m_auth = sctp_get_mbuf_for_msg(sizeof(*auth), 0, M_NOWAIT, 1, MT_HEADER);
13735 	if (m_auth == NULL) {
13736 		/* no mbuf's */
13737 		return (m);
13738 	}
13739 	/* reserve some space if this will be the first mbuf */
13740 	if (m == NULL)
13741 		SCTP_BUF_RESV_UF(m_auth, SCTP_MIN_OVERHEAD);
13742 	/* fill in the AUTH chunk details */
13743 	auth = mtod(m_auth, struct sctp_auth_chunk *);
13744 	memset(auth, 0, sizeof(*auth));
13745 	auth->ch.chunk_type = SCTP_AUTHENTICATION;
13746 	auth->ch.chunk_flags = 0;
13747 	chunk_len = sizeof(*auth) +
13748 	    sctp_get_hmac_digest_len(stcb->asoc.peer_hmac_id);
13749 	auth->ch.chunk_length = htons(chunk_len);
13750 	auth->hmac_id = htons(stcb->asoc.peer_hmac_id);
13751 	/* key id and hmac digest will be computed and filled in upon send */
13752 
13753 	/* save the offset where the auth was inserted into the chain */
13754 	*offset = 0;
13755 	for (cn = m; cn; cn = SCTP_BUF_NEXT(cn)) {
13756 		*offset += SCTP_BUF_LEN(cn);
13757 	}
13758 
13759 	/* update length and return pointer to the auth chunk */
13760 	SCTP_BUF_LEN(m_auth) = chunk_len;
13761 	m = sctp_copy_mbufchain(m_auth, m, m_end, 1, chunk_len, 0);
13762 	if (auth_ret != NULL)
13763 		*auth_ret = auth;
13764 
13765 	return (m);
13766 }
13767 
13768 #ifdef INET6
13769 int
13770 sctp_v6src_match_nexthop(struct sockaddr_in6 *src6, sctp_route_t *ro)
13771 {
13772 	struct nd_prefix *pfx = NULL;
13773 	struct nd_pfxrouter *pfxrtr = NULL;
13774 	struct sockaddr_in6 gw6;
13775 
13776 	if (ro == NULL || ro->ro_rt == NULL || src6->sin6_family != AF_INET6)
13777 		return (0);
13778 
13779 	/* get prefix entry of address */
13780 	ND6_RLOCK();
13781 	LIST_FOREACH(pfx, &MODULE_GLOBAL(nd_prefix), ndpr_entry) {
13782 		if (pfx->ndpr_stateflags & NDPRF_DETACHED)
13783 			continue;
13784 		if (IN6_ARE_MASKED_ADDR_EQUAL(&pfx->ndpr_prefix.sin6_addr,
13785 		    &src6->sin6_addr, &pfx->ndpr_mask))
13786 			break;
13787 	}
13788 	/* no prefix entry in the prefix list */
13789 	if (pfx == NULL) {
13790 		ND6_RUNLOCK();
13791 		SCTPDBG(SCTP_DEBUG_OUTPUT2, "No prefix entry for ");
13792 		SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT2, (struct sockaddr *)src6);
13793 		return (0);
13794 	}
13795 
13796 	SCTPDBG(SCTP_DEBUG_OUTPUT2, "v6src_match_nexthop(), Prefix entry is ");
13797 	SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT2, (struct sockaddr *)src6);
13798 
13799 	/* search installed gateway from prefix entry */
13800 	LIST_FOREACH(pfxrtr, &pfx->ndpr_advrtrs, pfr_entry) {
13801 		memset(&gw6, 0, sizeof(struct sockaddr_in6));
13802 		gw6.sin6_family = AF_INET6;
13803 		gw6.sin6_len = sizeof(struct sockaddr_in6);
13804 		memcpy(&gw6.sin6_addr, &pfxrtr->router->rtaddr,
13805 		    sizeof(struct in6_addr));
13806 		SCTPDBG(SCTP_DEBUG_OUTPUT2, "prefix router is ");
13807 		SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT2, (struct sockaddr *)&gw6);
13808 		SCTPDBG(SCTP_DEBUG_OUTPUT2, "installed router is ");
13809 		SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT2, ro->ro_rt->rt_gateway);
13810 		if (sctp_cmpaddr((struct sockaddr *)&gw6, ro->ro_rt->rt_gateway)) {
13811 			ND6_RUNLOCK();
13812 			SCTPDBG(SCTP_DEBUG_OUTPUT2, "pfxrouter is installed\n");
13813 			return (1);
13814 		}
13815 	}
13816 	ND6_RUNLOCK();
13817 	SCTPDBG(SCTP_DEBUG_OUTPUT2, "pfxrouter is not installed\n");
13818 	return (0);
13819 }
13820 #endif
13821 
13822 int
13823 sctp_v4src_match_nexthop(struct sctp_ifa *sifa, sctp_route_t *ro)
13824 {
13825 #ifdef INET
13826 	struct sockaddr_in *sin, *mask;
13827 	struct ifaddr *ifa;
13828 	struct in_addr srcnetaddr, gwnetaddr;
13829 
13830 	if (ro == NULL || ro->ro_rt == NULL ||
13831 	    sifa->address.sa.sa_family != AF_INET) {
13832 		return (0);
13833 	}
13834 	ifa = (struct ifaddr *)sifa->ifa;
13835 	mask = (struct sockaddr_in *)(ifa->ifa_netmask);
13836 	sin = &sifa->address.sin;
13837 	srcnetaddr.s_addr = (sin->sin_addr.s_addr & mask->sin_addr.s_addr);
13838 	SCTPDBG(SCTP_DEBUG_OUTPUT1, "match_nexthop4: src address is ");
13839 	SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT2, &sifa->address.sa);
13840 	SCTPDBG(SCTP_DEBUG_OUTPUT1, "network address is %x\n", srcnetaddr.s_addr);
13841 
13842 	sin = (struct sockaddr_in *)ro->ro_rt->rt_gateway;
13843 	gwnetaddr.s_addr = (sin->sin_addr.s_addr & mask->sin_addr.s_addr);
13844 	SCTPDBG(SCTP_DEBUG_OUTPUT1, "match_nexthop4: nexthop is ");
13845 	SCTPDBG_ADDR(SCTP_DEBUG_OUTPUT2, ro->ro_rt->rt_gateway);
13846 	SCTPDBG(SCTP_DEBUG_OUTPUT1, "network address is %x\n", gwnetaddr.s_addr);
13847 	if (srcnetaddr.s_addr == gwnetaddr.s_addr) {
13848 		return (1);
13849 	}
13850 #endif
13851 	return (0);
13852 }
13853