xref: /freebsd/sys/netinet/sctp_timer.c (revision c0020399a650364d0134f79f3fa319f84064372d)
1 /*-
2  * Copyright (c) 2001-2007, by Cisco Systems, Inc. All rights reserved.
3  *
4  * Redistribution and use in source and binary forms, with or without
5  * modification, are permitted provided that the following conditions are met:
6  *
7  * a) Redistributions of source code must retain the above copyright notice,
8  *   this list of conditions and the following disclaimer.
9  *
10  * b) Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in
12  *   the documentation and/or other materials provided with the distribution.
13  *
14  * c) Neither the name of Cisco Systems, Inc. nor the names of its
15  *    contributors may be used to endorse or promote products derived
16  *    from this software without specific prior written permission.
17  *
18  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
19  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
20  * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
21  * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
22  * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
23  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
24  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
25  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
26  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
27  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
28  * THE POSSIBILITY OF SUCH DAMAGE.
29  */
30 
31 /* $KAME: sctp_timer.c,v 1.29 2005/03/06 16:04:18 itojun Exp $	 */
32 
33 #include <sys/cdefs.h>
34 __FBSDID("$FreeBSD$");
35 
36 #define _IP_VHL
37 #include <netinet/sctp_os.h>
38 #include <netinet/sctp_pcb.h>
39 #ifdef INET6
40 #endif
41 #include <netinet/sctp_var.h>
42 #include <netinet/sctp_sysctl.h>
43 #include <netinet/sctp_timer.h>
44 #include <netinet/sctputil.h>
45 #include <netinet/sctp_output.h>
46 #include <netinet/sctp_header.h>
47 #include <netinet/sctp_indata.h>
48 #include <netinet/sctp_asconf.h>
49 #include <netinet/sctp_input.h>
50 #include <netinet/sctp.h>
51 #include <netinet/sctp_uio.h>
52 #include <netinet/udp.h>
53 
54 
55 void
56 sctp_early_fr_timer(struct sctp_inpcb *inp,
57     struct sctp_tcb *stcb,
58     struct sctp_nets *net)
59 {
60 	struct sctp_tmit_chunk *chk, *tp2;
61 	struct timeval now, min_wait, tv;
62 	unsigned int cur_rtt, cnt = 0, cnt_resend = 0;
63 
64 	/* an early FR is occuring. */
65 	(void)SCTP_GETTIME_TIMEVAL(&now);
66 	/* get cur rto in micro-seconds */
67 	if (net->lastsa == 0) {
68 		/* Hmm no rtt estimate yet? */
69 		cur_rtt = stcb->asoc.initial_rto >> 2;
70 	} else {
71 
72 		cur_rtt = ((net->lastsa >> 2) + net->lastsv) >> 1;
73 	}
74 	if (cur_rtt < SCTP_BASE_SYSCTL(sctp_early_fr_msec)) {
75 		cur_rtt = SCTP_BASE_SYSCTL(sctp_early_fr_msec);
76 	}
77 	cur_rtt *= 1000;
78 	tv.tv_sec = cur_rtt / 1000000;
79 	tv.tv_usec = cur_rtt % 1000000;
80 	min_wait = now;
81 	timevalsub(&min_wait, &tv);
82 	if (min_wait.tv_sec < 0 || min_wait.tv_usec < 0) {
83 		/*
84 		 * if we hit here, we don't have enough seconds on the clock
85 		 * to account for the RTO. We just let the lower seconds be
86 		 * the bounds and don't worry about it. This may mean we
87 		 * will mark a lot more than we should.
88 		 */
89 		min_wait.tv_sec = min_wait.tv_usec = 0;
90 	}
91 	chk = TAILQ_LAST(&stcb->asoc.sent_queue, sctpchunk_listhead);
92 	for (; chk != NULL; chk = tp2) {
93 		tp2 = TAILQ_PREV(chk, sctpchunk_listhead, sctp_next);
94 		if (chk->whoTo != net) {
95 			continue;
96 		}
97 		if (chk->sent == SCTP_DATAGRAM_RESEND)
98 			cnt_resend++;
99 		else if ((chk->sent > SCTP_DATAGRAM_UNSENT) &&
100 		    (chk->sent < SCTP_DATAGRAM_RESEND)) {
101 			/* pending, may need retran */
102 			if (chk->sent_rcv_time.tv_sec > min_wait.tv_sec) {
103 				/*
104 				 * we have reached a chunk that was sent
105 				 * some seconds past our min.. forget it we
106 				 * will find no more to send.
107 				 */
108 				continue;
109 			} else if (chk->sent_rcv_time.tv_sec == min_wait.tv_sec) {
110 				/*
111 				 * we must look at the micro seconds to
112 				 * know.
113 				 */
114 				if (chk->sent_rcv_time.tv_usec >= min_wait.tv_usec) {
115 					/*
116 					 * ok it was sent after our boundary
117 					 * time.
118 					 */
119 					continue;
120 				}
121 			}
122 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_EARLYFR_LOGGING_ENABLE) {
123 				sctp_log_fr(chk->rec.data.TSN_seq, chk->snd_count,
124 				    4, SCTP_FR_MARKED_EARLY);
125 			}
126 			SCTP_STAT_INCR(sctps_earlyfrmrkretrans);
127 			chk->sent = SCTP_DATAGRAM_RESEND;
128 			sctp_ucount_incr(stcb->asoc.sent_queue_retran_cnt);
129 			/* double book size since we are doing an early FR */
130 			chk->book_size_scale++;
131 			cnt += chk->send_size;
132 			if ((cnt + net->flight_size) > net->cwnd) {
133 				/* Mark all we could possibly resend */
134 				break;
135 			}
136 		}
137 	}
138 	if (cnt) {
139 		/*
140 		 * JRS - Use the congestion control given in the congestion
141 		 * control module
142 		 */
143 		stcb->asoc.cc_functions.sctp_cwnd_update_after_fr_timer(inp, stcb, net);
144 	} else if (cnt_resend) {
145 		sctp_chunk_output(inp, stcb, SCTP_OUTPUT_FROM_EARLY_FR_TMR, SCTP_SO_NOT_LOCKED);
146 	}
147 	/* Restart it? */
148 	if (net->flight_size < net->cwnd) {
149 		SCTP_STAT_INCR(sctps_earlyfrstrtmr);
150 		sctp_timer_start(SCTP_TIMER_TYPE_EARLYFR, stcb->sctp_ep, stcb, net);
151 	}
152 }
153 
154 void
155 sctp_audit_retranmission_queue(struct sctp_association *asoc)
156 {
157 	struct sctp_tmit_chunk *chk;
158 
159 	SCTPDBG(SCTP_DEBUG_TIMER4, "Audit invoked on send queue cnt:%d onqueue:%d\n",
160 	    asoc->sent_queue_retran_cnt,
161 	    asoc->sent_queue_cnt);
162 	asoc->sent_queue_retran_cnt = 0;
163 	asoc->sent_queue_cnt = 0;
164 	TAILQ_FOREACH(chk, &asoc->sent_queue, sctp_next) {
165 		if (chk->sent == SCTP_DATAGRAM_RESEND) {
166 			sctp_ucount_incr(asoc->sent_queue_retran_cnt);
167 		}
168 		asoc->sent_queue_cnt++;
169 	}
170 	TAILQ_FOREACH(chk, &asoc->control_send_queue, sctp_next) {
171 		if (chk->sent == SCTP_DATAGRAM_RESEND) {
172 			sctp_ucount_incr(asoc->sent_queue_retran_cnt);
173 		}
174 	}
175 	TAILQ_FOREACH(chk, &asoc->asconf_send_queue, sctp_next) {
176 		if (chk->sent == SCTP_DATAGRAM_RESEND) {
177 			sctp_ucount_incr(asoc->sent_queue_retran_cnt);
178 		}
179 	}
180 	SCTPDBG(SCTP_DEBUG_TIMER4, "Audit completes retran:%d onqueue:%d\n",
181 	    asoc->sent_queue_retran_cnt,
182 	    asoc->sent_queue_cnt);
183 }
184 
185 int
186 sctp_threshold_management(struct sctp_inpcb *inp, struct sctp_tcb *stcb,
187     struct sctp_nets *net, uint16_t threshold)
188 {
189 	if (net) {
190 		net->error_count++;
191 		SCTPDBG(SCTP_DEBUG_TIMER4, "Error count for %p now %d thresh:%d\n",
192 		    net, net->error_count,
193 		    net->failure_threshold);
194 		if (net->error_count > net->failure_threshold) {
195 			/* We had a threshold failure */
196 			if (net->dest_state & SCTP_ADDR_REACHABLE) {
197 				net->dest_state &= ~SCTP_ADDR_REACHABLE;
198 				net->dest_state |= SCTP_ADDR_NOT_REACHABLE;
199 				net->dest_state &= ~SCTP_ADDR_REQ_PRIMARY;
200 				if (net == stcb->asoc.primary_destination) {
201 					net->dest_state |= SCTP_ADDR_WAS_PRIMARY;
202 				}
203 				/*
204 				 * JRS 5/14/07 - If a destination is
205 				 * unreachable, the PF bit is turned off.
206 				 * This allows an unambiguous use of the PF
207 				 * bit for destinations that are reachable
208 				 * but potentially failed. If the
209 				 * destination is set to the unreachable
210 				 * state, also set the destination to the PF
211 				 * state.
212 				 */
213 				/*
214 				 * Add debug message here if destination is
215 				 * not in PF state.
216 				 */
217 				/* Stop any running T3 timers here? */
218 				if (SCTP_BASE_SYSCTL(sctp_cmt_on_off) && SCTP_BASE_SYSCTL(sctp_cmt_pf)) {
219 					net->dest_state &= ~SCTP_ADDR_PF;
220 					SCTPDBG(SCTP_DEBUG_TIMER4, "Destination %p moved from PF to unreachable.\n",
221 					    net);
222 				}
223 				sctp_ulp_notify(SCTP_NOTIFY_INTERFACE_DOWN,
224 				    stcb,
225 				    SCTP_FAILED_THRESHOLD,
226 				    (void *)net, SCTP_SO_NOT_LOCKED);
227 			}
228 		}
229 		/*********HOLD THIS COMMENT FOR PATCH OF ALTERNATE
230 		 *********ROUTING CODE
231 		 */
232 		/*********HOLD THIS COMMENT FOR END OF PATCH OF ALTERNATE
233 		 *********ROUTING CODE
234 		 */
235 	}
236 	if (stcb == NULL)
237 		return (0);
238 
239 	if (net) {
240 		if ((net->dest_state & SCTP_ADDR_UNCONFIRMED) == 0) {
241 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_THRESHOLD_LOGGING) {
242 				sctp_misc_ints(SCTP_THRESHOLD_INCR,
243 				    stcb->asoc.overall_error_count,
244 				    (stcb->asoc.overall_error_count + 1),
245 				    SCTP_FROM_SCTP_TIMER,
246 				    __LINE__);
247 			}
248 			stcb->asoc.overall_error_count++;
249 		}
250 	} else {
251 		if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_THRESHOLD_LOGGING) {
252 			sctp_misc_ints(SCTP_THRESHOLD_INCR,
253 			    stcb->asoc.overall_error_count,
254 			    (stcb->asoc.overall_error_count + 1),
255 			    SCTP_FROM_SCTP_TIMER,
256 			    __LINE__);
257 		}
258 		stcb->asoc.overall_error_count++;
259 	}
260 	SCTPDBG(SCTP_DEBUG_TIMER4, "Overall error count for %p now %d thresh:%u state:%x\n",
261 	    &stcb->asoc, stcb->asoc.overall_error_count,
262 	    (uint32_t) threshold,
263 	    ((net == NULL) ? (uint32_t) 0 : (uint32_t) net->dest_state));
264 	/*
265 	 * We specifically do not do >= to give the assoc one more change
266 	 * before we fail it.
267 	 */
268 	if (stcb->asoc.overall_error_count > threshold) {
269 		/* Abort notification sends a ULP notify */
270 		struct mbuf *oper;
271 
272 		oper = sctp_get_mbuf_for_msg((sizeof(struct sctp_paramhdr) + sizeof(uint32_t)),
273 		    0, M_DONTWAIT, 1, MT_DATA);
274 		if (oper) {
275 			struct sctp_paramhdr *ph;
276 			uint32_t *ippp;
277 
278 			SCTP_BUF_LEN(oper) = sizeof(struct sctp_paramhdr) +
279 			    sizeof(uint32_t);
280 			ph = mtod(oper, struct sctp_paramhdr *);
281 			ph->param_type = htons(SCTP_CAUSE_PROTOCOL_VIOLATION);
282 			ph->param_length = htons(SCTP_BUF_LEN(oper));
283 			ippp = (uint32_t *) (ph + 1);
284 			*ippp = htonl(SCTP_FROM_SCTP_TIMER + SCTP_LOC_1);
285 		}
286 		inp->last_abort_code = SCTP_FROM_SCTP_TIMER + SCTP_LOC_1;
287 		sctp_abort_an_association(inp, stcb, SCTP_FAILED_THRESHOLD, oper, SCTP_SO_NOT_LOCKED);
288 		return (1);
289 	}
290 	return (0);
291 }
292 
293 struct sctp_nets *
294 sctp_find_alternate_net(struct sctp_tcb *stcb,
295     struct sctp_nets *net,
296     int mode)
297 {
298 	/* Find and return an alternate network if possible */
299 	struct sctp_nets *alt, *mnet, *min_errors_net = NULL, *max_cwnd_net = NULL;
300 	int once;
301 
302 	/* JRS 5/14/07 - Initialize min_errors to an impossible value. */
303 	int min_errors = -1;
304 	uint32_t max_cwnd = 0;
305 
306 	if (stcb->asoc.numnets == 1) {
307 		/* No others but net */
308 		return (TAILQ_FIRST(&stcb->asoc.nets));
309 	}
310 	/*
311 	 * JRS 5/14/07 - If mode is set to 2, use the CMT PF find alternate
312 	 * net algorithm. This algorithm chooses the active destination (not
313 	 * in PF state) with the largest cwnd value. If all destinations are
314 	 * in PF state, unreachable, or unconfirmed, choose the desination
315 	 * that is in PF state with the lowest error count. In case of a
316 	 * tie, choose the destination that was most recently active.
317 	 */
318 	if (mode == 2) {
319 		TAILQ_FOREACH(mnet, &stcb->asoc.nets, sctp_next) {
320 			/*
321 			 * JRS 5/14/07 - If the destination is unreachable
322 			 * or unconfirmed, skip it.
323 			 */
324 			if (((mnet->dest_state & SCTP_ADDR_REACHABLE) != SCTP_ADDR_REACHABLE) ||
325 			    (mnet->dest_state & SCTP_ADDR_UNCONFIRMED)) {
326 				continue;
327 			}
328 			/*
329 			 * JRS 5/14/07 -  If the destination is reachable
330 			 * but in PF state, compare the error count of the
331 			 * destination to the minimum error count seen thus
332 			 * far. Store the destination with the lower error
333 			 * count.  If the error counts are equal, store the
334 			 * destination that was most recently active.
335 			 */
336 			if (mnet->dest_state & SCTP_ADDR_PF) {
337 				/*
338 				 * JRS 5/14/07 - If the destination under
339 				 * consideration is the current destination,
340 				 * work as if the error count is one higher.
341 				 * The actual error count will not be
342 				 * incremented until later in the t3
343 				 * handler.
344 				 */
345 				if (mnet == net) {
346 					if (min_errors == -1) {
347 						min_errors = mnet->error_count + 1;
348 						min_errors_net = mnet;
349 					} else if (mnet->error_count + 1 < min_errors) {
350 						min_errors = mnet->error_count + 1;
351 						min_errors_net = mnet;
352 					} else if (mnet->error_count + 1 == min_errors
353 					    && mnet->last_active > min_errors_net->last_active) {
354 						min_errors_net = mnet;
355 						min_errors = mnet->error_count + 1;
356 					}
357 					continue;
358 				} else {
359 					if (min_errors == -1) {
360 						min_errors = mnet->error_count;
361 						min_errors_net = mnet;
362 					} else if (mnet->error_count < min_errors) {
363 						min_errors = mnet->error_count;
364 						min_errors_net = mnet;
365 					} else if (mnet->error_count == min_errors
366 					    && mnet->last_active > min_errors_net->last_active) {
367 						min_errors_net = mnet;
368 						min_errors = mnet->error_count;
369 					}
370 					continue;
371 				}
372 			}
373 			/*
374 			 * JRS 5/14/07 - If the destination is reachable and
375 			 * not in PF state, compare the cwnd of the
376 			 * destination to the highest cwnd seen thus far.
377 			 * Store the destination with the higher cwnd value.
378 			 * If the cwnd values are equal, randomly choose one
379 			 * of the two destinations.
380 			 */
381 			if (max_cwnd < mnet->cwnd) {
382 				max_cwnd_net = mnet;
383 				max_cwnd = mnet->cwnd;
384 			} else if (max_cwnd == mnet->cwnd) {
385 				uint32_t rndval;
386 				uint8_t this_random;
387 
388 				if (stcb->asoc.hb_random_idx > 3) {
389 					rndval = sctp_select_initial_TSN(&stcb->sctp_ep->sctp_ep);
390 					memcpy(stcb->asoc.hb_random_values, &rndval, sizeof(stcb->asoc.hb_random_values));
391 					this_random = stcb->asoc.hb_random_values[0];
392 					stcb->asoc.hb_random_idx++;
393 					stcb->asoc.hb_ect_randombit = 0;
394 				} else {
395 					this_random = stcb->asoc.hb_random_values[stcb->asoc.hb_random_idx];
396 					stcb->asoc.hb_random_idx++;
397 					stcb->asoc.hb_ect_randombit = 0;
398 				}
399 				if (this_random % 2 == 1) {
400 					max_cwnd_net = mnet;
401 					max_cwnd = mnet->cwnd;	/* Useless? */
402 				}
403 			}
404 		}
405 		/*
406 		 * JRS 5/14/07 - After all destination have been considered
407 		 * as alternates, check to see if there was some active
408 		 * destination (not in PF state).  If not, check to see if
409 		 * there was some PF destination with the minimum number of
410 		 * errors.  If not, return the original destination.  If
411 		 * there is a min_errors_net, remove the PF flag from that
412 		 * destination, set the cwnd to one or two MTUs, and return
413 		 * the destination as an alt. If there was some active
414 		 * destination with a highest cwnd, return the destination
415 		 * as an alt.
416 		 */
417 		if (max_cwnd_net == NULL) {
418 			if (min_errors_net == NULL) {
419 				return (net);
420 			}
421 			min_errors_net->dest_state &= ~SCTP_ADDR_PF;
422 			min_errors_net->cwnd = min_errors_net->mtu * SCTP_BASE_SYSCTL(sctp_cmt_pf);
423 			if (SCTP_OS_TIMER_PENDING(&min_errors_net->rxt_timer.timer)) {
424 				sctp_timer_stop(SCTP_TIMER_TYPE_SEND, stcb->sctp_ep,
425 				    stcb, min_errors_net,
426 				    SCTP_FROM_SCTP_TIMER + SCTP_LOC_2);
427 			}
428 			SCTPDBG(SCTP_DEBUG_TIMER4, "Destination %p moved from PF to active with %d errors.\n",
429 			    min_errors_net, min_errors_net->error_count);
430 			return (min_errors_net);
431 		} else {
432 			return (max_cwnd_net);
433 		}
434 	}
435 	/*
436 	 * JRS 5/14/07 - If mode is set to 1, use the CMT policy for
437 	 * choosing an alternate net.
438 	 */
439 	else if (mode == 1) {
440 		TAILQ_FOREACH(mnet, &stcb->asoc.nets, sctp_next) {
441 			if (((mnet->dest_state & SCTP_ADDR_REACHABLE) != SCTP_ADDR_REACHABLE) ||
442 			    (mnet->dest_state & SCTP_ADDR_UNCONFIRMED)
443 			    ) {
444 				/*
445 				 * will skip ones that are not-reachable or
446 				 * unconfirmed
447 				 */
448 				continue;
449 			}
450 			if (max_cwnd < mnet->cwnd) {
451 				max_cwnd_net = mnet;
452 				max_cwnd = mnet->cwnd;
453 			} else if (max_cwnd == mnet->cwnd) {
454 				uint32_t rndval;
455 				uint8_t this_random;
456 
457 				if (stcb->asoc.hb_random_idx > 3) {
458 					rndval = sctp_select_initial_TSN(&stcb->sctp_ep->sctp_ep);
459 					memcpy(stcb->asoc.hb_random_values, &rndval,
460 					    sizeof(stcb->asoc.hb_random_values));
461 					this_random = stcb->asoc.hb_random_values[0];
462 					stcb->asoc.hb_random_idx = 0;
463 					stcb->asoc.hb_ect_randombit = 0;
464 				} else {
465 					this_random = stcb->asoc.hb_random_values[stcb->asoc.hb_random_idx];
466 					stcb->asoc.hb_random_idx++;
467 					stcb->asoc.hb_ect_randombit = 0;
468 				}
469 				if (this_random % 2) {
470 					max_cwnd_net = mnet;
471 					max_cwnd = mnet->cwnd;
472 				}
473 			}
474 		}
475 		if (max_cwnd_net) {
476 			return (max_cwnd_net);
477 		}
478 	}
479 	mnet = net;
480 	once = 0;
481 
482 	if (mnet == NULL) {
483 		mnet = TAILQ_FIRST(&stcb->asoc.nets);
484 	}
485 	do {
486 		alt = TAILQ_NEXT(mnet, sctp_next);
487 		if (alt == NULL) {
488 			once++;
489 			if (once > 1) {
490 				break;
491 			}
492 			alt = TAILQ_FIRST(&stcb->asoc.nets);
493 		}
494 		if (alt->ro.ro_rt == NULL) {
495 			if (alt->ro._s_addr) {
496 				sctp_free_ifa(alt->ro._s_addr);
497 				alt->ro._s_addr = NULL;
498 			}
499 			alt->src_addr_selected = 0;
500 		}
501 		if (
502 		    ((alt->dest_state & SCTP_ADDR_REACHABLE) == SCTP_ADDR_REACHABLE) &&
503 		    (alt->ro.ro_rt != NULL) &&
504 		/* sa_ignore NO_NULL_CHK */
505 		    (!(alt->dest_state & SCTP_ADDR_UNCONFIRMED))
506 		    ) {
507 			/* Found a reachable address */
508 			break;
509 		}
510 		mnet = alt;
511 	} while (alt != NULL);
512 
513 	if (alt == NULL) {
514 		/* Case where NO insv network exists (dormant state) */
515 		/* we rotate destinations */
516 		once = 0;
517 		mnet = net;
518 		do {
519 			alt = TAILQ_NEXT(mnet, sctp_next);
520 			if (alt == NULL) {
521 				once++;
522 				if (once > 1) {
523 					break;
524 				}
525 				alt = TAILQ_FIRST(&stcb->asoc.nets);
526 			}
527 			/* sa_ignore NO_NULL_CHK */
528 			if ((!(alt->dest_state & SCTP_ADDR_UNCONFIRMED)) &&
529 			    (alt != net)) {
530 				/* Found an alternate address */
531 				break;
532 			}
533 			mnet = alt;
534 		} while (alt != NULL);
535 	}
536 	if (alt == NULL) {
537 		return (net);
538 	}
539 	return (alt);
540 }
541 
542 
543 
544 static void
545 sctp_backoff_on_timeout(struct sctp_tcb *stcb,
546     struct sctp_nets *net,
547     int win_probe,
548     int num_marked)
549 {
550 	if (net->RTO == 0) {
551 		net->RTO = stcb->asoc.minrto;
552 	}
553 	net->RTO <<= 1;
554 	if (net->RTO > stcb->asoc.maxrto) {
555 		net->RTO = stcb->asoc.maxrto;
556 	}
557 	if ((win_probe == 0) && num_marked) {
558 		/* We don't apply penalty to window probe scenarios */
559 		/* JRS - Use the congestion control given in the CC module */
560 		stcb->asoc.cc_functions.sctp_cwnd_update_after_timeout(stcb, net);
561 	}
562 }
563 
564 #ifndef INVARIANTS
565 static void
566 sctp_recover_sent_list(struct sctp_tcb *stcb)
567 {
568 	struct sctp_tmit_chunk *chk, *tp2;
569 	struct sctp_association *asoc;
570 
571 	asoc = &stcb->asoc;
572 	chk = TAILQ_FIRST(&stcb->asoc.sent_queue);
573 	for (; chk != NULL; chk = tp2) {
574 		tp2 = TAILQ_NEXT(chk, sctp_next);
575 		if ((compare_with_wrap(stcb->asoc.last_acked_seq,
576 		    chk->rec.data.TSN_seq,
577 		    MAX_TSN)) ||
578 		    (stcb->asoc.last_acked_seq == chk->rec.data.TSN_seq)) {
579 
580 			SCTP_PRINTF("Found chk:%p tsn:%x <= last_acked_seq:%x\n",
581 			    chk, chk->rec.data.TSN_seq, stcb->asoc.last_acked_seq);
582 			TAILQ_REMOVE(&asoc->sent_queue, chk, sctp_next);
583 			if (chk->pr_sctp_on) {
584 				if (asoc->pr_sctp_cnt != 0)
585 					asoc->pr_sctp_cnt--;
586 			}
587 			if (chk->data) {
588 				/* sa_ignore NO_NULL_CHK */
589 				sctp_free_bufspace(stcb, asoc, chk, 1);
590 				sctp_m_freem(chk->data);
591 				if (PR_SCTP_BUF_ENABLED(chk->flags)) {
592 					asoc->sent_queue_cnt_removeable--;
593 				}
594 			}
595 			chk->data = NULL;
596 			asoc->sent_queue_cnt--;
597 			sctp_free_a_chunk(stcb, chk);
598 		}
599 	}
600 	SCTP_PRINTF("after recover order is as follows\n");
601 	chk = TAILQ_FIRST(&stcb->asoc.sent_queue);
602 	for (; chk != NULL; chk = tp2) {
603 		tp2 = TAILQ_NEXT(chk, sctp_next);
604 		SCTP_PRINTF("chk:%p TSN:%x\n", chk, chk->rec.data.TSN_seq);
605 	}
606 }
607 
608 #endif
609 
610 static int
611 sctp_mark_all_for_resend(struct sctp_tcb *stcb,
612     struct sctp_nets *net,
613     struct sctp_nets *alt,
614     int window_probe,
615     int *num_marked)
616 {
617 
618 	/*
619 	 * Mark all chunks (well not all) that were sent to *net for
620 	 * retransmission. Move them to alt for there destination as well...
621 	 * We only mark chunks that have been outstanding long enough to
622 	 * have received feed-back.
623 	 */
624 	struct sctp_tmit_chunk *chk, *tp2, *could_be_sent = NULL;
625 	struct sctp_nets *lnets;
626 	struct timeval now, min_wait, tv;
627 	int cur_rtt;
628 	int audit_tf, num_mk, fir;
629 	unsigned int cnt_mk;
630 	uint32_t orig_flight, orig_tf;
631 	uint32_t tsnlast, tsnfirst;
632 	int recovery_cnt = 0;
633 
634 
635 	/* none in flight now */
636 	audit_tf = 0;
637 	fir = 0;
638 	/*
639 	 * figure out how long a data chunk must be pending before we can
640 	 * mark it ..
641 	 */
642 	(void)SCTP_GETTIME_TIMEVAL(&now);
643 	/* get cur rto in micro-seconds */
644 	cur_rtt = (((net->lastsa >> 2) + net->lastsv) >> 1);
645 	cur_rtt *= 1000;
646 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & (SCTP_EARLYFR_LOGGING_ENABLE | SCTP_FR_LOGGING_ENABLE)) {
647 		sctp_log_fr(cur_rtt,
648 		    stcb->asoc.peers_rwnd,
649 		    window_probe,
650 		    SCTP_FR_T3_MARK_TIME);
651 		sctp_log_fr(net->flight_size,
652 		    SCTP_OS_TIMER_PENDING(&net->fr_timer.timer),
653 		    SCTP_OS_TIMER_ACTIVE(&net->fr_timer.timer),
654 		    SCTP_FR_CWND_REPORT);
655 		sctp_log_fr(net->flight_size, net->cwnd, stcb->asoc.total_flight, SCTP_FR_CWND_REPORT);
656 	}
657 	tv.tv_sec = cur_rtt / 1000000;
658 	tv.tv_usec = cur_rtt % 1000000;
659 	min_wait = now;
660 	timevalsub(&min_wait, &tv);
661 	if (min_wait.tv_sec < 0 || min_wait.tv_usec < 0) {
662 		/*
663 		 * if we hit here, we don't have enough seconds on the clock
664 		 * to account for the RTO. We just let the lower seconds be
665 		 * the bounds and don't worry about it. This may mean we
666 		 * will mark a lot more than we should.
667 		 */
668 		min_wait.tv_sec = min_wait.tv_usec = 0;
669 	}
670 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & (SCTP_EARLYFR_LOGGING_ENABLE | SCTP_FR_LOGGING_ENABLE)) {
671 		sctp_log_fr(cur_rtt, now.tv_sec, now.tv_usec, SCTP_FR_T3_MARK_TIME);
672 		sctp_log_fr(0, min_wait.tv_sec, min_wait.tv_usec, SCTP_FR_T3_MARK_TIME);
673 	}
674 	/*
675 	 * Our rwnd will be incorrect here since we are not adding back the
676 	 * cnt * mbuf but we will fix that down below.
677 	 */
678 	orig_flight = net->flight_size;
679 	orig_tf = stcb->asoc.total_flight;
680 
681 	net->fast_retran_ip = 0;
682 	/* Now on to each chunk */
683 	num_mk = cnt_mk = 0;
684 	tsnfirst = tsnlast = 0;
685 #ifndef INVARIANTS
686 start_again:
687 #endif
688 	chk = TAILQ_FIRST(&stcb->asoc.sent_queue);
689 	for (; chk != NULL; chk = tp2) {
690 		tp2 = TAILQ_NEXT(chk, sctp_next);
691 		if ((compare_with_wrap(stcb->asoc.last_acked_seq,
692 		    chk->rec.data.TSN_seq,
693 		    MAX_TSN)) ||
694 		    (stcb->asoc.last_acked_seq == chk->rec.data.TSN_seq)) {
695 			/* Strange case our list got out of order? */
696 			SCTP_PRINTF("Our list is out of order? last_acked:%x chk:%x",
697 			    (unsigned int)stcb->asoc.last_acked_seq, (unsigned int)chk->rec.data.TSN_seq);
698 			recovery_cnt++;
699 #ifdef INVARIANTS
700 			panic("last acked >= chk on sent-Q");
701 #else
702 			SCTP_PRINTF("Recover attempts a restart cnt:%d\n", recovery_cnt);
703 			sctp_recover_sent_list(stcb);
704 			if (recovery_cnt < 10) {
705 				goto start_again;
706 			} else {
707 				SCTP_PRINTF("Recovery fails %d times??\n", recovery_cnt);
708 			}
709 #endif
710 		}
711 		if ((chk->whoTo == net) && (chk->sent < SCTP_DATAGRAM_ACKED)) {
712 			/*
713 			 * found one to mark: If it is less than
714 			 * DATAGRAM_ACKED it MUST not be a skipped or marked
715 			 * TSN but instead one that is either already set
716 			 * for retransmission OR one that needs
717 			 * retransmission.
718 			 */
719 
720 			/* validate its been outstanding long enough */
721 			if (SCTP_BASE_SYSCTL(sctp_logging_level) & (SCTP_EARLYFR_LOGGING_ENABLE | SCTP_FR_LOGGING_ENABLE)) {
722 				sctp_log_fr(chk->rec.data.TSN_seq,
723 				    chk->sent_rcv_time.tv_sec,
724 				    chk->sent_rcv_time.tv_usec,
725 				    SCTP_FR_T3_MARK_TIME);
726 			}
727 			if ((chk->sent_rcv_time.tv_sec > min_wait.tv_sec) && (window_probe == 0)) {
728 				/*
729 				 * we have reached a chunk that was sent
730 				 * some seconds past our min.. forget it we
731 				 * will find no more to send.
732 				 */
733 				if (SCTP_BASE_SYSCTL(sctp_logging_level) & (SCTP_EARLYFR_LOGGING_ENABLE | SCTP_FR_LOGGING_ENABLE)) {
734 					sctp_log_fr(0,
735 					    chk->sent_rcv_time.tv_sec,
736 					    chk->sent_rcv_time.tv_usec,
737 					    SCTP_FR_T3_STOPPED);
738 				}
739 				continue;
740 			} else if ((chk->sent_rcv_time.tv_sec == min_wait.tv_sec) &&
741 			    (window_probe == 0)) {
742 				/*
743 				 * we must look at the micro seconds to
744 				 * know.
745 				 */
746 				if (chk->sent_rcv_time.tv_usec >= min_wait.tv_usec) {
747 					/*
748 					 * ok it was sent after our boundary
749 					 * time.
750 					 */
751 					if (SCTP_BASE_SYSCTL(sctp_logging_level) & (SCTP_EARLYFR_LOGGING_ENABLE | SCTP_FR_LOGGING_ENABLE)) {
752 						sctp_log_fr(0,
753 						    chk->sent_rcv_time.tv_sec,
754 						    chk->sent_rcv_time.tv_usec,
755 						    SCTP_FR_T3_STOPPED);
756 					}
757 					continue;
758 				}
759 			}
760 			if (PR_SCTP_TTL_ENABLED(chk->flags)) {
761 				/* Is it expired? */
762 				if ((now.tv_sec > chk->rec.data.timetodrop.tv_sec) ||
763 				    ((chk->rec.data.timetodrop.tv_sec == now.tv_sec) &&
764 				    (now.tv_usec > chk->rec.data.timetodrop.tv_usec))) {
765 					/* Yes so drop it */
766 					if (chk->data) {
767 						(void)sctp_release_pr_sctp_chunk(stcb,
768 						    chk,
769 						    (SCTP_RESPONSE_TO_USER_REQ | SCTP_NOTIFY_DATAGRAM_SENT),
770 						    SCTP_SO_NOT_LOCKED);
771 					}
772 					continue;
773 				}
774 			}
775 			if (PR_SCTP_RTX_ENABLED(chk->flags)) {
776 				/* Has it been retransmitted tv_sec times? */
777 				if (chk->snd_count > chk->rec.data.timetodrop.tv_sec) {
778 					if (chk->data) {
779 						(void)sctp_release_pr_sctp_chunk(stcb,
780 						    chk,
781 						    (SCTP_RESPONSE_TO_USER_REQ | SCTP_NOTIFY_DATAGRAM_SENT),
782 						    SCTP_SO_NOT_LOCKED);
783 					}
784 					continue;
785 				}
786 			}
787 			if (chk->sent < SCTP_DATAGRAM_RESEND) {
788 				sctp_ucount_incr(stcb->asoc.sent_queue_retran_cnt);
789 				num_mk++;
790 				if (fir == 0) {
791 					fir = 1;
792 					tsnfirst = chk->rec.data.TSN_seq;
793 				}
794 				tsnlast = chk->rec.data.TSN_seq;
795 				if (SCTP_BASE_SYSCTL(sctp_logging_level) & (SCTP_EARLYFR_LOGGING_ENABLE | SCTP_FR_LOGGING_ENABLE)) {
796 					sctp_log_fr(chk->rec.data.TSN_seq, chk->snd_count,
797 					    0, SCTP_FR_T3_MARKED);
798 				}
799 				if (chk->rec.data.chunk_was_revoked) {
800 					/* deflate the cwnd */
801 					chk->whoTo->cwnd -= chk->book_size;
802 					chk->rec.data.chunk_was_revoked = 0;
803 				}
804 				net->marked_retrans++;
805 				stcb->asoc.marked_retrans++;
806 				if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_FLIGHT_LOGGING_ENABLE) {
807 					sctp_misc_ints(SCTP_FLIGHT_LOG_DOWN_RSND_TO,
808 					    chk->whoTo->flight_size,
809 					    chk->book_size,
810 					    (uintptr_t) chk->whoTo,
811 					    chk->rec.data.TSN_seq);
812 				}
813 				sctp_flight_size_decrease(chk);
814 				sctp_total_flight_decrease(stcb, chk);
815 				stcb->asoc.peers_rwnd += chk->send_size;
816 				stcb->asoc.peers_rwnd += SCTP_BASE_SYSCTL(sctp_peer_chunk_oh);
817 			}
818 			chk->sent = SCTP_DATAGRAM_RESEND;
819 			SCTP_STAT_INCR(sctps_markedretrans);
820 
821 			/* reset the TSN for striking and other FR stuff */
822 			chk->rec.data.doing_fast_retransmit = 0;
823 			/* Clear any time so NO RTT is being done */
824 			chk->do_rtt = 0;
825 			if (alt != net) {
826 				sctp_free_remote_addr(chk->whoTo);
827 				chk->no_fr_allowed = 1;
828 				chk->whoTo = alt;
829 				atomic_add_int(&alt->ref_count, 1);
830 			} else {
831 				chk->no_fr_allowed = 0;
832 				if (TAILQ_EMPTY(&stcb->asoc.send_queue)) {
833 					chk->rec.data.fast_retran_tsn = stcb->asoc.sending_seq;
834 				} else {
835 					chk->rec.data.fast_retran_tsn = (TAILQ_FIRST(&stcb->asoc.send_queue))->rec.data.TSN_seq;
836 				}
837 			}
838 			/*
839 			 * CMT: Do not allow FRs on retransmitted TSNs.
840 			 */
841 			if (SCTP_BASE_SYSCTL(sctp_cmt_on_off) == 1) {
842 				chk->no_fr_allowed = 1;
843 			}
844 		} else if (chk->sent == SCTP_DATAGRAM_ACKED) {
845 			/* remember highest acked one */
846 			could_be_sent = chk;
847 		}
848 		if (chk->sent == SCTP_DATAGRAM_RESEND) {
849 			cnt_mk++;
850 		}
851 	}
852 	if ((orig_flight - net->flight_size) != (orig_tf - stcb->asoc.total_flight)) {
853 		/* we did not subtract the same things? */
854 		audit_tf = 1;
855 	}
856 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & (SCTP_EARLYFR_LOGGING_ENABLE | SCTP_FR_LOGGING_ENABLE)) {
857 		sctp_log_fr(tsnfirst, tsnlast, num_mk, SCTP_FR_T3_TIMEOUT);
858 	}
859 #ifdef SCTP_DEBUG
860 	if (num_mk) {
861 		SCTPDBG(SCTP_DEBUG_TIMER1, "LAST TSN marked was %x\n",
862 		    tsnlast);
863 		SCTPDBG(SCTP_DEBUG_TIMER1, "Num marked for retransmission was %d peer-rwd:%ld\n",
864 		    num_mk, (u_long)stcb->asoc.peers_rwnd);
865 		SCTPDBG(SCTP_DEBUG_TIMER1, "LAST TSN marked was %x\n",
866 		    tsnlast);
867 		SCTPDBG(SCTP_DEBUG_TIMER1, "Num marked for retransmission was %d peer-rwd:%d\n",
868 		    num_mk,
869 		    (int)stcb->asoc.peers_rwnd);
870 	}
871 #endif
872 	*num_marked = num_mk;
873 	if ((stcb->asoc.sent_queue_retran_cnt == 0) && (could_be_sent)) {
874 		/* fix it so we retransmit the highest acked anyway */
875 		sctp_ucount_incr(stcb->asoc.sent_queue_retran_cnt);
876 		cnt_mk++;
877 		could_be_sent->sent = SCTP_DATAGRAM_RESEND;
878 	}
879 	if (stcb->asoc.sent_queue_retran_cnt != cnt_mk) {
880 #ifdef INVARIANTS
881 		SCTP_PRINTF("Local Audit says there are %d for retran asoc cnt:%d we marked:%d this time\n",
882 		    cnt_mk, stcb->asoc.sent_queue_retran_cnt, num_mk);
883 #endif
884 #ifndef SCTP_AUDITING_ENABLED
885 		stcb->asoc.sent_queue_retran_cnt = cnt_mk;
886 #endif
887 	}
888 	/* Now check for a ECN Echo that may be stranded */
889 	TAILQ_FOREACH(chk, &stcb->asoc.control_send_queue, sctp_next) {
890 		if ((chk->whoTo == net) &&
891 		    (chk->rec.chunk_id.id == SCTP_ECN_ECHO)) {
892 			sctp_free_remote_addr(chk->whoTo);
893 			chk->whoTo = alt;
894 			if (chk->sent != SCTP_DATAGRAM_RESEND) {
895 				chk->sent = SCTP_DATAGRAM_RESEND;
896 				sctp_ucount_incr(stcb->asoc.sent_queue_retran_cnt);
897 			}
898 			atomic_add_int(&alt->ref_count, 1);
899 		}
900 	}
901 	if (audit_tf) {
902 		SCTPDBG(SCTP_DEBUG_TIMER4,
903 		    "Audit total flight due to negative value net:%p\n",
904 		    net);
905 		stcb->asoc.total_flight = 0;
906 		stcb->asoc.total_flight_count = 0;
907 		/* Clear all networks flight size */
908 		TAILQ_FOREACH(lnets, &stcb->asoc.nets, sctp_next) {
909 			lnets->flight_size = 0;
910 			SCTPDBG(SCTP_DEBUG_TIMER4,
911 			    "Net:%p c-f cwnd:%d ssthresh:%d\n",
912 			    lnets, lnets->cwnd, lnets->ssthresh);
913 		}
914 		TAILQ_FOREACH(chk, &stcb->asoc.sent_queue, sctp_next) {
915 			if (chk->sent < SCTP_DATAGRAM_RESEND) {
916 				if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_FLIGHT_LOGGING_ENABLE) {
917 					sctp_misc_ints(SCTP_FLIGHT_LOG_UP,
918 					    chk->whoTo->flight_size,
919 					    chk->book_size,
920 					    (uintptr_t) chk->whoTo,
921 					    chk->rec.data.TSN_seq);
922 				}
923 				sctp_flight_size_increase(chk);
924 				sctp_total_flight_increase(stcb, chk);
925 			}
926 		}
927 	}
928 	/*
929 	 * Setup the ecn nonce re-sync point. We do this since
930 	 * retranmissions are NOT setup for ECN. This means that do to
931 	 * Karn's rule, we don't know the total of the peers ecn bits.
932 	 */
933 	chk = TAILQ_FIRST(&stcb->asoc.send_queue);
934 	if (chk == NULL) {
935 		stcb->asoc.nonce_resync_tsn = stcb->asoc.sending_seq;
936 	} else {
937 		stcb->asoc.nonce_resync_tsn = chk->rec.data.TSN_seq;
938 	}
939 	stcb->asoc.nonce_wait_for_ecne = 0;
940 	stcb->asoc.nonce_sum_check = 0;
941 	/* We return 1 if we only have a window probe outstanding */
942 	return (0);
943 }
944 
945 static void
946 sctp_move_all_chunks_to_alt(struct sctp_tcb *stcb,
947     struct sctp_nets *net,
948     struct sctp_nets *alt)
949 {
950 	struct sctp_association *asoc;
951 	struct sctp_stream_out *outs;
952 	struct sctp_tmit_chunk *chk;
953 	struct sctp_stream_queue_pending *sp;
954 
955 	if (net == alt)
956 		/* nothing to do */
957 		return;
958 
959 	asoc = &stcb->asoc;
960 
961 	/*
962 	 * now through all the streams checking for chunks sent to our bad
963 	 * network.
964 	 */
965 	TAILQ_FOREACH(outs, &asoc->out_wheel, next_spoke) {
966 		/* now clean up any chunks here */
967 		TAILQ_FOREACH(sp, &outs->outqueue, next) {
968 			if (sp->net == net) {
969 				sctp_free_remote_addr(sp->net);
970 				sp->net = alt;
971 				atomic_add_int(&alt->ref_count, 1);
972 			}
973 		}
974 	}
975 	/* Now check the pending queue */
976 	TAILQ_FOREACH(chk, &asoc->send_queue, sctp_next) {
977 		if (chk->whoTo == net) {
978 			sctp_free_remote_addr(chk->whoTo);
979 			chk->whoTo = alt;
980 			atomic_add_int(&alt->ref_count, 1);
981 		}
982 	}
983 
984 }
985 
986 int
987 sctp_t3rxt_timer(struct sctp_inpcb *inp,
988     struct sctp_tcb *stcb,
989     struct sctp_nets *net)
990 {
991 	struct sctp_nets *alt;
992 	int win_probe, num_mk;
993 
994 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_FR_LOGGING_ENABLE) {
995 		sctp_log_fr(0, 0, 0, SCTP_FR_T3_TIMEOUT);
996 	}
997 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_CWND_LOGGING_ENABLE) {
998 		struct sctp_nets *lnet;
999 
1000 		TAILQ_FOREACH(lnet, &stcb->asoc.nets, sctp_next) {
1001 			if (net == lnet) {
1002 				sctp_log_cwnd(stcb, lnet, 1, SCTP_CWND_LOG_FROM_T3);
1003 			} else {
1004 				sctp_log_cwnd(stcb, lnet, 0, SCTP_CWND_LOG_FROM_T3);
1005 			}
1006 		}
1007 	}
1008 	/* Find an alternate and mark those for retransmission */
1009 	if ((stcb->asoc.peers_rwnd == 0) &&
1010 	    (stcb->asoc.total_flight < net->mtu)) {
1011 		SCTP_STAT_INCR(sctps_timowindowprobe);
1012 		win_probe = 1;
1013 	} else {
1014 		win_probe = 0;
1015 	}
1016 
1017 	/*
1018 	 * JRS 5/14/07 - If CMT PF is on and the destination if not already
1019 	 * in PF state, set the destination to PF state and store the
1020 	 * current time as the time that the destination was last active. In
1021 	 * addition, find an alternate destination with PF-based
1022 	 * find_alt_net().
1023 	 */
1024 	if (SCTP_BASE_SYSCTL(sctp_cmt_on_off) && SCTP_BASE_SYSCTL(sctp_cmt_pf)) {
1025 		if ((net->dest_state & SCTP_ADDR_PF) != SCTP_ADDR_PF) {
1026 			net->dest_state |= SCTP_ADDR_PF;
1027 			net->last_active = sctp_get_tick_count();
1028 			SCTPDBG(SCTP_DEBUG_TIMER4, "Destination %p moved from active to PF.\n",
1029 			    net);
1030 		}
1031 		alt = sctp_find_alternate_net(stcb, net, 2);
1032 	} else if (SCTP_BASE_SYSCTL(sctp_cmt_on_off)) {
1033 		/*
1034 		 * CMT: Using RTX_SSTHRESH policy for CMT. If CMT is being
1035 		 * used, then pick dest with largest ssthresh for any
1036 		 * retransmission.
1037 		 */
1038 		alt = net;
1039 		alt = sctp_find_alternate_net(stcb, alt, 1);
1040 		/*
1041 		 * CUCv2: If a different dest is picked for the
1042 		 * retransmission, then new (rtx-)pseudo_cumack needs to be
1043 		 * tracked for orig dest. Let CUCv2 track new (rtx-)
1044 		 * pseudo-cumack always.
1045 		 */
1046 		net->find_pseudo_cumack = 1;
1047 		net->find_rtx_pseudo_cumack = 1;
1048 	} else {		/* CMT is OFF */
1049 		alt = sctp_find_alternate_net(stcb, net, 0);
1050 	}
1051 
1052 	(void)sctp_mark_all_for_resend(stcb, net, alt, win_probe, &num_mk);
1053 	/* FR Loss recovery just ended with the T3. */
1054 	stcb->asoc.fast_retran_loss_recovery = 0;
1055 
1056 	/* CMT FR loss recovery ended with the T3 */
1057 	net->fast_retran_loss_recovery = 0;
1058 
1059 	/*
1060 	 * setup the sat loss recovery that prevents satellite cwnd advance.
1061 	 */
1062 	stcb->asoc.sat_t3_loss_recovery = 1;
1063 	stcb->asoc.sat_t3_recovery_tsn = stcb->asoc.sending_seq;
1064 
1065 	/* Backoff the timer and cwnd */
1066 	sctp_backoff_on_timeout(stcb, net, win_probe, num_mk);
1067 	if (win_probe == 0) {
1068 		/* We don't do normal threshold management on window probes */
1069 		if (sctp_threshold_management(inp, stcb, net,
1070 		    stcb->asoc.max_send_times)) {
1071 			/* Association was destroyed */
1072 			return (1);
1073 		} else {
1074 			if (net != stcb->asoc.primary_destination) {
1075 				/* send a immediate HB if our RTO is stale */
1076 				struct timeval now;
1077 				unsigned int ms_goneby;
1078 
1079 				(void)SCTP_GETTIME_TIMEVAL(&now);
1080 				if (net->last_sent_time.tv_sec) {
1081 					ms_goneby = (now.tv_sec - net->last_sent_time.tv_sec) * 1000;
1082 				} else {
1083 					ms_goneby = 0;
1084 				}
1085 				if ((ms_goneby > net->RTO) || (net->RTO == 0)) {
1086 					/*
1087 					 * no recent feed back in an RTO or
1088 					 * more, request a RTT update
1089 					 */
1090 					if (sctp_send_hb(stcb, 1, net) < 0)
1091 						/*
1092 						 * Less than 0 means we lost
1093 						 * the assoc
1094 						 */
1095 						return (1);
1096 				}
1097 			}
1098 		}
1099 	} else {
1100 		/*
1101 		 * For a window probe we don't penalize the net's but only
1102 		 * the association. This may fail it if SACKs are not coming
1103 		 * back. If sack's are coming with rwnd locked at 0, we will
1104 		 * continue to hold things waiting for rwnd to raise
1105 		 */
1106 		if (sctp_threshold_management(inp, stcb, NULL,
1107 		    stcb->asoc.max_send_times)) {
1108 			/* Association was destroyed */
1109 			return (1);
1110 		}
1111 	}
1112 	if (net->dest_state & SCTP_ADDR_NOT_REACHABLE) {
1113 		/* Move all pending over too */
1114 		sctp_move_all_chunks_to_alt(stcb, net, alt);
1115 
1116 		/*
1117 		 * Get the address that failed, to force a new src address
1118 		 * selecton and a route allocation.
1119 		 */
1120 		if (net->ro._s_addr) {
1121 			sctp_free_ifa(net->ro._s_addr);
1122 			net->ro._s_addr = NULL;
1123 		}
1124 		net->src_addr_selected = 0;
1125 
1126 		/* Force a route allocation too */
1127 		if (net->ro.ro_rt) {
1128 			RTFREE(net->ro.ro_rt);
1129 			net->ro.ro_rt = NULL;
1130 		}
1131 		/* Was it our primary? */
1132 		if ((stcb->asoc.primary_destination == net) && (alt != net)) {
1133 			/*
1134 			 * Yes, note it as such and find an alternate note:
1135 			 * this means HB code must use this to resent the
1136 			 * primary if it goes active AND if someone does a
1137 			 * change-primary then this flag must be cleared
1138 			 * from any net structures.
1139 			 */
1140 			if (sctp_set_primary_addr(stcb,
1141 			    (struct sockaddr *)NULL,
1142 			    alt) == 0) {
1143 				net->dest_state |= SCTP_ADDR_WAS_PRIMARY;
1144 			}
1145 		}
1146 	} else if (SCTP_BASE_SYSCTL(sctp_cmt_on_off) && SCTP_BASE_SYSCTL(sctp_cmt_pf) && (net->dest_state & SCTP_ADDR_PF) == SCTP_ADDR_PF) {
1147 		/*
1148 		 * JRS 5/14/07 - If the destination hasn't failed completely
1149 		 * but is in PF state, a PF-heartbeat needs to be sent
1150 		 * manually.
1151 		 */
1152 		if (sctp_send_hb(stcb, 1, net) < 0)
1153 			/* Return less than 0 means we lost the association */
1154 			return (1);
1155 	}
1156 	/*
1157 	 * Special case for cookie-echo'ed case, we don't do output but must
1158 	 * await the COOKIE-ACK before retransmission
1159 	 */
1160 	if (SCTP_GET_STATE(&stcb->asoc) == SCTP_STATE_COOKIE_ECHOED) {
1161 		/*
1162 		 * Here we just reset the timer and start again since we
1163 		 * have not established the asoc
1164 		 */
1165 		sctp_timer_start(SCTP_TIMER_TYPE_SEND, inp, stcb, net);
1166 		return (0);
1167 	}
1168 	if (stcb->asoc.peer_supports_prsctp) {
1169 		struct sctp_tmit_chunk *lchk;
1170 
1171 		lchk = sctp_try_advance_peer_ack_point(stcb, &stcb->asoc);
1172 		/* C3. See if we need to send a Fwd-TSN */
1173 		if (compare_with_wrap(stcb->asoc.advanced_peer_ack_point,
1174 		    stcb->asoc.last_acked_seq, MAX_TSN)) {
1175 			/*
1176 			 * ISSUE with ECN, see FWD-TSN processing for notes
1177 			 * on issues that will occur when the ECN NONCE
1178 			 * stuff is put into SCTP for cross checking.
1179 			 */
1180 			send_forward_tsn(stcb, &stcb->asoc);
1181 			if (lchk) {
1182 				/* Assure a timer is up */
1183 				sctp_timer_start(SCTP_TIMER_TYPE_SEND, stcb->sctp_ep, stcb, lchk->whoTo);
1184 			}
1185 		}
1186 	}
1187 	if (SCTP_BASE_SYSCTL(sctp_logging_level) & SCTP_CWND_MONITOR_ENABLE) {
1188 		sctp_log_cwnd(stcb, net, net->cwnd, SCTP_CWND_LOG_FROM_RTX);
1189 	}
1190 	return (0);
1191 }
1192 
1193 int
1194 sctp_t1init_timer(struct sctp_inpcb *inp,
1195     struct sctp_tcb *stcb,
1196     struct sctp_nets *net)
1197 {
1198 	/* bump the thresholds */
1199 	if (stcb->asoc.delayed_connection) {
1200 		/*
1201 		 * special hook for delayed connection. The library did NOT
1202 		 * complete the rest of its sends.
1203 		 */
1204 		stcb->asoc.delayed_connection = 0;
1205 		sctp_send_initiate(inp, stcb, SCTP_SO_NOT_LOCKED);
1206 		return (0);
1207 	}
1208 	if (SCTP_GET_STATE((&stcb->asoc)) != SCTP_STATE_COOKIE_WAIT) {
1209 		return (0);
1210 	}
1211 	if (sctp_threshold_management(inp, stcb, net,
1212 	    stcb->asoc.max_init_times)) {
1213 		/* Association was destroyed */
1214 		return (1);
1215 	}
1216 	stcb->asoc.dropped_special_cnt = 0;
1217 	sctp_backoff_on_timeout(stcb, stcb->asoc.primary_destination, 1, 0);
1218 	if (stcb->asoc.initial_init_rto_max < net->RTO) {
1219 		net->RTO = stcb->asoc.initial_init_rto_max;
1220 	}
1221 	if (stcb->asoc.numnets > 1) {
1222 		/* If we have more than one addr use it */
1223 		struct sctp_nets *alt;
1224 
1225 		alt = sctp_find_alternate_net(stcb, stcb->asoc.primary_destination, 0);
1226 		if ((alt != NULL) && (alt != stcb->asoc.primary_destination)) {
1227 			sctp_move_all_chunks_to_alt(stcb, stcb->asoc.primary_destination, alt);
1228 			stcb->asoc.primary_destination = alt;
1229 		}
1230 	}
1231 	/* Send out a new init */
1232 	sctp_send_initiate(inp, stcb, SCTP_SO_NOT_LOCKED);
1233 	return (0);
1234 }
1235 
1236 /*
1237  * For cookie and asconf we actually need to find and mark for resend, then
1238  * increment the resend counter (after all the threshold management stuff of
1239  * course).
1240  */
1241 int
1242 sctp_cookie_timer(struct sctp_inpcb *inp,
1243     struct sctp_tcb *stcb,
1244     struct sctp_nets *net)
1245 {
1246 	struct sctp_nets *alt;
1247 	struct sctp_tmit_chunk *cookie;
1248 
1249 	/* first before all else we must find the cookie */
1250 	TAILQ_FOREACH(cookie, &stcb->asoc.control_send_queue, sctp_next) {
1251 		if (cookie->rec.chunk_id.id == SCTP_COOKIE_ECHO) {
1252 			break;
1253 		}
1254 	}
1255 	if (cookie == NULL) {
1256 		if (SCTP_GET_STATE(&stcb->asoc) == SCTP_STATE_COOKIE_ECHOED) {
1257 			/* FOOBAR! */
1258 			struct mbuf *oper;
1259 
1260 			oper = sctp_get_mbuf_for_msg((sizeof(struct sctp_paramhdr) + sizeof(uint32_t)),
1261 			    0, M_DONTWAIT, 1, MT_DATA);
1262 			if (oper) {
1263 				struct sctp_paramhdr *ph;
1264 				uint32_t *ippp;
1265 
1266 				SCTP_BUF_LEN(oper) = sizeof(struct sctp_paramhdr) +
1267 				    sizeof(uint32_t);
1268 				ph = mtod(oper, struct sctp_paramhdr *);
1269 				ph->param_type = htons(SCTP_CAUSE_PROTOCOL_VIOLATION);
1270 				ph->param_length = htons(SCTP_BUF_LEN(oper));
1271 				ippp = (uint32_t *) (ph + 1);
1272 				*ippp = htonl(SCTP_FROM_SCTP_TIMER + SCTP_LOC_3);
1273 			}
1274 			inp->last_abort_code = SCTP_FROM_SCTP_TIMER + SCTP_LOC_4;
1275 			sctp_abort_an_association(inp, stcb, SCTP_INTERNAL_ERROR,
1276 			    oper, SCTP_SO_NOT_LOCKED);
1277 		} else {
1278 #ifdef INVARIANTS
1279 			panic("Cookie timer expires in wrong state?");
1280 #else
1281 			SCTP_PRINTF("Strange in state %d not cookie-echoed yet c-e timer expires?\n", SCTP_GET_STATE(&stcb->asoc));
1282 			return (0);
1283 #endif
1284 		}
1285 		return (0);
1286 	}
1287 	/* Ok we found the cookie, threshold management next */
1288 	if (sctp_threshold_management(inp, stcb, cookie->whoTo,
1289 	    stcb->asoc.max_init_times)) {
1290 		/* Assoc is over */
1291 		return (1);
1292 	}
1293 	/*
1294 	 * cleared theshold management now lets backoff the address & select
1295 	 * an alternate
1296 	 */
1297 	stcb->asoc.dropped_special_cnt = 0;
1298 	sctp_backoff_on_timeout(stcb, cookie->whoTo, 1, 0);
1299 	alt = sctp_find_alternate_net(stcb, cookie->whoTo, 0);
1300 	if (alt != cookie->whoTo) {
1301 		sctp_free_remote_addr(cookie->whoTo);
1302 		cookie->whoTo = alt;
1303 		atomic_add_int(&alt->ref_count, 1);
1304 	}
1305 	/* Now mark the retran info */
1306 	if (cookie->sent != SCTP_DATAGRAM_RESEND) {
1307 		sctp_ucount_incr(stcb->asoc.sent_queue_retran_cnt);
1308 	}
1309 	cookie->sent = SCTP_DATAGRAM_RESEND;
1310 	/*
1311 	 * Now call the output routine to kick out the cookie again, Note we
1312 	 * don't mark any chunks for retran so that FR will need to kick in
1313 	 * to move these (or a send timer).
1314 	 */
1315 	return (0);
1316 }
1317 
1318 int
1319 sctp_strreset_timer(struct sctp_inpcb *inp, struct sctp_tcb *stcb,
1320     struct sctp_nets *net)
1321 {
1322 	struct sctp_nets *alt;
1323 	struct sctp_tmit_chunk *strrst = NULL, *chk = NULL;
1324 
1325 	if (stcb->asoc.stream_reset_outstanding == 0) {
1326 		return (0);
1327 	}
1328 	/* find the existing STRRESET, we use the seq number we sent out on */
1329 	(void)sctp_find_stream_reset(stcb, stcb->asoc.str_reset_seq_out, &strrst);
1330 	if (strrst == NULL) {
1331 		return (0);
1332 	}
1333 	/* do threshold management */
1334 	if (sctp_threshold_management(inp, stcb, strrst->whoTo,
1335 	    stcb->asoc.max_send_times)) {
1336 		/* Assoc is over */
1337 		return (1);
1338 	}
1339 	/*
1340 	 * cleared theshold management now lets backoff the address & select
1341 	 * an alternate
1342 	 */
1343 	sctp_backoff_on_timeout(stcb, strrst->whoTo, 1, 0);
1344 	alt = sctp_find_alternate_net(stcb, strrst->whoTo, 0);
1345 	sctp_free_remote_addr(strrst->whoTo);
1346 	strrst->whoTo = alt;
1347 	atomic_add_int(&alt->ref_count, 1);
1348 
1349 	/* See if a ECN Echo is also stranded */
1350 	TAILQ_FOREACH(chk, &stcb->asoc.control_send_queue, sctp_next) {
1351 		if ((chk->whoTo == net) &&
1352 		    (chk->rec.chunk_id.id == SCTP_ECN_ECHO)) {
1353 			sctp_free_remote_addr(chk->whoTo);
1354 			if (chk->sent != SCTP_DATAGRAM_RESEND) {
1355 				chk->sent = SCTP_DATAGRAM_RESEND;
1356 				sctp_ucount_incr(stcb->asoc.sent_queue_retran_cnt);
1357 			}
1358 			chk->whoTo = alt;
1359 			atomic_add_int(&alt->ref_count, 1);
1360 		}
1361 	}
1362 	if (net->dest_state & SCTP_ADDR_NOT_REACHABLE) {
1363 		/*
1364 		 * If the address went un-reachable, we need to move to
1365 		 * alternates for ALL chk's in queue
1366 		 */
1367 		sctp_move_all_chunks_to_alt(stcb, net, alt);
1368 	}
1369 	/* mark the retran info */
1370 	if (strrst->sent != SCTP_DATAGRAM_RESEND)
1371 		sctp_ucount_incr(stcb->asoc.sent_queue_retran_cnt);
1372 	strrst->sent = SCTP_DATAGRAM_RESEND;
1373 
1374 	/* restart the timer */
1375 	sctp_timer_start(SCTP_TIMER_TYPE_STRRESET, inp, stcb, strrst->whoTo);
1376 	return (0);
1377 }
1378 
1379 int
1380 sctp_asconf_timer(struct sctp_inpcb *inp, struct sctp_tcb *stcb,
1381     struct sctp_nets *net)
1382 {
1383 	struct sctp_nets *alt;
1384 	struct sctp_tmit_chunk *asconf, *chk, *nchk;
1385 
1386 	/* is this a first send, or a retransmission? */
1387 	if (TAILQ_EMPTY(&stcb->asoc.asconf_send_queue)) {
1388 		/* compose a new ASCONF chunk and send it */
1389 		sctp_send_asconf(stcb, net, SCTP_ADDR_NOT_LOCKED);
1390 	} else {
1391 		/*
1392 		 * Retransmission of the existing ASCONF is needed
1393 		 */
1394 
1395 		/* find the existing ASCONF */
1396 		asconf = TAILQ_FIRST(&stcb->asoc.asconf_send_queue);
1397 		if (asconf == NULL) {
1398 			return (0);
1399 		}
1400 		/* do threshold management */
1401 		if (sctp_threshold_management(inp, stcb, asconf->whoTo,
1402 		    stcb->asoc.max_send_times)) {
1403 			/* Assoc is over */
1404 			return (1);
1405 		}
1406 		if (asconf->snd_count > stcb->asoc.max_send_times) {
1407 			/*
1408 			 * Something is rotten: our peer is not responding
1409 			 * to ASCONFs but apparently is to other chunks.
1410 			 * i.e. it is not properly handling the chunk type
1411 			 * upper bits. Mark this peer as ASCONF incapable
1412 			 * and cleanup.
1413 			 */
1414 			SCTPDBG(SCTP_DEBUG_TIMER1, "asconf_timer: Peer has not responded to our repeated ASCONFs\n");
1415 			sctp_asconf_cleanup(stcb, net);
1416 			return (0);
1417 		}
1418 		/*
1419 		 * cleared threshold management, so now backoff the net and
1420 		 * select an alternate
1421 		 */
1422 		sctp_backoff_on_timeout(stcb, asconf->whoTo, 1, 0);
1423 		alt = sctp_find_alternate_net(stcb, asconf->whoTo, 0);
1424 		if (asconf->whoTo != alt) {
1425 			sctp_free_remote_addr(asconf->whoTo);
1426 			asconf->whoTo = alt;
1427 			atomic_add_int(&alt->ref_count, 1);
1428 		}
1429 		/* See if an ECN Echo is also stranded */
1430 		TAILQ_FOREACH(chk, &stcb->asoc.control_send_queue, sctp_next) {
1431 			if ((chk->whoTo == net) &&
1432 			    (chk->rec.chunk_id.id == SCTP_ECN_ECHO)) {
1433 				sctp_free_remote_addr(chk->whoTo);
1434 				chk->whoTo = alt;
1435 				if (chk->sent != SCTP_DATAGRAM_RESEND) {
1436 					chk->sent = SCTP_DATAGRAM_RESEND;
1437 					sctp_ucount_incr(stcb->asoc.sent_queue_retran_cnt);
1438 				}
1439 				atomic_add_int(&alt->ref_count, 1);
1440 			}
1441 		}
1442 		for (chk = asconf; chk; chk = nchk) {
1443 			nchk = TAILQ_NEXT(chk, sctp_next);
1444 			if (chk->whoTo != alt) {
1445 				sctp_free_remote_addr(chk->whoTo);
1446 				chk->whoTo = alt;
1447 				atomic_add_int(&alt->ref_count, 1);
1448 			}
1449 			if (asconf->sent != SCTP_DATAGRAM_RESEND && chk->sent != SCTP_DATAGRAM_UNSENT)
1450 				sctp_ucount_incr(stcb->asoc.sent_queue_retran_cnt);
1451 			chk->sent = SCTP_DATAGRAM_RESEND;
1452 		}
1453 		if (net->dest_state & SCTP_ADDR_NOT_REACHABLE) {
1454 			/*
1455 			 * If the address went un-reachable, we need to move
1456 			 * to the alternate for ALL chunks in queue
1457 			 */
1458 			sctp_move_all_chunks_to_alt(stcb, net, alt);
1459 			net = alt;
1460 		}
1461 		/* mark the retran info */
1462 		if (asconf->sent != SCTP_DATAGRAM_RESEND)
1463 			sctp_ucount_incr(stcb->asoc.sent_queue_retran_cnt);
1464 		asconf->sent = SCTP_DATAGRAM_RESEND;
1465 
1466 		/* send another ASCONF if any and we can do */
1467 		sctp_send_asconf(stcb, alt, SCTP_ADDR_NOT_LOCKED);
1468 	}
1469 	return (0);
1470 }
1471 
1472 /* Mobility adaptation */
1473 void
1474 sctp_delete_prim_timer(struct sctp_inpcb *inp, struct sctp_tcb *stcb,
1475     struct sctp_nets *net)
1476 {
1477 	if (stcb->asoc.deleted_primary == NULL) {
1478 		SCTPDBG(SCTP_DEBUG_ASCONF1, "delete_prim_timer: deleted_primary is not stored...\n");
1479 		sctp_mobility_feature_off(inp, SCTP_MOBILITY_PRIM_DELETED);
1480 		return;
1481 	}
1482 	SCTPDBG(SCTP_DEBUG_ASCONF1, "delete_prim_timer: finished to keep deleted primary ");
1483 	SCTPDBG_ADDR(SCTP_DEBUG_ASCONF1, &stcb->asoc.deleted_primary->ro._l_addr.sa);
1484 	sctp_free_remote_addr(stcb->asoc.deleted_primary);
1485 	stcb->asoc.deleted_primary = NULL;
1486 	sctp_mobility_feature_off(inp, SCTP_MOBILITY_PRIM_DELETED);
1487 	return;
1488 }
1489 
1490 /*
1491  * For the shutdown and shutdown-ack, we do not keep one around on the
1492  * control queue. This means we must generate a new one and call the general
1493  * chunk output routine, AFTER having done threshold management.
1494  */
1495 int
1496 sctp_shutdown_timer(struct sctp_inpcb *inp, struct sctp_tcb *stcb,
1497     struct sctp_nets *net)
1498 {
1499 	struct sctp_nets *alt;
1500 
1501 	/* first threshold managment */
1502 	if (sctp_threshold_management(inp, stcb, net, stcb->asoc.max_send_times)) {
1503 		/* Assoc is over */
1504 		return (1);
1505 	}
1506 	/* second select an alternative */
1507 	alt = sctp_find_alternate_net(stcb, net, 0);
1508 
1509 	/* third generate a shutdown into the queue for out net */
1510 	if (alt) {
1511 		sctp_send_shutdown(stcb, alt);
1512 	} else {
1513 		/*
1514 		 * if alt is NULL, there is no dest to send to??
1515 		 */
1516 		return (0);
1517 	}
1518 	/* fourth restart timer */
1519 	sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWN, inp, stcb, alt);
1520 	return (0);
1521 }
1522 
1523 int
1524 sctp_shutdownack_timer(struct sctp_inpcb *inp, struct sctp_tcb *stcb,
1525     struct sctp_nets *net)
1526 {
1527 	struct sctp_nets *alt;
1528 
1529 	/* first threshold managment */
1530 	if (sctp_threshold_management(inp, stcb, net, stcb->asoc.max_send_times)) {
1531 		/* Assoc is over */
1532 		return (1);
1533 	}
1534 	/* second select an alternative */
1535 	alt = sctp_find_alternate_net(stcb, net, 0);
1536 
1537 	/* third generate a shutdown into the queue for out net */
1538 	sctp_send_shutdown_ack(stcb, alt);
1539 
1540 	/* fourth restart timer */
1541 	sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWNACK, inp, stcb, alt);
1542 	return (0);
1543 }
1544 
1545 static void
1546 sctp_audit_stream_queues_for_size(struct sctp_inpcb *inp,
1547     struct sctp_tcb *stcb)
1548 {
1549 	struct sctp_stream_out *outs;
1550 	struct sctp_stream_queue_pending *sp;
1551 	unsigned int chks_in_queue = 0;
1552 	int being_filled = 0;
1553 
1554 	/*
1555 	 * This function is ONLY called when the send/sent queues are empty.
1556 	 */
1557 	if ((stcb == NULL) || (inp == NULL))
1558 		return;
1559 
1560 	if (stcb->asoc.sent_queue_retran_cnt) {
1561 		SCTP_PRINTF("Hmm, sent_queue_retran_cnt is non-zero %d\n",
1562 		    stcb->asoc.sent_queue_retran_cnt);
1563 		stcb->asoc.sent_queue_retran_cnt = 0;
1564 	}
1565 	SCTP_TCB_SEND_LOCK(stcb);
1566 	if (TAILQ_EMPTY(&stcb->asoc.out_wheel)) {
1567 		int i, cnt = 0;
1568 
1569 		/* Check to see if a spoke fell off the wheel */
1570 		for (i = 0; i < stcb->asoc.streamoutcnt; i++) {
1571 			if (!TAILQ_EMPTY(&stcb->asoc.strmout[i].outqueue)) {
1572 				sctp_insert_on_wheel(stcb, &stcb->asoc, &stcb->asoc.strmout[i], 1);
1573 				cnt++;
1574 			}
1575 		}
1576 		if (cnt) {
1577 			/* yep, we lost a spoke or two */
1578 			SCTP_PRINTF("Found an additional %d streams NOT on outwheel, corrected\n", cnt);
1579 		} else {
1580 			/* no spokes lost, */
1581 			stcb->asoc.total_output_queue_size = 0;
1582 		}
1583 		SCTP_TCB_SEND_UNLOCK(stcb);
1584 		return;
1585 	}
1586 	SCTP_TCB_SEND_UNLOCK(stcb);
1587 	/* Check to see if some data queued, if so report it */
1588 	TAILQ_FOREACH(outs, &stcb->asoc.out_wheel, next_spoke) {
1589 		if (!TAILQ_EMPTY(&outs->outqueue)) {
1590 			TAILQ_FOREACH(sp, &outs->outqueue, next) {
1591 				if (sp->msg_is_complete)
1592 					being_filled++;
1593 				chks_in_queue++;
1594 			}
1595 		}
1596 	}
1597 	if (chks_in_queue != stcb->asoc.stream_queue_cnt) {
1598 		SCTP_PRINTF("Hmm, stream queue cnt at %d I counted %d in stream out wheel\n",
1599 		    stcb->asoc.stream_queue_cnt, chks_in_queue);
1600 	}
1601 	if (chks_in_queue) {
1602 		/* call the output queue function */
1603 		sctp_chunk_output(inp, stcb, SCTP_OUTPUT_FROM_T3, SCTP_SO_NOT_LOCKED);
1604 		if ((TAILQ_EMPTY(&stcb->asoc.send_queue)) &&
1605 		    (TAILQ_EMPTY(&stcb->asoc.sent_queue))) {
1606 			/*
1607 			 * Probably should go in and make it go back through
1608 			 * and add fragments allowed
1609 			 */
1610 			if (being_filled == 0) {
1611 				SCTP_PRINTF("Still nothing moved %d chunks are stuck\n",
1612 				    chks_in_queue);
1613 			}
1614 		}
1615 	} else {
1616 		SCTP_PRINTF("Found no chunks on any queue tot:%lu\n",
1617 		    (u_long)stcb->asoc.total_output_queue_size);
1618 		stcb->asoc.total_output_queue_size = 0;
1619 	}
1620 }
1621 
1622 int
1623 sctp_heartbeat_timer(struct sctp_inpcb *inp, struct sctp_tcb *stcb,
1624     struct sctp_nets *net, int cnt_of_unconf)
1625 {
1626 	int ret;
1627 
1628 	if (net) {
1629 		if (net->hb_responded == 0) {
1630 			if (net->ro._s_addr) {
1631 				/*
1632 				 * Invalidate the src address if we did not
1633 				 * get a response last time.
1634 				 */
1635 				sctp_free_ifa(net->ro._s_addr);
1636 				net->ro._s_addr = NULL;
1637 				net->src_addr_selected = 0;
1638 			}
1639 			sctp_backoff_on_timeout(stcb, net, 1, 0);
1640 		}
1641 		/* Zero PBA, if it needs it */
1642 		if (net->partial_bytes_acked) {
1643 			net->partial_bytes_acked = 0;
1644 		}
1645 	}
1646 	if ((stcb->asoc.total_output_queue_size > 0) &&
1647 	    (TAILQ_EMPTY(&stcb->asoc.send_queue)) &&
1648 	    (TAILQ_EMPTY(&stcb->asoc.sent_queue))) {
1649 		sctp_audit_stream_queues_for_size(inp, stcb);
1650 	}
1651 	/* Send a new HB, this will do threshold managment, pick a new dest */
1652 	if (cnt_of_unconf == 0) {
1653 		if (sctp_send_hb(stcb, 0, NULL) < 0) {
1654 			return (1);
1655 		}
1656 	} else {
1657 		/*
1658 		 * this will send out extra hb's up to maxburst if there are
1659 		 * any unconfirmed addresses.
1660 		 */
1661 		uint32_t cnt_sent = 0;
1662 
1663 		TAILQ_FOREACH(net, &stcb->asoc.nets, sctp_next) {
1664 			if ((net->dest_state & SCTP_ADDR_UNCONFIRMED) &&
1665 			    (net->dest_state & SCTP_ADDR_REACHABLE)) {
1666 				cnt_sent++;
1667 				if (net->hb_responded == 0) {
1668 					/* Did we respond last time? */
1669 					if (net->ro._s_addr) {
1670 						sctp_free_ifa(net->ro._s_addr);
1671 						net->ro._s_addr = NULL;
1672 						net->src_addr_selected = 0;
1673 					}
1674 				}
1675 				ret = sctp_send_hb(stcb, 1, net);
1676 				if (ret < 0)
1677 					return 1;
1678 				else if (ret == 0) {
1679 					break;
1680 				}
1681 				if (cnt_sent >= SCTP_BASE_SYSCTL(sctp_hb_maxburst))
1682 					break;
1683 			}
1684 		}
1685 	}
1686 	return (0);
1687 }
1688 
1689 int
1690 sctp_is_hb_timer_running(struct sctp_tcb *stcb)
1691 {
1692 	if (SCTP_OS_TIMER_PENDING(&stcb->asoc.hb_timer.timer)) {
1693 		/* its running */
1694 		return (1);
1695 	} else {
1696 		/* nope */
1697 		return (0);
1698 	}
1699 }
1700 
1701 int
1702 sctp_is_sack_timer_running(struct sctp_tcb *stcb)
1703 {
1704 	if (SCTP_OS_TIMER_PENDING(&stcb->asoc.dack_timer.timer)) {
1705 		/* its running */
1706 		return (1);
1707 	} else {
1708 		/* nope */
1709 		return (0);
1710 	}
1711 }
1712 
1713 #define SCTP_NUMBER_OF_MTU_SIZES 18
1714 static uint32_t mtu_sizes[] = {
1715 	68,
1716 	296,
1717 	508,
1718 	512,
1719 	544,
1720 	576,
1721 	1006,
1722 	1492,
1723 	1500,
1724 	1536,
1725 	2002,
1726 	2048,
1727 	4352,
1728 	4464,
1729 	8166,
1730 	17914,
1731 	32000,
1732 	65535
1733 };
1734 
1735 
1736 static uint32_t
1737 sctp_getnext_mtu(struct sctp_inpcb *inp, uint32_t cur_mtu)
1738 {
1739 	/* select another MTU that is just bigger than this one */
1740 	int i;
1741 
1742 	for (i = 0; i < SCTP_NUMBER_OF_MTU_SIZES; i++) {
1743 		if (cur_mtu < mtu_sizes[i]) {
1744 			/* no max_mtu is bigger than this one */
1745 			return (mtu_sizes[i]);
1746 		}
1747 	}
1748 	/* here return the highest allowable */
1749 	return (cur_mtu);
1750 }
1751 
1752 
1753 void
1754 sctp_pathmtu_timer(struct sctp_inpcb *inp,
1755     struct sctp_tcb *stcb,
1756     struct sctp_nets *net)
1757 {
1758 	uint32_t next_mtu, mtu;
1759 
1760 	next_mtu = sctp_getnext_mtu(inp, net->mtu);
1761 
1762 	if ((next_mtu > net->mtu) && (net->port == 0)) {
1763 		if ((net->src_addr_selected == 0) ||
1764 		    (net->ro._s_addr == NULL) ||
1765 		    (net->ro._s_addr->localifa_flags & SCTP_BEING_DELETED)) {
1766 			if ((net->ro._s_addr != NULL) && (net->ro._s_addr->localifa_flags & SCTP_BEING_DELETED)) {
1767 				sctp_free_ifa(net->ro._s_addr);
1768 				net->ro._s_addr = NULL;
1769 				net->src_addr_selected = 0;
1770 			} else if (net->ro._s_addr == NULL) {
1771 #if defined(INET6) && defined(SCTP_EMBEDDED_V6_SCOPE)
1772 				if (net->ro._l_addr.sa.sa_family == AF_INET6) {
1773 					struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)&net->ro._l_addr;
1774 
1775 					/* KAME hack: embed scopeid */
1776 					(void)sa6_embedscope(sin6, MODULE_GLOBAL(MOD_INET6, ip6_use_defzone));
1777 				}
1778 #endif
1779 
1780 				net->ro._s_addr = sctp_source_address_selection(inp,
1781 				    stcb,
1782 				    (sctp_route_t *) & net->ro,
1783 				    net, 0, stcb->asoc.vrf_id);
1784 #if defined(INET6) && defined(SCTP_EMBEDDED_V6_SCOPE)
1785 				if (net->ro._l_addr.sa.sa_family == AF_INET6) {
1786 					struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)&net->ro._l_addr;
1787 
1788 					(void)sa6_recoverscope(sin6);
1789 				}
1790 #endif				/* INET6 */
1791 			}
1792 			if (net->ro._s_addr)
1793 				net->src_addr_selected = 1;
1794 		}
1795 		if (net->ro._s_addr) {
1796 			mtu = SCTP_GATHER_MTU_FROM_ROUTE(net->ro._s_addr, &net->ro._s_addr.sa, net->ro.ro_rt);
1797 			if (net->port) {
1798 				mtu -= sizeof(struct udphdr);
1799 			}
1800 			if (mtu > next_mtu) {
1801 				net->mtu = next_mtu;
1802 			}
1803 		}
1804 	}
1805 	/* restart the timer */
1806 	sctp_timer_start(SCTP_TIMER_TYPE_PATHMTURAISE, inp, stcb, net);
1807 }
1808 
1809 void
1810 sctp_autoclose_timer(struct sctp_inpcb *inp,
1811     struct sctp_tcb *stcb,
1812     struct sctp_nets *net)
1813 {
1814 	struct timeval tn, *tim_touse;
1815 	struct sctp_association *asoc;
1816 	int ticks_gone_by;
1817 
1818 	(void)SCTP_GETTIME_TIMEVAL(&tn);
1819 	if (stcb->asoc.sctp_autoclose_ticks &&
1820 	    sctp_is_feature_on(inp, SCTP_PCB_FLAGS_AUTOCLOSE)) {
1821 		/* Auto close is on */
1822 		asoc = &stcb->asoc;
1823 		/* pick the time to use */
1824 		if (asoc->time_last_rcvd.tv_sec >
1825 		    asoc->time_last_sent.tv_sec) {
1826 			tim_touse = &asoc->time_last_rcvd;
1827 		} else {
1828 			tim_touse = &asoc->time_last_sent;
1829 		}
1830 		/* Now has long enough transpired to autoclose? */
1831 		ticks_gone_by = SEC_TO_TICKS(tn.tv_sec - tim_touse->tv_sec);
1832 		if ((ticks_gone_by > 0) &&
1833 		    (ticks_gone_by >= (int)asoc->sctp_autoclose_ticks)) {
1834 			/*
1835 			 * autoclose time has hit, call the output routine,
1836 			 * which should do nothing just to be SURE we don't
1837 			 * have hanging data. We can then safely check the
1838 			 * queues and know that we are clear to send
1839 			 * shutdown
1840 			 */
1841 			sctp_chunk_output(inp, stcb, SCTP_OUTPUT_FROM_AUTOCLOSE_TMR, SCTP_SO_NOT_LOCKED);
1842 			/* Are we clean? */
1843 			if (TAILQ_EMPTY(&asoc->send_queue) &&
1844 			    TAILQ_EMPTY(&asoc->sent_queue)) {
1845 				/*
1846 				 * there is nothing queued to send, so I'm
1847 				 * done...
1848 				 */
1849 				if (SCTP_GET_STATE(asoc) != SCTP_STATE_SHUTDOWN_SENT) {
1850 					/* only send SHUTDOWN 1st time thru */
1851 					sctp_send_shutdown(stcb, stcb->asoc.primary_destination);
1852 					if ((SCTP_GET_STATE(asoc) == SCTP_STATE_OPEN) ||
1853 					    (SCTP_GET_STATE(asoc) == SCTP_STATE_SHUTDOWN_RECEIVED)) {
1854 						SCTP_STAT_DECR_GAUGE32(sctps_currestab);
1855 					}
1856 					SCTP_SET_STATE(asoc, SCTP_STATE_SHUTDOWN_SENT);
1857 					SCTP_CLEAR_SUBSTATE(asoc, SCTP_STATE_SHUTDOWN_PENDING);
1858 					sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWN,
1859 					    stcb->sctp_ep, stcb,
1860 					    asoc->primary_destination);
1861 					sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWNGUARD,
1862 					    stcb->sctp_ep, stcb,
1863 					    asoc->primary_destination);
1864 				}
1865 			}
1866 		} else {
1867 			/*
1868 			 * No auto close at this time, reset t-o to check
1869 			 * later
1870 			 */
1871 			int tmp;
1872 
1873 			/* fool the timer startup to use the time left */
1874 			tmp = asoc->sctp_autoclose_ticks;
1875 			asoc->sctp_autoclose_ticks -= ticks_gone_by;
1876 			sctp_timer_start(SCTP_TIMER_TYPE_AUTOCLOSE, inp, stcb,
1877 			    net);
1878 			/* restore the real tick value */
1879 			asoc->sctp_autoclose_ticks = tmp;
1880 		}
1881 	}
1882 }
1883 
1884 void
1885 sctp_iterator_timer(struct sctp_iterator *it)
1886 {
1887 	int iteration_count = 0;
1888 	int inp_skip = 0;
1889 
1890 	/*
1891 	 * only one iterator can run at a time. This is the only way we can
1892 	 * cleanly pull ep's from underneath all the running interators when
1893 	 * a ep is freed.
1894 	 */
1895 	SCTP_ITERATOR_LOCK();
1896 	if (it->inp == NULL) {
1897 		/* iterator is complete */
1898 done_with_iterator:
1899 		SCTP_ITERATOR_UNLOCK();
1900 		SCTP_INP_INFO_WLOCK();
1901 		TAILQ_REMOVE(&SCTP_BASE_INFO(iteratorhead), it, sctp_nxt_itr);
1902 		/* stopping the callout is not needed, in theory */
1903 		SCTP_INP_INFO_WUNLOCK();
1904 		(void)SCTP_OS_TIMER_STOP(&it->tmr.timer);
1905 		if (it->function_atend != NULL) {
1906 			(*it->function_atend) (it->pointer, it->val);
1907 		}
1908 		SCTP_FREE(it, SCTP_M_ITER);
1909 		return;
1910 	}
1911 select_a_new_ep:
1912 	SCTP_INP_WLOCK(it->inp);
1913 	while (((it->pcb_flags) &&
1914 	    ((it->inp->sctp_flags & it->pcb_flags) != it->pcb_flags)) ||
1915 	    ((it->pcb_features) &&
1916 	    ((it->inp->sctp_features & it->pcb_features) != it->pcb_features))) {
1917 		/* endpoint flags or features don't match, so keep looking */
1918 		if (it->iterator_flags & SCTP_ITERATOR_DO_SINGLE_INP) {
1919 			SCTP_INP_WUNLOCK(it->inp);
1920 			goto done_with_iterator;
1921 		}
1922 		SCTP_INP_WUNLOCK(it->inp);
1923 		it->inp = LIST_NEXT(it->inp, sctp_list);
1924 		if (it->inp == NULL) {
1925 			goto done_with_iterator;
1926 		}
1927 		SCTP_INP_WLOCK(it->inp);
1928 	}
1929 	if ((it->inp->inp_starting_point_for_iterator != NULL) &&
1930 	    (it->inp->inp_starting_point_for_iterator != it)) {
1931 		SCTP_PRINTF("Iterator collision, waiting for one at %p\n",
1932 		    it->inp);
1933 		SCTP_INP_WUNLOCK(it->inp);
1934 		goto start_timer_return;
1935 	}
1936 	/* mark the current iterator on the endpoint */
1937 	it->inp->inp_starting_point_for_iterator = it;
1938 	SCTP_INP_WUNLOCK(it->inp);
1939 	SCTP_INP_RLOCK(it->inp);
1940 	/* now go through each assoc which is in the desired state */
1941 	if (it->done_current_ep == 0) {
1942 		if (it->function_inp != NULL)
1943 			inp_skip = (*it->function_inp) (it->inp, it->pointer, it->val);
1944 		it->done_current_ep = 1;
1945 	}
1946 	if (it->stcb == NULL) {
1947 		/* run the per instance function */
1948 		it->stcb = LIST_FIRST(&it->inp->sctp_asoc_list);
1949 	}
1950 	SCTP_INP_RUNLOCK(it->inp);
1951 	if ((inp_skip) || it->stcb == NULL) {
1952 		if (it->function_inp_end != NULL) {
1953 			inp_skip = (*it->function_inp_end) (it->inp,
1954 			    it->pointer,
1955 			    it->val);
1956 		}
1957 		goto no_stcb;
1958 	}
1959 	if ((it->stcb) &&
1960 	    (it->stcb->asoc.stcb_starting_point_for_iterator == it)) {
1961 		it->stcb->asoc.stcb_starting_point_for_iterator = NULL;
1962 	}
1963 	while (it->stcb) {
1964 		SCTP_TCB_LOCK(it->stcb);
1965 		if (it->asoc_state && ((it->stcb->asoc.state & it->asoc_state) != it->asoc_state)) {
1966 			/* not in the right state... keep looking */
1967 			SCTP_TCB_UNLOCK(it->stcb);
1968 			goto next_assoc;
1969 		}
1970 		/* mark the current iterator on the assoc */
1971 		it->stcb->asoc.stcb_starting_point_for_iterator = it;
1972 		/* see if we have limited out the iterator loop */
1973 		iteration_count++;
1974 		if (iteration_count > SCTP_ITERATOR_MAX_AT_ONCE) {
1975 	start_timer_return:
1976 			/* set a timer to continue this later */
1977 			if (it->stcb)
1978 				SCTP_TCB_UNLOCK(it->stcb);
1979 			sctp_timer_start(SCTP_TIMER_TYPE_ITERATOR,
1980 			    (struct sctp_inpcb *)it, NULL, NULL);
1981 			SCTP_ITERATOR_UNLOCK();
1982 			return;
1983 		}
1984 		/* run function on this one */
1985 		(*it->function_assoc) (it->inp, it->stcb, it->pointer, it->val);
1986 
1987 		/*
1988 		 * we lie here, it really needs to have its own type but
1989 		 * first I must verify that this won't effect things :-0
1990 		 */
1991 		if (it->no_chunk_output == 0)
1992 			sctp_chunk_output(it->inp, it->stcb, SCTP_OUTPUT_FROM_T3, SCTP_SO_NOT_LOCKED);
1993 
1994 		SCTP_TCB_UNLOCK(it->stcb);
1995 next_assoc:
1996 		it->stcb = LIST_NEXT(it->stcb, sctp_tcblist);
1997 		if (it->stcb == NULL) {
1998 			if (it->function_inp_end != NULL) {
1999 				inp_skip = (*it->function_inp_end) (it->inp,
2000 				    it->pointer,
2001 				    it->val);
2002 			}
2003 		}
2004 	}
2005 no_stcb:
2006 	/* done with all assocs on this endpoint, move on to next endpoint */
2007 	it->done_current_ep = 0;
2008 	SCTP_INP_WLOCK(it->inp);
2009 	it->inp->inp_starting_point_for_iterator = NULL;
2010 	SCTP_INP_WUNLOCK(it->inp);
2011 	if (it->iterator_flags & SCTP_ITERATOR_DO_SINGLE_INP) {
2012 		it->inp = NULL;
2013 	} else {
2014 		SCTP_INP_INFO_RLOCK();
2015 		it->inp = LIST_NEXT(it->inp, sctp_list);
2016 		SCTP_INP_INFO_RUNLOCK();
2017 	}
2018 	if (it->inp == NULL) {
2019 		goto done_with_iterator;
2020 	}
2021 	goto select_a_new_ep;
2022 }
2023