/linux/include/net/ |
H A D | inet_connection_sock.h | 124 } icsk_ack; member 179 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_SCHED; in inet_csk_schedule_ack() 184 return inet_csk(sk)->icsk_ack.pending & ICSK_ACK_SCHED; in inet_csk_ack_scheduled() 189 memset(&inet_csk(sk)->icsk_ack, 0, sizeof(inet_csk(sk)->icsk_ack)); in inet_csk_delack_init() 205 smp_store_release(&icsk->icsk_ack.pending, 0); in inet_csk_clear_xmit_timer() 206 icsk->icsk_ack.retry = 0; in inet_csk_clear_xmit_timer() 236 smp_store_release(&icsk->icsk_ack.pending, in inet_csk_reset_xmit_timer() 237 icsk->icsk_ack.pending | ICSK_ACK_TIMER); in inet_csk_reset_xmit_timer() 238 icsk->icsk_ack.timeout = jiffies + when; in inet_csk_reset_xmit_timer() 239 sk_reset_timer(sk, &icsk->icsk_delack_timer, icsk->icsk_ack.timeout); in inet_csk_reset_xmit_timer() [all …]
|
H A D | tcp.h | 362 if (icsk->icsk_ack.quick) { in tcp_dec_quickack_mode() 366 if (pkts >= icsk->icsk_ack.quick) { in tcp_dec_quickack_mode() 367 icsk->icsk_ack.quick = 0; in tcp_dec_quickack_mode() 369 icsk->icsk_ack.ato = TCP_ATO_MIN; in tcp_dec_quickack_mode() 371 icsk->icsk_ack.quick -= pkts; in tcp_dec_quickack_mode() 1636 (tcp_receive_window(tp) <= inet_csk(sk)->icsk_ack.rcv_mss); in tcp_epollin_ready() 1687 return min_t(u32, tcp_jiffies32 - icsk->icsk_ack.lrcvtime, in keepalive_time_elapsed()
|
/linux/net/dccp/ |
H A D | timer.c | 186 !(icsk->icsk_ack.pending & ICSK_ACK_TIMER)) in dccp_delack_timer() 188 if (time_after(icsk->icsk_ack.timeout, jiffies)) { in dccp_delack_timer() 190 icsk->icsk_ack.timeout); in dccp_delack_timer() 194 icsk->icsk_ack.pending &= ~ICSK_ACK_TIMER; in dccp_delack_timer() 199 icsk->icsk_ack.ato = min_t(u32, icsk->icsk_ack.ato << 1, in dccp_delack_timer() 206 icsk->icsk_ack.ato = TCP_ATO_MIN; in dccp_delack_timer()
|
H A D | output.c | 590 inet_csk(sk)->icsk_ack.ato = TCP_ATO_MIN; in dccp_send_ack() 619 if (icsk->icsk_ack.pending & ICSK_ACK_TIMER) { 625 if (icsk->icsk_ack.blocked) { 630 if (!time_before(timeout, icsk->icsk_ack.timeout)) 631 timeout = icsk->icsk_ack.timeout; 633 icsk->icsk_ack.pending |= ICSK_ACK_SCHED | ICSK_ACK_TIMER; 634 icsk->icsk_ack.timeout = timeout;
|
/linux/net/ipv4/ |
H A D | tcp_dctcp.h | 29 if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) { in dctcp_ece_ack_update() 33 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in dctcp_ece_ack_update()
|
H A D | tcp_timer.c | 322 if (!(icsk->icsk_ack.pending & ICSK_ACK_TIMER)) in tcp_delack_timer_handler() 325 if (time_after(icsk->icsk_ack.timeout, jiffies)) { in tcp_delack_timer_handler() 326 sk_reset_timer(sk, &icsk->icsk_delack_timer, icsk->icsk_ack.timeout); in tcp_delack_timer_handler() 329 icsk->icsk_ack.pending &= ~ICSK_ACK_TIMER; in tcp_delack_timer_handler() 334 icsk->icsk_ack.ato = min_t(u32, icsk->icsk_ack.ato << 1, icsk->icsk_rto); in tcp_delack_timer_handler() 340 icsk->icsk_ack.ato = TCP_ATO_MIN; in tcp_delack_timer_handler() 368 if (!(smp_load_acquire(&icsk->icsk_ack.pending) & ICSK_ACK_TIMER) && in tcp_delack_timer()
|
H A D | tcp_input.c | 225 const unsigned int lss = icsk->icsk_ack.last_seg_size; in tcp_measure_rcv_mss() 228 icsk->icsk_ack.last_seg_size = 0; in tcp_measure_rcv_mss() 234 if (len >= icsk->icsk_ack.rcv_mss) { in tcp_measure_rcv_mss() 239 if (unlikely(len != icsk->icsk_ack.rcv_mss)) { in tcp_measure_rcv_mss() 256 icsk->icsk_ack.rcv_mss = min_t(unsigned int, len, in tcp_measure_rcv_mss() 259 DO_ONCE_LITE_IF(len > icsk->icsk_ack.rcv_mss + MAX_TCP_OPTION_SPACE, in tcp_measure_rcv_mss() 273 icsk->icsk_ack.pending |= ICSK_ACK_PUSHED; in tcp_measure_rcv_mss() 294 icsk->icsk_ack.last_seg_size = len; in tcp_measure_rcv_mss() 296 icsk->icsk_ack.rcv_mss = len; in tcp_measure_rcv_mss() 300 if (icsk->icsk_ack.pending & ICSK_ACK_PUSHED) in tcp_measure_rcv_mss() [all …]
|
H A D | bpf_tcp_ca.c | 84 case offsetof(struct inet_connection_sock, icsk_ack.pending): in bpf_tcp_ca_btf_struct_access() 86 icsk_ack.pending); in bpf_tcp_ca_btf_struct_access()
|
H A D | tcp_output.c | 177 if ((u32)(now - icsk->icsk_ack.lrcvtime) < icsk->icsk_ack.ato) in tcp_event_data_sent() 270 if (unlikely(inet_csk(sk)->icsk_ack.pending & ICSK_ACK_NOMEM)) { in tcp_select_window() 3095 int mss = icsk->icsk_ack.rcv_mss; in __tcp_select_window() 3120 icsk->icsk_ack.quick = 0; in __tcp_select_window() 3179 icsk->icsk_ack.quick = 0; in __tcp_select_window() 4186 int ato = icsk->icsk_ack.ato; in tcp_send_delayed_ack() 4194 (icsk->icsk_ack.pending & ICSK_ACK_PUSHED)) in tcp_send_delayed_ack() 4220 if (icsk->icsk_ack.pending & ICSK_ACK_TIMER) { in tcp_send_delayed_ack() 4222 if (time_before_eq(icsk->icsk_ack.timeout, jiffies + (ato >> 2))) { in tcp_send_delayed_ack() 4227 if (!time_before(timeout, icsk->icsk_ack.timeout)) in tcp_send_delayed_ack() [all …]
|
H A D | tcp_minisocks.c | 559 newicsk->icsk_ack.lrcvtime = tcp_jiffies32; in tcp_create_openreq_child() 629 newicsk->icsk_ack.last_seg_size = skb->len - newtp->tcp_header_len; in tcp_create_openreq_child()
|
H A D | tcp.c | 1468 tp->rcv_nxt - tp->rcv_wup > icsk->icsk_ack.rcv_mss || in __tcp_cleanup_rbuf() 1476 ((icsk->icsk_ack.pending & ICSK_ACK_PUSHED2) || in __tcp_cleanup_rbuf() 1477 ((icsk->icsk_ack.pending & ICSK_ACK_PUSHED) && in __tcp_cleanup_rbuf() 3364 icsk->icsk_ack.rcv_mss = TCP_MIN_MSS; in tcp_disconnect() 3588 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_PUSHED; in __tcp_sock_set_quickack() 4133 info->tcpi_ato = jiffies_to_usecs(min_t(u32, icsk->icsk_ack.ato, in tcp_get_info() 4136 info->tcpi_rcv_mss = icsk->icsk_ack.rcv_mss; in tcp_get_info() 4146 info->tcpi_last_data_recv = jiffies_to_msecs(now - icsk->icsk_ack.lrcvtime); in tcp_get_info()
|
H A D | tcp_ipv4.c | 2950 jiffies_to_clock_t(icsk->icsk_ack.ato), 2951 (icsk->icsk_ack.quick << 1) | inet_csk_in_pingpong_mode(sk),
|
/linux/tools/testing/selftests/bpf/progs/ |
H A D | bpf_dctcp.c | 199 if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) { in dctcp_ece_ack_update() 203 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in dctcp_ece_ack_update()
|
/linux/include/linux/ |
H A D | dccp.h | 267 #define dccps_syn_rtt dccps_inet_connection.icsk_ack.lrcvtime
|
/linux/net/mptcp/ |
H A D | options.c | 435 icsk->icsk_ack.timeout = 0; in clear_3rdack_retransmission() 436 icsk->icsk_ack.ato = 0; in clear_3rdack_retransmission() 437 icsk->icsk_ack.pending &= ~(ICSK_ACK_SCHED | ICSK_ACK_TIMER); in clear_3rdack_retransmission()
|
H A D | protocol.c | 545 u8 ack_pending = READ_ONCE(icsk->icsk_ack.pending); in mptcp_subflow_could_cleanup() 550 READ_ONCE(icsk->icsk_ack.rcv_mss)) || in mptcp_subflow_could_cleanup() 3533 WARN_ON_ONCE(icsk->icsk_ack.pending & ICSK_ACK_TIMER); in schedule_3rdack_retransmission() 3534 smp_store_release(&icsk->icsk_ack.pending, in schedule_3rdack_retransmission() 3535 icsk->icsk_ack.pending | ICSK_ACK_SCHED | ICSK_ACK_TIMER); in schedule_3rdack_retransmission() 3536 icsk->icsk_ack.timeout = timeout; in schedule_3rdack_retransmission()
|
H A D | subflow.c | 1500 (tcp_sk(sk)->rcv_nxt - tcp_sk(sk)->rcv_wup) > inet_csk(sk)->icsk_ack.rcv_mss) in subflow_data_ready() 1501 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in subflow_data_ready()
|
/linux/net/ipv6/ |
H A D | tcp_ipv6.c | 2236 jiffies_to_clock_t(icsk->icsk_ack.ato), 2237 (icsk->icsk_ack.quick << 1) | inet_csk_in_pingpong_mode(sp),
|