Lines Matching refs:inet_csk
229 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_measure_rcv_mss()
313 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_incr_quickack()
325 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_enter_quickack_mode()
338 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_in_quickack_mode()
538 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in __tcp_accecn_process()
555 ((1 << inet_csk(sk)->icsk_ca_state) & in __tcp_accecn_process()
608 const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops; in tcp_sndbuf_expand()
673 return 2 * inet_csk(sk)->icsk_ack.rcv_mss; in __tcp_grow_window()
727 inet_csk(sk)->icsk_ack.quick |= 1; in tcp_grow_window()
781 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_window()
815 inet_csk(sk)->icsk_ack.rcv_mss = hint; in tcp_initialize_rcv_mss()
903 TCP_SKB_CB(skb)->seq >= inet_csk(sk)->icsk_ack.rcv_mss) { in tcp_rcv_rtt_measure_ts()
998 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_save_lrcv_flowlabel()
1018 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_data_recv()
1188 inet_csk(sk)->icsk_rto = __tcp_set_rto(tp); in tcp_set_rto()
1290 tp->rx_opt.sack_ok, inet_csk(sk)->icsk_ca_state, in tcp_check_sack_reordering()
1784 inet_csk(sk)->icsk_ca_state, in tcp_rate_gen()
2400 if (inet_csk(sk)->icsk_ca_state != TCP_CA_Loss || tp->undo_marker) in tcp_sacktag_write_queue()
2556 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_enter_loss()
2598 !inet_csk(sk)->icsk_mtup.probe_size; in tcp_enter_loss()
2858 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_undo_cwnd_reduction()
2903 DBGUNDO(sk, inet_csk(sk)->icsk_ca_state == TCP_CA_Loss ? "loss" : "retrans"); in tcp_try_undo_recovery()
2905 if (inet_csk(sk)->icsk_ca_state == TCP_CA_Loss) in tcp_try_undo_recovery()
2950 WRITE_ONCE(inet_csk(sk)->icsk_retransmits, 0); in tcp_try_undo_loss()
2981 WRITE_ONCE(tp->snd_ssthresh, inet_csk(sk)->icsk_ca_ops->ssthresh(sk)); in tcp_init_cwnd_reduction()
3017 if (inet_csk(sk)->icsk_ca_ops->cong_control) in tcp_end_cwnd_reduction()
3022 (inet_csk(sk)->icsk_ca_state == TCP_CA_CWR || tp->undo_marker)) { in tcp_end_cwnd_reduction()
3035 if (inet_csk(sk)->icsk_ca_state < TCP_CA_CWR) { in tcp_enter_cwr()
3051 if (inet_csk(sk)->icsk_ca_state != state) { in tcp_try_keep_open()
3069 if (inet_csk(sk)->icsk_ca_state != TCP_CA_CWR) { in tcp_try_to_open()
3076 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_probe_failed()
3086 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_probe_success()
3118 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_non_congestion_loss_retransmit()
3331 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastretrans_alert()
3496 inet_csk(sk)->icsk_backoff = 0; in tcp_ack_update_rtt()
3515 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cong_avoid()
3526 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rearm_rto()
3538 u32 rto = inet_csk(sk)->icsk_rto; in tcp_rearm_rto()
3606 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clean_rtx_queue()
3786 icsk = inet_csk(sk); in tcp_clean_rtx_queue()
3809 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ack_probe()
3834 inet_csk(sk)->icsk_ca_state != TCP_CA_Open; in tcp_ack_is_dubious()
3861 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cong_control()
3979 tcp_sync_mss(sk, inet_csk(sk)->icsk_pmtu_cookie); in tcp_ack_update_window()
4144 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_in_ack_event()
4248 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ack()
4753 return inet_csk(sk)->icsk_rto * (USEC_PER_SEC / HZ); in tcp_tsval_replay()
4758 return inet_csk(sk)->icsk_rto * 1200 / HZ; in tcp_tsval_replay()
5019 if (inet_csk(sk)->icsk_ca_state != TCP_CA_Loss && in tcp_rcv_spurious_retrans()
5624 inet_csk(sk)->icsk_ack.pending |= in tcp_data_queue()
5650 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in tcp_data_queue()
6090 if (((tp->rcv_nxt - tp->rcv_wup) > inet_csk(sk)->icsk_ack.rcv_mss && in __tcp_ack_snd_check()
6101 inet_csk(sk)->icsk_ack.pending & ICSK_ACK_NOW) { in __tcp_ack_snd_check()
6479 inet_csk(sk)->icsk_af_ops->sk_rx_dst_set(sk, skb); in tcp_rcv_established()
6674 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_transfer()
6703 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_finish_connect()
6822 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rcv_synsent_state_process()
7078 if (inet_csk(sk)->icsk_ca_state == TCP_CA_Loss && !tp->packets_out) in tcp_rcv_synrecv_state_fastopen()
7082 WRITE_ONCE(inet_csk(sk)->icsk_retransmits, 0); in tcp_rcv_synrecv_state_fastopen()
7122 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rcv_state_process()
7251 if (!inet_csk(sk)->icsk_ca_ops->cong_control) in tcp_rcv_state_process()
7495 struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; in tcp_syn_flood_action()