| /linux/net/dccp/ |
| H A D | timer.c | |
| H A D | output.c | |
| H A D | diag.c | |
| H A D | minisocks.c | |
| H A D | input.c | |
| H A D | proto.c | |
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | bpf_iter_tcp4.c | 64 inet_csk_in_pingpong_mode(const struct inet_connection_sock *icsk) in inet_csk_in_pingpong_mode() argument 66 return icsk->icsk_ack.pingpong >= TCP_PINGPONG_THRESH; in inet_csk_in_pingpong_mode() 77 const struct inet_connection_sock *icsk; in dump_tcp_sock() local 88 icsk = &tp->inet_conn; in dump_tcp_sock() 89 inet = &icsk->icsk_inet; in dump_tcp_sock() 91 fastopenq = &icsk->icsk_accept_queue.fastopenq; in dump_tcp_sock() 98 if (icsk->icsk_pending == ICSK_TIME_RETRANS || in dump_tcp_sock() 99 icsk->icsk_pending == ICSK_TIME_REO_TIMEOUT || in dump_tcp_sock() 100 icsk->icsk_pending == ICSK_TIME_LOSS_PROBE) { in dump_tcp_sock() 103 } else if (icsk in dump_tcp_sock() [all...] |
| H A D | bpf_iter_tcp6.c | 64 inet_csk_in_pingpong_mode(const struct inet_connection_sock *icsk) in inet_csk_in_pingpong_mode() argument 66 return icsk->icsk_ack.pingpong >= TCP_PINGPONG_THRESH; in inet_csk_in_pingpong_mode() 77 const struct inet_connection_sock *icsk; in dump_tcp6_sock() local 88 icsk = &tp->tcp.inet_conn; in dump_tcp6_sock() 89 inet = &icsk->icsk_inet; in dump_tcp6_sock() 91 fastopenq = &icsk->icsk_accept_queue.fastopenq; in dump_tcp6_sock() 98 if (icsk->icsk_pending == ICSK_TIME_RETRANS || in dump_tcp6_sock() 99 icsk->icsk_pending == ICSK_TIME_REO_TIMEOUT || in dump_tcp6_sock() 100 icsk->icsk_pending == ICSK_TIME_LOSS_PROBE) { in dump_tcp6_sock() 103 } else if (icsk in dump_tcp6_sock() [all...] |
| H A D | sock_destroy_prog.c | 77 const struct inet_connection_sock *icsk; in iter_tcp6_server() local 92 icsk = &tcp_sk->tcp.inet_conn; in iter_tcp6_server() 93 inet = &icsk->icsk_inet; in iter_tcp6_server()
|
| /linux/include/net/ |
| H A D | inet_connection_sock.h | 200 icsk_delack_timeout(const struct inet_connection_sock *icsk) in icsk_delack_timeout() argument 202 return READ_ONCE(icsk->icsk_delack_timer.expires); in icsk_delack_timeout() 207 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timer() local 210 smp_store_release(&icsk->icsk_pending, 0); in inet_csk_clear_xmit_timer() 215 smp_store_release(&icsk->icsk_ack.pending, 0); in inet_csk_clear_xmit_timer() 216 icsk->icsk_ack.retry = 0; in inet_csk_clear_xmit_timer() 218 sk_stop_timer(sk, &icsk->icsk_delack_timer); in inet_csk_clear_xmit_timer() 232 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_reset_xmit_timer() local 243 smp_store_release(&icsk->icsk_pending, what); in inet_csk_reset_xmit_timer() 246 smp_store_release(&icsk->icsk_ack.pending, in inet_csk_reset_xmit_timer() [all …]
|
| H A D | espintcp.h | 35 const struct inet_connection_sock *icsk = inet_csk(sk); in espintcp_getctx() local 38 return (__force void *)icsk->icsk_ulp_data; in espintcp_getctx()
|
| H A D | tcp.h | 400 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_dec_quickack_mode() local 402 if (icsk->icsk_ack.quick) { in tcp_dec_quickack_mode() 406 if (pkts >= icsk->icsk_ack.quick) { in tcp_dec_quickack_mode() 407 icsk->icsk_ack.quick = 0; in tcp_dec_quickack_mode() 409 icsk->icsk_ack.ato = TCP_ATO_MIN; in tcp_dec_quickack_mode() 411 icsk->icsk_ack.quick -= pkts; in tcp_dec_quickack_mode() 1373 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_needs_ecn() local 1375 return icsk->icsk_ca_ops->flags & TCP_CONG_NEEDS_ECN; in tcp_ca_needs_ecn() 1380 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_needs_accecn() local 1382 return icsk->icsk_ca_ops->flags & TCP_CONG_NEEDS_ACCECN; in tcp_ca_needs_accecn() [all …]
|
| /linux/net/ipv4/ |
| H A D | tcp_ulp.c | 106 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_update_ulp() local 108 if (icsk->icsk_ulp_ops->update) in tcp_update_ulp() 109 icsk->icsk_ulp_ops->update(sk, proto, write_space); in tcp_update_ulp() 114 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_ulp() local 120 if (!icsk->icsk_ulp_ops) in tcp_cleanup_ulp() 123 if (icsk->icsk_ulp_ops->release) in tcp_cleanup_ulp() 124 icsk->icsk_ulp_ops->release(sk); in tcp_cleanup_ulp() 125 module_put(icsk->icsk_ulp_ops->owner); in tcp_cleanup_ulp() 127 icsk->icsk_ulp_ops = NULL; in tcp_cleanup_ulp() 132 struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_set_ulp() local [all …]
|
| H A D | tcp_output.c | 91 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_new_data_sent() 104 if (!prior_packets || icsk->icsk_pending == ICSK_TIME_LOSS_PROBE) in tcp_acceptable_seq() 188 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_ack_sent() 199 if ((u32)(now - icsk->icsk_ack.lrcvtime) < icsk->icsk_ack.ato) in tcp_event_ack_sent() 1515 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_transmit_skb() 1662 INDIRECT_CALL_INET(icsk->icsk_af_ops->send_check, in tcp_queue_skb() 1693 err = INDIRECT_CALL_INET(icsk->icsk_af_ops->queue_xmit, in tcp_adjust_pcount() 1976 const struct inet_connection_sock *icsk = inet_csk(sk); 1982 mss_now = pmtu - icsk 72 struct inet_connection_sock *icsk = inet_csk(sk); tcp_event_new_data_sent() local 169 struct inet_connection_sock *icsk = inet_csk(sk); tcp_event_data_sent() local 1453 const struct inet_connection_sock *icsk = inet_csk(sk); __tcp_transmit_skb() local 1914 const struct inet_connection_sock *icsk = inet_csk(sk); __tcp_mtu_to_mss() local 1948 const struct inet_connection_sock *icsk = inet_csk(sk); tcp_mss_to_mtu() local 1961 struct inet_connection_sock *icsk = inet_csk(sk); tcp_mtup_init() local 1998 struct inet_connection_sock *icsk = inet_csk(sk); tcp_sync_mss() local 2374 const struct inet_connection_sock *icsk = inet_csk(sk); tcp_tso_should_defer() local 2484 struct inet_connection_sock *icsk = inet_csk(sk); tcp_mtu_check_reprobe() local 2602 struct inet_connection_sock *icsk = inet_csk(sk); tcp_mtu_probe() local 3041 struct inet_connection_sock *icsk = inet_csk(sk); tcp_schedule_loss_probe() local 3256 struct inet_connection_sock *icsk = inet_csk(sk); __tcp_select_window() local 3492 struct inet_connection_sock *icsk = inet_csk(sk); __tcp_retransmit_skb() local 3664 const struct inet_connection_sock *icsk = inet_csk(sk); tcp_xmit_retransmit_queue() local 4047 struct inet_connection_sock *icsk = inet_csk(sk); tcp_ca_dst_init() local 4162 struct inet_connection_sock *icsk = inet_csk(sk); tcp_send_syn_data() local 4374 struct inet_connection_sock *icsk = inet_csk(sk); tcp_send_delayed_ack() local 4440 struct inet_connection_sock *icsk = inet_csk(sk); __tcp_send_ack() local 4565 struct inet_connection_sock *icsk = inet_csk(sk); tcp_send_probe0() local [all...] |
| H A D | tcp_input.c | 229 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_measure_rcv_mss() local 230 const unsigned int lss = icsk->icsk_ack.last_seg_size; in tcp_measure_rcv_mss() 233 icsk->icsk_ack.last_seg_size = 0; in tcp_measure_rcv_mss() 239 if (len >= icsk->icsk_ack.rcv_mss) { in tcp_measure_rcv_mss() 244 if (unlikely(len != icsk->icsk_ack.rcv_mss)) { in tcp_measure_rcv_mss() 261 icsk->icsk_ack.rcv_mss = min_t(unsigned int, len, in tcp_measure_rcv_mss() 264 DO_ONCE_LITE_IF(len > icsk->icsk_ack.rcv_mss + MAX_TCP_OPTION_SPACE, in tcp_measure_rcv_mss() 278 icsk->icsk_ack.pending |= ICSK_ACK_PUSHED; in tcp_measure_rcv_mss() 299 icsk->icsk_ack.last_seg_size = len; in tcp_measure_rcv_mss() 301 icsk->icsk_ack.rcv_mss = len; in tcp_measure_rcv_mss() [all …]
|
| H A D | tcp_htcp.c | 84 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_rtt() local 92 if (icsk->icsk_ca_state == TCP_CA_Open) { in measure_rtt() 104 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_achieved_throughput() local 109 if (icsk->icsk_ca_state == TCP_CA_Open) in measure_achieved_throughput() 119 if (!((1 << icsk->icsk_ca_state) & (TCPF_CA_Open | TCPF_CA_Disorder))) { in measure_achieved_throughput()
|
| H A D | tcp.c | 430 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_sock() local 440 icsk->icsk_rto = TCP_TIMEOUT_INIT; in tcp_init_sock() 443 icsk->icsk_rto_max = msecs_to_jiffies(rto_max_ms); in tcp_init_sock() 446 icsk->icsk_rto_min = usecs_to_jiffies(rto_min_us); in tcp_init_sock() 447 icsk->icsk_delack_max = TCP_DELACK_MAX; in tcp_init_sock() 478 icsk->icsk_sync_mss = tcp_sync_mss; in tcp_init_sock() 1571 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_cleanup_rbuf() local 1574 tp->rcv_nxt - tp->rcv_wup > icsk->icsk_ack.rcv_mss || in __tcp_cleanup_rbuf() 1582 ((icsk->icsk_ack.pending & ICSK_ACK_PUSHED2) || in __tcp_cleanup_rbuf() 1583 ((icsk->icsk_ack.pending & ICSK_ACK_PUSHED) && in __tcp_cleanup_rbuf() [all …]
|
| H A D | tcp_minisocks.c | 329 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_time_wait() local 338 const int rto = (icsk->icsk_rto << 2) - (icsk->icsk_rto >> 1); in tcp_time_wait() 501 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_openreq_child() local 511 icsk->icsk_ca_dst_locked = tcp_ca_dst_locked(dst); in tcp_ca_openreq_child() 512 icsk->icsk_ca_ops = ca; in tcp_ca_openreq_child() 520 (!icsk->icsk_ca_setsockopt || in tcp_ca_openreq_child() 521 !bpf_try_module_get(icsk->icsk_ca_ops, icsk->icsk_ca_ops->owner))) in tcp_ca_openreq_child()
|
| H A D | inet_diag.c | 209 int inet_sk_diag_fill(struct sock *sk, struct inet_connection_sock *icsk, in inet_sk_diag_fill() argument 278 if (!icsk) { in inet_sk_diag_fill() 283 icsk_pending = smp_load_acquire(&icsk->icsk_pending); in inet_sk_diag_fill() 288 r->idiag_retrans = READ_ONCE(icsk->icsk_retransmits); in inet_sk_diag_fill() 293 r->idiag_retrans = READ_ONCE(icsk->icsk_probes_out); in inet_sk_diag_fill() 296 } else if (timer_pending(&icsk->icsk_keepalive_timer)) { in inet_sk_diag_fill() 298 r->idiag_retrans = READ_ONCE(icsk->icsk_probes_out); in inet_sk_diag_fill() 300 jiffies_delta_to_msecs(icsk->icsk_keepalive_timer.expires - jiffies); in inet_sk_diag_fill() 317 ca_ops = READ_ONCE(icsk->icsk_ca_ops); in inet_sk_diag_fill() 337 ca_ops = READ_ONCE(icsk->icsk_ca_ops); in inet_sk_diag_fill()
|
| H A D | tcp_nv.c | 241 const struct inet_connection_sock *icsk = inet_csk(sk); in tcpnv_acked() local 255 if (icsk->icsk_ca_state != TCP_CA_Open && in tcpnv_acked() 256 icsk->icsk_ca_state != TCP_CA_Disorder) in tcpnv_acked()
|
| H A D | tcp_bic.c | 192 const struct inet_connection_sock *icsk = inet_csk(sk); in bictcp_acked() local 194 if (icsk->icsk_ca_state == TCP_CA_Open) { in bictcp_acked()
|
| /linux/net/mptcp/ |
| H A D | token_test.c | 34 struct inet_connection_sock *icsk; in build_icsk() local 36 icsk = kunit_kzalloc(test, sizeof(struct inet_connection_sock), in build_icsk() 38 KUNIT_EXPECT_NOT_ERR_OR_NULL(test, icsk); in build_icsk() 39 return icsk; in build_icsk() 73 struct inet_connection_sock *icsk = build_icsk(test); in mptcp_token_test_msk_basic() local 79 rcu_assign_pointer(icsk->icsk_ulp_data, ctx); in mptcp_token_test_msk_basic() 84 mptcp_token_new_connect((struct sock *)icsk)); in mptcp_token_test_msk_basic()
|
| /linux/net/psp/ |
| H A D | psp_sock.c | 181 struct inet_connection_sock *icsk; in psp_sock_assoc_set_tx() local 238 icsk = inet_csk(sk); in psp_sock_assoc_set_tx() 239 icsk->icsk_ext_hdr_len += psp_sk_overhead(sk); in psp_sock_assoc_set_tx() 240 icsk->icsk_sync_mss(sk, icsk->icsk_pmtu_cookie); in psp_sock_assoc_set_tx()
|
| /linux/net/tls/ |
| H A D | tls_toe.c | 48 struct inet_connection_sock *icsk = inet_csk(sk); in tls_toe_sk_destruct() local 53 rcu_assign_pointer(icsk->icsk_ulp_data, NULL); in tls_toe_sk_destruct()
|
| /linux/net/ipv6/ |
| H A D | tcp_ipv6.c | 140 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_v6_connect() local 223 u32 exthdrlen = icsk->icsk_ext_hdr_len; in tcp_v6_connect() 234 WRITE_ONCE(icsk->icsk_af_ops, &ipv6_mapped); in tcp_v6_connect() 245 icsk->icsk_ext_hdr_len = exthdrlen; in tcp_v6_connect() 247 WRITE_ONCE(icsk->icsk_af_ops, &ipv6_specific); in tcp_v6_connect() 305 icsk->icsk_ext_hdr_len = psp_sk_overhead(sk); in tcp_v6_connect() 307 icsk->icsk_ext_hdr_len += opt->opt_flen + in tcp_v6_connect() 2082 struct inet_connection_sock *icsk = inet_csk(sk); local 2086 icsk->icsk_af_ops = &ipv6_specific; 2138 const struct inet_connection_sock *icsk = inet_csk(sp); local [all …]
|