| /linux/samples/bpf/ |
| H A D | tcp_dumpstats_kern.c | 28 struct bpf_tcp_sock *tcp_sk; in _sockops() local 56 tcp_sk = bpf_tcp_sock(sk); in _sockops() 57 if (!tcp_sk) in _sockops() 63 tcp_sk->dsack_dups, tcp_sk->delivered); in _sockops() 65 tcp_sk->delivered_ce, tcp_sk->icsk_retransmits); in _sockops()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | cgrp_ls_attach_cgroup.c | 27 struct tcp_sock *tcp_sk; in set_cookie() local 37 tcp_sk = bpf_skc_to_tcp_sock(sk); in set_cookie() 38 if (!tcp_sk) in set_cookie() 42 tcp_sk->inet_conn.icsk_inet.sk.sk_cgrp_data.cgroup, 0, in set_cookie() 56 struct tcp_sock *tcp_sk; in update_cookie_sockops() local 66 tcp_sk = bpf_skc_to_tcp_sock(sk); in update_cookie_sockops() 67 if (!tcp_sk) in update_cookie_sockops() 71 tcp_sk->inet_conn.icsk_inet.sk.sk_cgrp_data.cgroup, 0, 0); in update_cookie_sockops()
|
| H A D | tcp_rtt.c | 29 struct bpf_tcp_sock *tcp_sk; in _sockops() local 50 tcp_sk = bpf_tcp_sock(sk); in _sockops() 51 if (!tcp_sk) in _sockops() 56 storage->dsack_dups = tcp_sk->dsack_dups; in _sockops() 57 storage->delivered = tcp_sk->delivered; in _sockops() 58 storage->delivered_ce = tcp_sk->delivered_ce; in _sockops() 59 storage->icsk_retransmits = tcp_sk->icsk_retransmits; in _sockops()
|
| H A D | bpf_dctcp.c | 64 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_reset() 112 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 121 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 151 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 170 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 199 *prior_rcv_nxt = tcp_sk(sk)->rcv_nxt; in dctcp_ece_ack_update() 228 return max(tcp_sk(sk)->snd_cwnd, ca->loss_cwnd);
|
| H A D | sock_destroy_prog.c | 79 struct tcp6_sock *tcp_sk; in iter_tcp6_server() local 88 tcp_sk = bpf_skc_to_tcp6_sock(sk_common); in iter_tcp6_server() 89 if (!tcp_sk) in iter_tcp6_server() 92 icsk = &tcp_sk->tcp.inet_conn; in iter_tcp6_server()
|
| H A D | tcp_ca_write_sk_pacing.c | 36 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 47 return tcp_sk(sk)->snd_ssthresh; in BPF_PROG() 53 return tcp_sk(sk)->snd_cwnd; in BPF_PROG()
|
| H A D | bpf_cubic.c | 160 return tcp_sk(sk)->tcp_mstamp; in bictcp_clock_us() 165 struct tcp_sock *tp = tcp_sk(sk); in bictcp_hystart_reset() 193 tcp_sk(sk)->snd_ssthresh = initial_ssthresh; in BPF_PROG() 208 delta = now - tcp_sk(sk)->lsndtime; in BPF_PROG() 392 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 412 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 460 struct tcp_sock *tp = tcp_sk(sk); in hystart_update() 509 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
|
| H A D | bpf_cc_cubic.c | 40 const struct tcp_sock *tp = tcp_sk(sk); in before() 70 struct tcp_sock *tp = tcp_sk(sk); in tcp_update_pacing_rate() 98 if (tcp_sk(sk)->reordering > TCP_REORDERING) in tcp_cwnd_reduction() 120 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
|
| H A D | tcp_ca_incompl_cong_ops.c | 12 return tcp_sk(sk)->snd_ssthresh; in BPF_PROG() 18 return tcp_sk(sk)->snd_cwnd; in BPF_PROG()
|
| H A D | tcp_ca_update.c | 33 return tcp_sk(sk)->snd_ssthresh; in BPF_PROG() 39 return tcp_sk(sk)->snd_cwnd; in BPF_PROG()
|
| /linux/net/ipv4/ |
| H A D | tcp_input.c | 149 bool unknown_opt = tcp_sk(sk)->rx_opt.saw_unknown && in bpf_skops_parse_hdr() 150 BPF_SOCK_OPS_TEST_FLAG(tcp_sk(sk), in bpf_skops_parse_hdr() 152 bool parse_all_opt = BPF_SOCK_OPS_TEST_FLAG(tcp_sk(sk), in bpf_skops_parse_hdr() 246 u8 old_ratio = tcp_sk(sk)->scaling_ratio; in tcp_measure_rcv_mss() 249 tcp_sk(sk)->scaling_ratio = val ? val : 1; in tcp_measure_rcv_mss() 251 if (old_ratio != tcp_sk(sk)->scaling_ratio) { in tcp_measure_rcv_mss() 252 struct tcp_sock *tp = tcp_sk(sk); in tcp_measure_rcv_mss() 262 tcp_sk(sk)->advmss); in tcp_measure_rcv_mss() 298 len -= tcp_sk(sk)->tcp_header_len; in tcp_measure_rcv_mss() 314 unsigned int quickacks = tcp_sk(sk)->rcv_wnd / (2 * icsk->icsk_ack.rcv_mss); in tcp_incr_quickack() [all …]
|
| H A D | tcp_timer.c | 31 const struct tcp_sock *tp = tcp_sk(sk); in tcp_clamp_rto_to_user_timeout() 109 struct tcp_sock *tp = tcp_sk(sk); in tcp_out_of_resources() 219 struct tcp_sock *tp = tcp_sk(sk); in retransmits_timed_out() 246 struct tcp_sock *tp = tcp_sk(sk); in tcp_write_timeout() 312 struct tcp_sock *tp = tcp_sk(sk); in tcp_delack_timer_handler() 372 !READ_ONCE(tcp_sk(sk)->compressed_ack)) in tcp_delack_timer() 393 struct tcp_sock *tp = tcp_sk(sk); in tcp_probe_timer() 443 struct tcp_sock *tp = tcp_sk(sk); in tcp_update_rto_stats() 460 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_synack_timer() 500 const struct tcp_sock *tp = tcp_sk(sk); in tcp_rtx_probe0_timed_out() [all …]
|
| H A D | tcp_westwood.c | 73 w->snd_una = tcp_sk(sk)->snd_una; in tcp_westwood_init() 127 w->snd_una = tcp_sk(sk)->snd_una; in westwood_update_window() 165 const struct tcp_sock *tp = tcp_sk(sk); in westwood_fast_bw() 182 const struct tcp_sock *tp = tcp_sk(sk); in westwood_acked_count() 219 const struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_bw_rttmin() 242 struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_event()
|
| H A D | tcp_fastopen.c | 59 RCU_INIT_POINTER(tcp_sk(sk)->fastopen_rsk, NULL); in reqsk_fastopen_remove() 259 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_add_skb() 348 tp = tcp_sk(child); in tcp_fastopen_create_child() 430 tcp_sk(sk)->fastopen_no_cookie || in tcp_fastopen_no_cookie() 494 tcp_sk(child)->syn_fastopen_child = 1; in tcp_try_fastopen() 527 tcp_sk(sk)->fastopen_client_fail = TFO_COOKIE_UNAVAILABLE; in tcp_fastopen_cookie_check() 541 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_defer_connect() 635 tcp_sk(sk)->syn_fastopen_ch = 1; in tcp_fastopen_active_should_disable() 647 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_active_disable_ofo_check() 677 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_active_detect_blackhole()
|
| H A D | tcp_cdg.c | 143 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_hystart_update() 244 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_backoff() 265 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cong_avoid() 302 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_acked() 331 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_ssthresh() 348 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cwnd_event() 376 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_init()
|
| H A D | tcp_highspeed.c | 102 struct tcp_sock *tp = tcp_sk(sk); in hstcp_init() 114 struct tcp_sock *tp = tcp_sk(sk); in hstcp_cong_avoid() 153 const struct tcp_sock *tp = tcp_sk(sk); in hstcp_ssthresh()
|
| H A D | tcp_scalable.c | 20 struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_cong_avoid() 36 const struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_ssthresh()
|
| H A D | tcp.c | 408 struct tcp_sock *tp = tcp_sk(sk); in tcp_md5_destruct_sock() 427 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_sock() 515 const struct tcp_sock *tp = tcp_sk(sk); in tcp_stream_memory_free() 540 const struct tcp_sock *tp = tcp_sk(sk); in tcp_poll() 645 struct tcp_sock *tp = tcp_sk(sk); in tcp_ioctl() 702 struct tcp_sock *tp = tcp_sk(sk); in tcp_skb_entail() 747 struct tcp_sock *tp = tcp_sk(sk); in tcp_push() 964 struct tcp_sock *tp = tcp_sk(sk); in tcp_xmit_size_goal() 1058 struct tcp_sock *tp = tcp_sk(sk); in tcp_sendmsg_fastopen() 1105 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_check_app_limited() [all …]
|
| H A D | tcp_hybla.c | 39 tcp_sk(sk)->srtt_us / (rtt0 * USEC_PER_MSEC), in hybla_recalc_param() 48 struct tcp_sock *tp = tcp_sk(sk); in hybla_init() 92 struct tcp_sock *tp = tcp_sk(sk); in hybla_cong_avoid()
|
| H A D | tcp_illinois.c | 59 struct tcp_sock *tp = tcp_sk(sk); in rtt_reset() 224 struct tcp_sock *tp = tcp_sk(sk); in update_params() 262 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_cong_avoid() 297 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_ssthresh()
|
| H A D | tcp_ao.c | 96 ao = rcu_dereference(tcp_sk(sk)->ao_info); in tcp_ao_ignore_icmp() 205 ao = rcu_dereference_check(tcp_sk(sk)->ao_info, in __tcp_ao_do_lookup() 304 ao = rcu_dereference_protected(tcp_sk(sk)->ao_info, 1); in tcp_ao_destroy_sock() 305 rcu_assign_pointer(tcp_sk(sk)->ao_info, NULL); in tcp_ao_destroy_sock() 768 ao_info = rcu_dereference(tcp_sk(sk)->ao_info); in tcp_ao_prepare_reset() 791 ao_info = rcu_dereference(tcp_sk(sk)->ao_info); in tcp_ao_prepare_reset() 792 snd_basis = tcp_sk(sk)->snd_una; in tcp_ao_prepare_reset() 814 struct tcp_sock *tp = tcp_sk(sk); in tcp_ao_transmit_skb() 820 ao = rcu_dereference_protected(tcp_sk(sk)->ao_info, in tcp_ao_transmit_skb() 958 info = rcu_dereference(tcp_sk(sk)->ao_info); in tcp_inbound_ao_hash() [all …]
|
| H A D | tcp_bic.c | 77 WRITE_ONCE(tcp_sk(sk)->snd_ssthresh, initial_ssthresh); in bictcp_init() 142 struct tcp_sock *tp = tcp_sk(sk); in bictcp_cong_avoid() 163 const struct tcp_sock *tp = tcp_sk(sk); in bictcp_recalc_ssthresh()
|
| H A D | tcp_lp.c | 141 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_remote_hz_estimator() 195 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_owd_calculator() 276 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_pkts_acked()
|
| /linux/include/net/ |
| H A D | tcp.h | 653 last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp); in tcp_synq_overflow() 676 last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp); in tcp_synq_no_recent_overflow() 817 if (hrtimer_try_to_cancel(&tcp_sk(sk)->pacing_timer) == 1) in tcp_clear_xmit_timers() 820 if (hrtimer_try_to_cancel(&tcp_sk(sk)->compressed_ack_timer) == 1) in tcp_clear_xmit_timers() 1541 const struct tcp_sock *tp = tcp_sk(sk); in tcp_current_ssthresh() 1586 const struct tcp_sock *tp = tcp_sk(sk); in tcp_is_cwnd_limited() 1614 s64 delay = tcp_sk(sk)->tcp_wstamp_ns - tcp_sk(sk)->tcp_clock_cache; in tcp_pacing_delay() 1654 if (!tcp_sk(sk)->packets_out && !inet_csk(sk)->icsk_pending) in tcp_check_probe_timer() 1709 struct tcp_sock *tp = tcp_sk(sk); in tcp_slow_start_after_idle_check() 1735 return __tcp_win_from_space(tcp_sk(sk)->scaling_ratio, space); in tcp_win_from_space() [all …]
|
| /linux/net/tls/ |
| H A D | tls_device.c | 209 clean_acked_data_disable(tcp_sk(sk)); in tls_device_sk_destruct() 286 struct tcp_sock *tp = tcp_sk(sk); in tls_push_record() 877 tls_device_resync_rx(tls_ctx, sk, tcp_sk(sk)->copied_seq, in tls_device_core_ctrl_rx_resync() 990 trace_tls_device_decrypted(sk, tcp_sk(sk)->copied_seq - rxm->full_len, in tls_device_decrypted() 1129 start_marker_record->end_seq = tcp_sk(sk)->write_seq; in tls_set_device_offload() 1134 clean_acked_data_enable(tcp_sk(sk), &tls_tcp_clean_acked); in tls_set_device_offload() 1160 tcp_sk(sk)->write_seq); in tls_set_device_offload() 1162 tcp_sk(sk)->write_seq, rec_seq, rc); in tls_set_device_offload() 1180 clean_acked_data_disable(tcp_sk(sk)); in tls_set_device_offload() 1241 tcp_sk(sk)->copied_seq); in tls_set_device_offload_rx() [all …]
|