| /linux/net/ipv4/ |
| H A D | bpf_tcp_ca.c | 88 case offsetof(struct tcp_sock, snd_cwnd): in bpf_tcp_ca_btf_struct_access() 89 end = offsetofend(struct tcp_sock, snd_cwnd); in bpf_tcp_ca_btf_struct_access() 91 case offsetof(struct tcp_sock, snd_cwnd_cnt): in bpf_tcp_ca_btf_struct_access() 92 end = offsetofend(struct tcp_sock, snd_cwnd_cnt); in bpf_tcp_ca_btf_struct_access() 94 case offsetof(struct tcp_sock, snd_cwnd_stamp): in bpf_tcp_ca_btf_struct_access() 95 end = offsetofend(struct tcp_sock, snd_cwnd_stamp); in bpf_tcp_ca_btf_struct_access() 97 case offsetof(struct tcp_sock, snd_ssthresh): in bpf_tcp_ca_btf_struct_access() 98 end = offsetofend(struct tcp_sock, snd_ssthresh); in bpf_tcp_ca_btf_struct_access() 100 case offsetof(struct tcp_sock, ecn_flags): in bpf_tcp_ca_btf_struct_access() 101 end = offsetofend(struct tcp_sock, ecn_flags); in bpf_tcp_ca_btf_struct_access() [all …]
|
| H A D | tcp.c | 392 static u64 tcp_compute_delivery_rate(const struct tcp_sock *tp) in tcp_compute_delivery_rate() 408 struct tcp_sock *tp = tcp_sk(sk); in tcp_md5_destruct_sock() 427 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_sock() 515 const struct tcp_sock *tp = tcp_sk(sk); in tcp_stream_memory_free() 540 const struct tcp_sock *tp = tcp_sk(sk); in tcp_poll() 645 struct tcp_sock *tp = tcp_sk(sk); in tcp_ioctl() 689 void tcp_mark_push(struct tcp_sock *tp, struct sk_buff *skb) in tcp_mark_push() 695 static inline bool forced_push(const struct tcp_sock *tp) in forced_push() 702 struct tcp_sock *tp = tcp_sk(sk); in tcp_skb_entail() 718 static inline void tcp_mark_urg(struct tcp_sock *tp, int flags) in tcp_mark_urg() [all …]
|
| H A D | tcp_input.c | 124 void clean_acked_data_enable(struct tcp_sock *tp, in clean_acked_data_enable() 132 void clean_acked_data_disable(struct tcp_sock *tp) in clean_acked_data_disable() 252 struct tcp_sock *tp = tcp_sk(sk); in tcp_measure_rcv_mss() 346 struct tcp_sock *tp = tcp_sk(sk); in tcp_data_ecn_check() 389 static bool tcp_accecn_process_option(struct tcp_sock *tp, in tcp_accecn_process_option() 477 static void tcp_count_delivered_ce(struct tcp_sock *tp, u32 ecn_count) in tcp_count_delivered_ce() 483 static void tcp_count_delivered(struct tcp_sock *tp, u32 delivered, in tcp_count_delivered() 502 struct tcp_sock *tp = tcp_sk(sk); in __tcp_accecn_process() 585 struct tcp_sock *tp = tcp_sk(sk); in tcp_accecn_process() 607 const struct tcp_sock *tp = tcp_sk(sk); in tcp_sndbuf_expand() [all …]
|
| H A D | tcp_timer.c | 31 const struct tcp_sock *tp = tcp_sk(sk); in tcp_clamp_rto_to_user_timeout() 109 struct tcp_sock *tp = tcp_sk(sk); in tcp_out_of_resources() 219 struct tcp_sock *tp = tcp_sk(sk); in retransmits_timed_out() 246 struct tcp_sock *tp = tcp_sk(sk); in tcp_write_timeout() 312 struct tcp_sock *tp = tcp_sk(sk); in tcp_delack_timer_handler() 393 struct tcp_sock *tp = tcp_sk(sk); in tcp_probe_timer() 443 struct tcp_sock *tp = tcp_sk(sk); in tcp_update_rto_stats() 460 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_synack_timer() 500 const struct tcp_sock *tp = tcp_sk(sk); in tcp_rtx_probe0_timed_out() 537 struct tcp_sock *tp = tcp_sk(sk); in tcp_retransmit_timer() [all …]
|
| H A D | tcp_cdg.c | 143 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_hystart_update() 244 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_backoff() 265 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cong_avoid() 302 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_acked() 331 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_ssthresh() 348 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cwnd_event() 376 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_init()
|
| H A D | tcp_highspeed.c | 102 struct tcp_sock *tp = tcp_sk(sk); in hstcp_init() 114 struct tcp_sock *tp = tcp_sk(sk); in hstcp_cong_avoid() 153 const struct tcp_sock *tp = tcp_sk(sk); in hstcp_ssthresh()
|
| H A D | tcp_westwood.c | 165 const struct tcp_sock *tp = tcp_sk(sk); in westwood_fast_bw() 182 const struct tcp_sock *tp = tcp_sk(sk); in westwood_acked_count() 219 const struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_bw_rttmin() 242 struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_event()
|
| H A D | tcp_scalable.c | 20 struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_cong_avoid() 36 const struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_ssthresh()
|
| H A D | tcp_illinois.c | 59 struct tcp_sock *tp = tcp_sk(sk); in rtt_reset() 224 struct tcp_sock *tp = tcp_sk(sk); in update_params() 262 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_cong_avoid() 297 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_ssthresh()
|
| H A D | tcp_lp.c | 141 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_remote_hz_estimator() 195 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_owd_calculator() 276 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_pkts_acked()
|
| H A D | tcp_nv.c | 125 struct tcp_sock *tp = tcp_sk(sk); in tcpnv_reset() 182 struct tcp_sock *tp = tcp_sk(sk); in tcpnv_cong_avoid() 210 const struct tcp_sock *tp = tcp_sk(sk); in tcpnv_recalc_ssthresh() 242 struct tcp_sock *tp = tcp_sk(sk); in tcpnv_acked()
|
| H A D | tcp_minisocks.c | 294 const struct tcp_sock *tp = tcp_sk(sk); in tcp_time_wait_init() 329 struct tcp_sock *tp = tcp_sk(sk); in tcp_time_wait() 437 const struct tcp_sock *tp = tcp_sk(sk_listener); in tcp_openreq_init_rwin() 476 struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_openreq_child() 526 static void smc_check_reset_syn_req(const struct tcp_sock *oldtp, in smc_check_reset_syn_req() 528 struct tcp_sock *newtp) in smc_check_reset_syn_req() 555 const struct tcp_sock *oldtp; in tcp_create_openreq_child() 556 struct tcp_sock *newtp; in tcp_create_openreq_child()
|
| H A D | tcp_fastopen.c | 259 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_add_skb() 330 struct tcp_sock *tp; in tcp_fastopen_create_child() 541 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_defer_connect() 647 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_active_disable_ofo_check() 677 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_active_detect_blackhole()
|
| H A D | tcp_hybla.c | 48 struct tcp_sock *tp = tcp_sk(sk); in hybla_init() 92 struct tcp_sock *tp = tcp_sk(sk); in hybla_cong_avoid()
|
| H A D | tcp_htcp.c | 105 const struct tcp_sock *tp = tcp_sk(sk); in measure_achieved_throughput() 224 const struct tcp_sock *tp = tcp_sk(sk); in htcp_recalc_ssthresh() 233 struct tcp_sock *tp = tcp_sk(sk); in htcp_cong_avoid()
|
| /linux/include/net/ |
| H A D | tcp.h | 439 static inline bool tcp_ecn_mode_any(const struct tcp_sock *tp) in tcp_ecn_mode_any() 444 static inline bool tcp_ecn_mode_rfc3168(const struct tcp_sock *tp) in tcp_ecn_mode_rfc3168() 449 static inline bool tcp_ecn_mode_accecn(const struct tcp_sock *tp) in tcp_ecn_mode_accecn() 454 static inline bool tcp_ecn_disabled(const struct tcp_sock *tp) in tcp_ecn_disabled() 459 static inline bool tcp_ecn_mode_pending(const struct tcp_sock *tp) in tcp_ecn_mode_pending() 464 static inline void tcp_ecn_mode_set(struct tcp_sock *tp, u8 mode) in tcp_ecn_mode_set() 491 void tcp_clear_retrans(struct tcp_sock *tp); 750 void tcp_mark_push(struct tcp_sock *tp, struct sk_buff *skb); 831 static inline int tcp_bound_to_half_wnd(struct tcp_sock *tp, int pktsize) in tcp_bound_to_half_wnd() 883 static inline u32 __tcp_set_rto(const struct tcp_sock *tp) in __tcp_set_rto() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | bpf_dctcp.c | 53 static void dctcp_reset(const struct tcp_sock *tp, struct bpf_dctcp *ca) 64 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_reset() 112 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 121 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 151 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 170 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
|
| H A D | tcp_ca_write_sk_pacing.c | 11 static unsigned int tcp_left_out(const struct tcp_sock *tp) 16 static unsigned int tcp_packets_in_flight(const struct tcp_sock *tp) in tcp_left_out() 36 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
|
| H A D | bpf_tracing_net.h | 173 static inline struct tcp_sock *tcp_sk(const struct sock *sk) in tcp_is_cwnd_limited() 175 return (struct tcp_sock *)sk; 178 static inline bool tcp_in_slow_start(const struct tcp_sock *tp) 185 const struct tcp_sock *tp = tcp_sk(sk);
|
| H A D | bpf_cubic.c | 25 extern __u32 tcp_slow_start(struct tcp_sock *tp, __u32 acked) __ksym; 26 extern void tcp_cong_avoid_ai(struct tcp_sock *tp, __u32 w, __u32 acked) __ksym; 165 struct tcp_sock *tp = tcp_sk(sk); in bictcp_hystart_reset() 392 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 412 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 460 struct tcp_sock *tp = tcp_sk(sk); in hystart_update() 509 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
|
| H A D | bpf_cc_cubic.c | 40 const struct tcp_sock *tp = tcp_sk(sk); in before() 70 struct tcp_sock *tp = tcp_sk(sk); in tcp_update_pacing_rate() 120 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
|
| H A D | cgrp_ls_attach_cgroup.c | 27 struct tcp_sock *tcp_sk; in set_cookie() 56 struct tcp_sock *tcp_sk; in update_cookie_sockops()
|
| H A D | bpf_iter_tcp4.c | 69 static bool tcp_in_initial_slowstart(const struct tcp_sock *tcp) in tcp_in_initial_slowstart() 74 static int dump_tcp_sock(struct seq_file *seq, struct tcp_sock *tp, in dump_tcp_sock() 202 struct tcp_sock *tp; in dump_tcp4()
|
| /linux/net/mptcp/ |
| H A D | fastopen.c | 14 struct tcp_sock *tp; in mptcp_fastopen_subflow_synack_set_params() 23 ssk = subflow->tcp_sock; in mptcp_fastopen_subflow_synack_set_params()
|
| /linux/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
| H A D | chtls_cm.c | 145 struct tcp_sock *tp; in assign_rxopt() 227 struct tcp_sock *tp; in chtls_send_abort() 257 struct tcp_sock *tp = tcp_sk(sk); in chtls_send_reset() 288 struct tcp_sock *tp = tcp_sk(sk); in tcp_uncork() 448 struct tcp_sock *tp; in chtls_disconnect() 951 struct tcp_sock *tp; in chtls_select_mss() 1015 const struct tcp_sock *tp; in chtls_pass_accept_rpl() 1124 struct tcp_sock *tp; in chtls_recv_sock() 1465 struct tcp_sock *tp = tcp_sk(sk); in make_established() 1640 struct tcp_sock *tp = tcp_sk(sk); in handle_urg_ptr() [all …]
|