/linux/net/tls/ |
H A D | trace.h | 17 TP_PROTO(struct sock *sk, int dir, u32 tcp_seq, u8 *rec_no, int ret), 19 TP_ARGS(sk, dir, tcp_seq, rec_no, ret), 25 __field( u32, tcp_seq ) 33 __entry->tcp_seq = tcp_seq; 39 __entry->sk, __entry->dir, __entry->tcp_seq, __entry->rec_no, 46 TP_PROTO(struct sock *sk, u32 tcp_seq, u8 *rec_no, u32 rec_len, 49 TP_ARGS(sk, tcp_seq, rec_no, rec_len, encrypted, decrypted), 54 __field( u32, tcp_seq ) 63 __entry->tcp_seq = tcp_seq; 71 __entry->sk, __entry->tcp_seq, [all …]
|
H A D | tls_device_fallback.c | 244 u32 tcp_seq = ntohl(tcp_hdr(skb)->seq); in fill_sg_in() local 251 record = tls_get_record(ctx, tcp_seq, rcd_sn); in fill_sg_in() 257 *sync_size = tcp_seq - tls_record_start_seq(record); in fill_sg_in()
|
/linux/drivers/net/ethernet/netronome/nfp/crypto/ |
H A D | fw.h | 14 __be32 tcp_seq; member 54 __be32 tcp_seq; member 90 __be32 tcp_seq; member
|
H A D | tls.c | 335 back->tcp_seq = cpu_to_be32(start_offload_tcp_sn); in nfp_net_tls_add() 445 req->tcp_seq = cpu_to_be32(seq); in nfp_net_tls_resync() 482 __be32 tcp_seq; in nfp_net_tls_rx_resync_req() local 531 memcpy(&tcp_seq, &req->tcp_seq, sizeof(tcp_seq)); in nfp_net_tls_rx_resync_req() 532 tls_offload_rx_resync_request(sk, tcp_seq); in nfp_net_tls_rx_resync_req()
|
/linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/ |
H A D | chcr_ktls.c | 839 struct sge_eth_txq *q, u64 tcp_seq, in chcr_ktls_xmit_tcb_cpls() argument 872 if (first_wr || tcp_seq != tx_info->prev_seq) { in chcr_ktls_xmit_tcb_cpls() 876 TCB_TX_MAX_V(tcp_seq), 0); in chcr_ktls_xmit_tcb_cpls() 880 if (tcp_seq != tx_info->prev_seq || offset) { in chcr_ktls_xmit_tcb_cpls() 889 if (tcp_seq != tx_info->prev_seq) in chcr_ktls_xmit_tcb_cpls() 1098 struct sge_eth_txq *q, u32 tcp_seq, in chcr_ktls_xmit_wr_complete() argument 1224 tx_data->rsvd = htonl(tcp_seq); in chcr_ktls_xmit_wr_complete() 1273 u32 tcp_seq, bool tcp_push, u32 mss, in chcr_ktls_xmit_wr_short() argument 1400 tx_data->rsvd = htonl(tcp_seq); in chcr_ktls_xmit_wr_short() 1458 struct sk_buff *skb, u32 tcp_seq, u32 mss, in chcr_ktls_tx_plaintxt() argument [all …]
|
/linux/net/core/ |
H A D | tso.c | 30 put_unaligned_be32(tso->tcp_seq, &tcph->seq); in tso_build_hdr() 48 tso->tcp_seq += size; /* not worth avoiding this operation for UDP */ in tso_build_data() 71 tso->tcp_seq = (tlen != sizeof(struct udphdr)) ? ntohl(tcp_hdr(skb)->seq) : 0; in tso_start()
|
/linux/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_ktls.c | 29 .tcp_seq = cpu_to_be32(start_offload_tcp_sn), in fun_ktls_add() 85 offsetof(struct fun_admin_ktls_modify_req, tcp_seq)); in fun_ktls_del() 114 req.tcp_seq = cpu_to_be32(seq); in fun_ktls_resync()
|
/linux/include/net/ |
H A D | tso.h | 17 u32 tcp_seq; member
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
H A D | ktls_tx.c | 623 u32 tcp_seq, int datalen, struct tx_sync_info *info) in tx_sync_info_get() argument 633 record = tls_get_record(tx_ctx, tcp_seq, &info->rcd_sn); in tx_sync_info_get() 648 ends_before = before(tcp_seq + datalen - 1, tls_record_start_seq(record)); in tx_sync_info_get() 658 info->sync_len = tcp_seq - tls_record_start_seq(record); in tx_sync_info_get()
|
/linux/drivers/net/ethernet/fungible/funcore/ |
H A D | fun_hci.h | 982 __be32 tcp_seq; member 1001 .tcp_seq = cpu_to_be32(_tcp_seq), .version = _version, \
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
H A D | mlx5hws_definer.h | 534 u8 tcp_seq[0x20]; member
|
H A D | mlx5hws_definer.c | 1117 misc_parameters_3.inner_tcp_seq_num, tcp_icmp.tcp_seq); in hws_definer_conv_misc3() 1126 misc_parameters_3.outer_tcp_seq_num, tcp_icmp.tcp_seq); in hws_definer_conv_misc3()
|