xref: /linux/net/rxrpc/ar-internal.h (revision 9ee0034b8f49aaaa7e7c2da8db1038915db99c19)
1 /* AF_RXRPC internal definitions
2  *
3  * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved.
4  * Written by David Howells (dhowells@redhat.com)
5  *
6  * This program is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU General Public License
8  * as published by the Free Software Foundation; either version
9  * 2 of the License, or (at your option) any later version.
10  */
11 
12 #include <linux/atomic.h>
13 #include <linux/seqlock.h>
14 #include <net/sock.h>
15 #include <net/af_rxrpc.h>
16 #include <rxrpc/packet.h>
17 
18 #if 0
19 #define CHECK_SLAB_OKAY(X)				     \
20 	BUG_ON(atomic_read((X)) >> (sizeof(atomic_t) - 2) == \
21 	       (POISON_FREE << 8 | POISON_FREE))
22 #else
23 #define CHECK_SLAB_OKAY(X) do {} while (0)
24 #endif
25 
26 #define FCRYPT_BSIZE 8
27 struct rxrpc_crypt {
28 	union {
29 		u8	x[FCRYPT_BSIZE];
30 		__be32	n[2];
31 	};
32 } __attribute__((aligned(8)));
33 
34 #define rxrpc_queue_work(WS)	queue_work(rxrpc_workqueue, (WS))
35 #define rxrpc_queue_delayed_work(WS,D)	\
36 	queue_delayed_work(rxrpc_workqueue, (WS), (D))
37 
38 struct rxrpc_connection;
39 
40 /*
41  * Mark applied to socket buffers.
42  */
43 enum rxrpc_skb_mark {
44 	RXRPC_SKB_MARK_DATA,		/* data message */
45 	RXRPC_SKB_MARK_FINAL_ACK,	/* final ACK received message */
46 	RXRPC_SKB_MARK_BUSY,		/* server busy message */
47 	RXRPC_SKB_MARK_REMOTE_ABORT,	/* remote abort message */
48 	RXRPC_SKB_MARK_LOCAL_ABORT,	/* local abort message */
49 	RXRPC_SKB_MARK_NET_ERROR,	/* network error message */
50 	RXRPC_SKB_MARK_LOCAL_ERROR,	/* local error message */
51 	RXRPC_SKB_MARK_NEW_CALL,	/* local error message */
52 };
53 
54 /*
55  * sk_state for RxRPC sockets
56  */
57 enum {
58 	RXRPC_UNBOUND = 0,
59 	RXRPC_CLIENT_UNBOUND,		/* Unbound socket used as client */
60 	RXRPC_CLIENT_BOUND,		/* client local address bound */
61 	RXRPC_SERVER_BOUND,		/* server local address bound */
62 	RXRPC_SERVER_LISTENING,		/* server listening for connections */
63 	RXRPC_CLOSE,			/* socket is being closed */
64 };
65 
66 /*
67  * Service backlog preallocation.
68  *
69  * This contains circular buffers of preallocated peers, connections and calls
70  * for incoming service calls and their head and tail pointers.  This allows
71  * calls to be set up in the data_ready handler, thereby avoiding the need to
72  * shuffle packets around so much.
73  */
74 struct rxrpc_backlog {
75 	unsigned short		peer_backlog_head;
76 	unsigned short		peer_backlog_tail;
77 	unsigned short		conn_backlog_head;
78 	unsigned short		conn_backlog_tail;
79 	unsigned short		call_backlog_head;
80 	unsigned short		call_backlog_tail;
81 #define RXRPC_BACKLOG_MAX	32
82 	struct rxrpc_peer	*peer_backlog[RXRPC_BACKLOG_MAX];
83 	struct rxrpc_connection	*conn_backlog[RXRPC_BACKLOG_MAX];
84 	struct rxrpc_call	*call_backlog[RXRPC_BACKLOG_MAX];
85 };
86 
87 /*
88  * RxRPC socket definition
89  */
90 struct rxrpc_sock {
91 	/* WARNING: sk has to be the first member */
92 	struct sock		sk;
93 	rxrpc_notify_new_call_t	notify_new_call; /* Func to notify of new call */
94 	rxrpc_discard_new_call_t discard_new_call; /* Func to discard a new call */
95 	struct rxrpc_local	*local;		/* local endpoint */
96 	struct hlist_node	listen_link;	/* link in the local endpoint's listen list */
97 	struct rxrpc_backlog	*backlog;	/* Preallocation for services */
98 	spinlock_t		incoming_lock;	/* Incoming call vs service shutdown lock */
99 	struct list_head	sock_calls;	/* List of calls owned by this socket */
100 	struct list_head	to_be_accepted;	/* calls awaiting acceptance */
101 	struct list_head	recvmsg_q;	/* Calls awaiting recvmsg's attention  */
102 	rwlock_t		recvmsg_lock;	/* Lock for recvmsg_q */
103 	struct key		*key;		/* security for this socket */
104 	struct key		*securities;	/* list of server security descriptors */
105 	struct rb_root		calls;		/* User ID -> call mapping */
106 	unsigned long		flags;
107 #define RXRPC_SOCK_CONNECTED		0	/* connect_srx is set */
108 	rwlock_t		call_lock;	/* lock for calls */
109 	u32			min_sec_level;	/* minimum security level */
110 #define RXRPC_SECURITY_MAX	RXRPC_SECURITY_ENCRYPT
111 	bool			exclusive;	/* Exclusive connection for a client socket */
112 	sa_family_t		family;		/* Protocol family created with */
113 	struct sockaddr_rxrpc	srx;		/* local address */
114 	struct sockaddr_rxrpc	connect_srx;	/* Default client address from connect() */
115 };
116 
117 #define rxrpc_sk(__sk) container_of((__sk), struct rxrpc_sock, sk)
118 
119 /*
120  * CPU-byteorder normalised Rx packet header.
121  */
122 struct rxrpc_host_header {
123 	u32		epoch;		/* client boot timestamp */
124 	u32		cid;		/* connection and channel ID */
125 	u32		callNumber;	/* call ID (0 for connection-level packets) */
126 	u32		seq;		/* sequence number of pkt in call stream */
127 	u32		serial;		/* serial number of pkt sent to network */
128 	u8		type;		/* packet type */
129 	u8		flags;		/* packet flags */
130 	u8		userStatus;	/* app-layer defined status */
131 	u8		securityIndex;	/* security protocol ID */
132 	union {
133 		u16	_rsvd;		/* reserved */
134 		u16	cksum;		/* kerberos security checksum */
135 	};
136 	u16		serviceId;	/* service ID */
137 } __packed;
138 
139 /*
140  * RxRPC socket buffer private variables
141  * - max 48 bytes (struct sk_buff::cb)
142  */
143 struct rxrpc_skb_priv {
144 	union {
145 		unsigned long	resend_at;	/* time in jiffies at which to resend */
146 		struct {
147 			u8	nr_jumbo;	/* Number of jumbo subpackets */
148 		};
149 	};
150 	union {
151 		unsigned int	offset;		/* offset into buffer of next read */
152 		int		remain;		/* amount of space remaining for next write */
153 		u32		error;		/* network error code */
154 	};
155 
156 	struct rxrpc_host_header hdr;		/* RxRPC packet header from this packet */
157 };
158 
159 #define rxrpc_skb(__skb) ((struct rxrpc_skb_priv *) &(__skb)->cb)
160 
161 /*
162  * RxRPC security module interface
163  */
164 struct rxrpc_security {
165 	const char		*name;		/* name of this service */
166 	u8			security_index;	/* security type provided */
167 
168 	/* Initialise a security service */
169 	int (*init)(void);
170 
171 	/* Clean up a security service */
172 	void (*exit)(void);
173 
174 	/* initialise a connection's security */
175 	int (*init_connection_security)(struct rxrpc_connection *);
176 
177 	/* prime a connection's packet security */
178 	int (*prime_packet_security)(struct rxrpc_connection *);
179 
180 	/* impose security on a packet */
181 	int (*secure_packet)(struct rxrpc_call *,
182 			     struct sk_buff *,
183 			     size_t,
184 			     void *);
185 
186 	/* verify the security on a received packet */
187 	int (*verify_packet)(struct rxrpc_call *, struct sk_buff *,
188 			     unsigned int, unsigned int, rxrpc_seq_t, u16);
189 
190 	/* Locate the data in a received packet that has been verified. */
191 	void (*locate_data)(struct rxrpc_call *, struct sk_buff *,
192 			    unsigned int *, unsigned int *);
193 
194 	/* issue a challenge */
195 	int (*issue_challenge)(struct rxrpc_connection *);
196 
197 	/* respond to a challenge */
198 	int (*respond_to_challenge)(struct rxrpc_connection *,
199 				    struct sk_buff *,
200 				    u32 *);
201 
202 	/* verify a response */
203 	int (*verify_response)(struct rxrpc_connection *,
204 			       struct sk_buff *,
205 			       u32 *);
206 
207 	/* clear connection security */
208 	void (*clear)(struct rxrpc_connection *);
209 };
210 
211 /*
212  * RxRPC local transport endpoint description
213  * - owned by a single AF_RXRPC socket
214  * - pointed to by transport socket struct sk_user_data
215  */
216 struct rxrpc_local {
217 	struct rcu_head		rcu;
218 	atomic_t		usage;
219 	struct list_head	link;
220 	struct socket		*socket;	/* my UDP socket */
221 	struct work_struct	processor;
222 	struct hlist_head	services;	/* services listening on this endpoint */
223 	struct rw_semaphore	defrag_sem;	/* control re-enablement of IP DF bit */
224 	struct sk_buff_head	reject_queue;	/* packets awaiting rejection */
225 	struct sk_buff_head	event_queue;	/* endpoint event packets awaiting processing */
226 	struct rb_root		client_conns;	/* Client connections by socket params */
227 	spinlock_t		client_conns_lock; /* Lock for client_conns */
228 	spinlock_t		lock;		/* access lock */
229 	rwlock_t		services_lock;	/* lock for services list */
230 	int			debug_id;	/* debug ID for printks */
231 	bool			dead;
232 	struct sockaddr_rxrpc	srx;		/* local address */
233 };
234 
235 /*
236  * RxRPC remote transport endpoint definition
237  * - matched by local endpoint, remote port, address and protocol type
238  */
239 struct rxrpc_peer {
240 	struct rcu_head		rcu;		/* This must be first */
241 	atomic_t		usage;
242 	unsigned long		hash_key;
243 	struct hlist_node	hash_link;
244 	struct rxrpc_local	*local;
245 	struct hlist_head	error_targets;	/* targets for net error distribution */
246 	struct work_struct	error_distributor;
247 	struct rb_root		service_conns;	/* Service connections */
248 	seqlock_t		service_conn_lock;
249 	spinlock_t		lock;		/* access lock */
250 	unsigned int		if_mtu;		/* interface MTU for this peer */
251 	unsigned int		mtu;		/* network MTU for this peer */
252 	unsigned int		maxdata;	/* data size (MTU - hdrsize) */
253 	unsigned short		hdrsize;	/* header size (IP + UDP + RxRPC) */
254 	int			debug_id;	/* debug ID for printks */
255 	int			error_report;	/* Net (+0) or local (+1000000) to distribute */
256 #define RXRPC_LOCAL_ERROR_OFFSET 1000000
257 	struct sockaddr_rxrpc	srx;		/* remote address */
258 
259 	/* calculated RTT cache */
260 #define RXRPC_RTT_CACHE_SIZE 32
261 	suseconds_t		rtt;		/* current RTT estimate (in uS) */
262 	unsigned int		rtt_point;	/* next entry at which to insert */
263 	unsigned int		rtt_usage;	/* amount of cache actually used */
264 	suseconds_t		rtt_cache[RXRPC_RTT_CACHE_SIZE]; /* calculated RTT cache */
265 };
266 
267 /*
268  * Keys for matching a connection.
269  */
270 struct rxrpc_conn_proto {
271 	union {
272 		struct {
273 			u32	epoch;		/* epoch of this connection */
274 			u32	cid;		/* connection ID */
275 		};
276 		u64		index_key;
277 	};
278 };
279 
280 struct rxrpc_conn_parameters {
281 	struct rxrpc_local	*local;		/* Representation of local endpoint */
282 	struct rxrpc_peer	*peer;		/* Remote endpoint */
283 	struct key		*key;		/* Security details */
284 	bool			exclusive;	/* T if conn is exclusive */
285 	u16			service_id;	/* Service ID for this connection */
286 	u32			security_level;	/* Security level selected */
287 };
288 
289 /*
290  * Bits in the connection flags.
291  */
292 enum rxrpc_conn_flag {
293 	RXRPC_CONN_HAS_IDR,		/* Has a client conn ID assigned */
294 	RXRPC_CONN_IN_SERVICE_CONNS,	/* Conn is in peer->service_conns */
295 	RXRPC_CONN_IN_CLIENT_CONNS,	/* Conn is in local->client_conns */
296 	RXRPC_CONN_EXPOSED,		/* Conn has extra ref for exposure */
297 	RXRPC_CONN_DONT_REUSE,		/* Don't reuse this connection */
298 	RXRPC_CONN_COUNTED,		/* Counted by rxrpc_nr_client_conns */
299 };
300 
301 /*
302  * Events that can be raised upon a connection.
303  */
304 enum rxrpc_conn_event {
305 	RXRPC_CONN_EV_CHALLENGE,	/* Send challenge packet */
306 };
307 
308 /*
309  * The connection cache state.
310  */
311 enum rxrpc_conn_cache_state {
312 	RXRPC_CONN_CLIENT_INACTIVE,	/* Conn is not yet listed */
313 	RXRPC_CONN_CLIENT_WAITING,	/* Conn is on wait list, waiting for capacity */
314 	RXRPC_CONN_CLIENT_ACTIVE,	/* Conn is on active list, doing calls */
315 	RXRPC_CONN_CLIENT_CULLED,	/* Conn is culled and delisted, doing calls */
316 	RXRPC_CONN_CLIENT_IDLE,		/* Conn is on idle list, doing mostly nothing */
317 };
318 
319 /*
320  * The connection protocol state.
321  */
322 enum rxrpc_conn_proto_state {
323 	RXRPC_CONN_UNUSED,		/* Connection not yet attempted */
324 	RXRPC_CONN_CLIENT,		/* Client connection */
325 	RXRPC_CONN_SERVICE_PREALLOC,	/* Service connection preallocation */
326 	RXRPC_CONN_SERVICE_UNSECURED,	/* Service unsecured connection */
327 	RXRPC_CONN_SERVICE_CHALLENGING,	/* Service challenging for security */
328 	RXRPC_CONN_SERVICE,		/* Service secured connection */
329 	RXRPC_CONN_REMOTELY_ABORTED,	/* Conn aborted by peer */
330 	RXRPC_CONN_LOCALLY_ABORTED,	/* Conn aborted locally */
331 	RXRPC_CONN__NR_STATES
332 };
333 
334 /*
335  * RxRPC connection definition
336  * - matched by { local, peer, epoch, conn_id, direction }
337  * - each connection can only handle four simultaneous calls
338  */
339 struct rxrpc_connection {
340 	struct rxrpc_conn_proto	proto;
341 	struct rxrpc_conn_parameters params;
342 
343 	atomic_t		usage;
344 	struct rcu_head		rcu;
345 	struct list_head	cache_link;
346 
347 	spinlock_t		channel_lock;
348 	unsigned char		active_chans;	/* Mask of active channels */
349 #define RXRPC_ACTIVE_CHANS_MASK	((1 << RXRPC_MAXCALLS) - 1)
350 	struct list_head	waiting_calls;	/* Calls waiting for channels */
351 	struct rxrpc_channel {
352 		struct rxrpc_call __rcu	*call;		/* Active call */
353 		u32			call_id;	/* ID of current call */
354 		u32			call_counter;	/* Call ID counter */
355 		u32			last_call;	/* ID of last call */
356 		u8			last_type;	/* Type of last packet */
357 		u16			last_service_id;
358 		union {
359 			u32		last_seq;
360 			u32		last_abort;
361 		};
362 	} channels[RXRPC_MAXCALLS];
363 
364 	struct work_struct	processor;	/* connection event processor */
365 	union {
366 		struct rb_node	client_node;	/* Node in local->client_conns */
367 		struct rb_node	service_node;	/* Node in peer->service_conns */
368 	};
369 	struct list_head	proc_link;	/* link in procfs list */
370 	struct list_head	link;		/* link in master connection list */
371 	struct sk_buff_head	rx_queue;	/* received conn-level packets */
372 	const struct rxrpc_security *security;	/* applied security module */
373 	struct key		*server_key;	/* security for this service */
374 	struct crypto_skcipher	*cipher;	/* encryption handle */
375 	struct rxrpc_crypt	csum_iv;	/* packet checksum base */
376 	unsigned long		flags;
377 	unsigned long		events;
378 	unsigned long		idle_timestamp;	/* Time at which last became idle */
379 	spinlock_t		state_lock;	/* state-change lock */
380 	enum rxrpc_conn_cache_state cache_state;
381 	enum rxrpc_conn_proto_state state;	/* current state of connection */
382 	u32			local_abort;	/* local abort code */
383 	u32			remote_abort;	/* remote abort code */
384 	int			debug_id;	/* debug ID for printks */
385 	atomic_t		serial;		/* packet serial number counter */
386 	unsigned int		hi_serial;	/* highest serial number received */
387 	u8			size_align;	/* data size alignment (for security) */
388 	u8			header_size;	/* rxrpc + security header size */
389 	u8			security_size;	/* security header size */
390 	u32			security_nonce;	/* response re-use preventer */
391 	u8			security_ix;	/* security type */
392 	u8			out_clientflag;	/* RXRPC_CLIENT_INITIATED if we are client */
393 };
394 
395 /*
396  * Flags in call->flags.
397  */
398 enum rxrpc_call_flag {
399 	RXRPC_CALL_RELEASED,		/* call has been released - no more message to userspace */
400 	RXRPC_CALL_HAS_USERID,		/* has a user ID attached */
401 	RXRPC_CALL_IS_SERVICE,		/* Call is service call */
402 	RXRPC_CALL_EXPOSED,		/* The call was exposed to the world */
403 	RXRPC_CALL_RX_LAST,		/* Received the last packet (at rxtx_top) */
404 	RXRPC_CALL_TX_LAST,		/* Last packet in Tx buffer (at rxtx_top) */
405 };
406 
407 /*
408  * Events that can be raised on a call.
409  */
410 enum rxrpc_call_event {
411 	RXRPC_CALL_EV_ACK,		/* need to generate ACK */
412 	RXRPC_CALL_EV_ABORT,		/* need to generate abort */
413 	RXRPC_CALL_EV_TIMER,		/* Timer expired */
414 	RXRPC_CALL_EV_RESEND,		/* Tx resend required */
415 };
416 
417 /*
418  * The states that a call can be in.
419  */
420 enum rxrpc_call_state {
421 	RXRPC_CALL_UNINITIALISED,
422 	RXRPC_CALL_CLIENT_AWAIT_CONN,	/* - client waiting for connection to become available */
423 	RXRPC_CALL_CLIENT_SEND_REQUEST,	/* - client sending request phase */
424 	RXRPC_CALL_CLIENT_AWAIT_REPLY,	/* - client awaiting reply */
425 	RXRPC_CALL_CLIENT_RECV_REPLY,	/* - client receiving reply phase */
426 	RXRPC_CALL_SERVER_PREALLOC,	/* - service preallocation */
427 	RXRPC_CALL_SERVER_SECURING,	/* - server securing request connection */
428 	RXRPC_CALL_SERVER_ACCEPTING,	/* - server accepting request */
429 	RXRPC_CALL_SERVER_RECV_REQUEST,	/* - server receiving request */
430 	RXRPC_CALL_SERVER_ACK_REQUEST,	/* - server pending ACK of request */
431 	RXRPC_CALL_SERVER_SEND_REPLY,	/* - server sending reply */
432 	RXRPC_CALL_SERVER_AWAIT_ACK,	/* - server awaiting final ACK */
433 	RXRPC_CALL_COMPLETE,		/* - call complete */
434 	NR__RXRPC_CALL_STATES
435 };
436 
437 /*
438  * Call completion condition (state == RXRPC_CALL_COMPLETE).
439  */
440 enum rxrpc_call_completion {
441 	RXRPC_CALL_SUCCEEDED,		/* - Normal termination */
442 	RXRPC_CALL_REMOTELY_ABORTED,	/* - call aborted by peer */
443 	RXRPC_CALL_LOCALLY_ABORTED,	/* - call aborted locally on error or close */
444 	RXRPC_CALL_LOCAL_ERROR,		/* - call failed due to local error */
445 	RXRPC_CALL_NETWORK_ERROR,	/* - call terminated by network error */
446 	NR__RXRPC_CALL_COMPLETIONS
447 };
448 
449 /*
450  * RxRPC call definition
451  * - matched by { connection, call_id }
452  */
453 struct rxrpc_call {
454 	struct rcu_head		rcu;
455 	struct rxrpc_connection	*conn;		/* connection carrying call */
456 	struct rxrpc_peer	*peer;		/* Peer record for remote address */
457 	struct rxrpc_sock __rcu	*socket;	/* socket responsible */
458 	unsigned long		ack_at;		/* When deferred ACK needs to happen */
459 	unsigned long		resend_at;	/* When next resend needs to happen */
460 	unsigned long		expire_at;	/* When the call times out */
461 	struct timer_list	timer;		/* Combined event timer */
462 	struct work_struct	processor;	/* Event processor */
463 	rxrpc_notify_rx_t	notify_rx;	/* kernel service Rx notification function */
464 	struct list_head	link;		/* link in master call list */
465 	struct list_head	chan_wait_link;	/* Link in conn->waiting_calls */
466 	struct hlist_node	error_link;	/* link in error distribution list */
467 	struct list_head	accept_link;	/* Link in rx->acceptq */
468 	struct list_head	recvmsg_link;	/* Link in rx->recvmsg_q */
469 	struct list_head	sock_link;	/* Link in rx->sock_calls */
470 	struct rb_node		sock_node;	/* Node in rx->calls */
471 	struct sk_buff		*tx_pending;	/* Tx socket buffer being filled */
472 	wait_queue_head_t	waitq;		/* Wait queue for channel or Tx */
473 	__be32			crypto_buf[2];	/* Temporary packet crypto buffer */
474 	unsigned long		user_call_ID;	/* user-defined call ID */
475 	unsigned long		flags;
476 	unsigned long		events;
477 	spinlock_t		lock;
478 	rwlock_t		state_lock;	/* lock for state transition */
479 	u32			abort_code;	/* Local/remote abort code */
480 	int			error;		/* Local error incurred */
481 	enum rxrpc_call_state	state;		/* current state of call */
482 	enum rxrpc_call_completion completion;	/* Call completion condition */
483 	atomic_t		usage;
484 	u16			service_id;	/* service ID */
485 	u8			security_ix;	/* Security type */
486 	u32			call_id;	/* call ID on connection  */
487 	u32			cid;		/* connection ID plus channel index */
488 	int			debug_id;	/* debug ID for printks */
489 
490 	/* Rx/Tx circular buffer, depending on phase.
491 	 *
492 	 * In the Rx phase, packets are annotated with 0 or the number of the
493 	 * segment of a jumbo packet each buffer refers to.  There can be up to
494 	 * 47 segments in a maximum-size UDP packet.
495 	 *
496 	 * In the Tx phase, packets are annotated with which buffers have been
497 	 * acked.
498 	 */
499 #define RXRPC_RXTX_BUFF_SIZE	64
500 #define RXRPC_RXTX_BUFF_MASK	(RXRPC_RXTX_BUFF_SIZE - 1)
501 	struct sk_buff		**rxtx_buffer;
502 	u8			*rxtx_annotations;
503 #define RXRPC_TX_ANNO_ACK	0
504 #define RXRPC_TX_ANNO_UNACK	1
505 #define RXRPC_TX_ANNO_NAK	2
506 #define RXRPC_TX_ANNO_RETRANS	3
507 #define RXRPC_RX_ANNO_JUMBO	0x3f		/* Jumbo subpacket number + 1 if not zero */
508 #define RXRPC_RX_ANNO_JLAST	0x40		/* Set if last element of a jumbo packet */
509 #define RXRPC_RX_ANNO_VERIFIED	0x80		/* Set if verified and decrypted */
510 	rxrpc_seq_t		tx_hard_ack;	/* Dead slot in buffer; the first transmitted but
511 						 * not hard-ACK'd packet follows this.
512 						 */
513 	rxrpc_seq_t		tx_top;		/* Highest Tx slot allocated. */
514 	rxrpc_seq_t		rx_hard_ack;	/* Dead slot in buffer; the first received but not
515 						 * consumed packet follows this.
516 						 */
517 	rxrpc_seq_t		rx_top;		/* Highest Rx slot allocated. */
518 	rxrpc_seq_t		rx_expect_next;	/* Expected next packet sequence number */
519 	u8			rx_winsize;	/* Size of Rx window */
520 	u8			tx_winsize;	/* Maximum size of Tx window */
521 	u8			nr_jumbo_dup;	/* Number of jumbo duplicates */
522 
523 	/* receive-phase ACK management */
524 	u8			ackr_reason;	/* reason to ACK */
525 	u16			ackr_skew;	/* skew on packet being ACK'd */
526 	rxrpc_serial_t		ackr_serial;	/* serial of packet being ACK'd */
527 	rxrpc_seq_t		ackr_prev_seq;	/* previous sequence number received */
528 	unsigned short		rx_pkt_offset;	/* Current recvmsg packet offset */
529 	unsigned short		rx_pkt_len;	/* Current recvmsg packet len */
530 
531 	/* transmission-phase ACK management */
532 	rxrpc_serial_t		acks_latest;	/* serial number of latest ACK received */
533 };
534 
535 enum rxrpc_call_trace {
536 	rxrpc_call_new_client,
537 	rxrpc_call_new_service,
538 	rxrpc_call_queued,
539 	rxrpc_call_queued_ref,
540 	rxrpc_call_seen,
541 	rxrpc_call_got,
542 	rxrpc_call_got_userid,
543 	rxrpc_call_put,
544 	rxrpc_call_put_userid,
545 	rxrpc_call_put_noqueue,
546 	rxrpc_call__nr_trace
547 };
548 
549 extern const char rxrpc_call_traces[rxrpc_call__nr_trace][4];
550 
551 #include <trace/events/rxrpc.h>
552 
553 /*
554  * af_rxrpc.c
555  */
556 extern atomic_t rxrpc_n_skbs;
557 extern u32 rxrpc_epoch;
558 extern atomic_t rxrpc_debug_id;
559 extern struct workqueue_struct *rxrpc_workqueue;
560 
561 /*
562  * call_accept.c
563  */
564 int rxrpc_service_prealloc(struct rxrpc_sock *, gfp_t);
565 void rxrpc_discard_prealloc(struct rxrpc_sock *);
566 struct rxrpc_call *rxrpc_new_incoming_call(struct rxrpc_local *,
567 					   struct rxrpc_connection *,
568 					   struct sk_buff *);
569 void rxrpc_accept_incoming_calls(struct rxrpc_local *);
570 struct rxrpc_call *rxrpc_accept_call(struct rxrpc_sock *, unsigned long,
571 				     rxrpc_notify_rx_t);
572 int rxrpc_reject_call(struct rxrpc_sock *);
573 
574 /*
575  * call_event.c
576  */
577 void rxrpc_propose_ACK(struct rxrpc_call *, u8, u16, u32, bool, bool);
578 void rxrpc_process_call(struct work_struct *);
579 
580 /*
581  * call_object.c
582  */
583 extern const char *const rxrpc_call_states[];
584 extern const char *const rxrpc_call_completions[];
585 extern unsigned int rxrpc_max_call_lifetime;
586 extern struct kmem_cache *rxrpc_call_jar;
587 extern struct list_head rxrpc_calls;
588 extern rwlock_t rxrpc_call_lock;
589 
590 struct rxrpc_call *rxrpc_find_call_by_user_ID(struct rxrpc_sock *, unsigned long);
591 struct rxrpc_call *rxrpc_alloc_call(gfp_t);
592 struct rxrpc_call *rxrpc_new_client_call(struct rxrpc_sock *,
593 					 struct rxrpc_conn_parameters *,
594 					 struct sockaddr_rxrpc *,
595 					 unsigned long, gfp_t);
596 void rxrpc_incoming_call(struct rxrpc_sock *, struct rxrpc_call *,
597 			 struct sk_buff *);
598 void rxrpc_release_call(struct rxrpc_sock *, struct rxrpc_call *);
599 void rxrpc_release_calls_on_socket(struct rxrpc_sock *);
600 bool __rxrpc_queue_call(struct rxrpc_call *);
601 bool rxrpc_queue_call(struct rxrpc_call *);
602 void rxrpc_see_call(struct rxrpc_call *);
603 void rxrpc_get_call(struct rxrpc_call *, enum rxrpc_call_trace);
604 void rxrpc_put_call(struct rxrpc_call *, enum rxrpc_call_trace);
605 void rxrpc_cleanup_call(struct rxrpc_call *);
606 void __exit rxrpc_destroy_all_calls(void);
607 
608 static inline bool rxrpc_is_service_call(const struct rxrpc_call *call)
609 {
610 	return test_bit(RXRPC_CALL_IS_SERVICE, &call->flags);
611 }
612 
613 static inline bool rxrpc_is_client_call(const struct rxrpc_call *call)
614 {
615 	return !rxrpc_is_service_call(call);
616 }
617 
618 /*
619  * Transition a call to the complete state.
620  */
621 static inline bool __rxrpc_set_call_completion(struct rxrpc_call *call,
622 					       enum rxrpc_call_completion compl,
623 					       u32 abort_code,
624 					       int error)
625 {
626 	if (call->state < RXRPC_CALL_COMPLETE) {
627 		call->abort_code = abort_code;
628 		call->error = error;
629 		call->completion = compl,
630 		call->state = RXRPC_CALL_COMPLETE;
631 		return true;
632 	}
633 	return false;
634 }
635 
636 static inline bool rxrpc_set_call_completion(struct rxrpc_call *call,
637 					     enum rxrpc_call_completion compl,
638 					     u32 abort_code,
639 					     int error)
640 {
641 	bool ret;
642 
643 	write_lock_bh(&call->state_lock);
644 	ret = __rxrpc_set_call_completion(call, compl, abort_code, error);
645 	write_unlock_bh(&call->state_lock);
646 	return ret;
647 }
648 
649 /*
650  * Record that a call successfully completed.
651  */
652 static inline bool __rxrpc_call_completed(struct rxrpc_call *call)
653 {
654 	return __rxrpc_set_call_completion(call, RXRPC_CALL_SUCCEEDED, 0, 0);
655 }
656 
657 static inline bool rxrpc_call_completed(struct rxrpc_call *call)
658 {
659 	bool ret;
660 
661 	write_lock_bh(&call->state_lock);
662 	ret = __rxrpc_call_completed(call);
663 	write_unlock_bh(&call->state_lock);
664 	return ret;
665 }
666 
667 /*
668  * Record that a call is locally aborted.
669  */
670 static inline bool __rxrpc_abort_call(const char *why, struct rxrpc_call *call,
671 				      rxrpc_seq_t seq,
672 				      u32 abort_code, int error)
673 {
674 	trace_rxrpc_abort(why, call->cid, call->call_id, seq,
675 			  abort_code, error);
676 	return __rxrpc_set_call_completion(call, RXRPC_CALL_LOCALLY_ABORTED,
677 					   abort_code, error);
678 }
679 
680 static inline bool rxrpc_abort_call(const char *why, struct rxrpc_call *call,
681 				    rxrpc_seq_t seq, u32 abort_code, int error)
682 {
683 	bool ret;
684 
685 	write_lock_bh(&call->state_lock);
686 	ret = __rxrpc_abort_call(why, call, seq, abort_code, error);
687 	write_unlock_bh(&call->state_lock);
688 	return ret;
689 }
690 
691 /*
692  * conn_client.c
693  */
694 extern unsigned int rxrpc_max_client_connections;
695 extern unsigned int rxrpc_reap_client_connections;
696 extern unsigned int rxrpc_conn_idle_client_expiry;
697 extern unsigned int rxrpc_conn_idle_client_fast_expiry;
698 extern struct idr rxrpc_client_conn_ids;
699 
700 void rxrpc_destroy_client_conn_ids(void);
701 int rxrpc_connect_call(struct rxrpc_call *, struct rxrpc_conn_parameters *,
702 		       struct sockaddr_rxrpc *, gfp_t);
703 void rxrpc_expose_client_call(struct rxrpc_call *);
704 void rxrpc_disconnect_client_call(struct rxrpc_call *);
705 void rxrpc_put_client_conn(struct rxrpc_connection *);
706 void __exit rxrpc_destroy_all_client_connections(void);
707 
708 /*
709  * conn_event.c
710  */
711 void rxrpc_process_connection(struct work_struct *);
712 
713 /*
714  * conn_object.c
715  */
716 extern unsigned int rxrpc_connection_expiry;
717 extern struct list_head rxrpc_connections;
718 extern struct list_head rxrpc_connection_proc_list;
719 extern rwlock_t rxrpc_connection_lock;
720 
721 int rxrpc_extract_addr_from_skb(struct sockaddr_rxrpc *, struct sk_buff *);
722 struct rxrpc_connection *rxrpc_alloc_connection(gfp_t);
723 struct rxrpc_connection *rxrpc_find_connection_rcu(struct rxrpc_local *,
724 						   struct sk_buff *);
725 void __rxrpc_disconnect_call(struct rxrpc_connection *, struct rxrpc_call *);
726 void rxrpc_disconnect_call(struct rxrpc_call *);
727 void rxrpc_kill_connection(struct rxrpc_connection *);
728 void __rxrpc_put_connection(struct rxrpc_connection *);
729 void __exit rxrpc_destroy_all_connections(void);
730 
731 static inline bool rxrpc_conn_is_client(const struct rxrpc_connection *conn)
732 {
733 	return conn->out_clientflag;
734 }
735 
736 static inline bool rxrpc_conn_is_service(const struct rxrpc_connection *conn)
737 {
738 	return !rxrpc_conn_is_client(conn);
739 }
740 
741 static inline void rxrpc_get_connection(struct rxrpc_connection *conn)
742 {
743 	atomic_inc(&conn->usage);
744 }
745 
746 static inline
747 struct rxrpc_connection *rxrpc_get_connection_maybe(struct rxrpc_connection *conn)
748 {
749 	return atomic_inc_not_zero(&conn->usage) ? conn : NULL;
750 }
751 
752 static inline void rxrpc_put_connection(struct rxrpc_connection *conn)
753 {
754 	if (!conn)
755 		return;
756 
757 	if (rxrpc_conn_is_client(conn)) {
758 		if (atomic_dec_and_test(&conn->usage))
759 			rxrpc_put_client_conn(conn);
760 	} else {
761 		if (atomic_dec_return(&conn->usage) == 1)
762 			__rxrpc_put_connection(conn);
763 	}
764 }
765 
766 static inline bool rxrpc_queue_conn(struct rxrpc_connection *conn)
767 {
768 	if (!rxrpc_get_connection_maybe(conn))
769 		return false;
770 	if (!rxrpc_queue_work(&conn->processor))
771 		rxrpc_put_connection(conn);
772 	return true;
773 }
774 
775 /*
776  * conn_service.c
777  */
778 struct rxrpc_connection *rxrpc_find_service_conn_rcu(struct rxrpc_peer *,
779 						     struct sk_buff *);
780 struct rxrpc_connection *rxrpc_prealloc_service_connection(gfp_t);
781 void rxrpc_new_incoming_connection(struct rxrpc_connection *, struct sk_buff *);
782 void rxrpc_unpublish_service_conn(struct rxrpc_connection *);
783 
784 /*
785  * input.c
786  */
787 void rxrpc_data_ready(struct sock *);
788 
789 /*
790  * insecure.c
791  */
792 extern const struct rxrpc_security rxrpc_no_security;
793 
794 /*
795  * key.c
796  */
797 extern struct key_type key_type_rxrpc;
798 extern struct key_type key_type_rxrpc_s;
799 
800 int rxrpc_request_key(struct rxrpc_sock *, char __user *, int);
801 int rxrpc_server_keyring(struct rxrpc_sock *, char __user *, int);
802 int rxrpc_get_server_data_key(struct rxrpc_connection *, const void *, time_t,
803 			      u32);
804 
805 /*
806  * local_event.c
807  */
808 extern void rxrpc_process_local_events(struct rxrpc_local *);
809 
810 /*
811  * local_object.c
812  */
813 struct rxrpc_local *rxrpc_lookup_local(const struct sockaddr_rxrpc *);
814 void __rxrpc_put_local(struct rxrpc_local *);
815 void __exit rxrpc_destroy_all_locals(void);
816 
817 static inline void rxrpc_get_local(struct rxrpc_local *local)
818 {
819 	atomic_inc(&local->usage);
820 }
821 
822 static inline
823 struct rxrpc_local *rxrpc_get_local_maybe(struct rxrpc_local *local)
824 {
825 	return atomic_inc_not_zero(&local->usage) ? local : NULL;
826 }
827 
828 static inline void rxrpc_put_local(struct rxrpc_local *local)
829 {
830 	if (local && atomic_dec_and_test(&local->usage))
831 		__rxrpc_put_local(local);
832 }
833 
834 static inline void rxrpc_queue_local(struct rxrpc_local *local)
835 {
836 	rxrpc_queue_work(&local->processor);
837 }
838 
839 /*
840  * misc.c
841  */
842 extern unsigned int rxrpc_max_backlog __read_mostly;
843 extern unsigned int rxrpc_requested_ack_delay;
844 extern unsigned int rxrpc_soft_ack_delay;
845 extern unsigned int rxrpc_idle_ack_delay;
846 extern unsigned int rxrpc_rx_window_size;
847 extern unsigned int rxrpc_rx_mtu;
848 extern unsigned int rxrpc_rx_jumbo_max;
849 extern unsigned int rxrpc_resend_timeout;
850 
851 extern const char *const rxrpc_pkts[];
852 extern const s8 rxrpc_ack_priority[];
853 
854 extern const char *rxrpc_acks(u8 reason);
855 
856 /*
857  * output.c
858  */
859 int rxrpc_send_call_packet(struct rxrpc_call *, u8);
860 int rxrpc_send_data_packet(struct rxrpc_connection *, struct sk_buff *);
861 void rxrpc_reject_packets(struct rxrpc_local *);
862 
863 /*
864  * peer_event.c
865  */
866 void rxrpc_error_report(struct sock *);
867 void rxrpc_peer_error_distributor(struct work_struct *);
868 
869 /*
870  * peer_object.c
871  */
872 struct rxrpc_peer *rxrpc_lookup_peer_rcu(struct rxrpc_local *,
873 					 const struct sockaddr_rxrpc *);
874 struct rxrpc_peer *rxrpc_lookup_peer(struct rxrpc_local *,
875 				     struct sockaddr_rxrpc *, gfp_t);
876 struct rxrpc_peer *rxrpc_alloc_peer(struct rxrpc_local *, gfp_t);
877 struct rxrpc_peer *rxrpc_lookup_incoming_peer(struct rxrpc_local *,
878 					      struct rxrpc_peer *);
879 
880 static inline struct rxrpc_peer *rxrpc_get_peer(struct rxrpc_peer *peer)
881 {
882 	atomic_inc(&peer->usage);
883 	return peer;
884 }
885 
886 static inline
887 struct rxrpc_peer *rxrpc_get_peer_maybe(struct rxrpc_peer *peer)
888 {
889 	return atomic_inc_not_zero(&peer->usage) ? peer : NULL;
890 }
891 
892 extern void __rxrpc_put_peer(struct rxrpc_peer *peer);
893 static inline void rxrpc_put_peer(struct rxrpc_peer *peer)
894 {
895 	if (peer && atomic_dec_and_test(&peer->usage))
896 		__rxrpc_put_peer(peer);
897 }
898 
899 /*
900  * proc.c
901  */
902 extern const struct file_operations rxrpc_call_seq_fops;
903 extern const struct file_operations rxrpc_connection_seq_fops;
904 
905 /*
906  * recvmsg.c
907  */
908 void rxrpc_notify_socket(struct rxrpc_call *);
909 int rxrpc_recvmsg(struct socket *, struct msghdr *, size_t, int);
910 
911 /*
912  * rxkad.c
913  */
914 #ifdef CONFIG_RXKAD
915 extern const struct rxrpc_security rxkad;
916 #endif
917 
918 /*
919  * security.c
920  */
921 int __init rxrpc_init_security(void);
922 void rxrpc_exit_security(void);
923 int rxrpc_init_client_conn_security(struct rxrpc_connection *);
924 int rxrpc_init_server_conn_security(struct rxrpc_connection *);
925 
926 /*
927  * sendmsg.c
928  */
929 int rxrpc_do_sendmsg(struct rxrpc_sock *, struct msghdr *, size_t);
930 
931 /*
932  * skbuff.c
933  */
934 void rxrpc_kernel_data_consumed(struct rxrpc_call *, struct sk_buff *);
935 void rxrpc_packet_destructor(struct sk_buff *);
936 void rxrpc_new_skb(struct sk_buff *);
937 void rxrpc_see_skb(struct sk_buff *);
938 void rxrpc_get_skb(struct sk_buff *);
939 void rxrpc_free_skb(struct sk_buff *);
940 void rxrpc_purge_queue(struct sk_buff_head *);
941 
942 /*
943  * sysctl.c
944  */
945 #ifdef CONFIG_SYSCTL
946 extern int __init rxrpc_sysctl_init(void);
947 extern void rxrpc_sysctl_exit(void);
948 #else
949 static inline int __init rxrpc_sysctl_init(void) { return 0; }
950 static inline void rxrpc_sysctl_exit(void) {}
951 #endif
952 
953 /*
954  * utils.c
955  */
956 int rxrpc_extract_addr_from_skb(struct sockaddr_rxrpc *, struct sk_buff *);
957 
958 static inline bool before(u32 seq1, u32 seq2)
959 {
960         return (s32)(seq1 - seq2) < 0;
961 }
962 static inline bool before_eq(u32 seq1, u32 seq2)
963 {
964         return (s32)(seq1 - seq2) <= 0;
965 }
966 static inline bool after(u32 seq1, u32 seq2)
967 {
968         return (s32)(seq1 - seq2) > 0;
969 }
970 static inline bool after_eq(u32 seq1, u32 seq2)
971 {
972         return (s32)(seq1 - seq2) >= 0;
973 }
974 
975 /*
976  * debug tracing
977  */
978 extern unsigned int rxrpc_debug;
979 
980 #define dbgprintk(FMT,...) \
981 	printk("[%-6.6s] "FMT"\n", current->comm ,##__VA_ARGS__)
982 
983 #define kenter(FMT,...)	dbgprintk("==> %s("FMT")",__func__ ,##__VA_ARGS__)
984 #define kleave(FMT,...)	dbgprintk("<== %s()"FMT"",__func__ ,##__VA_ARGS__)
985 #define kdebug(FMT,...)	dbgprintk("    "FMT ,##__VA_ARGS__)
986 #define kproto(FMT,...)	dbgprintk("### "FMT ,##__VA_ARGS__)
987 #define knet(FMT,...)	dbgprintk("@@@ "FMT ,##__VA_ARGS__)
988 
989 
990 #if defined(__KDEBUG)
991 #define _enter(FMT,...)	kenter(FMT,##__VA_ARGS__)
992 #define _leave(FMT,...)	kleave(FMT,##__VA_ARGS__)
993 #define _debug(FMT,...)	kdebug(FMT,##__VA_ARGS__)
994 #define _proto(FMT,...)	kproto(FMT,##__VA_ARGS__)
995 #define _net(FMT,...)	knet(FMT,##__VA_ARGS__)
996 
997 #elif defined(CONFIG_AF_RXRPC_DEBUG)
998 #define RXRPC_DEBUG_KENTER	0x01
999 #define RXRPC_DEBUG_KLEAVE	0x02
1000 #define RXRPC_DEBUG_KDEBUG	0x04
1001 #define RXRPC_DEBUG_KPROTO	0x08
1002 #define RXRPC_DEBUG_KNET	0x10
1003 
1004 #define _enter(FMT,...)					\
1005 do {							\
1006 	if (unlikely(rxrpc_debug & RXRPC_DEBUG_KENTER))	\
1007 		kenter(FMT,##__VA_ARGS__);		\
1008 } while (0)
1009 
1010 #define _leave(FMT,...)					\
1011 do {							\
1012 	if (unlikely(rxrpc_debug & RXRPC_DEBUG_KLEAVE))	\
1013 		kleave(FMT,##__VA_ARGS__);		\
1014 } while (0)
1015 
1016 #define _debug(FMT,...)					\
1017 do {							\
1018 	if (unlikely(rxrpc_debug & RXRPC_DEBUG_KDEBUG))	\
1019 		kdebug(FMT,##__VA_ARGS__);		\
1020 } while (0)
1021 
1022 #define _proto(FMT,...)					\
1023 do {							\
1024 	if (unlikely(rxrpc_debug & RXRPC_DEBUG_KPROTO))	\
1025 		kproto(FMT,##__VA_ARGS__);		\
1026 } while (0)
1027 
1028 #define _net(FMT,...)					\
1029 do {							\
1030 	if (unlikely(rxrpc_debug & RXRPC_DEBUG_KNET))	\
1031 		knet(FMT,##__VA_ARGS__);		\
1032 } while (0)
1033 
1034 #else
1035 #define _enter(FMT,...)	no_printk("==> %s("FMT")",__func__ ,##__VA_ARGS__)
1036 #define _leave(FMT,...)	no_printk("<== %s()"FMT"",__func__ ,##__VA_ARGS__)
1037 #define _debug(FMT,...)	no_printk("    "FMT ,##__VA_ARGS__)
1038 #define _proto(FMT,...)	no_printk("### "FMT ,##__VA_ARGS__)
1039 #define _net(FMT,...)	no_printk("@@@ "FMT ,##__VA_ARGS__)
1040 #endif
1041 
1042 /*
1043  * debug assertion checking
1044  */
1045 #if 1 // defined(__KDEBUGALL)
1046 
1047 #define ASSERT(X)						\
1048 do {								\
1049 	if (unlikely(!(X))) {					\
1050 		pr_err("Assertion failed\n");			\
1051 		BUG();						\
1052 	}							\
1053 } while (0)
1054 
1055 #define ASSERTCMP(X, OP, Y)						\
1056 do {									\
1057 	__typeof__(X) _x = (X);						\
1058 	__typeof__(Y) _y = (__typeof__(X))(Y);				\
1059 	if (unlikely(!(_x OP _y))) {					\
1060 		pr_err("Assertion failed - %lu(0x%lx) %s %lu(0x%lx) is false\n", \
1061 		       (unsigned long)_x, (unsigned long)_x, #OP,	\
1062 		       (unsigned long)_y, (unsigned long)_y);		\
1063 		BUG();							\
1064 	}								\
1065 } while (0)
1066 
1067 #define ASSERTIF(C, X)						\
1068 do {								\
1069 	if (unlikely((C) && !(X))) {				\
1070 		pr_err("Assertion failed\n");			\
1071 		BUG();						\
1072 	}							\
1073 } while (0)
1074 
1075 #define ASSERTIFCMP(C, X, OP, Y)					\
1076 do {									\
1077 	__typeof__(X) _x = (X);						\
1078 	__typeof__(Y) _y = (__typeof__(X))(Y);				\
1079 	if (unlikely((C) && !(_x OP _y))) {				\
1080 		pr_err("Assertion failed - %lu(0x%lx) %s %lu(0x%lx) is false\n", \
1081 		       (unsigned long)_x, (unsigned long)_x, #OP,	\
1082 		       (unsigned long)_y, (unsigned long)_y);		\
1083 		BUG();							\
1084 	}								\
1085 } while (0)
1086 
1087 #else
1088 
1089 #define ASSERT(X)				\
1090 do {						\
1091 } while (0)
1092 
1093 #define ASSERTCMP(X, OP, Y)			\
1094 do {						\
1095 } while (0)
1096 
1097 #define ASSERTIF(C, X)				\
1098 do {						\
1099 } while (0)
1100 
1101 #define ASSERTIFCMP(C, X, OP, Y)		\
1102 do {						\
1103 } while (0)
1104 
1105 #endif /* __KDEBUGALL */
1106