Lines Matching full:peer

28  * Hash a peer key.
62 /* Step through the peer address in 16-bit portions for speed */ in rxrpc_peer_hash_key()
71 * Compare a peer to a key. Return -ve, 0 or +ve to indicate less than, same
78 static long rxrpc_peer_cmp_key(const struct rxrpc_peer *peer, in rxrpc_peer_cmp_key() argument
85 diff = ((peer->hash_key - hash_key) ?: in rxrpc_peer_cmp_key()
86 ((unsigned long)peer->local - (unsigned long)local) ?: in rxrpc_peer_cmp_key()
87 (peer->srx.transport_type - srx->transport_type) ?: in rxrpc_peer_cmp_key()
88 (peer->srx.transport_len - srx->transport_len) ?: in rxrpc_peer_cmp_key()
89 (peer->srx.transport.family - srx->transport.family)); in rxrpc_peer_cmp_key()
95 return ((u16 __force)peer->srx.transport.sin.sin_port - in rxrpc_peer_cmp_key()
97 memcmp(&peer->srx.transport.sin.sin_addr, in rxrpc_peer_cmp_key()
102 return ((u16 __force)peer->srx.transport.sin6.sin6_port - in rxrpc_peer_cmp_key()
104 memcmp(&peer->srx.transport.sin6.sin6_addr, in rxrpc_peer_cmp_key()
121 struct rxrpc_peer *peer; in __rxrpc_lookup_peer_rcu() local
124 hash_for_each_possible_rcu(rxnet->peer_hash, peer, hash_link, hash_key) { in __rxrpc_lookup_peer_rcu()
125 if (rxrpc_peer_cmp_key(peer, local, srx, hash_key) == 0 && in __rxrpc_lookup_peer_rcu()
126 refcount_read(&peer->ref) > 0) in __rxrpc_lookup_peer_rcu()
127 return peer; in __rxrpc_lookup_peer_rcu()
139 struct rxrpc_peer *peer; in rxrpc_lookup_peer_rcu() local
142 peer = __rxrpc_lookup_peer_rcu(local, srx, hash_key); in rxrpc_lookup_peer_rcu()
143 if (peer) in rxrpc_lookup_peer_rcu()
144 _leave(" = %p {u=%d}", peer, refcount_read(&peer->ref)); in rxrpc_lookup_peer_rcu()
145 return peer; in rxrpc_lookup_peer_rcu()
149 * assess the MTU size for the network interface through which this peer is
153 struct rxrpc_peer *peer) in rxrpc_assess_MTU_size() argument
164 peer->if_mtu = 1500; in rxrpc_assess_MTU_size()
165 if (peer->max_data < peer->if_mtu - peer->hdrsize) { in rxrpc_assess_MTU_size()
166 trace_rxrpc_pmtud_reduce(peer, 0, peer->if_mtu - peer->hdrsize, in rxrpc_assess_MTU_size()
168 peer->max_data = peer->if_mtu - peer->hdrsize; in rxrpc_assess_MTU_size()
172 switch (peer->srx.transport.family) { in rxrpc_assess_MTU_size()
176 peer->srx.transport.sin.sin_addr.s_addr, 0, in rxrpc_assess_MTU_size()
190 memcpy(&fl6->daddr, &peer->srx.transport.sin6.sin6_addr, in rxrpc_assess_MTU_size()
206 peer->if_mtu = dst_mtu(dst); in rxrpc_assess_MTU_size()
207 peer->hdrsize += dst->header_len + dst->trailer_len; in rxrpc_assess_MTU_size()
208 peer->tx_seg_max = dst->dev->gso_max_segs; in rxrpc_assess_MTU_size()
211 peer->max_data = umin(RXRPC_JUMBO(1), peer->if_mtu - peer->hdrsize); in rxrpc_assess_MTU_size()
212 peer->pmtud_good = 500; in rxrpc_assess_MTU_size()
213 peer->pmtud_bad = peer->if_mtu - peer->hdrsize + 1; in rxrpc_assess_MTU_size()
214 peer->pmtud_trial = umin(peer->max_data, peer->pmtud_bad - 1); in rxrpc_assess_MTU_size()
215 peer->pmtud_pending = true; in rxrpc_assess_MTU_size()
217 _leave(" [if_mtu %u]", peer->if_mtu); in rxrpc_assess_MTU_size()
221 * Allocate a peer.
226 struct rxrpc_peer *peer; in rxrpc_alloc_peer() local
230 peer = kzalloc(sizeof(struct rxrpc_peer), gfp); in rxrpc_alloc_peer()
231 if (peer) { in rxrpc_alloc_peer()
232 refcount_set(&peer->ref, 1); in rxrpc_alloc_peer()
233 peer->local = rxrpc_get_local(local, rxrpc_local_get_peer); in rxrpc_alloc_peer()
234 INIT_HLIST_HEAD(&peer->error_targets); in rxrpc_alloc_peer()
235 peer->service_conns = RB_ROOT; in rxrpc_alloc_peer()
236 seqlock_init(&peer->service_conn_lock); in rxrpc_alloc_peer()
237 spin_lock_init(&peer->lock); in rxrpc_alloc_peer()
238 peer->debug_id = atomic_inc_return(&rxrpc_debug_id); in rxrpc_alloc_peer()
239 peer->recent_srtt_us = UINT_MAX; in rxrpc_alloc_peer()
240 peer->cong_ssthresh = RXRPC_TX_MAX_WINDOW; in rxrpc_alloc_peer()
241 trace_rxrpc_peer(peer->debug_id, 1, why); in rxrpc_alloc_peer()
244 _leave(" = %p", peer); in rxrpc_alloc_peer()
245 return peer; in rxrpc_alloc_peer()
249 * Initialise peer record.
251 static void rxrpc_init_peer(struct rxrpc_local *local, struct rxrpc_peer *peer, in rxrpc_init_peer() argument
254 peer->hash_key = hash_key; in rxrpc_init_peer()
257 switch (peer->srx.transport.family) { in rxrpc_init_peer()
259 peer->hdrsize = sizeof(struct iphdr); in rxrpc_init_peer()
263 peer->hdrsize = sizeof(struct ipv6hdr); in rxrpc_init_peer()
270 switch (peer->srx.transport_type) { in rxrpc_init_peer()
272 peer->hdrsize += sizeof(struct udphdr); in rxrpc_init_peer()
278 peer->hdrsize += sizeof(struct rxrpc_wire_header); in rxrpc_init_peer()
279 peer->max_data = peer->if_mtu - peer->hdrsize; in rxrpc_init_peer()
281 rxrpc_assess_MTU_size(local, peer); in rxrpc_init_peer()
285 * Set up a new peer.
292 struct rxrpc_peer *peer; in rxrpc_create_peer() local
296 peer = rxrpc_alloc_peer(local, gfp, rxrpc_peer_new_client); in rxrpc_create_peer()
297 if (peer) { in rxrpc_create_peer()
298 memcpy(&peer->srx, srx, sizeof(*srx)); in rxrpc_create_peer()
299 rxrpc_init_peer(local, peer, hash_key); in rxrpc_create_peer()
302 _leave(" = %p", peer); in rxrpc_create_peer()
303 return peer; in rxrpc_create_peer()
306 static void rxrpc_free_peer(struct rxrpc_peer *peer) in rxrpc_free_peer() argument
308 trace_rxrpc_peer(peer->debug_id, 0, rxrpc_peer_free); in rxrpc_free_peer()
309 rxrpc_put_local(peer->local, rxrpc_local_put_peer); in rxrpc_free_peer()
310 kfree_rcu(peer, rcu); in rxrpc_free_peer()
314 * Set up a new incoming peer. There shouldn't be any other matching peers
319 void rxrpc_new_incoming_peer(struct rxrpc_local *local, struct rxrpc_peer *peer) in rxrpc_new_incoming_peer() argument
324 hash_key = rxrpc_peer_hash_key(local, &peer->srx); in rxrpc_new_incoming_peer()
325 rxrpc_init_peer(local, peer, hash_key); in rxrpc_new_incoming_peer()
328 hash_add_rcu(rxnet->peer_hash, &peer->hash_link, hash_key); in rxrpc_new_incoming_peer()
329 list_add_tail(&peer->keepalive_link, &rxnet->peer_keepalive_new); in rxrpc_new_incoming_peer()
339 struct rxrpc_peer *peer, *candidate; in rxrpc_lookup_peer() local
345 /* search the peer list first */ in rxrpc_lookup_peer()
347 peer = __rxrpc_lookup_peer_rcu(local, srx, hash_key); in rxrpc_lookup_peer()
348 if (peer && !rxrpc_get_peer_maybe(peer, rxrpc_peer_get_lookup_client)) in rxrpc_lookup_peer()
349 peer = NULL; in rxrpc_lookup_peer()
352 if (!peer) { in rxrpc_lookup_peer()
353 /* The peer is not yet present in hash - create a candidate in rxrpc_lookup_peer()
365 peer = __rxrpc_lookup_peer_rcu(local, srx, hash_key); in rxrpc_lookup_peer()
366 if (peer && !rxrpc_get_peer_maybe(peer, rxrpc_peer_get_lookup_client)) in rxrpc_lookup_peer()
367 peer = NULL; in rxrpc_lookup_peer()
368 if (!peer) { in rxrpc_lookup_peer()
377 if (peer) in rxrpc_lookup_peer()
380 peer = candidate; in rxrpc_lookup_peer()
383 _leave(" = %p {u=%d}", peer, refcount_read(&peer->ref)); in rxrpc_lookup_peer()
384 return peer; in rxrpc_lookup_peer()
388 * Get a ref on a peer record.
390 struct rxrpc_peer *rxrpc_get_peer(struct rxrpc_peer *peer, enum rxrpc_peer_trace why) in rxrpc_get_peer() argument
394 __refcount_inc(&peer->ref, &r); in rxrpc_get_peer()
395 trace_rxrpc_peer(peer->debug_id, r + 1, why); in rxrpc_get_peer()
396 return peer; in rxrpc_get_peer()
400 * Get a ref on a peer record unless its usage has already reached 0.
402 struct rxrpc_peer *rxrpc_get_peer_maybe(struct rxrpc_peer *peer, in rxrpc_get_peer_maybe() argument
407 if (peer) { in rxrpc_get_peer_maybe()
408 if (__refcount_inc_not_zero(&peer->ref, &r)) in rxrpc_get_peer_maybe()
409 trace_rxrpc_peer(peer->debug_id, r + 1, why); in rxrpc_get_peer_maybe()
411 peer = NULL; in rxrpc_get_peer_maybe()
413 return peer; in rxrpc_get_peer_maybe()
417 * Discard a peer record.
419 static void __rxrpc_put_peer(struct rxrpc_peer *peer) in __rxrpc_put_peer() argument
421 struct rxrpc_net *rxnet = peer->local->rxnet; in __rxrpc_put_peer()
423 ASSERT(hlist_empty(&peer->error_targets)); in __rxrpc_put_peer()
426 hash_del_rcu(&peer->hash_link); in __rxrpc_put_peer()
427 list_del_init(&peer->keepalive_link); in __rxrpc_put_peer()
430 rxrpc_free_peer(peer); in __rxrpc_put_peer()
434 * Drop a ref on a peer record.
436 void rxrpc_put_peer(struct rxrpc_peer *peer, enum rxrpc_peer_trace why) in rxrpc_put_peer() argument
442 if (peer) { in rxrpc_put_peer()
443 debug_id = peer->debug_id; in rxrpc_put_peer()
444 dead = __refcount_dec_and_test(&peer->ref, &r); in rxrpc_put_peer()
447 __rxrpc_put_peer(peer); in rxrpc_put_peer()
452 * Make sure all peer records have been discarded.
456 struct rxrpc_peer *peer; in rxrpc_destroy_all_peers() local
463 hlist_for_each_entry(peer, &rxnet->peer_hash[i], hash_link) { in rxrpc_destroy_all_peers()
464 pr_err("Leaked peer %u {%u} %pISp\n", in rxrpc_destroy_all_peers()
465 peer->debug_id, in rxrpc_destroy_all_peers()
466 refcount_read(&peer->ref), in rxrpc_destroy_all_peers()
467 &peer->srx.transport); in rxrpc_destroy_all_peers()
473 * rxrpc_kernel_get_call_peer - Get the peer address of a call
477 * Get a record for the remote peer in a call.
481 return call->peer; in rxrpc_kernel_get_call_peer()
486 * rxrpc_kernel_get_srtt - Get a call's peer smoothed RTT
487 * @peer: The peer to query
489 * Get the call's peer smoothed RTT in uS or UINT_MAX if we have no samples.
491 unsigned int rxrpc_kernel_get_srtt(const struct rxrpc_peer *peer) in rxrpc_kernel_get_srtt() argument
493 return READ_ONCE(peer->recent_srtt_us); in rxrpc_kernel_get_srtt()
498 * rxrpc_kernel_remote_srx - Get the address of a peer
499 * @peer: The peer to query
501 * Get a pointer to the address from a peer record. The caller is responsible
504 const struct sockaddr_rxrpc *rxrpc_kernel_remote_srx(const struct rxrpc_peer *peer) in rxrpc_kernel_remote_srx() argument
506 return peer ? &peer->srx : &rxrpc_null_addr; in rxrpc_kernel_remote_srx()
511 * rxrpc_kernel_remote_addr - Get the peer transport address of a call
512 * @peer: The peer to query
514 * Get a pointer to the transport address from a peer record. The caller is
517 const struct sockaddr *rxrpc_kernel_remote_addr(const struct rxrpc_peer *peer) in rxrpc_kernel_remote_addr() argument
520 (peer ? &peer->srx.transport : &rxrpc_null_addr.transport); in rxrpc_kernel_remote_addr()