xref: /freebsd/sys/ofed/drivers/infiniband/ulp/sdp/sdp_cma.c (revision 95ee2897e98f5d444f26ed2334cc7c439f9c16c6)
1 /*-
2  * SPDX-License-Identifier: BSD-2-Clause OR GPL-2.0
3  *
4  * Copyright (c) 2006 Mellanox Technologies Ltd.  All rights reserved.
5  *
6  * This software is available to you under a choice of one of two
7  * licenses.  You may choose to be licensed under the terms of the GNU
8  * General Public License (GPL) Version 2, available from the file
9  * COPYING in the main directory of this source tree, or the
10  * OpenIB.org BSD license below:
11  *
12  *     Redistribution and use in source and binary forms, with or
13  *     without modification, are permitted provided that the following
14  *     conditions are met:
15  *
16  *      - Redistributions of source code must retain the above
17  *        copyright notice, this list of conditions and the following
18  *        disclaimer.
19  *
20  *      - Redistributions in binary form must reproduce the above
21  *        copyright notice, this list of conditions and the following
22  *        disclaimer in the documentation and/or other materials
23  *        provided with the distribution.
24  *
25  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
26  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
27  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
28  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
29  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
30  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
31  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
32  * SOFTWARE.
33  *
34  * $Id$
35  */
36 #include "sdp.h"
37 
38 #define SDP_MAJV_MINV 0x22
39 
40 SDP_MODPARAM_SINT(sdp_link_layer_ib_only, 1, "Support only link layer of "
41 		"type Infiniband");
42 
43 enum {
44 	SDP_HH_SIZE = 76,
45 	SDP_HAH_SIZE = 180,
46 };
47 
48 static void
49 sdp_qp_event_handler(struct ib_event *event, void *data)
50 {
51 	struct socket *sk = data;
52 
53 	sdp_dbg(sk, "QP Event: %s (%d)", ib_event_msg(event->event),
54 	    event->event);
55 }
56 
57 static int
58 sdp_get_max_dev_sge(struct ib_device *dev)
59 {
60 	struct ib_device_attr *device_attr;
61 	static int max_sges = -1;
62 
63 	if (max_sges > 0)
64 		goto out;
65 
66 	device_attr = &dev->attrs;
67 	max_sges = device_attr->max_sge;
68 
69 out:
70 	return max_sges;
71 }
72 
73 static int
74 sdp_init_qp(struct socket *sk, struct rdma_cm_id *id)
75 {
76 	struct ib_qp_init_attr qp_init_attr = {
77 		.event_handler = sdp_qp_event_handler,
78 		.qp_context = sk,
79 		.cap.max_send_wr = SDP_TX_SIZE,
80 		.cap.max_recv_wr = SDP_RX_SIZE,
81         	.sq_sig_type = IB_SIGNAL_REQ_WR,
82         	.qp_type = IB_QPT_RC,
83 	};
84 	struct ib_device *device = id->device;
85 	struct sdp_sock *ssk;
86 	int rc;
87 
88 	sdp_dbg(sk, "%s\n", __func__);
89 
90 	ssk = sdp_sk(sk);
91 	ssk->max_sge = sdp_get_max_dev_sge(device);
92 	sdp_dbg(sk, "Max sges: %d\n", ssk->max_sge);
93 
94 	qp_init_attr.cap.max_send_sge = MIN(ssk->max_sge, SDP_MAX_SEND_SGES);
95 	sdp_dbg(sk, "Setting max send sge to: %d\n",
96 	    qp_init_attr.cap.max_send_sge);
97 
98 	qp_init_attr.cap.max_recv_sge = MIN(ssk->max_sge, SDP_MAX_RECV_SGES);
99 	sdp_dbg(sk, "Setting max recv sge to: %d\n",
100 	    qp_init_attr.cap.max_recv_sge);
101 
102 	ssk->sdp_dev = ib_get_client_data(device, &sdp_client);
103 	if (!ssk->sdp_dev) {
104 		sdp_warn(sk, "SDP not available on device %s\n", device->name);
105 		rc = -ENODEV;
106 		goto err_rx;
107 	}
108 
109 	rc = sdp_rx_ring_create(ssk, device);
110 	if (rc)
111 		goto err_rx;
112 
113 	rc = sdp_tx_ring_create(ssk, device);
114 	if (rc)
115 		goto err_tx;
116 
117 	qp_init_attr.recv_cq = ssk->rx_ring.cq;
118 	qp_init_attr.send_cq = ssk->tx_ring.cq;
119 
120 	rc = rdma_create_qp(id, ssk->sdp_dev->pd, &qp_init_attr);
121 	if (rc) {
122 		sdp_warn(sk, "Unable to create QP: %d.\n", rc);
123 		goto err_qp;
124 	}
125 	ssk->qp = id->qp;
126 	ssk->ib_device = device;
127 	ssk->qp_active = 1;
128 	ssk->context.device = device;
129 
130 	sdp_dbg(sk, "%s done\n", __func__);
131 	return 0;
132 
133 err_qp:
134 	sdp_tx_ring_destroy(ssk);
135 err_tx:
136 	sdp_rx_ring_destroy(ssk);
137 err_rx:
138 	return rc;
139 }
140 
141 static int
142 sdp_connect_handler(struct socket *sk, struct rdma_cm_id *id,
143     struct rdma_cm_event *event)
144 {
145 	struct sockaddr_in *src_addr;
146 	struct sockaddr_in *dst_addr;
147 	struct socket *child;
148 	const struct sdp_hh *h;
149 	struct sdp_sock *ssk;
150 	int rc;
151 
152 	sdp_dbg(sk, "%s %p -> %p\n", __func__, sdp_sk(sk)->id, id);
153 
154 	h = event->param.conn.private_data;
155 	SDP_DUMP_PACKET(sk, "RX", NULL, &h->bsdh);
156 
157 	if (!h->max_adverts)
158 		return -EINVAL;
159 
160 	child = sonewconn(sk, SS_ISCONNECTED);
161 	if (!child)
162 		return -ENOMEM;
163 
164 	ssk = sdp_sk(child);
165 	rc = sdp_init_qp(child, id);
166 	if (rc)
167 		return rc;
168 	SDP_WLOCK(ssk);
169 	id->context = ssk;
170 	ssk->id = id;
171 	ssk->socket = child;
172 	ssk->cred = crhold(child->so_cred);
173 	dst_addr = (struct sockaddr_in *)&id->route.addr.dst_addr;
174 	src_addr = (struct sockaddr_in *)&id->route.addr.src_addr;
175 	ssk->fport = dst_addr->sin_port;
176 	ssk->faddr = dst_addr->sin_addr.s_addr;
177 	ssk->lport = src_addr->sin_port;
178 	ssk->max_bufs = ntohs(h->bsdh.bufs);
179 	atomic_set(&ssk->tx_ring.credits, ssk->max_bufs);
180 	ssk->min_bufs = tx_credits(ssk) / 4;
181 	ssk->xmit_size_goal = ntohl(h->localrcvsz) - sizeof(struct sdp_bsdh);
182 	sdp_init_buffers(ssk, rcvbuf_initial_size);
183 	ssk->state = TCPS_SYN_RECEIVED;
184 	SDP_WUNLOCK(ssk);
185 
186 	return 0;
187 }
188 
189 static int
190 sdp_response_handler(struct socket *sk, struct rdma_cm_id *id,
191     struct rdma_cm_event *event)
192 {
193 	const struct sdp_hah *h;
194 	struct sockaddr_in *dst_addr;
195 	struct sdp_sock *ssk;
196 	sdp_dbg(sk, "%s\n", __func__);
197 
198 	ssk = sdp_sk(sk);
199 	SDP_WLOCK(ssk);
200 	ssk->state = TCPS_ESTABLISHED;
201 	sdp_set_default_moderation(ssk);
202 	if (ssk->flags & SDP_DROPPED) {
203 		SDP_WUNLOCK(ssk);
204 		return 0;
205 	}
206 	if (sk->so_options & SO_KEEPALIVE)
207 		sdp_start_keepalive_timer(sk);
208 	h = event->param.conn.private_data;
209 	SDP_DUMP_PACKET(sk, "RX", NULL, &h->bsdh);
210 	ssk->max_bufs = ntohs(h->bsdh.bufs);
211 	atomic_set(&ssk->tx_ring.credits, ssk->max_bufs);
212 	ssk->min_bufs = tx_credits(ssk) / 4;
213 	ssk->xmit_size_goal =
214 		ntohl(h->actrcvsz) - sizeof(struct sdp_bsdh);
215 	ssk->poll_cq = 1;
216 
217 	dst_addr = (struct sockaddr_in *)&id->route.addr.dst_addr;
218 	ssk->fport = dst_addr->sin_port;
219 	ssk->faddr = dst_addr->sin_addr.s_addr;
220 	soisconnected(sk);
221 	SDP_WUNLOCK(ssk);
222 
223 	return 0;
224 }
225 
226 static int
227 sdp_connected_handler(struct socket *sk, struct rdma_cm_event *event)
228 {
229 	struct sdp_sock *ssk;
230 
231 	sdp_dbg(sk, "%s\n", __func__);
232 
233 	ssk = sdp_sk(sk);
234 	SDP_WLOCK(ssk);
235 	ssk->state = TCPS_ESTABLISHED;
236 
237 	sdp_set_default_moderation(ssk);
238 
239 	if (sk->so_options & SO_KEEPALIVE)
240 		sdp_start_keepalive_timer(sk);
241 
242 	if ((ssk->flags & SDP_DROPPED) == 0)
243 		soisconnected(sk);
244 	SDP_WUNLOCK(ssk);
245 	return 0;
246 }
247 
248 static int
249 sdp_disconnected_handler(struct socket *sk)
250 {
251 	struct sdp_sock *ssk;
252 
253 	ssk = sdp_sk(sk);
254 	sdp_dbg(sk, "%s\n", __func__);
255 
256 	SDP_WLOCK_ASSERT(ssk);
257 	if (sdp_sk(sk)->state == TCPS_SYN_RECEIVED) {
258 		sdp_connected_handler(sk, NULL);
259 
260 		if (rcv_nxt(ssk))
261 			return 0;
262 	}
263 
264 	return -ECONNRESET;
265 }
266 
267 int
268 sdp_cma_handler(struct rdma_cm_id *id, struct rdma_cm_event *event)
269 {
270 	struct rdma_conn_param conn_param;
271 	struct socket *sk;
272 	struct sdp_sock *ssk;
273 	struct sdp_hah hah;
274 	struct sdp_hh hh;
275 
276 	int rc = 0;
277 
278 	ssk = id->context;
279 	sk = NULL;
280 	if (ssk)
281 		sk = ssk->socket;
282 	if (!ssk || !sk || !ssk->id) {
283 		sdp_dbg(sk,
284 		    "cm_id is being torn down, event %d, ssk %p, sk %p, id %p\n",
285 		       	event->event, ssk, sk, id);
286 		return event->event == RDMA_CM_EVENT_CONNECT_REQUEST ?
287 			-EINVAL : 0;
288 	}
289 
290 	sdp_dbg(sk, "%s event %d id %p\n", __func__, event->event, id);
291 	switch (event->event) {
292 	case RDMA_CM_EVENT_ADDR_RESOLVED:
293 		sdp_dbg(sk, "RDMA_CM_EVENT_ADDR_RESOLVED\n");
294 
295 		if (sdp_link_layer_ib_only &&
296 			rdma_node_get_transport(id->device->node_type) ==
297 				RDMA_TRANSPORT_IB &&
298 			rdma_port_get_link_layer(id->device, id->port_num) !=
299 				IB_LINK_LAYER_INFINIBAND) {
300 			sdp_dbg(sk, "Link layer is: %d. Only IB link layer "
301 				"is allowed\n",
302 				rdma_port_get_link_layer(id->device, id->port_num));
303 			rc = -ENETUNREACH;
304 			break;
305 		}
306 
307 		rc = rdma_resolve_route(id, SDP_ROUTE_TIMEOUT);
308 		break;
309 	case RDMA_CM_EVENT_ADDR_ERROR:
310 		sdp_dbg(sk, "RDMA_CM_EVENT_ADDR_ERROR\n");
311 		rc = -ENETUNREACH;
312 		break;
313 	case RDMA_CM_EVENT_ROUTE_RESOLVED:
314 		sdp_dbg(sk, "RDMA_CM_EVENT_ROUTE_RESOLVED : %p\n", id);
315 		rc = sdp_init_qp(sk, id);
316 		if (rc)
317 			break;
318 		atomic_set(&sdp_sk(sk)->remote_credits,
319 				rx_ring_posted(sdp_sk(sk)));
320 		memset(&hh, 0, sizeof hh);
321 		hh.bsdh.mid = SDP_MID_HELLO;
322 		hh.bsdh.len = htonl(sizeof(struct sdp_hh));
323 		hh.max_adverts = 1;
324 		hh.ipv_cap = 0x40;
325 		hh.majv_minv = SDP_MAJV_MINV;
326 		sdp_init_buffers(sdp_sk(sk), rcvbuf_initial_size);
327 		hh.bsdh.bufs = htons(rx_ring_posted(sdp_sk(sk)));
328 		hh.localrcvsz = hh.desremrcvsz = htonl(sdp_sk(sk)->recv_bytes);
329 		hh.max_adverts = 0x1;
330 		sdp_sk(sk)->laddr =
331 			((struct sockaddr_in *)&id->route.addr.src_addr)->sin_addr.s_addr;
332 		memset(&conn_param, 0, sizeof conn_param);
333 		conn_param.private_data_len = sizeof hh;
334 		conn_param.private_data = &hh;
335 		conn_param.responder_resources = 4 /* TODO */;
336 		conn_param.initiator_depth = 4 /* TODO */;
337 		conn_param.retry_count = SDP_RETRY_COUNT;
338 		SDP_DUMP_PACKET(NULL, "TX", NULL, &hh.bsdh);
339 		rc = rdma_connect(id, &conn_param);
340 		break;
341 	case RDMA_CM_EVENT_ROUTE_ERROR:
342 		sdp_dbg(sk, "RDMA_CM_EVENT_ROUTE_ERROR : %p\n", id);
343 		rc = -ETIMEDOUT;
344 		break;
345 	case RDMA_CM_EVENT_CONNECT_REQUEST:
346 		sdp_dbg(sk, "RDMA_CM_EVENT_CONNECT_REQUEST\n");
347 		rc = sdp_connect_handler(sk, id, event);
348 		if (rc) {
349 			sdp_dbg(sk, "Destroying qp\n");
350 			rdma_reject(id, NULL, 0);
351 			break;
352 		}
353 		ssk = id->context;
354 		atomic_set(&ssk->remote_credits, rx_ring_posted(ssk));
355 		memset(&hah, 0, sizeof hah);
356 		hah.bsdh.mid = SDP_MID_HELLO_ACK;
357 		hah.bsdh.bufs = htons(rx_ring_posted(ssk));
358 		hah.bsdh.len = htonl(sizeof(struct sdp_hah));
359 		hah.majv_minv = SDP_MAJV_MINV;
360 		hah.ext_max_adverts = 1; /* Doesn't seem to be mandated by spec,
361 					    but just in case */
362 		hah.actrcvsz = htonl(ssk->recv_bytes);
363 		memset(&conn_param, 0, sizeof conn_param);
364 		conn_param.private_data_len = sizeof hah;
365 		conn_param.private_data = &hah;
366 		conn_param.responder_resources = 4 /* TODO */;
367 		conn_param.initiator_depth = 4 /* TODO */;
368 		conn_param.retry_count = SDP_RETRY_COUNT;
369 		SDP_DUMP_PACKET(sk, "TX", NULL, &hah.bsdh);
370 		rc = rdma_accept(id, &conn_param);
371 		if (rc) {
372 			ssk->id = NULL;
373 			id->qp = NULL;
374 			id->context = NULL;
375 		}
376 		break;
377 	case RDMA_CM_EVENT_CONNECT_RESPONSE:
378 		sdp_dbg(sk, "RDMA_CM_EVENT_CONNECT_RESPONSE\n");
379 		rc = sdp_response_handler(sk, id, event);
380 		if (rc) {
381 			sdp_dbg(sk, "Destroying qp\n");
382 			rdma_reject(id, NULL, 0);
383 		} else
384 			rc = rdma_accept(id, NULL);
385 		break;
386 	case RDMA_CM_EVENT_CONNECT_ERROR:
387 		sdp_dbg(sk, "RDMA_CM_EVENT_CONNECT_ERROR\n");
388 		rc = -ETIMEDOUT;
389 		break;
390 	case RDMA_CM_EVENT_UNREACHABLE:
391 		sdp_dbg(sk, "RDMA_CM_EVENT_UNREACHABLE\n");
392 		rc = -ENETUNREACH;
393 		break;
394 	case RDMA_CM_EVENT_REJECTED:
395 		sdp_dbg(sk, "RDMA_CM_EVENT_REJECTED\n");
396 		rc = -ECONNREFUSED;
397 		break;
398 	case RDMA_CM_EVENT_ESTABLISHED:
399 		sdp_dbg(sk, "RDMA_CM_EVENT_ESTABLISHED\n");
400 		sdp_sk(sk)->laddr =
401 			((struct sockaddr_in *)&id->route.addr.src_addr)->sin_addr.s_addr;
402 		rc = sdp_connected_handler(sk, event);
403 		break;
404 	case RDMA_CM_EVENT_DISCONNECTED: /* This means DREQ/DREP received */
405 		sdp_dbg(sk, "RDMA_CM_EVENT_DISCONNECTED\n");
406 
407 		SDP_WLOCK(ssk);
408 		if (ssk->state == TCPS_LAST_ACK) {
409 			sdp_cancel_dreq_wait_timeout(ssk);
410 
411 			sdp_dbg(sk, "%s: waiting for Infiniband tear down\n",
412 				__func__);
413 		}
414 		ssk->qp_active = 0;
415 		SDP_WUNLOCK(ssk);
416 		rdma_disconnect(id);
417 		SDP_WLOCK(ssk);
418 		if (ssk->state != TCPS_TIME_WAIT) {
419 			if (ssk->state == TCPS_CLOSE_WAIT) {
420 				sdp_dbg(sk, "IB teardown while in "
421 					"TCPS_CLOSE_WAIT taking reference to "
422 					"let close() finish the work\n");
423 			}
424 			rc = sdp_disconnected_handler(sk);
425 			if (rc)
426 				rc = -EPIPE;
427 		}
428 		SDP_WUNLOCK(ssk);
429 		break;
430 	case RDMA_CM_EVENT_TIMEWAIT_EXIT:
431 		sdp_dbg(sk, "RDMA_CM_EVENT_TIMEWAIT_EXIT\n");
432 		SDP_WLOCK(ssk);
433 		rc = sdp_disconnected_handler(sk);
434 		SDP_WUNLOCK(ssk);
435 		break;
436 	case RDMA_CM_EVENT_DEVICE_REMOVAL:
437 		sdp_dbg(sk, "RDMA_CM_EVENT_DEVICE_REMOVAL\n");
438 		rc = -ENETRESET;
439 		break;
440 	default:
441 		printk(KERN_ERR "SDP: Unexpected CMA event: %d\n",
442 		       event->event);
443 		rc = -ECONNABORTED;
444 		break;
445 	}
446 
447 	sdp_dbg(sk, "event %s (%d) done. status %d\n",
448 	    rdma_event_msg(event->event), event->event, rc);
449 
450 	if (rc) {
451 		SDP_WLOCK(ssk);
452 		if (ssk->id == id) {
453 			ssk->id = NULL;
454 			id->qp = NULL;
455 			id->context = NULL;
456 			if (sdp_notify(ssk, -rc))
457 				SDP_WUNLOCK(ssk);
458 		} else
459 			SDP_WUNLOCK(ssk);
460 	}
461 
462 	return rc;
463 }
464