xref: /freebsd/sys/dev/irdma/irdma_cm.c (revision 7b6644e160ed63b633e7c68a3cacf2c71d216cd5)
1 /*-
2  * SPDX-License-Identifier: GPL-2.0 or Linux-OpenIB
3  *
4  * Copyright (c) 2015 - 2025 Intel Corporation
5  *
6  * This software is available to you under a choice of one of two
7  * licenses.  You may choose to be licensed under the terms of the GNU
8  * General Public License (GPL) Version 2, available from the file
9  * COPYING in the main directory of this source tree, or the
10  * OpenFabrics.org BSD license below:
11  *
12  *   Redistribution and use in source and binary forms, with or
13  *   without modification, are permitted provided that the following
14  *   conditions are met:
15  *
16  *    - Redistributions of source code must retain the above
17  *	copyright notice, this list of conditions and the following
18  *	disclaimer.
19  *
20  *    - Redistributions in binary form must reproduce the above
21  *	copyright notice, this list of conditions and the following
22  *	disclaimer in the documentation and/or other materials
23  *	provided with the distribution.
24  *
25  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
26  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
27  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
28  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
29  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
30  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
31  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
32  * SOFTWARE.
33  */
34 
35 #include "irdma_main.h"
36 
37 static void irdma_cm_post_event(struct irdma_cm_event *event);
38 static void irdma_disconnect_worker(struct work_struct *work);
39 
40 /**
41  * irdma_free_sqbuf - put back puda buffer if refcount is 0
42  * @vsi: The VSI structure of the device
43  * @bufp: puda buffer to free
44  */
45 void
irdma_free_sqbuf(struct irdma_sc_vsi * vsi,void * bufp)46 irdma_free_sqbuf(struct irdma_sc_vsi *vsi, void *bufp)
47 {
48 	struct irdma_puda_buf *buf = bufp;
49 	struct irdma_puda_rsrc *ilq = vsi->ilq;
50 
51 	if (atomic_dec_and_test(&buf->refcount))
52 		irdma_puda_ret_bufpool(ilq, buf);
53 }
54 
55 /**
56  * irdma_record_ird_ord - Record IRD/ORD passed in
57  * @cm_node: connection's node
58  * @conn_ird: connection IRD
59  * @conn_ord: connection ORD
60  */
61 static void
irdma_record_ird_ord(struct irdma_cm_node * cm_node,u32 conn_ird,u32 conn_ord)62 irdma_record_ird_ord(struct irdma_cm_node *cm_node, u32 conn_ird,
63 		     u32 conn_ord)
64 {
65 	if (conn_ird > cm_node->dev->hw_attrs.max_hw_ird)
66 		conn_ird = cm_node->dev->hw_attrs.max_hw_ird;
67 
68 	if (conn_ord > cm_node->dev->hw_attrs.max_hw_ord)
69 		conn_ord = cm_node->dev->hw_attrs.max_hw_ord;
70 	else if (!conn_ord && cm_node->send_rdma0_op == SEND_RDMA_READ_ZERO)
71 		conn_ord = 1;
72 	cm_node->ird_size = conn_ird;
73 	cm_node->ord_size = conn_ord;
74 }
75 
76 /**
77  * irdma_copy_ip_ntohl - copy IP address from  network to host
78  * @dst: IP address in host order
79  * @src: IP address in network order (big endian)
80  */
81 void
irdma_copy_ip_ntohl(u32 * dst,__be32 * src)82 irdma_copy_ip_ntohl(u32 *dst, __be32 *src)
83 {
84 	*dst++ = ntohl(*src++);
85 	*dst++ = ntohl(*src++);
86 	*dst++ = ntohl(*src++);
87 	*dst = ntohl(*src);
88 }
89 
90 /**
91  * irdma_copy_ip_htonl - copy IP address from host to network order
92  * @dst: IP address in network order (big endian)
93  * @src: IP address in host order
94  */
95 void
irdma_copy_ip_htonl(__be32 * dst,u32 * src)96 irdma_copy_ip_htonl(__be32 *dst, u32 *src)
97 {
98 	*dst++ = htonl(*src++);
99 	*dst++ = htonl(*src++);
100 	*dst++ = htonl(*src++);
101 	*dst = htonl(*src);
102 }
103 
104 /**
105  * irdma_get_addr_info
106  * @cm_node: contains ip/tcp info
107  * @cm_info: to get a copy of the cm_node ip/tcp info
108  */
109 static void
irdma_get_addr_info(struct irdma_cm_node * cm_node,struct irdma_cm_info * cm_info)110 irdma_get_addr_info(struct irdma_cm_node *cm_node,
111 		    struct irdma_cm_info *cm_info)
112 {
113 	memset(cm_info, 0, sizeof(*cm_info));
114 	cm_info->ipv4 = cm_node->ipv4;
115 	cm_info->vlan_id = cm_node->vlan_id;
116 	memcpy(cm_info->loc_addr, cm_node->loc_addr, sizeof(cm_info->loc_addr));
117 	memcpy(cm_info->rem_addr, cm_node->rem_addr, sizeof(cm_info->rem_addr));
118 	cm_info->loc_port = cm_node->loc_port;
119 	cm_info->rem_port = cm_node->rem_port;
120 }
121 
122 /**
123  * irdma_fill_sockaddr4 - fill in addr info for IPv4 connection
124  * @cm_node: connection's node
125  * @event: upper layer's cm event
126  */
127 static inline void
irdma_fill_sockaddr4(struct irdma_cm_node * cm_node,struct iw_cm_event * event)128 irdma_fill_sockaddr4(struct irdma_cm_node *cm_node,
129 		     struct iw_cm_event *event)
130 {
131 	struct sockaddr_in *laddr = (struct sockaddr_in *)&event->local_addr;
132 	struct sockaddr_in *raddr = (struct sockaddr_in *)&event->remote_addr;
133 
134 	laddr->sin_family = AF_INET;
135 	raddr->sin_family = AF_INET;
136 
137 	laddr->sin_port = htons(cm_node->loc_port);
138 	raddr->sin_port = htons(cm_node->rem_port);
139 
140 	laddr->sin_addr.s_addr = htonl(cm_node->loc_addr[0]);
141 	raddr->sin_addr.s_addr = htonl(cm_node->rem_addr[0]);
142 }
143 
144 /**
145  * irdma_fill_sockaddr6 - fill in addr info for IPv6 connection
146  * @cm_node: connection's node
147  * @event: upper layer's cm event
148  */
149 static inline void
irdma_fill_sockaddr6(struct irdma_cm_node * cm_node,struct iw_cm_event * event)150 irdma_fill_sockaddr6(struct irdma_cm_node *cm_node,
151 		     struct iw_cm_event *event)
152 {
153 	struct sockaddr_in6 *laddr6 = (struct sockaddr_in6 *)&event->local_addr;
154 	struct sockaddr_in6 *raddr6 = (struct sockaddr_in6 *)&event->remote_addr;
155 
156 	laddr6->sin6_family = AF_INET6;
157 	raddr6->sin6_family = AF_INET6;
158 
159 	laddr6->sin6_port = htons(cm_node->loc_port);
160 	raddr6->sin6_port = htons(cm_node->rem_port);
161 
162 	irdma_copy_ip_htonl(laddr6->sin6_addr.__u6_addr.__u6_addr32,
163 			    cm_node->loc_addr);
164 	irdma_copy_ip_htonl(raddr6->sin6_addr.__u6_addr.__u6_addr32,
165 			    cm_node->rem_addr);
166 }
167 
168 /**
169  * irdma_get_cmevent_info - for cm event upcall
170  * @cm_node: connection's node
171  * @cm_id: upper layers cm struct for the event
172  * @event: upper layer's cm event
173  */
174 static inline void
irdma_get_cmevent_info(struct irdma_cm_node * cm_node,struct iw_cm_id * cm_id,struct iw_cm_event * event)175 irdma_get_cmevent_info(struct irdma_cm_node *cm_node,
176 		       struct iw_cm_id *cm_id,
177 		       struct iw_cm_event *event)
178 {
179 	memcpy(&event->local_addr, &cm_id->m_local_addr,
180 	       sizeof(event->local_addr));
181 	memcpy(&event->remote_addr, &cm_id->m_remote_addr,
182 	       sizeof(event->remote_addr));
183 	if (cm_node) {
184 		event->private_data = cm_node->pdata_buf;
185 		event->private_data_len = (u8)cm_node->pdata.size;
186 		event->ird = cm_node->ird_size;
187 		event->ord = cm_node->ord_size;
188 	}
189 }
190 
191 /**
192  * irdma_send_cm_event - upcall cm's event handler
193  * @cm_node: connection's node
194  * @cm_id: upper layer's cm info struct
195  * @type: Event type to indicate
196  * @status: status for the event type
197  */
198 static int
irdma_send_cm_event(struct irdma_cm_node * cm_node,struct iw_cm_id * cm_id,enum iw_cm_event_type type,int status)199 irdma_send_cm_event(struct irdma_cm_node *cm_node,
200 		    struct iw_cm_id *cm_id,
201 		    enum iw_cm_event_type type, int status)
202 {
203 	struct iw_cm_event event = {0};
204 
205 	event.event = type;
206 	event.status = status;
207 
208 	irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
209 		    "cm_node %p cm_id=%p state=%d accel=%d event_type=%d status=%d\n",
210 		    cm_node, cm_id, cm_node->accelerated, cm_node->state, type,
211 		    status);
212 
213 	switch (type) {
214 	case IW_CM_EVENT_CONNECT_REQUEST:
215 		if (cm_node->ipv4)
216 			irdma_fill_sockaddr4(cm_node, &event);
217 		else
218 			irdma_fill_sockaddr6(cm_node, &event);
219 		event.provider_data = cm_node;
220 		event.private_data = cm_node->pdata_buf;
221 		event.private_data_len = (u8)cm_node->pdata.size;
222 		event.ird = cm_node->ird_size;
223 		break;
224 	case IW_CM_EVENT_CONNECT_REPLY:
225 		irdma_get_cmevent_info(cm_node, cm_id, &event);
226 		break;
227 	case IW_CM_EVENT_ESTABLISHED:
228 		event.ird = cm_node->ird_size;
229 		event.ord = cm_node->ord_size;
230 		break;
231 	case IW_CM_EVENT_DISCONNECT:
232 	case IW_CM_EVENT_CLOSE:
233 		/* Wait if we are in RTS but havent issued the iwcm event upcall */
234 		if (!cm_node->accelerated)
235 			wait_for_completion(&cm_node->establish_comp);
236 		break;
237 	default:
238 		return -EINVAL;
239 	}
240 
241 	return cm_id->event_handler(cm_id, &event);
242 }
243 
244 /**
245  * irdma_timer_list_prep - add connection nodes to a list to perform timer tasks
246  * @cm_core: cm's core
247  * @timer_list: a timer list to which cm_node will be selected
248  */
249 static void
irdma_timer_list_prep(struct irdma_cm_core * cm_core,struct list_head * timer_list)250 irdma_timer_list_prep(struct irdma_cm_core *cm_core,
251 		      struct list_head *timer_list)
252 {
253 	struct irdma_cm_node *cm_node;
254 	int bkt;
255 
256 	HASH_FOR_EACH_RCU(cm_core->cm_hash_tbl, bkt, cm_node, list) {
257 		if ((cm_node->close_entry || cm_node->send_entry) &&
258 		    atomic_inc_not_zero(&cm_node->refcnt))
259 			list_add(&cm_node->timer_entry, timer_list);
260 	}
261 }
262 
263 /**
264  * irdma_create_event - create cm event
265  * @cm_node: connection's node
266  * @type: Event type to generate
267  */
268 static struct irdma_cm_event *
irdma_create_event(struct irdma_cm_node * cm_node,enum irdma_cm_event_type type)269 irdma_create_event(struct irdma_cm_node *cm_node,
270 		   enum irdma_cm_event_type type)
271 {
272 	struct irdma_cm_event *event;
273 
274 	if (!cm_node->cm_id)
275 		return NULL;
276 
277 	event = kzalloc(sizeof(*event), GFP_ATOMIC);
278 
279 	if (!event)
280 		return NULL;
281 
282 	event->type = type;
283 	event->cm_node = cm_node;
284 	memcpy(event->cm_info.rem_addr, cm_node->rem_addr,
285 	       sizeof(event->cm_info.rem_addr));
286 	memcpy(event->cm_info.loc_addr, cm_node->loc_addr,
287 	       sizeof(event->cm_info.loc_addr));
288 	event->cm_info.rem_port = cm_node->rem_port;
289 	event->cm_info.loc_port = cm_node->loc_port;
290 	event->cm_info.cm_id = cm_node->cm_id;
291 	irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
292 		    "node=%p event=%p type=%u dst=%x src=%x\n", cm_node, event,
293 		    type, event->cm_info.loc_addr[0],
294 		    event->cm_info.rem_addr[0]);
295 	irdma_cm_post_event(event);
296 
297 	return event;
298 }
299 
300 /**
301  * irdma_free_retrans_entry - free send entry
302  * @cm_node: connection's node
303  */
304 static void
irdma_free_retrans_entry(struct irdma_cm_node * cm_node)305 irdma_free_retrans_entry(struct irdma_cm_node *cm_node)
306 {
307 	struct irdma_device *iwdev = cm_node->iwdev;
308 	struct irdma_timer_entry *send_entry;
309 
310 	send_entry = cm_node->send_entry;
311 	if (!send_entry)
312 		return;
313 
314 	cm_node->send_entry = NULL;
315 	irdma_free_sqbuf(&iwdev->vsi, send_entry->sqbuf);
316 	kfree(send_entry);
317 	atomic_dec(&cm_node->refcnt);
318 }
319 
320 /**
321  * irdma_cleanup_retrans_entry - free send entry with lock
322  * @cm_node: connection's node
323  */
324 static void
irdma_cleanup_retrans_entry(struct irdma_cm_node * cm_node)325 irdma_cleanup_retrans_entry(struct irdma_cm_node *cm_node)
326 {
327 	unsigned long flags;
328 
329 	spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
330 	irdma_free_retrans_entry(cm_node);
331 	spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
332 }
333 
334 /**
335  * irdma_form_ah_cm_frame - get a free packet and build frame with address handle
336  * @cm_node: connection's node ionfo to use in frame
337  * @options: pointer to options info
338  * @hdr: pointer mpa header
339  * @pdata: pointer to private data
340  * @flags:  indicates FIN or ACK
341  */
342 static struct irdma_puda_buf *
irdma_form_ah_cm_frame(struct irdma_cm_node * cm_node,struct irdma_kmem_info * options,struct irdma_kmem_info * hdr,struct irdma_mpa_priv_info * pdata,u8 flags)343 irdma_form_ah_cm_frame(struct irdma_cm_node *cm_node,
344 		       struct irdma_kmem_info *options,
345 		       struct irdma_kmem_info *hdr,
346 		       struct irdma_mpa_priv_info *pdata,
347 		       u8 flags)
348 {
349 	struct irdma_puda_buf *sqbuf;
350 	struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
351 	u8 *buf;
352 	struct tcphdr *tcph;
353 	u16 pktsize;
354 	u32 opts_len = 0;
355 	u32 pd_len = 0;
356 	u32 hdr_len = 0;
357 
358 	if (!cm_node->ah || !cm_node->ah->ah_info.ah_valid) {
359 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM, "AH invalid\n");
360 		return NULL;
361 	}
362 
363 	sqbuf = irdma_puda_get_bufpool(vsi->ilq);
364 	if (!sqbuf) {
365 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM, "SQ buf NULL\n");
366 		return NULL;
367 	}
368 
369 	sqbuf->ah_id = cm_node->ah->ah_info.ah_idx;
370 	buf = sqbuf->mem.va;
371 	if (options)
372 		opts_len = (u32)options->size;
373 
374 	if (hdr)
375 		hdr_len = hdr->size;
376 
377 	if (pdata)
378 		pd_len = pdata->size;
379 
380 	pktsize = sizeof(*tcph) + opts_len + hdr_len + pd_len;
381 
382 	memset(buf, 0, pktsize);
383 
384 	sqbuf->totallen = pktsize;
385 	sqbuf->tcphlen = sizeof(*tcph) + opts_len;
386 	sqbuf->scratch = cm_node;
387 
388 	tcph = (struct tcphdr *)buf;
389 	buf += sizeof(*tcph);
390 
391 	tcph->th_sport = htons(cm_node->loc_port);
392 	tcph->th_dport = htons(cm_node->rem_port);
393 	tcph->th_seq = htonl(cm_node->tcp_cntxt.loc_seq_num);
394 
395 	if (flags & SET_ACK) {
396 		cm_node->tcp_cntxt.loc_ack_num = cm_node->tcp_cntxt.rcv_nxt;
397 		tcph->th_ack = htonl(cm_node->tcp_cntxt.loc_ack_num);
398 		tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_ACK);
399 	} else {
400 		tcph->th_ack = 0;
401 	}
402 
403 	if (flags & SET_SYN) {
404 		cm_node->tcp_cntxt.loc_seq_num++;
405 		tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_SYN);
406 	} else {
407 		cm_node->tcp_cntxt.loc_seq_num += hdr_len + pd_len;
408 	}
409 
410 	if (flags & SET_FIN) {
411 		cm_node->tcp_cntxt.loc_seq_num++;
412 		tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_FIN);
413 	}
414 
415 	if (flags & SET_RST)
416 		tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_RST);
417 
418 	tcph->th_off = (u16)((sizeof(*tcph) + opts_len + 3) >> 2);
419 	sqbuf->tcphlen = tcph->th_off << 2;
420 	tcph->th_win = htons(cm_node->tcp_cntxt.rcv_wnd);
421 	tcph->th_urp = 0;
422 
423 	if (opts_len) {
424 		memcpy(buf, options->addr, opts_len);
425 		buf += opts_len;
426 	}
427 
428 	if (hdr_len) {
429 		memcpy(buf, hdr->addr, hdr_len);
430 		buf += hdr_len;
431 	}
432 
433 	if (pdata && pdata->addr)
434 		memcpy(buf, pdata->addr, pdata->size);
435 
436 	atomic_set(&sqbuf->refcount, 1);
437 
438 	irdma_debug_buf(vsi->dev, IRDMA_DEBUG_ILQ, "TRANSMIT ILQ BUFFER",
439 			sqbuf->mem.va, sqbuf->totallen);
440 
441 	return sqbuf;
442 }
443 
444 /**
445  * irdma_form_uda_cm_frame - get a free packet and build frame full tcpip packet
446  * @cm_node: connection's node ionfo to use in frame
447  * @options: pointer to options info
448  * @hdr: pointer mpa header
449  * @pdata: pointer to private data
450  * @flags:  indicates FIN or ACK
451  */
452 static struct irdma_puda_buf *
irdma_form_uda_cm_frame(struct irdma_cm_node * cm_node,struct irdma_kmem_info * options,struct irdma_kmem_info * hdr,struct irdma_mpa_priv_info * pdata,u8 flags)453 irdma_form_uda_cm_frame(struct irdma_cm_node *cm_node,
454 			struct irdma_kmem_info *options,
455 			struct irdma_kmem_info *hdr,
456 			struct irdma_mpa_priv_info *pdata,
457 			u8 flags)
458 {
459 	struct irdma_puda_buf *sqbuf;
460 	struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
461 	u8 *buf;
462 
463 	struct tcphdr *tcph;
464 	struct ip *iph;
465 	struct ip6_hdr *ip6h;
466 	struct ether_header *ethh;
467 	u16 pktsize;
468 	u16 eth_hlen = ETH_HLEN;
469 	u32 opts_len = 0;
470 	u32 pd_len = 0;
471 	u32 hdr_len = 0;
472 
473 	u16 vtag;
474 
475 	sqbuf = irdma_puda_get_bufpool(vsi->ilq);
476 	if (!sqbuf)
477 		return NULL;
478 
479 	buf = sqbuf->mem.va;
480 
481 	if (options)
482 		opts_len = (u32)options->size;
483 
484 	if (hdr)
485 		hdr_len = hdr->size;
486 
487 	if (pdata)
488 		pd_len = pdata->size;
489 
490 	if (cm_node->vlan_id < VLAN_N_VID)
491 		eth_hlen += 4;
492 
493 	if (cm_node->ipv4)
494 		pktsize = sizeof(*iph) + sizeof(*tcph);
495 	else
496 		pktsize = sizeof(*ip6h) + sizeof(*tcph);
497 	pktsize += opts_len + hdr_len + pd_len;
498 
499 	memset(buf, 0, eth_hlen + pktsize);
500 
501 	sqbuf->totallen = pktsize + eth_hlen;
502 	sqbuf->maclen = eth_hlen;
503 	sqbuf->tcphlen = sizeof(*tcph) + opts_len;
504 	sqbuf->scratch = cm_node;
505 
506 	ethh = (struct ether_header *)buf;
507 	buf += eth_hlen;
508 
509 	if (cm_node->do_lpb)
510 		sqbuf->do_lpb = true;
511 
512 	if (cm_node->ipv4) {
513 		sqbuf->ipv4 = true;
514 
515 		iph = (struct ip *)buf;
516 		buf += sizeof(*iph);
517 		tcph = (struct tcphdr *)buf;
518 		buf += sizeof(*tcph);
519 
520 		ether_addr_copy(ethh->ether_dhost, cm_node->rem_mac);
521 		ether_addr_copy(ethh->ether_shost, cm_node->loc_mac);
522 		if (cm_node->vlan_id < VLAN_N_VID) {
523 			((struct ether_vlan_header *)ethh)->evl_proto =
524 			    htons(ETH_P_8021Q);
525 			vtag = (cm_node->user_pri << VLAN_PRIO_SHIFT) |
526 			    cm_node->vlan_id;
527 			((struct ether_vlan_header *)ethh)->evl_tag = htons(vtag);
528 
529 			((struct ether_vlan_header *)ethh)->evl_encap_proto =
530 			    htons(ETH_P_IP);
531 		} else {
532 			ethh->ether_type = htons(ETH_P_IP);
533 		}
534 
535 		iph->ip_v = IPVERSION;
536 		iph->ip_hl = 5;	/* 5 * 4Byte words, IP headr len */
537 		iph->ip_tos = cm_node->tos;
538 		iph->ip_len = htons(pktsize);
539 		iph->ip_id = htons(++cm_node->tcp_cntxt.loc_id);
540 
541 		iph->ip_off = htons(0x4000);
542 		iph->ip_ttl = 0x40;
543 		iph->ip_p = IPPROTO_TCP;
544 		iph->ip_src.s_addr = htonl(cm_node->loc_addr[0]);
545 		iph->ip_dst.s_addr = htonl(cm_node->rem_addr[0]);
546 	} else {
547 		sqbuf->ipv4 = false;
548 		ip6h = (struct ip6_hdr *)buf;
549 		buf += sizeof(*ip6h);
550 		tcph = (struct tcphdr *)buf;
551 		buf += sizeof(*tcph);
552 
553 		ether_addr_copy(ethh->ether_dhost, cm_node->rem_mac);
554 		ether_addr_copy(ethh->ether_shost, cm_node->loc_mac);
555 		if (cm_node->vlan_id < VLAN_N_VID) {
556 			((struct ether_vlan_header *)ethh)->evl_proto =
557 			    htons(ETH_P_8021Q);
558 			vtag = (cm_node->user_pri << VLAN_PRIO_SHIFT) |
559 			    cm_node->vlan_id;
560 			((struct ether_vlan_header *)ethh)->evl_tag = htons(vtag);
561 			((struct ether_vlan_header *)ethh)->evl_encap_proto =
562 			    htons(ETH_P_IPV6);
563 		} else {
564 			ethh->ether_type = htons(ETH_P_IPV6);
565 		}
566 		ip6h->ip6_vfc = 6 << 4;
567 		ip6h->ip6_vfc |= cm_node->tos >> 4;
568 		ip6h->ip6_flow = cm_node->tos << 20;
569 		ip6h->ip6_plen = htons(pktsize - sizeof(*ip6h));
570 		ip6h->ip6_nxt = 6;
571 		ip6h->ip6_hops = 128;
572 		irdma_copy_ip_htonl(ip6h->ip6_src.__u6_addr.__u6_addr32,
573 				    cm_node->loc_addr);
574 		irdma_copy_ip_htonl(ip6h->ip6_dst.__u6_addr.__u6_addr32,
575 				    cm_node->rem_addr);
576 	}
577 
578 	tcph->th_sport = htons(cm_node->loc_port);
579 	tcph->th_dport = htons(cm_node->rem_port);
580 	tcph->th_seq = htonl(cm_node->tcp_cntxt.loc_seq_num);
581 
582 	if (flags & SET_ACK) {
583 		cm_node->tcp_cntxt.loc_ack_num = cm_node->tcp_cntxt.rcv_nxt;
584 		tcph->th_ack = htonl(cm_node->tcp_cntxt.loc_ack_num);
585 		tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_ACK);
586 	} else {
587 		tcph->th_ack = 0;
588 	}
589 
590 	if (flags & SET_SYN) {
591 		cm_node->tcp_cntxt.loc_seq_num++;
592 		tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_SYN);
593 	} else {
594 		cm_node->tcp_cntxt.loc_seq_num += hdr_len + pd_len;
595 	}
596 
597 	if (flags & SET_FIN) {
598 		cm_node->tcp_cntxt.loc_seq_num++;
599 		tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_FIN);
600 	}
601 
602 	if (flags & SET_RST)
603 		tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_RST);
604 
605 	tcph->th_off = (u16)((sizeof(*tcph) + opts_len + 3) >> 2);
606 	sqbuf->tcphlen = tcph->th_off << 2;
607 	tcph->th_win = htons(cm_node->tcp_cntxt.rcv_wnd);
608 	tcph->th_urp = 0;
609 
610 	if (opts_len) {
611 		memcpy(buf, options->addr, opts_len);
612 		buf += opts_len;
613 	}
614 
615 	if (hdr_len) {
616 		memcpy(buf, hdr->addr, hdr_len);
617 		buf += hdr_len;
618 	}
619 
620 	if (pdata && pdata->addr)
621 		memcpy(buf, pdata->addr, pdata->size);
622 
623 	atomic_set(&sqbuf->refcount, 1);
624 
625 	irdma_debug_buf(vsi->dev, IRDMA_DEBUG_ILQ, "TRANSMIT ILQ BUFFER",
626 			sqbuf->mem.va, sqbuf->totallen);
627 
628 	return sqbuf;
629 }
630 
631 /**
632  * irdma_send_reset - Send RST packet
633  * @cm_node: connection's node
634  */
635 int
irdma_send_reset(struct irdma_cm_node * cm_node)636 irdma_send_reset(struct irdma_cm_node *cm_node)
637 {
638 	struct irdma_puda_buf *sqbuf;
639 	int flags = SET_RST | SET_ACK;
640 
641 	sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL, NULL, NULL,
642 						flags);
643 	if (!sqbuf)
644 		return -ENOMEM;
645 
646 	irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
647 		    "caller: %pS cm_node %p cm_id=%p accel=%d state=%d rem_port=0x%04x, loc_port=0x%04x rem_addr=%x loc_addr=%x\n",
648 		    __builtin_return_address(0), cm_node, cm_node->cm_id,
649 		    cm_node->accelerated, cm_node->state, cm_node->rem_port,
650 		    cm_node->loc_port, cm_node->rem_addr[0],
651 		    cm_node->loc_addr[0]);
652 
653 	return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 0,
654 				       1);
655 }
656 
657 /**
658  * irdma_active_open_err - send event for active side cm error
659  * @cm_node: connection's node
660  * @reset: Flag to send reset or not
661  */
662 static void
irdma_active_open_err(struct irdma_cm_node * cm_node,bool reset)663 irdma_active_open_err(struct irdma_cm_node *cm_node, bool reset)
664 {
665 	irdma_cleanup_retrans_entry(cm_node);
666 	cm_node->cm_core->stats_connect_errs++;
667 	if (reset) {
668 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
669 			    "cm_node=%p state=%d\n", cm_node, cm_node->state);
670 		atomic_inc(&cm_node->refcnt);
671 		irdma_send_reset(cm_node);
672 	}
673 
674 	cm_node->state = IRDMA_CM_STATE_CLOSED;
675 	irdma_create_event(cm_node, IRDMA_CM_EVENT_ABORTED);
676 }
677 
678 /**
679  * irdma_passive_open_err - handle passive side cm error
680  * @cm_node: connection's node
681  * @reset: send reset or just free cm_node
682  */
683 static void
irdma_passive_open_err(struct irdma_cm_node * cm_node,bool reset)684 irdma_passive_open_err(struct irdma_cm_node *cm_node, bool reset)
685 {
686 	irdma_cleanup_retrans_entry(cm_node);
687 	cm_node->cm_core->stats_passive_errs++;
688 	cm_node->state = IRDMA_CM_STATE_CLOSED;
689 	irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
690 		    "cm_node=%p state=%d\n", cm_node, cm_node->state);
691 	if (reset)
692 		irdma_send_reset(cm_node);
693 	else
694 		irdma_rem_ref_cm_node(cm_node);
695 }
696 
697 /**
698  * irdma_event_connect_error - to create connect error event
699  * @event: cm information for connect event
700  */
701 static void
irdma_event_connect_error(struct irdma_cm_event * event)702 irdma_event_connect_error(struct irdma_cm_event *event)
703 {
704 	struct irdma_qp *iwqp;
705 	struct iw_cm_id *cm_id;
706 
707 	cm_id = event->cm_node->cm_id;
708 	if (!cm_id)
709 		return;
710 
711 	iwqp = cm_id->provider_data;
712 
713 	if (!iwqp || !iwqp->iwdev)
714 		return;
715 
716 	iwqp->cm_id = NULL;
717 	cm_id->provider_data = NULL;
718 	irdma_send_cm_event(event->cm_node, cm_id, IW_CM_EVENT_CONNECT_REPLY,
719 			    -ECONNRESET);
720 	irdma_rem_ref_cm_node(event->cm_node);
721 }
722 
723 /**
724  * irdma_process_options - process options from TCP header
725  * @cm_node: connection's node
726  * @optionsloc: point to start of options
727  * @optionsize: size of all options
728  * @syn_pkt: flag if syn packet
729  */
730 static int
irdma_process_options(struct irdma_cm_node * cm_node,u8 * optionsloc,u32 optionsize,u32 syn_pkt)731 irdma_process_options(struct irdma_cm_node *cm_node, u8 *optionsloc,
732 		      u32 optionsize, u32 syn_pkt)
733 {
734 	u32 tmp;
735 	u32 offset = 0;
736 	union all_known_options *all_options;
737 	char got_mss_option = 0;
738 
739 	while (offset < optionsize) {
740 		all_options = (union all_known_options *)(optionsloc + offset);
741 		switch (all_options->base.optionnum) {
742 		case OPTION_NUM_EOL:
743 			offset = optionsize;
744 			break;
745 		case OPTION_NUM_NONE:
746 			offset += 1;
747 			continue;
748 		case OPTION_NUM_MSS:
749 			irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
750 				    "MSS Length: %d Offset: %d Size: %d\n",
751 				    all_options->mss.len, offset, optionsize);
752 			got_mss_option = 1;
753 			if (all_options->mss.len != 4)
754 				return -EINVAL;
755 			tmp = ntohs(all_options->mss.mss);
756 			if ((cm_node->ipv4 &&
757 			     (tmp + IRDMA_MTU_TO_MSS_IPV4) < IRDMA_MIN_MTU_IPV4) ||
758 			    (!cm_node->ipv4 &&
759 			     (tmp + IRDMA_MTU_TO_MSS_IPV6) < IRDMA_MIN_MTU_IPV6))
760 				return -EINVAL;
761 			if (tmp < cm_node->tcp_cntxt.mss)
762 				cm_node->tcp_cntxt.mss = tmp;
763 			break;
764 		case OPTION_NUM_WINDOW_SCALE:
765 			cm_node->tcp_cntxt.snd_wscale =
766 			    all_options->windowscale.shiftcount;
767 			break;
768 		default:
769 			irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
770 				    "Unsupported TCP Option: %x\n",
771 				    all_options->base.optionnum);
772 			break;
773 		}
774 		offset += all_options->base.len;
775 	}
776 	if (!got_mss_option && syn_pkt)
777 		cm_node->tcp_cntxt.mss = IRDMA_CM_DEFAULT_MSS;
778 
779 	return 0;
780 }
781 
782 /**
783  * irdma_handle_tcp_options - setup TCP context info after parsing TCP options
784  * @cm_node: connection's node
785  * @tcph: pointer tcp header
786  * @optionsize: size of options rcvd
787  * @passive: active or passive flag
788  */
789 static int
irdma_handle_tcp_options(struct irdma_cm_node * cm_node,struct tcphdr * tcph,int optionsize,int passive)790 irdma_handle_tcp_options(struct irdma_cm_node *cm_node,
791 			 struct tcphdr *tcph, int optionsize,
792 			 int passive)
793 {
794 	u8 *optionsloc = (u8 *)&tcph[1];
795 	int ret;
796 
797 	if (optionsize) {
798 		ret = irdma_process_options(cm_node, optionsloc, optionsize,
799 					    (u32)tcp_get_flags(tcph) & TH_SYN);
800 		if (ret) {
801 			irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
802 				    "Node %p, Sending Reset\n", cm_node);
803 			if (passive)
804 				irdma_passive_open_err(cm_node, true);
805 			else
806 				irdma_active_open_err(cm_node, true);
807 			return ret;
808 		}
809 	}
810 
811 	cm_node->tcp_cntxt.snd_wnd = ntohs(tcph->th_win)
812 	    << cm_node->tcp_cntxt.snd_wscale;
813 
814 	if (cm_node->tcp_cntxt.snd_wnd > cm_node->tcp_cntxt.max_snd_wnd)
815 		cm_node->tcp_cntxt.max_snd_wnd = cm_node->tcp_cntxt.snd_wnd;
816 
817 	return 0;
818 }
819 
820 /**
821  * irdma_build_mpa_v1 - build a MPA V1 frame
822  * @cm_node: connection's node
823  * @start_addr: address where to build frame
824  * @mpa_key: to do read0 or write0
825  */
826 static void
irdma_build_mpa_v1(struct irdma_cm_node * cm_node,void * start_addr,u8 mpa_key)827 irdma_build_mpa_v1(struct irdma_cm_node *cm_node, void *start_addr,
828 		   u8 mpa_key)
829 {
830 	struct ietf_mpa_v1 *mpa_frame = start_addr;
831 
832 	switch (mpa_key) {
833 	case MPA_KEY_REQUEST:
834 		memcpy(mpa_frame->key, IEFT_MPA_KEY_REQ, IETF_MPA_KEY_SIZE);
835 		break;
836 	case MPA_KEY_REPLY:
837 		memcpy(mpa_frame->key, IEFT_MPA_KEY_REP, IETF_MPA_KEY_SIZE);
838 		break;
839 	default:
840 		break;
841 	}
842 	mpa_frame->flags = IETF_MPA_FLAGS_CRC;
843 	mpa_frame->rev = cm_node->mpa_frame_rev;
844 	mpa_frame->priv_data_len = htons(cm_node->pdata.size);
845 }
846 
847 /**
848  * irdma_build_mpa_v2 - build a MPA V2 frame
849  * @cm_node: connection's node
850  * @start_addr: buffer start address
851  * @mpa_key: to do read0 or write0
852  */
853 static void
irdma_build_mpa_v2(struct irdma_cm_node * cm_node,void * start_addr,u8 mpa_key)854 irdma_build_mpa_v2(struct irdma_cm_node *cm_node, void *start_addr,
855 		   u8 mpa_key)
856 {
857 	struct ietf_mpa_v2 *mpa_frame = start_addr;
858 	struct ietf_rtr_msg *rtr_msg = &mpa_frame->rtr_msg;
859 	u16 ctrl_ird, ctrl_ord;
860 
861 	/* initialize the upper 5 bytes of the frame */
862 	irdma_build_mpa_v1(cm_node, start_addr, mpa_key);
863 	mpa_frame->flags |= IETF_MPA_V2_FLAG;
864 	if (cm_node->iwdev->iw_ooo) {
865 		mpa_frame->flags |= IETF_MPA_FLAGS_MARKERS;
866 		cm_node->rcv_mark_en = true;
867 	}
868 	mpa_frame->priv_data_len = cpu_to_be16(be16_to_cpu(mpa_frame->priv_data_len) +
869 					       IETF_RTR_MSG_SIZE);
870 
871 	/* initialize RTR msg */
872 	if (cm_node->mpav2_ird_ord == IETF_NO_IRD_ORD) {
873 		ctrl_ird = IETF_NO_IRD_ORD;
874 		ctrl_ord = IETF_NO_IRD_ORD;
875 	} else {
876 		ctrl_ird = (cm_node->ird_size > IETF_NO_IRD_ORD) ?
877 		    IETF_NO_IRD_ORD :
878 		    cm_node->ird_size;
879 		ctrl_ord = (cm_node->ord_size > IETF_NO_IRD_ORD) ?
880 		    IETF_NO_IRD_ORD :
881 		    cm_node->ord_size;
882 	}
883 	ctrl_ird |= IETF_PEER_TO_PEER;
884 
885 	switch (mpa_key) {
886 	case MPA_KEY_REQUEST:
887 		ctrl_ord |= IETF_RDMA0_WRITE;
888 		ctrl_ord |= IETF_RDMA0_READ;
889 		break;
890 	case MPA_KEY_REPLY:
891 		switch (cm_node->send_rdma0_op) {
892 		case SEND_RDMA_WRITE_ZERO:
893 			ctrl_ord |= IETF_RDMA0_WRITE;
894 			break;
895 		case SEND_RDMA_READ_ZERO:
896 			ctrl_ord |= IETF_RDMA0_READ;
897 			break;
898 		}
899 		break;
900 	default:
901 		break;
902 	}
903 	rtr_msg->ctrl_ird = htons(ctrl_ird);
904 	rtr_msg->ctrl_ord = htons(ctrl_ord);
905 }
906 
907 /**
908  * irdma_cm_build_mpa_frame - build mpa frame for mpa version 1 or version 2
909  * @cm_node: connection's node
910  * @mpa: mpa: data buffer
911  * @mpa_key: to do read0 or write0
912  */
913 static int
irdma_cm_build_mpa_frame(struct irdma_cm_node * cm_node,struct irdma_kmem_info * mpa,u8 mpa_key)914 irdma_cm_build_mpa_frame(struct irdma_cm_node *cm_node,
915 			 struct irdma_kmem_info *mpa, u8 mpa_key)
916 {
917 	int hdr_len = 0;
918 
919 	switch (cm_node->mpa_frame_rev) {
920 	case IETF_MPA_V1:
921 		hdr_len = sizeof(struct ietf_mpa_v1);
922 		irdma_build_mpa_v1(cm_node, mpa->addr, mpa_key);
923 		break;
924 	case IETF_MPA_V2:
925 		hdr_len = sizeof(struct ietf_mpa_v2);
926 		irdma_build_mpa_v2(cm_node, mpa->addr, mpa_key);
927 		break;
928 	default:
929 		break;
930 	}
931 
932 	return hdr_len;
933 }
934 
935 /**
936  * irdma_send_mpa_request - active node send mpa request to passive node
937  * @cm_node: connection's node
938  */
939 static int
irdma_send_mpa_request(struct irdma_cm_node * cm_node)940 irdma_send_mpa_request(struct irdma_cm_node *cm_node)
941 {
942 	struct irdma_puda_buf *sqbuf;
943 
944 	cm_node->mpa_hdr.addr = &cm_node->mpa_v2_frame;
945 	cm_node->mpa_hdr.size = irdma_cm_build_mpa_frame(cm_node,
946 							 &cm_node->mpa_hdr,
947 							 MPA_KEY_REQUEST);
948 	if (!cm_node->mpa_hdr.size) {
949 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
950 			    "mpa size = %d\n", cm_node->mpa_hdr.size);
951 		return -EINVAL;
952 	}
953 
954 	sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL,
955 						&cm_node->mpa_hdr,
956 						&cm_node->pdata, SET_ACK);
957 	if (!sqbuf)
958 		return -ENOMEM;
959 
960 	return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
961 				       0);
962 }
963 
964 /**
965  * irdma_send_mpa_reject -
966  * @cm_node: connection's node
967  * @pdata: reject data for connection
968  * @plen: length of reject data
969  */
970 static int
irdma_send_mpa_reject(struct irdma_cm_node * cm_node,const void * pdata,u8 plen)971 irdma_send_mpa_reject(struct irdma_cm_node *cm_node,
972 		      const void *pdata, u8 plen)
973 {
974 	struct irdma_puda_buf *sqbuf;
975 	struct irdma_mpa_priv_info priv_info;
976 
977 	cm_node->mpa_hdr.addr = &cm_node->mpa_v2_frame;
978 	cm_node->mpa_hdr.size = irdma_cm_build_mpa_frame(cm_node,
979 							 &cm_node->mpa_hdr,
980 							 MPA_KEY_REPLY);
981 
982 	cm_node->mpa_v2_frame.flags |= IETF_MPA_FLAGS_REJECT;
983 	priv_info.addr = pdata;
984 	priv_info.size = plen;
985 
986 	sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL,
987 						&cm_node->mpa_hdr, &priv_info,
988 						SET_ACK | SET_FIN);
989 	if (!sqbuf)
990 		return -ENOMEM;
991 
992 	cm_node->state = IRDMA_CM_STATE_FIN_WAIT1;
993 
994 	return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
995 				       0);
996 }
997 
998 /**
999  * irdma_negotiate_mpa_v2_ird_ord - negotiate MPAv2 IRD/ORD
1000  * @cm_node: connection's node
1001  * @buf: Data pointer
1002  */
1003 static int
irdma_negotiate_mpa_v2_ird_ord(struct irdma_cm_node * cm_node,u8 * buf)1004 irdma_negotiate_mpa_v2_ird_ord(struct irdma_cm_node *cm_node,
1005 			       u8 *buf)
1006 {
1007 	struct ietf_mpa_v2 *mpa_v2_frame;
1008 	struct ietf_rtr_msg *rtr_msg;
1009 	u16 ird_size;
1010 	u16 ord_size;
1011 	u16 ctrl_ord;
1012 	u16 ctrl_ird;
1013 
1014 	mpa_v2_frame = (struct ietf_mpa_v2 *)buf;
1015 	rtr_msg = &mpa_v2_frame->rtr_msg;
1016 
1017 	/* parse rtr message */
1018 	ctrl_ord = ntohs(rtr_msg->ctrl_ord);
1019 	ctrl_ird = ntohs(rtr_msg->ctrl_ird);
1020 	ird_size = ctrl_ird & IETF_NO_IRD_ORD;
1021 	ord_size = ctrl_ord & IETF_NO_IRD_ORD;
1022 
1023 	if (!(ctrl_ird & IETF_PEER_TO_PEER))
1024 		return -EOPNOTSUPP;
1025 
1026 	if (ird_size == IETF_NO_IRD_ORD || ord_size == IETF_NO_IRD_ORD) {
1027 		cm_node->mpav2_ird_ord = IETF_NO_IRD_ORD;
1028 		goto negotiate_done;
1029 	}
1030 
1031 	if (cm_node->state != IRDMA_CM_STATE_MPAREQ_SENT) {
1032 		/* responder */
1033 		if (!ord_size && (ctrl_ord & IETF_RDMA0_READ))
1034 			cm_node->ird_size = 1;
1035 		if (cm_node->ord_size > ird_size)
1036 			cm_node->ord_size = ird_size;
1037 	} else {
1038 		/* initiator */
1039 		if (!ird_size && (ctrl_ord & IETF_RDMA0_READ))
1040 			/* Remote peer doesn't support RDMA0_READ */
1041 			return -EOPNOTSUPP;
1042 
1043 		if (cm_node->ord_size > ird_size)
1044 			cm_node->ord_size = ird_size;
1045 
1046 		if (cm_node->ird_size < ord_size)
1047 			/* no resources available */
1048 			return -EINVAL;
1049 	}
1050 
1051 negotiate_done:
1052 	if (ctrl_ord & IETF_RDMA0_READ)
1053 		cm_node->send_rdma0_op = SEND_RDMA_READ_ZERO;
1054 	else if (ctrl_ord & IETF_RDMA0_WRITE)
1055 		cm_node->send_rdma0_op = SEND_RDMA_WRITE_ZERO;
1056 	else
1057 		/* Not supported RDMA0 operation */
1058 		return -EOPNOTSUPP;
1059 
1060 	irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1061 		    "MPAV2 Negotiated ORD: %d, IRD: %d\n", cm_node->ord_size,
1062 		    cm_node->ird_size);
1063 	return 0;
1064 }
1065 
1066 /**
1067  * irdma_parse_mpa - process an IETF MPA frame
1068  * @cm_node: connection's node
1069  * @buf: Data pointer
1070  * @type: to return accept or reject
1071  * @len: Len of mpa buffer
1072  */
1073 static int
irdma_parse_mpa(struct irdma_cm_node * cm_node,u8 * buf,u32 * type,u32 len)1074 irdma_parse_mpa(struct irdma_cm_node *cm_node, u8 *buf, u32 *type,
1075 		u32 len)
1076 {
1077 	struct ietf_mpa_v1 *mpa_frame;
1078 	int mpa_hdr_len, priv_data_len, ret;
1079 
1080 	*type = IRDMA_MPA_REQUEST_ACCEPT;
1081 
1082 	if (len < sizeof(*mpa_frame)) {
1083 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1084 			    "ietf buffer small (%x)\n", len);
1085 		return -EINVAL;
1086 	}
1087 
1088 	mpa_frame = (struct ietf_mpa_v1 *)buf;
1089 	mpa_hdr_len = sizeof(*mpa_frame);
1090 	priv_data_len = ntohs(mpa_frame->priv_data_len);
1091 
1092 	if (priv_data_len > IETF_MAX_PRIV_DATA_LEN) {
1093 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1094 			    "private_data too big %d\n", priv_data_len);
1095 		return -EOVERFLOW;
1096 	}
1097 
1098 	if (mpa_frame->rev != IETF_MPA_V1 && mpa_frame->rev != IETF_MPA_V2) {
1099 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1100 			    "unsupported mpa rev = %d\n", mpa_frame->rev);
1101 		return -EINVAL;
1102 	}
1103 
1104 	if (mpa_frame->rev > cm_node->mpa_frame_rev) {
1105 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1106 			    "rev %d\n", mpa_frame->rev);
1107 		return -EINVAL;
1108 	}
1109 
1110 	cm_node->mpa_frame_rev = mpa_frame->rev;
1111 	if (cm_node->state != IRDMA_CM_STATE_MPAREQ_SENT) {
1112 		if (memcmp(mpa_frame->key, IEFT_MPA_KEY_REQ,
1113 			   IETF_MPA_KEY_SIZE)) {
1114 			irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1115 				    "Unexpected MPA Key received\n");
1116 			return -EINVAL;
1117 		}
1118 	} else {
1119 		if (memcmp(mpa_frame->key, IEFT_MPA_KEY_REP,
1120 			   IETF_MPA_KEY_SIZE)) {
1121 			irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1122 				    "Unexpected MPA Key received\n");
1123 			return -EINVAL;
1124 		}
1125 	}
1126 
1127 	if (priv_data_len + mpa_hdr_len > len) {
1128 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1129 			    "ietf buffer len(%x + %x != %x)\n", priv_data_len,
1130 			    mpa_hdr_len, len);
1131 		return -EOVERFLOW;
1132 	}
1133 
1134 	if (len > IRDMA_MAX_CM_BUF) {
1135 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1136 			    "ietf buffer large len = %d\n", len);
1137 		return -EOVERFLOW;
1138 	}
1139 
1140 	switch (mpa_frame->rev) {
1141 	case IETF_MPA_V2:
1142 		mpa_hdr_len += IETF_RTR_MSG_SIZE;
1143 		ret = irdma_negotiate_mpa_v2_ird_ord(cm_node, buf);
1144 		if (ret)
1145 			return ret;
1146 		break;
1147 	case IETF_MPA_V1:
1148 	default:
1149 		break;
1150 	}
1151 
1152 	memcpy(cm_node->pdata_buf, buf + mpa_hdr_len, priv_data_len);
1153 	cm_node->pdata.size = priv_data_len;
1154 
1155 	if (mpa_frame->flags & IETF_MPA_FLAGS_REJECT)
1156 		*type = IRDMA_MPA_REQUEST_REJECT;
1157 
1158 	if (mpa_frame->flags & IETF_MPA_FLAGS_MARKERS)
1159 		cm_node->snd_mark_en = true;
1160 
1161 	return 0;
1162 }
1163 
1164 /**
1165  * irdma_schedule_cm_timer
1166  * @cm_node: connection's node
1167  * @sqbuf: buffer to send
1168  * @type: if it is send or close
1169  * @send_retrans: if rexmits to be done
1170  * @close_when_complete: is cm_node to be removed
1171  *
1172  * note - cm_node needs to be protected before calling this. Encase in:
1173  *		irdma_rem_ref_cm_node(cm_core, cm_node);
1174  *		irdma_schedule_cm_timer(...)
1175  *		atomic_inc(&cm_node->refcnt);
1176  */
1177 int
irdma_schedule_cm_timer(struct irdma_cm_node * cm_node,struct irdma_puda_buf * sqbuf,enum irdma_timer_type type,int send_retrans,int close_when_complete)1178 irdma_schedule_cm_timer(struct irdma_cm_node *cm_node,
1179 			struct irdma_puda_buf *sqbuf,
1180 			enum irdma_timer_type type, int send_retrans,
1181 			int close_when_complete)
1182 {
1183 	struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
1184 	struct irdma_cm_core *cm_core = cm_node->cm_core;
1185 	struct irdma_timer_entry *new_send;
1186 	u32 was_timer_set;
1187 	unsigned long flags;
1188 
1189 	new_send = kzalloc(sizeof(*new_send), GFP_ATOMIC);
1190 	if (!new_send) {
1191 		if (type != IRDMA_TIMER_TYPE_CLOSE)
1192 			irdma_free_sqbuf(vsi, sqbuf);
1193 		return -ENOMEM;
1194 	}
1195 
1196 	new_send->retrycount = IRDMA_DEFAULT_RETRYS;
1197 	new_send->retranscount = IRDMA_DEFAULT_RETRANS;
1198 	new_send->sqbuf = sqbuf;
1199 	new_send->timetosend = jiffies;
1200 	new_send->type = type;
1201 	new_send->send_retrans = send_retrans;
1202 	new_send->close_when_complete = close_when_complete;
1203 
1204 	if (type == IRDMA_TIMER_TYPE_CLOSE) {
1205 		new_send->timetosend += (HZ / 10);
1206 		if (cm_node->close_entry) {
1207 			kfree(new_send);
1208 			irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1209 				    "already close entry\n");
1210 			return -EINVAL;
1211 		}
1212 
1213 		cm_node->close_entry = new_send;
1214 	} else {		/* type == IRDMA_TIMER_TYPE_SEND */
1215 		spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
1216 		cm_node->send_entry = new_send;
1217 		atomic_inc(&cm_node->refcnt);
1218 		spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
1219 		new_send->timetosend = jiffies + IRDMA_RETRY_TIMEOUT;
1220 
1221 		atomic_inc(&sqbuf->refcount);
1222 		irdma_puda_send_buf(vsi->ilq, sqbuf);
1223 		if (!send_retrans) {
1224 			irdma_cleanup_retrans_entry(cm_node);
1225 			if (close_when_complete)
1226 				irdma_rem_ref_cm_node(cm_node);
1227 			return 0;
1228 		}
1229 	}
1230 
1231 	spin_lock_irqsave(&cm_core->ht_lock, flags);
1232 	was_timer_set = timer_pending(&cm_core->tcp_timer);
1233 
1234 	if (!was_timer_set) {
1235 		cm_core->tcp_timer.expires = new_send->timetosend;
1236 		add_timer(&cm_core->tcp_timer);
1237 	}
1238 	spin_unlock_irqrestore(&cm_core->ht_lock, flags);
1239 
1240 	return 0;
1241 }
1242 
1243 /**
1244  * irdma_retrans_expired - Could not rexmit the packet
1245  * @cm_node: connection's node
1246  */
1247 static void
irdma_retrans_expired(struct irdma_cm_node * cm_node)1248 irdma_retrans_expired(struct irdma_cm_node *cm_node)
1249 {
1250 	enum irdma_cm_node_state state = cm_node->state;
1251 
1252 	cm_node->state = IRDMA_CM_STATE_CLOSED;
1253 	switch (state) {
1254 	case IRDMA_CM_STATE_SYN_RCVD:
1255 	case IRDMA_CM_STATE_CLOSING:
1256 		irdma_rem_ref_cm_node(cm_node);
1257 		break;
1258 	case IRDMA_CM_STATE_FIN_WAIT1:
1259 	case IRDMA_CM_STATE_LAST_ACK:
1260 		irdma_send_reset(cm_node);
1261 		break;
1262 	default:
1263 		atomic_inc(&cm_node->refcnt);
1264 		irdma_send_reset(cm_node);
1265 		irdma_create_event(cm_node, IRDMA_CM_EVENT_ABORTED);
1266 		break;
1267 	}
1268 }
1269 
1270 /**
1271  * irdma_handle_close_entry - for handling retry/timeouts
1272  * @cm_node: connection's node
1273  * @rem_node: flag for remove cm_node
1274  */
1275 static void
irdma_handle_close_entry(struct irdma_cm_node * cm_node,u32 rem_node)1276 irdma_handle_close_entry(struct irdma_cm_node *cm_node,
1277 			 u32 rem_node)
1278 {
1279 	struct irdma_timer_entry *close_entry = cm_node->close_entry;
1280 	struct irdma_qp *iwqp;
1281 	unsigned long flags;
1282 
1283 	if (!close_entry)
1284 		return;
1285 	iwqp = (struct irdma_qp *)close_entry->sqbuf;
1286 	if (iwqp) {
1287 		spin_lock_irqsave(&iwqp->lock, flags);
1288 		if (iwqp->cm_id) {
1289 			iwqp->hw_tcp_state = IRDMA_TCP_STATE_CLOSED;
1290 			iwqp->hw_iwarp_state = IRDMA_QP_STATE_ERROR;
1291 			iwqp->last_aeq = IRDMA_AE_RESET_SENT;
1292 			iwqp->ibqp_state = IB_QPS_ERR;
1293 			spin_unlock_irqrestore(&iwqp->lock, flags);
1294 			irdma_cm_disconn(iwqp);
1295 		} else {
1296 			spin_unlock_irqrestore(&iwqp->lock, flags);
1297 		}
1298 	} else if (rem_node) {
1299 		/* TIME_WAIT state */
1300 		irdma_rem_ref_cm_node(cm_node);
1301 	}
1302 
1303 	kfree(close_entry);
1304 	cm_node->close_entry = NULL;
1305 }
1306 
1307 /**
1308  * irdma_cm_timer_tick - system's timer expired callback
1309  * @t: Pointer to timer_list
1310  */
1311 static void
irdma_cm_timer_tick(struct timer_list * t)1312 irdma_cm_timer_tick(struct timer_list *t)
1313 {
1314 	unsigned long nexttimeout = jiffies + IRDMA_LONG_TIME;
1315 	struct irdma_cm_node *cm_node;
1316 	struct irdma_timer_entry *send_entry, *close_entry;
1317 	struct list_head *list_core_temp;
1318 	struct list_head *list_node;
1319 	struct irdma_cm_core *cm_core = timer_container_of(cm_core, t, tcp_timer);
1320 	struct irdma_sc_vsi *vsi;
1321 	u32 settimer = 0;
1322 	unsigned long timetosend;
1323 	unsigned long flags;
1324 	struct list_head timer_list;
1325 
1326 	INIT_LIST_HEAD(&timer_list);
1327 
1328 	rcu_read_lock();
1329 	irdma_timer_list_prep(cm_core, &timer_list);
1330 	rcu_read_unlock();
1331 
1332 	list_for_each_safe(list_node, list_core_temp, &timer_list) {
1333 		cm_node = container_of(list_node, struct irdma_cm_node,
1334 				       timer_entry);
1335 		close_entry = cm_node->close_entry;
1336 
1337 		if (close_entry) {
1338 			if (time_after(close_entry->timetosend, jiffies)) {
1339 				if (nexttimeout > close_entry->timetosend ||
1340 				    !settimer) {
1341 					nexttimeout = close_entry->timetosend;
1342 					settimer = 1;
1343 				}
1344 			} else {
1345 				irdma_handle_close_entry(cm_node, 1);
1346 			}
1347 		}
1348 
1349 		spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
1350 
1351 		send_entry = cm_node->send_entry;
1352 		if (!send_entry)
1353 			goto done;
1354 		if (time_after(send_entry->timetosend, jiffies)) {
1355 			if (cm_node->state != IRDMA_CM_STATE_OFFLOADED) {
1356 				if (nexttimeout > send_entry->timetosend ||
1357 				    !settimer) {
1358 					nexttimeout = send_entry->timetosend;
1359 					settimer = 1;
1360 				}
1361 			} else {
1362 				irdma_free_retrans_entry(cm_node);
1363 			}
1364 			goto done;
1365 		}
1366 
1367 		if (cm_node->state == IRDMA_CM_STATE_OFFLOADED ||
1368 		    cm_node->state == IRDMA_CM_STATE_CLOSED) {
1369 			irdma_free_retrans_entry(cm_node);
1370 			goto done;
1371 		}
1372 
1373 		if (!send_entry->retranscount || !send_entry->retrycount) {
1374 			irdma_free_retrans_entry(cm_node);
1375 
1376 			spin_unlock_irqrestore(&cm_node->retrans_list_lock,
1377 					       flags);
1378 			irdma_retrans_expired(cm_node);
1379 			cm_node->state = IRDMA_CM_STATE_CLOSED;
1380 			spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
1381 			goto done;
1382 		}
1383 		spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
1384 
1385 		vsi = &cm_node->iwdev->vsi;
1386 		if (!cm_node->ack_rcvd) {
1387 			atomic_inc(&send_entry->sqbuf->refcount);
1388 			irdma_puda_send_buf(vsi->ilq, send_entry->sqbuf);
1389 			cm_node->cm_core->stats_pkt_retrans++;
1390 		}
1391 
1392 		spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
1393 		if (send_entry->send_retrans) {
1394 			send_entry->retranscount--;
1395 			timetosend = (IRDMA_RETRY_TIMEOUT <<
1396 				      (IRDMA_DEFAULT_RETRANS -
1397 				       send_entry->retranscount));
1398 
1399 			send_entry->timetosend = jiffies +
1400 			    min(timetosend, IRDMA_MAX_TIMEOUT);
1401 			if (nexttimeout > send_entry->timetosend || !settimer) {
1402 				nexttimeout = send_entry->timetosend;
1403 				settimer = 1;
1404 			}
1405 		} else {
1406 			int close_when_complete;
1407 
1408 			close_when_complete = send_entry->close_when_complete;
1409 			irdma_free_retrans_entry(cm_node);
1410 			if (close_when_complete)
1411 				irdma_rem_ref_cm_node(cm_node);
1412 		}
1413 done:
1414 		spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
1415 		irdma_rem_ref_cm_node(cm_node);
1416 	}
1417 
1418 	if (settimer) {
1419 		spin_lock_irqsave(&cm_core->ht_lock, flags);
1420 		if (!timer_pending(&cm_core->tcp_timer)) {
1421 			cm_core->tcp_timer.expires = nexttimeout;
1422 			add_timer(&cm_core->tcp_timer);
1423 		}
1424 		spin_unlock_irqrestore(&cm_core->ht_lock, flags);
1425 	}
1426 }
1427 
1428 /**
1429  * irdma_send_syn - send SYN packet
1430  * @cm_node: connection's node
1431  * @sendack: flag to set ACK bit or not
1432  */
1433 int
irdma_send_syn(struct irdma_cm_node * cm_node,u32 sendack)1434 irdma_send_syn(struct irdma_cm_node *cm_node, u32 sendack)
1435 {
1436 	struct irdma_puda_buf *sqbuf;
1437 	int flags = SET_SYN;
1438 	char optionsbuf[sizeof(struct option_mss) +
1439 			sizeof(struct option_windowscale) +
1440 			sizeof(struct option_base) + TCP_OPTIONS_PADDING];
1441 	struct irdma_kmem_info opts;
1442 	int optionssize = 0;
1443 
1444 	/* Sending MSS option */
1445 	union all_known_options *options;
1446 
1447 	opts.addr = optionsbuf;
1448 	if (!cm_node)
1449 		return -EINVAL;
1450 
1451 	options = (union all_known_options *)&optionsbuf[optionssize];
1452 	options->mss.optionnum = OPTION_NUM_MSS;
1453 	options->mss.len = sizeof(struct option_mss);
1454 	options->mss.mss = htons(cm_node->tcp_cntxt.mss);
1455 	optionssize += sizeof(struct option_mss);
1456 
1457 	options = (union all_known_options *)&optionsbuf[optionssize];
1458 	options->windowscale.optionnum = OPTION_NUM_WINDOW_SCALE;
1459 	options->windowscale.len = sizeof(struct option_windowscale);
1460 	options->windowscale.shiftcount = cm_node->tcp_cntxt.rcv_wscale;
1461 	optionssize += sizeof(struct option_windowscale);
1462 	options = (union all_known_options *)&optionsbuf[optionssize];
1463 	options->eol = OPTION_NUM_EOL;
1464 	optionssize += 1;
1465 
1466 	if (sendack)
1467 		flags |= SET_ACK;
1468 
1469 	opts.size = optionssize;
1470 
1471 	sqbuf = cm_node->cm_core->form_cm_frame(cm_node, &opts, NULL, NULL,
1472 						flags);
1473 	if (!sqbuf)
1474 		return -ENOMEM;
1475 
1476 	return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
1477 				       0);
1478 }
1479 
1480 /**
1481  * irdma_send_ack - Send ACK packet
1482  * @cm_node: connection's node
1483  */
1484 void
irdma_send_ack(struct irdma_cm_node * cm_node)1485 irdma_send_ack(struct irdma_cm_node *cm_node)
1486 {
1487 	struct irdma_puda_buf *sqbuf;
1488 	struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
1489 
1490 	sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL, NULL, NULL,
1491 						SET_ACK);
1492 	if (sqbuf)
1493 		irdma_puda_send_buf(vsi->ilq, sqbuf);
1494 }
1495 
1496 /**
1497  * irdma_send_fin - Send FIN pkt
1498  * @cm_node: connection's node
1499  */
1500 static int
irdma_send_fin(struct irdma_cm_node * cm_node)1501 irdma_send_fin(struct irdma_cm_node *cm_node)
1502 {
1503 	struct irdma_puda_buf *sqbuf;
1504 
1505 	sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL, NULL, NULL,
1506 						SET_ACK | SET_FIN);
1507 	if (!sqbuf)
1508 		return -ENOMEM;
1509 
1510 	return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
1511 				       0);
1512 }
1513 
1514 /**
1515  * irdma_find_listener - find a cm node listening on this addr-port pair
1516  * @cm_core: cm's core
1517  * @dst_addr: listener ip addr
1518  * @ipv4: flag indicating IPv4 when true
1519  * @dst_port: listener tcp port num
1520  * @vlan_id: virtual LAN ID
1521  * @listener_state: state to match with listen node's
1522  */
1523 static struct irdma_cm_listener *
irdma_find_listener(struct irdma_cm_core * cm_core,u32 * dst_addr,bool ipv4,u16 dst_port,u16 vlan_id,enum irdma_cm_listener_state listener_state)1524 irdma_find_listener(struct irdma_cm_core *cm_core, u32 *dst_addr, bool ipv4, u16 dst_port,
1525 		    u16 vlan_id, enum irdma_cm_listener_state listener_state)
1526 {
1527 	struct irdma_cm_listener *listen_node;
1528 	static const u32 ip_zero[4] = {0, 0, 0, 0};
1529 	u32 listen_addr[4];
1530 	u16 listen_port;
1531 	unsigned long flags;
1532 
1533 	/* walk list and find cm_node associated with this session ID */
1534 	spin_lock_irqsave(&cm_core->listen_list_lock, flags);
1535 	list_for_each_entry(listen_node, &cm_core->listen_list, list) {
1536 		memcpy(listen_addr, listen_node->loc_addr, sizeof(listen_addr));
1537 		listen_port = listen_node->loc_port;
1538 		if (listen_node->ipv4 != ipv4 || listen_port != dst_port ||
1539 		    !(listener_state & listen_node->listener_state))
1540 			continue;
1541 		/* compare node pair, return node handle if a match */
1542 		if (!memcmp(listen_addr, ip_zero, sizeof(listen_addr)) ||
1543 		    (!memcmp(listen_addr, dst_addr, sizeof(listen_addr)) &&
1544 		     vlan_id == listen_node->vlan_id)) {
1545 			atomic_inc(&listen_node->refcnt);
1546 			spin_unlock_irqrestore(&cm_core->listen_list_lock,
1547 					       flags);
1548 			return listen_node;
1549 		}
1550 	}
1551 	spin_unlock_irqrestore(&cm_core->listen_list_lock, flags);
1552 
1553 	return NULL;
1554 }
1555 
1556 /**
1557  * irdma_del_multiple_qhash - Remove qhash and child listens
1558  * @iwdev: iWarp device
1559  * @cm_info: CM info for parent listen node
1560  * @cm_parent_listen_node: The parent listen node
1561  */
1562 static int
irdma_del_multiple_qhash(struct irdma_device * iwdev,struct irdma_cm_info * cm_info,struct irdma_cm_listener * cm_parent_listen_node)1563 irdma_del_multiple_qhash(struct irdma_device *iwdev,
1564 			 struct irdma_cm_info *cm_info,
1565 			 struct irdma_cm_listener *cm_parent_listen_node)
1566 {
1567 	struct irdma_cm_listener *child_listen_node;
1568 	struct list_head *pos, *tpos;
1569 	unsigned long flags;
1570 	int ret = -EINVAL;
1571 
1572 	spin_lock_irqsave(&iwdev->cm_core.listen_list_lock, flags);
1573 	list_for_each_safe(pos, tpos,
1574 			   &cm_parent_listen_node->child_listen_list) {
1575 		child_listen_node = list_entry(pos, struct irdma_cm_listener,
1576 					       child_listen_list);
1577 		if (child_listen_node->ipv4)
1578 			irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1579 				    "removing child listen for IP=%x, port=%d, vlan=%d\n",
1580 				    child_listen_node->loc_addr[0],
1581 				    child_listen_node->loc_port,
1582 				    child_listen_node->vlan_id);
1583 		else
1584 			irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1585 				    "removing child listen for IP=%x:%x:%x:%x, port=%d, vlan=%d\n",
1586 				    IRDMA_PRINT_IP6(child_listen_node->loc_addr),
1587 				    child_listen_node->loc_port,
1588 				    child_listen_node->vlan_id);
1589 		list_del(pos);
1590 		memcpy(cm_info->loc_addr, child_listen_node->loc_addr,
1591 		       sizeof(cm_info->loc_addr));
1592 		cm_info->vlan_id = child_listen_node->vlan_id;
1593 		if (child_listen_node->qhash_set) {
1594 			ret = irdma_manage_qhash(iwdev, cm_info,
1595 						 IRDMA_QHASH_TYPE_TCP_SYN,
1596 						 IRDMA_QHASH_MANAGE_TYPE_DELETE,
1597 						 NULL, false);
1598 			child_listen_node->qhash_set = false;
1599 		} else {
1600 			ret = 0;
1601 		}
1602 		irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1603 			    "Child listen node freed = %p\n",
1604 			    child_listen_node);
1605 		kfree(child_listen_node);
1606 		cm_parent_listen_node->cm_core->stats_listen_nodes_destroyed++;
1607 	}
1608 	spin_unlock_irqrestore(&iwdev->cm_core.listen_list_lock, flags);
1609 
1610 	return ret;
1611 }
1612 
irdma_iw_get_vlan_prio(u32 * loc_addr,u8 prio,bool ipv4)1613 static u8 irdma_iw_get_vlan_prio(u32 *loc_addr, u8 prio, bool ipv4)
1614 {
1615 	return prio;
1616 }
1617 
1618 /**
1619  * irdma_get_vlan_mac_ipv6 - Get the vlan and mac for an IPv6
1620  * address
1621  * @addr: local IPv6 address
1622  * @vlan_id: vlan id for the given IPv6 address
1623  * @mac: mac address for the given IPv6 address
1624  *
1625  * Returns the net_device of the IPv6 address and also sets the
1626  * vlan id and mac for that address.
1627  */
1628 void
irdma_get_vlan_mac_ipv6(struct iw_cm_id * cm_id,u32 * addr,u16 * vlan_id,u8 * mac)1629 irdma_get_vlan_mac_ipv6(struct iw_cm_id *cm_id, u32 *addr, u16 *vlan_id, u8 *mac)
1630 {
1631 	if_t ip_dev = NULL;
1632 	struct in6_addr laddr6;
1633 	struct vnet *vnet = &init_net;
1634 	struct ifaddr *ifa;
1635 	u16 scope_id = 0;
1636 
1637 	irdma_copy_ip_htonl(laddr6.__u6_addr.__u6_addr32, addr);
1638 	if (vlan_id)
1639 		*vlan_id = 0xFFFF;	/* Match rdma_vlan_dev_vlan_id() */
1640 	if (mac)
1641 		eth_zero_addr(mac);
1642 
1643 	if (IN6_IS_SCOPE_LINKLOCAL(&laddr6) ||
1644 	    IN6_IS_ADDR_MC_INTFACELOCAL(&laddr6))
1645 		scope_id = ntohs(laddr6.__u6_addr.__u6_addr16[1]);
1646 
1647 #ifdef VIMAGE
1648 	vnet = irdma_cmid_to_vnet(cm_id);
1649 #endif
1650 	ip_dev = ip6_ifp_find(vnet, laddr6, scope_id);
1651 	if (ip_dev) {
1652 		if (vlan_id)
1653 			*vlan_id = rdma_vlan_dev_vlan_id(ip_dev);
1654 		ifa = if_getifaddr(ip_dev);
1655 		if (ifa && ifa->ifa_addr && mac)
1656 			ether_addr_copy(mac, if_getlladdr(ip_dev));
1657 	}
1658 }
1659 
1660 /**
1661  * irdma_get_vlan_ipv4 - Returns the vlan_id for IPv4 address
1662  * @addr: local IPv4 address
1663  */
1664 u16
irdma_get_vlan_ipv4(struct iw_cm_id * cm_id,u32 * addr)1665 irdma_get_vlan_ipv4(struct iw_cm_id *cm_id, u32 *addr)
1666 {
1667 	u16 vlan_id = 0xFFFF;
1668 
1669 #ifdef INET
1670 	if_t netdev;
1671 	struct vnet *vnet = &init_net;
1672 
1673 #ifdef VIMAGE
1674 	vnet = irdma_cmid_to_vnet(cm_id);
1675 #endif
1676 	netdev = ip_ifp_find(vnet, htonl(addr[0]));
1677 	if (netdev) {
1678 		vlan_id = rdma_vlan_dev_vlan_id(netdev);
1679 		dev_put(netdev);
1680 	}
1681 #endif
1682 
1683 	return vlan_id;
1684 }
1685 
1686 /**
1687  * irdma_add_mqh_ifa_cb - Adds multiple qhashes for IPv4/IPv6
1688  * @arg: Calback argument structure from irdma_add_mqh
1689  * @ifa: Current address to compute against
1690  * @count: Current cumulative output of all callbacks in this iteration
1691  *
1692  * Adds a qhash and a child listen node for a single IPv4/IPv6 address
1693  * on the adapter and adds the associated qhash filter
1694  */
1695 static u_int
irdma_add_mqh_ifa_cb(void * arg,struct ifaddr * ifa,u_int count)1696 irdma_add_mqh_ifa_cb(void *arg, struct ifaddr *ifa, u_int count)
1697 {
1698 	struct irdma_add_mqh_cbs *cbs = arg;
1699 	struct irdma_cm_listener *child_listen_node;
1700 	struct irdma_cm_info *cm_info = cbs->cm_info;
1701 	struct irdma_device *iwdev = cbs->iwdev;
1702 	struct irdma_cm_listener *cm_parent_listen_node = cbs->cm_listen_node;
1703 	if_t ip_dev = ifa->ifa_ifp;
1704 	unsigned long flags;
1705 	int ret;
1706 
1707 	if (count)
1708 		return 0;
1709 
1710 	child_listen_node = kzalloc(sizeof(*child_listen_node), GFP_ATOMIC);
1711 	if (!child_listen_node) {
1712 		irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1713 			    "listener memory allocation\n");
1714 		return -ENOMEM;
1715 	}
1716 
1717 	memcpy(child_listen_node, cm_parent_listen_node,
1718 	       sizeof(*child_listen_node));
1719 	cm_info->vlan_id = rdma_vlan_dev_vlan_id(ip_dev);
1720 	child_listen_node->vlan_id = cm_info->vlan_id;
1721 	if (cm_info->ipv4) {
1722 		irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1723 			    "Allocating child CM Listener forIP=%x, vlan_id=%d, MAC=%x:%x:%x:%x:%x:%x\n",
1724 			    ((struct sockaddr_in *)&ifa->ifa_addr)->sin_addr.s_addr,
1725 			    rdma_vlan_dev_vlan_id(ip_dev),
1726 			    if_getlladdr(ip_dev)[0], if_getlladdr(ip_dev)[1],
1727 			    if_getlladdr(ip_dev)[2], if_getlladdr(ip_dev)[3],
1728 			    if_getlladdr(ip_dev)[4], if_getlladdr(ip_dev)[5]);
1729 		child_listen_node->loc_addr[0] =
1730 		    ntohl(((struct sockaddr_in *)ifa->ifa_addr)->sin_addr.s_addr);
1731 	} else {
1732 		irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1733 			    "IP=%x:%x:%x:%x, vlan_id=%d, MAC=%x:%x:%x:%x:%x:%x\n",
1734 			    IRDMA_PRINT_IP6(&((struct sockaddr_in6 *)ifa->ifa_addr)->sin6_addr),
1735 			    rdma_vlan_dev_vlan_id(ip_dev),
1736 			    if_getlladdr(ip_dev)[0], if_getlladdr(ip_dev)[1],
1737 			    if_getlladdr(ip_dev)[2], if_getlladdr(ip_dev)[3],
1738 			    if_getlladdr(ip_dev)[4], if_getlladdr(ip_dev)[5]);
1739 		irdma_copy_ip_ntohl(child_listen_node->loc_addr,
1740 				    ((struct sockaddr_in6 *)ifa->ifa_addr)->sin6_addr.__u6_addr.__u6_addr32);
1741 	}
1742 	memcpy(cm_info->loc_addr, child_listen_node->loc_addr,
1743 	       sizeof(cm_info->loc_addr));
1744 	if (!iwdev->vsi.dscp_mode)
1745 		cm_info->user_pri =
1746 		    irdma_iw_get_vlan_prio(child_listen_node->loc_addr,
1747 					   cm_info->user_pri,
1748 					   cm_info->ipv4);
1749 	ret = irdma_add_qhash_wait_no_lock(iwdev, cm_info);
1750 	if (ret) {
1751 		kfree(child_listen_node);
1752 		return ret;
1753 	}
1754 
1755 	child_listen_node->qhash_set = true;
1756 	spin_lock_irqsave(&iwdev->cm_core.listen_list_lock, flags);
1757 	list_add(&child_listen_node->child_listen_list,
1758 		 &cm_parent_listen_node->child_listen_list);
1759 	spin_unlock_irqrestore(&iwdev->cm_core.listen_list_lock, flags);
1760 	cm_parent_listen_node->cm_core->stats_listen_nodes_created++;
1761 
1762 	return 0;
1763 }
1764 
1765 /**
1766  * irdma_add_mqh - Adds multiple qhashes
1767  * @iwdev: iWarp device
1768  * @cm_info: CM info for parent listen node
1769  * @cm_listen_node: The parent listen node
1770  */
1771 static int
irdma_add_mqh(struct irdma_device * iwdev,struct irdma_cm_info * cm_info,struct irdma_cm_listener * cm_listen_node)1772 irdma_add_mqh(struct irdma_device *iwdev,
1773 	      struct irdma_cm_info *cm_info,
1774 	      struct irdma_cm_listener *cm_listen_node)
1775 {
1776 	struct epoch_tracker et;
1777 	struct irdma_add_mqh_cbs cbs;
1778 	struct if_iter iter;
1779 	if_t ifp;
1780 	int err = -ENOENT;
1781 
1782 	cbs.iwdev = iwdev;
1783 	cbs.cm_info = cm_info;
1784 	cbs.cm_listen_node = cm_listen_node;
1785 
1786 	VNET_ITERATOR_DECL(vnet_iter);
1787 
1788 	VNET_LIST_RLOCK();
1789 	NET_EPOCH_ENTER(et);
1790 	VNET_FOREACH(vnet_iter) {
1791 		CURVNET_SET_QUIET(vnet_iter);
1792 		for (ifp = if_iter_start(&iter); ifp != NULL; ifp = if_iter_next(&iter)) {
1793 			if (!(if_getflags(ifp) & IFF_UP))
1794 				continue;
1795 
1796 			if (((rdma_vlan_dev_vlan_id(ifp) >= VLAN_N_VID) ||
1797 			     (rdma_vlan_dev_real_dev(ifp) != iwdev->netdev)) &&
1798 			    ifp != iwdev->netdev)
1799 				continue;
1800 
1801 			if_addr_rlock(ifp);
1802 			if (cm_info->ipv4)
1803 				err = if_foreach_addr_type(ifp, AF_INET, irdma_add_mqh_ifa_cb, &cbs);
1804 			else
1805 				err = if_foreach_addr_type(ifp, AF_INET6, irdma_add_mqh_ifa_cb, &cbs);
1806 			if_addr_runlock(ifp);
1807 		}
1808 		if_iter_finish(&iter);
1809 		CURVNET_RESTORE();
1810 	}
1811 	NET_EPOCH_EXIT(et);
1812 	VNET_LIST_RUNLOCK();
1813 
1814 	return err;
1815 }
1816 
1817 /**
1818  * irdma_reset_list_prep - add connection nodes slated for reset to list
1819  * @cm_core: cm's core
1820  * @listener: pointer to listener node
1821  * @reset_list: a list to which cm_node will be selected
1822  */
1823 static void
irdma_reset_list_prep(struct irdma_cm_core * cm_core,struct irdma_cm_listener * listener,struct list_head * reset_list)1824 irdma_reset_list_prep(struct irdma_cm_core *cm_core,
1825 		      struct irdma_cm_listener *listener,
1826 		      struct list_head *reset_list)
1827 {
1828 	struct irdma_cm_node *cm_node;
1829 	int bkt;
1830 
1831 	HASH_FOR_EACH_RCU(cm_core->cm_hash_tbl, bkt, cm_node, list) {
1832 		if (cm_node->listener == listener &&
1833 		    !cm_node->accelerated &&
1834 		    atomic_inc_not_zero(&cm_node->refcnt))
1835 			list_add(&cm_node->reset_entry, reset_list);
1836 	}
1837 }
1838 
1839 /**
1840  * irdma_dec_refcnt_listen - delete listener and associated cm nodes
1841  * @cm_core: cm's core
1842  * @listener: pointer to listener node
1843  * @free_hanging_nodes: to free associated cm_nodes
1844  * @apbvt_del: flag to delete the apbvt
1845  */
1846 static int
irdma_dec_refcnt_listen(struct irdma_cm_core * cm_core,struct irdma_cm_listener * listener,int free_hanging_nodes,bool apbvt_del)1847 irdma_dec_refcnt_listen(struct irdma_cm_core *cm_core,
1848 			struct irdma_cm_listener *listener,
1849 			int free_hanging_nodes, bool apbvt_del)
1850 {
1851 	struct list_head *list_pos;
1852 	struct list_head *list_temp;
1853 	struct irdma_cm_node *cm_node;
1854 	struct list_head reset_list;
1855 	struct irdma_cm_info nfo;
1856 	enum irdma_cm_node_state old_state;
1857 	unsigned long flags;
1858 	int err;
1859 
1860 	/* free non-accelerated child nodes for this listener */
1861 	INIT_LIST_HEAD(&reset_list);
1862 	if (free_hanging_nodes) {
1863 		rcu_read_lock();
1864 		irdma_reset_list_prep(cm_core, listener, &reset_list);
1865 		rcu_read_unlock();
1866 	}
1867 
1868 	list_for_each_safe(list_pos, list_temp, &reset_list) {
1869 		cm_node = container_of(list_pos, struct irdma_cm_node,
1870 				       reset_entry);
1871 		if (cm_node->state >= IRDMA_CM_STATE_FIN_WAIT1) {
1872 			irdma_rem_ref_cm_node(cm_node);
1873 			continue;
1874 		}
1875 
1876 		irdma_cleanup_retrans_entry(cm_node);
1877 		err = irdma_send_reset(cm_node);
1878 		if (err) {
1879 			cm_node->state = IRDMA_CM_STATE_CLOSED;
1880 			irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1881 				    "send reset failed\n");
1882 		} else {
1883 			old_state = cm_node->state;
1884 			cm_node->state = IRDMA_CM_STATE_LISTENER_DESTROYED;
1885 			if (old_state != IRDMA_CM_STATE_MPAREQ_RCVD)
1886 				irdma_rem_ref_cm_node(cm_node);
1887 		}
1888 	}
1889 
1890 	if (atomic_dec_and_test(&listener->refcnt)) {
1891 		spin_lock_irqsave(&cm_core->listen_list_lock, flags);
1892 		list_del(&listener->list);
1893 		spin_unlock_irqrestore(&cm_core->listen_list_lock, flags);
1894 
1895 		if (apbvt_del)
1896 			irdma_del_apbvt(listener->iwdev,
1897 					listener->apbvt_entry);
1898 		memcpy(nfo.loc_addr, listener->loc_addr, sizeof(nfo.loc_addr));
1899 		nfo.loc_port = listener->loc_port;
1900 		nfo.ipv4 = listener->ipv4;
1901 		nfo.vlan_id = listener->vlan_id;
1902 		nfo.user_pri = listener->user_pri;
1903 		nfo.qh_qpid = listener->iwdev->vsi.ilq->qp_id;
1904 
1905 		if (!list_empty(&listener->child_listen_list)) {
1906 			irdma_del_multiple_qhash(listener->iwdev, &nfo,
1907 						 listener);
1908 		} else {
1909 			if (listener->qhash_set)
1910 				irdma_manage_qhash(listener->iwdev,
1911 						   &nfo,
1912 						   IRDMA_QHASH_TYPE_TCP_SYN,
1913 						   IRDMA_QHASH_MANAGE_TYPE_DELETE,
1914 						   NULL, false);
1915 		}
1916 
1917 		cm_core->stats_listen_destroyed++;
1918 		cm_core->stats_listen_nodes_destroyed++;
1919 		irdma_debug(&listener->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1920 			    "loc_port=0x%04x loc_addr=%x cm_listen_node=%p cm_id=%p qhash_set=%d vlan_id=%d apbvt_del=%d\n",
1921 			    listener->loc_port, listener->loc_addr[0], listener,
1922 			    listener->cm_id, listener->qhash_set,
1923 			    listener->vlan_id, apbvt_del);
1924 		kfree(listener);
1925 		listener = NULL;
1926 		return 0;
1927 	}
1928 
1929 	return -EINVAL;
1930 }
1931 
1932 /**
1933  * irdma_cm_del_listen - delete a listener
1934  * @cm_core: cm's core
1935  * @listener: passive connection's listener
1936  * @apbvt_del: flag to delete apbvt
1937  */
1938 static int
irdma_cm_del_listen(struct irdma_cm_core * cm_core,struct irdma_cm_listener * listener,bool apbvt_del)1939 irdma_cm_del_listen(struct irdma_cm_core *cm_core,
1940 		    struct irdma_cm_listener *listener,
1941 		    bool apbvt_del)
1942 {
1943 	listener->listener_state = IRDMA_CM_LISTENER_PASSIVE_STATE;
1944 	listener->cm_id = NULL;
1945 
1946 	return irdma_dec_refcnt_listen(cm_core, listener, 1, apbvt_del);
1947 }
1948 
1949 /**
1950  * irdma_find_node - find a cm node that matches the reference cm node
1951  * @cm_core: cm's core
1952  * @rem_port: remote tcp port num
1953  * @rem_addr: remote ip addr
1954  * @loc_port: local tcp port num
1955  * @loc_addr: local ip addr
1956  * @vlan_id: local VLAN ID
1957  */
1958 struct irdma_cm_node *
irdma_find_node(struct irdma_cm_core * cm_core,u16 rem_port,u32 * rem_addr,u16 loc_port,u32 * loc_addr,u16 vlan_id)1959 irdma_find_node(struct irdma_cm_core *cm_core,
1960 		u16 rem_port, u32 *rem_addr, u16 loc_port,
1961 		u32 *loc_addr, u16 vlan_id)
1962 {
1963 	struct irdma_cm_node *cm_node;
1964 	u32 key = (rem_port << 16) | loc_port;
1965 
1966 	rcu_read_lock();
1967 	HASH_FOR_EACH_POSSIBLE_RCU(cm_core->cm_hash_tbl, cm_node, list, key) {
1968 		if (cm_node->vlan_id == vlan_id &&
1969 		    cm_node->loc_port == loc_port && cm_node->rem_port == rem_port &&
1970 		    !memcmp(cm_node->loc_addr, loc_addr, sizeof(cm_node->loc_addr)) &&
1971 		    !memcmp(cm_node->rem_addr, rem_addr, sizeof(cm_node->rem_addr))) {
1972 			if (!atomic_inc_not_zero(&cm_node->refcnt))
1973 				goto exit;
1974 			rcu_read_unlock();
1975 			return cm_node;
1976 		}
1977 	}
1978 
1979 exit:
1980 	rcu_read_unlock();
1981 
1982 	/* no owner node */
1983 	return NULL;
1984 }
1985 
1986 /**
1987  * irdma_add_hte_node - add a cm node to the hash table
1988  * @cm_core: cm's core
1989  * @cm_node: connection's node
1990  */
1991 static void
irdma_add_hte_node(struct irdma_cm_core * cm_core,struct irdma_cm_node * cm_node)1992 irdma_add_hte_node(struct irdma_cm_core *cm_core,
1993 		   struct irdma_cm_node *cm_node)
1994 {
1995 	unsigned long flags;
1996 	u32 key = (cm_node->rem_port << 16) | cm_node->loc_port;
1997 
1998 	spin_lock_irqsave(&cm_core->ht_lock, flags);
1999 	HASH_ADD_RCU(cm_core->cm_hash_tbl, &cm_node->list, key);
2000 	spin_unlock_irqrestore(&cm_core->ht_lock, flags);
2001 }
2002 
2003 /**
2004  * irdma_ipv4_is_lpb - check if loopback
2005  * @loc_addr: local addr to compare
2006  * @rem_addr: remote address
2007  */
2008 bool
irdma_ipv4_is_lpb(u32 loc_addr,u32 rem_addr)2009 irdma_ipv4_is_lpb(u32 loc_addr, u32 rem_addr)
2010 {
2011 	return ipv4_is_loopback(htonl(rem_addr)) || (loc_addr == rem_addr);
2012 }
2013 
2014 /**
2015  * irdma_ipv6_is_lpb - check if loopback
2016  * @loc_addr: local addr to compare
2017  * @rem_addr: remote address
2018  */
2019 bool
irdma_ipv6_is_lpb(u32 * loc_addr,u32 * rem_addr)2020 irdma_ipv6_is_lpb(u32 *loc_addr, u32 *rem_addr)
2021 {
2022 	struct in6_addr raddr6;
2023 
2024 	irdma_copy_ip_htonl(raddr6.__u6_addr.__u6_addr32, rem_addr);
2025 
2026 	return !memcmp(loc_addr, rem_addr, 16) || ipv6_addr_loopback(&raddr6);
2027 }
2028 
2029 /**
2030  * irdma_cm_create_ah - create a cm address handle
2031  * @cm_node: The connection manager node to create AH for
2032  * @wait: Provides option to wait for ah creation or not
2033  */
2034 static int
irdma_cm_create_ah(struct irdma_cm_node * cm_node,bool wait)2035 irdma_cm_create_ah(struct irdma_cm_node *cm_node, bool wait)
2036 {
2037 	struct irdma_ah_info ah_info = {0};
2038 	struct irdma_device *iwdev = cm_node->iwdev;
2039 #ifdef VIMAGE
2040 	struct vnet *vnet = irdma_cmid_to_vnet(cm_node->cm_id);
2041 #endif
2042 
2043 	ether_addr_copy(ah_info.mac_addr, if_getlladdr(iwdev->netdev));
2044 
2045 	ah_info.hop_ttl = 0x40;
2046 	ah_info.tc_tos = cm_node->tos;
2047 	ah_info.vsi = &iwdev->vsi;
2048 
2049 	if (cm_node->ipv4) {
2050 		ah_info.ipv4_valid = true;
2051 		ah_info.dest_ip_addr[0] = cm_node->rem_addr[0];
2052 		ah_info.src_ip_addr[0] = cm_node->loc_addr[0];
2053 		CURVNET_SET_QUIET(vnet);
2054 		ah_info.do_lpbk = irdma_ipv4_is_lpb(ah_info.src_ip_addr[0],
2055 						    ah_info.dest_ip_addr[0]);
2056 		CURVNET_RESTORE();
2057 	} else {
2058 		memcpy(ah_info.dest_ip_addr, cm_node->rem_addr,
2059 		       sizeof(ah_info.dest_ip_addr));
2060 		memcpy(ah_info.src_ip_addr, cm_node->loc_addr,
2061 		       sizeof(ah_info.src_ip_addr));
2062 		ah_info.do_lpbk = irdma_ipv6_is_lpb(ah_info.src_ip_addr,
2063 						    ah_info.dest_ip_addr);
2064 	}
2065 
2066 	ah_info.vlan_tag = cm_node->vlan_id;
2067 	if (cm_node->vlan_id < VLAN_N_VID) {
2068 		ah_info.insert_vlan_tag = 1;
2069 		ah_info.vlan_tag |= cm_node->user_pri << VLAN_PRIO_SHIFT;
2070 	}
2071 
2072 	ah_info.dst_arpindex =
2073 	    irdma_arp_table(iwdev->rf, ah_info.dest_ip_addr,
2074 			    NULL, IRDMA_ARP_RESOLVE);
2075 
2076 	if (irdma_puda_create_ah(&iwdev->rf->sc_dev, &ah_info, wait,
2077 				 IRDMA_PUDA_RSRC_TYPE_ILQ, cm_node,
2078 				 &cm_node->ah))
2079 		return -ENOMEM;
2080 
2081 	return 0;
2082 }
2083 
2084 /**
2085  * irdma_cm_free_ah - free a cm address handle
2086  * @cm_node: The connection manager node to create AH for
2087  */
2088 static void
irdma_cm_free_ah(struct irdma_cm_node * cm_node)2089 irdma_cm_free_ah(struct irdma_cm_node *cm_node)
2090 {
2091 	struct irdma_device *iwdev = cm_node->iwdev;
2092 
2093 	irdma_puda_free_ah(&iwdev->rf->sc_dev, cm_node->ah);
2094 	cm_node->ah = NULL;
2095 }
2096 
2097 /**
2098  * irdma_make_cm_node - create a new instance of a cm node
2099  * @cm_core: cm's core
2100  * @iwdev: iwarp device structure
2101  * @cm_info: quad info for connection
2102  * @listener: passive connection's listener
2103  */
2104 static struct irdma_cm_node *
irdma_make_cm_node(struct irdma_cm_core * cm_core,struct irdma_device * iwdev,struct irdma_cm_info * cm_info,struct irdma_cm_listener * listener)2105 irdma_make_cm_node(struct irdma_cm_core *cm_core, struct irdma_device *iwdev,
2106 		   struct irdma_cm_info *cm_info,
2107 		   struct irdma_cm_listener *listener)
2108 {
2109 	struct irdma_cm_node *cm_node;
2110 	int arpindex;
2111 	if_t netdev = iwdev->netdev;
2112 
2113 	/* create an hte and cm_node for this instance */
2114 	cm_node = kzalloc(sizeof(*cm_node), GFP_ATOMIC);
2115 	if (!cm_node)
2116 		return NULL;
2117 
2118 	/* set our node specific transport info */
2119 	cm_node->ipv4 = cm_info->ipv4;
2120 	cm_node->vlan_id = cm_info->vlan_id;
2121 	if (cm_node->vlan_id >= VLAN_N_VID && iwdev->dcb_vlan_mode)
2122 		cm_node->vlan_id = 0;
2123 	cm_node->tos = cm_info->tos;
2124 	cm_node->user_pri = cm_info->user_pri;
2125 	if (listener) {
2126 		if (listener->tos != cm_info->tos)
2127 			irdma_dev_warn(&iwdev->ibdev,
2128 				       "application TOS[%d] and remote client TOS[%d] mismatch\n",
2129 				       listener->tos, cm_info->tos);
2130 		if (iwdev->vsi.dscp_mode) {
2131 			cm_node->user_pri = listener->user_pri;
2132 		} else {
2133 			cm_node->tos = max(listener->tos, cm_info->tos);
2134 			cm_node->user_pri = rt_tos2priority(cm_node->tos);
2135 			cm_node->user_pri =
2136 			    irdma_iw_get_vlan_prio(cm_info->loc_addr,
2137 						   cm_node->user_pri,
2138 						   cm_info->ipv4);
2139 		}
2140 		irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_DCB,
2141 			    "listener: TOS:[%d] UP:[%d]\n", cm_node->tos,
2142 			    cm_node->user_pri);
2143 	}
2144 	memcpy(cm_node->loc_addr, cm_info->loc_addr, sizeof(cm_node->loc_addr));
2145 	memcpy(cm_node->rem_addr, cm_info->rem_addr, sizeof(cm_node->rem_addr));
2146 	cm_node->loc_port = cm_info->loc_port;
2147 	cm_node->rem_port = cm_info->rem_port;
2148 
2149 	cm_node->mpa_frame_rev = IRDMA_CM_DEFAULT_MPA_VER;
2150 	cm_node->send_rdma0_op = SEND_RDMA_READ_ZERO;
2151 	cm_node->iwdev = iwdev;
2152 	cm_node->dev = &iwdev->rf->sc_dev;
2153 
2154 	cm_node->ird_size = cm_node->dev->hw_attrs.max_hw_ird;
2155 	cm_node->ord_size = cm_node->dev->hw_attrs.max_hw_ord;
2156 
2157 	cm_node->listener = listener;
2158 	cm_node->cm_id = cm_info->cm_id;
2159 	ether_addr_copy(cm_node->loc_mac, if_getlladdr(netdev));
2160 	spin_lock_init(&cm_node->retrans_list_lock);
2161 	cm_node->ack_rcvd = false;
2162 
2163 	init_completion(&cm_node->establish_comp);
2164 	atomic_set(&cm_node->refcnt, 1);
2165 	/* associate our parent CM core */
2166 	cm_node->cm_core = cm_core;
2167 	cm_node->tcp_cntxt.loc_id = IRDMA_CM_DEFAULT_LOCAL_ID;
2168 	cm_node->tcp_cntxt.rcv_wscale = iwdev->rcv_wscale;
2169 	cm_node->tcp_cntxt.rcv_wnd = iwdev->rcv_wnd >> cm_node->tcp_cntxt.rcv_wscale;
2170 	kc_set_loc_seq_num_mss(cm_node);
2171 
2172 	arpindex = irdma_resolve_neigh_lpb_chk(iwdev, cm_node, cm_info);
2173 	if (arpindex < 0)
2174 		goto err;
2175 
2176 	ether_addr_copy(cm_node->rem_mac, iwdev->rf->arp_table[arpindex].mac_addr);
2177 	irdma_add_hte_node(cm_core, cm_node);
2178 	cm_core->stats_nodes_created++;
2179 	return cm_node;
2180 
2181 err:
2182 	kfree(cm_node);
2183 
2184 	return NULL;
2185 }
2186 
2187 static void
irdma_destroy_connection(struct irdma_cm_node * cm_node)2188 irdma_destroy_connection(struct irdma_cm_node *cm_node)
2189 {
2190 	struct irdma_cm_core *cm_core = cm_node->cm_core;
2191 	struct irdma_qp *iwqp;
2192 	struct irdma_cm_info nfo;
2193 
2194 	/* if the node is destroyed before connection was accelerated */
2195 	if (!cm_node->accelerated && cm_node->accept_pend) {
2196 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2197 			    "node destroyed before established\n");
2198 		atomic_dec(&cm_node->listener->pend_accepts_cnt);
2199 	}
2200 	if (cm_node->close_entry)
2201 		irdma_handle_close_entry(cm_node, 0);
2202 	if (cm_node->listener) {
2203 		irdma_dec_refcnt_listen(cm_core, cm_node->listener, 0, true);
2204 	} else {
2205 		if (cm_node->apbvt_set) {
2206 			irdma_del_apbvt(cm_node->iwdev, cm_node->apbvt_entry);
2207 			cm_node->apbvt_set = 0;
2208 		}
2209 		irdma_get_addr_info(cm_node, &nfo);
2210 		if (cm_node->qhash_set) {
2211 			nfo.qh_qpid = cm_node->iwdev->vsi.ilq->qp_id;
2212 			irdma_manage_qhash(cm_node->iwdev, &nfo,
2213 					   IRDMA_QHASH_TYPE_TCP_ESTABLISHED,
2214 					   IRDMA_QHASH_MANAGE_TYPE_DELETE, NULL,
2215 					   false);
2216 			cm_node->qhash_set = 0;
2217 		}
2218 	}
2219 
2220 	iwqp = cm_node->iwqp;
2221 	if (iwqp) {
2222 		cm_node->cm_id->rem_ref(cm_node->cm_id);
2223 		cm_node->cm_id = NULL;
2224 		iwqp->cm_id = NULL;
2225 		irdma_qp_rem_ref(&iwqp->ibqp);
2226 		cm_node->iwqp = NULL;
2227 	} else if (cm_node->qhash_set) {
2228 		irdma_get_addr_info(cm_node, &nfo);
2229 		nfo.qh_qpid = cm_node->iwdev->vsi.ilq->qp_id;
2230 		irdma_manage_qhash(cm_node->iwdev, &nfo,
2231 				   IRDMA_QHASH_TYPE_TCP_ESTABLISHED,
2232 				   IRDMA_QHASH_MANAGE_TYPE_DELETE, NULL, false);
2233 		cm_node->qhash_set = 0;
2234 	}
2235 
2236 	cm_core->cm_free_ah(cm_node);
2237 }
2238 
2239 /**
2240  * irdma_rem_ref_cm_node - destroy an instance of a cm node
2241  * @cm_node: connection's node
2242  */
2243 void
irdma_rem_ref_cm_node(struct irdma_cm_node * cm_node)2244 irdma_rem_ref_cm_node(struct irdma_cm_node *cm_node)
2245 {
2246 	struct irdma_cm_core *cm_core = cm_node->cm_core;
2247 	unsigned long flags;
2248 
2249 	spin_lock_irqsave(&cm_core->ht_lock, flags);
2250 
2251 	if (!atomic_dec_and_test(&cm_node->refcnt)) {
2252 		spin_unlock_irqrestore(&cm_core->ht_lock, flags);
2253 		return;
2254 	}
2255 	if (cm_node->iwqp) {
2256 		cm_node->iwqp->cm_node = NULL;
2257 		cm_node->iwqp->cm_id = NULL;
2258 	}
2259 	HASH_DEL_RCU(cm_core->cm_hash_tbl, &cm_node->list);
2260 	cm_node->cm_core->stats_nodes_destroyed++;
2261 
2262 	spin_unlock_irqrestore(&cm_core->ht_lock, flags);
2263 
2264 	irdma_destroy_connection(cm_node);
2265 
2266 	kfree_rcu(cm_node, rcu_head);
2267 }
2268 
2269 /**
2270  * irdma_handle_fin_pkt - FIN packet received
2271  * @cm_node: connection's node
2272  */
2273 static void
irdma_handle_fin_pkt(struct irdma_cm_node * cm_node)2274 irdma_handle_fin_pkt(struct irdma_cm_node *cm_node)
2275 {
2276 	switch (cm_node->state) {
2277 	case IRDMA_CM_STATE_SYN_RCVD:
2278 	case IRDMA_CM_STATE_SYN_SENT:
2279 	case IRDMA_CM_STATE_ESTABLISHED:
2280 	case IRDMA_CM_STATE_MPAREJ_RCVD:
2281 		cm_node->tcp_cntxt.rcv_nxt++;
2282 		irdma_cleanup_retrans_entry(cm_node);
2283 		cm_node->state = IRDMA_CM_STATE_LAST_ACK;
2284 		irdma_send_fin(cm_node);
2285 		break;
2286 	case IRDMA_CM_STATE_MPAREQ_SENT:
2287 		irdma_create_event(cm_node, IRDMA_CM_EVENT_ABORTED);
2288 		cm_node->tcp_cntxt.rcv_nxt++;
2289 		irdma_cleanup_retrans_entry(cm_node);
2290 		cm_node->state = IRDMA_CM_STATE_CLOSED;
2291 		atomic_inc(&cm_node->refcnt);
2292 		irdma_send_reset(cm_node);
2293 		break;
2294 	case IRDMA_CM_STATE_FIN_WAIT1:
2295 		cm_node->tcp_cntxt.rcv_nxt++;
2296 		irdma_cleanup_retrans_entry(cm_node);
2297 		cm_node->state = IRDMA_CM_STATE_CLOSING;
2298 		irdma_send_ack(cm_node);
2299 		/*
2300 		 * Wait for ACK as this is simultaneous close. After we receive ACK, do not send anything. Just rm the
2301 		 * node.
2302 		 */
2303 		break;
2304 	case IRDMA_CM_STATE_FIN_WAIT2:
2305 		cm_node->tcp_cntxt.rcv_nxt++;
2306 		irdma_cleanup_retrans_entry(cm_node);
2307 		cm_node->state = IRDMA_CM_STATE_TIME_WAIT;
2308 		irdma_send_ack(cm_node);
2309 		irdma_schedule_cm_timer(cm_node, NULL, IRDMA_TIMER_TYPE_CLOSE,
2310 					1, 0);
2311 		break;
2312 	case IRDMA_CM_STATE_TIME_WAIT:
2313 		cm_node->tcp_cntxt.rcv_nxt++;
2314 		irdma_cleanup_retrans_entry(cm_node);
2315 		cm_node->state = IRDMA_CM_STATE_CLOSED;
2316 		irdma_rem_ref_cm_node(cm_node);
2317 		break;
2318 	case IRDMA_CM_STATE_OFFLOADED:
2319 	default:
2320 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2321 			    "bad state node state = %d\n", cm_node->state);
2322 		break;
2323 	}
2324 }
2325 
2326 /**
2327  * irdma_handle_rst_pkt - process received RST packet
2328  * @cm_node: connection's node
2329  * @rbuf: receive buffer
2330  */
2331 static void
irdma_handle_rst_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2332 irdma_handle_rst_pkt(struct irdma_cm_node *cm_node,
2333 		     struct irdma_puda_buf *rbuf)
2334 {
2335 	irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2336 		    "caller: %pS cm_node=%p state=%d rem_port=0x%04x loc_port=0x%04x rem_addr=%x loc_addr=%x\n",
2337 		    __builtin_return_address(0), cm_node, cm_node->state,
2338 		    cm_node->rem_port, cm_node->loc_port, cm_node->rem_addr[0],
2339 		    cm_node->loc_addr[0]);
2340 
2341 	irdma_cleanup_retrans_entry(cm_node);
2342 	switch (cm_node->state) {
2343 	case IRDMA_CM_STATE_SYN_SENT:
2344 	case IRDMA_CM_STATE_MPAREQ_SENT:
2345 		switch (cm_node->mpa_frame_rev) {
2346 		case IETF_MPA_V2:
2347 			/* Drop down to MPA_V1 */
2348 			cm_node->mpa_frame_rev = IETF_MPA_V1;
2349 			/* send a syn and goto syn sent state */
2350 			cm_node->state = IRDMA_CM_STATE_SYN_SENT;
2351 			if (irdma_send_syn(cm_node, 0))
2352 				irdma_active_open_err(cm_node, false);
2353 			break;
2354 		case IETF_MPA_V1:
2355 		default:
2356 			irdma_active_open_err(cm_node, false);
2357 			break;
2358 		}
2359 		break;
2360 	case IRDMA_CM_STATE_MPAREQ_RCVD:
2361 		atomic_inc(&cm_node->passive_state);
2362 		break;
2363 	case IRDMA_CM_STATE_ESTABLISHED:
2364 	case IRDMA_CM_STATE_SYN_RCVD:
2365 	case IRDMA_CM_STATE_LISTENING:
2366 		irdma_passive_open_err(cm_node, false);
2367 		break;
2368 	case IRDMA_CM_STATE_OFFLOADED:
2369 		irdma_active_open_err(cm_node, false);
2370 		break;
2371 	case IRDMA_CM_STATE_CLOSED:
2372 		break;
2373 	case IRDMA_CM_STATE_FIN_WAIT2:
2374 	case IRDMA_CM_STATE_FIN_WAIT1:
2375 	case IRDMA_CM_STATE_LAST_ACK:
2376 	case IRDMA_CM_STATE_TIME_WAIT:
2377 		cm_node->state = IRDMA_CM_STATE_CLOSED;
2378 		irdma_rem_ref_cm_node(cm_node);
2379 		break;
2380 	default:
2381 		break;
2382 	}
2383 }
2384 
2385 /**
2386  * irdma_handle_rcv_mpa - Process a recv'd mpa buffer
2387  * @cm_node: connection's node
2388  * @rbuf: receive buffer
2389  */
2390 static void
irdma_handle_rcv_mpa(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2391 irdma_handle_rcv_mpa(struct irdma_cm_node *cm_node,
2392 		     struct irdma_puda_buf *rbuf)
2393 {
2394 	int err;
2395 	int datasize = rbuf->datalen;
2396 	u8 *dataloc = rbuf->data;
2397 
2398 	enum irdma_cm_event_type type = IRDMA_CM_EVENT_UNKNOWN;
2399 	u32 res_type;
2400 
2401 	err = irdma_parse_mpa(cm_node, dataloc, &res_type, datasize);
2402 	if (err) {
2403 		if (cm_node->state == IRDMA_CM_STATE_MPAREQ_SENT)
2404 			irdma_active_open_err(cm_node, true);
2405 		else
2406 			irdma_passive_open_err(cm_node, true);
2407 		return;
2408 	}
2409 
2410 	switch (cm_node->state) {
2411 	case IRDMA_CM_STATE_ESTABLISHED:
2412 		if (res_type == IRDMA_MPA_REQUEST_REJECT)
2413 			irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2414 				    "state for reject\n");
2415 		cm_node->state = IRDMA_CM_STATE_MPAREQ_RCVD;
2416 		type = IRDMA_CM_EVENT_MPA_REQ;
2417 		irdma_send_ack(cm_node);	/* ACK received MPA request */
2418 		atomic_set(&cm_node->passive_state,
2419 			   IRDMA_PASSIVE_STATE_INDICATED);
2420 		break;
2421 	case IRDMA_CM_STATE_MPAREQ_SENT:
2422 		irdma_cleanup_retrans_entry(cm_node);
2423 		if (res_type == IRDMA_MPA_REQUEST_REJECT) {
2424 			type = IRDMA_CM_EVENT_MPA_REJECT;
2425 			cm_node->state = IRDMA_CM_STATE_MPAREJ_RCVD;
2426 		} else {
2427 			type = IRDMA_CM_EVENT_CONNECTED;
2428 			cm_node->state = IRDMA_CM_STATE_OFFLOADED;
2429 		}
2430 		irdma_send_ack(cm_node);
2431 		break;
2432 	default:
2433 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2434 			    "wrong cm_node state=%d\n", cm_node->state);
2435 		break;
2436 	}
2437 	irdma_create_event(cm_node, type);
2438 }
2439 
2440 /**
2441  * irdma_check_syn - Check for error on received syn ack
2442  * @cm_node: connection's node
2443  * @tcph: pointer tcp header
2444  */
2445 static int
irdma_check_syn(struct irdma_cm_node * cm_node,struct tcphdr * tcph)2446 irdma_check_syn(struct irdma_cm_node *cm_node, struct tcphdr *tcph)
2447 {
2448 	if (ntohl(tcph->th_ack) != cm_node->tcp_cntxt.loc_seq_num) {
2449 		irdma_active_open_err(cm_node, true);
2450 		return 1;
2451 	}
2452 
2453 	return 0;
2454 }
2455 
2456 /**
2457  * irdma_check_seq - check seq numbers if OK
2458  * @cm_node: connection's node
2459  * @tcph: pointer tcp header
2460  */
2461 static int
irdma_check_seq(struct irdma_cm_node * cm_node,struct tcphdr * tcph)2462 irdma_check_seq(struct irdma_cm_node *cm_node, struct tcphdr *tcph)
2463 {
2464 	u32 seq;
2465 	u32 ack_seq;
2466 	u32 loc_seq_num = cm_node->tcp_cntxt.loc_seq_num;
2467 	u32 rcv_nxt = cm_node->tcp_cntxt.rcv_nxt;
2468 	u32 rcv_wnd;
2469 	int err = 0;
2470 
2471 	seq = ntohl(tcph->th_seq);
2472 	ack_seq = ntohl(tcph->th_ack);
2473 	rcv_wnd = cm_node->tcp_cntxt.rcv_wnd;
2474 	if (ack_seq != loc_seq_num ||
2475 	    !between(seq, rcv_nxt, (rcv_nxt + rcv_wnd)))
2476 		err = -1;
2477 	if (err)
2478 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2479 			    "seq number err\n");
2480 
2481 	return err;
2482 }
2483 
2484 void
irdma_add_conn_est_qh(struct irdma_cm_node * cm_node)2485 irdma_add_conn_est_qh(struct irdma_cm_node *cm_node)
2486 {
2487 	struct irdma_cm_info nfo;
2488 
2489 	irdma_get_addr_info(cm_node, &nfo);
2490 	nfo.qh_qpid = cm_node->iwdev->vsi.ilq->qp_id;
2491 	irdma_manage_qhash(cm_node->iwdev, &nfo,
2492 			   IRDMA_QHASH_TYPE_TCP_ESTABLISHED,
2493 			   IRDMA_QHASH_MANAGE_TYPE_ADD,
2494 			   cm_node, false);
2495 	cm_node->qhash_set = true;
2496 }
2497 
2498 /**
2499  * irdma_handle_syn_pkt - is for Passive node
2500  * @cm_node: connection's node
2501  * @rbuf: receive buffer
2502  */
2503 static void
irdma_handle_syn_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2504 irdma_handle_syn_pkt(struct irdma_cm_node *cm_node,
2505 		     struct irdma_puda_buf *rbuf)
2506 {
2507 	struct tcphdr *tcph = (struct tcphdr *)rbuf->tcph;
2508 	int err;
2509 	u32 inc_sequence;
2510 	int optionsize;
2511 
2512 	optionsize = (tcph->th_off << 2) - sizeof(*tcph);
2513 	inc_sequence = ntohl(tcph->th_seq);
2514 
2515 	switch (cm_node->state) {
2516 	case IRDMA_CM_STATE_SYN_SENT:
2517 	case IRDMA_CM_STATE_MPAREQ_SENT:
2518 		/* Rcvd syn on active open connection */
2519 		irdma_active_open_err(cm_node, 1);
2520 		break;
2521 	case IRDMA_CM_STATE_LISTENING:
2522 		/* Passive OPEN */
2523 		if (atomic_read(&cm_node->listener->pend_accepts_cnt) >
2524 		    cm_node->listener->backlog) {
2525 			cm_node->cm_core->stats_backlog_drops++;
2526 			irdma_passive_open_err(cm_node, false);
2527 			break;
2528 		}
2529 		err = irdma_handle_tcp_options(cm_node, tcph, optionsize, 1);
2530 		if (err) {
2531 			irdma_passive_open_err(cm_node, false);
2532 			/* drop pkt */
2533 			break;
2534 		}
2535 		err = cm_node->cm_core->cm_create_ah(cm_node, false);
2536 		if (err) {
2537 			irdma_passive_open_err(cm_node, false);
2538 			/* drop pkt */
2539 			break;
2540 		}
2541 		cm_node->tcp_cntxt.rcv_nxt = inc_sequence + 1;
2542 		cm_node->accept_pend = 1;
2543 		atomic_inc(&cm_node->listener->pend_accepts_cnt);
2544 
2545 		cm_node->state = IRDMA_CM_STATE_SYN_RCVD;
2546 		break;
2547 	case IRDMA_CM_STATE_CLOSED:
2548 		irdma_cleanup_retrans_entry(cm_node);
2549 		atomic_inc(&cm_node->refcnt);
2550 		irdma_send_reset(cm_node);
2551 		break;
2552 	case IRDMA_CM_STATE_OFFLOADED:
2553 	case IRDMA_CM_STATE_ESTABLISHED:
2554 	case IRDMA_CM_STATE_FIN_WAIT1:
2555 	case IRDMA_CM_STATE_FIN_WAIT2:
2556 	case IRDMA_CM_STATE_MPAREQ_RCVD:
2557 	case IRDMA_CM_STATE_LAST_ACK:
2558 	case IRDMA_CM_STATE_CLOSING:
2559 	case IRDMA_CM_STATE_UNKNOWN:
2560 	default:
2561 		break;
2562 	}
2563 }
2564 
2565 /**
2566  * irdma_handle_synack_pkt - Process SYN+ACK packet (active side)
2567  * @cm_node: connection's node
2568  * @rbuf: receive buffer
2569  */
2570 static void
irdma_handle_synack_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2571 irdma_handle_synack_pkt(struct irdma_cm_node *cm_node,
2572 			struct irdma_puda_buf *rbuf)
2573 {
2574 	struct tcphdr *tcph = (struct tcphdr *)rbuf->tcph;
2575 	int err;
2576 	u32 inc_sequence;
2577 	int optionsize;
2578 
2579 	optionsize = (tcph->th_off << 2) - sizeof(*tcph);
2580 	inc_sequence = ntohl(tcph->th_seq);
2581 	switch (cm_node->state) {
2582 	case IRDMA_CM_STATE_SYN_SENT:
2583 		irdma_cleanup_retrans_entry(cm_node);
2584 		/* active open */
2585 		if (irdma_check_syn(cm_node, tcph)) {
2586 			irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2587 				    "check syn fail\n");
2588 			return;
2589 		}
2590 		cm_node->tcp_cntxt.rem_ack_num = ntohl(tcph->th_ack);
2591 		/* setup options */
2592 		err = irdma_handle_tcp_options(cm_node, tcph, optionsize, 0);
2593 		if (err) {
2594 			irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2595 				    "cm_node=%p tcp_options failed\n", cm_node);
2596 			break;
2597 		}
2598 		irdma_cleanup_retrans_entry(cm_node);
2599 		cm_node->tcp_cntxt.rcv_nxt = inc_sequence + 1;
2600 		irdma_send_ack(cm_node);	/* ACK  for the syn_ack */
2601 		err = irdma_send_mpa_request(cm_node);
2602 		if (err) {
2603 			irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2604 				    "cm_node=%p irdma_send_mpa_request failed\n",
2605 				    cm_node);
2606 			break;
2607 		}
2608 		cm_node->state = IRDMA_CM_STATE_MPAREQ_SENT;
2609 		break;
2610 	case IRDMA_CM_STATE_MPAREQ_RCVD:
2611 		irdma_passive_open_err(cm_node, true);
2612 		break;
2613 	case IRDMA_CM_STATE_LISTENING:
2614 		cm_node->tcp_cntxt.loc_seq_num = ntohl(tcph->th_ack);
2615 		irdma_cleanup_retrans_entry(cm_node);
2616 		cm_node->state = IRDMA_CM_STATE_CLOSED;
2617 		irdma_send_reset(cm_node);
2618 		break;
2619 	case IRDMA_CM_STATE_CLOSED:
2620 		cm_node->tcp_cntxt.loc_seq_num = ntohl(tcph->th_ack);
2621 		irdma_cleanup_retrans_entry(cm_node);
2622 		atomic_inc(&cm_node->refcnt);
2623 		irdma_send_reset(cm_node);
2624 		break;
2625 	case IRDMA_CM_STATE_ESTABLISHED:
2626 	case IRDMA_CM_STATE_FIN_WAIT1:
2627 	case IRDMA_CM_STATE_FIN_WAIT2:
2628 	case IRDMA_CM_STATE_LAST_ACK:
2629 	case IRDMA_CM_STATE_OFFLOADED:
2630 	case IRDMA_CM_STATE_CLOSING:
2631 	case IRDMA_CM_STATE_UNKNOWN:
2632 	case IRDMA_CM_STATE_MPAREQ_SENT:
2633 	default:
2634 		break;
2635 	}
2636 }
2637 
2638 /**
2639  * irdma_handle_ack_pkt - process packet with ACK
2640  * @cm_node: connection's node
2641  * @rbuf: receive buffer
2642  */
2643 static int
irdma_handle_ack_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2644 irdma_handle_ack_pkt(struct irdma_cm_node *cm_node,
2645 		     struct irdma_puda_buf *rbuf)
2646 {
2647 	struct tcphdr *tcph = (struct tcphdr *)rbuf->tcph;
2648 	u32 inc_sequence;
2649 	int ret;
2650 	int optionsize;
2651 	u32 datasize = rbuf->datalen;
2652 
2653 	optionsize = (tcph->th_off << 2) - sizeof(*tcph);
2654 
2655 	if (irdma_check_seq(cm_node, tcph))
2656 		return -EINVAL;
2657 
2658 	inc_sequence = ntohl(tcph->th_seq);
2659 	switch (cm_node->state) {
2660 	case IRDMA_CM_STATE_SYN_RCVD:
2661 		irdma_cleanup_retrans_entry(cm_node);
2662 		ret = irdma_handle_tcp_options(cm_node, tcph, optionsize, 1);
2663 		if (ret)
2664 			return ret;
2665 		cm_node->tcp_cntxt.rem_ack_num = ntohl(tcph->th_ack);
2666 		cm_node->state = IRDMA_CM_STATE_ESTABLISHED;
2667 		if (datasize) {
2668 			cm_node->tcp_cntxt.rcv_nxt = inc_sequence + datasize;
2669 			irdma_handle_rcv_mpa(cm_node, rbuf);
2670 		}
2671 		break;
2672 	case IRDMA_CM_STATE_ESTABLISHED:
2673 		irdma_cleanup_retrans_entry(cm_node);
2674 		if (datasize) {
2675 			cm_node->tcp_cntxt.rcv_nxt = inc_sequence + datasize;
2676 			irdma_handle_rcv_mpa(cm_node, rbuf);
2677 		}
2678 		break;
2679 	case IRDMA_CM_STATE_MPAREQ_SENT:
2680 		cm_node->tcp_cntxt.rem_ack_num = ntohl(tcph->th_ack);
2681 		if (datasize) {
2682 			cm_node->tcp_cntxt.rcv_nxt = inc_sequence + datasize;
2683 			cm_node->ack_rcvd = false;
2684 			irdma_handle_rcv_mpa(cm_node, rbuf);
2685 		} else {
2686 			cm_node->ack_rcvd = true;
2687 		}
2688 		break;
2689 	case IRDMA_CM_STATE_LISTENING:
2690 		irdma_cleanup_retrans_entry(cm_node);
2691 		cm_node->state = IRDMA_CM_STATE_CLOSED;
2692 		irdma_send_reset(cm_node);
2693 		break;
2694 	case IRDMA_CM_STATE_CLOSED:
2695 		irdma_cleanup_retrans_entry(cm_node);
2696 		atomic_inc(&cm_node->refcnt);
2697 		irdma_send_reset(cm_node);
2698 		break;
2699 	case IRDMA_CM_STATE_LAST_ACK:
2700 	case IRDMA_CM_STATE_CLOSING:
2701 		irdma_cleanup_retrans_entry(cm_node);
2702 		cm_node->state = IRDMA_CM_STATE_CLOSED;
2703 		irdma_rem_ref_cm_node(cm_node);
2704 		break;
2705 	case IRDMA_CM_STATE_FIN_WAIT1:
2706 		irdma_cleanup_retrans_entry(cm_node);
2707 		cm_node->state = IRDMA_CM_STATE_FIN_WAIT2;
2708 		break;
2709 	case IRDMA_CM_STATE_SYN_SENT:
2710 	case IRDMA_CM_STATE_FIN_WAIT2:
2711 	case IRDMA_CM_STATE_OFFLOADED:
2712 	case IRDMA_CM_STATE_MPAREQ_RCVD:
2713 	case IRDMA_CM_STATE_UNKNOWN:
2714 	default:
2715 		irdma_cleanup_retrans_entry(cm_node);
2716 		break;
2717 	}
2718 
2719 	return 0;
2720 }
2721 
2722 /**
2723  * irdma_process_pkt - process cm packet
2724  * @cm_node: connection's node
2725  * @rbuf: receive buffer
2726  */
2727 static void
irdma_process_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2728 irdma_process_pkt(struct irdma_cm_node *cm_node,
2729 		  struct irdma_puda_buf *rbuf)
2730 {
2731 	enum irdma_tcpip_pkt_type pkt_type = IRDMA_PKT_TYPE_UNKNOWN;
2732 	struct tcphdr *tcph = (struct tcphdr *)rbuf->tcph;
2733 	u32 fin_set = 0;
2734 	int err;
2735 
2736 	if (tcp_get_flags(tcph) & TH_RST) {
2737 		pkt_type = IRDMA_PKT_TYPE_RST;
2738 	} else if (tcp_get_flags(tcph) & TH_SYN) {
2739 		pkt_type = IRDMA_PKT_TYPE_SYN;
2740 		if (tcp_get_flags(tcph) & TH_ACK)
2741 			pkt_type = IRDMA_PKT_TYPE_SYNACK;
2742 	} else if (tcp_get_flags(tcph) & TH_ACK) {
2743 		pkt_type = IRDMA_PKT_TYPE_ACK;
2744 	}
2745 	if (tcp_get_flags(tcph) & TH_FIN)
2746 		fin_set = 1;
2747 
2748 	switch (pkt_type) {
2749 	case IRDMA_PKT_TYPE_SYN:
2750 		irdma_handle_syn_pkt(cm_node, rbuf);
2751 		break;
2752 	case IRDMA_PKT_TYPE_SYNACK:
2753 		irdma_handle_synack_pkt(cm_node, rbuf);
2754 		break;
2755 	case IRDMA_PKT_TYPE_ACK:
2756 		err = irdma_handle_ack_pkt(cm_node, rbuf);
2757 		if (fin_set && !err)
2758 			irdma_handle_fin_pkt(cm_node);
2759 		break;
2760 	case IRDMA_PKT_TYPE_RST:
2761 		irdma_handle_rst_pkt(cm_node, rbuf);
2762 		break;
2763 	default:
2764 		if (fin_set &&
2765 		    (!irdma_check_seq(cm_node, (struct tcphdr *)rbuf->tcph)))
2766 			irdma_handle_fin_pkt(cm_node);
2767 		break;
2768 	}
2769 }
2770 
2771 /**
2772  * irdma_make_listen_node - create a listen node with params
2773  * @cm_core: cm's core
2774  * @iwdev: iwarp device structure
2775  * @cm_info: quad info for connection
2776  */
2777 static struct irdma_cm_listener *
irdma_make_listen_node(struct irdma_cm_core * cm_core,struct irdma_device * iwdev,struct irdma_cm_info * cm_info)2778 irdma_make_listen_node(struct irdma_cm_core *cm_core,
2779 		       struct irdma_device *iwdev,
2780 		       struct irdma_cm_info *cm_info)
2781 {
2782 	struct irdma_cm_listener *listener;
2783 	unsigned long flags;
2784 
2785 	/* cannot have multiple matching listeners */
2786 	listener = irdma_find_listener(cm_core, cm_info->loc_addr, cm_info->ipv4,
2787 				       cm_info->loc_port, cm_info->vlan_id,
2788 				       IRDMA_CM_LISTENER_EITHER_STATE);
2789 	if (listener &&
2790 	    listener->listener_state == IRDMA_CM_LISTENER_ACTIVE_STATE) {
2791 		atomic_dec(&listener->refcnt);
2792 		return NULL;
2793 	}
2794 
2795 	if (!listener) {
2796 		/*
2797 		 * create a CM listen node 1/2 node to compare incoming traffic to
2798 		 */
2799 		listener = kzalloc(sizeof(*listener), GFP_KERNEL);
2800 		if (!listener)
2801 			return NULL;
2802 		cm_core->stats_listen_nodes_created++;
2803 		memcpy(listener->loc_addr, cm_info->loc_addr,
2804 		       sizeof(listener->loc_addr));
2805 		listener->loc_port = cm_info->loc_port;
2806 
2807 		INIT_LIST_HEAD(&listener->child_listen_list);
2808 
2809 		atomic_set(&listener->refcnt, 1);
2810 	} else {
2811 		listener->reused_node = 1;
2812 	}
2813 
2814 	listener->cm_id = cm_info->cm_id;
2815 	listener->ipv4 = cm_info->ipv4;
2816 	listener->vlan_id = cm_info->vlan_id;
2817 	atomic_set(&listener->pend_accepts_cnt, 0);
2818 	listener->cm_core = cm_core;
2819 	listener->iwdev = iwdev;
2820 
2821 	listener->backlog = cm_info->backlog;
2822 	listener->listener_state = IRDMA_CM_LISTENER_ACTIVE_STATE;
2823 
2824 	if (!listener->reused_node) {
2825 		spin_lock_irqsave(&cm_core->listen_list_lock, flags);
2826 		list_add(&listener->list, &cm_core->listen_list);
2827 		spin_unlock_irqrestore(&cm_core->listen_list_lock, flags);
2828 	}
2829 
2830 	return listener;
2831 }
2832 
2833 /**
2834  * irdma_create_cm_node - make a connection node with params
2835  * @cm_core: cm's core
2836  * @iwdev: iwarp device structure
2837  * @conn_param: connection parameters
2838  * @cm_info: quad info for connection
2839  * @caller_cm_node: pointer to cm_node structure to return
2840  */
2841 static int
irdma_create_cm_node(struct irdma_cm_core * cm_core,struct irdma_device * iwdev,struct iw_cm_conn_param * conn_param,struct irdma_cm_info * cm_info,struct irdma_cm_node ** caller_cm_node)2842 irdma_create_cm_node(struct irdma_cm_core *cm_core,
2843 		     struct irdma_device *iwdev,
2844 		     struct iw_cm_conn_param *conn_param,
2845 		     struct irdma_cm_info *cm_info,
2846 		     struct irdma_cm_node **caller_cm_node)
2847 {
2848 	struct irdma_cm_node *cm_node;
2849 	u16 private_data_len = conn_param->private_data_len;
2850 	const void *private_data = conn_param->private_data;
2851 
2852 	/* create a CM connection node */
2853 	cm_node = irdma_make_cm_node(cm_core, iwdev, cm_info, NULL);
2854 	if (!cm_node)
2855 		return -ENOMEM;
2856 
2857 	/* set our node side to client (active) side */
2858 	cm_node->tcp_cntxt.client = 1;
2859 	cm_node->tcp_cntxt.rcv_wscale = IRDMA_CM_DEFAULT_RCV_WND_SCALE;
2860 
2861 	irdma_record_ird_ord(cm_node, conn_param->ird, conn_param->ord);
2862 
2863 	cm_node->pdata.size = private_data_len;
2864 	cm_node->pdata.addr = cm_node->pdata_buf;
2865 
2866 	memcpy(cm_node->pdata_buf, private_data, private_data_len);
2867 	*caller_cm_node = cm_node;
2868 
2869 	return 0;
2870 }
2871 
2872 /**
2873  * irdma_cm_reject - reject and teardown a connection
2874  * @cm_node: connection's node
2875  * @pdata: ptr to private data for reject
2876  * @plen: size of private data
2877  */
2878 static int
irdma_cm_reject(struct irdma_cm_node * cm_node,const void * pdata,u8 plen)2879 irdma_cm_reject(struct irdma_cm_node *cm_node, const void *pdata,
2880 		u8 plen)
2881 {
2882 	int ret;
2883 	int passive_state;
2884 
2885 	if (cm_node->tcp_cntxt.client)
2886 		return 0;
2887 
2888 	irdma_cleanup_retrans_entry(cm_node);
2889 
2890 	passive_state = atomic_add_return(1, &cm_node->passive_state);
2891 	if (passive_state == IRDMA_SEND_RESET_EVENT) {
2892 		cm_node->state = IRDMA_CM_STATE_CLOSED;
2893 		irdma_rem_ref_cm_node(cm_node);
2894 		return 0;
2895 	}
2896 
2897 	if (cm_node->state == IRDMA_CM_STATE_LISTENER_DESTROYED) {
2898 		irdma_rem_ref_cm_node(cm_node);
2899 		return 0;
2900 	}
2901 
2902 	ret = irdma_send_mpa_reject(cm_node, pdata, plen);
2903 	if (!ret)
2904 		return 0;
2905 
2906 	cm_node->state = IRDMA_CM_STATE_CLOSED;
2907 	if (irdma_send_reset(cm_node))
2908 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2909 			    "send reset failed\n");
2910 
2911 	return ret;
2912 }
2913 
2914 /**
2915  * irdma_cm_close - close of cm connection
2916  * @cm_node: connection's node
2917  */
2918 static int
irdma_cm_close(struct irdma_cm_node * cm_node)2919 irdma_cm_close(struct irdma_cm_node *cm_node)
2920 {
2921 	switch (cm_node->state) {
2922 	case IRDMA_CM_STATE_SYN_RCVD:
2923 	case IRDMA_CM_STATE_SYN_SENT:
2924 	case IRDMA_CM_STATE_ONE_SIDE_ESTABLISHED:
2925 	case IRDMA_CM_STATE_ESTABLISHED:
2926 	case IRDMA_CM_STATE_ACCEPTING:
2927 	case IRDMA_CM_STATE_MPAREQ_SENT:
2928 	case IRDMA_CM_STATE_MPAREQ_RCVD:
2929 		irdma_cleanup_retrans_entry(cm_node);
2930 		irdma_send_reset(cm_node);
2931 		break;
2932 	case IRDMA_CM_STATE_CLOSE_WAIT:
2933 		cm_node->state = IRDMA_CM_STATE_LAST_ACK;
2934 		irdma_send_fin(cm_node);
2935 		break;
2936 	case IRDMA_CM_STATE_FIN_WAIT1:
2937 	case IRDMA_CM_STATE_FIN_WAIT2:
2938 	case IRDMA_CM_STATE_LAST_ACK:
2939 	case IRDMA_CM_STATE_TIME_WAIT:
2940 	case IRDMA_CM_STATE_CLOSING:
2941 		return -EINVAL;
2942 	case IRDMA_CM_STATE_LISTENING:
2943 		irdma_cleanup_retrans_entry(cm_node);
2944 		irdma_send_reset(cm_node);
2945 		break;
2946 	case IRDMA_CM_STATE_MPAREJ_RCVD:
2947 	case IRDMA_CM_STATE_UNKNOWN:
2948 	case IRDMA_CM_STATE_INITED:
2949 	case IRDMA_CM_STATE_CLOSED:
2950 	case IRDMA_CM_STATE_LISTENER_DESTROYED:
2951 		irdma_rem_ref_cm_node(cm_node);
2952 		break;
2953 	case IRDMA_CM_STATE_OFFLOADED:
2954 		if (cm_node->send_entry)
2955 			irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2956 				    "CM send_entry in OFFLOADED state\n");
2957 		irdma_rem_ref_cm_node(cm_node);
2958 		break;
2959 	}
2960 
2961 	return 0;
2962 }
2963 
2964 /**
2965  * irdma_receive_ilq - recv an ETHERNET packet, and process it
2966  * through CM
2967  * @vsi: VSI structure of dev
2968  * @rbuf: receive buffer
2969  */
2970 void
irdma_receive_ilq(struct irdma_sc_vsi * vsi,struct irdma_puda_buf * rbuf)2971 irdma_receive_ilq(struct irdma_sc_vsi *vsi, struct irdma_puda_buf *rbuf)
2972 {
2973 	struct irdma_cm_node *cm_node;
2974 	struct irdma_cm_listener *listener;
2975 	struct ip *iph;
2976 	struct ip6_hdr *ip6h;
2977 	struct tcphdr *tcph;
2978 	struct irdma_cm_info cm_info = {0};
2979 	struct irdma_device *iwdev = vsi->back_vsi;
2980 	struct irdma_cm_core *cm_core = &iwdev->cm_core;
2981 	struct ether_vlan_header *ethh;
2982 	u16 vtag;
2983 
2984 	/* if vlan, then maclen = 18 else 14 */
2985 	iph = (struct ip *)rbuf->iph;
2986 	irdma_debug_buf(vsi->dev, IRDMA_DEBUG_ILQ, "RECEIVE ILQ BUFFER",
2987 			rbuf->mem.va, rbuf->totallen);
2988 	if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) {
2989 		if (rbuf->vlan_valid) {
2990 			vtag = rbuf->vlan_id;
2991 			cm_info.user_pri = (vtag & EVL_PRI_MASK) >>
2992 			    VLAN_PRIO_SHIFT;
2993 			cm_info.vlan_id = vtag & EVL_VLID_MASK;
2994 		} else {
2995 			cm_info.vlan_id = 0xFFFF;
2996 		}
2997 	} else {
2998 		ethh = rbuf->mem.va;
2999 
3000 		if (ethh->evl_proto == htons(ETH_P_8021Q)) {
3001 			vtag = ntohs(ethh->evl_tag);
3002 			cm_info.user_pri = (vtag & EVL_PRI_MASK) >>
3003 			    VLAN_PRIO_SHIFT;
3004 			cm_info.vlan_id = vtag & EVL_VLID_MASK;
3005 			irdma_debug(&cm_core->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3006 				    "vlan_id=%d\n", cm_info.vlan_id);
3007 		} else {
3008 			cm_info.vlan_id = 0xFFFF;
3009 		}
3010 	}
3011 	tcph = (struct tcphdr *)rbuf->tcph;
3012 
3013 	if (rbuf->ipv4) {
3014 		cm_info.loc_addr[0] = ntohl(iph->ip_dst.s_addr);
3015 		cm_info.rem_addr[0] = ntohl(iph->ip_src.s_addr);
3016 		cm_info.ipv4 = true;
3017 		cm_info.tos = iph->ip_tos;
3018 	} else {
3019 		ip6h = (struct ip6_hdr *)rbuf->iph;
3020 		irdma_copy_ip_ntohl(cm_info.loc_addr,
3021 				    ip6h->ip6_dst.__u6_addr.__u6_addr32);
3022 		irdma_copy_ip_ntohl(cm_info.rem_addr,
3023 				    ip6h->ip6_src.__u6_addr.__u6_addr32);
3024 		cm_info.ipv4 = false;
3025 		cm_info.tos = (ip6h->ip6_vfc << 4) | ip6h->ip6_flow;
3026 	}
3027 	cm_info.loc_port = ntohs(tcph->th_dport);
3028 	cm_info.rem_port = ntohs(tcph->th_sport);
3029 	cm_node = irdma_find_node(cm_core, cm_info.rem_port, cm_info.rem_addr,
3030 				  cm_info.loc_port, cm_info.loc_addr, cm_info.vlan_id);
3031 
3032 	if (!cm_node) {
3033 		/*
3034 		 * Only type of packet accepted are for the PASSIVE open (syn only)
3035 		 */
3036 		if (!(tcp_get_flags(tcph) & TH_SYN) || tcp_get_flags(tcph) & TH_ACK)
3037 			return;
3038 
3039 		listener = irdma_find_listener(cm_core,
3040 					       cm_info.loc_addr,
3041 					       cm_info.ipv4,
3042 					       cm_info.loc_port,
3043 					       cm_info.vlan_id,
3044 					       IRDMA_CM_LISTENER_ACTIVE_STATE);
3045 		if (!listener) {
3046 			cm_info.cm_id = NULL;
3047 			irdma_debug(&cm_core->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3048 				    "no listener found\n");
3049 			return;
3050 		}
3051 
3052 		cm_info.cm_id = listener->cm_id;
3053 		cm_node = irdma_make_cm_node(cm_core, iwdev, &cm_info,
3054 					     listener);
3055 		if (!cm_node) {
3056 			irdma_debug(&cm_core->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3057 				    "allocate node failed\n");
3058 			atomic_dec(&listener->refcnt);
3059 			return;
3060 		}
3061 
3062 		if (!(tcp_get_flags(tcph) & (TH_RST | TH_FIN))) {
3063 			cm_node->state = IRDMA_CM_STATE_LISTENING;
3064 		} else {
3065 			irdma_rem_ref_cm_node(cm_node);
3066 			return;
3067 		}
3068 
3069 		atomic_inc(&cm_node->refcnt);
3070 	} else if (cm_node->state == IRDMA_CM_STATE_OFFLOADED) {
3071 		irdma_rem_ref_cm_node(cm_node);
3072 		return;
3073 	}
3074 
3075 	irdma_process_pkt(cm_node, rbuf);
3076 	irdma_rem_ref_cm_node(cm_node);
3077 }
3078 
3079 static int
irdma_add_qh(struct irdma_cm_node * cm_node,bool active)3080 irdma_add_qh(struct irdma_cm_node *cm_node, bool active)
3081 {
3082 	if (!active)
3083 		irdma_add_conn_est_qh(cm_node);
3084 	return 0;
3085 }
3086 
3087 static void
irdma_cm_free_ah_nop(struct irdma_cm_node * cm_node)3088 irdma_cm_free_ah_nop(struct irdma_cm_node *cm_node)
3089 {
3090 }
3091 
3092 /**
3093  * irdma_setup_cm_core - setup top level instance of a cm core
3094  * @iwdev: iwarp device structure
3095  * @rdma_ver: HW version
3096  */
3097 int
irdma_setup_cm_core(struct irdma_device * iwdev,u8 rdma_ver)3098 irdma_setup_cm_core(struct irdma_device *iwdev, u8 rdma_ver)
3099 {
3100 	struct irdma_cm_core *cm_core = &iwdev->cm_core;
3101 
3102 	cm_core->iwdev = iwdev;
3103 	cm_core->dev = &iwdev->rf->sc_dev;
3104 
3105 	/* Handles CM event work items send to Iwarp core */
3106 	cm_core->event_wq = alloc_ordered_workqueue("iwarp-event-wq", 0);
3107 	if (!cm_core->event_wq)
3108 		return -ENOMEM;
3109 
3110 	INIT_LIST_HEAD(&cm_core->listen_list);
3111 
3112 	timer_setup(&cm_core->tcp_timer, irdma_cm_timer_tick, 0);
3113 
3114 	spin_lock_init(&cm_core->ht_lock);
3115 	spin_lock_init(&cm_core->listen_list_lock);
3116 	spin_lock_init(&cm_core->apbvt_lock);
3117 	switch (rdma_ver) {
3118 	case IRDMA_GEN_1:
3119 		cm_core->form_cm_frame = irdma_form_uda_cm_frame;
3120 		cm_core->cm_create_ah = irdma_add_qh;
3121 		cm_core->cm_free_ah = irdma_cm_free_ah_nop;
3122 		break;
3123 	case IRDMA_GEN_2:
3124 	default:
3125 		cm_core->form_cm_frame = irdma_form_ah_cm_frame;
3126 		cm_core->cm_create_ah = irdma_cm_create_ah;
3127 		cm_core->cm_free_ah = irdma_cm_free_ah;
3128 	}
3129 
3130 	return 0;
3131 }
3132 
3133 /**
3134  * irdma_cleanup_cm_core - deallocate a top level instance of a
3135  * cm core
3136  * @cm_core: cm's core
3137  */
3138 void
irdma_cleanup_cm_core(struct irdma_cm_core * cm_core)3139 irdma_cleanup_cm_core(struct irdma_cm_core *cm_core)
3140 {
3141 	if (!cm_core)
3142 		return;
3143 
3144 	del_timer_sync(&cm_core->tcp_timer);
3145 
3146 	destroy_workqueue(cm_core->event_wq);
3147 	cm_core->dev->ws_reset(&cm_core->iwdev->vsi);
3148 }
3149 
3150 /**
3151  * irdma_init_tcp_ctx - setup qp context
3152  * @cm_node: connection's node
3153  * @tcp_info: offload info for tcp
3154  * @iwqp: associate qp for the connection
3155  */
3156 static void
irdma_init_tcp_ctx(struct irdma_cm_node * cm_node,struct irdma_tcp_offload_info * tcp_info,struct irdma_qp * iwqp)3157 irdma_init_tcp_ctx(struct irdma_cm_node *cm_node,
3158 		   struct irdma_tcp_offload_info *tcp_info,
3159 		   struct irdma_qp *iwqp)
3160 {
3161 	tcp_info->ipv4 = cm_node->ipv4;
3162 	tcp_info->drop_ooo_seg = !iwqp->iwdev->iw_ooo;
3163 	tcp_info->wscale = true;
3164 	tcp_info->ignore_tcp_opt = true;
3165 	tcp_info->ignore_tcp_uns_opt = true;
3166 	tcp_info->no_nagle = false;
3167 
3168 	tcp_info->ttl = IRDMA_DEFAULT_TTL;
3169 	tcp_info->rtt_var = IRDMA_DEFAULT_RTT_VAR;
3170 	tcp_info->ss_thresh = IRDMA_DEFAULT_SS_THRESH;
3171 	tcp_info->rexmit_thresh = IRDMA_DEFAULT_REXMIT_THRESH;
3172 
3173 	tcp_info->tcp_state = IRDMA_TCP_STATE_ESTABLISHED;
3174 	tcp_info->snd_wscale = cm_node->tcp_cntxt.snd_wscale;
3175 	tcp_info->rcv_wscale = cm_node->tcp_cntxt.rcv_wscale;
3176 
3177 	tcp_info->snd_nxt = cm_node->tcp_cntxt.loc_seq_num;
3178 	tcp_info->snd_wnd = cm_node->tcp_cntxt.snd_wnd;
3179 	tcp_info->rcv_nxt = cm_node->tcp_cntxt.rcv_nxt;
3180 	tcp_info->snd_max = cm_node->tcp_cntxt.loc_seq_num;
3181 
3182 	tcp_info->snd_una = cm_node->tcp_cntxt.loc_seq_num;
3183 	tcp_info->cwnd = 2 * cm_node->tcp_cntxt.mss;
3184 	tcp_info->snd_wl1 = cm_node->tcp_cntxt.rcv_nxt;
3185 	tcp_info->snd_wl2 = cm_node->tcp_cntxt.loc_seq_num;
3186 	tcp_info->max_snd_window = cm_node->tcp_cntxt.max_snd_wnd;
3187 	tcp_info->rcv_wnd = cm_node->tcp_cntxt.rcv_wnd
3188 	    << cm_node->tcp_cntxt.rcv_wscale;
3189 
3190 	tcp_info->flow_label = 0;
3191 	tcp_info->snd_mss = (u32)cm_node->tcp_cntxt.mss;
3192 	tcp_info->tos = cm_node->tos;
3193 	if (cm_node->vlan_id < VLAN_N_VID) {
3194 		tcp_info->insert_vlan_tag = true;
3195 		tcp_info->vlan_tag = cm_node->vlan_id;
3196 		tcp_info->vlan_tag |= cm_node->user_pri << VLAN_PRIO_SHIFT;
3197 	}
3198 	tcp_info->src_port = cm_node->loc_port;
3199 	tcp_info->dst_port = cm_node->rem_port;
3200 	tcp_info->arp_idx = (u16)irdma_arp_table(iwqp->iwdev->rf,
3201 						  cm_node->rem_addr, NULL,
3202 						  IRDMA_ARP_RESOLVE);
3203 	if (cm_node->ipv4) {
3204 		tcp_info->dest_ip_addr[3] = cm_node->rem_addr[0];
3205 		tcp_info->local_ipaddr[3] = cm_node->loc_addr[0];
3206 	} else {
3207 		memcpy(tcp_info->dest_ip_addr, cm_node->rem_addr,
3208 		       sizeof(tcp_info->dest_ip_addr));
3209 		memcpy(tcp_info->local_ipaddr, cm_node->loc_addr,
3210 		       sizeof(tcp_info->local_ipaddr));
3211 	}
3212 }
3213 
3214 /**
3215  * irdma_cm_init_tsa_conn - setup qp for RTS
3216  * @iwqp: associate qp for the connection
3217  * @cm_node: connection's node
3218  */
3219 static void
irdma_cm_init_tsa_conn(struct irdma_qp * iwqp,struct irdma_cm_node * cm_node)3220 irdma_cm_init_tsa_conn(struct irdma_qp *iwqp,
3221 		       struct irdma_cm_node *cm_node)
3222 {
3223 	struct irdma_iwarp_offload_info *iwarp_info;
3224 	struct irdma_qp_host_ctx_info *ctx_info;
3225 
3226 	iwarp_info = &iwqp->iwarp_info;
3227 	ctx_info = &iwqp->ctx_info;
3228 
3229 	ctx_info->tcp_info = &iwqp->tcp_info;
3230 	ctx_info->send_cq_num = iwqp->iwscq->sc_cq.cq_uk.cq_id;
3231 	ctx_info->rcv_cq_num = iwqp->iwrcq->sc_cq.cq_uk.cq_id;
3232 
3233 	iwarp_info->ord_size = cm_node->ord_size;
3234 	iwarp_info->ird_size = cm_node->ird_size;
3235 	iwarp_info->rd_en = true;
3236 	iwarp_info->rdmap_ver = 1;
3237 	iwarp_info->ddp_ver = 1;
3238 	iwarp_info->pd_id = iwqp->iwpd->sc_pd.pd_id;
3239 
3240 	ctx_info->tcp_info_valid = true;
3241 	ctx_info->iwarp_info_valid = true;
3242 	ctx_info->user_pri = cm_node->user_pri;
3243 
3244 	irdma_init_tcp_ctx(cm_node, &iwqp->tcp_info, iwqp);
3245 	if (cm_node->snd_mark_en) {
3246 		iwarp_info->snd_mark_en = true;
3247 		iwarp_info->snd_mark_offset = (iwqp->tcp_info.snd_nxt & SNDMARKER_SEQNMASK) +
3248 		    cm_node->lsmm_size;
3249 	}
3250 
3251 	cm_node->state = IRDMA_CM_STATE_OFFLOADED;
3252 	iwqp->tcp_info.tcp_state = IRDMA_TCP_STATE_ESTABLISHED;
3253 	iwqp->tcp_info.src_mac_addr_idx = iwqp->iwdev->mac_ip_table_idx;
3254 
3255 	if (cm_node->rcv_mark_en) {
3256 		iwarp_info->rcv_mark_en = true;
3257 		iwarp_info->align_hdrs = true;
3258 	}
3259 
3260 	irdma_sc_qp_setctx(&iwqp->sc_qp, iwqp->host_ctx.va, ctx_info);
3261 
3262 	/* once tcp_info is set, no need to do it again */
3263 	ctx_info->tcp_info_valid = false;
3264 	ctx_info->iwarp_info_valid = false;
3265 }
3266 
3267 /**
3268  * irdma_cm_disconn - when a connection is being closed
3269  * @iwqp: associated qp for the connection
3270  */
3271 void
irdma_cm_disconn(struct irdma_qp * iwqp)3272 irdma_cm_disconn(struct irdma_qp *iwqp)
3273 {
3274 	struct irdma_device *iwdev = iwqp->iwdev;
3275 	struct disconn_work *work;
3276 	unsigned long flags;
3277 
3278 	work = kzalloc(sizeof(*work), GFP_ATOMIC);
3279 	if (!work)
3280 		return;
3281 
3282 	spin_lock_irqsave(&iwdev->rf->qptable_lock, flags);
3283 	if (!iwdev->rf->qp_table[iwqp->ibqp.qp_num]) {
3284 		spin_unlock_irqrestore(&iwdev->rf->qptable_lock, flags);
3285 		irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3286 			    "qp_id %d is already freed\n", iwqp->ibqp.qp_num);
3287 		kfree(work);
3288 		return;
3289 	}
3290 	irdma_qp_add_ref(&iwqp->ibqp);
3291 	spin_unlock_irqrestore(&iwdev->rf->qptable_lock, flags);
3292 
3293 	work->iwqp = iwqp;
3294 	INIT_WORK(&work->work, irdma_disconnect_worker);
3295 	queue_work(iwdev->cleanup_wq, &work->work);
3296 }
3297 
3298 /**
3299  * irdma_qp_disconnect - free qp and close cm
3300  * @iwqp: associate qp for the connection
3301  */
3302 static void
irdma_qp_disconnect(struct irdma_qp * iwqp)3303 irdma_qp_disconnect(struct irdma_qp *iwqp)
3304 {
3305 	struct irdma_device *iwdev = iwqp->iwdev;
3306 
3307 	iwqp->active_conn = 0;
3308 	/* close the CM node down if it is still active */
3309 	irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM, "Call close API\n");
3310 	irdma_cm_close(iwqp->cm_node);
3311 }
3312 
3313 /**
3314  * irdma_cm_disconn_true - called by worker thread to disconnect qp
3315  * @iwqp: associate qp for the connection
3316  */
3317 static void
irdma_cm_disconn_true(struct irdma_qp * iwqp)3318 irdma_cm_disconn_true(struct irdma_qp *iwqp)
3319 {
3320 	struct iw_cm_id *cm_id;
3321 	struct irdma_device *iwdev;
3322 	struct irdma_sc_qp *qp = &iwqp->sc_qp;
3323 	u16 last_ae;
3324 	u8 original_hw_tcp_state;
3325 	u8 original_ibqp_state;
3326 	int disconn_status = 0;
3327 	int issue_disconn = 0;
3328 	int issue_close = 0;
3329 	int issue_flush = 0;
3330 	unsigned long flags;
3331 	int err;
3332 
3333 	iwdev = iwqp->iwdev;
3334 	spin_lock_irqsave(&iwqp->lock, flags);
3335 	if (rdma_protocol_roce(&iwdev->ibdev, 1)) {
3336 		struct ib_qp_attr attr;
3337 
3338 		if (iwqp->flush_issued || iwqp->sc_qp.qp_uk.destroy_pending) {
3339 			spin_unlock_irqrestore(&iwqp->lock, flags);
3340 			return;
3341 		}
3342 
3343 		spin_unlock_irqrestore(&iwqp->lock, flags);
3344 
3345 		attr.qp_state = IB_QPS_ERR;
3346 		irdma_modify_qp_roce(&iwqp->ibqp, &attr, IB_QP_STATE, NULL);
3347 		irdma_ib_qp_event(iwqp, qp->event_type);
3348 		return;
3349 	}
3350 
3351 	cm_id = iwqp->cm_id;
3352 	original_hw_tcp_state = iwqp->hw_tcp_state;
3353 	original_ibqp_state = iwqp->ibqp_state;
3354 	last_ae = iwqp->last_aeq;
3355 
3356 	if (qp->term_flags) {
3357 		issue_disconn = 1;
3358 		issue_close = 1;
3359 		iwqp->cm_id = NULL;
3360 		irdma_terminate_del_timer(qp);
3361 		if (!iwqp->flush_issued) {
3362 			iwqp->flush_issued = 1;
3363 			issue_flush = 1;
3364 		}
3365 	} else if ((original_hw_tcp_state == IRDMA_TCP_STATE_CLOSE_WAIT) ||
3366 		   ((original_ibqp_state == IB_QPS_RTS) &&
3367 		    (last_ae == IRDMA_AE_LLP_CONNECTION_RESET))) {
3368 		issue_disconn = 1;
3369 		if (last_ae == IRDMA_AE_LLP_CONNECTION_RESET)
3370 			disconn_status = -ECONNRESET;
3371 	}
3372 
3373 	if (original_hw_tcp_state == IRDMA_TCP_STATE_CLOSED ||
3374 	    original_hw_tcp_state == IRDMA_TCP_STATE_TIME_WAIT ||
3375 	    last_ae == IRDMA_AE_RDMAP_ROE_BAD_LLP_CLOSE ||
3376 	    last_ae == IRDMA_AE_BAD_CLOSE ||
3377 	    last_ae == IRDMA_AE_LLP_CONNECTION_RESET || iwdev->rf->reset || !cm_id) {
3378 		issue_close = 1;
3379 		iwqp->cm_id = NULL;
3380 		qp->term_flags = 0;
3381 		if (!iwqp->flush_issued) {
3382 			iwqp->flush_issued = 1;
3383 			issue_flush = 1;
3384 		}
3385 	}
3386 
3387 	spin_unlock_irqrestore(&iwqp->lock, flags);
3388 	if (issue_flush && !iwqp->sc_qp.qp_uk.destroy_pending) {
3389 		irdma_flush_wqes(iwqp, IRDMA_FLUSH_SQ | IRDMA_FLUSH_RQ |
3390 				 IRDMA_FLUSH_WAIT);
3391 
3392 		if (qp->term_flags)
3393 			irdma_ib_qp_event(iwqp, qp->event_type);
3394 	}
3395 
3396 	if (!cm_id || !cm_id->event_handler)
3397 		return;
3398 
3399 	spin_lock_irqsave(&iwdev->cm_core.ht_lock, flags);
3400 	if (!iwqp->cm_node) {
3401 		spin_unlock_irqrestore(&iwdev->cm_core.ht_lock, flags);
3402 		return;
3403 	}
3404 	atomic_inc(&iwqp->cm_node->refcnt);
3405 
3406 	spin_unlock_irqrestore(&iwdev->cm_core.ht_lock, flags);
3407 
3408 	if (issue_disconn) {
3409 		err = irdma_send_cm_event(iwqp->cm_node, cm_id,
3410 					  IW_CM_EVENT_DISCONNECT,
3411 					  disconn_status);
3412 		if (err)
3413 			irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3414 				    "disconnect event failed: - cm_id = %p\n",
3415 				    cm_id);
3416 	}
3417 	if (issue_close) {
3418 		cm_id->provider_data = iwqp;
3419 		err = irdma_send_cm_event(iwqp->cm_node, cm_id,
3420 					  IW_CM_EVENT_CLOSE, 0);
3421 		if (err)
3422 			irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3423 				    "close event failed: - cm_id = %p\n",
3424 				    cm_id);
3425 		irdma_qp_disconnect(iwqp);
3426 	}
3427 	irdma_rem_ref_cm_node(iwqp->cm_node);
3428 }
3429 
3430 /**
3431  * irdma_disconnect_worker - worker for connection close
3432  * @work: points or disconn structure
3433  */
3434 static void
irdma_disconnect_worker(struct work_struct * work)3435 irdma_disconnect_worker(struct work_struct *work)
3436 {
3437 	struct disconn_work *dwork = container_of(work, struct disconn_work, work);
3438 	struct irdma_qp *iwqp = dwork->iwqp;
3439 
3440 	kfree(dwork);
3441 	irdma_cm_disconn_true(iwqp);
3442 	irdma_qp_rem_ref(&iwqp->ibqp);
3443 }
3444 
3445 /**
3446  * irdma_free_lsmm_rsrc - free lsmm memory and deregister
3447  * @iwqp: associate qp for the connection
3448  */
3449 void
irdma_free_lsmm_rsrc(struct irdma_qp * iwqp)3450 irdma_free_lsmm_rsrc(struct irdma_qp *iwqp)
3451 {
3452 	struct irdma_device *iwdev;
3453 
3454 	iwdev = iwqp->iwdev;
3455 
3456 	if (iwqp->ietf_mem.va) {
3457 		if (iwqp->lsmm_mr)
3458 			iwdev->ibdev.dereg_mr(iwqp->lsmm_mr, NULL);
3459 
3460 		irdma_free_dma_mem(iwdev->rf->sc_dev.hw,
3461 				   &iwqp->ietf_mem);
3462 		iwqp->ietf_mem.va = NULL;
3463 	}
3464 }
3465 
3466 /**
3467  * irdma_accept - registered call for connection to be accepted
3468  * @cm_id: cm information for passive connection
3469  * @conn_param: accpet parameters
3470  */
3471 int
irdma_accept(struct iw_cm_id * cm_id,struct iw_cm_conn_param * conn_param)3472 irdma_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param)
3473 {
3474 	struct ib_qp *ibqp;
3475 	struct irdma_qp *iwqp;
3476 	struct irdma_device *iwdev;
3477 	struct irdma_sc_dev *dev;
3478 	struct irdma_cm_node *cm_node;
3479 	struct ib_qp_attr attr = {0};
3480 	int passive_state;
3481 	struct ib_mr *ibmr;
3482 	struct irdma_pd *iwpd;
3483 	u16 buf_len = 0;
3484 	struct irdma_kmem_info accept;
3485 	u64 tagged_offset;
3486 	int wait_ret;
3487 	int ret = 0;
3488 
3489 	ibqp = irdma_get_qp(cm_id->device, conn_param->qpn);
3490 	if (!ibqp)
3491 		return -EINVAL;
3492 
3493 	iwqp = to_iwqp(ibqp);
3494 	iwdev = iwqp->iwdev;
3495 	dev = &iwdev->rf->sc_dev;
3496 	cm_node = cm_id->provider_data;
3497 
3498 	if (((struct sockaddr_in *)&cm_id->local_addr)->sin_family == AF_INET) {
3499 		cm_node->ipv4 = true;
3500 		cm_node->vlan_id = irdma_get_vlan_ipv4(cm_id, cm_node->loc_addr);
3501 	} else {
3502 		cm_node->ipv4 = false;
3503 		irdma_get_vlan_mac_ipv6(cm_id, cm_node->loc_addr, &cm_node->vlan_id,
3504 					NULL);
3505 	}
3506 	irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM, "Accept vlan_id=%d\n",
3507 		    cm_node->vlan_id);
3508 
3509 	if (cm_node->state == IRDMA_CM_STATE_LISTENER_DESTROYED) {
3510 		ret = -EINVAL;
3511 		goto error;
3512 	}
3513 
3514 	passive_state = atomic_add_return(1, &cm_node->passive_state);
3515 	if (passive_state == IRDMA_SEND_RESET_EVENT) {
3516 		ret = -ECONNRESET;
3517 		goto error;
3518 	}
3519 
3520 	buf_len = conn_param->private_data_len + IRDMA_MAX_IETF_SIZE;
3521 	iwqp->ietf_mem.size = buf_len;
3522 	iwqp->ietf_mem.va = irdma_allocate_dma_mem(dev->hw, &iwqp->ietf_mem,
3523 						   iwqp->ietf_mem.size, 1);
3524 	if (!iwqp->ietf_mem.va) {
3525 		ret = -ENOMEM;
3526 		goto error;
3527 	}
3528 
3529 	cm_node->pdata.size = conn_param->private_data_len;
3530 	accept.addr = iwqp->ietf_mem.va;
3531 	accept.size = irdma_cm_build_mpa_frame(cm_node, &accept, MPA_KEY_REPLY);
3532 	memcpy((u8 *)accept.addr + accept.size, conn_param->private_data,
3533 	       conn_param->private_data_len);
3534 
3535 	if (cm_node->dev->ws_add(iwqp->sc_qp.vsi, cm_node->user_pri)) {
3536 		ret = -ENOMEM;
3537 		goto error;
3538 	}
3539 	iwqp->sc_qp.user_pri = cm_node->user_pri;
3540 	irdma_qp_add_qos(&iwqp->sc_qp);
3541 	if (cm_node->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_2)
3542 		iwdev->rf->check_fc(&iwdev->vsi, &iwqp->sc_qp);
3543 	/* setup our first outgoing iWarp send WQE (the IETF frame response) */
3544 	iwpd = iwqp->iwpd;
3545 	tagged_offset = (uintptr_t)iwqp->ietf_mem.va;
3546 	ibmr = irdma_reg_phys_mr(&iwpd->ibpd, iwqp->ietf_mem.pa, buf_len,
3547 				 IB_ACCESS_LOCAL_WRITE, &tagged_offset);
3548 	if (IS_ERR(ibmr)) {
3549 		ret = -ENOMEM;
3550 		goto error;
3551 	}
3552 
3553 	ibmr->pd = &iwpd->ibpd;
3554 	ibmr->device = iwpd->ibpd.device;
3555 	iwqp->lsmm_mr = ibmr;
3556 	if (iwqp->page)
3557 		iwqp->sc_qp.qp_uk.sq_base = kmap_local_page(iwqp->page);
3558 
3559 	cm_node->lsmm_size = accept.size + conn_param->private_data_len;
3560 	irdma_sc_send_lsmm(&iwqp->sc_qp, iwqp->ietf_mem.va, cm_node->lsmm_size,
3561 			   ibmr->lkey);
3562 
3563 	if (iwqp->page)
3564 		kunmap_local(iwqp->sc_qp.qp_uk.sq_base);
3565 
3566 	iwqp->cm_id = cm_id;
3567 	cm_node->cm_id = cm_id;
3568 
3569 	cm_id->provider_data = iwqp;
3570 	iwqp->active_conn = 0;
3571 	iwqp->cm_node = cm_node;
3572 	cm_node->iwqp = iwqp;
3573 	irdma_cm_init_tsa_conn(iwqp, cm_node);
3574 	irdma_qp_add_ref(&iwqp->ibqp);
3575 	cm_id->add_ref(cm_id);
3576 
3577 	attr.qp_state = IB_QPS_RTS;
3578 	cm_node->qhash_set = false;
3579 	cm_node->cm_core->cm_free_ah(cm_node);
3580 
3581 	irdma_modify_qp(&iwqp->ibqp, &attr, IB_QP_STATE, NULL);
3582 	if (dev->hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE) {
3583 		wait_ret = wait_event_interruptible_timeout(iwqp->waitq,
3584 							    iwqp->rts_ae_rcvd,
3585 							    IRDMA_MAX_TIMEOUT);
3586 		if (!wait_ret) {
3587 			irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3588 				    "Slow Connection: cm_node=%p, loc_port=%d, rem_port=%d, cm_id=%p\n",
3589 				    cm_node, cm_node->loc_port,
3590 				    cm_node->rem_port, cm_node->cm_id);
3591 			ret = -ECONNRESET;
3592 			goto error;
3593 		}
3594 	}
3595 
3596 	irdma_send_cm_event(cm_node, cm_id, IW_CM_EVENT_ESTABLISHED, 0);
3597 	cm_node->accelerated = true;
3598 	complete(&cm_node->establish_comp);
3599 
3600 	if (cm_node->accept_pend) {
3601 		atomic_dec(&cm_node->listener->pend_accepts_cnt);
3602 		cm_node->accept_pend = 0;
3603 	}
3604 
3605 	irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3606 		    "rem_port=0x%04x, loc_port=0x%04x rem_addr=%x loc_addr=%x cm_node=%p cm_id=%p qp_id=%d\n\n",
3607 		    cm_node->rem_port, cm_node->loc_port, cm_node->rem_addr[0],
3608 		    cm_node->loc_addr[0], cm_node, cm_id, ibqp->qp_num);
3609 	cm_node->cm_core->stats_accepts++;
3610 
3611 	return 0;
3612 error:
3613 	irdma_free_lsmm_rsrc(iwqp);
3614 	irdma_rem_ref_cm_node(cm_node);
3615 
3616 	return ret;
3617 }
3618 
3619 /**
3620  * irdma_reject - registered call for connection to be rejected
3621  * @cm_id: cm information for passive connection
3622  * @pdata: private data to be sent
3623  * @pdata_len: private data length
3624  */
3625 int
irdma_reject(struct iw_cm_id * cm_id,const void * pdata,u8 pdata_len)3626 irdma_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len)
3627 {
3628 	struct irdma_device *iwdev;
3629 	struct irdma_cm_node *cm_node;
3630 
3631 	cm_node = cm_id->provider_data;
3632 	cm_node->pdata.size = pdata_len;
3633 
3634 	iwdev = to_iwdev(cm_id->device);
3635 	if (!iwdev)
3636 		return -EINVAL;
3637 
3638 	cm_node->cm_core->stats_rejects++;
3639 
3640 	if (pdata_len + sizeof(struct ietf_mpa_v2) > IRDMA_MAX_CM_BUF)
3641 		return -EINVAL;
3642 
3643 	return irdma_cm_reject(cm_node, pdata, pdata_len);
3644 }
3645 
3646 /**
3647  * irdma_connect - registered call for connection to be established
3648  * @cm_id: cm information for passive connection
3649  * @conn_param: Information about the connection
3650  */
3651 int
irdma_connect(struct iw_cm_id * cm_id,struct iw_cm_conn_param * conn_param)3652 irdma_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param)
3653 {
3654 	struct ib_qp *ibqp;
3655 	struct irdma_qp *iwqp;
3656 	struct irdma_device *iwdev;
3657 	struct irdma_cm_node *cm_node;
3658 	struct irdma_cm_info cm_info;
3659 	struct sockaddr_in *laddr;
3660 	struct sockaddr_in *raddr;
3661 	struct sockaddr_in6 *laddr6;
3662 	struct sockaddr_in6 *raddr6;
3663 	int ret = 0;
3664 
3665 	ibqp = irdma_get_qp(cm_id->device, conn_param->qpn);
3666 	if (!ibqp)
3667 		return -EINVAL;
3668 	iwqp = to_iwqp(ibqp);
3669 	if (!iwqp)
3670 		return -EINVAL;
3671 	iwdev = iwqp->iwdev;
3672 	if (!iwdev)
3673 		return -EINVAL;
3674 
3675 	laddr = (struct sockaddr_in *)&cm_id->m_local_addr;
3676 	raddr = (struct sockaddr_in *)&cm_id->m_remote_addr;
3677 	laddr6 = (struct sockaddr_in6 *)&cm_id->m_local_addr;
3678 	raddr6 = (struct sockaddr_in6 *)&cm_id->m_remote_addr;
3679 
3680 	if (!(laddr->sin_port) || !(raddr->sin_port))
3681 		return -EINVAL;
3682 
3683 	iwqp->active_conn = 1;
3684 	iwqp->cm_id = NULL;
3685 	cm_id->provider_data = iwqp;
3686 
3687 	/* set up the connection params for the node */
3688 	if (cm_id->remote_addr.ss_family == AF_INET) {
3689 		if (iwdev->vsi.mtu < IRDMA_MIN_MTU_IPV4)
3690 			return -EINVAL;
3691 
3692 		cm_info.ipv4 = true;
3693 		memset(cm_info.loc_addr, 0, sizeof(cm_info.loc_addr));
3694 		memset(cm_info.rem_addr, 0, sizeof(cm_info.rem_addr));
3695 		cm_info.loc_addr[0] = ntohl(laddr->sin_addr.s_addr);
3696 		cm_info.rem_addr[0] = ntohl(raddr->sin_addr.s_addr);
3697 		cm_info.loc_port = ntohs(laddr->sin_port);
3698 		cm_info.rem_port = ntohs(raddr->sin_port);
3699 		cm_info.vlan_id = irdma_get_vlan_ipv4(cm_id, cm_info.loc_addr);
3700 	} else {
3701 		if (iwdev->vsi.mtu < IRDMA_MIN_MTU_IPV6)
3702 			return -EINVAL;
3703 
3704 		cm_info.ipv4 = false;
3705 		irdma_copy_ip_ntohl(cm_info.loc_addr,
3706 				    laddr6->sin6_addr.__u6_addr.__u6_addr32);
3707 		irdma_copy_ip_ntohl(cm_info.rem_addr,
3708 				    raddr6->sin6_addr.__u6_addr.__u6_addr32);
3709 		cm_info.loc_port = ntohs(laddr6->sin6_port);
3710 		cm_info.rem_port = ntohs(raddr6->sin6_port);
3711 		irdma_get_vlan_mac_ipv6(cm_id, cm_info.loc_addr, &cm_info.vlan_id, NULL);
3712 	}
3713 	cm_info.cm_id = cm_id;
3714 	cm_info.qh_qpid = iwdev->vsi.ilq->qp_id;
3715 	cm_info.tos = cm_id->tos;
3716 	if (iwdev->vsi.dscp_mode) {
3717 		cm_info.user_pri =
3718 		    iwqp->sc_qp.vsi->dscp_map[irdma_tos2dscp(cm_info.tos)];
3719 	} else {
3720 		cm_info.user_pri = rt_tos2priority(cm_id->tos);
3721 		cm_info.user_pri =
3722 		    irdma_iw_get_vlan_prio(cm_info.loc_addr,
3723 					   cm_info.user_pri,
3724 					   cm_info.ipv4);
3725 	}
3726 
3727 	if (iwqp->sc_qp.dev->ws_add(iwqp->sc_qp.vsi, cm_info.user_pri))
3728 		return -ENOMEM;
3729 	iwqp->sc_qp.user_pri = cm_info.user_pri;
3730 	irdma_qp_add_qos(&iwqp->sc_qp);
3731 	if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_2)
3732 		iwdev->rf->check_fc(&iwdev->vsi, &iwqp->sc_qp);
3733 	irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_DCB, "TOS:[%d] UP:[%d]\n",
3734 		    cm_id->tos, cm_info.user_pri);
3735 
3736 	ret = irdma_create_cm_node(&iwdev->cm_core, iwdev, conn_param, &cm_info,
3737 				   &cm_node);
3738 	if (ret)
3739 		return ret;
3740 	ret = cm_node->cm_core->cm_create_ah(cm_node, true);
3741 	if (ret)
3742 		goto err;
3743 	if (irdma_manage_qhash(iwdev, &cm_info,
3744 			       IRDMA_QHASH_TYPE_TCP_ESTABLISHED,
3745 			       IRDMA_QHASH_MANAGE_TYPE_ADD, NULL, true)) {
3746 		ret = -EINVAL;
3747 		goto err;
3748 	}
3749 	cm_node->qhash_set = true;
3750 
3751 	cm_node->apbvt_entry = irdma_add_apbvt(iwdev, cm_info.loc_port);
3752 	if (!cm_node->apbvt_entry) {
3753 		ret = -EINVAL;
3754 		goto err;
3755 	}
3756 
3757 	cm_node->apbvt_set = true;
3758 	iwqp->cm_node = cm_node;
3759 	cm_node->iwqp = iwqp;
3760 	iwqp->cm_id = cm_id;
3761 	irdma_qp_add_ref(&iwqp->ibqp);
3762 	cm_id->add_ref(cm_id);
3763 
3764 	if (cm_node->state != IRDMA_CM_STATE_OFFLOADED) {
3765 		cm_node->state = IRDMA_CM_STATE_SYN_SENT;
3766 		ret = irdma_send_syn(cm_node, 0);
3767 		if (ret)
3768 			goto err;
3769 	}
3770 
3771 	irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3772 		    "rem_port=0x%04x, loc_port=0x%04x rem_addr=%x loc_addr=%x cm_node=%p cm_id=%p qp_id = %d\n\n",
3773 		    cm_node->rem_port, cm_node->loc_port, cm_node->rem_addr[0],
3774 		    cm_node->loc_addr[0], cm_node, cm_id, ibqp->qp_num);
3775 
3776 	return 0;
3777 
3778 err:
3779 	if (cm_info.ipv4)
3780 		irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3781 			    "connect() FAILED: dest addr=%x",
3782 			    cm_info.rem_addr[0]);
3783 	else
3784 		irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3785 			    "connect() FAILED: dest addr=%x:%x:%x:%x",
3786 			    IRDMA_PRINT_IP6(cm_info.rem_addr));
3787 	irdma_rem_ref_cm_node(cm_node);
3788 	iwdev->cm_core.stats_connect_errs++;
3789 
3790 	return ret;
3791 }
3792 
3793 /**
3794  * irdma_create_listen - registered call creating listener
3795  * @cm_id: cm information for passive connection
3796  * @backlog: to max accept pending count
3797  */
3798 int
irdma_create_listen(struct iw_cm_id * cm_id,int backlog)3799 irdma_create_listen(struct iw_cm_id *cm_id, int backlog)
3800 {
3801 	struct irdma_device *iwdev;
3802 	struct irdma_cm_listener *cm_listen_node;
3803 	struct irdma_cm_info cm_info = {0};
3804 	struct sockaddr_in *laddr;
3805 	struct sockaddr_in6 *laddr6;
3806 	bool wildcard = false;
3807 	int err;
3808 
3809 	iwdev = to_iwdev(cm_id->device);
3810 	if (!iwdev)
3811 		return -EINVAL;
3812 
3813 	laddr = (struct sockaddr_in *)&cm_id->m_local_addr;
3814 	laddr6 = (struct sockaddr_in6 *)&cm_id->m_local_addr;
3815 	cm_info.qh_qpid = iwdev->vsi.ilq->qp_id;
3816 
3817 	if (laddr->sin_family == AF_INET) {
3818 		if (iwdev->vsi.mtu < IRDMA_MIN_MTU_IPV4)
3819 			return -EINVAL;
3820 
3821 		cm_info.ipv4 = true;
3822 		cm_info.loc_addr[0] = ntohl(laddr->sin_addr.s_addr);
3823 		cm_info.loc_port = ntohs(laddr->sin_port);
3824 
3825 		if (laddr->sin_addr.s_addr != htonl(INADDR_ANY)) {
3826 			cm_info.vlan_id = irdma_get_vlan_ipv4(cm_id, cm_info.loc_addr);
3827 		} else {
3828 			cm_info.vlan_id = 0xFFFF;
3829 			wildcard = true;
3830 		}
3831 	} else {
3832 		if (iwdev->vsi.mtu < IRDMA_MIN_MTU_IPV6)
3833 			return -EINVAL;
3834 
3835 		cm_info.ipv4 = false;
3836 		irdma_copy_ip_ntohl(cm_info.loc_addr,
3837 				    laddr6->sin6_addr.__u6_addr.__u6_addr32);
3838 		cm_info.loc_port = ntohs(laddr6->sin6_port);
3839 		if (!IN6_IS_ADDR_UNSPECIFIED(&laddr6->sin6_addr)) {
3840 			irdma_get_vlan_mac_ipv6(cm_id, cm_info.loc_addr,
3841 						&cm_info.vlan_id, NULL);
3842 		} else {
3843 			cm_info.vlan_id = 0xFFFF;
3844 			wildcard = true;
3845 		}
3846 	}
3847 
3848 	if (cm_info.vlan_id >= VLAN_N_VID && iwdev->dcb_vlan_mode)
3849 		cm_info.vlan_id = 0;
3850 	cm_info.backlog = backlog;
3851 	cm_info.cm_id = cm_id;
3852 
3853 	cm_listen_node = irdma_make_listen_node(&iwdev->cm_core, iwdev,
3854 						&cm_info);
3855 	if (!cm_listen_node) {
3856 		irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3857 			    "cm_listen_node == NULL\n");
3858 		return -ENOMEM;
3859 	}
3860 
3861 	cm_id->provider_data = cm_listen_node;
3862 
3863 	cm_listen_node->tos = cm_id->tos;
3864 	if (iwdev->vsi.dscp_mode)
3865 		cm_listen_node->user_pri =
3866 		    iwdev->vsi.dscp_map[irdma_tos2dscp(cm_id->tos)];
3867 	else
3868 		cm_listen_node->user_pri = rt_tos2priority(cm_id->tos);
3869 	cm_info.user_pri = cm_listen_node->user_pri;
3870 	if (!cm_listen_node->reused_node) {
3871 		if (wildcard) {
3872 			err = irdma_add_mqh(iwdev, &cm_info, cm_listen_node);
3873 			if (err)
3874 				goto error;
3875 		} else {
3876 			if (!iwdev->vsi.dscp_mode)
3877 				cm_info.user_pri = cm_listen_node->user_pri =
3878 				    irdma_iw_get_vlan_prio(cm_info.loc_addr,
3879 							   cm_info.user_pri,
3880 							   cm_info.ipv4);
3881 			err = irdma_manage_qhash(iwdev, &cm_info,
3882 						 IRDMA_QHASH_TYPE_TCP_SYN,
3883 						 IRDMA_QHASH_MANAGE_TYPE_ADD,
3884 						 NULL, true);
3885 			if (err)
3886 				goto error;
3887 
3888 			cm_listen_node->qhash_set = true;
3889 		}
3890 
3891 		cm_listen_node->apbvt_entry = irdma_add_apbvt(iwdev,
3892 							      cm_info.loc_port);
3893 		if (!cm_listen_node->apbvt_entry)
3894 			goto error;
3895 	}
3896 	cm_id->add_ref(cm_id);
3897 	cm_listen_node->cm_core->stats_listen_created++;
3898 	irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3899 		    "loc_port=0x%04x loc_addr=%x cm_listen_node=%p cm_id=%p qhash_set=%d vlan_id=%d\n",
3900 		    cm_listen_node->loc_port, cm_listen_node->loc_addr[0],
3901 		    cm_listen_node, cm_listen_node->cm_id,
3902 		    cm_listen_node->qhash_set, cm_listen_node->vlan_id);
3903 
3904 	return 0;
3905 
3906 error:
3907 
3908 	irdma_cm_del_listen(&iwdev->cm_core, cm_listen_node, false);
3909 
3910 	return -EINVAL;
3911 }
3912 
3913 /**
3914  * irdma_destroy_listen - registered call to destroy listener
3915  * @cm_id: cm information for passive connection
3916  */
3917 int
irdma_destroy_listen(struct iw_cm_id * cm_id)3918 irdma_destroy_listen(struct iw_cm_id *cm_id)
3919 {
3920 	struct irdma_device *iwdev;
3921 
3922 	iwdev = to_iwdev(cm_id->device);
3923 	if (cm_id->provider_data)
3924 		irdma_cm_del_listen(&iwdev->cm_core, cm_id->provider_data,
3925 				    true);
3926 	else
3927 		irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3928 			    "cm_id->provider_data was NULL\n");
3929 
3930 	cm_id->rem_ref(cm_id);
3931 
3932 	return 0;
3933 }
3934 
3935 /**
3936  * irdma_iw_teardown_list_prep - add conn nodes slated for tear
3937  * down to list
3938  * @cm_core: cm's core
3939  * @teardown_list: a list to which cm_node will be selected
3940  * @ipaddr: pointer to ip address
3941  * @nfo: pointer to cm_info structure instance
3942  * @disconnect_all: flag indicating disconnect all QPs
3943  */
3944 static void
irdma_iw_teardown_list_prep(struct irdma_cm_core * cm_core,struct list_head * teardown_list,u32 * ipaddr,struct irdma_cm_info * nfo,bool disconnect_all)3945 irdma_iw_teardown_list_prep(struct irdma_cm_core *cm_core,
3946 			    struct list_head *teardown_list,
3947 			    u32 *ipaddr,
3948 			    struct irdma_cm_info *nfo,
3949 			    bool disconnect_all)
3950 {
3951 	struct irdma_cm_node *cm_node;
3952 	int bkt;
3953 
3954 	HASH_FOR_EACH_RCU(cm_core->cm_hash_tbl, bkt, cm_node, list) {
3955 		if ((disconnect_all ||
3956 		     (nfo->vlan_id == cm_node->vlan_id &&
3957 		      !memcmp(cm_node->loc_addr, ipaddr, nfo->ipv4 ? 4 : 16))) &&
3958 		    atomic_inc_not_zero(&cm_node->refcnt))
3959 			list_add(&cm_node->teardown_entry, teardown_list);
3960 	}
3961 }
3962 
3963 static inline bool
irdma_ip_vlan_match(u32 * ip1,u16 vlan_id1,bool check_vlan,u32 * ip2,u16 vlan_id2,bool ipv4)3964 irdma_ip_vlan_match(u32 *ip1, u16 vlan_id1,
3965 		    bool check_vlan, u32 *ip2,
3966 		    u16 vlan_id2, bool ipv4)
3967 {
3968 	return (!check_vlan || vlan_id1 == vlan_id2) &&
3969 	    !memcmp(ip1, ip2, ipv4 ? 4 : 16);
3970 }
3971 
3972 /**
3973  * irdma_roce_teardown_list_prep - add conn nodes slated for
3974  * tear down to list
3975  * @iwdev: RDMA device
3976  * @teardown_list: a list to which cm_node will be selected
3977  * @ipaddr: pointer to ip address
3978  * @nfo: pointer to cm_info structure instance
3979  * @disconnect_all: flag indicating disconnect all QPs
3980  */
3981 static void
irdma_roce_teardown_list_prep(struct irdma_device * iwdev,struct list_head * teardown_list,u32 * ipaddr,struct irdma_cm_info * nfo,bool disconnect_all)3982 irdma_roce_teardown_list_prep(struct irdma_device *iwdev,
3983 			      struct list_head *teardown_list,
3984 			      u32 *ipaddr,
3985 			      struct irdma_cm_info *nfo,
3986 			      bool disconnect_all)
3987 {
3988 	struct irdma_sc_vsi *vsi = &iwdev->vsi;
3989 	struct irdma_sc_qp *sc_qp;
3990 	struct list_head *list_node;
3991 	struct irdma_qp *qp;
3992 	unsigned long flags;
3993 	int i;
3994 
3995 	for (i = 0; i < IRDMA_MAX_USER_PRIORITY; i++) {
3996 		mutex_lock(&vsi->qos[i].qos_mutex);
3997 		list_for_each(list_node, &vsi->qos[i].qplist) {
3998 			u32 qp_ip[4];
3999 
4000 			sc_qp = container_of(list_node, struct irdma_sc_qp,
4001 					     list);
4002 			if (sc_qp->qp_uk.qp_type != IRDMA_QP_TYPE_ROCE_RC)
4003 				continue;
4004 
4005 			qp = sc_qp->qp_uk.back_qp;
4006 			if (!disconnect_all) {
4007 				if (nfo->ipv4)
4008 					qp_ip[0] = qp->udp_info.local_ipaddr[3];
4009 				else
4010 					memcpy(qp_ip,
4011 					       &qp->udp_info.local_ipaddr[0],
4012 					       sizeof(qp_ip));
4013 			}
4014 
4015 			if (disconnect_all ||
4016 			    irdma_ip_vlan_match(qp_ip,
4017 						qp->udp_info.vlan_tag & EVL_VLID_MASK,
4018 						qp->udp_info.insert_vlan_tag,
4019 						ipaddr, nfo->vlan_id, nfo->ipv4)) {
4020 				spin_lock_irqsave(&iwdev->rf->qptable_lock, flags);
4021 				if (iwdev->rf->qp_table[sc_qp->qp_uk.qp_id]) {
4022 					irdma_qp_add_ref(&qp->ibqp);
4023 					list_add(&qp->teardown_entry, teardown_list);
4024 				}
4025 				spin_unlock_irqrestore(&iwdev->rf->qptable_lock, flags);
4026 			}
4027 		}
4028 		mutex_unlock(&vsi->qos[i].qos_mutex);
4029 	}
4030 }
4031 
4032 /**
4033  * irdma_cm_event_connected - handle connected active node
4034  * @event: the info for cm_node of connection
4035  */
4036 static void
irdma_cm_event_connected(struct irdma_cm_event * event)4037 irdma_cm_event_connected(struct irdma_cm_event *event)
4038 {
4039 	struct irdma_qp *iwqp;
4040 	struct irdma_device *iwdev;
4041 	struct irdma_cm_node *cm_node;
4042 	struct irdma_sc_dev *dev;
4043 	struct ib_qp_attr attr = {0};
4044 	struct iw_cm_id *cm_id;
4045 	int status;
4046 	bool read0;
4047 	int wait_ret = 0;
4048 
4049 	cm_node = event->cm_node;
4050 	cm_id = cm_node->cm_id;
4051 	iwqp = cm_id->provider_data;
4052 	iwdev = iwqp->iwdev;
4053 	dev = &iwdev->rf->sc_dev;
4054 	if (iwqp->sc_qp.qp_uk.destroy_pending) {
4055 		status = -ETIMEDOUT;
4056 		goto error;
4057 	}
4058 
4059 	irdma_cm_init_tsa_conn(iwqp, cm_node);
4060 	read0 = (cm_node->send_rdma0_op == SEND_RDMA_READ_ZERO);
4061 	if (iwqp->page)
4062 		iwqp->sc_qp.qp_uk.sq_base = kmap_local_page(iwqp->page);
4063 	irdma_sc_send_rtt(&iwqp->sc_qp, read0);
4064 	if (iwqp->page)
4065 		kunmap_local(iwqp->sc_qp.qp_uk.sq_base);
4066 
4067 	attr.qp_state = IB_QPS_RTS;
4068 	cm_node->qhash_set = false;
4069 	irdma_modify_qp(&iwqp->ibqp, &attr, IB_QP_STATE, NULL);
4070 	if (dev->hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE) {
4071 		wait_ret = wait_event_interruptible_timeout(iwqp->waitq,
4072 							    iwqp->rts_ae_rcvd,
4073 							    IRDMA_MAX_TIMEOUT);
4074 		if (!wait_ret)
4075 			irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
4076 				    "Slow Connection: cm_node=%p, loc_port=%d, rem_port=%d, cm_id=%p\n",
4077 				    cm_node, cm_node->loc_port,
4078 				    cm_node->rem_port, cm_node->cm_id);
4079 	}
4080 
4081 	irdma_send_cm_event(cm_node, cm_id, IW_CM_EVENT_CONNECT_REPLY, 0);
4082 	cm_node->accelerated = true;
4083 	complete(&cm_node->establish_comp);
4084 	cm_node->cm_core->cm_free_ah(cm_node);
4085 	return;
4086 
4087 error:
4088 	iwqp->cm_id = NULL;
4089 	cm_id->provider_data = NULL;
4090 	irdma_send_cm_event(event->cm_node, cm_id, IW_CM_EVENT_CONNECT_REPLY,
4091 			    status);
4092 	irdma_rem_ref_cm_node(event->cm_node);
4093 }
4094 
4095 /**
4096  * irdma_cm_event_reset - handle reset
4097  * @event: the info for cm_node of connection
4098  */
4099 static void
irdma_cm_event_reset(struct irdma_cm_event * event)4100 irdma_cm_event_reset(struct irdma_cm_event *event)
4101 {
4102 	struct irdma_cm_node *cm_node = event->cm_node;
4103 	struct iw_cm_id *cm_id = cm_node->cm_id;
4104 	struct irdma_qp *iwqp;
4105 
4106 	if (!cm_id)
4107 		return;
4108 
4109 	iwqp = cm_id->provider_data;
4110 	if (!iwqp)
4111 		return;
4112 
4113 	irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
4114 		    "reset event %p - cm_id = %p\n", event->cm_node, cm_id);
4115 	iwqp->cm_id = NULL;
4116 
4117 	irdma_send_cm_event(cm_node, cm_node->cm_id, IW_CM_EVENT_DISCONNECT,
4118 			    -ECONNRESET);
4119 	irdma_send_cm_event(cm_node, cm_node->cm_id, IW_CM_EVENT_CLOSE, 0);
4120 }
4121 
4122 /**
4123  * irdma_cm_event_handler - send event to cm upper layer
4124  * @work: pointer of cm event info.
4125  */
4126 static void
irdma_cm_event_handler(struct work_struct * work)4127 irdma_cm_event_handler(struct work_struct *work)
4128 {
4129 	struct irdma_cm_event *event = container_of(work, struct irdma_cm_event, event_work);
4130 	struct irdma_cm_node *cm_node;
4131 
4132 	if (!event || !event->cm_node || !event->cm_node->cm_core)
4133 		return;
4134 
4135 	cm_node = event->cm_node;
4136 
4137 	switch (event->type) {
4138 	case IRDMA_CM_EVENT_MPA_REQ:
4139 		irdma_send_cm_event(cm_node, cm_node->cm_id,
4140 				    IW_CM_EVENT_CONNECT_REQUEST, 0);
4141 		break;
4142 	case IRDMA_CM_EVENT_RESET:
4143 		irdma_cm_event_reset(event);
4144 		break;
4145 	case IRDMA_CM_EVENT_CONNECTED:
4146 		if (!event->cm_node->cm_id ||
4147 		    event->cm_node->state != IRDMA_CM_STATE_OFFLOADED)
4148 			break;
4149 		irdma_cm_event_connected(event);
4150 		break;
4151 	case IRDMA_CM_EVENT_MPA_REJECT:
4152 		if (!event->cm_node->cm_id ||
4153 		    cm_node->state == IRDMA_CM_STATE_OFFLOADED)
4154 			break;
4155 		irdma_send_cm_event(cm_node, cm_node->cm_id,
4156 				    IW_CM_EVENT_CONNECT_REPLY, -ECONNREFUSED);
4157 		break;
4158 	case IRDMA_CM_EVENT_ABORTED:
4159 		if (!event->cm_node->cm_id ||
4160 		    event->cm_node->state == IRDMA_CM_STATE_OFFLOADED)
4161 			break;
4162 		irdma_event_connect_error(event);
4163 		break;
4164 	default:
4165 		irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
4166 			    "bad event type = %d\n", event->type);
4167 		break;
4168 	}
4169 
4170 	irdma_rem_ref_cm_node(event->cm_node);
4171 	kfree(event);
4172 }
4173 
4174 /**
4175  * irdma_cm_post_event - queue event request for worker thread
4176  * @event: cm node's info for up event call
4177  */
4178 static void
irdma_cm_post_event(struct irdma_cm_event * event)4179 irdma_cm_post_event(struct irdma_cm_event *event)
4180 {
4181 	atomic_inc(&event->cm_node->refcnt);
4182 	INIT_WORK(&event->event_work, irdma_cm_event_handler);
4183 	queue_work(event->cm_node->cm_core->event_wq, &event->event_work);
4184 }
4185 
4186 /**
4187  * irdma_cm_teardown_connections - teardown QPs
4188  * @iwdev: device pointer
4189  * @ipaddr: Pointer to IPv4 or IPv6 address
4190  * @nfo: Connection info
4191  * @disconnect_all: flag indicating disconnect all QPs
4192  *
4193  * teardown QPs where source or destination addr matches ip addr
4194  */
4195 static void __unused
irdma_cm_teardown_connections(struct irdma_device * iwdev,u32 * ipaddr,struct irdma_cm_info * nfo,bool disconnect_all)4196 irdma_cm_teardown_connections(struct irdma_device *iwdev,
4197 			      u32 *ipaddr,
4198 			      struct irdma_cm_info *nfo,
4199 			      bool disconnect_all)
4200 {
4201 	struct irdma_cm_core *cm_core = &iwdev->cm_core;
4202 	struct list_head *list_core_temp;
4203 	struct list_head *list_node;
4204 	struct irdma_cm_node *cm_node;
4205 	struct list_head teardown_list;
4206 	struct ib_qp_attr attr;
4207 	struct irdma_qp *qp;
4208 
4209 	INIT_LIST_HEAD(&teardown_list);
4210 
4211 	rcu_read_lock();
4212 	irdma_iw_teardown_list_prep(cm_core, &teardown_list, ipaddr, nfo, disconnect_all);
4213 	rcu_read_unlock();
4214 
4215 	attr.qp_state = IB_QPS_ERR;
4216 	list_for_each_safe(list_node, list_core_temp, &teardown_list) {
4217 		cm_node = container_of(list_node, struct irdma_cm_node,
4218 				       teardown_entry);
4219 		irdma_modify_qp(&cm_node->iwqp->ibqp, &attr, IB_QP_STATE, NULL);
4220 		if (iwdev->rf->reset)
4221 			irdma_cm_disconn(cm_node->iwqp);
4222 		irdma_rem_ref_cm_node(cm_node);
4223 	}
4224 
4225 	if (!rdma_protocol_roce(&iwdev->ibdev, 1))
4226 		return;
4227 
4228 	INIT_LIST_HEAD(&teardown_list);
4229 	irdma_roce_teardown_list_prep(iwdev, &teardown_list, ipaddr, nfo, disconnect_all);
4230 
4231 	list_for_each_safe(list_node, list_core_temp, &teardown_list) {
4232 		qp = container_of(list_node, struct irdma_qp, teardown_entry);
4233 		irdma_modify_qp_roce(&qp->ibqp, &attr, IB_QP_STATE, NULL);
4234 		irdma_ib_qp_event(qp, IRDMA_QP_EVENT_CATASTROPHIC);
4235 		irdma_qp_rem_ref(&qp->ibqp);
4236 	}
4237 }
4238