1 /*-
2 * SPDX-License-Identifier: GPL-2.0 or Linux-OpenIB
3 *
4 * Copyright (c) 2015 - 2025 Intel Corporation
5 *
6 * This software is available to you under a choice of one of two
7 * licenses. You may choose to be licensed under the terms of the GNU
8 * General Public License (GPL) Version 2, available from the file
9 * COPYING in the main directory of this source tree, or the
10 * OpenFabrics.org BSD license below:
11 *
12 * Redistribution and use in source and binary forms, with or
13 * without modification, are permitted provided that the following
14 * conditions are met:
15 *
16 * - Redistributions of source code must retain the above
17 * copyright notice, this list of conditions and the following
18 * disclaimer.
19 *
20 * - Redistributions in binary form must reproduce the above
21 * copyright notice, this list of conditions and the following
22 * disclaimer in the documentation and/or other materials
23 * provided with the distribution.
24 *
25 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
26 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
27 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
28 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
29 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
30 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
31 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
32 * SOFTWARE.
33 */
34
35 #include "irdma_main.h"
36
37 static void irdma_cm_post_event(struct irdma_cm_event *event);
38 static void irdma_disconnect_worker(struct work_struct *work);
39
40 /**
41 * irdma_free_sqbuf - put back puda buffer if refcount is 0
42 * @vsi: The VSI structure of the device
43 * @bufp: puda buffer to free
44 */
45 void
irdma_free_sqbuf(struct irdma_sc_vsi * vsi,void * bufp)46 irdma_free_sqbuf(struct irdma_sc_vsi *vsi, void *bufp)
47 {
48 struct irdma_puda_buf *buf = bufp;
49 struct irdma_puda_rsrc *ilq = vsi->ilq;
50
51 if (atomic_dec_and_test(&buf->refcount))
52 irdma_puda_ret_bufpool(ilq, buf);
53 }
54
55 /**
56 * irdma_record_ird_ord - Record IRD/ORD passed in
57 * @cm_node: connection's node
58 * @conn_ird: connection IRD
59 * @conn_ord: connection ORD
60 */
61 static void
irdma_record_ird_ord(struct irdma_cm_node * cm_node,u32 conn_ird,u32 conn_ord)62 irdma_record_ird_ord(struct irdma_cm_node *cm_node, u32 conn_ird,
63 u32 conn_ord)
64 {
65 if (conn_ird > cm_node->dev->hw_attrs.max_hw_ird)
66 conn_ird = cm_node->dev->hw_attrs.max_hw_ird;
67
68 if (conn_ord > cm_node->dev->hw_attrs.max_hw_ord)
69 conn_ord = cm_node->dev->hw_attrs.max_hw_ord;
70 else if (!conn_ord && cm_node->send_rdma0_op == SEND_RDMA_READ_ZERO)
71 conn_ord = 1;
72 cm_node->ird_size = conn_ird;
73 cm_node->ord_size = conn_ord;
74 }
75
76 /**
77 * irdma_copy_ip_ntohl - copy IP address from network to host
78 * @dst: IP address in host order
79 * @src: IP address in network order (big endian)
80 */
81 void
irdma_copy_ip_ntohl(u32 * dst,__be32 * src)82 irdma_copy_ip_ntohl(u32 *dst, __be32 *src)
83 {
84 *dst++ = ntohl(*src++);
85 *dst++ = ntohl(*src++);
86 *dst++ = ntohl(*src++);
87 *dst = ntohl(*src);
88 }
89
90 /**
91 * irdma_copy_ip_htonl - copy IP address from host to network order
92 * @dst: IP address in network order (big endian)
93 * @src: IP address in host order
94 */
95 void
irdma_copy_ip_htonl(__be32 * dst,u32 * src)96 irdma_copy_ip_htonl(__be32 *dst, u32 *src)
97 {
98 *dst++ = htonl(*src++);
99 *dst++ = htonl(*src++);
100 *dst++ = htonl(*src++);
101 *dst = htonl(*src);
102 }
103
104 /**
105 * irdma_get_addr_info
106 * @cm_node: contains ip/tcp info
107 * @cm_info: to get a copy of the cm_node ip/tcp info
108 */
109 static void
irdma_get_addr_info(struct irdma_cm_node * cm_node,struct irdma_cm_info * cm_info)110 irdma_get_addr_info(struct irdma_cm_node *cm_node,
111 struct irdma_cm_info *cm_info)
112 {
113 memset(cm_info, 0, sizeof(*cm_info));
114 cm_info->ipv4 = cm_node->ipv4;
115 cm_info->vlan_id = cm_node->vlan_id;
116 memcpy(cm_info->loc_addr, cm_node->loc_addr, sizeof(cm_info->loc_addr));
117 memcpy(cm_info->rem_addr, cm_node->rem_addr, sizeof(cm_info->rem_addr));
118 cm_info->loc_port = cm_node->loc_port;
119 cm_info->rem_port = cm_node->rem_port;
120 }
121
122 /**
123 * irdma_fill_sockaddr4 - fill in addr info for IPv4 connection
124 * @cm_node: connection's node
125 * @event: upper layer's cm event
126 */
127 static inline void
irdma_fill_sockaddr4(struct irdma_cm_node * cm_node,struct iw_cm_event * event)128 irdma_fill_sockaddr4(struct irdma_cm_node *cm_node,
129 struct iw_cm_event *event)
130 {
131 struct sockaddr_in *laddr = (struct sockaddr_in *)&event->local_addr;
132 struct sockaddr_in *raddr = (struct sockaddr_in *)&event->remote_addr;
133
134 laddr->sin_family = AF_INET;
135 raddr->sin_family = AF_INET;
136
137 laddr->sin_port = htons(cm_node->loc_port);
138 raddr->sin_port = htons(cm_node->rem_port);
139
140 laddr->sin_addr.s_addr = htonl(cm_node->loc_addr[0]);
141 raddr->sin_addr.s_addr = htonl(cm_node->rem_addr[0]);
142 }
143
144 /**
145 * irdma_fill_sockaddr6 - fill in addr info for IPv6 connection
146 * @cm_node: connection's node
147 * @event: upper layer's cm event
148 */
149 static inline void
irdma_fill_sockaddr6(struct irdma_cm_node * cm_node,struct iw_cm_event * event)150 irdma_fill_sockaddr6(struct irdma_cm_node *cm_node,
151 struct iw_cm_event *event)
152 {
153 struct sockaddr_in6 *laddr6 = (struct sockaddr_in6 *)&event->local_addr;
154 struct sockaddr_in6 *raddr6 = (struct sockaddr_in6 *)&event->remote_addr;
155
156 laddr6->sin6_family = AF_INET6;
157 raddr6->sin6_family = AF_INET6;
158
159 laddr6->sin6_port = htons(cm_node->loc_port);
160 raddr6->sin6_port = htons(cm_node->rem_port);
161
162 irdma_copy_ip_htonl(laddr6->sin6_addr.__u6_addr.__u6_addr32,
163 cm_node->loc_addr);
164 irdma_copy_ip_htonl(raddr6->sin6_addr.__u6_addr.__u6_addr32,
165 cm_node->rem_addr);
166 }
167
168 /**
169 * irdma_get_cmevent_info - for cm event upcall
170 * @cm_node: connection's node
171 * @cm_id: upper layers cm struct for the event
172 * @event: upper layer's cm event
173 */
174 static inline void
irdma_get_cmevent_info(struct irdma_cm_node * cm_node,struct iw_cm_id * cm_id,struct iw_cm_event * event)175 irdma_get_cmevent_info(struct irdma_cm_node *cm_node,
176 struct iw_cm_id *cm_id,
177 struct iw_cm_event *event)
178 {
179 memcpy(&event->local_addr, &cm_id->m_local_addr,
180 sizeof(event->local_addr));
181 memcpy(&event->remote_addr, &cm_id->m_remote_addr,
182 sizeof(event->remote_addr));
183 if (cm_node) {
184 event->private_data = cm_node->pdata_buf;
185 event->private_data_len = (u8)cm_node->pdata.size;
186 event->ird = cm_node->ird_size;
187 event->ord = cm_node->ord_size;
188 }
189 }
190
191 /**
192 * irdma_send_cm_event - upcall cm's event handler
193 * @cm_node: connection's node
194 * @cm_id: upper layer's cm info struct
195 * @type: Event type to indicate
196 * @status: status for the event type
197 */
198 static int
irdma_send_cm_event(struct irdma_cm_node * cm_node,struct iw_cm_id * cm_id,enum iw_cm_event_type type,int status)199 irdma_send_cm_event(struct irdma_cm_node *cm_node,
200 struct iw_cm_id *cm_id,
201 enum iw_cm_event_type type, int status)
202 {
203 struct iw_cm_event event = {0};
204
205 event.event = type;
206 event.status = status;
207
208 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
209 "cm_node %p cm_id=%p state=%d accel=%d event_type=%d status=%d\n",
210 cm_node, cm_id, cm_node->accelerated, cm_node->state, type,
211 status);
212
213 switch (type) {
214 case IW_CM_EVENT_CONNECT_REQUEST:
215 if (cm_node->ipv4)
216 irdma_fill_sockaddr4(cm_node, &event);
217 else
218 irdma_fill_sockaddr6(cm_node, &event);
219 event.provider_data = cm_node;
220 event.private_data = cm_node->pdata_buf;
221 event.private_data_len = (u8)cm_node->pdata.size;
222 event.ird = cm_node->ird_size;
223 break;
224 case IW_CM_EVENT_CONNECT_REPLY:
225 irdma_get_cmevent_info(cm_node, cm_id, &event);
226 break;
227 case IW_CM_EVENT_ESTABLISHED:
228 event.ird = cm_node->ird_size;
229 event.ord = cm_node->ord_size;
230 break;
231 case IW_CM_EVENT_DISCONNECT:
232 case IW_CM_EVENT_CLOSE:
233 /* Wait if we are in RTS but havent issued the iwcm event upcall */
234 if (!cm_node->accelerated)
235 wait_for_completion(&cm_node->establish_comp);
236 break;
237 default:
238 return -EINVAL;
239 }
240
241 return cm_id->event_handler(cm_id, &event);
242 }
243
244 /**
245 * irdma_timer_list_prep - add connection nodes to a list to perform timer tasks
246 * @cm_core: cm's core
247 * @timer_list: a timer list to which cm_node will be selected
248 */
249 static void
irdma_timer_list_prep(struct irdma_cm_core * cm_core,struct list_head * timer_list)250 irdma_timer_list_prep(struct irdma_cm_core *cm_core,
251 struct list_head *timer_list)
252 {
253 struct irdma_cm_node *cm_node;
254 int bkt;
255
256 HASH_FOR_EACH_RCU(cm_core->cm_hash_tbl, bkt, cm_node, list) {
257 if ((cm_node->close_entry || cm_node->send_entry) &&
258 atomic_inc_not_zero(&cm_node->refcnt))
259 list_add(&cm_node->timer_entry, timer_list);
260 }
261 }
262
263 /**
264 * irdma_create_event - create cm event
265 * @cm_node: connection's node
266 * @type: Event type to generate
267 */
268 static struct irdma_cm_event *
irdma_create_event(struct irdma_cm_node * cm_node,enum irdma_cm_event_type type)269 irdma_create_event(struct irdma_cm_node *cm_node,
270 enum irdma_cm_event_type type)
271 {
272 struct irdma_cm_event *event;
273
274 if (!cm_node->cm_id)
275 return NULL;
276
277 event = kzalloc(sizeof(*event), GFP_ATOMIC);
278
279 if (!event)
280 return NULL;
281
282 event->type = type;
283 event->cm_node = cm_node;
284 memcpy(event->cm_info.rem_addr, cm_node->rem_addr,
285 sizeof(event->cm_info.rem_addr));
286 memcpy(event->cm_info.loc_addr, cm_node->loc_addr,
287 sizeof(event->cm_info.loc_addr));
288 event->cm_info.rem_port = cm_node->rem_port;
289 event->cm_info.loc_port = cm_node->loc_port;
290 event->cm_info.cm_id = cm_node->cm_id;
291 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
292 "node=%p event=%p type=%u dst=%x src=%x\n", cm_node, event,
293 type, event->cm_info.loc_addr[0],
294 event->cm_info.rem_addr[0]);
295 irdma_cm_post_event(event);
296
297 return event;
298 }
299
300 /**
301 * irdma_free_retrans_entry - free send entry
302 * @cm_node: connection's node
303 */
304 static void
irdma_free_retrans_entry(struct irdma_cm_node * cm_node)305 irdma_free_retrans_entry(struct irdma_cm_node *cm_node)
306 {
307 struct irdma_device *iwdev = cm_node->iwdev;
308 struct irdma_timer_entry *send_entry;
309
310 send_entry = cm_node->send_entry;
311 if (!send_entry)
312 return;
313
314 cm_node->send_entry = NULL;
315 irdma_free_sqbuf(&iwdev->vsi, send_entry->sqbuf);
316 kfree(send_entry);
317 atomic_dec(&cm_node->refcnt);
318 }
319
320 /**
321 * irdma_cleanup_retrans_entry - free send entry with lock
322 * @cm_node: connection's node
323 */
324 static void
irdma_cleanup_retrans_entry(struct irdma_cm_node * cm_node)325 irdma_cleanup_retrans_entry(struct irdma_cm_node *cm_node)
326 {
327 unsigned long flags;
328
329 spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
330 irdma_free_retrans_entry(cm_node);
331 spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
332 }
333
334 /**
335 * irdma_form_ah_cm_frame - get a free packet and build frame with address handle
336 * @cm_node: connection's node ionfo to use in frame
337 * @options: pointer to options info
338 * @hdr: pointer mpa header
339 * @pdata: pointer to private data
340 * @flags: indicates FIN or ACK
341 */
342 static struct irdma_puda_buf *
irdma_form_ah_cm_frame(struct irdma_cm_node * cm_node,struct irdma_kmem_info * options,struct irdma_kmem_info * hdr,struct irdma_mpa_priv_info * pdata,u8 flags)343 irdma_form_ah_cm_frame(struct irdma_cm_node *cm_node,
344 struct irdma_kmem_info *options,
345 struct irdma_kmem_info *hdr,
346 struct irdma_mpa_priv_info *pdata,
347 u8 flags)
348 {
349 struct irdma_puda_buf *sqbuf;
350 struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
351 u8 *buf;
352 struct tcphdr *tcph;
353 u16 pktsize;
354 u32 opts_len = 0;
355 u32 pd_len = 0;
356 u32 hdr_len = 0;
357
358 if (!cm_node->ah || !cm_node->ah->ah_info.ah_valid) {
359 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM, "AH invalid\n");
360 return NULL;
361 }
362
363 sqbuf = irdma_puda_get_bufpool(vsi->ilq);
364 if (!sqbuf) {
365 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM, "SQ buf NULL\n");
366 return NULL;
367 }
368
369 sqbuf->ah_id = cm_node->ah->ah_info.ah_idx;
370 buf = sqbuf->mem.va;
371 if (options)
372 opts_len = (u32)options->size;
373
374 if (hdr)
375 hdr_len = hdr->size;
376
377 if (pdata)
378 pd_len = pdata->size;
379
380 pktsize = sizeof(*tcph) + opts_len + hdr_len + pd_len;
381
382 memset(buf, 0, pktsize);
383
384 sqbuf->totallen = pktsize;
385 sqbuf->tcphlen = sizeof(*tcph) + opts_len;
386 sqbuf->scratch = cm_node;
387
388 tcph = (struct tcphdr *)buf;
389 buf += sizeof(*tcph);
390
391 tcph->th_sport = htons(cm_node->loc_port);
392 tcph->th_dport = htons(cm_node->rem_port);
393 tcph->th_seq = htonl(cm_node->tcp_cntxt.loc_seq_num);
394
395 if (flags & SET_ACK) {
396 cm_node->tcp_cntxt.loc_ack_num = cm_node->tcp_cntxt.rcv_nxt;
397 tcph->th_ack = htonl(cm_node->tcp_cntxt.loc_ack_num);
398 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_ACK);
399 } else {
400 tcph->th_ack = 0;
401 }
402
403 if (flags & SET_SYN) {
404 cm_node->tcp_cntxt.loc_seq_num++;
405 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_SYN);
406 } else {
407 cm_node->tcp_cntxt.loc_seq_num += hdr_len + pd_len;
408 }
409
410 if (flags & SET_FIN) {
411 cm_node->tcp_cntxt.loc_seq_num++;
412 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_FIN);
413 }
414
415 if (flags & SET_RST)
416 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_RST);
417
418 tcph->th_off = (u16)((sizeof(*tcph) + opts_len + 3) >> 2);
419 sqbuf->tcphlen = tcph->th_off << 2;
420 tcph->th_win = htons(cm_node->tcp_cntxt.rcv_wnd);
421 tcph->th_urp = 0;
422
423 if (opts_len) {
424 memcpy(buf, options->addr, opts_len);
425 buf += opts_len;
426 }
427
428 if (hdr_len) {
429 memcpy(buf, hdr->addr, hdr_len);
430 buf += hdr_len;
431 }
432
433 if (pdata && pdata->addr)
434 memcpy(buf, pdata->addr, pdata->size);
435
436 atomic_set(&sqbuf->refcount, 1);
437
438 irdma_debug_buf(vsi->dev, IRDMA_DEBUG_ILQ, "TRANSMIT ILQ BUFFER",
439 sqbuf->mem.va, sqbuf->totallen);
440
441 return sqbuf;
442 }
443
444 /**
445 * irdma_form_uda_cm_frame - get a free packet and build frame full tcpip packet
446 * @cm_node: connection's node ionfo to use in frame
447 * @options: pointer to options info
448 * @hdr: pointer mpa header
449 * @pdata: pointer to private data
450 * @flags: indicates FIN or ACK
451 */
452 static struct irdma_puda_buf *
irdma_form_uda_cm_frame(struct irdma_cm_node * cm_node,struct irdma_kmem_info * options,struct irdma_kmem_info * hdr,struct irdma_mpa_priv_info * pdata,u8 flags)453 irdma_form_uda_cm_frame(struct irdma_cm_node *cm_node,
454 struct irdma_kmem_info *options,
455 struct irdma_kmem_info *hdr,
456 struct irdma_mpa_priv_info *pdata,
457 u8 flags)
458 {
459 struct irdma_puda_buf *sqbuf;
460 struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
461 u8 *buf;
462
463 struct tcphdr *tcph;
464 struct ip *iph;
465 struct ip6_hdr *ip6h;
466 struct ether_header *ethh;
467 u16 pktsize;
468 u16 eth_hlen = ETH_HLEN;
469 u32 opts_len = 0;
470 u32 pd_len = 0;
471 u32 hdr_len = 0;
472
473 u16 vtag;
474
475 sqbuf = irdma_puda_get_bufpool(vsi->ilq);
476 if (!sqbuf)
477 return NULL;
478
479 buf = sqbuf->mem.va;
480
481 if (options)
482 opts_len = (u32)options->size;
483
484 if (hdr)
485 hdr_len = hdr->size;
486
487 if (pdata)
488 pd_len = pdata->size;
489
490 if (cm_node->vlan_id < VLAN_N_VID)
491 eth_hlen += 4;
492
493 if (cm_node->ipv4)
494 pktsize = sizeof(*iph) + sizeof(*tcph);
495 else
496 pktsize = sizeof(*ip6h) + sizeof(*tcph);
497 pktsize += opts_len + hdr_len + pd_len;
498
499 memset(buf, 0, eth_hlen + pktsize);
500
501 sqbuf->totallen = pktsize + eth_hlen;
502 sqbuf->maclen = eth_hlen;
503 sqbuf->tcphlen = sizeof(*tcph) + opts_len;
504 sqbuf->scratch = cm_node;
505
506 ethh = (struct ether_header *)buf;
507 buf += eth_hlen;
508
509 if (cm_node->do_lpb)
510 sqbuf->do_lpb = true;
511
512 if (cm_node->ipv4) {
513 sqbuf->ipv4 = true;
514
515 iph = (struct ip *)buf;
516 buf += sizeof(*iph);
517 tcph = (struct tcphdr *)buf;
518 buf += sizeof(*tcph);
519
520 ether_addr_copy(ethh->ether_dhost, cm_node->rem_mac);
521 ether_addr_copy(ethh->ether_shost, cm_node->loc_mac);
522 if (cm_node->vlan_id < VLAN_N_VID) {
523 ((struct ether_vlan_header *)ethh)->evl_proto =
524 htons(ETH_P_8021Q);
525 vtag = (cm_node->user_pri << VLAN_PRIO_SHIFT) |
526 cm_node->vlan_id;
527 ((struct ether_vlan_header *)ethh)->evl_tag = htons(vtag);
528
529 ((struct ether_vlan_header *)ethh)->evl_encap_proto =
530 htons(ETH_P_IP);
531 } else {
532 ethh->ether_type = htons(ETH_P_IP);
533 }
534
535 iph->ip_v = IPVERSION;
536 iph->ip_hl = 5; /* 5 * 4Byte words, IP headr len */
537 iph->ip_tos = cm_node->tos;
538 iph->ip_len = htons(pktsize);
539 iph->ip_id = htons(++cm_node->tcp_cntxt.loc_id);
540
541 iph->ip_off = htons(0x4000);
542 iph->ip_ttl = 0x40;
543 iph->ip_p = IPPROTO_TCP;
544 iph->ip_src.s_addr = htonl(cm_node->loc_addr[0]);
545 iph->ip_dst.s_addr = htonl(cm_node->rem_addr[0]);
546 } else {
547 sqbuf->ipv4 = false;
548 ip6h = (struct ip6_hdr *)buf;
549 buf += sizeof(*ip6h);
550 tcph = (struct tcphdr *)buf;
551 buf += sizeof(*tcph);
552
553 ether_addr_copy(ethh->ether_dhost, cm_node->rem_mac);
554 ether_addr_copy(ethh->ether_shost, cm_node->loc_mac);
555 if (cm_node->vlan_id < VLAN_N_VID) {
556 ((struct ether_vlan_header *)ethh)->evl_proto =
557 htons(ETH_P_8021Q);
558 vtag = (cm_node->user_pri << VLAN_PRIO_SHIFT) |
559 cm_node->vlan_id;
560 ((struct ether_vlan_header *)ethh)->evl_tag = htons(vtag);
561 ((struct ether_vlan_header *)ethh)->evl_encap_proto =
562 htons(ETH_P_IPV6);
563 } else {
564 ethh->ether_type = htons(ETH_P_IPV6);
565 }
566 ip6h->ip6_vfc = 6 << 4;
567 ip6h->ip6_vfc |= cm_node->tos >> 4;
568 ip6h->ip6_flow = cm_node->tos << 20;
569 ip6h->ip6_plen = htons(pktsize - sizeof(*ip6h));
570 ip6h->ip6_nxt = 6;
571 ip6h->ip6_hops = 128;
572 irdma_copy_ip_htonl(ip6h->ip6_src.__u6_addr.__u6_addr32,
573 cm_node->loc_addr);
574 irdma_copy_ip_htonl(ip6h->ip6_dst.__u6_addr.__u6_addr32,
575 cm_node->rem_addr);
576 }
577
578 tcph->th_sport = htons(cm_node->loc_port);
579 tcph->th_dport = htons(cm_node->rem_port);
580 tcph->th_seq = htonl(cm_node->tcp_cntxt.loc_seq_num);
581
582 if (flags & SET_ACK) {
583 cm_node->tcp_cntxt.loc_ack_num = cm_node->tcp_cntxt.rcv_nxt;
584 tcph->th_ack = htonl(cm_node->tcp_cntxt.loc_ack_num);
585 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_ACK);
586 } else {
587 tcph->th_ack = 0;
588 }
589
590 if (flags & SET_SYN) {
591 cm_node->tcp_cntxt.loc_seq_num++;
592 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_SYN);
593 } else {
594 cm_node->tcp_cntxt.loc_seq_num += hdr_len + pd_len;
595 }
596
597 if (flags & SET_FIN) {
598 cm_node->tcp_cntxt.loc_seq_num++;
599 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_FIN);
600 }
601
602 if (flags & SET_RST)
603 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_RST);
604
605 tcph->th_off = (u16)((sizeof(*tcph) + opts_len + 3) >> 2);
606 sqbuf->tcphlen = tcph->th_off << 2;
607 tcph->th_win = htons(cm_node->tcp_cntxt.rcv_wnd);
608 tcph->th_urp = 0;
609
610 if (opts_len) {
611 memcpy(buf, options->addr, opts_len);
612 buf += opts_len;
613 }
614
615 if (hdr_len) {
616 memcpy(buf, hdr->addr, hdr_len);
617 buf += hdr_len;
618 }
619
620 if (pdata && pdata->addr)
621 memcpy(buf, pdata->addr, pdata->size);
622
623 atomic_set(&sqbuf->refcount, 1);
624
625 irdma_debug_buf(vsi->dev, IRDMA_DEBUG_ILQ, "TRANSMIT ILQ BUFFER",
626 sqbuf->mem.va, sqbuf->totallen);
627
628 return sqbuf;
629 }
630
631 /**
632 * irdma_send_reset - Send RST packet
633 * @cm_node: connection's node
634 */
635 int
irdma_send_reset(struct irdma_cm_node * cm_node)636 irdma_send_reset(struct irdma_cm_node *cm_node)
637 {
638 struct irdma_puda_buf *sqbuf;
639 int flags = SET_RST | SET_ACK;
640
641 sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL, NULL, NULL,
642 flags);
643 if (!sqbuf)
644 return -ENOMEM;
645
646 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
647 "caller: %pS cm_node %p cm_id=%p accel=%d state=%d rem_port=0x%04x, loc_port=0x%04x rem_addr=%x loc_addr=%x\n",
648 __builtin_return_address(0), cm_node, cm_node->cm_id,
649 cm_node->accelerated, cm_node->state, cm_node->rem_port,
650 cm_node->loc_port, cm_node->rem_addr[0],
651 cm_node->loc_addr[0]);
652
653 return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 0,
654 1);
655 }
656
657 /**
658 * irdma_active_open_err - send event for active side cm error
659 * @cm_node: connection's node
660 * @reset: Flag to send reset or not
661 */
662 static void
irdma_active_open_err(struct irdma_cm_node * cm_node,bool reset)663 irdma_active_open_err(struct irdma_cm_node *cm_node, bool reset)
664 {
665 irdma_cleanup_retrans_entry(cm_node);
666 cm_node->cm_core->stats_connect_errs++;
667 if (reset) {
668 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
669 "cm_node=%p state=%d\n", cm_node, cm_node->state);
670 atomic_inc(&cm_node->refcnt);
671 irdma_send_reset(cm_node);
672 }
673
674 cm_node->state = IRDMA_CM_STATE_CLOSED;
675 irdma_create_event(cm_node, IRDMA_CM_EVENT_ABORTED);
676 }
677
678 /**
679 * irdma_passive_open_err - handle passive side cm error
680 * @cm_node: connection's node
681 * @reset: send reset or just free cm_node
682 */
683 static void
irdma_passive_open_err(struct irdma_cm_node * cm_node,bool reset)684 irdma_passive_open_err(struct irdma_cm_node *cm_node, bool reset)
685 {
686 irdma_cleanup_retrans_entry(cm_node);
687 cm_node->cm_core->stats_passive_errs++;
688 cm_node->state = IRDMA_CM_STATE_CLOSED;
689 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
690 "cm_node=%p state=%d\n", cm_node, cm_node->state);
691 if (reset)
692 irdma_send_reset(cm_node);
693 else
694 irdma_rem_ref_cm_node(cm_node);
695 }
696
697 /**
698 * irdma_event_connect_error - to create connect error event
699 * @event: cm information for connect event
700 */
701 static void
irdma_event_connect_error(struct irdma_cm_event * event)702 irdma_event_connect_error(struct irdma_cm_event *event)
703 {
704 struct irdma_qp *iwqp;
705 struct iw_cm_id *cm_id;
706
707 cm_id = event->cm_node->cm_id;
708 if (!cm_id)
709 return;
710
711 iwqp = cm_id->provider_data;
712
713 if (!iwqp || !iwqp->iwdev)
714 return;
715
716 iwqp->cm_id = NULL;
717 cm_id->provider_data = NULL;
718 irdma_send_cm_event(event->cm_node, cm_id, IW_CM_EVENT_CONNECT_REPLY,
719 -ECONNRESET);
720 irdma_rem_ref_cm_node(event->cm_node);
721 }
722
723 /**
724 * irdma_process_options - process options from TCP header
725 * @cm_node: connection's node
726 * @optionsloc: point to start of options
727 * @optionsize: size of all options
728 * @syn_pkt: flag if syn packet
729 */
730 static int
irdma_process_options(struct irdma_cm_node * cm_node,u8 * optionsloc,u32 optionsize,u32 syn_pkt)731 irdma_process_options(struct irdma_cm_node *cm_node, u8 *optionsloc,
732 u32 optionsize, u32 syn_pkt)
733 {
734 u32 tmp;
735 u32 offset = 0;
736 union all_known_options *all_options;
737 char got_mss_option = 0;
738
739 while (offset < optionsize) {
740 all_options = (union all_known_options *)(optionsloc + offset);
741 switch (all_options->base.optionnum) {
742 case OPTION_NUM_EOL:
743 offset = optionsize;
744 break;
745 case OPTION_NUM_NONE:
746 offset += 1;
747 continue;
748 case OPTION_NUM_MSS:
749 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
750 "MSS Length: %d Offset: %d Size: %d\n",
751 all_options->mss.len, offset, optionsize);
752 got_mss_option = 1;
753 if (all_options->mss.len != 4)
754 return -EINVAL;
755 tmp = ntohs(all_options->mss.mss);
756 if ((cm_node->ipv4 &&
757 (tmp + IRDMA_MTU_TO_MSS_IPV4) < IRDMA_MIN_MTU_IPV4) ||
758 (!cm_node->ipv4 &&
759 (tmp + IRDMA_MTU_TO_MSS_IPV6) < IRDMA_MIN_MTU_IPV6))
760 return -EINVAL;
761 if (tmp < cm_node->tcp_cntxt.mss)
762 cm_node->tcp_cntxt.mss = tmp;
763 break;
764 case OPTION_NUM_WINDOW_SCALE:
765 cm_node->tcp_cntxt.snd_wscale =
766 all_options->windowscale.shiftcount;
767 break;
768 default:
769 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
770 "Unsupported TCP Option: %x\n",
771 all_options->base.optionnum);
772 break;
773 }
774 offset += all_options->base.len;
775 }
776 if (!got_mss_option && syn_pkt)
777 cm_node->tcp_cntxt.mss = IRDMA_CM_DEFAULT_MSS;
778
779 return 0;
780 }
781
782 /**
783 * irdma_handle_tcp_options - setup TCP context info after parsing TCP options
784 * @cm_node: connection's node
785 * @tcph: pointer tcp header
786 * @optionsize: size of options rcvd
787 * @passive: active or passive flag
788 */
789 static int
irdma_handle_tcp_options(struct irdma_cm_node * cm_node,struct tcphdr * tcph,int optionsize,int passive)790 irdma_handle_tcp_options(struct irdma_cm_node *cm_node,
791 struct tcphdr *tcph, int optionsize,
792 int passive)
793 {
794 u8 *optionsloc = (u8 *)&tcph[1];
795 int ret;
796
797 if (optionsize) {
798 ret = irdma_process_options(cm_node, optionsloc, optionsize,
799 (u32)tcp_get_flags(tcph) & TH_SYN);
800 if (ret) {
801 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
802 "Node %p, Sending Reset\n", cm_node);
803 if (passive)
804 irdma_passive_open_err(cm_node, true);
805 else
806 irdma_active_open_err(cm_node, true);
807 return ret;
808 }
809 }
810
811 cm_node->tcp_cntxt.snd_wnd = ntohs(tcph->th_win)
812 << cm_node->tcp_cntxt.snd_wscale;
813
814 if (cm_node->tcp_cntxt.snd_wnd > cm_node->tcp_cntxt.max_snd_wnd)
815 cm_node->tcp_cntxt.max_snd_wnd = cm_node->tcp_cntxt.snd_wnd;
816
817 return 0;
818 }
819
820 /**
821 * irdma_build_mpa_v1 - build a MPA V1 frame
822 * @cm_node: connection's node
823 * @start_addr: address where to build frame
824 * @mpa_key: to do read0 or write0
825 */
826 static void
irdma_build_mpa_v1(struct irdma_cm_node * cm_node,void * start_addr,u8 mpa_key)827 irdma_build_mpa_v1(struct irdma_cm_node *cm_node, void *start_addr,
828 u8 mpa_key)
829 {
830 struct ietf_mpa_v1 *mpa_frame = start_addr;
831
832 switch (mpa_key) {
833 case MPA_KEY_REQUEST:
834 memcpy(mpa_frame->key, IEFT_MPA_KEY_REQ, IETF_MPA_KEY_SIZE);
835 break;
836 case MPA_KEY_REPLY:
837 memcpy(mpa_frame->key, IEFT_MPA_KEY_REP, IETF_MPA_KEY_SIZE);
838 break;
839 default:
840 break;
841 }
842 mpa_frame->flags = IETF_MPA_FLAGS_CRC;
843 mpa_frame->rev = cm_node->mpa_frame_rev;
844 mpa_frame->priv_data_len = htons(cm_node->pdata.size);
845 }
846
847 /**
848 * irdma_build_mpa_v2 - build a MPA V2 frame
849 * @cm_node: connection's node
850 * @start_addr: buffer start address
851 * @mpa_key: to do read0 or write0
852 */
853 static void
irdma_build_mpa_v2(struct irdma_cm_node * cm_node,void * start_addr,u8 mpa_key)854 irdma_build_mpa_v2(struct irdma_cm_node *cm_node, void *start_addr,
855 u8 mpa_key)
856 {
857 struct ietf_mpa_v2 *mpa_frame = start_addr;
858 struct ietf_rtr_msg *rtr_msg = &mpa_frame->rtr_msg;
859 u16 ctrl_ird, ctrl_ord;
860
861 /* initialize the upper 5 bytes of the frame */
862 irdma_build_mpa_v1(cm_node, start_addr, mpa_key);
863 mpa_frame->flags |= IETF_MPA_V2_FLAG;
864 if (cm_node->iwdev->iw_ooo) {
865 mpa_frame->flags |= IETF_MPA_FLAGS_MARKERS;
866 cm_node->rcv_mark_en = true;
867 }
868 mpa_frame->priv_data_len = cpu_to_be16(be16_to_cpu(mpa_frame->priv_data_len) +
869 IETF_RTR_MSG_SIZE);
870
871 /* initialize RTR msg */
872 if (cm_node->mpav2_ird_ord == IETF_NO_IRD_ORD) {
873 ctrl_ird = IETF_NO_IRD_ORD;
874 ctrl_ord = IETF_NO_IRD_ORD;
875 } else {
876 ctrl_ird = (cm_node->ird_size > IETF_NO_IRD_ORD) ?
877 IETF_NO_IRD_ORD :
878 cm_node->ird_size;
879 ctrl_ord = (cm_node->ord_size > IETF_NO_IRD_ORD) ?
880 IETF_NO_IRD_ORD :
881 cm_node->ord_size;
882 }
883 ctrl_ird |= IETF_PEER_TO_PEER;
884
885 switch (mpa_key) {
886 case MPA_KEY_REQUEST:
887 ctrl_ord |= IETF_RDMA0_WRITE;
888 ctrl_ord |= IETF_RDMA0_READ;
889 break;
890 case MPA_KEY_REPLY:
891 switch (cm_node->send_rdma0_op) {
892 case SEND_RDMA_WRITE_ZERO:
893 ctrl_ord |= IETF_RDMA0_WRITE;
894 break;
895 case SEND_RDMA_READ_ZERO:
896 ctrl_ord |= IETF_RDMA0_READ;
897 break;
898 }
899 break;
900 default:
901 break;
902 }
903 rtr_msg->ctrl_ird = htons(ctrl_ird);
904 rtr_msg->ctrl_ord = htons(ctrl_ord);
905 }
906
907 /**
908 * irdma_cm_build_mpa_frame - build mpa frame for mpa version 1 or version 2
909 * @cm_node: connection's node
910 * @mpa: mpa: data buffer
911 * @mpa_key: to do read0 or write0
912 */
913 static int
irdma_cm_build_mpa_frame(struct irdma_cm_node * cm_node,struct irdma_kmem_info * mpa,u8 mpa_key)914 irdma_cm_build_mpa_frame(struct irdma_cm_node *cm_node,
915 struct irdma_kmem_info *mpa, u8 mpa_key)
916 {
917 int hdr_len = 0;
918
919 switch (cm_node->mpa_frame_rev) {
920 case IETF_MPA_V1:
921 hdr_len = sizeof(struct ietf_mpa_v1);
922 irdma_build_mpa_v1(cm_node, mpa->addr, mpa_key);
923 break;
924 case IETF_MPA_V2:
925 hdr_len = sizeof(struct ietf_mpa_v2);
926 irdma_build_mpa_v2(cm_node, mpa->addr, mpa_key);
927 break;
928 default:
929 break;
930 }
931
932 return hdr_len;
933 }
934
935 /**
936 * irdma_send_mpa_request - active node send mpa request to passive node
937 * @cm_node: connection's node
938 */
939 static int
irdma_send_mpa_request(struct irdma_cm_node * cm_node)940 irdma_send_mpa_request(struct irdma_cm_node *cm_node)
941 {
942 struct irdma_puda_buf *sqbuf;
943
944 cm_node->mpa_hdr.addr = &cm_node->mpa_v2_frame;
945 cm_node->mpa_hdr.size = irdma_cm_build_mpa_frame(cm_node,
946 &cm_node->mpa_hdr,
947 MPA_KEY_REQUEST);
948 if (!cm_node->mpa_hdr.size) {
949 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
950 "mpa size = %d\n", cm_node->mpa_hdr.size);
951 return -EINVAL;
952 }
953
954 sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL,
955 &cm_node->mpa_hdr,
956 &cm_node->pdata, SET_ACK);
957 if (!sqbuf)
958 return -ENOMEM;
959
960 return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
961 0);
962 }
963
964 /**
965 * irdma_send_mpa_reject -
966 * @cm_node: connection's node
967 * @pdata: reject data for connection
968 * @plen: length of reject data
969 */
970 static int
irdma_send_mpa_reject(struct irdma_cm_node * cm_node,const void * pdata,u8 plen)971 irdma_send_mpa_reject(struct irdma_cm_node *cm_node,
972 const void *pdata, u8 plen)
973 {
974 struct irdma_puda_buf *sqbuf;
975 struct irdma_mpa_priv_info priv_info;
976
977 cm_node->mpa_hdr.addr = &cm_node->mpa_v2_frame;
978 cm_node->mpa_hdr.size = irdma_cm_build_mpa_frame(cm_node,
979 &cm_node->mpa_hdr,
980 MPA_KEY_REPLY);
981
982 cm_node->mpa_v2_frame.flags |= IETF_MPA_FLAGS_REJECT;
983 priv_info.addr = pdata;
984 priv_info.size = plen;
985
986 sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL,
987 &cm_node->mpa_hdr, &priv_info,
988 SET_ACK | SET_FIN);
989 if (!sqbuf)
990 return -ENOMEM;
991
992 cm_node->state = IRDMA_CM_STATE_FIN_WAIT1;
993
994 return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
995 0);
996 }
997
998 /**
999 * irdma_negotiate_mpa_v2_ird_ord - negotiate MPAv2 IRD/ORD
1000 * @cm_node: connection's node
1001 * @buf: Data pointer
1002 */
1003 static int
irdma_negotiate_mpa_v2_ird_ord(struct irdma_cm_node * cm_node,u8 * buf)1004 irdma_negotiate_mpa_v2_ird_ord(struct irdma_cm_node *cm_node,
1005 u8 *buf)
1006 {
1007 struct ietf_mpa_v2 *mpa_v2_frame;
1008 struct ietf_rtr_msg *rtr_msg;
1009 u16 ird_size;
1010 u16 ord_size;
1011 u16 ctrl_ord;
1012 u16 ctrl_ird;
1013
1014 mpa_v2_frame = (struct ietf_mpa_v2 *)buf;
1015 rtr_msg = &mpa_v2_frame->rtr_msg;
1016
1017 /* parse rtr message */
1018 ctrl_ord = ntohs(rtr_msg->ctrl_ord);
1019 ctrl_ird = ntohs(rtr_msg->ctrl_ird);
1020 ird_size = ctrl_ird & IETF_NO_IRD_ORD;
1021 ord_size = ctrl_ord & IETF_NO_IRD_ORD;
1022
1023 if (!(ctrl_ird & IETF_PEER_TO_PEER))
1024 return -EOPNOTSUPP;
1025
1026 if (ird_size == IETF_NO_IRD_ORD || ord_size == IETF_NO_IRD_ORD) {
1027 cm_node->mpav2_ird_ord = IETF_NO_IRD_ORD;
1028 goto negotiate_done;
1029 }
1030
1031 if (cm_node->state != IRDMA_CM_STATE_MPAREQ_SENT) {
1032 /* responder */
1033 if (!ord_size && (ctrl_ord & IETF_RDMA0_READ))
1034 cm_node->ird_size = 1;
1035 if (cm_node->ord_size > ird_size)
1036 cm_node->ord_size = ird_size;
1037 } else {
1038 /* initiator */
1039 if (!ird_size && (ctrl_ord & IETF_RDMA0_READ))
1040 /* Remote peer doesn't support RDMA0_READ */
1041 return -EOPNOTSUPP;
1042
1043 if (cm_node->ord_size > ird_size)
1044 cm_node->ord_size = ird_size;
1045
1046 if (cm_node->ird_size < ord_size)
1047 /* no resources available */
1048 return -EINVAL;
1049 }
1050
1051 negotiate_done:
1052 if (ctrl_ord & IETF_RDMA0_READ)
1053 cm_node->send_rdma0_op = SEND_RDMA_READ_ZERO;
1054 else if (ctrl_ord & IETF_RDMA0_WRITE)
1055 cm_node->send_rdma0_op = SEND_RDMA_WRITE_ZERO;
1056 else
1057 /* Not supported RDMA0 operation */
1058 return -EOPNOTSUPP;
1059
1060 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1061 "MPAV2 Negotiated ORD: %d, IRD: %d\n", cm_node->ord_size,
1062 cm_node->ird_size);
1063 return 0;
1064 }
1065
1066 /**
1067 * irdma_parse_mpa - process an IETF MPA frame
1068 * @cm_node: connection's node
1069 * @buf: Data pointer
1070 * @type: to return accept or reject
1071 * @len: Len of mpa buffer
1072 */
1073 static int
irdma_parse_mpa(struct irdma_cm_node * cm_node,u8 * buf,u32 * type,u32 len)1074 irdma_parse_mpa(struct irdma_cm_node *cm_node, u8 *buf, u32 *type,
1075 u32 len)
1076 {
1077 struct ietf_mpa_v1 *mpa_frame;
1078 int mpa_hdr_len, priv_data_len, ret;
1079
1080 *type = IRDMA_MPA_REQUEST_ACCEPT;
1081
1082 if (len < sizeof(*mpa_frame)) {
1083 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1084 "ietf buffer small (%x)\n", len);
1085 return -EINVAL;
1086 }
1087
1088 mpa_frame = (struct ietf_mpa_v1 *)buf;
1089 mpa_hdr_len = sizeof(*mpa_frame);
1090 priv_data_len = ntohs(mpa_frame->priv_data_len);
1091
1092 if (priv_data_len > IETF_MAX_PRIV_DATA_LEN) {
1093 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1094 "private_data too big %d\n", priv_data_len);
1095 return -EOVERFLOW;
1096 }
1097
1098 if (mpa_frame->rev != IETF_MPA_V1 && mpa_frame->rev != IETF_MPA_V2) {
1099 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1100 "unsupported mpa rev = %d\n", mpa_frame->rev);
1101 return -EINVAL;
1102 }
1103
1104 if (mpa_frame->rev > cm_node->mpa_frame_rev) {
1105 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1106 "rev %d\n", mpa_frame->rev);
1107 return -EINVAL;
1108 }
1109
1110 cm_node->mpa_frame_rev = mpa_frame->rev;
1111 if (cm_node->state != IRDMA_CM_STATE_MPAREQ_SENT) {
1112 if (memcmp(mpa_frame->key, IEFT_MPA_KEY_REQ,
1113 IETF_MPA_KEY_SIZE)) {
1114 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1115 "Unexpected MPA Key received\n");
1116 return -EINVAL;
1117 }
1118 } else {
1119 if (memcmp(mpa_frame->key, IEFT_MPA_KEY_REP,
1120 IETF_MPA_KEY_SIZE)) {
1121 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1122 "Unexpected MPA Key received\n");
1123 return -EINVAL;
1124 }
1125 }
1126
1127 if (priv_data_len + mpa_hdr_len > len) {
1128 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1129 "ietf buffer len(%x + %x != %x)\n", priv_data_len,
1130 mpa_hdr_len, len);
1131 return -EOVERFLOW;
1132 }
1133
1134 if (len > IRDMA_MAX_CM_BUF) {
1135 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1136 "ietf buffer large len = %d\n", len);
1137 return -EOVERFLOW;
1138 }
1139
1140 switch (mpa_frame->rev) {
1141 case IETF_MPA_V2:
1142 mpa_hdr_len += IETF_RTR_MSG_SIZE;
1143 ret = irdma_negotiate_mpa_v2_ird_ord(cm_node, buf);
1144 if (ret)
1145 return ret;
1146 break;
1147 case IETF_MPA_V1:
1148 default:
1149 break;
1150 }
1151
1152 memcpy(cm_node->pdata_buf, buf + mpa_hdr_len, priv_data_len);
1153 cm_node->pdata.size = priv_data_len;
1154
1155 if (mpa_frame->flags & IETF_MPA_FLAGS_REJECT)
1156 *type = IRDMA_MPA_REQUEST_REJECT;
1157
1158 if (mpa_frame->flags & IETF_MPA_FLAGS_MARKERS)
1159 cm_node->snd_mark_en = true;
1160
1161 return 0;
1162 }
1163
1164 /**
1165 * irdma_schedule_cm_timer
1166 * @cm_node: connection's node
1167 * @sqbuf: buffer to send
1168 * @type: if it is send or close
1169 * @send_retrans: if rexmits to be done
1170 * @close_when_complete: is cm_node to be removed
1171 *
1172 * note - cm_node needs to be protected before calling this. Encase in:
1173 * irdma_rem_ref_cm_node(cm_core, cm_node);
1174 * irdma_schedule_cm_timer(...)
1175 * atomic_inc(&cm_node->refcnt);
1176 */
1177 int
irdma_schedule_cm_timer(struct irdma_cm_node * cm_node,struct irdma_puda_buf * sqbuf,enum irdma_timer_type type,int send_retrans,int close_when_complete)1178 irdma_schedule_cm_timer(struct irdma_cm_node *cm_node,
1179 struct irdma_puda_buf *sqbuf,
1180 enum irdma_timer_type type, int send_retrans,
1181 int close_when_complete)
1182 {
1183 struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
1184 struct irdma_cm_core *cm_core = cm_node->cm_core;
1185 struct irdma_timer_entry *new_send;
1186 u32 was_timer_set;
1187 unsigned long flags;
1188
1189 new_send = kzalloc(sizeof(*new_send), GFP_ATOMIC);
1190 if (!new_send) {
1191 if (type != IRDMA_TIMER_TYPE_CLOSE)
1192 irdma_free_sqbuf(vsi, sqbuf);
1193 return -ENOMEM;
1194 }
1195
1196 new_send->retrycount = IRDMA_DEFAULT_RETRYS;
1197 new_send->retranscount = IRDMA_DEFAULT_RETRANS;
1198 new_send->sqbuf = sqbuf;
1199 new_send->timetosend = jiffies;
1200 new_send->type = type;
1201 new_send->send_retrans = send_retrans;
1202 new_send->close_when_complete = close_when_complete;
1203
1204 if (type == IRDMA_TIMER_TYPE_CLOSE) {
1205 new_send->timetosend += (HZ / 10);
1206 if (cm_node->close_entry) {
1207 kfree(new_send);
1208 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1209 "already close entry\n");
1210 return -EINVAL;
1211 }
1212
1213 cm_node->close_entry = new_send;
1214 } else { /* type == IRDMA_TIMER_TYPE_SEND */
1215 spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
1216 cm_node->send_entry = new_send;
1217 atomic_inc(&cm_node->refcnt);
1218 spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
1219 new_send->timetosend = jiffies + IRDMA_RETRY_TIMEOUT;
1220
1221 atomic_inc(&sqbuf->refcount);
1222 irdma_puda_send_buf(vsi->ilq, sqbuf);
1223 if (!send_retrans) {
1224 irdma_cleanup_retrans_entry(cm_node);
1225 if (close_when_complete)
1226 irdma_rem_ref_cm_node(cm_node);
1227 return 0;
1228 }
1229 }
1230
1231 spin_lock_irqsave(&cm_core->ht_lock, flags);
1232 was_timer_set = timer_pending(&cm_core->tcp_timer);
1233
1234 if (!was_timer_set) {
1235 cm_core->tcp_timer.expires = new_send->timetosend;
1236 add_timer(&cm_core->tcp_timer);
1237 }
1238 spin_unlock_irqrestore(&cm_core->ht_lock, flags);
1239
1240 return 0;
1241 }
1242
1243 /**
1244 * irdma_retrans_expired - Could not rexmit the packet
1245 * @cm_node: connection's node
1246 */
1247 static void
irdma_retrans_expired(struct irdma_cm_node * cm_node)1248 irdma_retrans_expired(struct irdma_cm_node *cm_node)
1249 {
1250 enum irdma_cm_node_state state = cm_node->state;
1251
1252 cm_node->state = IRDMA_CM_STATE_CLOSED;
1253 switch (state) {
1254 case IRDMA_CM_STATE_SYN_RCVD:
1255 case IRDMA_CM_STATE_CLOSING:
1256 irdma_rem_ref_cm_node(cm_node);
1257 break;
1258 case IRDMA_CM_STATE_FIN_WAIT1:
1259 case IRDMA_CM_STATE_LAST_ACK:
1260 irdma_send_reset(cm_node);
1261 break;
1262 default:
1263 atomic_inc(&cm_node->refcnt);
1264 irdma_send_reset(cm_node);
1265 irdma_create_event(cm_node, IRDMA_CM_EVENT_ABORTED);
1266 break;
1267 }
1268 }
1269
1270 /**
1271 * irdma_handle_close_entry - for handling retry/timeouts
1272 * @cm_node: connection's node
1273 * @rem_node: flag for remove cm_node
1274 */
1275 static void
irdma_handle_close_entry(struct irdma_cm_node * cm_node,u32 rem_node)1276 irdma_handle_close_entry(struct irdma_cm_node *cm_node,
1277 u32 rem_node)
1278 {
1279 struct irdma_timer_entry *close_entry = cm_node->close_entry;
1280 struct irdma_qp *iwqp;
1281 unsigned long flags;
1282
1283 if (!close_entry)
1284 return;
1285 iwqp = (struct irdma_qp *)close_entry->sqbuf;
1286 if (iwqp) {
1287 spin_lock_irqsave(&iwqp->lock, flags);
1288 if (iwqp->cm_id) {
1289 iwqp->hw_tcp_state = IRDMA_TCP_STATE_CLOSED;
1290 iwqp->hw_iwarp_state = IRDMA_QP_STATE_ERROR;
1291 iwqp->last_aeq = IRDMA_AE_RESET_SENT;
1292 iwqp->ibqp_state = IB_QPS_ERR;
1293 spin_unlock_irqrestore(&iwqp->lock, flags);
1294 irdma_cm_disconn(iwqp);
1295 } else {
1296 spin_unlock_irqrestore(&iwqp->lock, flags);
1297 }
1298 } else if (rem_node) {
1299 /* TIME_WAIT state */
1300 irdma_rem_ref_cm_node(cm_node);
1301 }
1302
1303 kfree(close_entry);
1304 cm_node->close_entry = NULL;
1305 }
1306
1307 /**
1308 * irdma_cm_timer_tick - system's timer expired callback
1309 * @t: Pointer to timer_list
1310 */
1311 static void
irdma_cm_timer_tick(struct timer_list * t)1312 irdma_cm_timer_tick(struct timer_list *t)
1313 {
1314 unsigned long nexttimeout = jiffies + IRDMA_LONG_TIME;
1315 struct irdma_cm_node *cm_node;
1316 struct irdma_timer_entry *send_entry, *close_entry;
1317 struct list_head *list_core_temp;
1318 struct list_head *list_node;
1319 struct irdma_cm_core *cm_core = timer_container_of(cm_core, t, tcp_timer);
1320 struct irdma_sc_vsi *vsi;
1321 u32 settimer = 0;
1322 unsigned long timetosend;
1323 unsigned long flags;
1324 struct list_head timer_list;
1325
1326 INIT_LIST_HEAD(&timer_list);
1327
1328 rcu_read_lock();
1329 irdma_timer_list_prep(cm_core, &timer_list);
1330 rcu_read_unlock();
1331
1332 list_for_each_safe(list_node, list_core_temp, &timer_list) {
1333 cm_node = container_of(list_node, struct irdma_cm_node,
1334 timer_entry);
1335 close_entry = cm_node->close_entry;
1336
1337 if (close_entry) {
1338 if (time_after(close_entry->timetosend, jiffies)) {
1339 if (nexttimeout > close_entry->timetosend ||
1340 !settimer) {
1341 nexttimeout = close_entry->timetosend;
1342 settimer = 1;
1343 }
1344 } else {
1345 irdma_handle_close_entry(cm_node, 1);
1346 }
1347 }
1348
1349 spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
1350
1351 send_entry = cm_node->send_entry;
1352 if (!send_entry)
1353 goto done;
1354 if (time_after(send_entry->timetosend, jiffies)) {
1355 if (cm_node->state != IRDMA_CM_STATE_OFFLOADED) {
1356 if (nexttimeout > send_entry->timetosend ||
1357 !settimer) {
1358 nexttimeout = send_entry->timetosend;
1359 settimer = 1;
1360 }
1361 } else {
1362 irdma_free_retrans_entry(cm_node);
1363 }
1364 goto done;
1365 }
1366
1367 if (cm_node->state == IRDMA_CM_STATE_OFFLOADED ||
1368 cm_node->state == IRDMA_CM_STATE_CLOSED) {
1369 irdma_free_retrans_entry(cm_node);
1370 goto done;
1371 }
1372
1373 if (!send_entry->retranscount || !send_entry->retrycount) {
1374 irdma_free_retrans_entry(cm_node);
1375
1376 spin_unlock_irqrestore(&cm_node->retrans_list_lock,
1377 flags);
1378 irdma_retrans_expired(cm_node);
1379 cm_node->state = IRDMA_CM_STATE_CLOSED;
1380 spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
1381 goto done;
1382 }
1383 spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
1384
1385 vsi = &cm_node->iwdev->vsi;
1386 if (!cm_node->ack_rcvd) {
1387 atomic_inc(&send_entry->sqbuf->refcount);
1388 irdma_puda_send_buf(vsi->ilq, send_entry->sqbuf);
1389 cm_node->cm_core->stats_pkt_retrans++;
1390 }
1391
1392 spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
1393 if (send_entry->send_retrans) {
1394 send_entry->retranscount--;
1395 timetosend = (IRDMA_RETRY_TIMEOUT <<
1396 (IRDMA_DEFAULT_RETRANS -
1397 send_entry->retranscount));
1398
1399 send_entry->timetosend = jiffies +
1400 min(timetosend, IRDMA_MAX_TIMEOUT);
1401 if (nexttimeout > send_entry->timetosend || !settimer) {
1402 nexttimeout = send_entry->timetosend;
1403 settimer = 1;
1404 }
1405 } else {
1406 int close_when_complete;
1407
1408 close_when_complete = send_entry->close_when_complete;
1409 irdma_free_retrans_entry(cm_node);
1410 if (close_when_complete)
1411 irdma_rem_ref_cm_node(cm_node);
1412 }
1413 done:
1414 spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
1415 irdma_rem_ref_cm_node(cm_node);
1416 }
1417
1418 if (settimer) {
1419 spin_lock_irqsave(&cm_core->ht_lock, flags);
1420 if (!timer_pending(&cm_core->tcp_timer)) {
1421 cm_core->tcp_timer.expires = nexttimeout;
1422 add_timer(&cm_core->tcp_timer);
1423 }
1424 spin_unlock_irqrestore(&cm_core->ht_lock, flags);
1425 }
1426 }
1427
1428 /**
1429 * irdma_send_syn - send SYN packet
1430 * @cm_node: connection's node
1431 * @sendack: flag to set ACK bit or not
1432 */
1433 int
irdma_send_syn(struct irdma_cm_node * cm_node,u32 sendack)1434 irdma_send_syn(struct irdma_cm_node *cm_node, u32 sendack)
1435 {
1436 struct irdma_puda_buf *sqbuf;
1437 int flags = SET_SYN;
1438 char optionsbuf[sizeof(struct option_mss) +
1439 sizeof(struct option_windowscale) +
1440 sizeof(struct option_base) + TCP_OPTIONS_PADDING];
1441 struct irdma_kmem_info opts;
1442 int optionssize = 0;
1443
1444 /* Sending MSS option */
1445 union all_known_options *options;
1446
1447 opts.addr = optionsbuf;
1448 if (!cm_node)
1449 return -EINVAL;
1450
1451 options = (union all_known_options *)&optionsbuf[optionssize];
1452 options->mss.optionnum = OPTION_NUM_MSS;
1453 options->mss.len = sizeof(struct option_mss);
1454 options->mss.mss = htons(cm_node->tcp_cntxt.mss);
1455 optionssize += sizeof(struct option_mss);
1456
1457 options = (union all_known_options *)&optionsbuf[optionssize];
1458 options->windowscale.optionnum = OPTION_NUM_WINDOW_SCALE;
1459 options->windowscale.len = sizeof(struct option_windowscale);
1460 options->windowscale.shiftcount = cm_node->tcp_cntxt.rcv_wscale;
1461 optionssize += sizeof(struct option_windowscale);
1462 options = (union all_known_options *)&optionsbuf[optionssize];
1463 options->eol = OPTION_NUM_EOL;
1464 optionssize += 1;
1465
1466 if (sendack)
1467 flags |= SET_ACK;
1468
1469 opts.size = optionssize;
1470
1471 sqbuf = cm_node->cm_core->form_cm_frame(cm_node, &opts, NULL, NULL,
1472 flags);
1473 if (!sqbuf)
1474 return -ENOMEM;
1475
1476 return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
1477 0);
1478 }
1479
1480 /**
1481 * irdma_send_ack - Send ACK packet
1482 * @cm_node: connection's node
1483 */
1484 void
irdma_send_ack(struct irdma_cm_node * cm_node)1485 irdma_send_ack(struct irdma_cm_node *cm_node)
1486 {
1487 struct irdma_puda_buf *sqbuf;
1488 struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
1489
1490 sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL, NULL, NULL,
1491 SET_ACK);
1492 if (sqbuf)
1493 irdma_puda_send_buf(vsi->ilq, sqbuf);
1494 }
1495
1496 /**
1497 * irdma_send_fin - Send FIN pkt
1498 * @cm_node: connection's node
1499 */
1500 static int
irdma_send_fin(struct irdma_cm_node * cm_node)1501 irdma_send_fin(struct irdma_cm_node *cm_node)
1502 {
1503 struct irdma_puda_buf *sqbuf;
1504
1505 sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL, NULL, NULL,
1506 SET_ACK | SET_FIN);
1507 if (!sqbuf)
1508 return -ENOMEM;
1509
1510 return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
1511 0);
1512 }
1513
1514 /**
1515 * irdma_find_listener - find a cm node listening on this addr-port pair
1516 * @cm_core: cm's core
1517 * @dst_addr: listener ip addr
1518 * @ipv4: flag indicating IPv4 when true
1519 * @dst_port: listener tcp port num
1520 * @vlan_id: virtual LAN ID
1521 * @listener_state: state to match with listen node's
1522 */
1523 static struct irdma_cm_listener *
irdma_find_listener(struct irdma_cm_core * cm_core,u32 * dst_addr,bool ipv4,u16 dst_port,u16 vlan_id,enum irdma_cm_listener_state listener_state)1524 irdma_find_listener(struct irdma_cm_core *cm_core, u32 *dst_addr, bool ipv4, u16 dst_port,
1525 u16 vlan_id, enum irdma_cm_listener_state listener_state)
1526 {
1527 struct irdma_cm_listener *listen_node;
1528 static const u32 ip_zero[4] = {0, 0, 0, 0};
1529 u32 listen_addr[4];
1530 u16 listen_port;
1531 unsigned long flags;
1532
1533 /* walk list and find cm_node associated with this session ID */
1534 spin_lock_irqsave(&cm_core->listen_list_lock, flags);
1535 list_for_each_entry(listen_node, &cm_core->listen_list, list) {
1536 memcpy(listen_addr, listen_node->loc_addr, sizeof(listen_addr));
1537 listen_port = listen_node->loc_port;
1538 if (listen_node->ipv4 != ipv4 || listen_port != dst_port ||
1539 !(listener_state & listen_node->listener_state))
1540 continue;
1541 /* compare node pair, return node handle if a match */
1542 if (!memcmp(listen_addr, ip_zero, sizeof(listen_addr)) ||
1543 (!memcmp(listen_addr, dst_addr, sizeof(listen_addr)) &&
1544 vlan_id == listen_node->vlan_id)) {
1545 atomic_inc(&listen_node->refcnt);
1546 spin_unlock_irqrestore(&cm_core->listen_list_lock,
1547 flags);
1548 return listen_node;
1549 }
1550 }
1551 spin_unlock_irqrestore(&cm_core->listen_list_lock, flags);
1552
1553 return NULL;
1554 }
1555
1556 /**
1557 * irdma_del_multiple_qhash - Remove qhash and child listens
1558 * @iwdev: iWarp device
1559 * @cm_info: CM info for parent listen node
1560 * @cm_parent_listen_node: The parent listen node
1561 */
1562 static int
irdma_del_multiple_qhash(struct irdma_device * iwdev,struct irdma_cm_info * cm_info,struct irdma_cm_listener * cm_parent_listen_node)1563 irdma_del_multiple_qhash(struct irdma_device *iwdev,
1564 struct irdma_cm_info *cm_info,
1565 struct irdma_cm_listener *cm_parent_listen_node)
1566 {
1567 struct irdma_cm_listener *child_listen_node;
1568 struct list_head *pos, *tpos;
1569 unsigned long flags;
1570 int ret = -EINVAL;
1571
1572 spin_lock_irqsave(&iwdev->cm_core.listen_list_lock, flags);
1573 list_for_each_safe(pos, tpos,
1574 &cm_parent_listen_node->child_listen_list) {
1575 child_listen_node = list_entry(pos, struct irdma_cm_listener,
1576 child_listen_list);
1577 if (child_listen_node->ipv4)
1578 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1579 "removing child listen for IP=%x, port=%d, vlan=%d\n",
1580 child_listen_node->loc_addr[0],
1581 child_listen_node->loc_port,
1582 child_listen_node->vlan_id);
1583 else
1584 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1585 "removing child listen for IP=%x:%x:%x:%x, port=%d, vlan=%d\n",
1586 IRDMA_PRINT_IP6(child_listen_node->loc_addr),
1587 child_listen_node->loc_port,
1588 child_listen_node->vlan_id);
1589 list_del(pos);
1590 memcpy(cm_info->loc_addr, child_listen_node->loc_addr,
1591 sizeof(cm_info->loc_addr));
1592 cm_info->vlan_id = child_listen_node->vlan_id;
1593 if (child_listen_node->qhash_set) {
1594 ret = irdma_manage_qhash(iwdev, cm_info,
1595 IRDMA_QHASH_TYPE_TCP_SYN,
1596 IRDMA_QHASH_MANAGE_TYPE_DELETE,
1597 NULL, false);
1598 child_listen_node->qhash_set = false;
1599 } else {
1600 ret = 0;
1601 }
1602 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1603 "Child listen node freed = %p\n",
1604 child_listen_node);
1605 kfree(child_listen_node);
1606 cm_parent_listen_node->cm_core->stats_listen_nodes_destroyed++;
1607 }
1608 spin_unlock_irqrestore(&iwdev->cm_core.listen_list_lock, flags);
1609
1610 return ret;
1611 }
1612
irdma_iw_get_vlan_prio(u32 * loc_addr,u8 prio,bool ipv4)1613 static u8 irdma_iw_get_vlan_prio(u32 *loc_addr, u8 prio, bool ipv4)
1614 {
1615 return prio;
1616 }
1617
1618 /**
1619 * irdma_get_vlan_mac_ipv6 - Get the vlan and mac for an IPv6
1620 * address
1621 * @addr: local IPv6 address
1622 * @vlan_id: vlan id for the given IPv6 address
1623 * @mac: mac address for the given IPv6 address
1624 *
1625 * Returns the net_device of the IPv6 address and also sets the
1626 * vlan id and mac for that address.
1627 */
1628 void
irdma_get_vlan_mac_ipv6(struct iw_cm_id * cm_id,u32 * addr,u16 * vlan_id,u8 * mac)1629 irdma_get_vlan_mac_ipv6(struct iw_cm_id *cm_id, u32 *addr, u16 *vlan_id, u8 *mac)
1630 {
1631 if_t ip_dev = NULL;
1632 struct in6_addr laddr6;
1633 struct vnet *vnet = &init_net;
1634 struct ifaddr *ifa;
1635 u16 scope_id = 0;
1636
1637 irdma_copy_ip_htonl(laddr6.__u6_addr.__u6_addr32, addr);
1638 if (vlan_id)
1639 *vlan_id = 0xFFFF; /* Match rdma_vlan_dev_vlan_id() */
1640 if (mac)
1641 eth_zero_addr(mac);
1642
1643 if (IN6_IS_SCOPE_LINKLOCAL(&laddr6) ||
1644 IN6_IS_ADDR_MC_INTFACELOCAL(&laddr6))
1645 scope_id = ntohs(laddr6.__u6_addr.__u6_addr16[1]);
1646
1647 #ifdef VIMAGE
1648 vnet = irdma_cmid_to_vnet(cm_id);
1649 #endif
1650 ip_dev = ip6_ifp_find(vnet, laddr6, scope_id);
1651 if (ip_dev) {
1652 if (vlan_id)
1653 *vlan_id = rdma_vlan_dev_vlan_id(ip_dev);
1654 ifa = if_getifaddr(ip_dev);
1655 if (ifa && ifa->ifa_addr && mac)
1656 ether_addr_copy(mac, if_getlladdr(ip_dev));
1657 }
1658 }
1659
1660 /**
1661 * irdma_get_vlan_ipv4 - Returns the vlan_id for IPv4 address
1662 * @addr: local IPv4 address
1663 */
1664 u16
irdma_get_vlan_ipv4(struct iw_cm_id * cm_id,u32 * addr)1665 irdma_get_vlan_ipv4(struct iw_cm_id *cm_id, u32 *addr)
1666 {
1667 u16 vlan_id = 0xFFFF;
1668
1669 #ifdef INET
1670 if_t netdev;
1671 struct vnet *vnet = &init_net;
1672
1673 #ifdef VIMAGE
1674 vnet = irdma_cmid_to_vnet(cm_id);
1675 #endif
1676 netdev = ip_ifp_find(vnet, htonl(addr[0]));
1677 if (netdev) {
1678 vlan_id = rdma_vlan_dev_vlan_id(netdev);
1679 dev_put(netdev);
1680 }
1681 #endif
1682
1683 return vlan_id;
1684 }
1685
1686 /**
1687 * irdma_add_mqh_ifa_cb - Adds multiple qhashes for IPv4/IPv6
1688 * @arg: Calback argument structure from irdma_add_mqh
1689 * @ifa: Current address to compute against
1690 * @count: Current cumulative output of all callbacks in this iteration
1691 *
1692 * Adds a qhash and a child listen node for a single IPv4/IPv6 address
1693 * on the adapter and adds the associated qhash filter
1694 */
1695 static u_int
irdma_add_mqh_ifa_cb(void * arg,struct ifaddr * ifa,u_int count)1696 irdma_add_mqh_ifa_cb(void *arg, struct ifaddr *ifa, u_int count)
1697 {
1698 struct irdma_add_mqh_cbs *cbs = arg;
1699 struct irdma_cm_listener *child_listen_node;
1700 struct irdma_cm_info *cm_info = cbs->cm_info;
1701 struct irdma_device *iwdev = cbs->iwdev;
1702 struct irdma_cm_listener *cm_parent_listen_node = cbs->cm_listen_node;
1703 if_t ip_dev = ifa->ifa_ifp;
1704 unsigned long flags;
1705 int ret;
1706
1707 if (count)
1708 return 0;
1709
1710 child_listen_node = kzalloc(sizeof(*child_listen_node), GFP_ATOMIC);
1711 if (!child_listen_node) {
1712 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1713 "listener memory allocation\n");
1714 return -ENOMEM;
1715 }
1716
1717 memcpy(child_listen_node, cm_parent_listen_node,
1718 sizeof(*child_listen_node));
1719 cm_info->vlan_id = rdma_vlan_dev_vlan_id(ip_dev);
1720 child_listen_node->vlan_id = cm_info->vlan_id;
1721 if (cm_info->ipv4) {
1722 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1723 "Allocating child CM Listener forIP=%x, vlan_id=%d, MAC=%x:%x:%x:%x:%x:%x\n",
1724 ((struct sockaddr_in *)&ifa->ifa_addr)->sin_addr.s_addr,
1725 rdma_vlan_dev_vlan_id(ip_dev),
1726 if_getlladdr(ip_dev)[0], if_getlladdr(ip_dev)[1],
1727 if_getlladdr(ip_dev)[2], if_getlladdr(ip_dev)[3],
1728 if_getlladdr(ip_dev)[4], if_getlladdr(ip_dev)[5]);
1729 child_listen_node->loc_addr[0] =
1730 ntohl(((struct sockaddr_in *)ifa->ifa_addr)->sin_addr.s_addr);
1731 } else {
1732 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1733 "IP=%x:%x:%x:%x, vlan_id=%d, MAC=%x:%x:%x:%x:%x:%x\n",
1734 IRDMA_PRINT_IP6(&((struct sockaddr_in6 *)ifa->ifa_addr)->sin6_addr),
1735 rdma_vlan_dev_vlan_id(ip_dev),
1736 if_getlladdr(ip_dev)[0], if_getlladdr(ip_dev)[1],
1737 if_getlladdr(ip_dev)[2], if_getlladdr(ip_dev)[3],
1738 if_getlladdr(ip_dev)[4], if_getlladdr(ip_dev)[5]);
1739 irdma_copy_ip_ntohl(child_listen_node->loc_addr,
1740 ((struct sockaddr_in6 *)ifa->ifa_addr)->sin6_addr.__u6_addr.__u6_addr32);
1741 }
1742 memcpy(cm_info->loc_addr, child_listen_node->loc_addr,
1743 sizeof(cm_info->loc_addr));
1744 if (!iwdev->vsi.dscp_mode)
1745 cm_info->user_pri =
1746 irdma_iw_get_vlan_prio(child_listen_node->loc_addr,
1747 cm_info->user_pri,
1748 cm_info->ipv4);
1749 ret = irdma_add_qhash_wait_no_lock(iwdev, cm_info);
1750 if (ret) {
1751 kfree(child_listen_node);
1752 return ret;
1753 }
1754
1755 child_listen_node->qhash_set = true;
1756 spin_lock_irqsave(&iwdev->cm_core.listen_list_lock, flags);
1757 list_add(&child_listen_node->child_listen_list,
1758 &cm_parent_listen_node->child_listen_list);
1759 spin_unlock_irqrestore(&iwdev->cm_core.listen_list_lock, flags);
1760 cm_parent_listen_node->cm_core->stats_listen_nodes_created++;
1761
1762 return 0;
1763 }
1764
1765 /**
1766 * irdma_add_mqh - Adds multiple qhashes
1767 * @iwdev: iWarp device
1768 * @cm_info: CM info for parent listen node
1769 * @cm_listen_node: The parent listen node
1770 */
1771 static int
irdma_add_mqh(struct irdma_device * iwdev,struct irdma_cm_info * cm_info,struct irdma_cm_listener * cm_listen_node)1772 irdma_add_mqh(struct irdma_device *iwdev,
1773 struct irdma_cm_info *cm_info,
1774 struct irdma_cm_listener *cm_listen_node)
1775 {
1776 struct epoch_tracker et;
1777 struct irdma_add_mqh_cbs cbs;
1778 struct if_iter iter;
1779 if_t ifp;
1780 int err = -ENOENT;
1781
1782 cbs.iwdev = iwdev;
1783 cbs.cm_info = cm_info;
1784 cbs.cm_listen_node = cm_listen_node;
1785
1786 VNET_ITERATOR_DECL(vnet_iter);
1787
1788 VNET_LIST_RLOCK();
1789 NET_EPOCH_ENTER(et);
1790 VNET_FOREACH(vnet_iter) {
1791 CURVNET_SET_QUIET(vnet_iter);
1792 for (ifp = if_iter_start(&iter); ifp != NULL; ifp = if_iter_next(&iter)) {
1793 if (!(if_getflags(ifp) & IFF_UP))
1794 continue;
1795
1796 if (((rdma_vlan_dev_vlan_id(ifp) >= VLAN_N_VID) ||
1797 (rdma_vlan_dev_real_dev(ifp) != iwdev->netdev)) &&
1798 ifp != iwdev->netdev)
1799 continue;
1800
1801 if_addr_rlock(ifp);
1802 if (cm_info->ipv4)
1803 err = if_foreach_addr_type(ifp, AF_INET, irdma_add_mqh_ifa_cb, &cbs);
1804 else
1805 err = if_foreach_addr_type(ifp, AF_INET6, irdma_add_mqh_ifa_cb, &cbs);
1806 if_addr_runlock(ifp);
1807 }
1808 if_iter_finish(&iter);
1809 CURVNET_RESTORE();
1810 }
1811 NET_EPOCH_EXIT(et);
1812 VNET_LIST_RUNLOCK();
1813
1814 return err;
1815 }
1816
1817 /**
1818 * irdma_reset_list_prep - add connection nodes slated for reset to list
1819 * @cm_core: cm's core
1820 * @listener: pointer to listener node
1821 * @reset_list: a list to which cm_node will be selected
1822 */
1823 static void
irdma_reset_list_prep(struct irdma_cm_core * cm_core,struct irdma_cm_listener * listener,struct list_head * reset_list)1824 irdma_reset_list_prep(struct irdma_cm_core *cm_core,
1825 struct irdma_cm_listener *listener,
1826 struct list_head *reset_list)
1827 {
1828 struct irdma_cm_node *cm_node;
1829 int bkt;
1830
1831 HASH_FOR_EACH_RCU(cm_core->cm_hash_tbl, bkt, cm_node, list) {
1832 if (cm_node->listener == listener &&
1833 !cm_node->accelerated &&
1834 atomic_inc_not_zero(&cm_node->refcnt))
1835 list_add(&cm_node->reset_entry, reset_list);
1836 }
1837 }
1838
1839 /**
1840 * irdma_dec_refcnt_listen - delete listener and associated cm nodes
1841 * @cm_core: cm's core
1842 * @listener: pointer to listener node
1843 * @free_hanging_nodes: to free associated cm_nodes
1844 * @apbvt_del: flag to delete the apbvt
1845 */
1846 static int
irdma_dec_refcnt_listen(struct irdma_cm_core * cm_core,struct irdma_cm_listener * listener,int free_hanging_nodes,bool apbvt_del)1847 irdma_dec_refcnt_listen(struct irdma_cm_core *cm_core,
1848 struct irdma_cm_listener *listener,
1849 int free_hanging_nodes, bool apbvt_del)
1850 {
1851 struct list_head *list_pos;
1852 struct list_head *list_temp;
1853 struct irdma_cm_node *cm_node;
1854 struct list_head reset_list;
1855 struct irdma_cm_info nfo;
1856 enum irdma_cm_node_state old_state;
1857 unsigned long flags;
1858 int err;
1859
1860 /* free non-accelerated child nodes for this listener */
1861 INIT_LIST_HEAD(&reset_list);
1862 if (free_hanging_nodes) {
1863 rcu_read_lock();
1864 irdma_reset_list_prep(cm_core, listener, &reset_list);
1865 rcu_read_unlock();
1866 }
1867
1868 list_for_each_safe(list_pos, list_temp, &reset_list) {
1869 cm_node = container_of(list_pos, struct irdma_cm_node,
1870 reset_entry);
1871 if (cm_node->state >= IRDMA_CM_STATE_FIN_WAIT1) {
1872 irdma_rem_ref_cm_node(cm_node);
1873 continue;
1874 }
1875
1876 irdma_cleanup_retrans_entry(cm_node);
1877 err = irdma_send_reset(cm_node);
1878 if (err) {
1879 cm_node->state = IRDMA_CM_STATE_CLOSED;
1880 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1881 "send reset failed\n");
1882 } else {
1883 old_state = cm_node->state;
1884 cm_node->state = IRDMA_CM_STATE_LISTENER_DESTROYED;
1885 if (old_state != IRDMA_CM_STATE_MPAREQ_RCVD)
1886 irdma_rem_ref_cm_node(cm_node);
1887 }
1888 }
1889
1890 if (atomic_dec_and_test(&listener->refcnt)) {
1891 spin_lock_irqsave(&cm_core->listen_list_lock, flags);
1892 list_del(&listener->list);
1893 spin_unlock_irqrestore(&cm_core->listen_list_lock, flags);
1894
1895 if (apbvt_del)
1896 irdma_del_apbvt(listener->iwdev,
1897 listener->apbvt_entry);
1898 memcpy(nfo.loc_addr, listener->loc_addr, sizeof(nfo.loc_addr));
1899 nfo.loc_port = listener->loc_port;
1900 nfo.ipv4 = listener->ipv4;
1901 nfo.vlan_id = listener->vlan_id;
1902 nfo.user_pri = listener->user_pri;
1903 nfo.qh_qpid = listener->iwdev->vsi.ilq->qp_id;
1904
1905 if (!list_empty(&listener->child_listen_list)) {
1906 irdma_del_multiple_qhash(listener->iwdev, &nfo,
1907 listener);
1908 } else {
1909 if (listener->qhash_set)
1910 irdma_manage_qhash(listener->iwdev,
1911 &nfo,
1912 IRDMA_QHASH_TYPE_TCP_SYN,
1913 IRDMA_QHASH_MANAGE_TYPE_DELETE,
1914 NULL, false);
1915 }
1916
1917 cm_core->stats_listen_destroyed++;
1918 cm_core->stats_listen_nodes_destroyed++;
1919 irdma_debug(&listener->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1920 "loc_port=0x%04x loc_addr=%x cm_listen_node=%p cm_id=%p qhash_set=%d vlan_id=%d apbvt_del=%d\n",
1921 listener->loc_port, listener->loc_addr[0], listener,
1922 listener->cm_id, listener->qhash_set,
1923 listener->vlan_id, apbvt_del);
1924 kfree(listener);
1925 listener = NULL;
1926 return 0;
1927 }
1928
1929 return -EINVAL;
1930 }
1931
1932 /**
1933 * irdma_cm_del_listen - delete a listener
1934 * @cm_core: cm's core
1935 * @listener: passive connection's listener
1936 * @apbvt_del: flag to delete apbvt
1937 */
1938 static int
irdma_cm_del_listen(struct irdma_cm_core * cm_core,struct irdma_cm_listener * listener,bool apbvt_del)1939 irdma_cm_del_listen(struct irdma_cm_core *cm_core,
1940 struct irdma_cm_listener *listener,
1941 bool apbvt_del)
1942 {
1943 listener->listener_state = IRDMA_CM_LISTENER_PASSIVE_STATE;
1944 listener->cm_id = NULL;
1945
1946 return irdma_dec_refcnt_listen(cm_core, listener, 1, apbvt_del);
1947 }
1948
1949 /**
1950 * irdma_find_node - find a cm node that matches the reference cm node
1951 * @cm_core: cm's core
1952 * @rem_port: remote tcp port num
1953 * @rem_addr: remote ip addr
1954 * @loc_port: local tcp port num
1955 * @loc_addr: local ip addr
1956 * @vlan_id: local VLAN ID
1957 */
1958 struct irdma_cm_node *
irdma_find_node(struct irdma_cm_core * cm_core,u16 rem_port,u32 * rem_addr,u16 loc_port,u32 * loc_addr,u16 vlan_id)1959 irdma_find_node(struct irdma_cm_core *cm_core,
1960 u16 rem_port, u32 *rem_addr, u16 loc_port,
1961 u32 *loc_addr, u16 vlan_id)
1962 {
1963 struct irdma_cm_node *cm_node;
1964 u32 key = (rem_port << 16) | loc_port;
1965
1966 rcu_read_lock();
1967 HASH_FOR_EACH_POSSIBLE_RCU(cm_core->cm_hash_tbl, cm_node, list, key) {
1968 if (cm_node->vlan_id == vlan_id &&
1969 cm_node->loc_port == loc_port && cm_node->rem_port == rem_port &&
1970 !memcmp(cm_node->loc_addr, loc_addr, sizeof(cm_node->loc_addr)) &&
1971 !memcmp(cm_node->rem_addr, rem_addr, sizeof(cm_node->rem_addr))) {
1972 if (!atomic_inc_not_zero(&cm_node->refcnt))
1973 goto exit;
1974 rcu_read_unlock();
1975 return cm_node;
1976 }
1977 }
1978
1979 exit:
1980 rcu_read_unlock();
1981
1982 /* no owner node */
1983 return NULL;
1984 }
1985
1986 /**
1987 * irdma_add_hte_node - add a cm node to the hash table
1988 * @cm_core: cm's core
1989 * @cm_node: connection's node
1990 */
1991 static void
irdma_add_hte_node(struct irdma_cm_core * cm_core,struct irdma_cm_node * cm_node)1992 irdma_add_hte_node(struct irdma_cm_core *cm_core,
1993 struct irdma_cm_node *cm_node)
1994 {
1995 unsigned long flags;
1996 u32 key = (cm_node->rem_port << 16) | cm_node->loc_port;
1997
1998 spin_lock_irqsave(&cm_core->ht_lock, flags);
1999 HASH_ADD_RCU(cm_core->cm_hash_tbl, &cm_node->list, key);
2000 spin_unlock_irqrestore(&cm_core->ht_lock, flags);
2001 }
2002
2003 /**
2004 * irdma_ipv4_is_lpb - check if loopback
2005 * @loc_addr: local addr to compare
2006 * @rem_addr: remote address
2007 */
2008 bool
irdma_ipv4_is_lpb(u32 loc_addr,u32 rem_addr)2009 irdma_ipv4_is_lpb(u32 loc_addr, u32 rem_addr)
2010 {
2011 return ipv4_is_loopback(htonl(rem_addr)) || (loc_addr == rem_addr);
2012 }
2013
2014 /**
2015 * irdma_ipv6_is_lpb - check if loopback
2016 * @loc_addr: local addr to compare
2017 * @rem_addr: remote address
2018 */
2019 bool
irdma_ipv6_is_lpb(u32 * loc_addr,u32 * rem_addr)2020 irdma_ipv6_is_lpb(u32 *loc_addr, u32 *rem_addr)
2021 {
2022 struct in6_addr raddr6;
2023
2024 irdma_copy_ip_htonl(raddr6.__u6_addr.__u6_addr32, rem_addr);
2025
2026 return !memcmp(loc_addr, rem_addr, 16) || ipv6_addr_loopback(&raddr6);
2027 }
2028
2029 /**
2030 * irdma_cm_create_ah - create a cm address handle
2031 * @cm_node: The connection manager node to create AH for
2032 * @wait: Provides option to wait for ah creation or not
2033 */
2034 static int
irdma_cm_create_ah(struct irdma_cm_node * cm_node,bool wait)2035 irdma_cm_create_ah(struct irdma_cm_node *cm_node, bool wait)
2036 {
2037 struct irdma_ah_info ah_info = {0};
2038 struct irdma_device *iwdev = cm_node->iwdev;
2039 #ifdef VIMAGE
2040 struct vnet *vnet = irdma_cmid_to_vnet(cm_node->cm_id);
2041 #endif
2042
2043 ether_addr_copy(ah_info.mac_addr, if_getlladdr(iwdev->netdev));
2044
2045 ah_info.hop_ttl = 0x40;
2046 ah_info.tc_tos = cm_node->tos;
2047 ah_info.vsi = &iwdev->vsi;
2048
2049 if (cm_node->ipv4) {
2050 ah_info.ipv4_valid = true;
2051 ah_info.dest_ip_addr[0] = cm_node->rem_addr[0];
2052 ah_info.src_ip_addr[0] = cm_node->loc_addr[0];
2053 CURVNET_SET_QUIET(vnet);
2054 ah_info.do_lpbk = irdma_ipv4_is_lpb(ah_info.src_ip_addr[0],
2055 ah_info.dest_ip_addr[0]);
2056 CURVNET_RESTORE();
2057 } else {
2058 memcpy(ah_info.dest_ip_addr, cm_node->rem_addr,
2059 sizeof(ah_info.dest_ip_addr));
2060 memcpy(ah_info.src_ip_addr, cm_node->loc_addr,
2061 sizeof(ah_info.src_ip_addr));
2062 ah_info.do_lpbk = irdma_ipv6_is_lpb(ah_info.src_ip_addr,
2063 ah_info.dest_ip_addr);
2064 }
2065
2066 ah_info.vlan_tag = cm_node->vlan_id;
2067 if (cm_node->vlan_id < VLAN_N_VID) {
2068 ah_info.insert_vlan_tag = 1;
2069 ah_info.vlan_tag |= cm_node->user_pri << VLAN_PRIO_SHIFT;
2070 }
2071
2072 ah_info.dst_arpindex =
2073 irdma_arp_table(iwdev->rf, ah_info.dest_ip_addr,
2074 NULL, IRDMA_ARP_RESOLVE);
2075
2076 if (irdma_puda_create_ah(&iwdev->rf->sc_dev, &ah_info, wait,
2077 IRDMA_PUDA_RSRC_TYPE_ILQ, cm_node,
2078 &cm_node->ah))
2079 return -ENOMEM;
2080
2081 return 0;
2082 }
2083
2084 /**
2085 * irdma_cm_free_ah - free a cm address handle
2086 * @cm_node: The connection manager node to create AH for
2087 */
2088 static void
irdma_cm_free_ah(struct irdma_cm_node * cm_node)2089 irdma_cm_free_ah(struct irdma_cm_node *cm_node)
2090 {
2091 struct irdma_device *iwdev = cm_node->iwdev;
2092
2093 irdma_puda_free_ah(&iwdev->rf->sc_dev, cm_node->ah);
2094 cm_node->ah = NULL;
2095 }
2096
2097 /**
2098 * irdma_make_cm_node - create a new instance of a cm node
2099 * @cm_core: cm's core
2100 * @iwdev: iwarp device structure
2101 * @cm_info: quad info for connection
2102 * @listener: passive connection's listener
2103 */
2104 static struct irdma_cm_node *
irdma_make_cm_node(struct irdma_cm_core * cm_core,struct irdma_device * iwdev,struct irdma_cm_info * cm_info,struct irdma_cm_listener * listener)2105 irdma_make_cm_node(struct irdma_cm_core *cm_core, struct irdma_device *iwdev,
2106 struct irdma_cm_info *cm_info,
2107 struct irdma_cm_listener *listener)
2108 {
2109 struct irdma_cm_node *cm_node;
2110 int arpindex;
2111 if_t netdev = iwdev->netdev;
2112
2113 /* create an hte and cm_node for this instance */
2114 cm_node = kzalloc(sizeof(*cm_node), GFP_ATOMIC);
2115 if (!cm_node)
2116 return NULL;
2117
2118 /* set our node specific transport info */
2119 cm_node->ipv4 = cm_info->ipv4;
2120 cm_node->vlan_id = cm_info->vlan_id;
2121 if (cm_node->vlan_id >= VLAN_N_VID && iwdev->dcb_vlan_mode)
2122 cm_node->vlan_id = 0;
2123 cm_node->tos = cm_info->tos;
2124 cm_node->user_pri = cm_info->user_pri;
2125 if (listener) {
2126 if (listener->tos != cm_info->tos)
2127 irdma_dev_warn(&iwdev->ibdev,
2128 "application TOS[%d] and remote client TOS[%d] mismatch\n",
2129 listener->tos, cm_info->tos);
2130 if (iwdev->vsi.dscp_mode) {
2131 cm_node->user_pri = listener->user_pri;
2132 } else {
2133 cm_node->tos = max(listener->tos, cm_info->tos);
2134 cm_node->user_pri = rt_tos2priority(cm_node->tos);
2135 cm_node->user_pri =
2136 irdma_iw_get_vlan_prio(cm_info->loc_addr,
2137 cm_node->user_pri,
2138 cm_info->ipv4);
2139 }
2140 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_DCB,
2141 "listener: TOS:[%d] UP:[%d]\n", cm_node->tos,
2142 cm_node->user_pri);
2143 }
2144 memcpy(cm_node->loc_addr, cm_info->loc_addr, sizeof(cm_node->loc_addr));
2145 memcpy(cm_node->rem_addr, cm_info->rem_addr, sizeof(cm_node->rem_addr));
2146 cm_node->loc_port = cm_info->loc_port;
2147 cm_node->rem_port = cm_info->rem_port;
2148
2149 cm_node->mpa_frame_rev = IRDMA_CM_DEFAULT_MPA_VER;
2150 cm_node->send_rdma0_op = SEND_RDMA_READ_ZERO;
2151 cm_node->iwdev = iwdev;
2152 cm_node->dev = &iwdev->rf->sc_dev;
2153
2154 cm_node->ird_size = cm_node->dev->hw_attrs.max_hw_ird;
2155 cm_node->ord_size = cm_node->dev->hw_attrs.max_hw_ord;
2156
2157 cm_node->listener = listener;
2158 cm_node->cm_id = cm_info->cm_id;
2159 ether_addr_copy(cm_node->loc_mac, if_getlladdr(netdev));
2160 spin_lock_init(&cm_node->retrans_list_lock);
2161 cm_node->ack_rcvd = false;
2162
2163 init_completion(&cm_node->establish_comp);
2164 atomic_set(&cm_node->refcnt, 1);
2165 /* associate our parent CM core */
2166 cm_node->cm_core = cm_core;
2167 cm_node->tcp_cntxt.loc_id = IRDMA_CM_DEFAULT_LOCAL_ID;
2168 cm_node->tcp_cntxt.rcv_wscale = iwdev->rcv_wscale;
2169 cm_node->tcp_cntxt.rcv_wnd = iwdev->rcv_wnd >> cm_node->tcp_cntxt.rcv_wscale;
2170 kc_set_loc_seq_num_mss(cm_node);
2171
2172 arpindex = irdma_resolve_neigh_lpb_chk(iwdev, cm_node, cm_info);
2173 if (arpindex < 0)
2174 goto err;
2175
2176 ether_addr_copy(cm_node->rem_mac, iwdev->rf->arp_table[arpindex].mac_addr);
2177 irdma_add_hte_node(cm_core, cm_node);
2178 cm_core->stats_nodes_created++;
2179 return cm_node;
2180
2181 err:
2182 kfree(cm_node);
2183
2184 return NULL;
2185 }
2186
2187 static void
irdma_destroy_connection(struct irdma_cm_node * cm_node)2188 irdma_destroy_connection(struct irdma_cm_node *cm_node)
2189 {
2190 struct irdma_cm_core *cm_core = cm_node->cm_core;
2191 struct irdma_qp *iwqp;
2192 struct irdma_cm_info nfo;
2193
2194 /* if the node is destroyed before connection was accelerated */
2195 if (!cm_node->accelerated && cm_node->accept_pend) {
2196 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2197 "node destroyed before established\n");
2198 atomic_dec(&cm_node->listener->pend_accepts_cnt);
2199 }
2200 if (cm_node->close_entry)
2201 irdma_handle_close_entry(cm_node, 0);
2202 if (cm_node->listener) {
2203 irdma_dec_refcnt_listen(cm_core, cm_node->listener, 0, true);
2204 } else {
2205 if (cm_node->apbvt_set) {
2206 irdma_del_apbvt(cm_node->iwdev, cm_node->apbvt_entry);
2207 cm_node->apbvt_set = 0;
2208 }
2209 irdma_get_addr_info(cm_node, &nfo);
2210 if (cm_node->qhash_set) {
2211 nfo.qh_qpid = cm_node->iwdev->vsi.ilq->qp_id;
2212 irdma_manage_qhash(cm_node->iwdev, &nfo,
2213 IRDMA_QHASH_TYPE_TCP_ESTABLISHED,
2214 IRDMA_QHASH_MANAGE_TYPE_DELETE, NULL,
2215 false);
2216 cm_node->qhash_set = 0;
2217 }
2218 }
2219
2220 iwqp = cm_node->iwqp;
2221 if (iwqp) {
2222 cm_node->cm_id->rem_ref(cm_node->cm_id);
2223 cm_node->cm_id = NULL;
2224 iwqp->cm_id = NULL;
2225 irdma_qp_rem_ref(&iwqp->ibqp);
2226 cm_node->iwqp = NULL;
2227 } else if (cm_node->qhash_set) {
2228 irdma_get_addr_info(cm_node, &nfo);
2229 nfo.qh_qpid = cm_node->iwdev->vsi.ilq->qp_id;
2230 irdma_manage_qhash(cm_node->iwdev, &nfo,
2231 IRDMA_QHASH_TYPE_TCP_ESTABLISHED,
2232 IRDMA_QHASH_MANAGE_TYPE_DELETE, NULL, false);
2233 cm_node->qhash_set = 0;
2234 }
2235
2236 cm_core->cm_free_ah(cm_node);
2237 }
2238
2239 /**
2240 * irdma_rem_ref_cm_node - destroy an instance of a cm node
2241 * @cm_node: connection's node
2242 */
2243 void
irdma_rem_ref_cm_node(struct irdma_cm_node * cm_node)2244 irdma_rem_ref_cm_node(struct irdma_cm_node *cm_node)
2245 {
2246 struct irdma_cm_core *cm_core = cm_node->cm_core;
2247 unsigned long flags;
2248
2249 spin_lock_irqsave(&cm_core->ht_lock, flags);
2250
2251 if (!atomic_dec_and_test(&cm_node->refcnt)) {
2252 spin_unlock_irqrestore(&cm_core->ht_lock, flags);
2253 return;
2254 }
2255 if (cm_node->iwqp) {
2256 cm_node->iwqp->cm_node = NULL;
2257 cm_node->iwqp->cm_id = NULL;
2258 }
2259 HASH_DEL_RCU(cm_core->cm_hash_tbl, &cm_node->list);
2260 cm_node->cm_core->stats_nodes_destroyed++;
2261
2262 spin_unlock_irqrestore(&cm_core->ht_lock, flags);
2263
2264 irdma_destroy_connection(cm_node);
2265
2266 kfree_rcu(cm_node, rcu_head);
2267 }
2268
2269 /**
2270 * irdma_handle_fin_pkt - FIN packet received
2271 * @cm_node: connection's node
2272 */
2273 static void
irdma_handle_fin_pkt(struct irdma_cm_node * cm_node)2274 irdma_handle_fin_pkt(struct irdma_cm_node *cm_node)
2275 {
2276 switch (cm_node->state) {
2277 case IRDMA_CM_STATE_SYN_RCVD:
2278 case IRDMA_CM_STATE_SYN_SENT:
2279 case IRDMA_CM_STATE_ESTABLISHED:
2280 case IRDMA_CM_STATE_MPAREJ_RCVD:
2281 cm_node->tcp_cntxt.rcv_nxt++;
2282 irdma_cleanup_retrans_entry(cm_node);
2283 cm_node->state = IRDMA_CM_STATE_LAST_ACK;
2284 irdma_send_fin(cm_node);
2285 break;
2286 case IRDMA_CM_STATE_MPAREQ_SENT:
2287 irdma_create_event(cm_node, IRDMA_CM_EVENT_ABORTED);
2288 cm_node->tcp_cntxt.rcv_nxt++;
2289 irdma_cleanup_retrans_entry(cm_node);
2290 cm_node->state = IRDMA_CM_STATE_CLOSED;
2291 atomic_inc(&cm_node->refcnt);
2292 irdma_send_reset(cm_node);
2293 break;
2294 case IRDMA_CM_STATE_FIN_WAIT1:
2295 cm_node->tcp_cntxt.rcv_nxt++;
2296 irdma_cleanup_retrans_entry(cm_node);
2297 cm_node->state = IRDMA_CM_STATE_CLOSING;
2298 irdma_send_ack(cm_node);
2299 /*
2300 * Wait for ACK as this is simultaneous close. After we receive ACK, do not send anything. Just rm the
2301 * node.
2302 */
2303 break;
2304 case IRDMA_CM_STATE_FIN_WAIT2:
2305 cm_node->tcp_cntxt.rcv_nxt++;
2306 irdma_cleanup_retrans_entry(cm_node);
2307 cm_node->state = IRDMA_CM_STATE_TIME_WAIT;
2308 irdma_send_ack(cm_node);
2309 irdma_schedule_cm_timer(cm_node, NULL, IRDMA_TIMER_TYPE_CLOSE,
2310 1, 0);
2311 break;
2312 case IRDMA_CM_STATE_TIME_WAIT:
2313 cm_node->tcp_cntxt.rcv_nxt++;
2314 irdma_cleanup_retrans_entry(cm_node);
2315 cm_node->state = IRDMA_CM_STATE_CLOSED;
2316 irdma_rem_ref_cm_node(cm_node);
2317 break;
2318 case IRDMA_CM_STATE_OFFLOADED:
2319 default:
2320 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2321 "bad state node state = %d\n", cm_node->state);
2322 break;
2323 }
2324 }
2325
2326 /**
2327 * irdma_handle_rst_pkt - process received RST packet
2328 * @cm_node: connection's node
2329 * @rbuf: receive buffer
2330 */
2331 static void
irdma_handle_rst_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2332 irdma_handle_rst_pkt(struct irdma_cm_node *cm_node,
2333 struct irdma_puda_buf *rbuf)
2334 {
2335 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2336 "caller: %pS cm_node=%p state=%d rem_port=0x%04x loc_port=0x%04x rem_addr=%x loc_addr=%x\n",
2337 __builtin_return_address(0), cm_node, cm_node->state,
2338 cm_node->rem_port, cm_node->loc_port, cm_node->rem_addr[0],
2339 cm_node->loc_addr[0]);
2340
2341 irdma_cleanup_retrans_entry(cm_node);
2342 switch (cm_node->state) {
2343 case IRDMA_CM_STATE_SYN_SENT:
2344 case IRDMA_CM_STATE_MPAREQ_SENT:
2345 switch (cm_node->mpa_frame_rev) {
2346 case IETF_MPA_V2:
2347 /* Drop down to MPA_V1 */
2348 cm_node->mpa_frame_rev = IETF_MPA_V1;
2349 /* send a syn and goto syn sent state */
2350 cm_node->state = IRDMA_CM_STATE_SYN_SENT;
2351 if (irdma_send_syn(cm_node, 0))
2352 irdma_active_open_err(cm_node, false);
2353 break;
2354 case IETF_MPA_V1:
2355 default:
2356 irdma_active_open_err(cm_node, false);
2357 break;
2358 }
2359 break;
2360 case IRDMA_CM_STATE_MPAREQ_RCVD:
2361 atomic_inc(&cm_node->passive_state);
2362 break;
2363 case IRDMA_CM_STATE_ESTABLISHED:
2364 case IRDMA_CM_STATE_SYN_RCVD:
2365 case IRDMA_CM_STATE_LISTENING:
2366 irdma_passive_open_err(cm_node, false);
2367 break;
2368 case IRDMA_CM_STATE_OFFLOADED:
2369 irdma_active_open_err(cm_node, false);
2370 break;
2371 case IRDMA_CM_STATE_CLOSED:
2372 break;
2373 case IRDMA_CM_STATE_FIN_WAIT2:
2374 case IRDMA_CM_STATE_FIN_WAIT1:
2375 case IRDMA_CM_STATE_LAST_ACK:
2376 case IRDMA_CM_STATE_TIME_WAIT:
2377 cm_node->state = IRDMA_CM_STATE_CLOSED;
2378 irdma_rem_ref_cm_node(cm_node);
2379 break;
2380 default:
2381 break;
2382 }
2383 }
2384
2385 /**
2386 * irdma_handle_rcv_mpa - Process a recv'd mpa buffer
2387 * @cm_node: connection's node
2388 * @rbuf: receive buffer
2389 */
2390 static void
irdma_handle_rcv_mpa(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2391 irdma_handle_rcv_mpa(struct irdma_cm_node *cm_node,
2392 struct irdma_puda_buf *rbuf)
2393 {
2394 int err;
2395 int datasize = rbuf->datalen;
2396 u8 *dataloc = rbuf->data;
2397
2398 enum irdma_cm_event_type type = IRDMA_CM_EVENT_UNKNOWN;
2399 u32 res_type;
2400
2401 err = irdma_parse_mpa(cm_node, dataloc, &res_type, datasize);
2402 if (err) {
2403 if (cm_node->state == IRDMA_CM_STATE_MPAREQ_SENT)
2404 irdma_active_open_err(cm_node, true);
2405 else
2406 irdma_passive_open_err(cm_node, true);
2407 return;
2408 }
2409
2410 switch (cm_node->state) {
2411 case IRDMA_CM_STATE_ESTABLISHED:
2412 if (res_type == IRDMA_MPA_REQUEST_REJECT)
2413 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2414 "state for reject\n");
2415 cm_node->state = IRDMA_CM_STATE_MPAREQ_RCVD;
2416 type = IRDMA_CM_EVENT_MPA_REQ;
2417 irdma_send_ack(cm_node); /* ACK received MPA request */
2418 atomic_set(&cm_node->passive_state,
2419 IRDMA_PASSIVE_STATE_INDICATED);
2420 break;
2421 case IRDMA_CM_STATE_MPAREQ_SENT:
2422 irdma_cleanup_retrans_entry(cm_node);
2423 if (res_type == IRDMA_MPA_REQUEST_REJECT) {
2424 type = IRDMA_CM_EVENT_MPA_REJECT;
2425 cm_node->state = IRDMA_CM_STATE_MPAREJ_RCVD;
2426 } else {
2427 type = IRDMA_CM_EVENT_CONNECTED;
2428 cm_node->state = IRDMA_CM_STATE_OFFLOADED;
2429 }
2430 irdma_send_ack(cm_node);
2431 break;
2432 default:
2433 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2434 "wrong cm_node state=%d\n", cm_node->state);
2435 break;
2436 }
2437 irdma_create_event(cm_node, type);
2438 }
2439
2440 /**
2441 * irdma_check_syn - Check for error on received syn ack
2442 * @cm_node: connection's node
2443 * @tcph: pointer tcp header
2444 */
2445 static int
irdma_check_syn(struct irdma_cm_node * cm_node,struct tcphdr * tcph)2446 irdma_check_syn(struct irdma_cm_node *cm_node, struct tcphdr *tcph)
2447 {
2448 if (ntohl(tcph->th_ack) != cm_node->tcp_cntxt.loc_seq_num) {
2449 irdma_active_open_err(cm_node, true);
2450 return 1;
2451 }
2452
2453 return 0;
2454 }
2455
2456 /**
2457 * irdma_check_seq - check seq numbers if OK
2458 * @cm_node: connection's node
2459 * @tcph: pointer tcp header
2460 */
2461 static int
irdma_check_seq(struct irdma_cm_node * cm_node,struct tcphdr * tcph)2462 irdma_check_seq(struct irdma_cm_node *cm_node, struct tcphdr *tcph)
2463 {
2464 u32 seq;
2465 u32 ack_seq;
2466 u32 loc_seq_num = cm_node->tcp_cntxt.loc_seq_num;
2467 u32 rcv_nxt = cm_node->tcp_cntxt.rcv_nxt;
2468 u32 rcv_wnd;
2469 int err = 0;
2470
2471 seq = ntohl(tcph->th_seq);
2472 ack_seq = ntohl(tcph->th_ack);
2473 rcv_wnd = cm_node->tcp_cntxt.rcv_wnd;
2474 if (ack_seq != loc_seq_num ||
2475 !between(seq, rcv_nxt, (rcv_nxt + rcv_wnd)))
2476 err = -1;
2477 if (err)
2478 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2479 "seq number err\n");
2480
2481 return err;
2482 }
2483
2484 void
irdma_add_conn_est_qh(struct irdma_cm_node * cm_node)2485 irdma_add_conn_est_qh(struct irdma_cm_node *cm_node)
2486 {
2487 struct irdma_cm_info nfo;
2488
2489 irdma_get_addr_info(cm_node, &nfo);
2490 nfo.qh_qpid = cm_node->iwdev->vsi.ilq->qp_id;
2491 irdma_manage_qhash(cm_node->iwdev, &nfo,
2492 IRDMA_QHASH_TYPE_TCP_ESTABLISHED,
2493 IRDMA_QHASH_MANAGE_TYPE_ADD,
2494 cm_node, false);
2495 cm_node->qhash_set = true;
2496 }
2497
2498 /**
2499 * irdma_handle_syn_pkt - is for Passive node
2500 * @cm_node: connection's node
2501 * @rbuf: receive buffer
2502 */
2503 static void
irdma_handle_syn_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2504 irdma_handle_syn_pkt(struct irdma_cm_node *cm_node,
2505 struct irdma_puda_buf *rbuf)
2506 {
2507 struct tcphdr *tcph = (struct tcphdr *)rbuf->tcph;
2508 int err;
2509 u32 inc_sequence;
2510 int optionsize;
2511
2512 optionsize = (tcph->th_off << 2) - sizeof(*tcph);
2513 inc_sequence = ntohl(tcph->th_seq);
2514
2515 switch (cm_node->state) {
2516 case IRDMA_CM_STATE_SYN_SENT:
2517 case IRDMA_CM_STATE_MPAREQ_SENT:
2518 /* Rcvd syn on active open connection */
2519 irdma_active_open_err(cm_node, 1);
2520 break;
2521 case IRDMA_CM_STATE_LISTENING:
2522 /* Passive OPEN */
2523 if (atomic_read(&cm_node->listener->pend_accepts_cnt) >
2524 cm_node->listener->backlog) {
2525 cm_node->cm_core->stats_backlog_drops++;
2526 irdma_passive_open_err(cm_node, false);
2527 break;
2528 }
2529 err = irdma_handle_tcp_options(cm_node, tcph, optionsize, 1);
2530 if (err) {
2531 irdma_passive_open_err(cm_node, false);
2532 /* drop pkt */
2533 break;
2534 }
2535 err = cm_node->cm_core->cm_create_ah(cm_node, false);
2536 if (err) {
2537 irdma_passive_open_err(cm_node, false);
2538 /* drop pkt */
2539 break;
2540 }
2541 cm_node->tcp_cntxt.rcv_nxt = inc_sequence + 1;
2542 cm_node->accept_pend = 1;
2543 atomic_inc(&cm_node->listener->pend_accepts_cnt);
2544
2545 cm_node->state = IRDMA_CM_STATE_SYN_RCVD;
2546 break;
2547 case IRDMA_CM_STATE_CLOSED:
2548 irdma_cleanup_retrans_entry(cm_node);
2549 atomic_inc(&cm_node->refcnt);
2550 irdma_send_reset(cm_node);
2551 break;
2552 case IRDMA_CM_STATE_OFFLOADED:
2553 case IRDMA_CM_STATE_ESTABLISHED:
2554 case IRDMA_CM_STATE_FIN_WAIT1:
2555 case IRDMA_CM_STATE_FIN_WAIT2:
2556 case IRDMA_CM_STATE_MPAREQ_RCVD:
2557 case IRDMA_CM_STATE_LAST_ACK:
2558 case IRDMA_CM_STATE_CLOSING:
2559 case IRDMA_CM_STATE_UNKNOWN:
2560 default:
2561 break;
2562 }
2563 }
2564
2565 /**
2566 * irdma_handle_synack_pkt - Process SYN+ACK packet (active side)
2567 * @cm_node: connection's node
2568 * @rbuf: receive buffer
2569 */
2570 static void
irdma_handle_synack_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2571 irdma_handle_synack_pkt(struct irdma_cm_node *cm_node,
2572 struct irdma_puda_buf *rbuf)
2573 {
2574 struct tcphdr *tcph = (struct tcphdr *)rbuf->tcph;
2575 int err;
2576 u32 inc_sequence;
2577 int optionsize;
2578
2579 optionsize = (tcph->th_off << 2) - sizeof(*tcph);
2580 inc_sequence = ntohl(tcph->th_seq);
2581 switch (cm_node->state) {
2582 case IRDMA_CM_STATE_SYN_SENT:
2583 irdma_cleanup_retrans_entry(cm_node);
2584 /* active open */
2585 if (irdma_check_syn(cm_node, tcph)) {
2586 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2587 "check syn fail\n");
2588 return;
2589 }
2590 cm_node->tcp_cntxt.rem_ack_num = ntohl(tcph->th_ack);
2591 /* setup options */
2592 err = irdma_handle_tcp_options(cm_node, tcph, optionsize, 0);
2593 if (err) {
2594 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2595 "cm_node=%p tcp_options failed\n", cm_node);
2596 break;
2597 }
2598 irdma_cleanup_retrans_entry(cm_node);
2599 cm_node->tcp_cntxt.rcv_nxt = inc_sequence + 1;
2600 irdma_send_ack(cm_node); /* ACK for the syn_ack */
2601 err = irdma_send_mpa_request(cm_node);
2602 if (err) {
2603 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2604 "cm_node=%p irdma_send_mpa_request failed\n",
2605 cm_node);
2606 break;
2607 }
2608 cm_node->state = IRDMA_CM_STATE_MPAREQ_SENT;
2609 break;
2610 case IRDMA_CM_STATE_MPAREQ_RCVD:
2611 irdma_passive_open_err(cm_node, true);
2612 break;
2613 case IRDMA_CM_STATE_LISTENING:
2614 cm_node->tcp_cntxt.loc_seq_num = ntohl(tcph->th_ack);
2615 irdma_cleanup_retrans_entry(cm_node);
2616 cm_node->state = IRDMA_CM_STATE_CLOSED;
2617 irdma_send_reset(cm_node);
2618 break;
2619 case IRDMA_CM_STATE_CLOSED:
2620 cm_node->tcp_cntxt.loc_seq_num = ntohl(tcph->th_ack);
2621 irdma_cleanup_retrans_entry(cm_node);
2622 atomic_inc(&cm_node->refcnt);
2623 irdma_send_reset(cm_node);
2624 break;
2625 case IRDMA_CM_STATE_ESTABLISHED:
2626 case IRDMA_CM_STATE_FIN_WAIT1:
2627 case IRDMA_CM_STATE_FIN_WAIT2:
2628 case IRDMA_CM_STATE_LAST_ACK:
2629 case IRDMA_CM_STATE_OFFLOADED:
2630 case IRDMA_CM_STATE_CLOSING:
2631 case IRDMA_CM_STATE_UNKNOWN:
2632 case IRDMA_CM_STATE_MPAREQ_SENT:
2633 default:
2634 break;
2635 }
2636 }
2637
2638 /**
2639 * irdma_handle_ack_pkt - process packet with ACK
2640 * @cm_node: connection's node
2641 * @rbuf: receive buffer
2642 */
2643 static int
irdma_handle_ack_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2644 irdma_handle_ack_pkt(struct irdma_cm_node *cm_node,
2645 struct irdma_puda_buf *rbuf)
2646 {
2647 struct tcphdr *tcph = (struct tcphdr *)rbuf->tcph;
2648 u32 inc_sequence;
2649 int ret;
2650 int optionsize;
2651 u32 datasize = rbuf->datalen;
2652
2653 optionsize = (tcph->th_off << 2) - sizeof(*tcph);
2654
2655 if (irdma_check_seq(cm_node, tcph))
2656 return -EINVAL;
2657
2658 inc_sequence = ntohl(tcph->th_seq);
2659 switch (cm_node->state) {
2660 case IRDMA_CM_STATE_SYN_RCVD:
2661 irdma_cleanup_retrans_entry(cm_node);
2662 ret = irdma_handle_tcp_options(cm_node, tcph, optionsize, 1);
2663 if (ret)
2664 return ret;
2665 cm_node->tcp_cntxt.rem_ack_num = ntohl(tcph->th_ack);
2666 cm_node->state = IRDMA_CM_STATE_ESTABLISHED;
2667 if (datasize) {
2668 cm_node->tcp_cntxt.rcv_nxt = inc_sequence + datasize;
2669 irdma_handle_rcv_mpa(cm_node, rbuf);
2670 }
2671 break;
2672 case IRDMA_CM_STATE_ESTABLISHED:
2673 irdma_cleanup_retrans_entry(cm_node);
2674 if (datasize) {
2675 cm_node->tcp_cntxt.rcv_nxt = inc_sequence + datasize;
2676 irdma_handle_rcv_mpa(cm_node, rbuf);
2677 }
2678 break;
2679 case IRDMA_CM_STATE_MPAREQ_SENT:
2680 cm_node->tcp_cntxt.rem_ack_num = ntohl(tcph->th_ack);
2681 if (datasize) {
2682 cm_node->tcp_cntxt.rcv_nxt = inc_sequence + datasize;
2683 cm_node->ack_rcvd = false;
2684 irdma_handle_rcv_mpa(cm_node, rbuf);
2685 } else {
2686 cm_node->ack_rcvd = true;
2687 }
2688 break;
2689 case IRDMA_CM_STATE_LISTENING:
2690 irdma_cleanup_retrans_entry(cm_node);
2691 cm_node->state = IRDMA_CM_STATE_CLOSED;
2692 irdma_send_reset(cm_node);
2693 break;
2694 case IRDMA_CM_STATE_CLOSED:
2695 irdma_cleanup_retrans_entry(cm_node);
2696 atomic_inc(&cm_node->refcnt);
2697 irdma_send_reset(cm_node);
2698 break;
2699 case IRDMA_CM_STATE_LAST_ACK:
2700 case IRDMA_CM_STATE_CLOSING:
2701 irdma_cleanup_retrans_entry(cm_node);
2702 cm_node->state = IRDMA_CM_STATE_CLOSED;
2703 irdma_rem_ref_cm_node(cm_node);
2704 break;
2705 case IRDMA_CM_STATE_FIN_WAIT1:
2706 irdma_cleanup_retrans_entry(cm_node);
2707 cm_node->state = IRDMA_CM_STATE_FIN_WAIT2;
2708 break;
2709 case IRDMA_CM_STATE_SYN_SENT:
2710 case IRDMA_CM_STATE_FIN_WAIT2:
2711 case IRDMA_CM_STATE_OFFLOADED:
2712 case IRDMA_CM_STATE_MPAREQ_RCVD:
2713 case IRDMA_CM_STATE_UNKNOWN:
2714 default:
2715 irdma_cleanup_retrans_entry(cm_node);
2716 break;
2717 }
2718
2719 return 0;
2720 }
2721
2722 /**
2723 * irdma_process_pkt - process cm packet
2724 * @cm_node: connection's node
2725 * @rbuf: receive buffer
2726 */
2727 static void
irdma_process_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2728 irdma_process_pkt(struct irdma_cm_node *cm_node,
2729 struct irdma_puda_buf *rbuf)
2730 {
2731 enum irdma_tcpip_pkt_type pkt_type = IRDMA_PKT_TYPE_UNKNOWN;
2732 struct tcphdr *tcph = (struct tcphdr *)rbuf->tcph;
2733 u32 fin_set = 0;
2734 int err;
2735
2736 if (tcp_get_flags(tcph) & TH_RST) {
2737 pkt_type = IRDMA_PKT_TYPE_RST;
2738 } else if (tcp_get_flags(tcph) & TH_SYN) {
2739 pkt_type = IRDMA_PKT_TYPE_SYN;
2740 if (tcp_get_flags(tcph) & TH_ACK)
2741 pkt_type = IRDMA_PKT_TYPE_SYNACK;
2742 } else if (tcp_get_flags(tcph) & TH_ACK) {
2743 pkt_type = IRDMA_PKT_TYPE_ACK;
2744 }
2745 if (tcp_get_flags(tcph) & TH_FIN)
2746 fin_set = 1;
2747
2748 switch (pkt_type) {
2749 case IRDMA_PKT_TYPE_SYN:
2750 irdma_handle_syn_pkt(cm_node, rbuf);
2751 break;
2752 case IRDMA_PKT_TYPE_SYNACK:
2753 irdma_handle_synack_pkt(cm_node, rbuf);
2754 break;
2755 case IRDMA_PKT_TYPE_ACK:
2756 err = irdma_handle_ack_pkt(cm_node, rbuf);
2757 if (fin_set && !err)
2758 irdma_handle_fin_pkt(cm_node);
2759 break;
2760 case IRDMA_PKT_TYPE_RST:
2761 irdma_handle_rst_pkt(cm_node, rbuf);
2762 break;
2763 default:
2764 if (fin_set &&
2765 (!irdma_check_seq(cm_node, (struct tcphdr *)rbuf->tcph)))
2766 irdma_handle_fin_pkt(cm_node);
2767 break;
2768 }
2769 }
2770
2771 /**
2772 * irdma_make_listen_node - create a listen node with params
2773 * @cm_core: cm's core
2774 * @iwdev: iwarp device structure
2775 * @cm_info: quad info for connection
2776 */
2777 static struct irdma_cm_listener *
irdma_make_listen_node(struct irdma_cm_core * cm_core,struct irdma_device * iwdev,struct irdma_cm_info * cm_info)2778 irdma_make_listen_node(struct irdma_cm_core *cm_core,
2779 struct irdma_device *iwdev,
2780 struct irdma_cm_info *cm_info)
2781 {
2782 struct irdma_cm_listener *listener;
2783 unsigned long flags;
2784
2785 /* cannot have multiple matching listeners */
2786 listener = irdma_find_listener(cm_core, cm_info->loc_addr, cm_info->ipv4,
2787 cm_info->loc_port, cm_info->vlan_id,
2788 IRDMA_CM_LISTENER_EITHER_STATE);
2789 if (listener &&
2790 listener->listener_state == IRDMA_CM_LISTENER_ACTIVE_STATE) {
2791 atomic_dec(&listener->refcnt);
2792 return NULL;
2793 }
2794
2795 if (!listener) {
2796 /*
2797 * create a CM listen node 1/2 node to compare incoming traffic to
2798 */
2799 listener = kzalloc(sizeof(*listener), GFP_KERNEL);
2800 if (!listener)
2801 return NULL;
2802 cm_core->stats_listen_nodes_created++;
2803 memcpy(listener->loc_addr, cm_info->loc_addr,
2804 sizeof(listener->loc_addr));
2805 listener->loc_port = cm_info->loc_port;
2806
2807 INIT_LIST_HEAD(&listener->child_listen_list);
2808
2809 atomic_set(&listener->refcnt, 1);
2810 } else {
2811 listener->reused_node = 1;
2812 }
2813
2814 listener->cm_id = cm_info->cm_id;
2815 listener->ipv4 = cm_info->ipv4;
2816 listener->vlan_id = cm_info->vlan_id;
2817 atomic_set(&listener->pend_accepts_cnt, 0);
2818 listener->cm_core = cm_core;
2819 listener->iwdev = iwdev;
2820
2821 listener->backlog = cm_info->backlog;
2822 listener->listener_state = IRDMA_CM_LISTENER_ACTIVE_STATE;
2823
2824 if (!listener->reused_node) {
2825 spin_lock_irqsave(&cm_core->listen_list_lock, flags);
2826 list_add(&listener->list, &cm_core->listen_list);
2827 spin_unlock_irqrestore(&cm_core->listen_list_lock, flags);
2828 }
2829
2830 return listener;
2831 }
2832
2833 /**
2834 * irdma_create_cm_node - make a connection node with params
2835 * @cm_core: cm's core
2836 * @iwdev: iwarp device structure
2837 * @conn_param: connection parameters
2838 * @cm_info: quad info for connection
2839 * @caller_cm_node: pointer to cm_node structure to return
2840 */
2841 static int
irdma_create_cm_node(struct irdma_cm_core * cm_core,struct irdma_device * iwdev,struct iw_cm_conn_param * conn_param,struct irdma_cm_info * cm_info,struct irdma_cm_node ** caller_cm_node)2842 irdma_create_cm_node(struct irdma_cm_core *cm_core,
2843 struct irdma_device *iwdev,
2844 struct iw_cm_conn_param *conn_param,
2845 struct irdma_cm_info *cm_info,
2846 struct irdma_cm_node **caller_cm_node)
2847 {
2848 struct irdma_cm_node *cm_node;
2849 u16 private_data_len = conn_param->private_data_len;
2850 const void *private_data = conn_param->private_data;
2851
2852 /* create a CM connection node */
2853 cm_node = irdma_make_cm_node(cm_core, iwdev, cm_info, NULL);
2854 if (!cm_node)
2855 return -ENOMEM;
2856
2857 /* set our node side to client (active) side */
2858 cm_node->tcp_cntxt.client = 1;
2859 cm_node->tcp_cntxt.rcv_wscale = IRDMA_CM_DEFAULT_RCV_WND_SCALE;
2860
2861 irdma_record_ird_ord(cm_node, conn_param->ird, conn_param->ord);
2862
2863 cm_node->pdata.size = private_data_len;
2864 cm_node->pdata.addr = cm_node->pdata_buf;
2865
2866 memcpy(cm_node->pdata_buf, private_data, private_data_len);
2867 *caller_cm_node = cm_node;
2868
2869 return 0;
2870 }
2871
2872 /**
2873 * irdma_cm_reject - reject and teardown a connection
2874 * @cm_node: connection's node
2875 * @pdata: ptr to private data for reject
2876 * @plen: size of private data
2877 */
2878 static int
irdma_cm_reject(struct irdma_cm_node * cm_node,const void * pdata,u8 plen)2879 irdma_cm_reject(struct irdma_cm_node *cm_node, const void *pdata,
2880 u8 plen)
2881 {
2882 int ret;
2883 int passive_state;
2884
2885 if (cm_node->tcp_cntxt.client)
2886 return 0;
2887
2888 irdma_cleanup_retrans_entry(cm_node);
2889
2890 passive_state = atomic_add_return(1, &cm_node->passive_state);
2891 if (passive_state == IRDMA_SEND_RESET_EVENT) {
2892 cm_node->state = IRDMA_CM_STATE_CLOSED;
2893 irdma_rem_ref_cm_node(cm_node);
2894 return 0;
2895 }
2896
2897 if (cm_node->state == IRDMA_CM_STATE_LISTENER_DESTROYED) {
2898 irdma_rem_ref_cm_node(cm_node);
2899 return 0;
2900 }
2901
2902 ret = irdma_send_mpa_reject(cm_node, pdata, plen);
2903 if (!ret)
2904 return 0;
2905
2906 cm_node->state = IRDMA_CM_STATE_CLOSED;
2907 if (irdma_send_reset(cm_node))
2908 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2909 "send reset failed\n");
2910
2911 return ret;
2912 }
2913
2914 /**
2915 * irdma_cm_close - close of cm connection
2916 * @cm_node: connection's node
2917 */
2918 static int
irdma_cm_close(struct irdma_cm_node * cm_node)2919 irdma_cm_close(struct irdma_cm_node *cm_node)
2920 {
2921 switch (cm_node->state) {
2922 case IRDMA_CM_STATE_SYN_RCVD:
2923 case IRDMA_CM_STATE_SYN_SENT:
2924 case IRDMA_CM_STATE_ONE_SIDE_ESTABLISHED:
2925 case IRDMA_CM_STATE_ESTABLISHED:
2926 case IRDMA_CM_STATE_ACCEPTING:
2927 case IRDMA_CM_STATE_MPAREQ_SENT:
2928 case IRDMA_CM_STATE_MPAREQ_RCVD:
2929 irdma_cleanup_retrans_entry(cm_node);
2930 irdma_send_reset(cm_node);
2931 break;
2932 case IRDMA_CM_STATE_CLOSE_WAIT:
2933 cm_node->state = IRDMA_CM_STATE_LAST_ACK;
2934 irdma_send_fin(cm_node);
2935 break;
2936 case IRDMA_CM_STATE_FIN_WAIT1:
2937 case IRDMA_CM_STATE_FIN_WAIT2:
2938 case IRDMA_CM_STATE_LAST_ACK:
2939 case IRDMA_CM_STATE_TIME_WAIT:
2940 case IRDMA_CM_STATE_CLOSING:
2941 return -EINVAL;
2942 case IRDMA_CM_STATE_LISTENING:
2943 irdma_cleanup_retrans_entry(cm_node);
2944 irdma_send_reset(cm_node);
2945 break;
2946 case IRDMA_CM_STATE_MPAREJ_RCVD:
2947 case IRDMA_CM_STATE_UNKNOWN:
2948 case IRDMA_CM_STATE_INITED:
2949 case IRDMA_CM_STATE_CLOSED:
2950 case IRDMA_CM_STATE_LISTENER_DESTROYED:
2951 irdma_rem_ref_cm_node(cm_node);
2952 break;
2953 case IRDMA_CM_STATE_OFFLOADED:
2954 if (cm_node->send_entry)
2955 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2956 "CM send_entry in OFFLOADED state\n");
2957 irdma_rem_ref_cm_node(cm_node);
2958 break;
2959 }
2960
2961 return 0;
2962 }
2963
2964 /**
2965 * irdma_receive_ilq - recv an ETHERNET packet, and process it
2966 * through CM
2967 * @vsi: VSI structure of dev
2968 * @rbuf: receive buffer
2969 */
2970 void
irdma_receive_ilq(struct irdma_sc_vsi * vsi,struct irdma_puda_buf * rbuf)2971 irdma_receive_ilq(struct irdma_sc_vsi *vsi, struct irdma_puda_buf *rbuf)
2972 {
2973 struct irdma_cm_node *cm_node;
2974 struct irdma_cm_listener *listener;
2975 struct ip *iph;
2976 struct ip6_hdr *ip6h;
2977 struct tcphdr *tcph;
2978 struct irdma_cm_info cm_info = {0};
2979 struct irdma_device *iwdev = vsi->back_vsi;
2980 struct irdma_cm_core *cm_core = &iwdev->cm_core;
2981 struct ether_vlan_header *ethh;
2982 u16 vtag;
2983
2984 /* if vlan, then maclen = 18 else 14 */
2985 iph = (struct ip *)rbuf->iph;
2986 irdma_debug_buf(vsi->dev, IRDMA_DEBUG_ILQ, "RECEIVE ILQ BUFFER",
2987 rbuf->mem.va, rbuf->totallen);
2988 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) {
2989 if (rbuf->vlan_valid) {
2990 vtag = rbuf->vlan_id;
2991 cm_info.user_pri = (vtag & EVL_PRI_MASK) >>
2992 VLAN_PRIO_SHIFT;
2993 cm_info.vlan_id = vtag & EVL_VLID_MASK;
2994 } else {
2995 cm_info.vlan_id = 0xFFFF;
2996 }
2997 } else {
2998 ethh = rbuf->mem.va;
2999
3000 if (ethh->evl_proto == htons(ETH_P_8021Q)) {
3001 vtag = ntohs(ethh->evl_tag);
3002 cm_info.user_pri = (vtag & EVL_PRI_MASK) >>
3003 VLAN_PRIO_SHIFT;
3004 cm_info.vlan_id = vtag & EVL_VLID_MASK;
3005 irdma_debug(&cm_core->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3006 "vlan_id=%d\n", cm_info.vlan_id);
3007 } else {
3008 cm_info.vlan_id = 0xFFFF;
3009 }
3010 }
3011 tcph = (struct tcphdr *)rbuf->tcph;
3012
3013 if (rbuf->ipv4) {
3014 cm_info.loc_addr[0] = ntohl(iph->ip_dst.s_addr);
3015 cm_info.rem_addr[0] = ntohl(iph->ip_src.s_addr);
3016 cm_info.ipv4 = true;
3017 cm_info.tos = iph->ip_tos;
3018 } else {
3019 ip6h = (struct ip6_hdr *)rbuf->iph;
3020 irdma_copy_ip_ntohl(cm_info.loc_addr,
3021 ip6h->ip6_dst.__u6_addr.__u6_addr32);
3022 irdma_copy_ip_ntohl(cm_info.rem_addr,
3023 ip6h->ip6_src.__u6_addr.__u6_addr32);
3024 cm_info.ipv4 = false;
3025 cm_info.tos = (ip6h->ip6_vfc << 4) | ip6h->ip6_flow;
3026 }
3027 cm_info.loc_port = ntohs(tcph->th_dport);
3028 cm_info.rem_port = ntohs(tcph->th_sport);
3029 cm_node = irdma_find_node(cm_core, cm_info.rem_port, cm_info.rem_addr,
3030 cm_info.loc_port, cm_info.loc_addr, cm_info.vlan_id);
3031
3032 if (!cm_node) {
3033 /*
3034 * Only type of packet accepted are for the PASSIVE open (syn only)
3035 */
3036 if (!(tcp_get_flags(tcph) & TH_SYN) || tcp_get_flags(tcph) & TH_ACK)
3037 return;
3038
3039 listener = irdma_find_listener(cm_core,
3040 cm_info.loc_addr,
3041 cm_info.ipv4,
3042 cm_info.loc_port,
3043 cm_info.vlan_id,
3044 IRDMA_CM_LISTENER_ACTIVE_STATE);
3045 if (!listener) {
3046 cm_info.cm_id = NULL;
3047 irdma_debug(&cm_core->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3048 "no listener found\n");
3049 return;
3050 }
3051
3052 cm_info.cm_id = listener->cm_id;
3053 cm_node = irdma_make_cm_node(cm_core, iwdev, &cm_info,
3054 listener);
3055 if (!cm_node) {
3056 irdma_debug(&cm_core->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3057 "allocate node failed\n");
3058 atomic_dec(&listener->refcnt);
3059 return;
3060 }
3061
3062 if (!(tcp_get_flags(tcph) & (TH_RST | TH_FIN))) {
3063 cm_node->state = IRDMA_CM_STATE_LISTENING;
3064 } else {
3065 irdma_rem_ref_cm_node(cm_node);
3066 return;
3067 }
3068
3069 atomic_inc(&cm_node->refcnt);
3070 } else if (cm_node->state == IRDMA_CM_STATE_OFFLOADED) {
3071 irdma_rem_ref_cm_node(cm_node);
3072 return;
3073 }
3074
3075 irdma_process_pkt(cm_node, rbuf);
3076 irdma_rem_ref_cm_node(cm_node);
3077 }
3078
3079 static int
irdma_add_qh(struct irdma_cm_node * cm_node,bool active)3080 irdma_add_qh(struct irdma_cm_node *cm_node, bool active)
3081 {
3082 if (!active)
3083 irdma_add_conn_est_qh(cm_node);
3084 return 0;
3085 }
3086
3087 static void
irdma_cm_free_ah_nop(struct irdma_cm_node * cm_node)3088 irdma_cm_free_ah_nop(struct irdma_cm_node *cm_node)
3089 {
3090 }
3091
3092 /**
3093 * irdma_setup_cm_core - setup top level instance of a cm core
3094 * @iwdev: iwarp device structure
3095 * @rdma_ver: HW version
3096 */
3097 int
irdma_setup_cm_core(struct irdma_device * iwdev,u8 rdma_ver)3098 irdma_setup_cm_core(struct irdma_device *iwdev, u8 rdma_ver)
3099 {
3100 struct irdma_cm_core *cm_core = &iwdev->cm_core;
3101
3102 cm_core->iwdev = iwdev;
3103 cm_core->dev = &iwdev->rf->sc_dev;
3104
3105 /* Handles CM event work items send to Iwarp core */
3106 cm_core->event_wq = alloc_ordered_workqueue("iwarp-event-wq", 0);
3107 if (!cm_core->event_wq)
3108 return -ENOMEM;
3109
3110 INIT_LIST_HEAD(&cm_core->listen_list);
3111
3112 timer_setup(&cm_core->tcp_timer, irdma_cm_timer_tick, 0);
3113
3114 spin_lock_init(&cm_core->ht_lock);
3115 spin_lock_init(&cm_core->listen_list_lock);
3116 spin_lock_init(&cm_core->apbvt_lock);
3117 switch (rdma_ver) {
3118 case IRDMA_GEN_1:
3119 cm_core->form_cm_frame = irdma_form_uda_cm_frame;
3120 cm_core->cm_create_ah = irdma_add_qh;
3121 cm_core->cm_free_ah = irdma_cm_free_ah_nop;
3122 break;
3123 case IRDMA_GEN_2:
3124 default:
3125 cm_core->form_cm_frame = irdma_form_ah_cm_frame;
3126 cm_core->cm_create_ah = irdma_cm_create_ah;
3127 cm_core->cm_free_ah = irdma_cm_free_ah;
3128 }
3129
3130 return 0;
3131 }
3132
3133 /**
3134 * irdma_cleanup_cm_core - deallocate a top level instance of a
3135 * cm core
3136 * @cm_core: cm's core
3137 */
3138 void
irdma_cleanup_cm_core(struct irdma_cm_core * cm_core)3139 irdma_cleanup_cm_core(struct irdma_cm_core *cm_core)
3140 {
3141 if (!cm_core)
3142 return;
3143
3144 del_timer_sync(&cm_core->tcp_timer);
3145
3146 destroy_workqueue(cm_core->event_wq);
3147 cm_core->dev->ws_reset(&cm_core->iwdev->vsi);
3148 }
3149
3150 /**
3151 * irdma_init_tcp_ctx - setup qp context
3152 * @cm_node: connection's node
3153 * @tcp_info: offload info for tcp
3154 * @iwqp: associate qp for the connection
3155 */
3156 static void
irdma_init_tcp_ctx(struct irdma_cm_node * cm_node,struct irdma_tcp_offload_info * tcp_info,struct irdma_qp * iwqp)3157 irdma_init_tcp_ctx(struct irdma_cm_node *cm_node,
3158 struct irdma_tcp_offload_info *tcp_info,
3159 struct irdma_qp *iwqp)
3160 {
3161 tcp_info->ipv4 = cm_node->ipv4;
3162 tcp_info->drop_ooo_seg = !iwqp->iwdev->iw_ooo;
3163 tcp_info->wscale = true;
3164 tcp_info->ignore_tcp_opt = true;
3165 tcp_info->ignore_tcp_uns_opt = true;
3166 tcp_info->no_nagle = false;
3167
3168 tcp_info->ttl = IRDMA_DEFAULT_TTL;
3169 tcp_info->rtt_var = IRDMA_DEFAULT_RTT_VAR;
3170 tcp_info->ss_thresh = IRDMA_DEFAULT_SS_THRESH;
3171 tcp_info->rexmit_thresh = IRDMA_DEFAULT_REXMIT_THRESH;
3172
3173 tcp_info->tcp_state = IRDMA_TCP_STATE_ESTABLISHED;
3174 tcp_info->snd_wscale = cm_node->tcp_cntxt.snd_wscale;
3175 tcp_info->rcv_wscale = cm_node->tcp_cntxt.rcv_wscale;
3176
3177 tcp_info->snd_nxt = cm_node->tcp_cntxt.loc_seq_num;
3178 tcp_info->snd_wnd = cm_node->tcp_cntxt.snd_wnd;
3179 tcp_info->rcv_nxt = cm_node->tcp_cntxt.rcv_nxt;
3180 tcp_info->snd_max = cm_node->tcp_cntxt.loc_seq_num;
3181
3182 tcp_info->snd_una = cm_node->tcp_cntxt.loc_seq_num;
3183 tcp_info->cwnd = 2 * cm_node->tcp_cntxt.mss;
3184 tcp_info->snd_wl1 = cm_node->tcp_cntxt.rcv_nxt;
3185 tcp_info->snd_wl2 = cm_node->tcp_cntxt.loc_seq_num;
3186 tcp_info->max_snd_window = cm_node->tcp_cntxt.max_snd_wnd;
3187 tcp_info->rcv_wnd = cm_node->tcp_cntxt.rcv_wnd
3188 << cm_node->tcp_cntxt.rcv_wscale;
3189
3190 tcp_info->flow_label = 0;
3191 tcp_info->snd_mss = (u32)cm_node->tcp_cntxt.mss;
3192 tcp_info->tos = cm_node->tos;
3193 if (cm_node->vlan_id < VLAN_N_VID) {
3194 tcp_info->insert_vlan_tag = true;
3195 tcp_info->vlan_tag = cm_node->vlan_id;
3196 tcp_info->vlan_tag |= cm_node->user_pri << VLAN_PRIO_SHIFT;
3197 }
3198 tcp_info->src_port = cm_node->loc_port;
3199 tcp_info->dst_port = cm_node->rem_port;
3200 tcp_info->arp_idx = (u16)irdma_arp_table(iwqp->iwdev->rf,
3201 cm_node->rem_addr, NULL,
3202 IRDMA_ARP_RESOLVE);
3203 if (cm_node->ipv4) {
3204 tcp_info->dest_ip_addr[3] = cm_node->rem_addr[0];
3205 tcp_info->local_ipaddr[3] = cm_node->loc_addr[0];
3206 } else {
3207 memcpy(tcp_info->dest_ip_addr, cm_node->rem_addr,
3208 sizeof(tcp_info->dest_ip_addr));
3209 memcpy(tcp_info->local_ipaddr, cm_node->loc_addr,
3210 sizeof(tcp_info->local_ipaddr));
3211 }
3212 }
3213
3214 /**
3215 * irdma_cm_init_tsa_conn - setup qp for RTS
3216 * @iwqp: associate qp for the connection
3217 * @cm_node: connection's node
3218 */
3219 static void
irdma_cm_init_tsa_conn(struct irdma_qp * iwqp,struct irdma_cm_node * cm_node)3220 irdma_cm_init_tsa_conn(struct irdma_qp *iwqp,
3221 struct irdma_cm_node *cm_node)
3222 {
3223 struct irdma_iwarp_offload_info *iwarp_info;
3224 struct irdma_qp_host_ctx_info *ctx_info;
3225
3226 iwarp_info = &iwqp->iwarp_info;
3227 ctx_info = &iwqp->ctx_info;
3228
3229 ctx_info->tcp_info = &iwqp->tcp_info;
3230 ctx_info->send_cq_num = iwqp->iwscq->sc_cq.cq_uk.cq_id;
3231 ctx_info->rcv_cq_num = iwqp->iwrcq->sc_cq.cq_uk.cq_id;
3232
3233 iwarp_info->ord_size = cm_node->ord_size;
3234 iwarp_info->ird_size = cm_node->ird_size;
3235 iwarp_info->rd_en = true;
3236 iwarp_info->rdmap_ver = 1;
3237 iwarp_info->ddp_ver = 1;
3238 iwarp_info->pd_id = iwqp->iwpd->sc_pd.pd_id;
3239
3240 ctx_info->tcp_info_valid = true;
3241 ctx_info->iwarp_info_valid = true;
3242 ctx_info->user_pri = cm_node->user_pri;
3243
3244 irdma_init_tcp_ctx(cm_node, &iwqp->tcp_info, iwqp);
3245 if (cm_node->snd_mark_en) {
3246 iwarp_info->snd_mark_en = true;
3247 iwarp_info->snd_mark_offset = (iwqp->tcp_info.snd_nxt & SNDMARKER_SEQNMASK) +
3248 cm_node->lsmm_size;
3249 }
3250
3251 cm_node->state = IRDMA_CM_STATE_OFFLOADED;
3252 iwqp->tcp_info.tcp_state = IRDMA_TCP_STATE_ESTABLISHED;
3253 iwqp->tcp_info.src_mac_addr_idx = iwqp->iwdev->mac_ip_table_idx;
3254
3255 if (cm_node->rcv_mark_en) {
3256 iwarp_info->rcv_mark_en = true;
3257 iwarp_info->align_hdrs = true;
3258 }
3259
3260 irdma_sc_qp_setctx(&iwqp->sc_qp, iwqp->host_ctx.va, ctx_info);
3261
3262 /* once tcp_info is set, no need to do it again */
3263 ctx_info->tcp_info_valid = false;
3264 ctx_info->iwarp_info_valid = false;
3265 }
3266
3267 /**
3268 * irdma_cm_disconn - when a connection is being closed
3269 * @iwqp: associated qp for the connection
3270 */
3271 void
irdma_cm_disconn(struct irdma_qp * iwqp)3272 irdma_cm_disconn(struct irdma_qp *iwqp)
3273 {
3274 struct irdma_device *iwdev = iwqp->iwdev;
3275 struct disconn_work *work;
3276 unsigned long flags;
3277
3278 work = kzalloc(sizeof(*work), GFP_ATOMIC);
3279 if (!work)
3280 return;
3281
3282 spin_lock_irqsave(&iwdev->rf->qptable_lock, flags);
3283 if (!iwdev->rf->qp_table[iwqp->ibqp.qp_num]) {
3284 spin_unlock_irqrestore(&iwdev->rf->qptable_lock, flags);
3285 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3286 "qp_id %d is already freed\n", iwqp->ibqp.qp_num);
3287 kfree(work);
3288 return;
3289 }
3290 irdma_qp_add_ref(&iwqp->ibqp);
3291 spin_unlock_irqrestore(&iwdev->rf->qptable_lock, flags);
3292
3293 work->iwqp = iwqp;
3294 INIT_WORK(&work->work, irdma_disconnect_worker);
3295 queue_work(iwdev->cleanup_wq, &work->work);
3296 }
3297
3298 /**
3299 * irdma_qp_disconnect - free qp and close cm
3300 * @iwqp: associate qp for the connection
3301 */
3302 static void
irdma_qp_disconnect(struct irdma_qp * iwqp)3303 irdma_qp_disconnect(struct irdma_qp *iwqp)
3304 {
3305 struct irdma_device *iwdev = iwqp->iwdev;
3306
3307 iwqp->active_conn = 0;
3308 /* close the CM node down if it is still active */
3309 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM, "Call close API\n");
3310 irdma_cm_close(iwqp->cm_node);
3311 }
3312
3313 /**
3314 * irdma_cm_disconn_true - called by worker thread to disconnect qp
3315 * @iwqp: associate qp for the connection
3316 */
3317 static void
irdma_cm_disconn_true(struct irdma_qp * iwqp)3318 irdma_cm_disconn_true(struct irdma_qp *iwqp)
3319 {
3320 struct iw_cm_id *cm_id;
3321 struct irdma_device *iwdev;
3322 struct irdma_sc_qp *qp = &iwqp->sc_qp;
3323 u16 last_ae;
3324 u8 original_hw_tcp_state;
3325 u8 original_ibqp_state;
3326 int disconn_status = 0;
3327 int issue_disconn = 0;
3328 int issue_close = 0;
3329 int issue_flush = 0;
3330 unsigned long flags;
3331 int err;
3332
3333 iwdev = iwqp->iwdev;
3334 spin_lock_irqsave(&iwqp->lock, flags);
3335 if (rdma_protocol_roce(&iwdev->ibdev, 1)) {
3336 struct ib_qp_attr attr;
3337
3338 if (iwqp->flush_issued || iwqp->sc_qp.qp_uk.destroy_pending) {
3339 spin_unlock_irqrestore(&iwqp->lock, flags);
3340 return;
3341 }
3342
3343 spin_unlock_irqrestore(&iwqp->lock, flags);
3344
3345 attr.qp_state = IB_QPS_ERR;
3346 irdma_modify_qp_roce(&iwqp->ibqp, &attr, IB_QP_STATE, NULL);
3347 irdma_ib_qp_event(iwqp, qp->event_type);
3348 return;
3349 }
3350
3351 cm_id = iwqp->cm_id;
3352 original_hw_tcp_state = iwqp->hw_tcp_state;
3353 original_ibqp_state = iwqp->ibqp_state;
3354 last_ae = iwqp->last_aeq;
3355
3356 if (qp->term_flags) {
3357 issue_disconn = 1;
3358 issue_close = 1;
3359 iwqp->cm_id = NULL;
3360 irdma_terminate_del_timer(qp);
3361 if (!iwqp->flush_issued) {
3362 iwqp->flush_issued = 1;
3363 issue_flush = 1;
3364 }
3365 } else if ((original_hw_tcp_state == IRDMA_TCP_STATE_CLOSE_WAIT) ||
3366 ((original_ibqp_state == IB_QPS_RTS) &&
3367 (last_ae == IRDMA_AE_LLP_CONNECTION_RESET))) {
3368 issue_disconn = 1;
3369 if (last_ae == IRDMA_AE_LLP_CONNECTION_RESET)
3370 disconn_status = -ECONNRESET;
3371 }
3372
3373 if (original_hw_tcp_state == IRDMA_TCP_STATE_CLOSED ||
3374 original_hw_tcp_state == IRDMA_TCP_STATE_TIME_WAIT ||
3375 last_ae == IRDMA_AE_RDMAP_ROE_BAD_LLP_CLOSE ||
3376 last_ae == IRDMA_AE_BAD_CLOSE ||
3377 last_ae == IRDMA_AE_LLP_CONNECTION_RESET || iwdev->rf->reset || !cm_id) {
3378 issue_close = 1;
3379 iwqp->cm_id = NULL;
3380 qp->term_flags = 0;
3381 if (!iwqp->flush_issued) {
3382 iwqp->flush_issued = 1;
3383 issue_flush = 1;
3384 }
3385 }
3386
3387 spin_unlock_irqrestore(&iwqp->lock, flags);
3388 if (issue_flush && !iwqp->sc_qp.qp_uk.destroy_pending) {
3389 irdma_flush_wqes(iwqp, IRDMA_FLUSH_SQ | IRDMA_FLUSH_RQ |
3390 IRDMA_FLUSH_WAIT);
3391
3392 if (qp->term_flags)
3393 irdma_ib_qp_event(iwqp, qp->event_type);
3394 }
3395
3396 if (!cm_id || !cm_id->event_handler)
3397 return;
3398
3399 spin_lock_irqsave(&iwdev->cm_core.ht_lock, flags);
3400 if (!iwqp->cm_node) {
3401 spin_unlock_irqrestore(&iwdev->cm_core.ht_lock, flags);
3402 return;
3403 }
3404 atomic_inc(&iwqp->cm_node->refcnt);
3405
3406 spin_unlock_irqrestore(&iwdev->cm_core.ht_lock, flags);
3407
3408 if (issue_disconn) {
3409 err = irdma_send_cm_event(iwqp->cm_node, cm_id,
3410 IW_CM_EVENT_DISCONNECT,
3411 disconn_status);
3412 if (err)
3413 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3414 "disconnect event failed: - cm_id = %p\n",
3415 cm_id);
3416 }
3417 if (issue_close) {
3418 cm_id->provider_data = iwqp;
3419 err = irdma_send_cm_event(iwqp->cm_node, cm_id,
3420 IW_CM_EVENT_CLOSE, 0);
3421 if (err)
3422 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3423 "close event failed: - cm_id = %p\n",
3424 cm_id);
3425 irdma_qp_disconnect(iwqp);
3426 }
3427 irdma_rem_ref_cm_node(iwqp->cm_node);
3428 }
3429
3430 /**
3431 * irdma_disconnect_worker - worker for connection close
3432 * @work: points or disconn structure
3433 */
3434 static void
irdma_disconnect_worker(struct work_struct * work)3435 irdma_disconnect_worker(struct work_struct *work)
3436 {
3437 struct disconn_work *dwork = container_of(work, struct disconn_work, work);
3438 struct irdma_qp *iwqp = dwork->iwqp;
3439
3440 kfree(dwork);
3441 irdma_cm_disconn_true(iwqp);
3442 irdma_qp_rem_ref(&iwqp->ibqp);
3443 }
3444
3445 /**
3446 * irdma_free_lsmm_rsrc - free lsmm memory and deregister
3447 * @iwqp: associate qp for the connection
3448 */
3449 void
irdma_free_lsmm_rsrc(struct irdma_qp * iwqp)3450 irdma_free_lsmm_rsrc(struct irdma_qp *iwqp)
3451 {
3452 struct irdma_device *iwdev;
3453
3454 iwdev = iwqp->iwdev;
3455
3456 if (iwqp->ietf_mem.va) {
3457 if (iwqp->lsmm_mr)
3458 iwdev->ibdev.dereg_mr(iwqp->lsmm_mr, NULL);
3459
3460 irdma_free_dma_mem(iwdev->rf->sc_dev.hw,
3461 &iwqp->ietf_mem);
3462 iwqp->ietf_mem.va = NULL;
3463 }
3464 }
3465
3466 /**
3467 * irdma_accept - registered call for connection to be accepted
3468 * @cm_id: cm information for passive connection
3469 * @conn_param: accpet parameters
3470 */
3471 int
irdma_accept(struct iw_cm_id * cm_id,struct iw_cm_conn_param * conn_param)3472 irdma_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param)
3473 {
3474 struct ib_qp *ibqp;
3475 struct irdma_qp *iwqp;
3476 struct irdma_device *iwdev;
3477 struct irdma_sc_dev *dev;
3478 struct irdma_cm_node *cm_node;
3479 struct ib_qp_attr attr = {0};
3480 int passive_state;
3481 struct ib_mr *ibmr;
3482 struct irdma_pd *iwpd;
3483 u16 buf_len = 0;
3484 struct irdma_kmem_info accept;
3485 u64 tagged_offset;
3486 int wait_ret;
3487 int ret = 0;
3488
3489 ibqp = irdma_get_qp(cm_id->device, conn_param->qpn);
3490 if (!ibqp)
3491 return -EINVAL;
3492
3493 iwqp = to_iwqp(ibqp);
3494 iwdev = iwqp->iwdev;
3495 dev = &iwdev->rf->sc_dev;
3496 cm_node = cm_id->provider_data;
3497
3498 if (((struct sockaddr_in *)&cm_id->local_addr)->sin_family == AF_INET) {
3499 cm_node->ipv4 = true;
3500 cm_node->vlan_id = irdma_get_vlan_ipv4(cm_id, cm_node->loc_addr);
3501 } else {
3502 cm_node->ipv4 = false;
3503 irdma_get_vlan_mac_ipv6(cm_id, cm_node->loc_addr, &cm_node->vlan_id,
3504 NULL);
3505 }
3506 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM, "Accept vlan_id=%d\n",
3507 cm_node->vlan_id);
3508
3509 if (cm_node->state == IRDMA_CM_STATE_LISTENER_DESTROYED) {
3510 ret = -EINVAL;
3511 goto error;
3512 }
3513
3514 passive_state = atomic_add_return(1, &cm_node->passive_state);
3515 if (passive_state == IRDMA_SEND_RESET_EVENT) {
3516 ret = -ECONNRESET;
3517 goto error;
3518 }
3519
3520 buf_len = conn_param->private_data_len + IRDMA_MAX_IETF_SIZE;
3521 iwqp->ietf_mem.size = buf_len;
3522 iwqp->ietf_mem.va = irdma_allocate_dma_mem(dev->hw, &iwqp->ietf_mem,
3523 iwqp->ietf_mem.size, 1);
3524 if (!iwqp->ietf_mem.va) {
3525 ret = -ENOMEM;
3526 goto error;
3527 }
3528
3529 cm_node->pdata.size = conn_param->private_data_len;
3530 accept.addr = iwqp->ietf_mem.va;
3531 accept.size = irdma_cm_build_mpa_frame(cm_node, &accept, MPA_KEY_REPLY);
3532 memcpy((u8 *)accept.addr + accept.size, conn_param->private_data,
3533 conn_param->private_data_len);
3534
3535 if (cm_node->dev->ws_add(iwqp->sc_qp.vsi, cm_node->user_pri)) {
3536 ret = -ENOMEM;
3537 goto error;
3538 }
3539 iwqp->sc_qp.user_pri = cm_node->user_pri;
3540 irdma_qp_add_qos(&iwqp->sc_qp);
3541 if (cm_node->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_2)
3542 iwdev->rf->check_fc(&iwdev->vsi, &iwqp->sc_qp);
3543 /* setup our first outgoing iWarp send WQE (the IETF frame response) */
3544 iwpd = iwqp->iwpd;
3545 tagged_offset = (uintptr_t)iwqp->ietf_mem.va;
3546 ibmr = irdma_reg_phys_mr(&iwpd->ibpd, iwqp->ietf_mem.pa, buf_len,
3547 IB_ACCESS_LOCAL_WRITE, &tagged_offset);
3548 if (IS_ERR(ibmr)) {
3549 ret = -ENOMEM;
3550 goto error;
3551 }
3552
3553 ibmr->pd = &iwpd->ibpd;
3554 ibmr->device = iwpd->ibpd.device;
3555 iwqp->lsmm_mr = ibmr;
3556 if (iwqp->page)
3557 iwqp->sc_qp.qp_uk.sq_base = kmap_local_page(iwqp->page);
3558
3559 cm_node->lsmm_size = accept.size + conn_param->private_data_len;
3560 irdma_sc_send_lsmm(&iwqp->sc_qp, iwqp->ietf_mem.va, cm_node->lsmm_size,
3561 ibmr->lkey);
3562
3563 if (iwqp->page)
3564 kunmap_local(iwqp->sc_qp.qp_uk.sq_base);
3565
3566 iwqp->cm_id = cm_id;
3567 cm_node->cm_id = cm_id;
3568
3569 cm_id->provider_data = iwqp;
3570 iwqp->active_conn = 0;
3571 iwqp->cm_node = cm_node;
3572 cm_node->iwqp = iwqp;
3573 irdma_cm_init_tsa_conn(iwqp, cm_node);
3574 irdma_qp_add_ref(&iwqp->ibqp);
3575 cm_id->add_ref(cm_id);
3576
3577 attr.qp_state = IB_QPS_RTS;
3578 cm_node->qhash_set = false;
3579 cm_node->cm_core->cm_free_ah(cm_node);
3580
3581 irdma_modify_qp(&iwqp->ibqp, &attr, IB_QP_STATE, NULL);
3582 if (dev->hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE) {
3583 wait_ret = wait_event_interruptible_timeout(iwqp->waitq,
3584 iwqp->rts_ae_rcvd,
3585 IRDMA_MAX_TIMEOUT);
3586 if (!wait_ret) {
3587 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3588 "Slow Connection: cm_node=%p, loc_port=%d, rem_port=%d, cm_id=%p\n",
3589 cm_node, cm_node->loc_port,
3590 cm_node->rem_port, cm_node->cm_id);
3591 ret = -ECONNRESET;
3592 goto error;
3593 }
3594 }
3595
3596 irdma_send_cm_event(cm_node, cm_id, IW_CM_EVENT_ESTABLISHED, 0);
3597 cm_node->accelerated = true;
3598 complete(&cm_node->establish_comp);
3599
3600 if (cm_node->accept_pend) {
3601 atomic_dec(&cm_node->listener->pend_accepts_cnt);
3602 cm_node->accept_pend = 0;
3603 }
3604
3605 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3606 "rem_port=0x%04x, loc_port=0x%04x rem_addr=%x loc_addr=%x cm_node=%p cm_id=%p qp_id=%d\n\n",
3607 cm_node->rem_port, cm_node->loc_port, cm_node->rem_addr[0],
3608 cm_node->loc_addr[0], cm_node, cm_id, ibqp->qp_num);
3609 cm_node->cm_core->stats_accepts++;
3610
3611 return 0;
3612 error:
3613 irdma_free_lsmm_rsrc(iwqp);
3614 irdma_rem_ref_cm_node(cm_node);
3615
3616 return ret;
3617 }
3618
3619 /**
3620 * irdma_reject - registered call for connection to be rejected
3621 * @cm_id: cm information for passive connection
3622 * @pdata: private data to be sent
3623 * @pdata_len: private data length
3624 */
3625 int
irdma_reject(struct iw_cm_id * cm_id,const void * pdata,u8 pdata_len)3626 irdma_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len)
3627 {
3628 struct irdma_device *iwdev;
3629 struct irdma_cm_node *cm_node;
3630
3631 cm_node = cm_id->provider_data;
3632 cm_node->pdata.size = pdata_len;
3633
3634 iwdev = to_iwdev(cm_id->device);
3635 if (!iwdev)
3636 return -EINVAL;
3637
3638 cm_node->cm_core->stats_rejects++;
3639
3640 if (pdata_len + sizeof(struct ietf_mpa_v2) > IRDMA_MAX_CM_BUF)
3641 return -EINVAL;
3642
3643 return irdma_cm_reject(cm_node, pdata, pdata_len);
3644 }
3645
3646 /**
3647 * irdma_connect - registered call for connection to be established
3648 * @cm_id: cm information for passive connection
3649 * @conn_param: Information about the connection
3650 */
3651 int
irdma_connect(struct iw_cm_id * cm_id,struct iw_cm_conn_param * conn_param)3652 irdma_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param)
3653 {
3654 struct ib_qp *ibqp;
3655 struct irdma_qp *iwqp;
3656 struct irdma_device *iwdev;
3657 struct irdma_cm_node *cm_node;
3658 struct irdma_cm_info cm_info;
3659 struct sockaddr_in *laddr;
3660 struct sockaddr_in *raddr;
3661 struct sockaddr_in6 *laddr6;
3662 struct sockaddr_in6 *raddr6;
3663 int ret = 0;
3664
3665 ibqp = irdma_get_qp(cm_id->device, conn_param->qpn);
3666 if (!ibqp)
3667 return -EINVAL;
3668 iwqp = to_iwqp(ibqp);
3669 if (!iwqp)
3670 return -EINVAL;
3671 iwdev = iwqp->iwdev;
3672 if (!iwdev)
3673 return -EINVAL;
3674
3675 laddr = (struct sockaddr_in *)&cm_id->m_local_addr;
3676 raddr = (struct sockaddr_in *)&cm_id->m_remote_addr;
3677 laddr6 = (struct sockaddr_in6 *)&cm_id->m_local_addr;
3678 raddr6 = (struct sockaddr_in6 *)&cm_id->m_remote_addr;
3679
3680 if (!(laddr->sin_port) || !(raddr->sin_port))
3681 return -EINVAL;
3682
3683 iwqp->active_conn = 1;
3684 iwqp->cm_id = NULL;
3685 cm_id->provider_data = iwqp;
3686
3687 /* set up the connection params for the node */
3688 if (cm_id->remote_addr.ss_family == AF_INET) {
3689 if (iwdev->vsi.mtu < IRDMA_MIN_MTU_IPV4)
3690 return -EINVAL;
3691
3692 cm_info.ipv4 = true;
3693 memset(cm_info.loc_addr, 0, sizeof(cm_info.loc_addr));
3694 memset(cm_info.rem_addr, 0, sizeof(cm_info.rem_addr));
3695 cm_info.loc_addr[0] = ntohl(laddr->sin_addr.s_addr);
3696 cm_info.rem_addr[0] = ntohl(raddr->sin_addr.s_addr);
3697 cm_info.loc_port = ntohs(laddr->sin_port);
3698 cm_info.rem_port = ntohs(raddr->sin_port);
3699 cm_info.vlan_id = irdma_get_vlan_ipv4(cm_id, cm_info.loc_addr);
3700 } else {
3701 if (iwdev->vsi.mtu < IRDMA_MIN_MTU_IPV6)
3702 return -EINVAL;
3703
3704 cm_info.ipv4 = false;
3705 irdma_copy_ip_ntohl(cm_info.loc_addr,
3706 laddr6->sin6_addr.__u6_addr.__u6_addr32);
3707 irdma_copy_ip_ntohl(cm_info.rem_addr,
3708 raddr6->sin6_addr.__u6_addr.__u6_addr32);
3709 cm_info.loc_port = ntohs(laddr6->sin6_port);
3710 cm_info.rem_port = ntohs(raddr6->sin6_port);
3711 irdma_get_vlan_mac_ipv6(cm_id, cm_info.loc_addr, &cm_info.vlan_id, NULL);
3712 }
3713 cm_info.cm_id = cm_id;
3714 cm_info.qh_qpid = iwdev->vsi.ilq->qp_id;
3715 cm_info.tos = cm_id->tos;
3716 if (iwdev->vsi.dscp_mode) {
3717 cm_info.user_pri =
3718 iwqp->sc_qp.vsi->dscp_map[irdma_tos2dscp(cm_info.tos)];
3719 } else {
3720 cm_info.user_pri = rt_tos2priority(cm_id->tos);
3721 cm_info.user_pri =
3722 irdma_iw_get_vlan_prio(cm_info.loc_addr,
3723 cm_info.user_pri,
3724 cm_info.ipv4);
3725 }
3726
3727 if (iwqp->sc_qp.dev->ws_add(iwqp->sc_qp.vsi, cm_info.user_pri))
3728 return -ENOMEM;
3729 iwqp->sc_qp.user_pri = cm_info.user_pri;
3730 irdma_qp_add_qos(&iwqp->sc_qp);
3731 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_2)
3732 iwdev->rf->check_fc(&iwdev->vsi, &iwqp->sc_qp);
3733 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_DCB, "TOS:[%d] UP:[%d]\n",
3734 cm_id->tos, cm_info.user_pri);
3735
3736 ret = irdma_create_cm_node(&iwdev->cm_core, iwdev, conn_param, &cm_info,
3737 &cm_node);
3738 if (ret)
3739 return ret;
3740 ret = cm_node->cm_core->cm_create_ah(cm_node, true);
3741 if (ret)
3742 goto err;
3743 if (irdma_manage_qhash(iwdev, &cm_info,
3744 IRDMA_QHASH_TYPE_TCP_ESTABLISHED,
3745 IRDMA_QHASH_MANAGE_TYPE_ADD, NULL, true)) {
3746 ret = -EINVAL;
3747 goto err;
3748 }
3749 cm_node->qhash_set = true;
3750
3751 cm_node->apbvt_entry = irdma_add_apbvt(iwdev, cm_info.loc_port);
3752 if (!cm_node->apbvt_entry) {
3753 ret = -EINVAL;
3754 goto err;
3755 }
3756
3757 cm_node->apbvt_set = true;
3758 iwqp->cm_node = cm_node;
3759 cm_node->iwqp = iwqp;
3760 iwqp->cm_id = cm_id;
3761 irdma_qp_add_ref(&iwqp->ibqp);
3762 cm_id->add_ref(cm_id);
3763
3764 if (cm_node->state != IRDMA_CM_STATE_OFFLOADED) {
3765 cm_node->state = IRDMA_CM_STATE_SYN_SENT;
3766 ret = irdma_send_syn(cm_node, 0);
3767 if (ret)
3768 goto err;
3769 }
3770
3771 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3772 "rem_port=0x%04x, loc_port=0x%04x rem_addr=%x loc_addr=%x cm_node=%p cm_id=%p qp_id = %d\n\n",
3773 cm_node->rem_port, cm_node->loc_port, cm_node->rem_addr[0],
3774 cm_node->loc_addr[0], cm_node, cm_id, ibqp->qp_num);
3775
3776 return 0;
3777
3778 err:
3779 if (cm_info.ipv4)
3780 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3781 "connect() FAILED: dest addr=%x",
3782 cm_info.rem_addr[0]);
3783 else
3784 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3785 "connect() FAILED: dest addr=%x:%x:%x:%x",
3786 IRDMA_PRINT_IP6(cm_info.rem_addr));
3787 irdma_rem_ref_cm_node(cm_node);
3788 iwdev->cm_core.stats_connect_errs++;
3789
3790 return ret;
3791 }
3792
3793 /**
3794 * irdma_create_listen - registered call creating listener
3795 * @cm_id: cm information for passive connection
3796 * @backlog: to max accept pending count
3797 */
3798 int
irdma_create_listen(struct iw_cm_id * cm_id,int backlog)3799 irdma_create_listen(struct iw_cm_id *cm_id, int backlog)
3800 {
3801 struct irdma_device *iwdev;
3802 struct irdma_cm_listener *cm_listen_node;
3803 struct irdma_cm_info cm_info = {0};
3804 struct sockaddr_in *laddr;
3805 struct sockaddr_in6 *laddr6;
3806 bool wildcard = false;
3807 int err;
3808
3809 iwdev = to_iwdev(cm_id->device);
3810 if (!iwdev)
3811 return -EINVAL;
3812
3813 laddr = (struct sockaddr_in *)&cm_id->m_local_addr;
3814 laddr6 = (struct sockaddr_in6 *)&cm_id->m_local_addr;
3815 cm_info.qh_qpid = iwdev->vsi.ilq->qp_id;
3816
3817 if (laddr->sin_family == AF_INET) {
3818 if (iwdev->vsi.mtu < IRDMA_MIN_MTU_IPV4)
3819 return -EINVAL;
3820
3821 cm_info.ipv4 = true;
3822 cm_info.loc_addr[0] = ntohl(laddr->sin_addr.s_addr);
3823 cm_info.loc_port = ntohs(laddr->sin_port);
3824
3825 if (laddr->sin_addr.s_addr != htonl(INADDR_ANY)) {
3826 cm_info.vlan_id = irdma_get_vlan_ipv4(cm_id, cm_info.loc_addr);
3827 } else {
3828 cm_info.vlan_id = 0xFFFF;
3829 wildcard = true;
3830 }
3831 } else {
3832 if (iwdev->vsi.mtu < IRDMA_MIN_MTU_IPV6)
3833 return -EINVAL;
3834
3835 cm_info.ipv4 = false;
3836 irdma_copy_ip_ntohl(cm_info.loc_addr,
3837 laddr6->sin6_addr.__u6_addr.__u6_addr32);
3838 cm_info.loc_port = ntohs(laddr6->sin6_port);
3839 if (!IN6_IS_ADDR_UNSPECIFIED(&laddr6->sin6_addr)) {
3840 irdma_get_vlan_mac_ipv6(cm_id, cm_info.loc_addr,
3841 &cm_info.vlan_id, NULL);
3842 } else {
3843 cm_info.vlan_id = 0xFFFF;
3844 wildcard = true;
3845 }
3846 }
3847
3848 if (cm_info.vlan_id >= VLAN_N_VID && iwdev->dcb_vlan_mode)
3849 cm_info.vlan_id = 0;
3850 cm_info.backlog = backlog;
3851 cm_info.cm_id = cm_id;
3852
3853 cm_listen_node = irdma_make_listen_node(&iwdev->cm_core, iwdev,
3854 &cm_info);
3855 if (!cm_listen_node) {
3856 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3857 "cm_listen_node == NULL\n");
3858 return -ENOMEM;
3859 }
3860
3861 cm_id->provider_data = cm_listen_node;
3862
3863 cm_listen_node->tos = cm_id->tos;
3864 if (iwdev->vsi.dscp_mode)
3865 cm_listen_node->user_pri =
3866 iwdev->vsi.dscp_map[irdma_tos2dscp(cm_id->tos)];
3867 else
3868 cm_listen_node->user_pri = rt_tos2priority(cm_id->tos);
3869 cm_info.user_pri = cm_listen_node->user_pri;
3870 if (!cm_listen_node->reused_node) {
3871 if (wildcard) {
3872 err = irdma_add_mqh(iwdev, &cm_info, cm_listen_node);
3873 if (err)
3874 goto error;
3875 } else {
3876 if (!iwdev->vsi.dscp_mode)
3877 cm_info.user_pri = cm_listen_node->user_pri =
3878 irdma_iw_get_vlan_prio(cm_info.loc_addr,
3879 cm_info.user_pri,
3880 cm_info.ipv4);
3881 err = irdma_manage_qhash(iwdev, &cm_info,
3882 IRDMA_QHASH_TYPE_TCP_SYN,
3883 IRDMA_QHASH_MANAGE_TYPE_ADD,
3884 NULL, true);
3885 if (err)
3886 goto error;
3887
3888 cm_listen_node->qhash_set = true;
3889 }
3890
3891 cm_listen_node->apbvt_entry = irdma_add_apbvt(iwdev,
3892 cm_info.loc_port);
3893 if (!cm_listen_node->apbvt_entry)
3894 goto error;
3895 }
3896 cm_id->add_ref(cm_id);
3897 cm_listen_node->cm_core->stats_listen_created++;
3898 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3899 "loc_port=0x%04x loc_addr=%x cm_listen_node=%p cm_id=%p qhash_set=%d vlan_id=%d\n",
3900 cm_listen_node->loc_port, cm_listen_node->loc_addr[0],
3901 cm_listen_node, cm_listen_node->cm_id,
3902 cm_listen_node->qhash_set, cm_listen_node->vlan_id);
3903
3904 return 0;
3905
3906 error:
3907
3908 irdma_cm_del_listen(&iwdev->cm_core, cm_listen_node, false);
3909
3910 return -EINVAL;
3911 }
3912
3913 /**
3914 * irdma_destroy_listen - registered call to destroy listener
3915 * @cm_id: cm information for passive connection
3916 */
3917 int
irdma_destroy_listen(struct iw_cm_id * cm_id)3918 irdma_destroy_listen(struct iw_cm_id *cm_id)
3919 {
3920 struct irdma_device *iwdev;
3921
3922 iwdev = to_iwdev(cm_id->device);
3923 if (cm_id->provider_data)
3924 irdma_cm_del_listen(&iwdev->cm_core, cm_id->provider_data,
3925 true);
3926 else
3927 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3928 "cm_id->provider_data was NULL\n");
3929
3930 cm_id->rem_ref(cm_id);
3931
3932 return 0;
3933 }
3934
3935 /**
3936 * irdma_iw_teardown_list_prep - add conn nodes slated for tear
3937 * down to list
3938 * @cm_core: cm's core
3939 * @teardown_list: a list to which cm_node will be selected
3940 * @ipaddr: pointer to ip address
3941 * @nfo: pointer to cm_info structure instance
3942 * @disconnect_all: flag indicating disconnect all QPs
3943 */
3944 static void
irdma_iw_teardown_list_prep(struct irdma_cm_core * cm_core,struct list_head * teardown_list,u32 * ipaddr,struct irdma_cm_info * nfo,bool disconnect_all)3945 irdma_iw_teardown_list_prep(struct irdma_cm_core *cm_core,
3946 struct list_head *teardown_list,
3947 u32 *ipaddr,
3948 struct irdma_cm_info *nfo,
3949 bool disconnect_all)
3950 {
3951 struct irdma_cm_node *cm_node;
3952 int bkt;
3953
3954 HASH_FOR_EACH_RCU(cm_core->cm_hash_tbl, bkt, cm_node, list) {
3955 if ((disconnect_all ||
3956 (nfo->vlan_id == cm_node->vlan_id &&
3957 !memcmp(cm_node->loc_addr, ipaddr, nfo->ipv4 ? 4 : 16))) &&
3958 atomic_inc_not_zero(&cm_node->refcnt))
3959 list_add(&cm_node->teardown_entry, teardown_list);
3960 }
3961 }
3962
3963 static inline bool
irdma_ip_vlan_match(u32 * ip1,u16 vlan_id1,bool check_vlan,u32 * ip2,u16 vlan_id2,bool ipv4)3964 irdma_ip_vlan_match(u32 *ip1, u16 vlan_id1,
3965 bool check_vlan, u32 *ip2,
3966 u16 vlan_id2, bool ipv4)
3967 {
3968 return (!check_vlan || vlan_id1 == vlan_id2) &&
3969 !memcmp(ip1, ip2, ipv4 ? 4 : 16);
3970 }
3971
3972 /**
3973 * irdma_roce_teardown_list_prep - add conn nodes slated for
3974 * tear down to list
3975 * @iwdev: RDMA device
3976 * @teardown_list: a list to which cm_node will be selected
3977 * @ipaddr: pointer to ip address
3978 * @nfo: pointer to cm_info structure instance
3979 * @disconnect_all: flag indicating disconnect all QPs
3980 */
3981 static void
irdma_roce_teardown_list_prep(struct irdma_device * iwdev,struct list_head * teardown_list,u32 * ipaddr,struct irdma_cm_info * nfo,bool disconnect_all)3982 irdma_roce_teardown_list_prep(struct irdma_device *iwdev,
3983 struct list_head *teardown_list,
3984 u32 *ipaddr,
3985 struct irdma_cm_info *nfo,
3986 bool disconnect_all)
3987 {
3988 struct irdma_sc_vsi *vsi = &iwdev->vsi;
3989 struct irdma_sc_qp *sc_qp;
3990 struct list_head *list_node;
3991 struct irdma_qp *qp;
3992 unsigned long flags;
3993 int i;
3994
3995 for (i = 0; i < IRDMA_MAX_USER_PRIORITY; i++) {
3996 mutex_lock(&vsi->qos[i].qos_mutex);
3997 list_for_each(list_node, &vsi->qos[i].qplist) {
3998 u32 qp_ip[4];
3999
4000 sc_qp = container_of(list_node, struct irdma_sc_qp,
4001 list);
4002 if (sc_qp->qp_uk.qp_type != IRDMA_QP_TYPE_ROCE_RC)
4003 continue;
4004
4005 qp = sc_qp->qp_uk.back_qp;
4006 if (!disconnect_all) {
4007 if (nfo->ipv4)
4008 qp_ip[0] = qp->udp_info.local_ipaddr[3];
4009 else
4010 memcpy(qp_ip,
4011 &qp->udp_info.local_ipaddr[0],
4012 sizeof(qp_ip));
4013 }
4014
4015 if (disconnect_all ||
4016 irdma_ip_vlan_match(qp_ip,
4017 qp->udp_info.vlan_tag & EVL_VLID_MASK,
4018 qp->udp_info.insert_vlan_tag,
4019 ipaddr, nfo->vlan_id, nfo->ipv4)) {
4020 spin_lock_irqsave(&iwdev->rf->qptable_lock, flags);
4021 if (iwdev->rf->qp_table[sc_qp->qp_uk.qp_id]) {
4022 irdma_qp_add_ref(&qp->ibqp);
4023 list_add(&qp->teardown_entry, teardown_list);
4024 }
4025 spin_unlock_irqrestore(&iwdev->rf->qptable_lock, flags);
4026 }
4027 }
4028 mutex_unlock(&vsi->qos[i].qos_mutex);
4029 }
4030 }
4031
4032 /**
4033 * irdma_cm_event_connected - handle connected active node
4034 * @event: the info for cm_node of connection
4035 */
4036 static void
irdma_cm_event_connected(struct irdma_cm_event * event)4037 irdma_cm_event_connected(struct irdma_cm_event *event)
4038 {
4039 struct irdma_qp *iwqp;
4040 struct irdma_device *iwdev;
4041 struct irdma_cm_node *cm_node;
4042 struct irdma_sc_dev *dev;
4043 struct ib_qp_attr attr = {0};
4044 struct iw_cm_id *cm_id;
4045 int status;
4046 bool read0;
4047 int wait_ret = 0;
4048
4049 cm_node = event->cm_node;
4050 cm_id = cm_node->cm_id;
4051 iwqp = cm_id->provider_data;
4052 iwdev = iwqp->iwdev;
4053 dev = &iwdev->rf->sc_dev;
4054 if (iwqp->sc_qp.qp_uk.destroy_pending) {
4055 status = -ETIMEDOUT;
4056 goto error;
4057 }
4058
4059 irdma_cm_init_tsa_conn(iwqp, cm_node);
4060 read0 = (cm_node->send_rdma0_op == SEND_RDMA_READ_ZERO);
4061 if (iwqp->page)
4062 iwqp->sc_qp.qp_uk.sq_base = kmap_local_page(iwqp->page);
4063 irdma_sc_send_rtt(&iwqp->sc_qp, read0);
4064 if (iwqp->page)
4065 kunmap_local(iwqp->sc_qp.qp_uk.sq_base);
4066
4067 attr.qp_state = IB_QPS_RTS;
4068 cm_node->qhash_set = false;
4069 irdma_modify_qp(&iwqp->ibqp, &attr, IB_QP_STATE, NULL);
4070 if (dev->hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE) {
4071 wait_ret = wait_event_interruptible_timeout(iwqp->waitq,
4072 iwqp->rts_ae_rcvd,
4073 IRDMA_MAX_TIMEOUT);
4074 if (!wait_ret)
4075 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
4076 "Slow Connection: cm_node=%p, loc_port=%d, rem_port=%d, cm_id=%p\n",
4077 cm_node, cm_node->loc_port,
4078 cm_node->rem_port, cm_node->cm_id);
4079 }
4080
4081 irdma_send_cm_event(cm_node, cm_id, IW_CM_EVENT_CONNECT_REPLY, 0);
4082 cm_node->accelerated = true;
4083 complete(&cm_node->establish_comp);
4084 cm_node->cm_core->cm_free_ah(cm_node);
4085 return;
4086
4087 error:
4088 iwqp->cm_id = NULL;
4089 cm_id->provider_data = NULL;
4090 irdma_send_cm_event(event->cm_node, cm_id, IW_CM_EVENT_CONNECT_REPLY,
4091 status);
4092 irdma_rem_ref_cm_node(event->cm_node);
4093 }
4094
4095 /**
4096 * irdma_cm_event_reset - handle reset
4097 * @event: the info for cm_node of connection
4098 */
4099 static void
irdma_cm_event_reset(struct irdma_cm_event * event)4100 irdma_cm_event_reset(struct irdma_cm_event *event)
4101 {
4102 struct irdma_cm_node *cm_node = event->cm_node;
4103 struct iw_cm_id *cm_id = cm_node->cm_id;
4104 struct irdma_qp *iwqp;
4105
4106 if (!cm_id)
4107 return;
4108
4109 iwqp = cm_id->provider_data;
4110 if (!iwqp)
4111 return;
4112
4113 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
4114 "reset event %p - cm_id = %p\n", event->cm_node, cm_id);
4115 iwqp->cm_id = NULL;
4116
4117 irdma_send_cm_event(cm_node, cm_node->cm_id, IW_CM_EVENT_DISCONNECT,
4118 -ECONNRESET);
4119 irdma_send_cm_event(cm_node, cm_node->cm_id, IW_CM_EVENT_CLOSE, 0);
4120 }
4121
4122 /**
4123 * irdma_cm_event_handler - send event to cm upper layer
4124 * @work: pointer of cm event info.
4125 */
4126 static void
irdma_cm_event_handler(struct work_struct * work)4127 irdma_cm_event_handler(struct work_struct *work)
4128 {
4129 struct irdma_cm_event *event = container_of(work, struct irdma_cm_event, event_work);
4130 struct irdma_cm_node *cm_node;
4131
4132 if (!event || !event->cm_node || !event->cm_node->cm_core)
4133 return;
4134
4135 cm_node = event->cm_node;
4136
4137 switch (event->type) {
4138 case IRDMA_CM_EVENT_MPA_REQ:
4139 irdma_send_cm_event(cm_node, cm_node->cm_id,
4140 IW_CM_EVENT_CONNECT_REQUEST, 0);
4141 break;
4142 case IRDMA_CM_EVENT_RESET:
4143 irdma_cm_event_reset(event);
4144 break;
4145 case IRDMA_CM_EVENT_CONNECTED:
4146 if (!event->cm_node->cm_id ||
4147 event->cm_node->state != IRDMA_CM_STATE_OFFLOADED)
4148 break;
4149 irdma_cm_event_connected(event);
4150 break;
4151 case IRDMA_CM_EVENT_MPA_REJECT:
4152 if (!event->cm_node->cm_id ||
4153 cm_node->state == IRDMA_CM_STATE_OFFLOADED)
4154 break;
4155 irdma_send_cm_event(cm_node, cm_node->cm_id,
4156 IW_CM_EVENT_CONNECT_REPLY, -ECONNREFUSED);
4157 break;
4158 case IRDMA_CM_EVENT_ABORTED:
4159 if (!event->cm_node->cm_id ||
4160 event->cm_node->state == IRDMA_CM_STATE_OFFLOADED)
4161 break;
4162 irdma_event_connect_error(event);
4163 break;
4164 default:
4165 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
4166 "bad event type = %d\n", event->type);
4167 break;
4168 }
4169
4170 irdma_rem_ref_cm_node(event->cm_node);
4171 kfree(event);
4172 }
4173
4174 /**
4175 * irdma_cm_post_event - queue event request for worker thread
4176 * @event: cm node's info for up event call
4177 */
4178 static void
irdma_cm_post_event(struct irdma_cm_event * event)4179 irdma_cm_post_event(struct irdma_cm_event *event)
4180 {
4181 atomic_inc(&event->cm_node->refcnt);
4182 INIT_WORK(&event->event_work, irdma_cm_event_handler);
4183 queue_work(event->cm_node->cm_core->event_wq, &event->event_work);
4184 }
4185
4186 /**
4187 * irdma_cm_teardown_connections - teardown QPs
4188 * @iwdev: device pointer
4189 * @ipaddr: Pointer to IPv4 or IPv6 address
4190 * @nfo: Connection info
4191 * @disconnect_all: flag indicating disconnect all QPs
4192 *
4193 * teardown QPs where source or destination addr matches ip addr
4194 */
4195 static void __unused
irdma_cm_teardown_connections(struct irdma_device * iwdev,u32 * ipaddr,struct irdma_cm_info * nfo,bool disconnect_all)4196 irdma_cm_teardown_connections(struct irdma_device *iwdev,
4197 u32 *ipaddr,
4198 struct irdma_cm_info *nfo,
4199 bool disconnect_all)
4200 {
4201 struct irdma_cm_core *cm_core = &iwdev->cm_core;
4202 struct list_head *list_core_temp;
4203 struct list_head *list_node;
4204 struct irdma_cm_node *cm_node;
4205 struct list_head teardown_list;
4206 struct ib_qp_attr attr;
4207 struct irdma_qp *qp;
4208
4209 INIT_LIST_HEAD(&teardown_list);
4210
4211 rcu_read_lock();
4212 irdma_iw_teardown_list_prep(cm_core, &teardown_list, ipaddr, nfo, disconnect_all);
4213 rcu_read_unlock();
4214
4215 attr.qp_state = IB_QPS_ERR;
4216 list_for_each_safe(list_node, list_core_temp, &teardown_list) {
4217 cm_node = container_of(list_node, struct irdma_cm_node,
4218 teardown_entry);
4219 irdma_modify_qp(&cm_node->iwqp->ibqp, &attr, IB_QP_STATE, NULL);
4220 if (iwdev->rf->reset)
4221 irdma_cm_disconn(cm_node->iwqp);
4222 irdma_rem_ref_cm_node(cm_node);
4223 }
4224
4225 if (!rdma_protocol_roce(&iwdev->ibdev, 1))
4226 return;
4227
4228 INIT_LIST_HEAD(&teardown_list);
4229 irdma_roce_teardown_list_prep(iwdev, &teardown_list, ipaddr, nfo, disconnect_all);
4230
4231 list_for_each_safe(list_node, list_core_temp, &teardown_list) {
4232 qp = container_of(list_node, struct irdma_qp, teardown_entry);
4233 irdma_modify_qp_roce(&qp->ibqp, &attr, IB_QP_STATE, NULL);
4234 irdma_ib_qp_event(qp, IRDMA_QP_EVENT_CATASTROPHIC);
4235 irdma_qp_rem_ref(&qp->ibqp);
4236 }
4237 }
4238