1 /*-
2 * SPDX-License-Identifier: GPL-2.0 or Linux-OpenIB
3 *
4 * Copyright (c) 2015 - 2023 Intel Corporation
5 *
6 * This software is available to you under a choice of one of two
7 * licenses. You may choose to be licensed under the terms of the GNU
8 * General Public License (GPL) Version 2, available from the file
9 * COPYING in the main directory of this source tree, or the
10 * OpenFabrics.org BSD license below:
11 *
12 * Redistribution and use in source and binary forms, with or
13 * without modification, are permitted provided that the following
14 * conditions are met:
15 *
16 * - Redistributions of source code must retain the above
17 * copyright notice, this list of conditions and the following
18 * disclaimer.
19 *
20 * - Redistributions in binary form must reproduce the above
21 * copyright notice, this list of conditions and the following
22 * disclaimer in the documentation and/or other materials
23 * provided with the distribution.
24 *
25 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
26 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
27 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
28 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
29 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
30 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
31 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
32 * SOFTWARE.
33 */
34
35 #include "irdma_main.h"
36
37 static void irdma_cm_post_event(struct irdma_cm_event *event);
38 static void irdma_disconnect_worker(struct work_struct *work);
39
40 /**
41 * irdma_free_sqbuf - put back puda buffer if refcount is 0
42 * @vsi: The VSI structure of the device
43 * @bufp: puda buffer to free
44 */
45 void
irdma_free_sqbuf(struct irdma_sc_vsi * vsi,void * bufp)46 irdma_free_sqbuf(struct irdma_sc_vsi *vsi, void *bufp)
47 {
48 struct irdma_puda_buf *buf = bufp;
49 struct irdma_puda_rsrc *ilq = vsi->ilq;
50
51 if (atomic_dec_and_test(&buf->refcount))
52 irdma_puda_ret_bufpool(ilq, buf);
53 }
54
55 /**
56 * irdma_record_ird_ord - Record IRD/ORD passed in
57 * @cm_node: connection's node
58 * @conn_ird: connection IRD
59 * @conn_ord: connection ORD
60 */
61 static void
irdma_record_ird_ord(struct irdma_cm_node * cm_node,u32 conn_ird,u32 conn_ord)62 irdma_record_ird_ord(struct irdma_cm_node *cm_node, u32 conn_ird,
63 u32 conn_ord)
64 {
65 if (conn_ird > cm_node->dev->hw_attrs.max_hw_ird)
66 conn_ird = cm_node->dev->hw_attrs.max_hw_ird;
67
68 if (conn_ord > cm_node->dev->hw_attrs.max_hw_ord)
69 conn_ord = cm_node->dev->hw_attrs.max_hw_ord;
70 else if (!conn_ord && cm_node->send_rdma0_op == SEND_RDMA_READ_ZERO)
71 conn_ord = 1;
72 cm_node->ird_size = conn_ird;
73 cm_node->ord_size = conn_ord;
74 }
75
76 /**
77 * irdma_copy_ip_ntohl - copy IP address from network to host
78 * @dst: IP address in host order
79 * @src: IP address in network order (big endian)
80 */
81 void
irdma_copy_ip_ntohl(u32 * dst,__be32 * src)82 irdma_copy_ip_ntohl(u32 *dst, __be32 *src)
83 {
84 *dst++ = ntohl(*src++);
85 *dst++ = ntohl(*src++);
86 *dst++ = ntohl(*src++);
87 *dst = ntohl(*src);
88 }
89
90 /**
91 * irdma_copy_ip_htonl - copy IP address from host to network order
92 * @dst: IP address in network order (big endian)
93 * @src: IP address in host order
94 */
95 void
irdma_copy_ip_htonl(__be32 * dst,u32 * src)96 irdma_copy_ip_htonl(__be32 *dst, u32 *src)
97 {
98 *dst++ = htonl(*src++);
99 *dst++ = htonl(*src++);
100 *dst++ = htonl(*src++);
101 *dst = htonl(*src);
102 }
103
104 /**
105 * irdma_get_addr_info
106 * @cm_node: contains ip/tcp info
107 * @cm_info: to get a copy of the cm_node ip/tcp info
108 */
109 static void
irdma_get_addr_info(struct irdma_cm_node * cm_node,struct irdma_cm_info * cm_info)110 irdma_get_addr_info(struct irdma_cm_node *cm_node,
111 struct irdma_cm_info *cm_info)
112 {
113 memset(cm_info, 0, sizeof(*cm_info));
114 cm_info->ipv4 = cm_node->ipv4;
115 cm_info->vlan_id = cm_node->vlan_id;
116 memcpy(cm_info->loc_addr, cm_node->loc_addr, sizeof(cm_info->loc_addr));
117 memcpy(cm_info->rem_addr, cm_node->rem_addr, sizeof(cm_info->rem_addr));
118 cm_info->loc_port = cm_node->loc_port;
119 cm_info->rem_port = cm_node->rem_port;
120 }
121
122 /**
123 * irdma_fill_sockaddr4 - fill in addr info for IPv4 connection
124 * @cm_node: connection's node
125 * @event: upper layer's cm event
126 */
127 static inline void
irdma_fill_sockaddr4(struct irdma_cm_node * cm_node,struct iw_cm_event * event)128 irdma_fill_sockaddr4(struct irdma_cm_node *cm_node,
129 struct iw_cm_event *event)
130 {
131 struct sockaddr_in *laddr = (struct sockaddr_in *)&event->local_addr;
132 struct sockaddr_in *raddr = (struct sockaddr_in *)&event->remote_addr;
133
134 laddr->sin_family = AF_INET;
135 raddr->sin_family = AF_INET;
136
137 laddr->sin_port = htons(cm_node->loc_port);
138 raddr->sin_port = htons(cm_node->rem_port);
139
140 laddr->sin_addr.s_addr = htonl(cm_node->loc_addr[0]);
141 raddr->sin_addr.s_addr = htonl(cm_node->rem_addr[0]);
142 }
143
144 /**
145 * irdma_fill_sockaddr6 - fill in addr info for IPv6 connection
146 * @cm_node: connection's node
147 * @event: upper layer's cm event
148 */
149 static inline void
irdma_fill_sockaddr6(struct irdma_cm_node * cm_node,struct iw_cm_event * event)150 irdma_fill_sockaddr6(struct irdma_cm_node *cm_node,
151 struct iw_cm_event *event)
152 {
153 struct sockaddr_in6 *laddr6 = (struct sockaddr_in6 *)&event->local_addr;
154 struct sockaddr_in6 *raddr6 = (struct sockaddr_in6 *)&event->remote_addr;
155
156 laddr6->sin6_family = AF_INET6;
157 raddr6->sin6_family = AF_INET6;
158
159 laddr6->sin6_port = htons(cm_node->loc_port);
160 raddr6->sin6_port = htons(cm_node->rem_port);
161
162 irdma_copy_ip_htonl(laddr6->sin6_addr.__u6_addr.__u6_addr32,
163 cm_node->loc_addr);
164 irdma_copy_ip_htonl(raddr6->sin6_addr.__u6_addr.__u6_addr32,
165 cm_node->rem_addr);
166 }
167
168 /**
169 * irdma_get_cmevent_info - for cm event upcall
170 * @cm_node: connection's node
171 * @cm_id: upper layers cm struct for the event
172 * @event: upper layer's cm event
173 */
174 static inline void
irdma_get_cmevent_info(struct irdma_cm_node * cm_node,struct iw_cm_id * cm_id,struct iw_cm_event * event)175 irdma_get_cmevent_info(struct irdma_cm_node *cm_node,
176 struct iw_cm_id *cm_id,
177 struct iw_cm_event *event)
178 {
179 memcpy(&event->local_addr, &cm_id->m_local_addr,
180 sizeof(event->local_addr));
181 memcpy(&event->remote_addr, &cm_id->m_remote_addr,
182 sizeof(event->remote_addr));
183 if (cm_node) {
184 event->private_data = cm_node->pdata_buf;
185 event->private_data_len = (u8)cm_node->pdata.size;
186 event->ird = cm_node->ird_size;
187 event->ord = cm_node->ord_size;
188 }
189 }
190
191 /**
192 * irdma_send_cm_event - upcall cm's event handler
193 * @cm_node: connection's node
194 * @cm_id: upper layer's cm info struct
195 * @type: Event type to indicate
196 * @status: status for the event type
197 */
198 static int
irdma_send_cm_event(struct irdma_cm_node * cm_node,struct iw_cm_id * cm_id,enum iw_cm_event_type type,int status)199 irdma_send_cm_event(struct irdma_cm_node *cm_node,
200 struct iw_cm_id *cm_id,
201 enum iw_cm_event_type type, int status)
202 {
203 struct iw_cm_event event = {0};
204
205 event.event = type;
206 event.status = status;
207
208 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
209 "cm_node %p cm_id=%p state=%d accel=%d event_type=%d status=%d\n",
210 cm_node, cm_id, cm_node->accelerated, cm_node->state, type,
211 status);
212
213 switch (type) {
214 case IW_CM_EVENT_CONNECT_REQUEST:
215 if (cm_node->ipv4)
216 irdma_fill_sockaddr4(cm_node, &event);
217 else
218 irdma_fill_sockaddr6(cm_node, &event);
219 event.provider_data = cm_node;
220 event.private_data = cm_node->pdata_buf;
221 event.private_data_len = (u8)cm_node->pdata.size;
222 event.ird = cm_node->ird_size;
223 break;
224 case IW_CM_EVENT_CONNECT_REPLY:
225 irdma_get_cmevent_info(cm_node, cm_id, &event);
226 break;
227 case IW_CM_EVENT_ESTABLISHED:
228 event.ird = cm_node->ird_size;
229 event.ord = cm_node->ord_size;
230 break;
231 case IW_CM_EVENT_DISCONNECT:
232 case IW_CM_EVENT_CLOSE:
233 /* Wait if we are in RTS but havent issued the iwcm event upcall */
234 if (!cm_node->accelerated)
235 wait_for_completion(&cm_node->establish_comp);
236 break;
237 default:
238 return -EINVAL;
239 }
240
241 return cm_id->event_handler(cm_id, &event);
242 }
243
244 /**
245 * irdma_timer_list_prep - add connection nodes to a list to perform timer tasks
246 * @cm_core: cm's core
247 * @timer_list: a timer list to which cm_node will be selected
248 */
249 static void
irdma_timer_list_prep(struct irdma_cm_core * cm_core,struct list_head * timer_list)250 irdma_timer_list_prep(struct irdma_cm_core *cm_core,
251 struct list_head *timer_list)
252 {
253 struct irdma_cm_node *cm_node;
254 int bkt;
255
256 HASH_FOR_EACH_RCU(cm_core->cm_hash_tbl, bkt, cm_node, list) {
257 if ((cm_node->close_entry || cm_node->send_entry) &&
258 atomic_inc_not_zero(&cm_node->refcnt))
259 list_add(&cm_node->timer_entry, timer_list);
260 }
261 }
262
263 /**
264 * irdma_create_event - create cm event
265 * @cm_node: connection's node
266 * @type: Event type to generate
267 */
268 static struct irdma_cm_event *
irdma_create_event(struct irdma_cm_node * cm_node,enum irdma_cm_event_type type)269 irdma_create_event(struct irdma_cm_node *cm_node,
270 enum irdma_cm_event_type type)
271 {
272 struct irdma_cm_event *event;
273
274 if (!cm_node->cm_id)
275 return NULL;
276
277 event = kzalloc(sizeof(*event), GFP_ATOMIC);
278
279 if (!event)
280 return NULL;
281
282 event->type = type;
283 event->cm_node = cm_node;
284 memcpy(event->cm_info.rem_addr, cm_node->rem_addr,
285 sizeof(event->cm_info.rem_addr));
286 memcpy(event->cm_info.loc_addr, cm_node->loc_addr,
287 sizeof(event->cm_info.loc_addr));
288 event->cm_info.rem_port = cm_node->rem_port;
289 event->cm_info.loc_port = cm_node->loc_port;
290 event->cm_info.cm_id = cm_node->cm_id;
291 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
292 "node=%p event=%p type=%u dst=%x src=%x\n", cm_node, event,
293 type, event->cm_info.loc_addr[0],
294 event->cm_info.rem_addr[0]);
295 irdma_cm_post_event(event);
296
297 return event;
298 }
299
300 /**
301 * irdma_free_retrans_entry - free send entry
302 * @cm_node: connection's node
303 */
304 static void
irdma_free_retrans_entry(struct irdma_cm_node * cm_node)305 irdma_free_retrans_entry(struct irdma_cm_node *cm_node)
306 {
307 struct irdma_device *iwdev = cm_node->iwdev;
308 struct irdma_timer_entry *send_entry;
309
310 send_entry = cm_node->send_entry;
311 if (!send_entry)
312 return;
313
314 cm_node->send_entry = NULL;
315 irdma_free_sqbuf(&iwdev->vsi, send_entry->sqbuf);
316 kfree(send_entry);
317 atomic_dec(&cm_node->refcnt);
318 }
319
320 /**
321 * irdma_cleanup_retrans_entry - free send entry with lock
322 * @cm_node: connection's node
323 */
324 static void
irdma_cleanup_retrans_entry(struct irdma_cm_node * cm_node)325 irdma_cleanup_retrans_entry(struct irdma_cm_node *cm_node)
326 {
327 unsigned long flags;
328
329 spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
330 irdma_free_retrans_entry(cm_node);
331 spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
332 }
333
334 /**
335 * irdma_form_ah_cm_frame - get a free packet and build frame with address handle
336 * @cm_node: connection's node ionfo to use in frame
337 * @options: pointer to options info
338 * @hdr: pointer mpa header
339 * @pdata: pointer to private data
340 * @flags: indicates FIN or ACK
341 */
342 static struct irdma_puda_buf *
irdma_form_ah_cm_frame(struct irdma_cm_node * cm_node,struct irdma_kmem_info * options,struct irdma_kmem_info * hdr,struct irdma_mpa_priv_info * pdata,u8 flags)343 irdma_form_ah_cm_frame(struct irdma_cm_node *cm_node,
344 struct irdma_kmem_info *options,
345 struct irdma_kmem_info *hdr,
346 struct irdma_mpa_priv_info *pdata,
347 u8 flags)
348 {
349 struct irdma_puda_buf *sqbuf;
350 struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
351 u8 *buf;
352 struct tcphdr *tcph;
353 u16 pktsize;
354 u32 opts_len = 0;
355 u32 pd_len = 0;
356 u32 hdr_len = 0;
357
358 if (!cm_node->ah || !cm_node->ah->ah_info.ah_valid) {
359 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM, "AH invalid\n");
360 return NULL;
361 }
362
363 sqbuf = irdma_puda_get_bufpool(vsi->ilq);
364 if (!sqbuf) {
365 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM, "SQ buf NULL\n");
366 return NULL;
367 }
368
369 sqbuf->ah_id = cm_node->ah->ah_info.ah_idx;
370 buf = sqbuf->mem.va;
371 if (options)
372 opts_len = (u32)options->size;
373
374 if (hdr)
375 hdr_len = hdr->size;
376
377 if (pdata)
378 pd_len = pdata->size;
379
380 pktsize = sizeof(*tcph) + opts_len + hdr_len + pd_len;
381
382 memset(buf, 0, pktsize);
383
384 sqbuf->totallen = pktsize;
385 sqbuf->tcphlen = sizeof(*tcph) + opts_len;
386 sqbuf->scratch = cm_node;
387
388 tcph = (struct tcphdr *)buf;
389 buf += sizeof(*tcph);
390
391 tcph->th_sport = htons(cm_node->loc_port);
392 tcph->th_dport = htons(cm_node->rem_port);
393 tcph->th_seq = htonl(cm_node->tcp_cntxt.loc_seq_num);
394
395 if (flags & SET_ACK) {
396 cm_node->tcp_cntxt.loc_ack_num = cm_node->tcp_cntxt.rcv_nxt;
397 tcph->th_ack = htonl(cm_node->tcp_cntxt.loc_ack_num);
398 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_ACK);
399 } else {
400 tcph->th_ack = 0;
401 }
402
403 if (flags & SET_SYN) {
404 cm_node->tcp_cntxt.loc_seq_num++;
405 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_SYN);
406 } else {
407 cm_node->tcp_cntxt.loc_seq_num += hdr_len + pd_len;
408 }
409
410 if (flags & SET_FIN) {
411 cm_node->tcp_cntxt.loc_seq_num++;
412 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_FIN);
413 }
414
415 if (flags & SET_RST)
416 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_RST);
417
418 tcph->th_off = (u16)((sizeof(*tcph) + opts_len + 3) >> 2);
419 sqbuf->tcphlen = tcph->th_off << 2;
420 tcph->th_win = htons(cm_node->tcp_cntxt.rcv_wnd);
421 tcph->th_urp = 0;
422
423 if (opts_len) {
424 memcpy(buf, options->addr, opts_len);
425 buf += opts_len;
426 }
427
428 if (hdr_len) {
429 memcpy(buf, hdr->addr, hdr_len);
430 buf += hdr_len;
431 }
432
433 if (pdata && pdata->addr)
434 memcpy(buf, pdata->addr, pdata->size);
435
436 atomic_set(&sqbuf->refcount, 1);
437
438 irdma_debug_buf(vsi->dev, IRDMA_DEBUG_ILQ, "TRANSMIT ILQ BUFFER",
439 sqbuf->mem.va, sqbuf->totallen);
440
441 return sqbuf;
442 }
443
444 /**
445 * irdma_form_uda_cm_frame - get a free packet and build frame full tcpip packet
446 * @cm_node: connection's node ionfo to use in frame
447 * @options: pointer to options info
448 * @hdr: pointer mpa header
449 * @pdata: pointer to private data
450 * @flags: indicates FIN or ACK
451 */
452 static struct irdma_puda_buf *
irdma_form_uda_cm_frame(struct irdma_cm_node * cm_node,struct irdma_kmem_info * options,struct irdma_kmem_info * hdr,struct irdma_mpa_priv_info * pdata,u8 flags)453 irdma_form_uda_cm_frame(struct irdma_cm_node *cm_node,
454 struct irdma_kmem_info *options,
455 struct irdma_kmem_info *hdr,
456 struct irdma_mpa_priv_info *pdata,
457 u8 flags)
458 {
459 struct irdma_puda_buf *sqbuf;
460 struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
461 u8 *buf;
462
463 struct tcphdr *tcph;
464 struct ip *iph;
465 struct ip6_hdr *ip6h;
466 struct ether_header *ethh;
467 u16 pktsize;
468 u16 eth_hlen = ETH_HLEN;
469 u32 opts_len = 0;
470 u32 pd_len = 0;
471 u32 hdr_len = 0;
472
473 u16 vtag;
474
475 sqbuf = irdma_puda_get_bufpool(vsi->ilq);
476 if (!sqbuf)
477 return NULL;
478
479 buf = sqbuf->mem.va;
480
481 if (options)
482 opts_len = (u32)options->size;
483
484 if (hdr)
485 hdr_len = hdr->size;
486
487 if (pdata)
488 pd_len = pdata->size;
489
490 if (cm_node->vlan_id < VLAN_N_VID)
491 eth_hlen += 4;
492
493 if (cm_node->ipv4)
494 pktsize = sizeof(*iph) + sizeof(*tcph);
495 else
496 pktsize = sizeof(*ip6h) + sizeof(*tcph);
497 pktsize += opts_len + hdr_len + pd_len;
498
499 memset(buf, 0, eth_hlen + pktsize);
500
501 sqbuf->totallen = pktsize + eth_hlen;
502 sqbuf->maclen = eth_hlen;
503 sqbuf->tcphlen = sizeof(*tcph) + opts_len;
504 sqbuf->scratch = cm_node;
505
506 ethh = (struct ether_header *)buf;
507 buf += eth_hlen;
508
509 if (cm_node->do_lpb)
510 sqbuf->do_lpb = true;
511
512 if (cm_node->ipv4) {
513 sqbuf->ipv4 = true;
514
515 iph = (struct ip *)buf;
516 buf += sizeof(*iph);
517 tcph = (struct tcphdr *)buf;
518 buf += sizeof(*tcph);
519
520 ether_addr_copy(ethh->ether_dhost, cm_node->rem_mac);
521 ether_addr_copy(ethh->ether_shost, cm_node->loc_mac);
522 if (cm_node->vlan_id < VLAN_N_VID) {
523 ((struct ether_vlan_header *)ethh)->evl_proto =
524 htons(ETH_P_8021Q);
525 vtag = (cm_node->user_pri << VLAN_PRIO_SHIFT) |
526 cm_node->vlan_id;
527 ((struct ether_vlan_header *)ethh)->evl_tag = htons(vtag);
528
529 ((struct ether_vlan_header *)ethh)->evl_encap_proto =
530 htons(ETH_P_IP);
531 } else {
532 ethh->ether_type = htons(ETH_P_IP);
533 }
534
535 iph->ip_v = IPVERSION;
536 iph->ip_hl = 5; /* 5 * 4Byte words, IP headr len */
537 iph->ip_tos = cm_node->tos;
538 iph->ip_len = htons(pktsize);
539 iph->ip_id = htons(++cm_node->tcp_cntxt.loc_id);
540
541 iph->ip_off = htons(0x4000);
542 iph->ip_ttl = 0x40;
543 iph->ip_p = IPPROTO_TCP;
544 iph->ip_src.s_addr = htonl(cm_node->loc_addr[0]);
545 iph->ip_dst.s_addr = htonl(cm_node->rem_addr[0]);
546 } else {
547 sqbuf->ipv4 = false;
548 ip6h = (struct ip6_hdr *)buf;
549 buf += sizeof(*ip6h);
550 tcph = (struct tcphdr *)buf;
551 buf += sizeof(*tcph);
552
553 ether_addr_copy(ethh->ether_dhost, cm_node->rem_mac);
554 ether_addr_copy(ethh->ether_shost, cm_node->loc_mac);
555 if (cm_node->vlan_id < VLAN_N_VID) {
556 ((struct ether_vlan_header *)ethh)->evl_proto =
557 htons(ETH_P_8021Q);
558 vtag = (cm_node->user_pri << VLAN_PRIO_SHIFT) |
559 cm_node->vlan_id;
560 ((struct ether_vlan_header *)ethh)->evl_tag = htons(vtag);
561 ((struct ether_vlan_header *)ethh)->evl_encap_proto =
562 htons(ETH_P_IPV6);
563 } else {
564 ethh->ether_type = htons(ETH_P_IPV6);
565 }
566 ip6h->ip6_vfc = 6 << 4;
567 ip6h->ip6_vfc |= cm_node->tos >> 4;
568 ip6h->ip6_flow = cm_node->tos << 20;
569 ip6h->ip6_plen = htons(pktsize - sizeof(*ip6h));
570 ip6h->ip6_nxt = 6;
571 ip6h->ip6_hops = 128;
572 irdma_copy_ip_htonl(ip6h->ip6_src.__u6_addr.__u6_addr32,
573 cm_node->loc_addr);
574 irdma_copy_ip_htonl(ip6h->ip6_dst.__u6_addr.__u6_addr32,
575 cm_node->rem_addr);
576 }
577
578 tcph->th_sport = htons(cm_node->loc_port);
579 tcph->th_dport = htons(cm_node->rem_port);
580 tcph->th_seq = htonl(cm_node->tcp_cntxt.loc_seq_num);
581
582 if (flags & SET_ACK) {
583 cm_node->tcp_cntxt.loc_ack_num = cm_node->tcp_cntxt.rcv_nxt;
584 tcph->th_ack = htonl(cm_node->tcp_cntxt.loc_ack_num);
585 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_ACK);
586 } else {
587 tcph->th_ack = 0;
588 }
589
590 if (flags & SET_SYN) {
591 cm_node->tcp_cntxt.loc_seq_num++;
592 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_SYN);
593 } else {
594 cm_node->tcp_cntxt.loc_seq_num += hdr_len + pd_len;
595 }
596
597 if (flags & SET_FIN) {
598 cm_node->tcp_cntxt.loc_seq_num++;
599 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_FIN);
600 }
601
602 if (flags & SET_RST)
603 tcp_set_flags(tcph, tcp_get_flags(tcph) | TH_RST);
604
605 tcph->th_off = (u16)((sizeof(*tcph) + opts_len + 3) >> 2);
606 sqbuf->tcphlen = tcph->th_off << 2;
607 tcph->th_win = htons(cm_node->tcp_cntxt.rcv_wnd);
608 tcph->th_urp = 0;
609
610 if (opts_len) {
611 memcpy(buf, options->addr, opts_len);
612 buf += opts_len;
613 }
614
615 if (hdr_len) {
616 memcpy(buf, hdr->addr, hdr_len);
617 buf += hdr_len;
618 }
619
620 if (pdata && pdata->addr)
621 memcpy(buf, pdata->addr, pdata->size);
622
623 atomic_set(&sqbuf->refcount, 1);
624
625 irdma_debug_buf(vsi->dev, IRDMA_DEBUG_ILQ, "TRANSMIT ILQ BUFFER",
626 sqbuf->mem.va, sqbuf->totallen);
627
628 return sqbuf;
629 }
630
631 /**
632 * irdma_send_reset - Send RST packet
633 * @cm_node: connection's node
634 */
635 int
irdma_send_reset(struct irdma_cm_node * cm_node)636 irdma_send_reset(struct irdma_cm_node *cm_node)
637 {
638 struct irdma_puda_buf *sqbuf;
639 int flags = SET_RST | SET_ACK;
640
641 sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL, NULL, NULL,
642 flags);
643 if (!sqbuf)
644 return -ENOMEM;
645
646 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
647 "caller: %pS cm_node %p cm_id=%p accel=%d state=%d rem_port=0x%04x, loc_port=0x%04x rem_addr=%x loc_addr=%x\n",
648 __builtin_return_address(0), cm_node, cm_node->cm_id,
649 cm_node->accelerated, cm_node->state, cm_node->rem_port,
650 cm_node->loc_port, cm_node->rem_addr[0],
651 cm_node->loc_addr[0]);
652
653 return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 0,
654 1);
655 }
656
657 /**
658 * irdma_active_open_err - send event for active side cm error
659 * @cm_node: connection's node
660 * @reset: Flag to send reset or not
661 */
662 static void
irdma_active_open_err(struct irdma_cm_node * cm_node,bool reset)663 irdma_active_open_err(struct irdma_cm_node *cm_node, bool reset)
664 {
665 irdma_cleanup_retrans_entry(cm_node);
666 cm_node->cm_core->stats_connect_errs++;
667 if (reset) {
668 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
669 "cm_node=%p state=%d\n", cm_node, cm_node->state);
670 atomic_inc(&cm_node->refcnt);
671 irdma_send_reset(cm_node);
672 }
673
674 cm_node->state = IRDMA_CM_STATE_CLOSED;
675 irdma_create_event(cm_node, IRDMA_CM_EVENT_ABORTED);
676 }
677
678 /**
679 * irdma_passive_open_err - handle passive side cm error
680 * @cm_node: connection's node
681 * @reset: send reset or just free cm_node
682 */
683 static void
irdma_passive_open_err(struct irdma_cm_node * cm_node,bool reset)684 irdma_passive_open_err(struct irdma_cm_node *cm_node, bool reset)
685 {
686 irdma_cleanup_retrans_entry(cm_node);
687 cm_node->cm_core->stats_passive_errs++;
688 cm_node->state = IRDMA_CM_STATE_CLOSED;
689 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
690 "cm_node=%p state=%d\n", cm_node, cm_node->state);
691 if (reset)
692 irdma_send_reset(cm_node);
693 else
694 irdma_rem_ref_cm_node(cm_node);
695 }
696
697 /**
698 * irdma_event_connect_error - to create connect error event
699 * @event: cm information for connect event
700 */
701 static void
irdma_event_connect_error(struct irdma_cm_event * event)702 irdma_event_connect_error(struct irdma_cm_event *event)
703 {
704 struct irdma_qp *iwqp;
705 struct iw_cm_id *cm_id;
706
707 cm_id = event->cm_node->cm_id;
708 if (!cm_id)
709 return;
710
711 iwqp = cm_id->provider_data;
712
713 if (!iwqp || !iwqp->iwdev)
714 return;
715
716 iwqp->cm_id = NULL;
717 cm_id->provider_data = NULL;
718 irdma_send_cm_event(event->cm_node, cm_id, IW_CM_EVENT_CONNECT_REPLY,
719 -ECONNRESET);
720 irdma_rem_ref_cm_node(event->cm_node);
721 }
722
723 /**
724 * irdma_process_options - process options from TCP header
725 * @cm_node: connection's node
726 * @optionsloc: point to start of options
727 * @optionsize: size of all options
728 * @syn_pkt: flag if syn packet
729 */
730 static int
irdma_process_options(struct irdma_cm_node * cm_node,u8 * optionsloc,u32 optionsize,u32 syn_pkt)731 irdma_process_options(struct irdma_cm_node *cm_node, u8 *optionsloc,
732 u32 optionsize, u32 syn_pkt)
733 {
734 u32 tmp;
735 u32 offset = 0;
736 union all_known_options *all_options;
737 char got_mss_option = 0;
738
739 while (offset < optionsize) {
740 all_options = (union all_known_options *)(optionsloc + offset);
741 switch (all_options->base.optionnum) {
742 case OPTION_NUM_EOL:
743 offset = optionsize;
744 break;
745 case OPTION_NUM_NONE:
746 offset += 1;
747 continue;
748 case OPTION_NUM_MSS:
749 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
750 "MSS Length: %d Offset: %d Size: %d\n",
751 all_options->mss.len, offset, optionsize);
752 got_mss_option = 1;
753 if (all_options->mss.len != 4)
754 return -EINVAL;
755 tmp = ntohs(all_options->mss.mss);
756 if ((cm_node->ipv4 &&
757 (tmp + IRDMA_MTU_TO_MSS_IPV4) < IRDMA_MIN_MTU_IPV4) ||
758 (!cm_node->ipv4 &&
759 (tmp + IRDMA_MTU_TO_MSS_IPV6) < IRDMA_MIN_MTU_IPV6))
760 return -EINVAL;
761 if (tmp < cm_node->tcp_cntxt.mss)
762 cm_node->tcp_cntxt.mss = tmp;
763 break;
764 case OPTION_NUM_WINDOW_SCALE:
765 cm_node->tcp_cntxt.snd_wscale =
766 all_options->windowscale.shiftcount;
767 break;
768 default:
769 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
770 "Unsupported TCP Option: %x\n",
771 all_options->base.optionnum);
772 break;
773 }
774 offset += all_options->base.len;
775 }
776 if (!got_mss_option && syn_pkt)
777 cm_node->tcp_cntxt.mss = IRDMA_CM_DEFAULT_MSS;
778
779 return 0;
780 }
781
782 /**
783 * irdma_handle_tcp_options - setup TCP context info after parsing TCP options
784 * @cm_node: connection's node
785 * @tcph: pointer tcp header
786 * @optionsize: size of options rcvd
787 * @passive: active or passive flag
788 */
789 static int
irdma_handle_tcp_options(struct irdma_cm_node * cm_node,struct tcphdr * tcph,int optionsize,int passive)790 irdma_handle_tcp_options(struct irdma_cm_node *cm_node,
791 struct tcphdr *tcph, int optionsize,
792 int passive)
793 {
794 u8 *optionsloc = (u8 *)&tcph[1];
795 int ret;
796
797 if (optionsize) {
798 ret = irdma_process_options(cm_node, optionsloc, optionsize,
799 (u32)tcp_get_flags(tcph) & TH_SYN);
800 if (ret) {
801 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
802 "Node %p, Sending Reset\n", cm_node);
803 if (passive)
804 irdma_passive_open_err(cm_node, true);
805 else
806 irdma_active_open_err(cm_node, true);
807 return ret;
808 }
809 }
810
811 cm_node->tcp_cntxt.snd_wnd = ntohs(tcph->th_win)
812 << cm_node->tcp_cntxt.snd_wscale;
813
814 if (cm_node->tcp_cntxt.snd_wnd > cm_node->tcp_cntxt.max_snd_wnd)
815 cm_node->tcp_cntxt.max_snd_wnd = cm_node->tcp_cntxt.snd_wnd;
816
817 return 0;
818 }
819
820 /**
821 * irdma_build_mpa_v1 - build a MPA V1 frame
822 * @cm_node: connection's node
823 * @start_addr: address where to build frame
824 * @mpa_key: to do read0 or write0
825 */
826 static void
irdma_build_mpa_v1(struct irdma_cm_node * cm_node,void * start_addr,u8 mpa_key)827 irdma_build_mpa_v1(struct irdma_cm_node *cm_node, void *start_addr,
828 u8 mpa_key)
829 {
830 struct ietf_mpa_v1 *mpa_frame = start_addr;
831
832 switch (mpa_key) {
833 case MPA_KEY_REQUEST:
834 memcpy(mpa_frame->key, IEFT_MPA_KEY_REQ, IETF_MPA_KEY_SIZE);
835 break;
836 case MPA_KEY_REPLY:
837 memcpy(mpa_frame->key, IEFT_MPA_KEY_REP, IETF_MPA_KEY_SIZE);
838 break;
839 default:
840 break;
841 }
842 mpa_frame->flags = IETF_MPA_FLAGS_CRC;
843 mpa_frame->rev = cm_node->mpa_frame_rev;
844 mpa_frame->priv_data_len = htons(cm_node->pdata.size);
845 }
846
847 /**
848 * irdma_build_mpa_v2 - build a MPA V2 frame
849 * @cm_node: connection's node
850 * @start_addr: buffer start address
851 * @mpa_key: to do read0 or write0
852 */
853 static void
irdma_build_mpa_v2(struct irdma_cm_node * cm_node,void * start_addr,u8 mpa_key)854 irdma_build_mpa_v2(struct irdma_cm_node *cm_node, void *start_addr,
855 u8 mpa_key)
856 {
857 struct ietf_mpa_v2 *mpa_frame = start_addr;
858 struct ietf_rtr_msg *rtr_msg = &mpa_frame->rtr_msg;
859 u16 ctrl_ird, ctrl_ord;
860
861 /* initialize the upper 5 bytes of the frame */
862 irdma_build_mpa_v1(cm_node, start_addr, mpa_key);
863 mpa_frame->flags |= IETF_MPA_V2_FLAG;
864 if (cm_node->iwdev->iw_ooo) {
865 mpa_frame->flags |= IETF_MPA_FLAGS_MARKERS;
866 cm_node->rcv_mark_en = true;
867 }
868 mpa_frame->priv_data_len = cpu_to_be16(be16_to_cpu(mpa_frame->priv_data_len) +
869 IETF_RTR_MSG_SIZE);
870
871 /* initialize RTR msg */
872 if (cm_node->mpav2_ird_ord == IETF_NO_IRD_ORD) {
873 ctrl_ird = IETF_NO_IRD_ORD;
874 ctrl_ord = IETF_NO_IRD_ORD;
875 } else {
876 ctrl_ird = (cm_node->ird_size > IETF_NO_IRD_ORD) ?
877 IETF_NO_IRD_ORD :
878 cm_node->ird_size;
879 ctrl_ord = (cm_node->ord_size > IETF_NO_IRD_ORD) ?
880 IETF_NO_IRD_ORD :
881 cm_node->ord_size;
882 }
883 ctrl_ird |= IETF_PEER_TO_PEER;
884
885 switch (mpa_key) {
886 case MPA_KEY_REQUEST:
887 ctrl_ord |= IETF_RDMA0_WRITE;
888 ctrl_ord |= IETF_RDMA0_READ;
889 break;
890 case MPA_KEY_REPLY:
891 switch (cm_node->send_rdma0_op) {
892 case SEND_RDMA_WRITE_ZERO:
893 ctrl_ord |= IETF_RDMA0_WRITE;
894 break;
895 case SEND_RDMA_READ_ZERO:
896 ctrl_ord |= IETF_RDMA0_READ;
897 break;
898 }
899 break;
900 default:
901 break;
902 }
903 rtr_msg->ctrl_ird = htons(ctrl_ird);
904 rtr_msg->ctrl_ord = htons(ctrl_ord);
905 }
906
907 /**
908 * irdma_cm_build_mpa_frame - build mpa frame for mpa version 1 or version 2
909 * @cm_node: connection's node
910 * @mpa: mpa: data buffer
911 * @mpa_key: to do read0 or write0
912 */
913 static int
irdma_cm_build_mpa_frame(struct irdma_cm_node * cm_node,struct irdma_kmem_info * mpa,u8 mpa_key)914 irdma_cm_build_mpa_frame(struct irdma_cm_node *cm_node,
915 struct irdma_kmem_info *mpa, u8 mpa_key)
916 {
917 int hdr_len = 0;
918
919 switch (cm_node->mpa_frame_rev) {
920 case IETF_MPA_V1:
921 hdr_len = sizeof(struct ietf_mpa_v1);
922 irdma_build_mpa_v1(cm_node, mpa->addr, mpa_key);
923 break;
924 case IETF_MPA_V2:
925 hdr_len = sizeof(struct ietf_mpa_v2);
926 irdma_build_mpa_v2(cm_node, mpa->addr, mpa_key);
927 break;
928 default:
929 break;
930 }
931
932 return hdr_len;
933 }
934
935 /**
936 * irdma_send_mpa_request - active node send mpa request to passive node
937 * @cm_node: connection's node
938 */
939 static int
irdma_send_mpa_request(struct irdma_cm_node * cm_node)940 irdma_send_mpa_request(struct irdma_cm_node *cm_node)
941 {
942 struct irdma_puda_buf *sqbuf;
943
944 cm_node->mpa_hdr.addr = &cm_node->mpa_v2_frame;
945 cm_node->mpa_hdr.size = irdma_cm_build_mpa_frame(cm_node,
946 &cm_node->mpa_hdr,
947 MPA_KEY_REQUEST);
948 if (!cm_node->mpa_hdr.size) {
949 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
950 "mpa size = %d\n", cm_node->mpa_hdr.size);
951 return -EINVAL;
952 }
953
954 sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL,
955 &cm_node->mpa_hdr,
956 &cm_node->pdata, SET_ACK);
957 if (!sqbuf)
958 return -ENOMEM;
959
960 return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
961 0);
962 }
963
964 /**
965 * irdma_send_mpa_reject -
966 * @cm_node: connection's node
967 * @pdata: reject data for connection
968 * @plen: length of reject data
969 */
970 static int
irdma_send_mpa_reject(struct irdma_cm_node * cm_node,const void * pdata,u8 plen)971 irdma_send_mpa_reject(struct irdma_cm_node *cm_node,
972 const void *pdata, u8 plen)
973 {
974 struct irdma_puda_buf *sqbuf;
975 struct irdma_mpa_priv_info priv_info;
976
977 cm_node->mpa_hdr.addr = &cm_node->mpa_v2_frame;
978 cm_node->mpa_hdr.size = irdma_cm_build_mpa_frame(cm_node,
979 &cm_node->mpa_hdr,
980 MPA_KEY_REPLY);
981
982 cm_node->mpa_v2_frame.flags |= IETF_MPA_FLAGS_REJECT;
983 priv_info.addr = pdata;
984 priv_info.size = plen;
985
986 sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL,
987 &cm_node->mpa_hdr, &priv_info,
988 SET_ACK | SET_FIN);
989 if (!sqbuf)
990 return -ENOMEM;
991
992 cm_node->state = IRDMA_CM_STATE_FIN_WAIT1;
993
994 return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
995 0);
996 }
997
998 /**
999 * irdma_negotiate_mpa_v2_ird_ord - negotiate MPAv2 IRD/ORD
1000 * @cm_node: connection's node
1001 * @buf: Data pointer
1002 */
1003 static int
irdma_negotiate_mpa_v2_ird_ord(struct irdma_cm_node * cm_node,u8 * buf)1004 irdma_negotiate_mpa_v2_ird_ord(struct irdma_cm_node *cm_node,
1005 u8 *buf)
1006 {
1007 struct ietf_mpa_v2 *mpa_v2_frame;
1008 struct ietf_rtr_msg *rtr_msg;
1009 u16 ird_size;
1010 u16 ord_size;
1011 u16 ctrl_ord;
1012 u16 ctrl_ird;
1013
1014 mpa_v2_frame = (struct ietf_mpa_v2 *)buf;
1015 rtr_msg = &mpa_v2_frame->rtr_msg;
1016
1017 /* parse rtr message */
1018 ctrl_ord = ntohs(rtr_msg->ctrl_ord);
1019 ctrl_ird = ntohs(rtr_msg->ctrl_ird);
1020 ird_size = ctrl_ird & IETF_NO_IRD_ORD;
1021 ord_size = ctrl_ord & IETF_NO_IRD_ORD;
1022
1023 if (!(ctrl_ird & IETF_PEER_TO_PEER))
1024 return -EOPNOTSUPP;
1025
1026 if (ird_size == IETF_NO_IRD_ORD || ord_size == IETF_NO_IRD_ORD) {
1027 cm_node->mpav2_ird_ord = IETF_NO_IRD_ORD;
1028 goto negotiate_done;
1029 }
1030
1031 if (cm_node->state != IRDMA_CM_STATE_MPAREQ_SENT) {
1032 /* responder */
1033 if (!ord_size && (ctrl_ord & IETF_RDMA0_READ))
1034 cm_node->ird_size = 1;
1035 if (cm_node->ord_size > ird_size)
1036 cm_node->ord_size = ird_size;
1037 } else {
1038 /* initiator */
1039 if (!ird_size && (ctrl_ord & IETF_RDMA0_READ))
1040 /* Remote peer doesn't support RDMA0_READ */
1041 return -EOPNOTSUPP;
1042
1043 if (cm_node->ord_size > ird_size)
1044 cm_node->ord_size = ird_size;
1045
1046 if (cm_node->ird_size < ord_size)
1047 /* no resources available */
1048 return -EINVAL;
1049 }
1050
1051 negotiate_done:
1052 if (ctrl_ord & IETF_RDMA0_READ)
1053 cm_node->send_rdma0_op = SEND_RDMA_READ_ZERO;
1054 else if (ctrl_ord & IETF_RDMA0_WRITE)
1055 cm_node->send_rdma0_op = SEND_RDMA_WRITE_ZERO;
1056 else
1057 /* Not supported RDMA0 operation */
1058 return -EOPNOTSUPP;
1059
1060 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1061 "MPAV2 Negotiated ORD: %d, IRD: %d\n", cm_node->ord_size,
1062 cm_node->ird_size);
1063 return 0;
1064 }
1065
1066 /**
1067 * irdma_parse_mpa - process an IETF MPA frame
1068 * @cm_node: connection's node
1069 * @buf: Data pointer
1070 * @type: to return accept or reject
1071 * @len: Len of mpa buffer
1072 */
1073 static int
irdma_parse_mpa(struct irdma_cm_node * cm_node,u8 * buf,u32 * type,u32 len)1074 irdma_parse_mpa(struct irdma_cm_node *cm_node, u8 *buf, u32 *type,
1075 u32 len)
1076 {
1077 struct ietf_mpa_v1 *mpa_frame;
1078 int mpa_hdr_len, priv_data_len, ret;
1079
1080 *type = IRDMA_MPA_REQUEST_ACCEPT;
1081
1082 if (len < sizeof(*mpa_frame)) {
1083 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1084 "ietf buffer small (%x)\n", len);
1085 return -EINVAL;
1086 }
1087
1088 mpa_frame = (struct ietf_mpa_v1 *)buf;
1089 mpa_hdr_len = sizeof(*mpa_frame);
1090 priv_data_len = ntohs(mpa_frame->priv_data_len);
1091
1092 if (priv_data_len > IETF_MAX_PRIV_DATA_LEN) {
1093 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1094 "private_data too big %d\n", priv_data_len);
1095 return -EOVERFLOW;
1096 }
1097
1098 if (mpa_frame->rev != IETF_MPA_V1 && mpa_frame->rev != IETF_MPA_V2) {
1099 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1100 "unsupported mpa rev = %d\n", mpa_frame->rev);
1101 return -EINVAL;
1102 }
1103
1104 if (mpa_frame->rev > cm_node->mpa_frame_rev) {
1105 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1106 "rev %d\n", mpa_frame->rev);
1107 return -EINVAL;
1108 }
1109
1110 cm_node->mpa_frame_rev = mpa_frame->rev;
1111 if (cm_node->state != IRDMA_CM_STATE_MPAREQ_SENT) {
1112 if (memcmp(mpa_frame->key, IEFT_MPA_KEY_REQ,
1113 IETF_MPA_KEY_SIZE)) {
1114 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1115 "Unexpected MPA Key received\n");
1116 return -EINVAL;
1117 }
1118 } else {
1119 if (memcmp(mpa_frame->key, IEFT_MPA_KEY_REP,
1120 IETF_MPA_KEY_SIZE)) {
1121 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1122 "Unexpected MPA Key received\n");
1123 return -EINVAL;
1124 }
1125 }
1126
1127 if (priv_data_len + mpa_hdr_len > len) {
1128 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1129 "ietf buffer len(%x + %x != %x)\n", priv_data_len,
1130 mpa_hdr_len, len);
1131 return -EOVERFLOW;
1132 }
1133
1134 if (len > IRDMA_MAX_CM_BUF) {
1135 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1136 "ietf buffer large len = %d\n", len);
1137 return -EOVERFLOW;
1138 }
1139
1140 switch (mpa_frame->rev) {
1141 case IETF_MPA_V2:
1142 mpa_hdr_len += IETF_RTR_MSG_SIZE;
1143 ret = irdma_negotiate_mpa_v2_ird_ord(cm_node, buf);
1144 if (ret)
1145 return ret;
1146 break;
1147 case IETF_MPA_V1:
1148 default:
1149 break;
1150 }
1151
1152 memcpy(cm_node->pdata_buf, buf + mpa_hdr_len, priv_data_len);
1153 cm_node->pdata.size = priv_data_len;
1154
1155 if (mpa_frame->flags & IETF_MPA_FLAGS_REJECT)
1156 *type = IRDMA_MPA_REQUEST_REJECT;
1157
1158 if (mpa_frame->flags & IETF_MPA_FLAGS_MARKERS)
1159 cm_node->snd_mark_en = true;
1160
1161 return 0;
1162 }
1163
1164 /**
1165 * irdma_schedule_cm_timer
1166 * @cm_node: connection's node
1167 * @sqbuf: buffer to send
1168 * @type: if it is send or close
1169 * @send_retrans: if rexmits to be done
1170 * @close_when_complete: is cm_node to be removed
1171 *
1172 * note - cm_node needs to be protected before calling this. Encase in:
1173 * irdma_rem_ref_cm_node(cm_core, cm_node);
1174 * irdma_schedule_cm_timer(...)
1175 * atomic_inc(&cm_node->refcnt);
1176 */
1177 int
irdma_schedule_cm_timer(struct irdma_cm_node * cm_node,struct irdma_puda_buf * sqbuf,enum irdma_timer_type type,int send_retrans,int close_when_complete)1178 irdma_schedule_cm_timer(struct irdma_cm_node *cm_node,
1179 struct irdma_puda_buf *sqbuf,
1180 enum irdma_timer_type type, int send_retrans,
1181 int close_when_complete)
1182 {
1183 struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
1184 struct irdma_cm_core *cm_core = cm_node->cm_core;
1185 struct irdma_timer_entry *new_send;
1186 u32 was_timer_set;
1187 unsigned long flags;
1188
1189 new_send = kzalloc(sizeof(*new_send), GFP_ATOMIC);
1190 if (!new_send) {
1191 if (type != IRDMA_TIMER_TYPE_CLOSE)
1192 irdma_free_sqbuf(vsi, sqbuf);
1193 return -ENOMEM;
1194 }
1195
1196 new_send->retrycount = IRDMA_DEFAULT_RETRYS;
1197 new_send->retranscount = IRDMA_DEFAULT_RETRANS;
1198 new_send->sqbuf = sqbuf;
1199 new_send->timetosend = jiffies;
1200 new_send->type = type;
1201 new_send->send_retrans = send_retrans;
1202 new_send->close_when_complete = close_when_complete;
1203
1204 if (type == IRDMA_TIMER_TYPE_CLOSE) {
1205 new_send->timetosend += (HZ / 10);
1206 if (cm_node->close_entry) {
1207 kfree(new_send);
1208 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1209 "already close entry\n");
1210 return -EINVAL;
1211 }
1212
1213 cm_node->close_entry = new_send;
1214 } else { /* type == IRDMA_TIMER_TYPE_SEND */
1215 spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
1216 cm_node->send_entry = new_send;
1217 atomic_inc(&cm_node->refcnt);
1218 spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
1219 new_send->timetosend = jiffies + IRDMA_RETRY_TIMEOUT;
1220
1221 atomic_inc(&sqbuf->refcount);
1222 irdma_puda_send_buf(vsi->ilq, sqbuf);
1223 if (!send_retrans) {
1224 irdma_cleanup_retrans_entry(cm_node);
1225 if (close_when_complete)
1226 irdma_rem_ref_cm_node(cm_node);
1227 return 0;
1228 }
1229 }
1230
1231 spin_lock_irqsave(&cm_core->ht_lock, flags);
1232 was_timer_set = timer_pending(&cm_core->tcp_timer);
1233
1234 if (!was_timer_set) {
1235 cm_core->tcp_timer.expires = new_send->timetosend;
1236 add_timer(&cm_core->tcp_timer);
1237 }
1238 spin_unlock_irqrestore(&cm_core->ht_lock, flags);
1239
1240 return 0;
1241 }
1242
1243 /**
1244 * irdma_retrans_expired - Could not rexmit the packet
1245 * @cm_node: connection's node
1246 */
1247 static void
irdma_retrans_expired(struct irdma_cm_node * cm_node)1248 irdma_retrans_expired(struct irdma_cm_node *cm_node)
1249 {
1250 enum irdma_cm_node_state state = cm_node->state;
1251
1252 cm_node->state = IRDMA_CM_STATE_CLOSED;
1253 switch (state) {
1254 case IRDMA_CM_STATE_SYN_RCVD:
1255 case IRDMA_CM_STATE_CLOSING:
1256 irdma_rem_ref_cm_node(cm_node);
1257 break;
1258 case IRDMA_CM_STATE_FIN_WAIT1:
1259 case IRDMA_CM_STATE_LAST_ACK:
1260 irdma_send_reset(cm_node);
1261 break;
1262 default:
1263 atomic_inc(&cm_node->refcnt);
1264 irdma_send_reset(cm_node);
1265 irdma_create_event(cm_node, IRDMA_CM_EVENT_ABORTED);
1266 break;
1267 }
1268 }
1269
1270 /**
1271 * irdma_handle_close_entry - for handling retry/timeouts
1272 * @cm_node: connection's node
1273 * @rem_node: flag for remove cm_node
1274 */
1275 static void
irdma_handle_close_entry(struct irdma_cm_node * cm_node,u32 rem_node)1276 irdma_handle_close_entry(struct irdma_cm_node *cm_node,
1277 u32 rem_node)
1278 {
1279 struct irdma_timer_entry *close_entry = cm_node->close_entry;
1280 struct irdma_qp *iwqp;
1281 unsigned long flags;
1282
1283 if (!close_entry)
1284 return;
1285 iwqp = (struct irdma_qp *)close_entry->sqbuf;
1286 if (iwqp) {
1287 spin_lock_irqsave(&iwqp->lock, flags);
1288 if (iwqp->cm_id) {
1289 iwqp->hw_tcp_state = IRDMA_TCP_STATE_CLOSED;
1290 iwqp->hw_iwarp_state = IRDMA_QP_STATE_ERROR;
1291 iwqp->last_aeq = IRDMA_AE_RESET_SENT;
1292 iwqp->ibqp_state = IB_QPS_ERR;
1293 spin_unlock_irqrestore(&iwqp->lock, flags);
1294 irdma_cm_disconn(iwqp);
1295 } else {
1296 spin_unlock_irqrestore(&iwqp->lock, flags);
1297 }
1298 } else if (rem_node) {
1299 /* TIME_WAIT state */
1300 irdma_rem_ref_cm_node(cm_node);
1301 }
1302
1303 kfree(close_entry);
1304 cm_node->close_entry = NULL;
1305 }
1306
1307 /**
1308 * irdma_cm_timer_tick - system's timer expired callback
1309 * @t: Pointer to timer_list
1310 */
1311 static void
irdma_cm_timer_tick(struct timer_list * t)1312 irdma_cm_timer_tick(struct timer_list *t)
1313 {
1314 unsigned long nexttimeout = jiffies + IRDMA_LONG_TIME;
1315 struct irdma_cm_node *cm_node;
1316 struct irdma_timer_entry *send_entry, *close_entry;
1317 struct list_head *list_core_temp;
1318 struct list_head *list_node;
1319 struct irdma_cm_core *cm_core = from_timer(cm_core, t, tcp_timer);
1320 struct irdma_sc_vsi *vsi;
1321 u32 settimer = 0;
1322 unsigned long timetosend;
1323 unsigned long flags;
1324 struct list_head timer_list;
1325
1326 INIT_LIST_HEAD(&timer_list);
1327
1328 rcu_read_lock();
1329 irdma_timer_list_prep(cm_core, &timer_list);
1330 rcu_read_unlock();
1331
1332 list_for_each_safe(list_node, list_core_temp, &timer_list) {
1333 cm_node = container_of(list_node, struct irdma_cm_node,
1334 timer_entry);
1335 close_entry = cm_node->close_entry;
1336
1337 if (close_entry) {
1338 if (time_after(close_entry->timetosend, jiffies)) {
1339 if (nexttimeout > close_entry->timetosend ||
1340 !settimer) {
1341 nexttimeout = close_entry->timetosend;
1342 settimer = 1;
1343 }
1344 } else {
1345 irdma_handle_close_entry(cm_node, 1);
1346 }
1347 }
1348
1349 spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
1350
1351 send_entry = cm_node->send_entry;
1352 if (!send_entry)
1353 goto done;
1354 if (time_after(send_entry->timetosend, jiffies)) {
1355 if (cm_node->state != IRDMA_CM_STATE_OFFLOADED) {
1356 if (nexttimeout > send_entry->timetosend ||
1357 !settimer) {
1358 nexttimeout = send_entry->timetosend;
1359 settimer = 1;
1360 }
1361 } else {
1362 irdma_free_retrans_entry(cm_node);
1363 }
1364 goto done;
1365 }
1366
1367 if (cm_node->state == IRDMA_CM_STATE_OFFLOADED ||
1368 cm_node->state == IRDMA_CM_STATE_CLOSED) {
1369 irdma_free_retrans_entry(cm_node);
1370 goto done;
1371 }
1372
1373 if (!send_entry->retranscount || !send_entry->retrycount) {
1374 irdma_free_retrans_entry(cm_node);
1375
1376 spin_unlock_irqrestore(&cm_node->retrans_list_lock,
1377 flags);
1378 irdma_retrans_expired(cm_node);
1379 cm_node->state = IRDMA_CM_STATE_CLOSED;
1380 spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
1381 goto done;
1382 }
1383 spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
1384
1385 vsi = &cm_node->iwdev->vsi;
1386 if (!cm_node->ack_rcvd) {
1387 atomic_inc(&send_entry->sqbuf->refcount);
1388 irdma_puda_send_buf(vsi->ilq, send_entry->sqbuf);
1389 cm_node->cm_core->stats_pkt_retrans++;
1390 }
1391
1392 spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
1393 if (send_entry->send_retrans) {
1394 send_entry->retranscount--;
1395 timetosend = (IRDMA_RETRY_TIMEOUT <<
1396 (IRDMA_DEFAULT_RETRANS -
1397 send_entry->retranscount));
1398
1399 send_entry->timetosend = jiffies +
1400 min(timetosend, IRDMA_MAX_TIMEOUT);
1401 if (nexttimeout > send_entry->timetosend || !settimer) {
1402 nexttimeout = send_entry->timetosend;
1403 settimer = 1;
1404 }
1405 } else {
1406 int close_when_complete;
1407
1408 close_when_complete = send_entry->close_when_complete;
1409 irdma_free_retrans_entry(cm_node);
1410 if (close_when_complete)
1411 irdma_rem_ref_cm_node(cm_node);
1412 }
1413 done:
1414 spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
1415 irdma_rem_ref_cm_node(cm_node);
1416 }
1417
1418 if (settimer) {
1419 spin_lock_irqsave(&cm_core->ht_lock, flags);
1420 if (!timer_pending(&cm_core->tcp_timer)) {
1421 cm_core->tcp_timer.expires = nexttimeout;
1422 add_timer(&cm_core->tcp_timer);
1423 }
1424 spin_unlock_irqrestore(&cm_core->ht_lock, flags);
1425 }
1426 }
1427
1428 /**
1429 * irdma_send_syn - send SYN packet
1430 * @cm_node: connection's node
1431 * @sendack: flag to set ACK bit or not
1432 */
1433 int
irdma_send_syn(struct irdma_cm_node * cm_node,u32 sendack)1434 irdma_send_syn(struct irdma_cm_node *cm_node, u32 sendack)
1435 {
1436 struct irdma_puda_buf *sqbuf;
1437 int flags = SET_SYN;
1438 char optionsbuf[sizeof(struct option_mss) +
1439 sizeof(struct option_windowscale) +
1440 sizeof(struct option_base) + TCP_OPTIONS_PADDING];
1441 struct irdma_kmem_info opts;
1442 int optionssize = 0;
1443
1444 /* Sending MSS option */
1445 union all_known_options *options;
1446
1447 opts.addr = optionsbuf;
1448 if (!cm_node)
1449 return -EINVAL;
1450
1451 options = (union all_known_options *)&optionsbuf[optionssize];
1452 options->mss.optionnum = OPTION_NUM_MSS;
1453 options->mss.len = sizeof(struct option_mss);
1454 options->mss.mss = htons(cm_node->tcp_cntxt.mss);
1455 optionssize += sizeof(struct option_mss);
1456
1457 options = (union all_known_options *)&optionsbuf[optionssize];
1458 options->windowscale.optionnum = OPTION_NUM_WINDOW_SCALE;
1459 options->windowscale.len = sizeof(struct option_windowscale);
1460 options->windowscale.shiftcount = cm_node->tcp_cntxt.rcv_wscale;
1461 optionssize += sizeof(struct option_windowscale);
1462 options = (union all_known_options *)&optionsbuf[optionssize];
1463 options->eol = OPTION_NUM_EOL;
1464 optionssize += 1;
1465
1466 if (sendack)
1467 flags |= SET_ACK;
1468
1469 opts.size = optionssize;
1470
1471 sqbuf = cm_node->cm_core->form_cm_frame(cm_node, &opts, NULL, NULL,
1472 flags);
1473 if (!sqbuf)
1474 return -ENOMEM;
1475
1476 return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
1477 0);
1478 }
1479
1480 /**
1481 * irdma_send_ack - Send ACK packet
1482 * @cm_node: connection's node
1483 */
1484 void
irdma_send_ack(struct irdma_cm_node * cm_node)1485 irdma_send_ack(struct irdma_cm_node *cm_node)
1486 {
1487 struct irdma_puda_buf *sqbuf;
1488 struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
1489
1490 sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL, NULL, NULL,
1491 SET_ACK);
1492 if (sqbuf)
1493 irdma_puda_send_buf(vsi->ilq, sqbuf);
1494 }
1495
1496 /**
1497 * irdma_send_fin - Send FIN pkt
1498 * @cm_node: connection's node
1499 */
1500 static int
irdma_send_fin(struct irdma_cm_node * cm_node)1501 irdma_send_fin(struct irdma_cm_node *cm_node)
1502 {
1503 struct irdma_puda_buf *sqbuf;
1504
1505 sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL, NULL, NULL,
1506 SET_ACK | SET_FIN);
1507 if (!sqbuf)
1508 return -ENOMEM;
1509
1510 return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
1511 0);
1512 }
1513
1514 /**
1515 * irdma_find_listener - find a cm node listening on this addr-port pair
1516 * @cm_core: cm's core
1517 * @dst_addr: listener ip addr
1518 * @ipv4: flag indicating IPv4 when true
1519 * @dst_port: listener tcp port num
1520 * @vlan_id: virtual LAN ID
1521 * @listener_state: state to match with listen node's
1522 */
1523 static struct irdma_cm_listener *
irdma_find_listener(struct irdma_cm_core * cm_core,u32 * dst_addr,bool ipv4,u16 dst_port,u16 vlan_id,enum irdma_cm_listener_state listener_state)1524 irdma_find_listener(struct irdma_cm_core *cm_core, u32 *dst_addr, bool ipv4, u16 dst_port,
1525 u16 vlan_id, enum irdma_cm_listener_state listener_state)
1526 {
1527 struct irdma_cm_listener *listen_node;
1528 static const u32 ip_zero[4] = {0, 0, 0, 0};
1529 u32 listen_addr[4];
1530 u16 listen_port;
1531 unsigned long flags;
1532
1533 /* walk list and find cm_node associated with this session ID */
1534 spin_lock_irqsave(&cm_core->listen_list_lock, flags);
1535 list_for_each_entry(listen_node, &cm_core->listen_list, list) {
1536 memcpy(listen_addr, listen_node->loc_addr, sizeof(listen_addr));
1537 listen_port = listen_node->loc_port;
1538 if (listen_node->ipv4 != ipv4 || listen_port != dst_port ||
1539 !(listener_state & listen_node->listener_state))
1540 continue;
1541 /* compare node pair, return node handle if a match */
1542 if (!memcmp(listen_addr, ip_zero, sizeof(listen_addr)) ||
1543 (!memcmp(listen_addr, dst_addr, sizeof(listen_addr)) &&
1544 vlan_id == listen_node->vlan_id)) {
1545 atomic_inc(&listen_node->refcnt);
1546 spin_unlock_irqrestore(&cm_core->listen_list_lock,
1547 flags);
1548 return listen_node;
1549 }
1550 }
1551 spin_unlock_irqrestore(&cm_core->listen_list_lock, flags);
1552
1553 return NULL;
1554 }
1555
1556 /**
1557 * irdma_del_multiple_qhash - Remove qhash and child listens
1558 * @iwdev: iWarp device
1559 * @cm_info: CM info for parent listen node
1560 * @cm_parent_listen_node: The parent listen node
1561 */
1562 static int
irdma_del_multiple_qhash(struct irdma_device * iwdev,struct irdma_cm_info * cm_info,struct irdma_cm_listener * cm_parent_listen_node)1563 irdma_del_multiple_qhash(struct irdma_device *iwdev,
1564 struct irdma_cm_info *cm_info,
1565 struct irdma_cm_listener *cm_parent_listen_node)
1566 {
1567 struct irdma_cm_listener *child_listen_node;
1568 struct list_head *pos, *tpos;
1569 unsigned long flags;
1570 int ret = -EINVAL;
1571
1572 spin_lock_irqsave(&iwdev->cm_core.listen_list_lock, flags);
1573 list_for_each_safe(pos, tpos,
1574 &cm_parent_listen_node->child_listen_list) {
1575 child_listen_node = list_entry(pos, struct irdma_cm_listener,
1576 child_listen_list);
1577 if (child_listen_node->ipv4)
1578 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1579 "removing child listen for IP=%x, port=%d, vlan=%d\n",
1580 child_listen_node->loc_addr[0],
1581 child_listen_node->loc_port,
1582 child_listen_node->vlan_id);
1583 else
1584 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1585 "removing child listen for IP=%x:%x:%x:%x, port=%d, vlan=%d\n",
1586 IRDMA_PRINT_IP6(child_listen_node->loc_addr),
1587 child_listen_node->loc_port,
1588 child_listen_node->vlan_id);
1589 list_del(pos);
1590 memcpy(cm_info->loc_addr, child_listen_node->loc_addr,
1591 sizeof(cm_info->loc_addr));
1592 cm_info->vlan_id = child_listen_node->vlan_id;
1593 if (child_listen_node->qhash_set) {
1594 ret = irdma_manage_qhash(iwdev, cm_info,
1595 IRDMA_QHASH_TYPE_TCP_SYN,
1596 IRDMA_QHASH_MANAGE_TYPE_DELETE,
1597 NULL, false);
1598 child_listen_node->qhash_set = false;
1599 } else {
1600 ret = 0;
1601 }
1602 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1603 "Child listen node freed = %p\n",
1604 child_listen_node);
1605 kfree(child_listen_node);
1606 cm_parent_listen_node->cm_core->stats_listen_nodes_destroyed++;
1607 }
1608 spin_unlock_irqrestore(&iwdev->cm_core.listen_list_lock, flags);
1609
1610 return ret;
1611 }
1612
irdma_iw_get_vlan_prio(u32 * loc_addr,u8 prio,bool ipv4)1613 static u8 irdma_iw_get_vlan_prio(u32 *loc_addr, u8 prio, bool ipv4)
1614 {
1615 return prio;
1616 }
1617
1618 /**
1619 * irdma_get_vlan_mac_ipv6 - Get the vlan and mac for an IPv6
1620 * address
1621 * @addr: local IPv6 address
1622 * @vlan_id: vlan id for the given IPv6 address
1623 * @mac: mac address for the given IPv6 address
1624 *
1625 * Returns the net_device of the IPv6 address and also sets the
1626 * vlan id and mac for that address.
1627 */
1628 void
irdma_get_vlan_mac_ipv6(struct iw_cm_id * cm_id,u32 * addr,u16 * vlan_id,u8 * mac)1629 irdma_get_vlan_mac_ipv6(struct iw_cm_id *cm_id, u32 *addr, u16 *vlan_id, u8 *mac)
1630 {
1631 if_t ip_dev = NULL;
1632 struct in6_addr laddr6;
1633 struct vnet *vnet = &init_net;
1634 struct ifaddr *ifa;
1635 u16 scope_id = 0;
1636
1637 irdma_copy_ip_htonl(laddr6.__u6_addr.__u6_addr32, addr);
1638 if (vlan_id)
1639 *vlan_id = 0xFFFF; /* Match rdma_vlan_dev_vlan_id() */
1640 if (mac)
1641 eth_zero_addr(mac);
1642
1643 if (IN6_IS_SCOPE_LINKLOCAL(&laddr6) ||
1644 IN6_IS_ADDR_MC_INTFACELOCAL(&laddr6))
1645 scope_id = ntohs(laddr6.__u6_addr.__u6_addr16[1]);
1646
1647 #ifdef VIMAGE
1648 vnet = irdma_cmid_to_vnet(cm_id);
1649 #endif
1650 ip_dev = ip6_ifp_find(vnet, laddr6, scope_id);
1651 if (ip_dev) {
1652 if (vlan_id)
1653 *vlan_id = rdma_vlan_dev_vlan_id(ip_dev);
1654 ifa = if_getifaddr(ip_dev);
1655 if (ifa && ifa->ifa_addr && mac)
1656 ether_addr_copy(mac, if_getlladdr(ip_dev));
1657 }
1658 }
1659
1660 /**
1661 * irdma_get_vlan_ipv4 - Returns the vlan_id for IPv4 address
1662 * @addr: local IPv4 address
1663 */
1664 u16
irdma_get_vlan_ipv4(struct iw_cm_id * cm_id,u32 * addr)1665 irdma_get_vlan_ipv4(struct iw_cm_id *cm_id, u32 *addr)
1666 {
1667 u16 vlan_id = 0xFFFF;
1668
1669 #ifdef INET
1670 if_t netdev;
1671 struct vnet *vnet = &init_net;
1672
1673 #ifdef VIMAGE
1674 vnet = irdma_cmid_to_vnet(cm_id);
1675 #endif
1676 netdev = ip_ifp_find(vnet, htonl(addr[0]));
1677 if (netdev) {
1678 vlan_id = rdma_vlan_dev_vlan_id(netdev);
1679 dev_put(netdev);
1680 }
1681 #endif
1682
1683 return vlan_id;
1684 }
1685
1686 static int
irdma_manage_qhash_wait(struct irdma_pci_f * rf,struct irdma_cm_info * cm_info)1687 irdma_manage_qhash_wait(struct irdma_pci_f *rf, struct irdma_cm_info *cm_info)
1688 {
1689 struct irdma_cqp_request *cqp_request = cm_info->cqp_request;
1690 int cnt = rf->sc_dev.hw_attrs.max_cqp_compl_wait_time_ms * CQP_TIMEOUT_THRESHOLD;
1691 u32 ret_val;
1692
1693 if (!cqp_request)
1694 return -ENOMEM;
1695 do {
1696 irdma_cqp_ce_handler(rf, &rf->ccq.sc_cq);
1697 mdelay(1);
1698 } while (!READ_ONCE(cqp_request->request_done) && --cnt);
1699
1700 ret_val = cqp_request->compl_info.op_ret_val;
1701 irdma_put_cqp_request(&rf->cqp, cqp_request);
1702 if (cnt) {
1703 if (!ret_val)
1704 return 0;
1705 return -EINVAL;
1706 }
1707
1708 return -ETIMEDOUT;
1709 }
1710
1711 /**
1712 * irdma_add_mqh_ifa_cb - Adds multiple qhashes for IPv4/IPv6
1713 * @arg: Calback argument structure from irdma_add_mqh
1714 * @ifa: Current address to compute against
1715 * @count: Current cumulative output of all callbacks in this iteration
1716 *
1717 * Adds a qhash and a child listen node for a single IPv4/IPv6 address
1718 * on the adapter and adds the associated qhash filter
1719 */
1720 static u_int
irdma_add_mqh_ifa_cb(void * arg,struct ifaddr * ifa,u_int count)1721 irdma_add_mqh_ifa_cb(void *arg, struct ifaddr *ifa, u_int count)
1722 {
1723 struct irdma_add_mqh_cbs *cbs = arg;
1724 struct irdma_cm_listener *child_listen_node;
1725 struct irdma_cm_info *cm_info = cbs->cm_info;
1726 struct irdma_device *iwdev = cbs->iwdev;
1727 struct irdma_cm_listener *cm_parent_listen_node = cbs->cm_listen_node;
1728 if_t ip_dev = ifa->ifa_ifp;
1729 unsigned long flags;
1730 int ret;
1731
1732 if (count)
1733 return 0;
1734
1735 child_listen_node = kzalloc(sizeof(*child_listen_node), GFP_ATOMIC);
1736 if (!child_listen_node) {
1737 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1738 "listener memory allocation\n");
1739 return -ENOMEM;
1740 }
1741
1742 memcpy(child_listen_node, cm_parent_listen_node,
1743 sizeof(*child_listen_node));
1744 cm_info->vlan_id = rdma_vlan_dev_vlan_id(ip_dev);
1745 child_listen_node->vlan_id = cm_info->vlan_id;
1746 if (cm_info->ipv4) {
1747 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1748 "Allocating child CM Listener forIP=%x, vlan_id=%d, MAC=%x:%x:%x:%x:%x:%x\n",
1749 ((struct sockaddr_in *)&ifa->ifa_addr)->sin_addr.s_addr,
1750 rdma_vlan_dev_vlan_id(ip_dev),
1751 if_getlladdr(ip_dev)[0], if_getlladdr(ip_dev)[1],
1752 if_getlladdr(ip_dev)[2], if_getlladdr(ip_dev)[3],
1753 if_getlladdr(ip_dev)[4], if_getlladdr(ip_dev)[5]);
1754 child_listen_node->loc_addr[0] =
1755 ntohl(((struct sockaddr_in *)ifa->ifa_addr)->sin_addr.s_addr);
1756 } else {
1757 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1758 "IP=%x:%x:%x:%x, vlan_id=%d, MAC=%x:%x:%x:%x:%x:%x\n",
1759 IRDMA_PRINT_IP6(&((struct sockaddr_in6 *)ifa->ifa_addr)->sin6_addr),
1760 rdma_vlan_dev_vlan_id(ip_dev),
1761 if_getlladdr(ip_dev)[0], if_getlladdr(ip_dev)[1],
1762 if_getlladdr(ip_dev)[2], if_getlladdr(ip_dev)[3],
1763 if_getlladdr(ip_dev)[4], if_getlladdr(ip_dev)[5]);
1764 irdma_copy_ip_ntohl(child_listen_node->loc_addr,
1765 ((struct sockaddr_in6 *)ifa->ifa_addr)->sin6_addr.__u6_addr.__u6_addr32);
1766 }
1767 memcpy(cm_info->loc_addr, child_listen_node->loc_addr,
1768 sizeof(cm_info->loc_addr));
1769 if (!iwdev->vsi.dscp_mode)
1770 cm_info->user_pri =
1771 irdma_iw_get_vlan_prio(child_listen_node->loc_addr,
1772 cm_info->user_pri,
1773 cm_info->ipv4);
1774 ret = irdma_manage_qhash(iwdev, cm_info,
1775 IRDMA_QHASH_TYPE_TCP_SYN,
1776 IRDMA_QHASH_MANAGE_TYPE_ADD,
1777 NULL, false);
1778 if (ret) {
1779 kfree(child_listen_node);
1780 return ret;
1781 }
1782 /* wait for qhash finish */
1783 ret = irdma_manage_qhash_wait(iwdev->rf, cm_info);
1784 if (ret) {
1785 kfree(child_listen_node);
1786 return ret;
1787 }
1788
1789 child_listen_node->qhash_set = true;
1790 spin_lock_irqsave(&iwdev->cm_core.listen_list_lock, flags);
1791 list_add(&child_listen_node->child_listen_list,
1792 &cm_parent_listen_node->child_listen_list);
1793 spin_unlock_irqrestore(&iwdev->cm_core.listen_list_lock, flags);
1794 cm_parent_listen_node->cm_core->stats_listen_nodes_created++;
1795
1796 return 0;
1797 }
1798
1799 /**
1800 * irdma_add_mqh - Adds multiple qhashes
1801 * @iwdev: iWarp device
1802 * @cm_info: CM info for parent listen node
1803 * @cm_listen_node: The parent listen node
1804 */
1805 static int
irdma_add_mqh(struct irdma_device * iwdev,struct irdma_cm_info * cm_info,struct irdma_cm_listener * cm_listen_node)1806 irdma_add_mqh(struct irdma_device *iwdev,
1807 struct irdma_cm_info *cm_info,
1808 struct irdma_cm_listener *cm_listen_node)
1809 {
1810 struct epoch_tracker et;
1811 struct irdma_add_mqh_cbs cbs;
1812 struct if_iter iter;
1813 if_t ifp;
1814 int err = -ENOENT;
1815
1816 cbs.iwdev = iwdev;
1817 cbs.cm_info = cm_info;
1818 cbs.cm_listen_node = cm_listen_node;
1819
1820 VNET_ITERATOR_DECL(vnet_iter);
1821
1822 VNET_LIST_RLOCK();
1823 NET_EPOCH_ENTER(et);
1824 VNET_FOREACH(vnet_iter) {
1825 CURVNET_SET_QUIET(vnet_iter);
1826 for (ifp = if_iter_start(&iter); ifp != NULL; ifp = if_iter_next(&iter)) {
1827 if (!(if_getflags(ifp) & IFF_UP))
1828 continue;
1829
1830 if (((rdma_vlan_dev_vlan_id(ifp) >= VLAN_N_VID) ||
1831 (rdma_vlan_dev_real_dev(ifp) != iwdev->netdev)) &&
1832 ifp != iwdev->netdev)
1833 continue;
1834
1835 if_addr_rlock(ifp);
1836 if (cm_info->ipv4)
1837 err = if_foreach_addr_type(ifp, AF_INET, irdma_add_mqh_ifa_cb, &cbs);
1838 else
1839 err = if_foreach_addr_type(ifp, AF_INET6, irdma_add_mqh_ifa_cb, &cbs);
1840 if_addr_runlock(ifp);
1841 }
1842 if_iter_finish(&iter);
1843 CURVNET_RESTORE();
1844 }
1845 NET_EPOCH_EXIT(et);
1846 VNET_LIST_RUNLOCK();
1847
1848 return err;
1849 }
1850
1851 /**
1852 * irdma_reset_list_prep - add connection nodes slated for reset to list
1853 * @cm_core: cm's core
1854 * @listener: pointer to listener node
1855 * @reset_list: a list to which cm_node will be selected
1856 */
1857 static void
irdma_reset_list_prep(struct irdma_cm_core * cm_core,struct irdma_cm_listener * listener,struct list_head * reset_list)1858 irdma_reset_list_prep(struct irdma_cm_core *cm_core,
1859 struct irdma_cm_listener *listener,
1860 struct list_head *reset_list)
1861 {
1862 struct irdma_cm_node *cm_node;
1863 int bkt;
1864
1865 HASH_FOR_EACH_RCU(cm_core->cm_hash_tbl, bkt, cm_node, list) {
1866 if (cm_node->listener == listener &&
1867 !cm_node->accelerated &&
1868 atomic_inc_not_zero(&cm_node->refcnt))
1869 list_add(&cm_node->reset_entry, reset_list);
1870 }
1871 }
1872
1873 /**
1874 * irdma_dec_refcnt_listen - delete listener and associated cm nodes
1875 * @cm_core: cm's core
1876 * @listener: pointer to listener node
1877 * @free_hanging_nodes: to free associated cm_nodes
1878 * @apbvt_del: flag to delete the apbvt
1879 */
1880 static int
irdma_dec_refcnt_listen(struct irdma_cm_core * cm_core,struct irdma_cm_listener * listener,int free_hanging_nodes,bool apbvt_del)1881 irdma_dec_refcnt_listen(struct irdma_cm_core *cm_core,
1882 struct irdma_cm_listener *listener,
1883 int free_hanging_nodes, bool apbvt_del)
1884 {
1885 struct list_head *list_pos;
1886 struct list_head *list_temp;
1887 struct irdma_cm_node *cm_node;
1888 struct list_head reset_list;
1889 struct irdma_cm_info nfo;
1890 enum irdma_cm_node_state old_state;
1891 unsigned long flags;
1892 int err;
1893
1894 /* free non-accelerated child nodes for this listener */
1895 INIT_LIST_HEAD(&reset_list);
1896 if (free_hanging_nodes) {
1897 rcu_read_lock();
1898 irdma_reset_list_prep(cm_core, listener, &reset_list);
1899 rcu_read_unlock();
1900 }
1901
1902 list_for_each_safe(list_pos, list_temp, &reset_list) {
1903 cm_node = container_of(list_pos, struct irdma_cm_node,
1904 reset_entry);
1905 if (cm_node->state >= IRDMA_CM_STATE_FIN_WAIT1) {
1906 irdma_rem_ref_cm_node(cm_node);
1907 continue;
1908 }
1909
1910 irdma_cleanup_retrans_entry(cm_node);
1911 err = irdma_send_reset(cm_node);
1912 if (err) {
1913 cm_node->state = IRDMA_CM_STATE_CLOSED;
1914 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1915 "send reset failed\n");
1916 } else {
1917 old_state = cm_node->state;
1918 cm_node->state = IRDMA_CM_STATE_LISTENER_DESTROYED;
1919 if (old_state != IRDMA_CM_STATE_MPAREQ_RCVD)
1920 irdma_rem_ref_cm_node(cm_node);
1921 }
1922 }
1923
1924 if (atomic_dec_and_test(&listener->refcnt)) {
1925 spin_lock_irqsave(&cm_core->listen_list_lock, flags);
1926 list_del(&listener->list);
1927 spin_unlock_irqrestore(&cm_core->listen_list_lock, flags);
1928
1929 if (apbvt_del)
1930 irdma_del_apbvt(listener->iwdev,
1931 listener->apbvt_entry);
1932 memcpy(nfo.loc_addr, listener->loc_addr, sizeof(nfo.loc_addr));
1933 nfo.loc_port = listener->loc_port;
1934 nfo.ipv4 = listener->ipv4;
1935 nfo.vlan_id = listener->vlan_id;
1936 nfo.user_pri = listener->user_pri;
1937 nfo.qh_qpid = listener->iwdev->vsi.ilq->qp_id;
1938
1939 if (!list_empty(&listener->child_listen_list)) {
1940 irdma_del_multiple_qhash(listener->iwdev, &nfo,
1941 listener);
1942 } else {
1943 if (listener->qhash_set)
1944 irdma_manage_qhash(listener->iwdev,
1945 &nfo,
1946 IRDMA_QHASH_TYPE_TCP_SYN,
1947 IRDMA_QHASH_MANAGE_TYPE_DELETE,
1948 NULL, false);
1949 }
1950
1951 cm_core->stats_listen_destroyed++;
1952 cm_core->stats_listen_nodes_destroyed++;
1953 irdma_debug(&listener->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
1954 "loc_port=0x%04x loc_addr=%x cm_listen_node=%p cm_id=%p qhash_set=%d vlan_id=%d apbvt_del=%d\n",
1955 listener->loc_port, listener->loc_addr[0], listener,
1956 listener->cm_id, listener->qhash_set,
1957 listener->vlan_id, apbvt_del);
1958 kfree(listener);
1959 listener = NULL;
1960 return 0;
1961 }
1962
1963 return -EINVAL;
1964 }
1965
1966 /**
1967 * irdma_cm_del_listen - delete a listener
1968 * @cm_core: cm's core
1969 * @listener: passive connection's listener
1970 * @apbvt_del: flag to delete apbvt
1971 */
1972 static int
irdma_cm_del_listen(struct irdma_cm_core * cm_core,struct irdma_cm_listener * listener,bool apbvt_del)1973 irdma_cm_del_listen(struct irdma_cm_core *cm_core,
1974 struct irdma_cm_listener *listener,
1975 bool apbvt_del)
1976 {
1977 listener->listener_state = IRDMA_CM_LISTENER_PASSIVE_STATE;
1978 listener->cm_id = NULL;
1979
1980 return irdma_dec_refcnt_listen(cm_core, listener, 1, apbvt_del);
1981 }
1982
1983 /**
1984 * irdma_find_node - find a cm node that matches the reference cm node
1985 * @cm_core: cm's core
1986 * @rem_port: remote tcp port num
1987 * @rem_addr: remote ip addr
1988 * @loc_port: local tcp port num
1989 * @loc_addr: local ip addr
1990 * @vlan_id: local VLAN ID
1991 */
1992 struct irdma_cm_node *
irdma_find_node(struct irdma_cm_core * cm_core,u16 rem_port,u32 * rem_addr,u16 loc_port,u32 * loc_addr,u16 vlan_id)1993 irdma_find_node(struct irdma_cm_core *cm_core,
1994 u16 rem_port, u32 *rem_addr, u16 loc_port,
1995 u32 *loc_addr, u16 vlan_id)
1996 {
1997 struct irdma_cm_node *cm_node;
1998 u32 key = (rem_port << 16) | loc_port;
1999
2000 rcu_read_lock();
2001 HASH_FOR_EACH_POSSIBLE_RCU(cm_core->cm_hash_tbl, cm_node, list, key) {
2002 if (cm_node->vlan_id == vlan_id &&
2003 cm_node->loc_port == loc_port && cm_node->rem_port == rem_port &&
2004 !memcmp(cm_node->loc_addr, loc_addr, sizeof(cm_node->loc_addr)) &&
2005 !memcmp(cm_node->rem_addr, rem_addr, sizeof(cm_node->rem_addr))) {
2006 if (!atomic_inc_not_zero(&cm_node->refcnt))
2007 goto exit;
2008 rcu_read_unlock();
2009 return cm_node;
2010 }
2011 }
2012
2013 exit:
2014 rcu_read_unlock();
2015
2016 /* no owner node */
2017 return NULL;
2018 }
2019
2020 /**
2021 * irdma_add_hte_node - add a cm node to the hash table
2022 * @cm_core: cm's core
2023 * @cm_node: connection's node
2024 */
2025 static void
irdma_add_hte_node(struct irdma_cm_core * cm_core,struct irdma_cm_node * cm_node)2026 irdma_add_hte_node(struct irdma_cm_core *cm_core,
2027 struct irdma_cm_node *cm_node)
2028 {
2029 unsigned long flags;
2030 u32 key = (cm_node->rem_port << 16) | cm_node->loc_port;
2031
2032 spin_lock_irqsave(&cm_core->ht_lock, flags);
2033 HASH_ADD_RCU(cm_core->cm_hash_tbl, &cm_node->list, key);
2034 spin_unlock_irqrestore(&cm_core->ht_lock, flags);
2035 }
2036
2037 /**
2038 * irdma_ipv4_is_lpb - check if loopback
2039 * @loc_addr: local addr to compare
2040 * @rem_addr: remote address
2041 */
2042 bool
irdma_ipv4_is_lpb(u32 loc_addr,u32 rem_addr)2043 irdma_ipv4_is_lpb(u32 loc_addr, u32 rem_addr)
2044 {
2045 return ipv4_is_loopback(htonl(rem_addr)) || (loc_addr == rem_addr);
2046 }
2047
2048 /**
2049 * irdma_ipv6_is_lpb - check if loopback
2050 * @loc_addr: local addr to compare
2051 * @rem_addr: remote address
2052 */
2053 bool
irdma_ipv6_is_lpb(u32 * loc_addr,u32 * rem_addr)2054 irdma_ipv6_is_lpb(u32 *loc_addr, u32 *rem_addr)
2055 {
2056 struct in6_addr raddr6;
2057
2058 irdma_copy_ip_htonl(raddr6.__u6_addr.__u6_addr32, rem_addr);
2059
2060 return !memcmp(loc_addr, rem_addr, 16) || ipv6_addr_loopback(&raddr6);
2061 }
2062
2063 /**
2064 * irdma_cm_create_ah - create a cm address handle
2065 * @cm_node: The connection manager node to create AH for
2066 * @wait: Provides option to wait for ah creation or not
2067 */
2068 static int
irdma_cm_create_ah(struct irdma_cm_node * cm_node,bool wait)2069 irdma_cm_create_ah(struct irdma_cm_node *cm_node, bool wait)
2070 {
2071 struct irdma_ah_info ah_info = {0};
2072 struct irdma_device *iwdev = cm_node->iwdev;
2073 #ifdef VIMAGE
2074 struct vnet *vnet = irdma_cmid_to_vnet(cm_node->cm_id);
2075 #endif
2076
2077 ether_addr_copy(ah_info.mac_addr, if_getlladdr(iwdev->netdev));
2078
2079 ah_info.hop_ttl = 0x40;
2080 ah_info.tc_tos = cm_node->tos;
2081 ah_info.vsi = &iwdev->vsi;
2082
2083 if (cm_node->ipv4) {
2084 ah_info.ipv4_valid = true;
2085 ah_info.dest_ip_addr[0] = cm_node->rem_addr[0];
2086 ah_info.src_ip_addr[0] = cm_node->loc_addr[0];
2087 CURVNET_SET_QUIET(vnet);
2088 ah_info.do_lpbk = irdma_ipv4_is_lpb(ah_info.src_ip_addr[0],
2089 ah_info.dest_ip_addr[0]);
2090 CURVNET_RESTORE();
2091 } else {
2092 memcpy(ah_info.dest_ip_addr, cm_node->rem_addr,
2093 sizeof(ah_info.dest_ip_addr));
2094 memcpy(ah_info.src_ip_addr, cm_node->loc_addr,
2095 sizeof(ah_info.src_ip_addr));
2096 ah_info.do_lpbk = irdma_ipv6_is_lpb(ah_info.src_ip_addr,
2097 ah_info.dest_ip_addr);
2098 }
2099
2100 ah_info.vlan_tag = cm_node->vlan_id;
2101 if (cm_node->vlan_id < VLAN_N_VID) {
2102 ah_info.insert_vlan_tag = 1;
2103 ah_info.vlan_tag |= cm_node->user_pri << VLAN_PRIO_SHIFT;
2104 }
2105
2106 ah_info.dst_arpindex =
2107 irdma_arp_table(iwdev->rf, ah_info.dest_ip_addr,
2108 NULL, IRDMA_ARP_RESOLVE);
2109
2110 if (irdma_puda_create_ah(&iwdev->rf->sc_dev, &ah_info, wait,
2111 IRDMA_PUDA_RSRC_TYPE_ILQ, cm_node,
2112 &cm_node->ah))
2113 return -ENOMEM;
2114
2115 return 0;
2116 }
2117
2118 /**
2119 * irdma_cm_free_ah - free a cm address handle
2120 * @cm_node: The connection manager node to create AH for
2121 */
2122 static void
irdma_cm_free_ah(struct irdma_cm_node * cm_node)2123 irdma_cm_free_ah(struct irdma_cm_node *cm_node)
2124 {
2125 struct irdma_device *iwdev = cm_node->iwdev;
2126
2127 irdma_puda_free_ah(&iwdev->rf->sc_dev, cm_node->ah);
2128 cm_node->ah = NULL;
2129 }
2130
2131 /**
2132 * irdma_make_cm_node - create a new instance of a cm node
2133 * @cm_core: cm's core
2134 * @iwdev: iwarp device structure
2135 * @cm_info: quad info for connection
2136 * @listener: passive connection's listener
2137 */
2138 static struct irdma_cm_node *
irdma_make_cm_node(struct irdma_cm_core * cm_core,struct irdma_device * iwdev,struct irdma_cm_info * cm_info,struct irdma_cm_listener * listener)2139 irdma_make_cm_node(struct irdma_cm_core *cm_core, struct irdma_device *iwdev,
2140 struct irdma_cm_info *cm_info,
2141 struct irdma_cm_listener *listener)
2142 {
2143 struct irdma_cm_node *cm_node;
2144 int arpindex;
2145 if_t netdev = iwdev->netdev;
2146
2147 /* create an hte and cm_node for this instance */
2148 cm_node = kzalloc(sizeof(*cm_node), GFP_ATOMIC);
2149 if (!cm_node)
2150 return NULL;
2151
2152 /* set our node specific transport info */
2153 cm_node->ipv4 = cm_info->ipv4;
2154 cm_node->vlan_id = cm_info->vlan_id;
2155 if (cm_node->vlan_id >= VLAN_N_VID && iwdev->dcb_vlan_mode)
2156 cm_node->vlan_id = 0;
2157 cm_node->tos = cm_info->tos;
2158 cm_node->user_pri = cm_info->user_pri;
2159 if (listener) {
2160 if (listener->tos != cm_info->tos)
2161 irdma_dev_warn(&iwdev->ibdev,
2162 "application TOS[%d] and remote client TOS[%d] mismatch\n",
2163 listener->tos, cm_info->tos);
2164 if (iwdev->vsi.dscp_mode) {
2165 cm_node->user_pri = listener->user_pri;
2166 } else {
2167 cm_node->tos = max(listener->tos, cm_info->tos);
2168 cm_node->user_pri = rt_tos2priority(cm_node->tos);
2169 cm_node->user_pri =
2170 irdma_iw_get_vlan_prio(cm_info->loc_addr,
2171 cm_node->user_pri,
2172 cm_info->ipv4);
2173 }
2174 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_DCB,
2175 "listener: TOS:[%d] UP:[%d]\n", cm_node->tos,
2176 cm_node->user_pri);
2177 }
2178 memcpy(cm_node->loc_addr, cm_info->loc_addr, sizeof(cm_node->loc_addr));
2179 memcpy(cm_node->rem_addr, cm_info->rem_addr, sizeof(cm_node->rem_addr));
2180 cm_node->loc_port = cm_info->loc_port;
2181 cm_node->rem_port = cm_info->rem_port;
2182
2183 cm_node->mpa_frame_rev = IRDMA_CM_DEFAULT_MPA_VER;
2184 cm_node->send_rdma0_op = SEND_RDMA_READ_ZERO;
2185 cm_node->iwdev = iwdev;
2186 cm_node->dev = &iwdev->rf->sc_dev;
2187
2188 cm_node->ird_size = cm_node->dev->hw_attrs.max_hw_ird;
2189 cm_node->ord_size = cm_node->dev->hw_attrs.max_hw_ord;
2190
2191 cm_node->listener = listener;
2192 cm_node->cm_id = cm_info->cm_id;
2193 ether_addr_copy(cm_node->loc_mac, if_getlladdr(netdev));
2194 spin_lock_init(&cm_node->retrans_list_lock);
2195 cm_node->ack_rcvd = false;
2196
2197 init_completion(&cm_node->establish_comp);
2198 atomic_set(&cm_node->refcnt, 1);
2199 /* associate our parent CM core */
2200 cm_node->cm_core = cm_core;
2201 cm_node->tcp_cntxt.loc_id = IRDMA_CM_DEFAULT_LOCAL_ID;
2202 cm_node->tcp_cntxt.rcv_wscale = iwdev->rcv_wscale;
2203 cm_node->tcp_cntxt.rcv_wnd = iwdev->rcv_wnd >> cm_node->tcp_cntxt.rcv_wscale;
2204 kc_set_loc_seq_num_mss(cm_node);
2205
2206 arpindex = irdma_resolve_neigh_lpb_chk(iwdev, cm_node, cm_info);
2207 if (arpindex < 0)
2208 goto err;
2209
2210 ether_addr_copy(cm_node->rem_mac, iwdev->rf->arp_table[arpindex].mac_addr);
2211 irdma_add_hte_node(cm_core, cm_node);
2212 cm_core->stats_nodes_created++;
2213 return cm_node;
2214
2215 err:
2216 kfree(cm_node);
2217
2218 return NULL;
2219 }
2220
2221 static void
irdma_destroy_connection(struct irdma_cm_node * cm_node)2222 irdma_destroy_connection(struct irdma_cm_node *cm_node)
2223 {
2224 struct irdma_cm_core *cm_core = cm_node->cm_core;
2225 struct irdma_qp *iwqp;
2226 struct irdma_cm_info nfo;
2227
2228 /* if the node is destroyed before connection was accelerated */
2229 if (!cm_node->accelerated && cm_node->accept_pend) {
2230 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2231 "node destroyed before established\n");
2232 atomic_dec(&cm_node->listener->pend_accepts_cnt);
2233 }
2234 if (cm_node->close_entry)
2235 irdma_handle_close_entry(cm_node, 0);
2236 if (cm_node->listener) {
2237 irdma_dec_refcnt_listen(cm_core, cm_node->listener, 0, true);
2238 } else {
2239 if (cm_node->apbvt_set) {
2240 irdma_del_apbvt(cm_node->iwdev, cm_node->apbvt_entry);
2241 cm_node->apbvt_set = 0;
2242 }
2243 irdma_get_addr_info(cm_node, &nfo);
2244 if (cm_node->qhash_set) {
2245 nfo.qh_qpid = cm_node->iwdev->vsi.ilq->qp_id;
2246 irdma_manage_qhash(cm_node->iwdev, &nfo,
2247 IRDMA_QHASH_TYPE_TCP_ESTABLISHED,
2248 IRDMA_QHASH_MANAGE_TYPE_DELETE, NULL,
2249 false);
2250 cm_node->qhash_set = 0;
2251 }
2252 }
2253
2254 iwqp = cm_node->iwqp;
2255 if (iwqp) {
2256 cm_node->cm_id->rem_ref(cm_node->cm_id);
2257 cm_node->cm_id = NULL;
2258 iwqp->cm_id = NULL;
2259 irdma_qp_rem_ref(&iwqp->ibqp);
2260 cm_node->iwqp = NULL;
2261 } else if (cm_node->qhash_set) {
2262 irdma_get_addr_info(cm_node, &nfo);
2263 nfo.qh_qpid = cm_node->iwdev->vsi.ilq->qp_id;
2264 irdma_manage_qhash(cm_node->iwdev, &nfo,
2265 IRDMA_QHASH_TYPE_TCP_ESTABLISHED,
2266 IRDMA_QHASH_MANAGE_TYPE_DELETE, NULL, false);
2267 cm_node->qhash_set = 0;
2268 }
2269
2270 cm_core->cm_free_ah(cm_node);
2271 }
2272
2273 /**
2274 * irdma_rem_ref_cm_node - destroy an instance of a cm node
2275 * @cm_node: connection's node
2276 */
2277 void
irdma_rem_ref_cm_node(struct irdma_cm_node * cm_node)2278 irdma_rem_ref_cm_node(struct irdma_cm_node *cm_node)
2279 {
2280 struct irdma_cm_core *cm_core = cm_node->cm_core;
2281 unsigned long flags;
2282
2283 spin_lock_irqsave(&cm_core->ht_lock, flags);
2284
2285 if (!atomic_dec_and_test(&cm_node->refcnt)) {
2286 spin_unlock_irqrestore(&cm_core->ht_lock, flags);
2287 return;
2288 }
2289 if (cm_node->iwqp) {
2290 cm_node->iwqp->cm_node = NULL;
2291 cm_node->iwqp->cm_id = NULL;
2292 }
2293 HASH_DEL_RCU(cm_core->cm_hash_tbl, &cm_node->list);
2294 cm_node->cm_core->stats_nodes_destroyed++;
2295
2296 spin_unlock_irqrestore(&cm_core->ht_lock, flags);
2297
2298 irdma_destroy_connection(cm_node);
2299
2300 kfree_rcu(cm_node, rcu_head);
2301 }
2302
2303 /**
2304 * irdma_handle_fin_pkt - FIN packet received
2305 * @cm_node: connection's node
2306 */
2307 static void
irdma_handle_fin_pkt(struct irdma_cm_node * cm_node)2308 irdma_handle_fin_pkt(struct irdma_cm_node *cm_node)
2309 {
2310 switch (cm_node->state) {
2311 case IRDMA_CM_STATE_SYN_RCVD:
2312 case IRDMA_CM_STATE_SYN_SENT:
2313 case IRDMA_CM_STATE_ESTABLISHED:
2314 case IRDMA_CM_STATE_MPAREJ_RCVD:
2315 cm_node->tcp_cntxt.rcv_nxt++;
2316 irdma_cleanup_retrans_entry(cm_node);
2317 cm_node->state = IRDMA_CM_STATE_LAST_ACK;
2318 irdma_send_fin(cm_node);
2319 break;
2320 case IRDMA_CM_STATE_MPAREQ_SENT:
2321 irdma_create_event(cm_node, IRDMA_CM_EVENT_ABORTED);
2322 cm_node->tcp_cntxt.rcv_nxt++;
2323 irdma_cleanup_retrans_entry(cm_node);
2324 cm_node->state = IRDMA_CM_STATE_CLOSED;
2325 atomic_inc(&cm_node->refcnt);
2326 irdma_send_reset(cm_node);
2327 break;
2328 case IRDMA_CM_STATE_FIN_WAIT1:
2329 cm_node->tcp_cntxt.rcv_nxt++;
2330 irdma_cleanup_retrans_entry(cm_node);
2331 cm_node->state = IRDMA_CM_STATE_CLOSING;
2332 irdma_send_ack(cm_node);
2333 /*
2334 * Wait for ACK as this is simultaneous close. After we receive ACK, do not send anything. Just rm the
2335 * node.
2336 */
2337 break;
2338 case IRDMA_CM_STATE_FIN_WAIT2:
2339 cm_node->tcp_cntxt.rcv_nxt++;
2340 irdma_cleanup_retrans_entry(cm_node);
2341 cm_node->state = IRDMA_CM_STATE_TIME_WAIT;
2342 irdma_send_ack(cm_node);
2343 irdma_schedule_cm_timer(cm_node, NULL, IRDMA_TIMER_TYPE_CLOSE,
2344 1, 0);
2345 break;
2346 case IRDMA_CM_STATE_TIME_WAIT:
2347 cm_node->tcp_cntxt.rcv_nxt++;
2348 irdma_cleanup_retrans_entry(cm_node);
2349 cm_node->state = IRDMA_CM_STATE_CLOSED;
2350 irdma_rem_ref_cm_node(cm_node);
2351 break;
2352 case IRDMA_CM_STATE_OFFLOADED:
2353 default:
2354 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2355 "bad state node state = %d\n", cm_node->state);
2356 break;
2357 }
2358 }
2359
2360 /**
2361 * irdma_handle_rst_pkt - process received RST packet
2362 * @cm_node: connection's node
2363 * @rbuf: receive buffer
2364 */
2365 static void
irdma_handle_rst_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2366 irdma_handle_rst_pkt(struct irdma_cm_node *cm_node,
2367 struct irdma_puda_buf *rbuf)
2368 {
2369 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2370 "caller: %pS cm_node=%p state=%d rem_port=0x%04x loc_port=0x%04x rem_addr=%x loc_addr=%x\n",
2371 __builtin_return_address(0), cm_node, cm_node->state,
2372 cm_node->rem_port, cm_node->loc_port, cm_node->rem_addr[0],
2373 cm_node->loc_addr[0]);
2374
2375 irdma_cleanup_retrans_entry(cm_node);
2376 switch (cm_node->state) {
2377 case IRDMA_CM_STATE_SYN_SENT:
2378 case IRDMA_CM_STATE_MPAREQ_SENT:
2379 switch (cm_node->mpa_frame_rev) {
2380 case IETF_MPA_V2:
2381 /* Drop down to MPA_V1 */
2382 cm_node->mpa_frame_rev = IETF_MPA_V1;
2383 /* send a syn and goto syn sent state */
2384 cm_node->state = IRDMA_CM_STATE_SYN_SENT;
2385 if (irdma_send_syn(cm_node, 0))
2386 irdma_active_open_err(cm_node, false);
2387 break;
2388 case IETF_MPA_V1:
2389 default:
2390 irdma_active_open_err(cm_node, false);
2391 break;
2392 }
2393 break;
2394 case IRDMA_CM_STATE_MPAREQ_RCVD:
2395 atomic_inc(&cm_node->passive_state);
2396 break;
2397 case IRDMA_CM_STATE_ESTABLISHED:
2398 case IRDMA_CM_STATE_SYN_RCVD:
2399 case IRDMA_CM_STATE_LISTENING:
2400 irdma_passive_open_err(cm_node, false);
2401 break;
2402 case IRDMA_CM_STATE_OFFLOADED:
2403 irdma_active_open_err(cm_node, false);
2404 break;
2405 case IRDMA_CM_STATE_CLOSED:
2406 break;
2407 case IRDMA_CM_STATE_FIN_WAIT2:
2408 case IRDMA_CM_STATE_FIN_WAIT1:
2409 case IRDMA_CM_STATE_LAST_ACK:
2410 case IRDMA_CM_STATE_TIME_WAIT:
2411 cm_node->state = IRDMA_CM_STATE_CLOSED;
2412 irdma_rem_ref_cm_node(cm_node);
2413 break;
2414 default:
2415 break;
2416 }
2417 }
2418
2419 /**
2420 * irdma_handle_rcv_mpa - Process a recv'd mpa buffer
2421 * @cm_node: connection's node
2422 * @rbuf: receive buffer
2423 */
2424 static void
irdma_handle_rcv_mpa(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2425 irdma_handle_rcv_mpa(struct irdma_cm_node *cm_node,
2426 struct irdma_puda_buf *rbuf)
2427 {
2428 int err;
2429 int datasize = rbuf->datalen;
2430 u8 *dataloc = rbuf->data;
2431
2432 enum irdma_cm_event_type type = IRDMA_CM_EVENT_UNKNOWN;
2433 u32 res_type;
2434
2435 err = irdma_parse_mpa(cm_node, dataloc, &res_type, datasize);
2436 if (err) {
2437 if (cm_node->state == IRDMA_CM_STATE_MPAREQ_SENT)
2438 irdma_active_open_err(cm_node, true);
2439 else
2440 irdma_passive_open_err(cm_node, true);
2441 return;
2442 }
2443
2444 switch (cm_node->state) {
2445 case IRDMA_CM_STATE_ESTABLISHED:
2446 if (res_type == IRDMA_MPA_REQUEST_REJECT)
2447 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2448 "state for reject\n");
2449 cm_node->state = IRDMA_CM_STATE_MPAREQ_RCVD;
2450 type = IRDMA_CM_EVENT_MPA_REQ;
2451 irdma_send_ack(cm_node); /* ACK received MPA request */
2452 atomic_set(&cm_node->passive_state,
2453 IRDMA_PASSIVE_STATE_INDICATED);
2454 break;
2455 case IRDMA_CM_STATE_MPAREQ_SENT:
2456 irdma_cleanup_retrans_entry(cm_node);
2457 if (res_type == IRDMA_MPA_REQUEST_REJECT) {
2458 type = IRDMA_CM_EVENT_MPA_REJECT;
2459 cm_node->state = IRDMA_CM_STATE_MPAREJ_RCVD;
2460 } else {
2461 type = IRDMA_CM_EVENT_CONNECTED;
2462 cm_node->state = IRDMA_CM_STATE_OFFLOADED;
2463 }
2464 irdma_send_ack(cm_node);
2465 break;
2466 default:
2467 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2468 "wrong cm_node state=%d\n", cm_node->state);
2469 break;
2470 }
2471 irdma_create_event(cm_node, type);
2472 }
2473
2474 /**
2475 * irdma_check_syn - Check for error on received syn ack
2476 * @cm_node: connection's node
2477 * @tcph: pointer tcp header
2478 */
2479 static int
irdma_check_syn(struct irdma_cm_node * cm_node,struct tcphdr * tcph)2480 irdma_check_syn(struct irdma_cm_node *cm_node, struct tcphdr *tcph)
2481 {
2482 if (ntohl(tcph->th_ack) != cm_node->tcp_cntxt.loc_seq_num) {
2483 irdma_active_open_err(cm_node, true);
2484 return 1;
2485 }
2486
2487 return 0;
2488 }
2489
2490 /**
2491 * irdma_check_seq - check seq numbers if OK
2492 * @cm_node: connection's node
2493 * @tcph: pointer tcp header
2494 */
2495 static int
irdma_check_seq(struct irdma_cm_node * cm_node,struct tcphdr * tcph)2496 irdma_check_seq(struct irdma_cm_node *cm_node, struct tcphdr *tcph)
2497 {
2498 u32 seq;
2499 u32 ack_seq;
2500 u32 loc_seq_num = cm_node->tcp_cntxt.loc_seq_num;
2501 u32 rcv_nxt = cm_node->tcp_cntxt.rcv_nxt;
2502 u32 rcv_wnd;
2503 int err = 0;
2504
2505 seq = ntohl(tcph->th_seq);
2506 ack_seq = ntohl(tcph->th_ack);
2507 rcv_wnd = cm_node->tcp_cntxt.rcv_wnd;
2508 if (ack_seq != loc_seq_num ||
2509 !between(seq, rcv_nxt, (rcv_nxt + rcv_wnd)))
2510 err = -1;
2511 if (err)
2512 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2513 "seq number err\n");
2514
2515 return err;
2516 }
2517
2518 void
irdma_add_conn_est_qh(struct irdma_cm_node * cm_node)2519 irdma_add_conn_est_qh(struct irdma_cm_node *cm_node)
2520 {
2521 struct irdma_cm_info nfo;
2522
2523 irdma_get_addr_info(cm_node, &nfo);
2524 nfo.qh_qpid = cm_node->iwdev->vsi.ilq->qp_id;
2525 irdma_manage_qhash(cm_node->iwdev, &nfo,
2526 IRDMA_QHASH_TYPE_TCP_ESTABLISHED,
2527 IRDMA_QHASH_MANAGE_TYPE_ADD,
2528 cm_node, false);
2529 cm_node->qhash_set = true;
2530 }
2531
2532 /**
2533 * irdma_handle_syn_pkt - is for Passive node
2534 * @cm_node: connection's node
2535 * @rbuf: receive buffer
2536 */
2537 static void
irdma_handle_syn_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2538 irdma_handle_syn_pkt(struct irdma_cm_node *cm_node,
2539 struct irdma_puda_buf *rbuf)
2540 {
2541 struct tcphdr *tcph = (struct tcphdr *)rbuf->tcph;
2542 int err;
2543 u32 inc_sequence;
2544 int optionsize;
2545
2546 optionsize = (tcph->th_off << 2) - sizeof(*tcph);
2547 inc_sequence = ntohl(tcph->th_seq);
2548
2549 switch (cm_node->state) {
2550 case IRDMA_CM_STATE_SYN_SENT:
2551 case IRDMA_CM_STATE_MPAREQ_SENT:
2552 /* Rcvd syn on active open connection */
2553 irdma_active_open_err(cm_node, 1);
2554 break;
2555 case IRDMA_CM_STATE_LISTENING:
2556 /* Passive OPEN */
2557 if (atomic_read(&cm_node->listener->pend_accepts_cnt) >
2558 cm_node->listener->backlog) {
2559 cm_node->cm_core->stats_backlog_drops++;
2560 irdma_passive_open_err(cm_node, false);
2561 break;
2562 }
2563 err = irdma_handle_tcp_options(cm_node, tcph, optionsize, 1);
2564 if (err) {
2565 irdma_passive_open_err(cm_node, false);
2566 /* drop pkt */
2567 break;
2568 }
2569 err = cm_node->cm_core->cm_create_ah(cm_node, false);
2570 if (err) {
2571 irdma_passive_open_err(cm_node, false);
2572 /* drop pkt */
2573 break;
2574 }
2575 cm_node->tcp_cntxt.rcv_nxt = inc_sequence + 1;
2576 cm_node->accept_pend = 1;
2577 atomic_inc(&cm_node->listener->pend_accepts_cnt);
2578
2579 cm_node->state = IRDMA_CM_STATE_SYN_RCVD;
2580 break;
2581 case IRDMA_CM_STATE_CLOSED:
2582 irdma_cleanup_retrans_entry(cm_node);
2583 atomic_inc(&cm_node->refcnt);
2584 irdma_send_reset(cm_node);
2585 break;
2586 case IRDMA_CM_STATE_OFFLOADED:
2587 case IRDMA_CM_STATE_ESTABLISHED:
2588 case IRDMA_CM_STATE_FIN_WAIT1:
2589 case IRDMA_CM_STATE_FIN_WAIT2:
2590 case IRDMA_CM_STATE_MPAREQ_RCVD:
2591 case IRDMA_CM_STATE_LAST_ACK:
2592 case IRDMA_CM_STATE_CLOSING:
2593 case IRDMA_CM_STATE_UNKNOWN:
2594 default:
2595 break;
2596 }
2597 }
2598
2599 /**
2600 * irdma_handle_synack_pkt - Process SYN+ACK packet (active side)
2601 * @cm_node: connection's node
2602 * @rbuf: receive buffer
2603 */
2604 static void
irdma_handle_synack_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2605 irdma_handle_synack_pkt(struct irdma_cm_node *cm_node,
2606 struct irdma_puda_buf *rbuf)
2607 {
2608 struct tcphdr *tcph = (struct tcphdr *)rbuf->tcph;
2609 int err;
2610 u32 inc_sequence;
2611 int optionsize;
2612
2613 optionsize = (tcph->th_off << 2) - sizeof(*tcph);
2614 inc_sequence = ntohl(tcph->th_seq);
2615 switch (cm_node->state) {
2616 case IRDMA_CM_STATE_SYN_SENT:
2617 irdma_cleanup_retrans_entry(cm_node);
2618 /* active open */
2619 if (irdma_check_syn(cm_node, tcph)) {
2620 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2621 "check syn fail\n");
2622 return;
2623 }
2624 cm_node->tcp_cntxt.rem_ack_num = ntohl(tcph->th_ack);
2625 /* setup options */
2626 err = irdma_handle_tcp_options(cm_node, tcph, optionsize, 0);
2627 if (err) {
2628 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2629 "cm_node=%p tcp_options failed\n", cm_node);
2630 break;
2631 }
2632 irdma_cleanup_retrans_entry(cm_node);
2633 cm_node->tcp_cntxt.rcv_nxt = inc_sequence + 1;
2634 irdma_send_ack(cm_node); /* ACK for the syn_ack */
2635 err = irdma_send_mpa_request(cm_node);
2636 if (err) {
2637 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2638 "cm_node=%p irdma_send_mpa_request failed\n",
2639 cm_node);
2640 break;
2641 }
2642 cm_node->state = IRDMA_CM_STATE_MPAREQ_SENT;
2643 break;
2644 case IRDMA_CM_STATE_MPAREQ_RCVD:
2645 irdma_passive_open_err(cm_node, true);
2646 break;
2647 case IRDMA_CM_STATE_LISTENING:
2648 cm_node->tcp_cntxt.loc_seq_num = ntohl(tcph->th_ack);
2649 irdma_cleanup_retrans_entry(cm_node);
2650 cm_node->state = IRDMA_CM_STATE_CLOSED;
2651 irdma_send_reset(cm_node);
2652 break;
2653 case IRDMA_CM_STATE_CLOSED:
2654 cm_node->tcp_cntxt.loc_seq_num = ntohl(tcph->th_ack);
2655 irdma_cleanup_retrans_entry(cm_node);
2656 atomic_inc(&cm_node->refcnt);
2657 irdma_send_reset(cm_node);
2658 break;
2659 case IRDMA_CM_STATE_ESTABLISHED:
2660 case IRDMA_CM_STATE_FIN_WAIT1:
2661 case IRDMA_CM_STATE_FIN_WAIT2:
2662 case IRDMA_CM_STATE_LAST_ACK:
2663 case IRDMA_CM_STATE_OFFLOADED:
2664 case IRDMA_CM_STATE_CLOSING:
2665 case IRDMA_CM_STATE_UNKNOWN:
2666 case IRDMA_CM_STATE_MPAREQ_SENT:
2667 default:
2668 break;
2669 }
2670 }
2671
2672 /**
2673 * irdma_handle_ack_pkt - process packet with ACK
2674 * @cm_node: connection's node
2675 * @rbuf: receive buffer
2676 */
2677 static int
irdma_handle_ack_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2678 irdma_handle_ack_pkt(struct irdma_cm_node *cm_node,
2679 struct irdma_puda_buf *rbuf)
2680 {
2681 struct tcphdr *tcph = (struct tcphdr *)rbuf->tcph;
2682 u32 inc_sequence;
2683 int ret;
2684 int optionsize;
2685 u32 datasize = rbuf->datalen;
2686
2687 optionsize = (tcph->th_off << 2) - sizeof(*tcph);
2688
2689 if (irdma_check_seq(cm_node, tcph))
2690 return -EINVAL;
2691
2692 inc_sequence = ntohl(tcph->th_seq);
2693 switch (cm_node->state) {
2694 case IRDMA_CM_STATE_SYN_RCVD:
2695 irdma_cleanup_retrans_entry(cm_node);
2696 ret = irdma_handle_tcp_options(cm_node, tcph, optionsize, 1);
2697 if (ret)
2698 return ret;
2699 cm_node->tcp_cntxt.rem_ack_num = ntohl(tcph->th_ack);
2700 cm_node->state = IRDMA_CM_STATE_ESTABLISHED;
2701 if (datasize) {
2702 cm_node->tcp_cntxt.rcv_nxt = inc_sequence + datasize;
2703 irdma_handle_rcv_mpa(cm_node, rbuf);
2704 }
2705 break;
2706 case IRDMA_CM_STATE_ESTABLISHED:
2707 irdma_cleanup_retrans_entry(cm_node);
2708 if (datasize) {
2709 cm_node->tcp_cntxt.rcv_nxt = inc_sequence + datasize;
2710 irdma_handle_rcv_mpa(cm_node, rbuf);
2711 }
2712 break;
2713 case IRDMA_CM_STATE_MPAREQ_SENT:
2714 cm_node->tcp_cntxt.rem_ack_num = ntohl(tcph->th_ack);
2715 if (datasize) {
2716 cm_node->tcp_cntxt.rcv_nxt = inc_sequence + datasize;
2717 cm_node->ack_rcvd = false;
2718 irdma_handle_rcv_mpa(cm_node, rbuf);
2719 } else {
2720 cm_node->ack_rcvd = true;
2721 }
2722 break;
2723 case IRDMA_CM_STATE_LISTENING:
2724 irdma_cleanup_retrans_entry(cm_node);
2725 cm_node->state = IRDMA_CM_STATE_CLOSED;
2726 irdma_send_reset(cm_node);
2727 break;
2728 case IRDMA_CM_STATE_CLOSED:
2729 irdma_cleanup_retrans_entry(cm_node);
2730 atomic_inc(&cm_node->refcnt);
2731 irdma_send_reset(cm_node);
2732 break;
2733 case IRDMA_CM_STATE_LAST_ACK:
2734 case IRDMA_CM_STATE_CLOSING:
2735 irdma_cleanup_retrans_entry(cm_node);
2736 cm_node->state = IRDMA_CM_STATE_CLOSED;
2737 irdma_rem_ref_cm_node(cm_node);
2738 break;
2739 case IRDMA_CM_STATE_FIN_WAIT1:
2740 irdma_cleanup_retrans_entry(cm_node);
2741 cm_node->state = IRDMA_CM_STATE_FIN_WAIT2;
2742 break;
2743 case IRDMA_CM_STATE_SYN_SENT:
2744 case IRDMA_CM_STATE_FIN_WAIT2:
2745 case IRDMA_CM_STATE_OFFLOADED:
2746 case IRDMA_CM_STATE_MPAREQ_RCVD:
2747 case IRDMA_CM_STATE_UNKNOWN:
2748 default:
2749 irdma_cleanup_retrans_entry(cm_node);
2750 break;
2751 }
2752
2753 return 0;
2754 }
2755
2756 /**
2757 * irdma_process_pkt - process cm packet
2758 * @cm_node: connection's node
2759 * @rbuf: receive buffer
2760 */
2761 static void
irdma_process_pkt(struct irdma_cm_node * cm_node,struct irdma_puda_buf * rbuf)2762 irdma_process_pkt(struct irdma_cm_node *cm_node,
2763 struct irdma_puda_buf *rbuf)
2764 {
2765 enum irdma_tcpip_pkt_type pkt_type = IRDMA_PKT_TYPE_UNKNOWN;
2766 struct tcphdr *tcph = (struct tcphdr *)rbuf->tcph;
2767 u32 fin_set = 0;
2768 int err;
2769
2770 if (tcp_get_flags(tcph) & TH_RST) {
2771 pkt_type = IRDMA_PKT_TYPE_RST;
2772 } else if (tcp_get_flags(tcph) & TH_SYN) {
2773 pkt_type = IRDMA_PKT_TYPE_SYN;
2774 if (tcp_get_flags(tcph) & TH_ACK)
2775 pkt_type = IRDMA_PKT_TYPE_SYNACK;
2776 } else if (tcp_get_flags(tcph) & TH_ACK) {
2777 pkt_type = IRDMA_PKT_TYPE_ACK;
2778 }
2779 if (tcp_get_flags(tcph) & TH_FIN)
2780 fin_set = 1;
2781
2782 switch (pkt_type) {
2783 case IRDMA_PKT_TYPE_SYN:
2784 irdma_handle_syn_pkt(cm_node, rbuf);
2785 break;
2786 case IRDMA_PKT_TYPE_SYNACK:
2787 irdma_handle_synack_pkt(cm_node, rbuf);
2788 break;
2789 case IRDMA_PKT_TYPE_ACK:
2790 err = irdma_handle_ack_pkt(cm_node, rbuf);
2791 if (fin_set && !err)
2792 irdma_handle_fin_pkt(cm_node);
2793 break;
2794 case IRDMA_PKT_TYPE_RST:
2795 irdma_handle_rst_pkt(cm_node, rbuf);
2796 break;
2797 default:
2798 if (fin_set &&
2799 (!irdma_check_seq(cm_node, (struct tcphdr *)rbuf->tcph)))
2800 irdma_handle_fin_pkt(cm_node);
2801 break;
2802 }
2803 }
2804
2805 /**
2806 * irdma_make_listen_node - create a listen node with params
2807 * @cm_core: cm's core
2808 * @iwdev: iwarp device structure
2809 * @cm_info: quad info for connection
2810 */
2811 static struct irdma_cm_listener *
irdma_make_listen_node(struct irdma_cm_core * cm_core,struct irdma_device * iwdev,struct irdma_cm_info * cm_info)2812 irdma_make_listen_node(struct irdma_cm_core *cm_core,
2813 struct irdma_device *iwdev,
2814 struct irdma_cm_info *cm_info)
2815 {
2816 struct irdma_cm_listener *listener;
2817 unsigned long flags;
2818
2819 /* cannot have multiple matching listeners */
2820 listener = irdma_find_listener(cm_core, cm_info->loc_addr, cm_info->ipv4,
2821 cm_info->loc_port, cm_info->vlan_id,
2822 IRDMA_CM_LISTENER_EITHER_STATE);
2823 if (listener &&
2824 listener->listener_state == IRDMA_CM_LISTENER_ACTIVE_STATE) {
2825 atomic_dec(&listener->refcnt);
2826 return NULL;
2827 }
2828
2829 if (!listener) {
2830 /*
2831 * create a CM listen node 1/2 node to compare incoming traffic to
2832 */
2833 listener = kzalloc(sizeof(*listener), GFP_KERNEL);
2834 if (!listener)
2835 return NULL;
2836 cm_core->stats_listen_nodes_created++;
2837 memcpy(listener->loc_addr, cm_info->loc_addr,
2838 sizeof(listener->loc_addr));
2839 listener->loc_port = cm_info->loc_port;
2840
2841 INIT_LIST_HEAD(&listener->child_listen_list);
2842
2843 atomic_set(&listener->refcnt, 1);
2844 } else {
2845 listener->reused_node = 1;
2846 }
2847
2848 listener->cm_id = cm_info->cm_id;
2849 listener->ipv4 = cm_info->ipv4;
2850 listener->vlan_id = cm_info->vlan_id;
2851 atomic_set(&listener->pend_accepts_cnt, 0);
2852 listener->cm_core = cm_core;
2853 listener->iwdev = iwdev;
2854
2855 listener->backlog = cm_info->backlog;
2856 listener->listener_state = IRDMA_CM_LISTENER_ACTIVE_STATE;
2857
2858 if (!listener->reused_node) {
2859 spin_lock_irqsave(&cm_core->listen_list_lock, flags);
2860 list_add(&listener->list, &cm_core->listen_list);
2861 spin_unlock_irqrestore(&cm_core->listen_list_lock, flags);
2862 }
2863
2864 return listener;
2865 }
2866
2867 /**
2868 * irdma_create_cm_node - make a connection node with params
2869 * @cm_core: cm's core
2870 * @iwdev: iwarp device structure
2871 * @conn_param: connection parameters
2872 * @cm_info: quad info for connection
2873 * @caller_cm_node: pointer to cm_node structure to return
2874 */
2875 static int
irdma_create_cm_node(struct irdma_cm_core * cm_core,struct irdma_device * iwdev,struct iw_cm_conn_param * conn_param,struct irdma_cm_info * cm_info,struct irdma_cm_node ** caller_cm_node)2876 irdma_create_cm_node(struct irdma_cm_core *cm_core,
2877 struct irdma_device *iwdev,
2878 struct iw_cm_conn_param *conn_param,
2879 struct irdma_cm_info *cm_info,
2880 struct irdma_cm_node **caller_cm_node)
2881 {
2882 struct irdma_cm_node *cm_node;
2883 u16 private_data_len = conn_param->private_data_len;
2884 const void *private_data = conn_param->private_data;
2885
2886 /* create a CM connection node */
2887 cm_node = irdma_make_cm_node(cm_core, iwdev, cm_info, NULL);
2888 if (!cm_node)
2889 return -ENOMEM;
2890
2891 /* set our node side to client (active) side */
2892 cm_node->tcp_cntxt.client = 1;
2893 cm_node->tcp_cntxt.rcv_wscale = IRDMA_CM_DEFAULT_RCV_WND_SCALE;
2894
2895 irdma_record_ird_ord(cm_node, conn_param->ird, conn_param->ord);
2896
2897 cm_node->pdata.size = private_data_len;
2898 cm_node->pdata.addr = cm_node->pdata_buf;
2899
2900 memcpy(cm_node->pdata_buf, private_data, private_data_len);
2901 *caller_cm_node = cm_node;
2902
2903 return 0;
2904 }
2905
2906 /**
2907 * irdma_cm_reject - reject and teardown a connection
2908 * @cm_node: connection's node
2909 * @pdata: ptr to private data for reject
2910 * @plen: size of private data
2911 */
2912 static int
irdma_cm_reject(struct irdma_cm_node * cm_node,const void * pdata,u8 plen)2913 irdma_cm_reject(struct irdma_cm_node *cm_node, const void *pdata,
2914 u8 plen)
2915 {
2916 int ret;
2917 int passive_state;
2918
2919 if (cm_node->tcp_cntxt.client)
2920 return 0;
2921
2922 irdma_cleanup_retrans_entry(cm_node);
2923
2924 passive_state = atomic_add_return(1, &cm_node->passive_state);
2925 if (passive_state == IRDMA_SEND_RESET_EVENT) {
2926 cm_node->state = IRDMA_CM_STATE_CLOSED;
2927 irdma_rem_ref_cm_node(cm_node);
2928 return 0;
2929 }
2930
2931 if (cm_node->state == IRDMA_CM_STATE_LISTENER_DESTROYED) {
2932 irdma_rem_ref_cm_node(cm_node);
2933 return 0;
2934 }
2935
2936 ret = irdma_send_mpa_reject(cm_node, pdata, plen);
2937 if (!ret)
2938 return 0;
2939
2940 cm_node->state = IRDMA_CM_STATE_CLOSED;
2941 if (irdma_send_reset(cm_node))
2942 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2943 "send reset failed\n");
2944
2945 return ret;
2946 }
2947
2948 /**
2949 * irdma_cm_close - close of cm connection
2950 * @cm_node: connection's node
2951 */
2952 static int
irdma_cm_close(struct irdma_cm_node * cm_node)2953 irdma_cm_close(struct irdma_cm_node *cm_node)
2954 {
2955 switch (cm_node->state) {
2956 case IRDMA_CM_STATE_SYN_RCVD:
2957 case IRDMA_CM_STATE_SYN_SENT:
2958 case IRDMA_CM_STATE_ONE_SIDE_ESTABLISHED:
2959 case IRDMA_CM_STATE_ESTABLISHED:
2960 case IRDMA_CM_STATE_ACCEPTING:
2961 case IRDMA_CM_STATE_MPAREQ_SENT:
2962 case IRDMA_CM_STATE_MPAREQ_RCVD:
2963 irdma_cleanup_retrans_entry(cm_node);
2964 irdma_send_reset(cm_node);
2965 break;
2966 case IRDMA_CM_STATE_CLOSE_WAIT:
2967 cm_node->state = IRDMA_CM_STATE_LAST_ACK;
2968 irdma_send_fin(cm_node);
2969 break;
2970 case IRDMA_CM_STATE_FIN_WAIT1:
2971 case IRDMA_CM_STATE_FIN_WAIT2:
2972 case IRDMA_CM_STATE_LAST_ACK:
2973 case IRDMA_CM_STATE_TIME_WAIT:
2974 case IRDMA_CM_STATE_CLOSING:
2975 return -EINVAL;
2976 case IRDMA_CM_STATE_LISTENING:
2977 irdma_cleanup_retrans_entry(cm_node);
2978 irdma_send_reset(cm_node);
2979 break;
2980 case IRDMA_CM_STATE_MPAREJ_RCVD:
2981 case IRDMA_CM_STATE_UNKNOWN:
2982 case IRDMA_CM_STATE_INITED:
2983 case IRDMA_CM_STATE_CLOSED:
2984 case IRDMA_CM_STATE_LISTENER_DESTROYED:
2985 irdma_rem_ref_cm_node(cm_node);
2986 break;
2987 case IRDMA_CM_STATE_OFFLOADED:
2988 if (cm_node->send_entry)
2989 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
2990 "CM send_entry in OFFLOADED state\n");
2991 irdma_rem_ref_cm_node(cm_node);
2992 break;
2993 }
2994
2995 return 0;
2996 }
2997
2998 /**
2999 * irdma_receive_ilq - recv an ETHERNET packet, and process it
3000 * through CM
3001 * @vsi: VSI structure of dev
3002 * @rbuf: receive buffer
3003 */
3004 void
irdma_receive_ilq(struct irdma_sc_vsi * vsi,struct irdma_puda_buf * rbuf)3005 irdma_receive_ilq(struct irdma_sc_vsi *vsi, struct irdma_puda_buf *rbuf)
3006 {
3007 struct irdma_cm_node *cm_node;
3008 struct irdma_cm_listener *listener;
3009 struct ip *iph;
3010 struct ip6_hdr *ip6h;
3011 struct tcphdr *tcph;
3012 struct irdma_cm_info cm_info = {0};
3013 struct irdma_device *iwdev = vsi->back_vsi;
3014 struct irdma_cm_core *cm_core = &iwdev->cm_core;
3015 struct ether_vlan_header *ethh;
3016 u16 vtag;
3017
3018 /* if vlan, then maclen = 18 else 14 */
3019 iph = (struct ip *)rbuf->iph;
3020 irdma_debug_buf(vsi->dev, IRDMA_DEBUG_ILQ, "RECEIVE ILQ BUFFER",
3021 rbuf->mem.va, rbuf->totallen);
3022 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) {
3023 if (rbuf->vlan_valid) {
3024 vtag = rbuf->vlan_id;
3025 cm_info.user_pri = (vtag & EVL_PRI_MASK) >>
3026 VLAN_PRIO_SHIFT;
3027 cm_info.vlan_id = vtag & EVL_VLID_MASK;
3028 } else {
3029 cm_info.vlan_id = 0xFFFF;
3030 }
3031 } else {
3032 ethh = rbuf->mem.va;
3033
3034 if (ethh->evl_proto == htons(ETH_P_8021Q)) {
3035 vtag = ntohs(ethh->evl_tag);
3036 cm_info.user_pri = (vtag & EVL_PRI_MASK) >>
3037 VLAN_PRIO_SHIFT;
3038 cm_info.vlan_id = vtag & EVL_VLID_MASK;
3039 irdma_debug(&cm_core->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3040 "vlan_id=%d\n", cm_info.vlan_id);
3041 } else {
3042 cm_info.vlan_id = 0xFFFF;
3043 }
3044 }
3045 tcph = (struct tcphdr *)rbuf->tcph;
3046
3047 if (rbuf->ipv4) {
3048 cm_info.loc_addr[0] = ntohl(iph->ip_dst.s_addr);
3049 cm_info.rem_addr[0] = ntohl(iph->ip_src.s_addr);
3050 cm_info.ipv4 = true;
3051 cm_info.tos = iph->ip_tos;
3052 } else {
3053 ip6h = (struct ip6_hdr *)rbuf->iph;
3054 irdma_copy_ip_ntohl(cm_info.loc_addr,
3055 ip6h->ip6_dst.__u6_addr.__u6_addr32);
3056 irdma_copy_ip_ntohl(cm_info.rem_addr,
3057 ip6h->ip6_src.__u6_addr.__u6_addr32);
3058 cm_info.ipv4 = false;
3059 cm_info.tos = (ip6h->ip6_vfc << 4) | ip6h->ip6_flow;
3060 }
3061 cm_info.loc_port = ntohs(tcph->th_dport);
3062 cm_info.rem_port = ntohs(tcph->th_sport);
3063 cm_node = irdma_find_node(cm_core, cm_info.rem_port, cm_info.rem_addr,
3064 cm_info.loc_port, cm_info.loc_addr, cm_info.vlan_id);
3065
3066 if (!cm_node) {
3067 /*
3068 * Only type of packet accepted are for the PASSIVE open (syn only)
3069 */
3070 if (!(tcp_get_flags(tcph) & TH_SYN) || tcp_get_flags(tcph) & TH_ACK)
3071 return;
3072
3073 listener = irdma_find_listener(cm_core,
3074 cm_info.loc_addr,
3075 cm_info.ipv4,
3076 cm_info.loc_port,
3077 cm_info.vlan_id,
3078 IRDMA_CM_LISTENER_ACTIVE_STATE);
3079 if (!listener) {
3080 cm_info.cm_id = NULL;
3081 irdma_debug(&cm_core->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3082 "no listener found\n");
3083 return;
3084 }
3085
3086 cm_info.cm_id = listener->cm_id;
3087 cm_node = irdma_make_cm_node(cm_core, iwdev, &cm_info,
3088 listener);
3089 if (!cm_node) {
3090 irdma_debug(&cm_core->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3091 "allocate node failed\n");
3092 atomic_dec(&listener->refcnt);
3093 return;
3094 }
3095
3096 if (!(tcp_get_flags(tcph) & (TH_RST | TH_FIN))) {
3097 cm_node->state = IRDMA_CM_STATE_LISTENING;
3098 } else {
3099 irdma_rem_ref_cm_node(cm_node);
3100 return;
3101 }
3102
3103 atomic_inc(&cm_node->refcnt);
3104 } else if (cm_node->state == IRDMA_CM_STATE_OFFLOADED) {
3105 irdma_rem_ref_cm_node(cm_node);
3106 return;
3107 }
3108
3109 irdma_process_pkt(cm_node, rbuf);
3110 irdma_rem_ref_cm_node(cm_node);
3111 }
3112
3113 static int
irdma_add_qh(struct irdma_cm_node * cm_node,bool active)3114 irdma_add_qh(struct irdma_cm_node *cm_node, bool active)
3115 {
3116 if (!active)
3117 irdma_add_conn_est_qh(cm_node);
3118 return 0;
3119 }
3120
3121 static void
irdma_cm_free_ah_nop(struct irdma_cm_node * cm_node)3122 irdma_cm_free_ah_nop(struct irdma_cm_node *cm_node)
3123 {
3124 }
3125
3126 /**
3127 * irdma_setup_cm_core - setup top level instance of a cm core
3128 * @iwdev: iwarp device structure
3129 * @rdma_ver: HW version
3130 */
3131 int
irdma_setup_cm_core(struct irdma_device * iwdev,u8 rdma_ver)3132 irdma_setup_cm_core(struct irdma_device *iwdev, u8 rdma_ver)
3133 {
3134 struct irdma_cm_core *cm_core = &iwdev->cm_core;
3135
3136 cm_core->iwdev = iwdev;
3137 cm_core->dev = &iwdev->rf->sc_dev;
3138
3139 /* Handles CM event work items send to Iwarp core */
3140 cm_core->event_wq = alloc_ordered_workqueue("iwarp-event-wq", 0);
3141 if (!cm_core->event_wq)
3142 return -ENOMEM;
3143
3144 INIT_LIST_HEAD(&cm_core->listen_list);
3145
3146 timer_setup(&cm_core->tcp_timer, irdma_cm_timer_tick, 0);
3147
3148 spin_lock_init(&cm_core->ht_lock);
3149 spin_lock_init(&cm_core->listen_list_lock);
3150 spin_lock_init(&cm_core->apbvt_lock);
3151 switch (rdma_ver) {
3152 case IRDMA_GEN_1:
3153 cm_core->form_cm_frame = irdma_form_uda_cm_frame;
3154 cm_core->cm_create_ah = irdma_add_qh;
3155 cm_core->cm_free_ah = irdma_cm_free_ah_nop;
3156 break;
3157 case IRDMA_GEN_2:
3158 default:
3159 cm_core->form_cm_frame = irdma_form_ah_cm_frame;
3160 cm_core->cm_create_ah = irdma_cm_create_ah;
3161 cm_core->cm_free_ah = irdma_cm_free_ah;
3162 }
3163
3164 return 0;
3165 }
3166
3167 /**
3168 * irdma_cleanup_cm_core - deallocate a top level instance of a
3169 * cm core
3170 * @cm_core: cm's core
3171 */
3172 void
irdma_cleanup_cm_core(struct irdma_cm_core * cm_core)3173 irdma_cleanup_cm_core(struct irdma_cm_core *cm_core)
3174 {
3175 if (!cm_core)
3176 return;
3177
3178 del_timer_sync(&cm_core->tcp_timer);
3179
3180 destroy_workqueue(cm_core->event_wq);
3181 cm_core->dev->ws_reset(&cm_core->iwdev->vsi);
3182 }
3183
3184 /**
3185 * irdma_init_tcp_ctx - setup qp context
3186 * @cm_node: connection's node
3187 * @tcp_info: offload info for tcp
3188 * @iwqp: associate qp for the connection
3189 */
3190 static void
irdma_init_tcp_ctx(struct irdma_cm_node * cm_node,struct irdma_tcp_offload_info * tcp_info,struct irdma_qp * iwqp)3191 irdma_init_tcp_ctx(struct irdma_cm_node *cm_node,
3192 struct irdma_tcp_offload_info *tcp_info,
3193 struct irdma_qp *iwqp)
3194 {
3195 tcp_info->ipv4 = cm_node->ipv4;
3196 tcp_info->drop_ooo_seg = !iwqp->iwdev->iw_ooo;
3197 tcp_info->wscale = true;
3198 tcp_info->ignore_tcp_opt = true;
3199 tcp_info->ignore_tcp_uns_opt = true;
3200 tcp_info->no_nagle = false;
3201
3202 tcp_info->ttl = IRDMA_DEFAULT_TTL;
3203 tcp_info->rtt_var = IRDMA_DEFAULT_RTT_VAR;
3204 tcp_info->ss_thresh = IRDMA_DEFAULT_SS_THRESH;
3205 tcp_info->rexmit_thresh = IRDMA_DEFAULT_REXMIT_THRESH;
3206
3207 tcp_info->tcp_state = IRDMA_TCP_STATE_ESTABLISHED;
3208 tcp_info->snd_wscale = cm_node->tcp_cntxt.snd_wscale;
3209 tcp_info->rcv_wscale = cm_node->tcp_cntxt.rcv_wscale;
3210
3211 tcp_info->snd_nxt = cm_node->tcp_cntxt.loc_seq_num;
3212 tcp_info->snd_wnd = cm_node->tcp_cntxt.snd_wnd;
3213 tcp_info->rcv_nxt = cm_node->tcp_cntxt.rcv_nxt;
3214 tcp_info->snd_max = cm_node->tcp_cntxt.loc_seq_num;
3215
3216 tcp_info->snd_una = cm_node->tcp_cntxt.loc_seq_num;
3217 tcp_info->cwnd = 2 * cm_node->tcp_cntxt.mss;
3218 tcp_info->snd_wl1 = cm_node->tcp_cntxt.rcv_nxt;
3219 tcp_info->snd_wl2 = cm_node->tcp_cntxt.loc_seq_num;
3220 tcp_info->max_snd_window = cm_node->tcp_cntxt.max_snd_wnd;
3221 tcp_info->rcv_wnd = cm_node->tcp_cntxt.rcv_wnd
3222 << cm_node->tcp_cntxt.rcv_wscale;
3223
3224 tcp_info->flow_label = 0;
3225 tcp_info->snd_mss = (u32)cm_node->tcp_cntxt.mss;
3226 tcp_info->tos = cm_node->tos;
3227 if (cm_node->vlan_id < VLAN_N_VID) {
3228 tcp_info->insert_vlan_tag = true;
3229 tcp_info->vlan_tag = cm_node->vlan_id;
3230 tcp_info->vlan_tag |= cm_node->user_pri << VLAN_PRIO_SHIFT;
3231 }
3232 tcp_info->src_port = cm_node->loc_port;
3233 tcp_info->dst_port = cm_node->rem_port;
3234 tcp_info->arp_idx = (u16)irdma_arp_table(iwqp->iwdev->rf,
3235 cm_node->rem_addr, NULL,
3236 IRDMA_ARP_RESOLVE);
3237 if (cm_node->ipv4) {
3238 tcp_info->dest_ip_addr[3] = cm_node->rem_addr[0];
3239 tcp_info->local_ipaddr[3] = cm_node->loc_addr[0];
3240 } else {
3241 memcpy(tcp_info->dest_ip_addr, cm_node->rem_addr,
3242 sizeof(tcp_info->dest_ip_addr));
3243 memcpy(tcp_info->local_ipaddr, cm_node->loc_addr,
3244 sizeof(tcp_info->local_ipaddr));
3245 }
3246 }
3247
3248 /**
3249 * irdma_cm_init_tsa_conn - setup qp for RTS
3250 * @iwqp: associate qp for the connection
3251 * @cm_node: connection's node
3252 */
3253 static void
irdma_cm_init_tsa_conn(struct irdma_qp * iwqp,struct irdma_cm_node * cm_node)3254 irdma_cm_init_tsa_conn(struct irdma_qp *iwqp,
3255 struct irdma_cm_node *cm_node)
3256 {
3257 struct irdma_iwarp_offload_info *iwarp_info;
3258 struct irdma_qp_host_ctx_info *ctx_info;
3259
3260 iwarp_info = &iwqp->iwarp_info;
3261 ctx_info = &iwqp->ctx_info;
3262
3263 ctx_info->tcp_info = &iwqp->tcp_info;
3264 ctx_info->send_cq_num = iwqp->iwscq->sc_cq.cq_uk.cq_id;
3265 ctx_info->rcv_cq_num = iwqp->iwrcq->sc_cq.cq_uk.cq_id;
3266
3267 iwarp_info->ord_size = cm_node->ord_size;
3268 iwarp_info->ird_size = cm_node->ird_size;
3269 iwarp_info->rd_en = true;
3270 iwarp_info->rdmap_ver = 1;
3271 iwarp_info->ddp_ver = 1;
3272 iwarp_info->pd_id = iwqp->iwpd->sc_pd.pd_id;
3273
3274 ctx_info->tcp_info_valid = true;
3275 ctx_info->iwarp_info_valid = true;
3276 ctx_info->user_pri = cm_node->user_pri;
3277
3278 irdma_init_tcp_ctx(cm_node, &iwqp->tcp_info, iwqp);
3279 if (cm_node->snd_mark_en) {
3280 iwarp_info->snd_mark_en = true;
3281 iwarp_info->snd_mark_offset = (iwqp->tcp_info.snd_nxt & SNDMARKER_SEQNMASK) +
3282 cm_node->lsmm_size;
3283 }
3284
3285 cm_node->state = IRDMA_CM_STATE_OFFLOADED;
3286 iwqp->tcp_info.tcp_state = IRDMA_TCP_STATE_ESTABLISHED;
3287 iwqp->tcp_info.src_mac_addr_idx = iwqp->iwdev->mac_ip_table_idx;
3288
3289 if (cm_node->rcv_mark_en) {
3290 iwarp_info->rcv_mark_en = true;
3291 iwarp_info->align_hdrs = true;
3292 }
3293
3294 irdma_sc_qp_setctx(&iwqp->sc_qp, iwqp->host_ctx.va, ctx_info);
3295
3296 /* once tcp_info is set, no need to do it again */
3297 ctx_info->tcp_info_valid = false;
3298 ctx_info->iwarp_info_valid = false;
3299 }
3300
3301 /**
3302 * irdma_cm_disconn - when a connection is being closed
3303 * @iwqp: associated qp for the connection
3304 */
3305 void
irdma_cm_disconn(struct irdma_qp * iwqp)3306 irdma_cm_disconn(struct irdma_qp *iwqp)
3307 {
3308 struct irdma_device *iwdev = iwqp->iwdev;
3309 struct disconn_work *work;
3310 unsigned long flags;
3311
3312 work = kzalloc(sizeof(*work), GFP_ATOMIC);
3313 if (!work)
3314 return;
3315
3316 spin_lock_irqsave(&iwdev->rf->qptable_lock, flags);
3317 if (!iwdev->rf->qp_table[iwqp->ibqp.qp_num]) {
3318 spin_unlock_irqrestore(&iwdev->rf->qptable_lock, flags);
3319 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3320 "qp_id %d is already freed\n", iwqp->ibqp.qp_num);
3321 kfree(work);
3322 return;
3323 }
3324 irdma_qp_add_ref(&iwqp->ibqp);
3325 spin_unlock_irqrestore(&iwdev->rf->qptable_lock, flags);
3326
3327 work->iwqp = iwqp;
3328 INIT_WORK(&work->work, irdma_disconnect_worker);
3329 queue_work(iwdev->cleanup_wq, &work->work);
3330 }
3331
3332 /**
3333 * irdma_qp_disconnect - free qp and close cm
3334 * @iwqp: associate qp for the connection
3335 */
3336 static void
irdma_qp_disconnect(struct irdma_qp * iwqp)3337 irdma_qp_disconnect(struct irdma_qp *iwqp)
3338 {
3339 struct irdma_device *iwdev = iwqp->iwdev;
3340
3341 iwqp->active_conn = 0;
3342 /* close the CM node down if it is still active */
3343 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM, "Call close API\n");
3344 irdma_cm_close(iwqp->cm_node);
3345 }
3346
3347 /**
3348 * irdma_cm_disconn_true - called by worker thread to disconnect qp
3349 * @iwqp: associate qp for the connection
3350 */
3351 static void
irdma_cm_disconn_true(struct irdma_qp * iwqp)3352 irdma_cm_disconn_true(struct irdma_qp *iwqp)
3353 {
3354 struct iw_cm_id *cm_id;
3355 struct irdma_device *iwdev;
3356 struct irdma_sc_qp *qp = &iwqp->sc_qp;
3357 u16 last_ae;
3358 u8 original_hw_tcp_state;
3359 u8 original_ibqp_state;
3360 int disconn_status = 0;
3361 int issue_disconn = 0;
3362 int issue_close = 0;
3363 int issue_flush = 0;
3364 unsigned long flags;
3365 int err;
3366
3367 iwdev = iwqp->iwdev;
3368 spin_lock_irqsave(&iwqp->lock, flags);
3369 if (rdma_protocol_roce(&iwdev->ibdev, 1)) {
3370 struct ib_qp_attr attr;
3371
3372 if (iwqp->flush_issued || iwqp->sc_qp.qp_uk.destroy_pending) {
3373 spin_unlock_irqrestore(&iwqp->lock, flags);
3374 return;
3375 }
3376
3377 spin_unlock_irqrestore(&iwqp->lock, flags);
3378
3379 attr.qp_state = IB_QPS_ERR;
3380 irdma_modify_qp_roce(&iwqp->ibqp, &attr, IB_QP_STATE, NULL);
3381 irdma_ib_qp_event(iwqp, qp->event_type);
3382 return;
3383 }
3384
3385 cm_id = iwqp->cm_id;
3386 original_hw_tcp_state = iwqp->hw_tcp_state;
3387 original_ibqp_state = iwqp->ibqp_state;
3388 last_ae = iwqp->last_aeq;
3389
3390 if (qp->term_flags) {
3391 issue_disconn = 1;
3392 issue_close = 1;
3393 iwqp->cm_id = NULL;
3394 irdma_terminate_del_timer(qp);
3395 if (!iwqp->flush_issued) {
3396 iwqp->flush_issued = 1;
3397 issue_flush = 1;
3398 }
3399 } else if ((original_hw_tcp_state == IRDMA_TCP_STATE_CLOSE_WAIT) ||
3400 ((original_ibqp_state == IB_QPS_RTS) &&
3401 (last_ae == IRDMA_AE_LLP_CONNECTION_RESET))) {
3402 issue_disconn = 1;
3403 if (last_ae == IRDMA_AE_LLP_CONNECTION_RESET)
3404 disconn_status = -ECONNRESET;
3405 }
3406
3407 if (original_hw_tcp_state == IRDMA_TCP_STATE_CLOSED ||
3408 original_hw_tcp_state == IRDMA_TCP_STATE_TIME_WAIT ||
3409 last_ae == IRDMA_AE_RDMAP_ROE_BAD_LLP_CLOSE ||
3410 last_ae == IRDMA_AE_BAD_CLOSE ||
3411 last_ae == IRDMA_AE_LLP_CONNECTION_RESET || iwdev->rf->reset || !cm_id) {
3412 issue_close = 1;
3413 iwqp->cm_id = NULL;
3414 qp->term_flags = 0;
3415 if (!iwqp->flush_issued) {
3416 iwqp->flush_issued = 1;
3417 issue_flush = 1;
3418 }
3419 }
3420
3421 spin_unlock_irqrestore(&iwqp->lock, flags);
3422 if (issue_flush && !iwqp->sc_qp.qp_uk.destroy_pending) {
3423 irdma_flush_wqes(iwqp, IRDMA_FLUSH_SQ | IRDMA_FLUSH_RQ |
3424 IRDMA_FLUSH_WAIT);
3425
3426 if (qp->term_flags)
3427 irdma_ib_qp_event(iwqp, qp->event_type);
3428 }
3429
3430 if (!cm_id || !cm_id->event_handler)
3431 return;
3432
3433 spin_lock_irqsave(&iwdev->cm_core.ht_lock, flags);
3434 if (!iwqp->cm_node) {
3435 spin_unlock_irqrestore(&iwdev->cm_core.ht_lock, flags);
3436 return;
3437 }
3438 atomic_inc(&iwqp->cm_node->refcnt);
3439
3440 spin_unlock_irqrestore(&iwdev->cm_core.ht_lock, flags);
3441
3442 if (issue_disconn) {
3443 err = irdma_send_cm_event(iwqp->cm_node, cm_id,
3444 IW_CM_EVENT_DISCONNECT,
3445 disconn_status);
3446 if (err)
3447 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3448 "disconnect event failed: - cm_id = %p\n",
3449 cm_id);
3450 }
3451 if (issue_close) {
3452 cm_id->provider_data = iwqp;
3453 err = irdma_send_cm_event(iwqp->cm_node, cm_id,
3454 IW_CM_EVENT_CLOSE, 0);
3455 if (err)
3456 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3457 "close event failed: - cm_id = %p\n",
3458 cm_id);
3459 irdma_qp_disconnect(iwqp);
3460 }
3461 irdma_rem_ref_cm_node(iwqp->cm_node);
3462 }
3463
3464 /**
3465 * irdma_disconnect_worker - worker for connection close
3466 * @work: points or disconn structure
3467 */
3468 static void
irdma_disconnect_worker(struct work_struct * work)3469 irdma_disconnect_worker(struct work_struct *work)
3470 {
3471 struct disconn_work *dwork = container_of(work, struct disconn_work, work);
3472 struct irdma_qp *iwqp = dwork->iwqp;
3473
3474 kfree(dwork);
3475 irdma_cm_disconn_true(iwqp);
3476 irdma_qp_rem_ref(&iwqp->ibqp);
3477 }
3478
3479 /**
3480 * irdma_free_lsmm_rsrc - free lsmm memory and deregister
3481 * @iwqp: associate qp for the connection
3482 */
3483 void
irdma_free_lsmm_rsrc(struct irdma_qp * iwqp)3484 irdma_free_lsmm_rsrc(struct irdma_qp *iwqp)
3485 {
3486 struct irdma_device *iwdev;
3487
3488 iwdev = iwqp->iwdev;
3489
3490 if (iwqp->ietf_mem.va) {
3491 if (iwqp->lsmm_mr)
3492 iwdev->ibdev.dereg_mr(iwqp->lsmm_mr, NULL);
3493
3494 irdma_free_dma_mem(iwdev->rf->sc_dev.hw,
3495 &iwqp->ietf_mem);
3496 iwqp->ietf_mem.va = NULL;
3497 }
3498 }
3499
3500 /**
3501 * irdma_accept - registered call for connection to be accepted
3502 * @cm_id: cm information for passive connection
3503 * @conn_param: accpet parameters
3504 */
3505 int
irdma_accept(struct iw_cm_id * cm_id,struct iw_cm_conn_param * conn_param)3506 irdma_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param)
3507 {
3508 struct ib_qp *ibqp;
3509 struct irdma_qp *iwqp;
3510 struct irdma_device *iwdev;
3511 struct irdma_sc_dev *dev;
3512 struct irdma_cm_node *cm_node;
3513 struct ib_qp_attr attr = {0};
3514 int passive_state;
3515 struct ib_mr *ibmr;
3516 struct irdma_pd *iwpd;
3517 u16 buf_len = 0;
3518 struct irdma_kmem_info accept;
3519 u64 tagged_offset;
3520 int wait_ret;
3521 int ret = 0;
3522
3523 ibqp = irdma_get_qp(cm_id->device, conn_param->qpn);
3524 if (!ibqp)
3525 return -EINVAL;
3526
3527 iwqp = to_iwqp(ibqp);
3528 iwdev = iwqp->iwdev;
3529 dev = &iwdev->rf->sc_dev;
3530 cm_node = cm_id->provider_data;
3531
3532 if (((struct sockaddr_in *)&cm_id->local_addr)->sin_family == AF_INET) {
3533 cm_node->ipv4 = true;
3534 cm_node->vlan_id = irdma_get_vlan_ipv4(cm_id, cm_node->loc_addr);
3535 } else {
3536 cm_node->ipv4 = false;
3537 irdma_get_vlan_mac_ipv6(cm_id, cm_node->loc_addr, &cm_node->vlan_id,
3538 NULL);
3539 }
3540 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM, "Accept vlan_id=%d\n",
3541 cm_node->vlan_id);
3542
3543 if (cm_node->state == IRDMA_CM_STATE_LISTENER_DESTROYED) {
3544 ret = -EINVAL;
3545 goto error;
3546 }
3547
3548 passive_state = atomic_add_return(1, &cm_node->passive_state);
3549 if (passive_state == IRDMA_SEND_RESET_EVENT) {
3550 ret = -ECONNRESET;
3551 goto error;
3552 }
3553
3554 buf_len = conn_param->private_data_len + IRDMA_MAX_IETF_SIZE;
3555 iwqp->ietf_mem.size = buf_len;
3556 iwqp->ietf_mem.va = irdma_allocate_dma_mem(dev->hw, &iwqp->ietf_mem,
3557 iwqp->ietf_mem.size, 1);
3558 if (!iwqp->ietf_mem.va) {
3559 ret = -ENOMEM;
3560 goto error;
3561 }
3562
3563 cm_node->pdata.size = conn_param->private_data_len;
3564 accept.addr = iwqp->ietf_mem.va;
3565 accept.size = irdma_cm_build_mpa_frame(cm_node, &accept, MPA_KEY_REPLY);
3566 memcpy((u8 *)accept.addr + accept.size, conn_param->private_data,
3567 conn_param->private_data_len);
3568
3569 if (cm_node->dev->ws_add(iwqp->sc_qp.vsi, cm_node->user_pri)) {
3570 ret = -ENOMEM;
3571 goto error;
3572 }
3573 iwqp->sc_qp.user_pri = cm_node->user_pri;
3574 irdma_qp_add_qos(&iwqp->sc_qp);
3575 if (cm_node->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_2)
3576 iwdev->rf->check_fc(&iwdev->vsi, &iwqp->sc_qp);
3577 /* setup our first outgoing iWarp send WQE (the IETF frame response) */
3578 iwpd = iwqp->iwpd;
3579 tagged_offset = (uintptr_t)iwqp->ietf_mem.va;
3580 ibmr = irdma_reg_phys_mr(&iwpd->ibpd, iwqp->ietf_mem.pa, buf_len,
3581 IB_ACCESS_LOCAL_WRITE, &tagged_offset);
3582 if (IS_ERR(ibmr)) {
3583 ret = -ENOMEM;
3584 goto error;
3585 }
3586
3587 ibmr->pd = &iwpd->ibpd;
3588 ibmr->device = iwpd->ibpd.device;
3589 iwqp->lsmm_mr = ibmr;
3590 if (iwqp->page)
3591 iwqp->sc_qp.qp_uk.sq_base = kmap_local_page(iwqp->page);
3592
3593 cm_node->lsmm_size = accept.size + conn_param->private_data_len;
3594 irdma_sc_send_lsmm(&iwqp->sc_qp, iwqp->ietf_mem.va, cm_node->lsmm_size,
3595 ibmr->lkey);
3596
3597 if (iwqp->page)
3598 kunmap_local(iwqp->sc_qp.qp_uk.sq_base);
3599
3600 iwqp->cm_id = cm_id;
3601 cm_node->cm_id = cm_id;
3602
3603 cm_id->provider_data = iwqp;
3604 iwqp->active_conn = 0;
3605 iwqp->cm_node = cm_node;
3606 cm_node->iwqp = iwqp;
3607 irdma_cm_init_tsa_conn(iwqp, cm_node);
3608 irdma_qp_add_ref(&iwqp->ibqp);
3609 cm_id->add_ref(cm_id);
3610
3611 attr.qp_state = IB_QPS_RTS;
3612 cm_node->qhash_set = false;
3613 cm_node->cm_core->cm_free_ah(cm_node);
3614
3615 irdma_modify_qp(&iwqp->ibqp, &attr, IB_QP_STATE, NULL);
3616 if (dev->hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE) {
3617 wait_ret = wait_event_interruptible_timeout(iwqp->waitq,
3618 iwqp->rts_ae_rcvd,
3619 IRDMA_MAX_TIMEOUT);
3620 if (!wait_ret) {
3621 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3622 "Slow Connection: cm_node=%p, loc_port=%d, rem_port=%d, cm_id=%p\n",
3623 cm_node, cm_node->loc_port,
3624 cm_node->rem_port, cm_node->cm_id);
3625 ret = -ECONNRESET;
3626 goto error;
3627 }
3628 }
3629
3630 irdma_send_cm_event(cm_node, cm_id, IW_CM_EVENT_ESTABLISHED, 0);
3631 cm_node->accelerated = true;
3632 complete(&cm_node->establish_comp);
3633
3634 if (cm_node->accept_pend) {
3635 atomic_dec(&cm_node->listener->pend_accepts_cnt);
3636 cm_node->accept_pend = 0;
3637 }
3638
3639 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3640 "rem_port=0x%04x, loc_port=0x%04x rem_addr=%x loc_addr=%x cm_node=%p cm_id=%p qp_id=%d\n\n",
3641 cm_node->rem_port, cm_node->loc_port, cm_node->rem_addr[0],
3642 cm_node->loc_addr[0], cm_node, cm_id, ibqp->qp_num);
3643 cm_node->cm_core->stats_accepts++;
3644
3645 return 0;
3646 error:
3647 irdma_free_lsmm_rsrc(iwqp);
3648 irdma_rem_ref_cm_node(cm_node);
3649
3650 return ret;
3651 }
3652
3653 /**
3654 * irdma_reject - registered call for connection to be rejected
3655 * @cm_id: cm information for passive connection
3656 * @pdata: private data to be sent
3657 * @pdata_len: private data length
3658 */
3659 int
irdma_reject(struct iw_cm_id * cm_id,const void * pdata,u8 pdata_len)3660 irdma_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len)
3661 {
3662 struct irdma_device *iwdev;
3663 struct irdma_cm_node *cm_node;
3664
3665 cm_node = cm_id->provider_data;
3666 cm_node->pdata.size = pdata_len;
3667
3668 iwdev = to_iwdev(cm_id->device);
3669 if (!iwdev)
3670 return -EINVAL;
3671
3672 cm_node->cm_core->stats_rejects++;
3673
3674 if (pdata_len + sizeof(struct ietf_mpa_v2) > IRDMA_MAX_CM_BUF)
3675 return -EINVAL;
3676
3677 return irdma_cm_reject(cm_node, pdata, pdata_len);
3678 }
3679
3680 /**
3681 * irdma_connect - registered call for connection to be established
3682 * @cm_id: cm information for passive connection
3683 * @conn_param: Information about the connection
3684 */
3685 int
irdma_connect(struct iw_cm_id * cm_id,struct iw_cm_conn_param * conn_param)3686 irdma_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param)
3687 {
3688 struct ib_qp *ibqp;
3689 struct irdma_qp *iwqp;
3690 struct irdma_device *iwdev;
3691 struct irdma_cm_node *cm_node;
3692 struct irdma_cm_info cm_info;
3693 struct sockaddr_in *laddr;
3694 struct sockaddr_in *raddr;
3695 struct sockaddr_in6 *laddr6;
3696 struct sockaddr_in6 *raddr6;
3697 int ret = 0;
3698
3699 ibqp = irdma_get_qp(cm_id->device, conn_param->qpn);
3700 if (!ibqp)
3701 return -EINVAL;
3702 iwqp = to_iwqp(ibqp);
3703 if (!iwqp)
3704 return -EINVAL;
3705 iwdev = iwqp->iwdev;
3706 if (!iwdev)
3707 return -EINVAL;
3708
3709 laddr = (struct sockaddr_in *)&cm_id->m_local_addr;
3710 raddr = (struct sockaddr_in *)&cm_id->m_remote_addr;
3711 laddr6 = (struct sockaddr_in6 *)&cm_id->m_local_addr;
3712 raddr6 = (struct sockaddr_in6 *)&cm_id->m_remote_addr;
3713
3714 if (!(laddr->sin_port) || !(raddr->sin_port))
3715 return -EINVAL;
3716
3717 iwqp->active_conn = 1;
3718 iwqp->cm_id = NULL;
3719 cm_id->provider_data = iwqp;
3720
3721 /* set up the connection params for the node */
3722 if (cm_id->remote_addr.ss_family == AF_INET) {
3723 if (iwdev->vsi.mtu < IRDMA_MIN_MTU_IPV4)
3724 return -EINVAL;
3725
3726 cm_info.ipv4 = true;
3727 memset(cm_info.loc_addr, 0, sizeof(cm_info.loc_addr));
3728 memset(cm_info.rem_addr, 0, sizeof(cm_info.rem_addr));
3729 cm_info.loc_addr[0] = ntohl(laddr->sin_addr.s_addr);
3730 cm_info.rem_addr[0] = ntohl(raddr->sin_addr.s_addr);
3731 cm_info.loc_port = ntohs(laddr->sin_port);
3732 cm_info.rem_port = ntohs(raddr->sin_port);
3733 cm_info.vlan_id = irdma_get_vlan_ipv4(cm_id, cm_info.loc_addr);
3734 } else {
3735 if (iwdev->vsi.mtu < IRDMA_MIN_MTU_IPV6)
3736 return -EINVAL;
3737
3738 cm_info.ipv4 = false;
3739 irdma_copy_ip_ntohl(cm_info.loc_addr,
3740 laddr6->sin6_addr.__u6_addr.__u6_addr32);
3741 irdma_copy_ip_ntohl(cm_info.rem_addr,
3742 raddr6->sin6_addr.__u6_addr.__u6_addr32);
3743 cm_info.loc_port = ntohs(laddr6->sin6_port);
3744 cm_info.rem_port = ntohs(raddr6->sin6_port);
3745 irdma_get_vlan_mac_ipv6(cm_id, cm_info.loc_addr, &cm_info.vlan_id, NULL);
3746 }
3747 cm_info.cm_id = cm_id;
3748 cm_info.qh_qpid = iwdev->vsi.ilq->qp_id;
3749 cm_info.tos = cm_id->tos;
3750 if (iwdev->vsi.dscp_mode) {
3751 cm_info.user_pri =
3752 iwqp->sc_qp.vsi->dscp_map[irdma_tos2dscp(cm_info.tos)];
3753 } else {
3754 cm_info.user_pri = rt_tos2priority(cm_id->tos);
3755 cm_info.user_pri =
3756 irdma_iw_get_vlan_prio(cm_info.loc_addr,
3757 cm_info.user_pri,
3758 cm_info.ipv4);
3759 }
3760
3761 if (iwqp->sc_qp.dev->ws_add(iwqp->sc_qp.vsi, cm_info.user_pri))
3762 return -ENOMEM;
3763 iwqp->sc_qp.user_pri = cm_info.user_pri;
3764 irdma_qp_add_qos(&iwqp->sc_qp);
3765 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_2)
3766 iwdev->rf->check_fc(&iwdev->vsi, &iwqp->sc_qp);
3767 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_DCB, "TOS:[%d] UP:[%d]\n",
3768 cm_id->tos, cm_info.user_pri);
3769
3770 ret = irdma_create_cm_node(&iwdev->cm_core, iwdev, conn_param, &cm_info,
3771 &cm_node);
3772 if (ret)
3773 return ret;
3774 ret = cm_node->cm_core->cm_create_ah(cm_node, true);
3775 if (ret)
3776 goto err;
3777 if (irdma_manage_qhash(iwdev, &cm_info,
3778 IRDMA_QHASH_TYPE_TCP_ESTABLISHED,
3779 IRDMA_QHASH_MANAGE_TYPE_ADD, NULL, true)) {
3780 ret = -EINVAL;
3781 goto err;
3782 }
3783 cm_node->qhash_set = true;
3784
3785 cm_node->apbvt_entry = irdma_add_apbvt(iwdev, cm_info.loc_port);
3786 if (!cm_node->apbvt_entry) {
3787 ret = -EINVAL;
3788 goto err;
3789 }
3790
3791 cm_node->apbvt_set = true;
3792 iwqp->cm_node = cm_node;
3793 cm_node->iwqp = iwqp;
3794 iwqp->cm_id = cm_id;
3795 irdma_qp_add_ref(&iwqp->ibqp);
3796 cm_id->add_ref(cm_id);
3797
3798 if (cm_node->state != IRDMA_CM_STATE_OFFLOADED) {
3799 cm_node->state = IRDMA_CM_STATE_SYN_SENT;
3800 ret = irdma_send_syn(cm_node, 0);
3801 if (ret)
3802 goto err;
3803 }
3804
3805 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3806 "rem_port=0x%04x, loc_port=0x%04x rem_addr=%x loc_addr=%x cm_node=%p cm_id=%p qp_id = %d\n\n",
3807 cm_node->rem_port, cm_node->loc_port, cm_node->rem_addr[0],
3808 cm_node->loc_addr[0], cm_node, cm_id, ibqp->qp_num);
3809
3810 return 0;
3811
3812 err:
3813 if (cm_info.ipv4)
3814 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3815 "connect() FAILED: dest addr=%x",
3816 cm_info.rem_addr[0]);
3817 else
3818 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3819 "connect() FAILED: dest addr=%x:%x:%x:%x",
3820 IRDMA_PRINT_IP6(cm_info.rem_addr));
3821 irdma_rem_ref_cm_node(cm_node);
3822 iwdev->cm_core.stats_connect_errs++;
3823
3824 return ret;
3825 }
3826
3827 /**
3828 * irdma_create_listen - registered call creating listener
3829 * @cm_id: cm information for passive connection
3830 * @backlog: to max accept pending count
3831 */
3832 int
irdma_create_listen(struct iw_cm_id * cm_id,int backlog)3833 irdma_create_listen(struct iw_cm_id *cm_id, int backlog)
3834 {
3835 struct irdma_device *iwdev;
3836 struct irdma_cm_listener *cm_listen_node;
3837 struct irdma_cm_info cm_info = {0};
3838 struct sockaddr_in *laddr;
3839 struct sockaddr_in6 *laddr6;
3840 bool wildcard = false;
3841 int err;
3842
3843 iwdev = to_iwdev(cm_id->device);
3844 if (!iwdev)
3845 return -EINVAL;
3846
3847 laddr = (struct sockaddr_in *)&cm_id->m_local_addr;
3848 laddr6 = (struct sockaddr_in6 *)&cm_id->m_local_addr;
3849 cm_info.qh_qpid = iwdev->vsi.ilq->qp_id;
3850
3851 if (laddr->sin_family == AF_INET) {
3852 if (iwdev->vsi.mtu < IRDMA_MIN_MTU_IPV4)
3853 return -EINVAL;
3854
3855 cm_info.ipv4 = true;
3856 cm_info.loc_addr[0] = ntohl(laddr->sin_addr.s_addr);
3857 cm_info.loc_port = ntohs(laddr->sin_port);
3858
3859 if (laddr->sin_addr.s_addr != htonl(INADDR_ANY)) {
3860 cm_info.vlan_id = irdma_get_vlan_ipv4(cm_id, cm_info.loc_addr);
3861 } else {
3862 cm_info.vlan_id = 0xFFFF;
3863 wildcard = true;
3864 }
3865 } else {
3866 if (iwdev->vsi.mtu < IRDMA_MIN_MTU_IPV6)
3867 return -EINVAL;
3868
3869 cm_info.ipv4 = false;
3870 irdma_copy_ip_ntohl(cm_info.loc_addr,
3871 laddr6->sin6_addr.__u6_addr.__u6_addr32);
3872 cm_info.loc_port = ntohs(laddr6->sin6_port);
3873 if (!IN6_IS_ADDR_UNSPECIFIED(&laddr6->sin6_addr)) {
3874 irdma_get_vlan_mac_ipv6(cm_id, cm_info.loc_addr,
3875 &cm_info.vlan_id, NULL);
3876 } else {
3877 cm_info.vlan_id = 0xFFFF;
3878 wildcard = true;
3879 }
3880 }
3881
3882 if (cm_info.vlan_id >= VLAN_N_VID && iwdev->dcb_vlan_mode)
3883 cm_info.vlan_id = 0;
3884 cm_info.backlog = backlog;
3885 cm_info.cm_id = cm_id;
3886
3887 cm_listen_node = irdma_make_listen_node(&iwdev->cm_core, iwdev,
3888 &cm_info);
3889 if (!cm_listen_node) {
3890 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3891 "cm_listen_node == NULL\n");
3892 return -ENOMEM;
3893 }
3894
3895 cm_id->provider_data = cm_listen_node;
3896
3897 cm_listen_node->tos = cm_id->tos;
3898 if (iwdev->vsi.dscp_mode)
3899 cm_listen_node->user_pri =
3900 iwdev->vsi.dscp_map[irdma_tos2dscp(cm_id->tos)];
3901 else
3902 cm_listen_node->user_pri = rt_tos2priority(cm_id->tos);
3903 cm_info.user_pri = cm_listen_node->user_pri;
3904 if (!cm_listen_node->reused_node) {
3905 if (wildcard) {
3906 err = irdma_add_mqh(iwdev, &cm_info, cm_listen_node);
3907 if (err)
3908 goto error;
3909 } else {
3910 if (!iwdev->vsi.dscp_mode)
3911 cm_info.user_pri = cm_listen_node->user_pri =
3912 irdma_iw_get_vlan_prio(cm_info.loc_addr,
3913 cm_info.user_pri,
3914 cm_info.ipv4);
3915 err = irdma_manage_qhash(iwdev, &cm_info,
3916 IRDMA_QHASH_TYPE_TCP_SYN,
3917 IRDMA_QHASH_MANAGE_TYPE_ADD,
3918 NULL, true);
3919 if (err)
3920 goto error;
3921
3922 cm_listen_node->qhash_set = true;
3923 }
3924
3925 cm_listen_node->apbvt_entry = irdma_add_apbvt(iwdev,
3926 cm_info.loc_port);
3927 if (!cm_listen_node->apbvt_entry)
3928 goto error;
3929 }
3930 cm_id->add_ref(cm_id);
3931 cm_listen_node->cm_core->stats_listen_created++;
3932 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3933 "loc_port=0x%04x loc_addr=%x cm_listen_node=%p cm_id=%p qhash_set=%d vlan_id=%d\n",
3934 cm_listen_node->loc_port, cm_listen_node->loc_addr[0],
3935 cm_listen_node, cm_listen_node->cm_id,
3936 cm_listen_node->qhash_set, cm_listen_node->vlan_id);
3937
3938 return 0;
3939
3940 error:
3941
3942 irdma_cm_del_listen(&iwdev->cm_core, cm_listen_node, false);
3943
3944 return -EINVAL;
3945 }
3946
3947 /**
3948 * irdma_destroy_listen - registered call to destroy listener
3949 * @cm_id: cm information for passive connection
3950 */
3951 int
irdma_destroy_listen(struct iw_cm_id * cm_id)3952 irdma_destroy_listen(struct iw_cm_id *cm_id)
3953 {
3954 struct irdma_device *iwdev;
3955
3956 iwdev = to_iwdev(cm_id->device);
3957 if (cm_id->provider_data)
3958 irdma_cm_del_listen(&iwdev->cm_core, cm_id->provider_data,
3959 true);
3960 else
3961 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
3962 "cm_id->provider_data was NULL\n");
3963
3964 cm_id->rem_ref(cm_id);
3965
3966 return 0;
3967 }
3968
3969 /**
3970 * irdma_iw_teardown_list_prep - add conn nodes slated for tear
3971 * down to list
3972 * @cm_core: cm's core
3973 * @teardown_list: a list to which cm_node will be selected
3974 * @ipaddr: pointer to ip address
3975 * @nfo: pointer to cm_info structure instance
3976 * @disconnect_all: flag indicating disconnect all QPs
3977 */
3978 static void
irdma_iw_teardown_list_prep(struct irdma_cm_core * cm_core,struct list_head * teardown_list,u32 * ipaddr,struct irdma_cm_info * nfo,bool disconnect_all)3979 irdma_iw_teardown_list_prep(struct irdma_cm_core *cm_core,
3980 struct list_head *teardown_list,
3981 u32 *ipaddr,
3982 struct irdma_cm_info *nfo,
3983 bool disconnect_all)
3984 {
3985 struct irdma_cm_node *cm_node;
3986 int bkt;
3987
3988 HASH_FOR_EACH_RCU(cm_core->cm_hash_tbl, bkt, cm_node, list) {
3989 if ((disconnect_all ||
3990 (nfo->vlan_id == cm_node->vlan_id &&
3991 !memcmp(cm_node->loc_addr, ipaddr, nfo->ipv4 ? 4 : 16))) &&
3992 atomic_inc_not_zero(&cm_node->refcnt))
3993 list_add(&cm_node->teardown_entry, teardown_list);
3994 }
3995 }
3996
3997 static inline bool
irdma_ip_vlan_match(u32 * ip1,u16 vlan_id1,bool check_vlan,u32 * ip2,u16 vlan_id2,bool ipv4)3998 irdma_ip_vlan_match(u32 *ip1, u16 vlan_id1,
3999 bool check_vlan, u32 *ip2,
4000 u16 vlan_id2, bool ipv4)
4001 {
4002 return (!check_vlan || vlan_id1 == vlan_id2) &&
4003 !memcmp(ip1, ip2, ipv4 ? 4 : 16);
4004 }
4005
4006 /**
4007 * irdma_roce_teardown_list_prep - add conn nodes slated for
4008 * tear down to list
4009 * @iwdev: RDMA device
4010 * @teardown_list: a list to which cm_node will be selected
4011 * @ipaddr: pointer to ip address
4012 * @nfo: pointer to cm_info structure instance
4013 * @disconnect_all: flag indicating disconnect all QPs
4014 */
4015 static void
irdma_roce_teardown_list_prep(struct irdma_device * iwdev,struct list_head * teardown_list,u32 * ipaddr,struct irdma_cm_info * nfo,bool disconnect_all)4016 irdma_roce_teardown_list_prep(struct irdma_device *iwdev,
4017 struct list_head *teardown_list,
4018 u32 *ipaddr,
4019 struct irdma_cm_info *nfo,
4020 bool disconnect_all)
4021 {
4022 struct irdma_sc_vsi *vsi = &iwdev->vsi;
4023 struct irdma_sc_qp *sc_qp;
4024 struct list_head *list_node;
4025 struct irdma_qp *qp;
4026 unsigned long flags;
4027 int i;
4028
4029 for (i = 0; i < IRDMA_MAX_USER_PRIORITY; i++) {
4030 mutex_lock(&vsi->qos[i].qos_mutex);
4031 list_for_each(list_node, &vsi->qos[i].qplist) {
4032 u32 qp_ip[4];
4033
4034 sc_qp = container_of(list_node, struct irdma_sc_qp,
4035 list);
4036 if (sc_qp->qp_uk.qp_type != IRDMA_QP_TYPE_ROCE_RC)
4037 continue;
4038
4039 qp = sc_qp->qp_uk.back_qp;
4040 if (!disconnect_all) {
4041 if (nfo->ipv4)
4042 qp_ip[0] = qp->udp_info.local_ipaddr[3];
4043 else
4044 memcpy(qp_ip,
4045 &qp->udp_info.local_ipaddr[0],
4046 sizeof(qp_ip));
4047 }
4048
4049 if (disconnect_all ||
4050 irdma_ip_vlan_match(qp_ip,
4051 qp->udp_info.vlan_tag & EVL_VLID_MASK,
4052 qp->udp_info.insert_vlan_tag,
4053 ipaddr, nfo->vlan_id, nfo->ipv4)) {
4054 spin_lock_irqsave(&iwdev->rf->qptable_lock, flags);
4055 if (iwdev->rf->qp_table[sc_qp->qp_uk.qp_id]) {
4056 irdma_qp_add_ref(&qp->ibqp);
4057 list_add(&qp->teardown_entry, teardown_list);
4058 }
4059 spin_unlock_irqrestore(&iwdev->rf->qptable_lock, flags);
4060 }
4061 }
4062 mutex_unlock(&vsi->qos[i].qos_mutex);
4063 }
4064 }
4065
4066 /**
4067 * irdma_cm_event_connected - handle connected active node
4068 * @event: the info for cm_node of connection
4069 */
4070 static void
irdma_cm_event_connected(struct irdma_cm_event * event)4071 irdma_cm_event_connected(struct irdma_cm_event *event)
4072 {
4073 struct irdma_qp *iwqp;
4074 struct irdma_device *iwdev;
4075 struct irdma_cm_node *cm_node;
4076 struct irdma_sc_dev *dev;
4077 struct ib_qp_attr attr = {0};
4078 struct iw_cm_id *cm_id;
4079 int status;
4080 bool read0;
4081 int wait_ret = 0;
4082
4083 cm_node = event->cm_node;
4084 cm_id = cm_node->cm_id;
4085 iwqp = cm_id->provider_data;
4086 iwdev = iwqp->iwdev;
4087 dev = &iwdev->rf->sc_dev;
4088 if (iwqp->sc_qp.qp_uk.destroy_pending) {
4089 status = -ETIMEDOUT;
4090 goto error;
4091 }
4092
4093 irdma_cm_init_tsa_conn(iwqp, cm_node);
4094 read0 = (cm_node->send_rdma0_op == SEND_RDMA_READ_ZERO);
4095 if (iwqp->page)
4096 iwqp->sc_qp.qp_uk.sq_base = kmap_local_page(iwqp->page);
4097 irdma_sc_send_rtt(&iwqp->sc_qp, read0);
4098 if (iwqp->page)
4099 kunmap_local(iwqp->sc_qp.qp_uk.sq_base);
4100
4101 attr.qp_state = IB_QPS_RTS;
4102 cm_node->qhash_set = false;
4103 irdma_modify_qp(&iwqp->ibqp, &attr, IB_QP_STATE, NULL);
4104 if (dev->hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE) {
4105 wait_ret = wait_event_interruptible_timeout(iwqp->waitq,
4106 iwqp->rts_ae_rcvd,
4107 IRDMA_MAX_TIMEOUT);
4108 if (!wait_ret)
4109 irdma_debug(&iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
4110 "Slow Connection: cm_node=%p, loc_port=%d, rem_port=%d, cm_id=%p\n",
4111 cm_node, cm_node->loc_port,
4112 cm_node->rem_port, cm_node->cm_id);
4113 }
4114
4115 irdma_send_cm_event(cm_node, cm_id, IW_CM_EVENT_CONNECT_REPLY, 0);
4116 cm_node->accelerated = true;
4117 complete(&cm_node->establish_comp);
4118 cm_node->cm_core->cm_free_ah(cm_node);
4119 return;
4120
4121 error:
4122 iwqp->cm_id = NULL;
4123 cm_id->provider_data = NULL;
4124 irdma_send_cm_event(event->cm_node, cm_id, IW_CM_EVENT_CONNECT_REPLY,
4125 status);
4126 irdma_rem_ref_cm_node(event->cm_node);
4127 }
4128
4129 /**
4130 * irdma_cm_event_reset - handle reset
4131 * @event: the info for cm_node of connection
4132 */
4133 static void
irdma_cm_event_reset(struct irdma_cm_event * event)4134 irdma_cm_event_reset(struct irdma_cm_event *event)
4135 {
4136 struct irdma_cm_node *cm_node = event->cm_node;
4137 struct iw_cm_id *cm_id = cm_node->cm_id;
4138 struct irdma_qp *iwqp;
4139
4140 if (!cm_id)
4141 return;
4142
4143 iwqp = cm_id->provider_data;
4144 if (!iwqp)
4145 return;
4146
4147 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
4148 "reset event %p - cm_id = %p\n", event->cm_node, cm_id);
4149 iwqp->cm_id = NULL;
4150
4151 irdma_send_cm_event(cm_node, cm_node->cm_id, IW_CM_EVENT_DISCONNECT,
4152 -ECONNRESET);
4153 irdma_send_cm_event(cm_node, cm_node->cm_id, IW_CM_EVENT_CLOSE, 0);
4154 }
4155
4156 /**
4157 * irdma_cm_event_handler - send event to cm upper layer
4158 * @work: pointer of cm event info.
4159 */
4160 static void
irdma_cm_event_handler(struct work_struct * work)4161 irdma_cm_event_handler(struct work_struct *work)
4162 {
4163 struct irdma_cm_event *event = container_of(work, struct irdma_cm_event, event_work);
4164 struct irdma_cm_node *cm_node;
4165
4166 if (!event || !event->cm_node || !event->cm_node->cm_core)
4167 return;
4168
4169 cm_node = event->cm_node;
4170
4171 switch (event->type) {
4172 case IRDMA_CM_EVENT_MPA_REQ:
4173 irdma_send_cm_event(cm_node, cm_node->cm_id,
4174 IW_CM_EVENT_CONNECT_REQUEST, 0);
4175 break;
4176 case IRDMA_CM_EVENT_RESET:
4177 irdma_cm_event_reset(event);
4178 break;
4179 case IRDMA_CM_EVENT_CONNECTED:
4180 if (!event->cm_node->cm_id ||
4181 event->cm_node->state != IRDMA_CM_STATE_OFFLOADED)
4182 break;
4183 irdma_cm_event_connected(event);
4184 break;
4185 case IRDMA_CM_EVENT_MPA_REJECT:
4186 if (!event->cm_node->cm_id ||
4187 cm_node->state == IRDMA_CM_STATE_OFFLOADED)
4188 break;
4189 irdma_send_cm_event(cm_node, cm_node->cm_id,
4190 IW_CM_EVENT_CONNECT_REPLY, -ECONNREFUSED);
4191 break;
4192 case IRDMA_CM_EVENT_ABORTED:
4193 if (!event->cm_node->cm_id ||
4194 event->cm_node->state == IRDMA_CM_STATE_OFFLOADED)
4195 break;
4196 irdma_event_connect_error(event);
4197 break;
4198 default:
4199 irdma_debug(&cm_node->iwdev->rf->sc_dev, IRDMA_DEBUG_CM,
4200 "bad event type = %d\n", event->type);
4201 break;
4202 }
4203
4204 irdma_rem_ref_cm_node(event->cm_node);
4205 kfree(event);
4206 }
4207
4208 /**
4209 * irdma_cm_post_event - queue event request for worker thread
4210 * @event: cm node's info for up event call
4211 */
4212 static void
irdma_cm_post_event(struct irdma_cm_event * event)4213 irdma_cm_post_event(struct irdma_cm_event *event)
4214 {
4215 atomic_inc(&event->cm_node->refcnt);
4216 INIT_WORK(&event->event_work, irdma_cm_event_handler);
4217 queue_work(event->cm_node->cm_core->event_wq, &event->event_work);
4218 }
4219
4220 /**
4221 * irdma_cm_teardown_connections - teardown QPs
4222 * @iwdev: device pointer
4223 * @ipaddr: Pointer to IPv4 or IPv6 address
4224 * @nfo: Connection info
4225 * @disconnect_all: flag indicating disconnect all QPs
4226 *
4227 * teardown QPs where source or destination addr matches ip addr
4228 */
4229 static void __unused
irdma_cm_teardown_connections(struct irdma_device * iwdev,u32 * ipaddr,struct irdma_cm_info * nfo,bool disconnect_all)4230 irdma_cm_teardown_connections(struct irdma_device *iwdev,
4231 u32 *ipaddr,
4232 struct irdma_cm_info *nfo,
4233 bool disconnect_all)
4234 {
4235 struct irdma_cm_core *cm_core = &iwdev->cm_core;
4236 struct list_head *list_core_temp;
4237 struct list_head *list_node;
4238 struct irdma_cm_node *cm_node;
4239 struct list_head teardown_list;
4240 struct ib_qp_attr attr;
4241 struct irdma_qp *qp;
4242
4243 INIT_LIST_HEAD(&teardown_list);
4244
4245 rcu_read_lock();
4246 irdma_iw_teardown_list_prep(cm_core, &teardown_list, ipaddr, nfo, disconnect_all);
4247 rcu_read_unlock();
4248
4249 attr.qp_state = IB_QPS_ERR;
4250 list_for_each_safe(list_node, list_core_temp, &teardown_list) {
4251 cm_node = container_of(list_node, struct irdma_cm_node,
4252 teardown_entry);
4253 irdma_modify_qp(&cm_node->iwqp->ibqp, &attr, IB_QP_STATE, NULL);
4254 if (iwdev->rf->reset)
4255 irdma_cm_disconn(cm_node->iwqp);
4256 irdma_rem_ref_cm_node(cm_node);
4257 }
4258
4259 if (!rdma_protocol_roce(&iwdev->ibdev, 1))
4260 return;
4261
4262 INIT_LIST_HEAD(&teardown_list);
4263 irdma_roce_teardown_list_prep(iwdev, &teardown_list, ipaddr, nfo, disconnect_all);
4264
4265 list_for_each_safe(list_node, list_core_temp, &teardown_list) {
4266 qp = container_of(list_node, struct irdma_qp, teardown_entry);
4267 irdma_modify_qp_roce(&qp->ibqp, &attr, IB_QP_STATE, NULL);
4268 irdma_ib_qp_event(qp, IRDMA_QP_EVENT_CATASTROPHIC);
4269 irdma_qp_rem_ref(&qp->ibqp);
4270 }
4271 }
4272