xref: /freebsd/sys/dev/qlnx/qlnxr/qlnxr_os.c (revision 685dc743dc3b5645e34836464128e1c0558b404b)
1 /*
2  * Copyright (c) 2018-2019 Cavium, Inc.
3  * All rights reserved.
4  *
5  *  Redistribution and use in source and binary forms, with or without
6  *  modification, are permitted provided that the following conditions
7  *  are met:
8  *
9  *  1. Redistributions of source code must retain the above copyright
10  *     notice, this list of conditions and the following disclaimer.
11  *  2. Redistributions in binary form must reproduce the above copyright
12  *     notice, this list of conditions and the following disclaimer in the
13  *     documentation and/or other materials provided with the distribution.
14  *
15  *  THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
16  *  AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17  *  IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18  *  ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
19  *  LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
20  *  CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
21  *  SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
22  *  INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
23  *  CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
24  *  ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
25  *  POSSIBILITY OF SUCH DAMAGE.
26  */
27 
28 /*
29  * File: qlnxr_os.c
30  */
31 #include <sys/cdefs.h>
32 #include "qlnxr_def.h"
33 
34 SYSCTL_NODE(_dev, OID_AUTO, qnxr, CTLFLAG_RW | CTLFLAG_MPSAFE, 0,
35     "Qlogic RDMA module");
36 
37 uint32_t delayed_ack = 0;
38 SYSCTL_UINT(_dev_qnxr, OID_AUTO, delayed_ack, CTLFLAG_RW, &delayed_ack, 1,
39 	"iWARP: Delayed Ack: 0 - Disabled 1 - Enabled. Default: Disabled");
40 
41 uint32_t timestamp = 1;
42 SYSCTL_UINT(_dev_qnxr, OID_AUTO, timestamp, CTLFLAG_RW, &timestamp, 1,
43 	"iWARP: Timestamp: 0 - Disabled 1 - Enabled. Default:Enabled");
44 
45 uint32_t rcv_wnd_size = 0;
46 SYSCTL_UINT(_dev_qnxr, OID_AUTO, rcv_wnd_size, CTLFLAG_RW, &rcv_wnd_size, 1,
47 	"iWARP: Receive Window Size in K. Default 1M");
48 
49 uint32_t crc_needed = 1;
50 SYSCTL_UINT(_dev_qnxr, OID_AUTO, crc_needed, CTLFLAG_RW, &crc_needed, 1,
51 	"iWARP: CRC needed 0 - Disabled 1 - Enabled. Default:Enabled");
52 
53 uint32_t peer2peer = 1;
54 SYSCTL_UINT(_dev_qnxr, OID_AUTO, peer2peer, CTLFLAG_RW, &peer2peer, 1,
55 	"iWARP: Support peer2peer ULPs 0 - Disabled 1 - Enabled. Default:Enabled");
56 
57 uint32_t mpa_enhanced = 1;
58 SYSCTL_UINT(_dev_qnxr, OID_AUTO, mpa_enhanced, CTLFLAG_RW, &mpa_enhanced, 1,
59 	"iWARP: MPA Enhanced mode. Default:1");
60 
61 uint32_t rtr_type = 7;
62 SYSCTL_UINT(_dev_qnxr, OID_AUTO, rtr_type, CTLFLAG_RW, &rtr_type, 1,
63 	"iWARP: RDMAP opcode to use for the RTR message: BITMAP 1: RDMA_SEND 2: RDMA_WRITE 4: RDMA_READ. Default: 7");
64 
65 #define QNXR_WQ_MULTIPLIER_MIN  (1)
66 #define QNXR_WQ_MULTIPLIER_MAX  (7)
67 #define QNXR_WQ_MULTIPLIER_DFT  (3)
68 
69 uint32_t wq_multiplier= QNXR_WQ_MULTIPLIER_DFT;
70 SYSCTL_UINT(_dev_qnxr, OID_AUTO, wq_multiplier, CTLFLAG_RW, &wq_multiplier, 1,
71 	" When creating a WQ the actual number of WQE created will"
72 	" be multiplied by this number (default is 3).");
73 static ssize_t
show_rev(struct device * device,struct device_attribute * attr,char * buf)74 show_rev(struct device *device, struct device_attribute *attr,
75 	char *buf)
76 {
77         struct qlnxr_dev *dev = dev_get_drvdata(device);
78 
79         return sprintf(buf, "0x%x\n", dev->cdev->vendor_id);
80 }
81 
82 static ssize_t
show_hca_type(struct device * device,struct device_attribute * attr,char * buf)83 show_hca_type(struct device *device,
84 	struct device_attribute *attr, char *buf)
85 {
86 	struct qlnxr_dev *dev = dev_get_drvdata(device);
87         return sprintf(buf, "QLogic0x%x\n", dev->cdev->device_id);
88 }
89 
90 static ssize_t
show_fw_ver(struct device * device,struct device_attribute * attr,char * buf)91 show_fw_ver(struct device *device,
92 	struct device_attribute *attr, char *buf)
93 {
94 	struct qlnxr_dev *dev = dev_get_drvdata(device);
95 	uint32_t fw_ver = (uint32_t) dev->attr.fw_ver;
96 
97 	return sprintf(buf, "%d.%d.%d\n",
98 		       (fw_ver >> 24) & 0xff, (fw_ver >> 16) & 0xff,
99 		       (fw_ver >> 8) & 0xff);
100 }
101 static ssize_t
show_board(struct device * device,struct device_attribute * attr,char * buf)102 show_board(struct device *device,
103 	struct device_attribute *attr, char *buf)
104 {
105 	struct qlnxr_dev *dev = dev_get_drvdata(device);
106 	return sprintf(buf, "%x\n", dev->cdev->device_id);
107 }
108 
109 static DEVICE_ATTR(hw_rev, S_IRUGO, show_rev, NULL);
110 static DEVICE_ATTR(hca_type, S_IRUGO, show_hca_type, NULL);
111 static DEVICE_ATTR(fw_ver, S_IRUGO, show_fw_ver, NULL);
112 static DEVICE_ATTR(board_id, S_IRUGO, show_board, NULL);
113 
114 static struct device_attribute *qlnxr_class_attributes[] = {
115 	&dev_attr_hw_rev,
116 	&dev_attr_hca_type,
117 	&dev_attr_fw_ver,
118 	&dev_attr_board_id
119 };
120 
121 static void
qlnxr_ib_dispatch_event(qlnxr_dev_t * dev,uint8_t port_num,enum ib_event_type type)122 qlnxr_ib_dispatch_event(qlnxr_dev_t *dev, uint8_t port_num,
123 	enum ib_event_type type)
124 {
125         struct ib_event ibev;
126 
127 	QL_DPRINT12(dev->ha, "enter\n");
128 
129         ibev.device = &dev->ibdev;
130         ibev.element.port_num = port_num;
131         ibev.event = type;
132 
133         ib_dispatch_event(&ibev);
134 
135 	QL_DPRINT12(dev->ha, "exit\n");
136 }
137 
138 static int
__qlnxr_iw_destroy_listen(struct iw_cm_id * cm_id)139 __qlnxr_iw_destroy_listen(struct iw_cm_id *cm_id)
140 {
141 	qlnxr_iw_destroy_listen(cm_id);
142 
143 	return (0);
144 }
145 
146 static int
qlnxr_register_device(qlnxr_dev_t * dev)147 qlnxr_register_device(qlnxr_dev_t *dev)
148 {
149 	struct ib_device *ibdev;
150 	struct iw_cm_verbs *iwcm;
151 	int ret;
152 
153 	QL_DPRINT12(dev->ha, "enter\n");
154 
155 	ibdev = &dev->ibdev;
156 
157 #define qlnxr_ib_ah qlnxr_ah
158 #define qlnxr_ib_cq qlnxr_cq
159 #define qlnxr_ib_pd qlnxr_pd
160 #define qlnxr_ib_qp qlnxr_qp
161 #define qlnxr_ib_srq qlnxr_srq
162 #define qlnxr_ib_ucontext qlnxr_ucontext
163 	INIT_IB_DEVICE_OPS(&ibdev->ops, qlnxr, QLNXR);
164 	strlcpy(ibdev->name, "qlnxr%d", IB_DEVICE_NAME_MAX);
165 
166 	memset(&ibdev->node_guid, 0, sizeof(ibdev->node_guid));
167 	memcpy(&ibdev->node_guid, dev->ha->primary_mac, ETHER_ADDR_LEN);
168 
169 	memcpy(ibdev->node_desc, QLNXR_NODE_DESC, sizeof(QLNXR_NODE_DESC));
170 
171 	ibdev->owner = THIS_MODULE;
172 	ibdev->uverbs_abi_ver = 7;
173 	ibdev->local_dma_lkey = 0;
174 
175 	ibdev->uverbs_cmd_mask =
176 		(1ull << IB_USER_VERBS_CMD_GET_CONTEXT) |
177 		(1ull << IB_USER_VERBS_CMD_QUERY_DEVICE) |
178 		(1ull << IB_USER_VERBS_CMD_QUERY_PORT) |
179 		(1ull << IB_USER_VERBS_CMD_ALLOC_PD) |
180 		(1ull << IB_USER_VERBS_CMD_DEALLOC_PD) |
181 		(1ull << IB_USER_VERBS_CMD_REG_MR) |
182 		(1ull << IB_USER_VERBS_CMD_DEREG_MR) |
183 		(1ull << IB_USER_VERBS_CMD_CREATE_COMP_CHANNEL) |
184 		(1ull << IB_USER_VERBS_CMD_CREATE_CQ) |
185 		(1ull << IB_USER_VERBS_CMD_DESTROY_CQ) |
186 		(1ull << IB_USER_VERBS_CMD_REQ_NOTIFY_CQ) |
187 		(1ull << IB_USER_VERBS_CMD_CREATE_QP) |
188 		(1ull << IB_USER_VERBS_CMD_MODIFY_QP) |
189 		(1ull << IB_USER_VERBS_CMD_QUERY_QP) |
190 		(1ull << IB_USER_VERBS_CMD_DESTROY_QP) |
191 		(1ull << IB_USER_VERBS_CMD_POLL_CQ) |
192 		(1ull << IB_USER_VERBS_CMD_POST_SEND) |
193 		(1ull << IB_USER_VERBS_CMD_POST_RECV);
194 
195         if (QLNX_IS_IWARP(dev)) {
196                 ibdev->node_type = RDMA_NODE_RNIC;
197                 ibdev->query_gid = qlnxr_iw_query_gid;
198         } else {
199                 ibdev->node_type = RDMA_NODE_IB_CA;
200                 ibdev->query_gid = qlnxr_query_gid;
201                 ibdev->uverbs_cmd_mask |=
202 			(1ull << IB_USER_VERBS_CMD_CREATE_SRQ) |
203 			(1ull << IB_USER_VERBS_CMD_DESTROY_SRQ) |
204 			(1ull << IB_USER_VERBS_CMD_QUERY_SRQ) |
205 			(1ull << IB_USER_VERBS_CMD_MODIFY_SRQ) |
206 			(1ull << IB_USER_VERBS_CMD_POST_SRQ_RECV);
207                 ibdev->create_srq = qlnxr_create_srq;
208                 ibdev->destroy_srq = qlnxr_destroy_srq;
209                 ibdev->modify_srq = qlnxr_modify_srq;
210                 ibdev->query_srq = qlnxr_query_srq;
211                 ibdev->post_srq_recv = qlnxr_post_srq_recv;
212         }
213 
214 	ibdev->phys_port_cnt = 1;
215 	ibdev->num_comp_vectors = dev->num_cnq;
216 
217         /* mandatory verbs. */
218         ibdev->query_device = qlnxr_query_device;
219         ibdev->query_port = qlnxr_query_port;
220         ibdev->modify_port = qlnxr_modify_port;
221 
222 	ibdev->alloc_ucontext = qlnxr_alloc_ucontext;
223 	ibdev->dealloc_ucontext = qlnxr_dealloc_ucontext;
224         /* mandatory to support user space verbs consumer. */
225         ibdev->mmap = qlnxr_mmap;
226 
227         ibdev->alloc_pd = qlnxr_alloc_pd;
228         ibdev->dealloc_pd = qlnxr_dealloc_pd;
229 
230         ibdev->create_cq = qlnxr_create_cq;
231         ibdev->destroy_cq = qlnxr_destroy_cq;
232         ibdev->resize_cq = qlnxr_resize_cq;
233         ibdev->req_notify_cq = qlnxr_arm_cq;
234 
235         ibdev->create_qp = qlnxr_create_qp;
236         ibdev->modify_qp = qlnxr_modify_qp;
237         ibdev->query_qp = qlnxr_query_qp;
238         ibdev->destroy_qp = qlnxr_destroy_qp;
239 
240         ibdev->query_pkey = qlnxr_query_pkey;
241         ibdev->create_ah = qlnxr_create_ah;
242         ibdev->destroy_ah = qlnxr_destroy_ah;
243         ibdev->query_ah = qlnxr_query_ah;
244         ibdev->modify_ah = qlnxr_modify_ah;
245         ibdev->get_dma_mr = qlnxr_get_dma_mr;
246         ibdev->dereg_mr = qlnxr_dereg_mr;
247         ibdev->reg_user_mr = qlnxr_reg_user_mr;
248 
249 	ibdev->alloc_mr = qlnxr_alloc_mr;
250 	ibdev->map_mr_sg = qlnxr_map_mr_sg;
251 	ibdev->get_port_immutable = qlnxr_get_port_immutable;
252 
253         ibdev->poll_cq = qlnxr_poll_cq;
254         ibdev->post_send = qlnxr_post_send;
255         ibdev->post_recv = qlnxr_post_recv;
256 	ibdev->process_mad = qlnxr_process_mad;
257 
258 	ibdev->dma_device = &dev->pdev.dev;
259 
260 	ibdev->get_link_layer = qlnxr_link_layer;
261 
262 	if (QLNX_IS_IWARP(dev)) {
263                 iwcm = kmalloc(sizeof(*iwcm), GFP_KERNEL);
264 
265 		device_printf(dev->ha->pci_dev, "device is IWARP\n");
266 		if (iwcm == NULL)
267 			return (-ENOMEM);
268 
269                 ibdev->iwcm = iwcm;
270 
271                 iwcm->connect = qlnxr_iw_connect;
272                 iwcm->accept = qlnxr_iw_accept;
273                 iwcm->reject = qlnxr_iw_reject;
274 
275                 iwcm->create_listen = qlnxr_iw_create_listen;
276                 iwcm->destroy_listen = __qlnxr_iw_destroy_listen;
277 
278                 iwcm->add_ref = qlnxr_iw_qp_add_ref;
279                 iwcm->rem_ref = qlnxr_iw_qp_rem_ref;
280                 iwcm->get_qp = qlnxr_iw_get_qp;
281         }
282 
283         ret = ib_register_device(ibdev, NULL);
284 	if (ret) {
285 		kfree(iwcm);
286 	}
287 
288 	QL_DPRINT12(dev->ha, "exit\n");
289         return ret;
290 }
291 
292 #define HILO_U64(hi, lo)                ((((u64)(hi)) << 32) + (lo))
293 
294 static void
qlnxr_intr(void * handle)295 qlnxr_intr(void *handle)
296 {
297         struct qlnxr_cnq *cnq = handle;
298         struct qlnxr_cq *cq;
299         struct regpair *cq_handle;
300         u16 hw_comp_cons, sw_comp_cons;
301 	qlnx_host_t *ha;
302 
303 	ha = cnq->dev->ha;
304 
305 	QL_DPRINT12(ha, "enter cnq = %p\n", handle);
306 
307         ecore_sb_ack(cnq->sb, IGU_INT_DISABLE, 0 /*do not update*/);
308 
309         ecore_sb_update_sb_idx(cnq->sb);
310 
311         hw_comp_cons = le16_to_cpu(*cnq->hw_cons_ptr);
312         sw_comp_cons = ecore_chain_get_cons_idx(&cnq->pbl);
313 
314         rmb();
315 
316 	QL_DPRINT12(ha, "enter cnq = %p hw_comp_cons = 0x%x sw_comp_cons = 0x%x\n",
317 		handle, hw_comp_cons, sw_comp_cons);
318 
319         while (sw_comp_cons != hw_comp_cons) {
320                 cq_handle = (struct regpair *)ecore_chain_consume(&cnq->pbl);
321                 cq = (struct qlnxr_cq *)(uintptr_t)HILO_U64(cq_handle->hi,
322                                 cq_handle->lo);
323 
324                 if (cq == NULL) {
325 			QL_DPRINT11(ha, "cq == NULL\n");
326                         break;
327                 }
328 
329                 if (cq->sig != QLNXR_CQ_MAGIC_NUMBER) {
330 			QL_DPRINT11(ha,
331 				"cq->sig = 0x%x QLNXR_CQ_MAGIC_NUMBER = 0x%x\n",
332 				cq->sig, QLNXR_CQ_MAGIC_NUMBER);
333                         break;
334                 }
335                 cq->arm_flags = 0;
336 
337                 if (!cq->destroyed && cq->ibcq.comp_handler) {
338 			QL_DPRINT11(ha, "calling comp_handler = %p "
339 				"ibcq = %p cq_context = 0x%x\n",
340 				&cq->ibcq, cq->ibcq.cq_context);
341 
342                         (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context);
343                 }
344 		cq->cnq_notif++;
345 
346                 sw_comp_cons = ecore_chain_get_cons_idx(&cnq->pbl);
347 
348                 cnq->n_comp++;
349         }
350 
351         ecore_rdma_cnq_prod_update(cnq->dev->rdma_ctx, cnq->index, sw_comp_cons);
352 
353         ecore_sb_ack(cnq->sb, IGU_INT_ENABLE, 1 /*update*/);
354 
355 	QL_DPRINT12(ha, "exit cnq = %p\n", handle);
356         return;
357 }
358 
359 static void
qlnxr_release_irqs(struct qlnxr_dev * dev)360 qlnxr_release_irqs(struct qlnxr_dev *dev)
361 {
362 	int i;
363 	qlnx_host_t *ha;
364 
365 	ha = dev->ha;
366 
367 	QL_DPRINT12(ha, "enter\n");
368 
369         for (i = 0; i < dev->num_cnq; i++) {
370                 if (dev->cnq_array[i].irq_handle)
371                         (void)bus_teardown_intr(dev->ha->pci_dev,
372 				dev->cnq_array[i].irq,
373                                 dev->cnq_array[i].irq_handle);
374 
375                 if (dev->cnq_array[i].irq)
376                         (void) bus_release_resource(dev->ha->pci_dev,
377 				SYS_RES_IRQ,
378                                 dev->cnq_array[i].irq_rid,
379 				dev->cnq_array[i].irq);
380 	}
381 	QL_DPRINT12(ha, "exit\n");
382 	return;
383 }
384 
385 static int
qlnxr_setup_irqs(struct qlnxr_dev * dev)386 qlnxr_setup_irqs(struct qlnxr_dev *dev)
387 {
388 	int start_irq_rid;
389 	int i;
390 	qlnx_host_t *ha;
391 
392 	ha = dev->ha;
393 
394 	start_irq_rid = dev->sb_start + 2;
395 
396 	QL_DPRINT12(ha, "enter start_irq_rid = %d num_rss = %d\n",
397 		start_irq_rid, dev->ha->num_rss);
398 
399         for (i = 0; i < dev->num_cnq; i++) {
400 		dev->cnq_array[i].irq_rid = start_irq_rid + i;
401 
402 		dev->cnq_array[i].irq = bus_alloc_resource_any(dev->ha->pci_dev,
403 						SYS_RES_IRQ,
404 						&dev->cnq_array[i].irq_rid,
405 						(RF_ACTIVE | RF_SHAREABLE));
406 
407 		if (dev->cnq_array[i].irq == NULL) {
408 			QL_DPRINT11(ha,
409 				"bus_alloc_resource_any failed irq_rid = %d\n",
410 				dev->cnq_array[i].irq_rid);
411 
412 			goto qlnxr_setup_irqs_err;
413 		}
414 
415                 if (bus_setup_intr(dev->ha->pci_dev,
416                                 dev->cnq_array[i].irq,
417                                 (INTR_TYPE_NET | INTR_MPSAFE),
418                                 NULL, qlnxr_intr, &dev->cnq_array[i],
419 				&dev->cnq_array[i].irq_handle)) {
420 			QL_DPRINT11(ha, "bus_setup_intr failed\n");
421 			goto qlnxr_setup_irqs_err;
422                 }
423 		QL_DPRINT12(ha, "irq_rid = %d irq = %p irq_handle = %p\n",
424 			dev->cnq_array[i].irq_rid, dev->cnq_array[i].irq,
425 			dev->cnq_array[i].irq_handle);
426 	}
427 
428 	QL_DPRINT12(ha, "exit\n");
429 	return (0);
430 
431 qlnxr_setup_irqs_err:
432 	qlnxr_release_irqs(dev);
433 
434 	QL_DPRINT12(ha, "exit -1\n");
435 	return (-1);
436 }
437 
438 static void
qlnxr_free_resources(struct qlnxr_dev * dev)439 qlnxr_free_resources(struct qlnxr_dev *dev)
440 {
441         int i;
442 	qlnx_host_t *ha;
443 
444 	ha = dev->ha;
445 
446 	QL_DPRINT12(ha, "enter dev->num_cnq = %d\n", dev->num_cnq);
447 
448 	if (QLNX_IS_IWARP(dev)) {
449 		if (dev->iwarp_wq != NULL)
450 			destroy_workqueue(dev->iwarp_wq);
451 	}
452 
453         for (i = 0; i < dev->num_cnq; i++) {
454                 qlnx_free_mem_sb(dev->ha, &dev->sb_array[i]);
455                 ecore_chain_free(&dev->ha->cdev, &dev->cnq_array[i].pbl);
456         }
457 
458 	bzero(dev->cnq_array, (sizeof(struct qlnxr_cnq) * QLNXR_MAX_MSIX));
459 	bzero(dev->sb_array, (sizeof(struct ecore_sb_info) * QLNXR_MAX_MSIX));
460 	bzero(dev->sgid_tbl, (sizeof(union ib_gid) * QLNXR_MAX_SGID));
461 
462 	if (mtx_initialized(&dev->idr_lock))
463 		mtx_destroy(&dev->idr_lock);
464 
465 	if (mtx_initialized(&dev->sgid_lock))
466 		mtx_destroy(&dev->sgid_lock);
467 
468 	QL_DPRINT12(ha, "exit\n");
469 	return;
470 }
471 
472 static int
qlnxr_alloc_resources(struct qlnxr_dev * dev)473 qlnxr_alloc_resources(struct qlnxr_dev *dev)
474 {
475 	uint16_t n_entries;
476 	int i, rc;
477 	qlnx_host_t *ha;
478 
479 	ha = dev->ha;
480 
481         QL_DPRINT12(ha, "enter\n");
482 
483         bzero(dev->sgid_tbl, (sizeof (union ib_gid) * QLNXR_MAX_SGID));
484 
485         mtx_init(&dev->idr_lock, "idr_lock", NULL, MTX_DEF);
486         mtx_init(&dev->sgid_lock, "sgid_lock", NULL, MTX_DEF);
487 
488         idr_init(&dev->qpidr);
489 
490         bzero(dev->sb_array, (sizeof (struct ecore_sb_info) * QLNXR_MAX_MSIX));
491         bzero(dev->cnq_array, (sizeof (struct qlnxr_cnq) * QLNXR_MAX_MSIX));
492 
493         dev->sb_start = ecore_rdma_get_sb_id(dev->rdma_ctx, 0);
494 
495         QL_DPRINT12(ha, "dev->sb_start = 0x%x\n", dev->sb_start);
496 
497         /* Allocate CNQ PBLs */
498 
499         n_entries = min_t(u32, ECORE_RDMA_MAX_CNQ_SIZE, QLNXR_ROCE_MAX_CNQ_SIZE);
500 
501         for (i = 0; i < dev->num_cnq; i++) {
502                 rc = qlnx_alloc_mem_sb(dev->ha, &dev->sb_array[i],
503                                        dev->sb_start + i);
504                 if (rc)
505                         goto qlnxr_alloc_resources_exit;
506 
507                 rc = ecore_chain_alloc(&dev->ha->cdev,
508                                 ECORE_CHAIN_USE_TO_CONSUME_PRODUCE,
509                                 ECORE_CHAIN_MODE_PBL,
510                                 ECORE_CHAIN_CNT_TYPE_U16,
511                                 n_entries,
512                                 sizeof(struct regpair *),
513                                 &dev->cnq_array[i].pbl,
514                                 NULL);
515 
516                 /* configure cnq, except name since ibdev.name is still NULL */
517                 dev->cnq_array[i].dev = dev;
518                 dev->cnq_array[i].sb = &dev->sb_array[i];
519                 dev->cnq_array[i].hw_cons_ptr =
520                         &(dev->sb_array[i].sb_virt->pi_array[ECORE_ROCE_PROTOCOL_INDEX]);
521                 dev->cnq_array[i].index = i;
522                 sprintf(dev->cnq_array[i].name, "qlnxr%d@pci:%d",
523                         i, (dev->ha->pci_func));
524         }
525 
526 	QL_DPRINT12(ha, "exit\n");
527         return 0;
528 
529 qlnxr_alloc_resources_exit:
530 
531 	qlnxr_free_resources(dev);
532 
533 	QL_DPRINT12(ha, "exit -ENOMEM\n");
534         return -ENOMEM;
535 }
536 
537 void
qlnxr_affiliated_event(void * context,u8 e_code,void * fw_handle)538 qlnxr_affiliated_event(void *context, u8 e_code, void *fw_handle)
539 {
540 #define EVENT_TYPE_NOT_DEFINED  0
541 #define EVENT_TYPE_CQ           1
542 #define EVENT_TYPE_QP           2
543 #define EVENT_TYPE_GENERAL      3
544 
545         struct qlnxr_dev *dev = (struct qlnxr_dev *)context;
546         struct regpair *async_handle = (struct regpair *)fw_handle;
547         u64 roceHandle64 = ((u64)async_handle->hi << 32) + async_handle->lo;
548         struct qlnxr_cq *cq =  (struct qlnxr_cq *)(uintptr_t)roceHandle64;
549         struct qlnxr_qp *qp =  (struct qlnxr_qp *)(uintptr_t)roceHandle64;
550         u8 event_type = EVENT_TYPE_NOT_DEFINED;
551         struct ib_event event;
552 	qlnx_host_t *ha;
553 
554 	ha = dev->ha;
555 
556 	QL_DPRINT12(ha, "enter context = %p e_code = 0x%x fw_handle = %p\n",
557 		context, e_code, fw_handle);
558 
559         if (QLNX_IS_IWARP(dev)) {
560 		switch (e_code) {
561 		case ECORE_IWARP_EVENT_CQ_OVERFLOW:
562 			event.event = IB_EVENT_CQ_ERR;
563 			event_type = EVENT_TYPE_CQ;
564 			break;
565 
566 		default:
567 			QL_DPRINT12(ha,
568 				"unsupported event %d on handle=%llx\n",
569 				e_code, roceHandle64);
570 			break;
571 		}
572         } else {
573 		switch (e_code) {
574 		case ROCE_ASYNC_EVENT_CQ_OVERFLOW_ERR:
575 			event.event = IB_EVENT_CQ_ERR;
576 			event_type = EVENT_TYPE_CQ;
577 			break;
578 
579 		case ROCE_ASYNC_EVENT_SQ_DRAINED:
580 			event.event = IB_EVENT_SQ_DRAINED;
581 			event_type = EVENT_TYPE_QP;
582 			break;
583 
584 		case ROCE_ASYNC_EVENT_QP_CATASTROPHIC_ERR:
585 			event.event = IB_EVENT_QP_FATAL;
586 			event_type = EVENT_TYPE_QP;
587 			break;
588 
589 		case ROCE_ASYNC_EVENT_LOCAL_INVALID_REQUEST_ERR:
590 			event.event = IB_EVENT_QP_REQ_ERR;
591 			event_type = EVENT_TYPE_QP;
592 			break;
593 
594 		case ROCE_ASYNC_EVENT_LOCAL_ACCESS_ERR:
595 			event.event = IB_EVENT_QP_ACCESS_ERR;
596 			event_type = EVENT_TYPE_QP;
597 			break;
598 
599 		/* NOTE the following are not implemented in FW
600 		 *      ROCE_ASYNC_EVENT_CQ_ERR
601 		 *      ROCE_ASYNC_EVENT_COMM_EST
602 		 */
603 		/* TODO associate the following events -
604 		 *      ROCE_ASYNC_EVENT_SRQ_LIMIT
605 		 *      ROCE_ASYNC_EVENT_LAST_WQE_REACHED
606 		 *      ROCE_ASYNC_EVENT_LOCAL_CATASTROPHIC_ERR (un-affiliated)
607 		 */
608 		default:
609 			QL_DPRINT12(ha,
610 				"unsupported event 0x%x on fw_handle = %p\n",
611 				e_code, fw_handle);
612 			break;
613 		}
614 	}
615 
616         switch (event_type) {
617         case EVENT_TYPE_CQ:
618                 if (cq && cq->sig == QLNXR_CQ_MAGIC_NUMBER) {
619                         struct ib_cq *ibcq = &cq->ibcq;
620 
621                         if (ibcq->event_handler) {
622                                 event.device     = ibcq->device;
623                                 event.element.cq = ibcq;
624                                 ibcq->event_handler(&event, ibcq->cq_context);
625                         }
626                 } else {
627 			QL_DPRINT11(ha,
628 				"CQ event with invalid CQ pointer"
629 				" Handle = %llx\n", roceHandle64);
630                 }
631 		QL_DPRINT12(ha,
632 			"CQ event 0x%x on handle = %p\n", e_code, cq);
633                 break;
634 
635         case EVENT_TYPE_QP:
636                 if (qp && qp->sig == QLNXR_QP_MAGIC_NUMBER) {
637                         struct ib_qp *ibqp = &qp->ibqp;
638 
639                         if (ibqp->event_handler) {
640                                 event.device     = ibqp->device;
641                                 event.element.qp = ibqp;
642                                 ibqp->event_handler(&event, ibqp->qp_context);
643                         }
644                 } else {
645 			QL_DPRINT11(ha,
646 				"QP event 0x%x with invalid QP pointer"
647 				" qp handle = %p\n",
648 				e_code, roceHandle64);
649                 }
650 		QL_DPRINT12(ha, "QP event 0x%x on qp handle = %p\n",
651 			e_code, qp);
652                 break;
653 
654         case EVENT_TYPE_GENERAL:
655                 break;
656 
657         default:
658                 break;
659 	}
660 
661 	QL_DPRINT12(ha, "exit\n");
662 
663 	return;
664 }
665 
666 void
qlnxr_unaffiliated_event(void * context,u8 e_code)667 qlnxr_unaffiliated_event(void *context, u8 e_code)
668 {
669         struct qlnxr_dev *dev = (struct qlnxr_dev *)context;
670 	qlnx_host_t *ha;
671 
672 	ha = dev->ha;
673 
674 	QL_DPRINT12(ha, "enter/exit \n");
675 	return;
676 }
677 
678 static int
qlnxr_set_device_attr(struct qlnxr_dev * dev)679 qlnxr_set_device_attr(struct qlnxr_dev *dev)
680 {
681 	struct ecore_rdma_device *ecore_attr;
682 	struct qlnxr_device_attr *attr;
683 	u32 page_size;
684 
685 	ecore_attr = ecore_rdma_query_device(dev->rdma_ctx);
686 
687 	page_size = ~dev->attr.page_size_caps + 1;
688 	if(page_size > PAGE_SIZE) {
689 		QL_DPRINT12(dev->ha, "Kernel page size : %ld is smaller than"
690 		    " minimum page size : %ld required by qlnxr\n",
691 		    PAGE_SIZE, page_size);
692 		return -ENODEV;
693 	}
694 	attr = &dev->attr;
695         attr->vendor_id = ecore_attr->vendor_id;
696         attr->vendor_part_id = ecore_attr->vendor_part_id;
697 
698         QL_DPRINT12(dev->ha, "in qlnxr_set_device_attr, vendor : %x device : %x\n",
699 		attr->vendor_id, attr->vendor_part_id);
700 
701 	attr->hw_ver = ecore_attr->hw_ver;
702         attr->fw_ver = ecore_attr->fw_ver;
703         attr->node_guid = ecore_attr->node_guid;
704         attr->sys_image_guid = ecore_attr->sys_image_guid;
705         attr->max_cnq = ecore_attr->max_cnq;
706         attr->max_sge = ecore_attr->max_sge;
707         attr->max_inline = ecore_attr->max_inline;
708         attr->max_sqe = min_t(u32, ecore_attr->max_wqe, QLNXR_MAX_SQE);
709         attr->max_rqe = min_t(u32, ecore_attr->max_wqe, QLNXR_MAX_RQE);
710         attr->max_qp_resp_rd_atomic_resc = ecore_attr->max_qp_resp_rd_atomic_resc;
711         attr->max_qp_req_rd_atomic_resc = ecore_attr->max_qp_req_rd_atomic_resc;
712         attr->max_dev_resp_rd_atomic_resc =
713             ecore_attr->max_dev_resp_rd_atomic_resc;
714         attr->max_cq = ecore_attr->max_cq;
715         attr->max_qp = ecore_attr->max_qp;
716         attr->max_mr = ecore_attr->max_mr;
717 	attr->max_mr_size = ecore_attr->max_mr_size;
718         attr->max_cqe = min_t(u64, ecore_attr->max_cqe, QLNXR_MAX_CQES);
719         attr->max_mw = ecore_attr->max_mw;
720         attr->max_fmr = ecore_attr->max_fmr;
721         attr->max_mr_mw_fmr_pbl = ecore_attr->max_mr_mw_fmr_pbl;
722         attr->max_mr_mw_fmr_size = ecore_attr->max_mr_mw_fmr_size;
723         attr->max_pd = ecore_attr->max_pd;
724         attr->max_ah = ecore_attr->max_ah;
725         attr->max_pkey = ecore_attr->max_pkey;
726         attr->max_srq = ecore_attr->max_srq;
727         attr->max_srq_wr = ecore_attr->max_srq_wr;
728         //attr->dev_caps = ecore_attr->dev_caps;
729         attr->page_size_caps = ecore_attr->page_size_caps;
730         attr->dev_ack_delay = ecore_attr->dev_ack_delay;
731         attr->reserved_lkey = ecore_attr->reserved_lkey;
732         attr->bad_pkey_counter = ecore_attr->bad_pkey_counter;
733         attr->max_stats_queues = ecore_attr->max_stats_queues;
734 
735         return 0;
736 }
737 
738 static int
qlnxr_init_hw(struct qlnxr_dev * dev)739 qlnxr_init_hw(struct qlnxr_dev *dev)
740 {
741         struct ecore_rdma_events events;
742         struct ecore_rdma_add_user_out_params out_params;
743         struct ecore_rdma_cnq_params *cur_pbl;
744         struct ecore_rdma_start_in_params *in_params;
745         dma_addr_t p_phys_table;
746         u32 page_cnt;
747         int rc = 0;
748         int i;
749 	qlnx_host_t *ha;
750 
751 	ha = dev->ha;
752 
753 	QL_DPRINT12(ha, "enter\n");
754 
755         in_params = kzalloc(sizeof(*in_params), GFP_KERNEL);
756         if (!in_params) {
757                 rc = -ENOMEM;
758                 goto out;
759         }
760 
761 	bzero(&out_params, sizeof(struct ecore_rdma_add_user_out_params));
762 	bzero(&events, sizeof(struct ecore_rdma_events));
763 
764         in_params->desired_cnq = dev->num_cnq;
765 
766         for (i = 0; i < dev->num_cnq; i++) {
767                 cur_pbl = &in_params->cnq_pbl_list[i];
768 
769                 page_cnt = ecore_chain_get_page_cnt(&dev->cnq_array[i].pbl);
770                 cur_pbl->num_pbl_pages = page_cnt;
771 
772                 p_phys_table = ecore_chain_get_pbl_phys(&dev->cnq_array[i].pbl);
773                 cur_pbl->pbl_ptr = (u64)p_phys_table;
774         }
775 
776         events.affiliated_event = qlnxr_affiliated_event;
777         events.unaffiliated_event = qlnxr_unaffiliated_event;
778         events.context = dev;
779 
780         in_params->events = &events;
781         in_params->roce.cq_mode = ECORE_RDMA_CQ_MODE_32_BITS;
782         in_params->max_mtu = dev->ha->max_frame_size;
783 
784 	if (QLNX_IS_IWARP(dev)) {
785 	        if (delayed_ack)
786         	        in_params->iwarp.flags |= ECORE_IWARP_DA_EN;
787 
788 	        if (timestamp)
789         	        in_params->iwarp.flags |= ECORE_IWARP_TS_EN;
790 
791 	        in_params->iwarp.rcv_wnd_size = rcv_wnd_size*1024;
792 	        in_params->iwarp.crc_needed = crc_needed;
793 	        in_params->iwarp.ooo_num_rx_bufs =
794         	        (MAX_RXMIT_CONNS * in_params->iwarp.rcv_wnd_size) /
795 	                in_params->max_mtu;
796 
797 	        in_params->iwarp.mpa_peer2peer = peer2peer;
798 	        in_params->iwarp.mpa_rev =
799 			mpa_enhanced ? ECORE_MPA_REV2 : ECORE_MPA_REV1;
800 	        in_params->iwarp.mpa_rtr = rtr_type;
801 	}
802 
803         memcpy(&in_params->mac_addr[0], dev->ha->primary_mac, ETH_ALEN);
804 
805         rc = ecore_rdma_start(dev->rdma_ctx, in_params);
806         if (rc)
807                 goto out;
808 
809         rc = ecore_rdma_add_user(dev->rdma_ctx, &out_params);
810         if (rc)
811                 goto out;
812 
813         dev->db_addr = (void *)(uintptr_t)out_params.dpi_addr;
814         dev->db_phys_addr = out_params.dpi_phys_addr;
815         dev->db_size = out_params.dpi_size;
816         dev->dpi = out_params.dpi;
817 
818 	qlnxr_set_device_attr(dev);
819 
820 	QL_DPRINT12(ha,
821 		"cdev->doorbells = %p, db_phys_addr = %p db_size = 0x%x\n",
822 		(void *)ha->cdev.doorbells,
823 		(void *)ha->cdev.db_phys_addr, ha->cdev.db_size);
824 
825 	QL_DPRINT12(ha,
826 		"db_addr = %p db_phys_addr = %p db_size = 0x%x dpi = 0x%x\n",
827 		(void *)dev->db_addr, (void *)dev->db_phys_addr,
828 		dev->db_size, dev->dpi);
829 out:
830         kfree(in_params);
831 
832 	QL_DPRINT12(ha, "exit\n");
833         return rc;
834 }
835 
836 static void
qlnxr_build_sgid_mac(union ib_gid * sgid,unsigned char * mac_addr,bool is_vlan,u16 vlan_id)837 qlnxr_build_sgid_mac(union ib_gid *sgid, unsigned char *mac_addr,
838 	bool is_vlan, u16 vlan_id)
839 {
840 	sgid->global.subnet_prefix = OSAL_CPU_TO_BE64(0xfe80000000000000LL);
841 	sgid->raw[8] = mac_addr[0] ^ 2;
842 	sgid->raw[9] = mac_addr[1];
843 	sgid->raw[10] = mac_addr[2];
844 	if (is_vlan) {
845 		sgid->raw[11] = vlan_id >> 8;
846 		sgid->raw[12] = vlan_id & 0xff;
847 	} else {
848 		sgid->raw[11] = 0xff;
849 		sgid->raw[12] = 0xfe;
850 	}
851 	sgid->raw[13] = mac_addr[3];
852 	sgid->raw[14] = mac_addr[4];
853 	sgid->raw[15] = mac_addr[5];
854 }
855 static bool
856 qlnxr_add_sgid(struct qlnxr_dev *dev, union ib_gid *new_sgid);
857 
858 struct qlnx_cb_s {
859 	struct qlnxr_dev *dev;
860 	union ib_gid gid;
861 };
862 
863 static u_int
qlnxr_add_ip_based_gid_cb(void * arg,struct ifaddr * ifa,u_int count)864 qlnxr_add_ip_based_gid_cb(void *arg, struct ifaddr *ifa, u_int count)
865 {
866 	struct qlnx_cb_s *cba = arg;
867 
868 	QL_DPRINT12(cba->dev->ha, "IP address : %x\n", ((struct sockaddr_in *) ifa->ifa_addr)->sin_addr.s_addr);
869 	ipv6_addr_set_v4mapped(
870 		((struct sockaddr_in *) ifa->ifa_addr)->sin_addr.s_addr,
871 		(struct in6_addr *)&cba->gid);
872 	QL_DPRINT12(cba->dev->ha, "gid generated : %llx\n", cba->gid);
873 
874 	qlnxr_add_sgid(cba->dev, &cba->gid);
875 	return (1);
876 }
877 
878 static void
qlnxr_add_ip_based_gid(struct qlnxr_dev * dev,if_t ifp)879 qlnxr_add_ip_based_gid(struct qlnxr_dev *dev, if_t ifp)
880 {
881 	struct qlnx_cb_s cba;
882 
883 	if_foreach_addr_type(ifp, AF_INET, qlnxr_add_ip_based_gid_cb, &cba);
884 	for (int i = 0; i < 16; i++) {
885 		QL_DPRINT12(dev->ha, "gid generated : %x\n", cba.gid.raw[i]);
886 	}
887 }
888 
889 static bool
qlnxr_add_sgid(struct qlnxr_dev * dev,union ib_gid * new_sgid)890 qlnxr_add_sgid(struct qlnxr_dev *dev, union ib_gid *new_sgid)
891 {
892 	union ib_gid zero_sgid = { { 0 } };
893 	int i;
894 	//unsigned long flags;
895 	mtx_lock(&dev->sgid_lock);
896 	for (i = 0; i < QLNXR_MAX_SGID; i++) {
897 		if (!memcmp(&dev->sgid_tbl[i], &zero_sgid,
898 				sizeof(union ib_gid))) {
899 			/* found free entry */
900 			memcpy(&dev->sgid_tbl[i], new_sgid,
901 				sizeof(union ib_gid));
902 			QL_DPRINT12(dev->ha, "copying sgid : %llx\n",
903 					*new_sgid);
904 			mtx_unlock(&dev->sgid_lock);
905 			//TODO ib_dispatch event here?
906 			return true;
907 		} else if (!memcmp(&dev->sgid_tbl[i], new_sgid,
908 				sizeof(union ib_gid))) {
909 			/* entry already present, no addition required */
910 			mtx_unlock(&dev->sgid_lock);
911 			QL_DPRINT12(dev->ha, "sgid present : %llx\n",
912 					*new_sgid);
913 			return false;
914 		}
915 	}
916 	if (i == QLNXR_MAX_SGID) {
917 		QL_DPRINT12(dev->ha, "didn't find an empty entry in sgid_tbl\n");
918 	}
919 	mtx_unlock(&dev->sgid_lock);
920 	return false;
921 }
922 
qlnxr_del_sgid(struct qlnxr_dev * dev,union ib_gid * gid)923 static bool qlnxr_del_sgid(struct qlnxr_dev *dev, union ib_gid *gid)
924 {
925 	int found = false;
926 	int i;
927 	//unsigned long flags;
928 
929 	QL_DPRINT12(dev->ha, "removing gid %llx %llx\n",
930 			gid->global.interface_id,
931 			gid->global.subnet_prefix);
932 	mtx_lock(&dev->sgid_lock);
933 	/* first is the default sgid which cannot be deleted */
934 	for (i = 1; i < QLNXR_MAX_SGID; i++) {
935 		if (!memcmp(&dev->sgid_tbl[i], gid, sizeof(union ib_gid))) {
936 			/* found matching entry */
937 			memset(&dev->sgid_tbl[i], 0, sizeof(union ib_gid));
938 			found = true;
939 			break;
940 		}
941 	}
942 	mtx_unlock(&dev->sgid_lock);
943 
944 	return found;
945 }
946 
947 static void
qlnxr_add_sgids(struct qlnxr_dev * dev)948 qlnxr_add_sgids(struct qlnxr_dev *dev)
949 {
950 	qlnx_host_t *ha = dev->ha;
951 	u16 vlan_id;
952 	bool is_vlan;
953 	union ib_gid vgid;
954 
955 	qlnxr_add_ip_based_gid(dev, ha->ifp);
956 	/* MAC/VLAN base GIDs */
957 	is_vlan = is_vlan_dev(ha->ifp);
958        	vlan_id = (is_vlan) ? vlan_dev_vlan_id(ha->ifp) : 0;
959 	qlnxr_build_sgid_mac(&vgid, ha->primary_mac, is_vlan, vlan_id);
960 	qlnxr_add_sgid(dev, &vgid);
961 }
962 
963 static int
qlnxr_add_default_sgid(struct qlnxr_dev * dev)964 qlnxr_add_default_sgid(struct qlnxr_dev *dev)
965 {
966 	/* GID Index 0 - Invariant manufacturer-assigned EUI-64 */
967 	union ib_gid *sgid = &dev->sgid_tbl[0];
968 	struct ecore_rdma_device        *qattr;
969 	qlnx_host_t *ha;
970 	ha = dev->ha;
971 
972 	qattr =	ecore_rdma_query_device(dev->rdma_ctx);
973 	if(sgid == NULL)
974 		QL_DPRINT12(ha, "sgid = NULL?\n");
975 
976 	sgid->global.subnet_prefix = OSAL_CPU_TO_BE64(0xfe80000000000000LL);
977 	QL_DPRINT12(ha, "node_guid = %llx", dev->attr.node_guid);
978 	memcpy(&sgid->raw[8], &qattr->node_guid,
979 		sizeof(qattr->node_guid));
980 	//memcpy(&sgid->raw[8], &dev->attr.node_guid,
981 	//	sizeof(dev->attr.node_guid));
982 	QL_DPRINT12(ha, "DEFAULT sgid=[%x][%x][%x][%x][%x][%x][%x][%x][%x][%x][%x][%x][%x][%x][%x][%x]\n",
983                    sgid->raw[0], sgid->raw[1], sgid->raw[2], sgid->raw[3], sgid->raw[4], sgid->raw[5],
984                    sgid->raw[6], sgid->raw[7], sgid->raw[8], sgid->raw[9], sgid->raw[10], sgid->raw[11],
985                    sgid->raw[12], sgid->raw[13], sgid->raw[14], sgid->raw[15]);
986 	return 0;
987 }
988 
qlnxr_addr_event(struct qlnxr_dev * dev,unsigned long event,if_t ifp,union ib_gid * gid)989 static int qlnxr_addr_event (struct qlnxr_dev *dev,
990 				unsigned long event,
991 				if_t ifp,
992 				union ib_gid *gid)
993 {
994 	bool is_vlan = false;
995 	union ib_gid vgid;
996 	u16 vlan_id = 0xffff;
997 
998 	QL_DPRINT12(dev->ha, "Link event occured\n");
999 	is_vlan = is_vlan_dev(dev->ha->ifp);
1000 	vlan_id = (is_vlan) ? vlan_dev_vlan_id(dev->ha->ifp) : 0;
1001 
1002 	switch (event) {
1003 	case NETDEV_UP :
1004 		qlnxr_add_sgid(dev, gid);
1005 		if (is_vlan) {
1006 			qlnxr_build_sgid_mac(&vgid, dev->ha->primary_mac, is_vlan, vlan_id);
1007 			qlnxr_add_sgid(dev, &vgid);
1008 		}
1009 		break;
1010 	case NETDEV_DOWN :
1011 		qlnxr_del_sgid(dev, gid);
1012 		if (is_vlan) {
1013 			qlnxr_build_sgid_mac(&vgid, dev->ha->primary_mac, is_vlan, vlan_id);
1014 			qlnxr_del_sgid(dev, &vgid);
1015 		}
1016 		break;
1017 	default :
1018 		break;
1019 	}
1020 	return 1;
1021 }
1022 
qlnxr_inetaddr_event(struct notifier_block * notifier,unsigned long event,void * ptr)1023 static int qlnxr_inetaddr_event(struct notifier_block *notifier,
1024 				unsigned long event, void *ptr)
1025 {
1026 	struct ifaddr *ifa = ptr;
1027 	union ib_gid gid;
1028 	struct qlnxr_dev *dev = container_of(notifier, struct qlnxr_dev, nb_inet);
1029 	qlnx_host_t *ha = dev->ha;
1030 
1031 	ipv6_addr_set_v4mapped(
1032 			((struct sockaddr_in *) ifa->ifa_addr)->sin_addr.s_addr,
1033 			(struct in6_addr *)&gid);
1034 	return qlnxr_addr_event(dev, event, ha->ifp, &gid);
1035 }
1036 
1037 static int
qlnxr_register_inet(struct qlnxr_dev * dev)1038 qlnxr_register_inet(struct qlnxr_dev *dev)
1039 {
1040 	int ret;
1041 	dev->nb_inet.notifier_call = qlnxr_inetaddr_event;
1042 	ret = register_inetaddr_notifier(&dev->nb_inet);
1043 	if (ret) {
1044 		QL_DPRINT12(dev->ha, "Failed to register inetaddr\n");
1045 		return ret;
1046 	}
1047 	/* TODO : add for CONFIG_IPV6) */
1048 	return 0;
1049 }
1050 
1051 static int
qlnxr_build_sgid_tbl(struct qlnxr_dev * dev)1052 qlnxr_build_sgid_tbl(struct qlnxr_dev *dev)
1053 {
1054 	qlnxr_add_default_sgid(dev);
1055 	qlnxr_add_sgids(dev);
1056 	return 0;
1057 }
1058 
1059 static struct qlnx_rdma_if qlnxr_drv;
1060 
1061 static void *
qlnxr_add(void * eth_dev)1062 qlnxr_add(void *eth_dev)
1063 {
1064 	struct qlnxr_dev *dev;
1065 	int ret;
1066 	//device_t pci_dev;
1067 	qlnx_host_t *ha;
1068 
1069 	ha = eth_dev;
1070 
1071 	QL_DPRINT12(ha, "enter [ha = %p]\n", ha);
1072 
1073 	dev = (struct qlnxr_dev *)ib_alloc_device(sizeof(struct qlnxr_dev));
1074 
1075 	if (dev == NULL)
1076 		return (NULL);
1077 
1078 	dev->ha = eth_dev;
1079 	dev->cdev = &ha->cdev;
1080 	/* Added to extend Application support */
1081 	linux_pci_attach_device(dev->ha->pci_dev, NULL, NULL, &dev->pdev);
1082 
1083 	dev->rdma_ctx = &ha->cdev.hwfns[0];
1084 	dev->wq_multiplier = wq_multiplier;
1085 	dev->num_cnq = QLNX_NUM_CNQ;
1086 
1087 	QL_DPRINT12(ha,
1088 		"ha = %p dev = %p ha->cdev = %p\n",
1089 		ha, dev, &ha->cdev);
1090 	QL_DPRINT12(ha,
1091 		"dev->cdev = %p dev->rdma_ctx = %p\n",
1092 		dev->cdev, dev->rdma_ctx);
1093 
1094 	ret = qlnxr_alloc_resources(dev);
1095 
1096 	if (ret)
1097 		goto qlnxr_add_err;
1098 
1099 	ret = qlnxr_setup_irqs(dev);
1100 
1101 	if (ret) {
1102 		qlnxr_free_resources(dev);
1103 		goto qlnxr_add_err;
1104 	}
1105 
1106 	ret = qlnxr_init_hw(dev);
1107 
1108 	if (ret) {
1109 		qlnxr_release_irqs(dev);
1110 		qlnxr_free_resources(dev);
1111 		goto qlnxr_add_err;
1112 	}
1113 
1114 	qlnxr_register_device(dev);
1115 	for (int i = 0; i < ARRAY_SIZE(qlnxr_class_attributes); ++i) {
1116 		if (device_create_file(&dev->ibdev.dev, qlnxr_class_attributes[i]))
1117 			goto sysfs_err;
1118 	}
1119 	qlnxr_build_sgid_tbl(dev);
1120 	//ret = qlnxr_register_inet(dev);
1121 	QL_DPRINT12(ha, "exit\n");
1122 	if (!test_and_set_bit(QLNXR_ENET_STATE_BIT, &dev->enet_state)) {
1123 		QL_DPRINT12(ha, "dispatching IB_PORT_ACITVE event\n");
1124 		qlnxr_ib_dispatch_event(dev, QLNXR_PORT,
1125 			IB_EVENT_PORT_ACTIVE);
1126 	}
1127 
1128 	return (dev);
1129 sysfs_err:
1130 	for (int i = 0; i < ARRAY_SIZE(qlnxr_class_attributes); ++i) {
1131 		device_remove_file(&dev->ibdev.dev, qlnxr_class_attributes[i]);
1132 	}
1133 	ib_unregister_device(&dev->ibdev);
1134 
1135 qlnxr_add_err:
1136 	ib_dealloc_device(&dev->ibdev);
1137 
1138 	QL_DPRINT12(ha, "exit failed\n");
1139 	return (NULL);
1140 }
1141 
1142 static void
qlnxr_remove_sysfiles(struct qlnxr_dev * dev)1143 qlnxr_remove_sysfiles(struct qlnxr_dev *dev)
1144 {
1145 	int i;
1146 	for (i = 0; i < ARRAY_SIZE(qlnxr_class_attributes); ++i)
1147 		device_remove_file(&dev->ibdev.dev, qlnxr_class_attributes[i]);
1148 }
1149 
1150 static int
qlnxr_remove(void * eth_dev,void * qlnx_rdma_dev)1151 qlnxr_remove(void *eth_dev, void *qlnx_rdma_dev)
1152 {
1153 	struct qlnxr_dev *dev;
1154 	qlnx_host_t *ha;
1155 
1156 	dev = qlnx_rdma_dev;
1157 	ha = eth_dev;
1158 
1159 	if ((ha == NULL) || (dev == NULL))
1160 		return (0);
1161 
1162 	QL_DPRINT12(ha, "enter ha = %p qlnx_rdma_dev = %p pd_count = %d\n",
1163 		ha, qlnx_rdma_dev, dev->pd_count);
1164 
1165 	qlnxr_ib_dispatch_event(dev, QLNXR_PORT,
1166 		IB_EVENT_PORT_ERR);
1167 
1168 	if (QLNX_IS_IWARP(dev)) {
1169 		if (dev->pd_count)
1170 			return (EBUSY);
1171 	}
1172 
1173 	ib_unregister_device(&dev->ibdev);
1174 
1175 	if (QLNX_IS_ROCE(dev)) {
1176 		if (dev->pd_count)
1177 			return (EBUSY);
1178 	}
1179 
1180 	ecore_rdma_remove_user(dev->rdma_ctx, dev->dpi);
1181 	ecore_rdma_stop(dev->rdma_ctx);
1182 
1183 	qlnxr_release_irqs(dev);
1184 
1185 	qlnxr_free_resources(dev);
1186 
1187 	qlnxr_remove_sysfiles(dev);
1188 	ib_dealloc_device(&dev->ibdev);
1189 
1190 	linux_pci_detach_device(&dev->pdev);
1191 
1192 	QL_DPRINT12(ha, "exit ha = %p qlnx_rdma_dev = %p\n", ha, qlnx_rdma_dev);
1193 	return (0);
1194 }
1195 
1196 int
qlnx_rdma_ll2_set_mac_filter(void * rdma_ctx,uint8_t * old_mac_address,uint8_t * new_mac_address)1197 qlnx_rdma_ll2_set_mac_filter(void *rdma_ctx, uint8_t *old_mac_address,
1198 	uint8_t *new_mac_address)
1199 {
1200         struct ecore_hwfn *p_hwfn = rdma_ctx;
1201         struct qlnx_host *ha;
1202         int ret = 0;
1203 
1204         ha = (struct qlnx_host *)(p_hwfn->p_dev);
1205         QL_DPRINT2(ha, "enter rdma_ctx (%p)\n", rdma_ctx);
1206 
1207         if (old_mac_address)
1208                 ecore_llh_remove_mac_filter(p_hwfn->p_dev, 0, old_mac_address);
1209 
1210         if (new_mac_address)
1211                 ret = ecore_llh_add_mac_filter(p_hwfn->p_dev, 0, new_mac_address);
1212 
1213         QL_DPRINT2(ha, "exit rdma_ctx (%p)\n", rdma_ctx);
1214         return (ret);
1215 }
1216 
1217 static void
qlnxr_mac_address_change(struct qlnxr_dev * dev)1218 qlnxr_mac_address_change(struct qlnxr_dev *dev)
1219 {
1220 	qlnx_host_t *ha;
1221 
1222 	ha = dev->ha;
1223 
1224 	QL_DPRINT12(ha, "enter/exit\n");
1225 
1226 	return;
1227 }
1228 
1229 static void
qlnxr_notify(void * eth_dev,void * qlnx_rdma_dev,enum qlnx_rdma_event event)1230 qlnxr_notify(void *eth_dev, void *qlnx_rdma_dev, enum qlnx_rdma_event event)
1231 {
1232 	struct qlnxr_dev *dev;
1233 	qlnx_host_t *ha;
1234 
1235 	dev = qlnx_rdma_dev;
1236 
1237 	if (dev == NULL)
1238 		return;
1239 
1240 	ha = dev->ha;
1241 
1242 	QL_DPRINT12(ha, "enter (%p, %d)\n", qlnx_rdma_dev, event);
1243 
1244         switch (event) {
1245         case QLNX_ETHDEV_UP:
1246 		if (!test_and_set_bit(QLNXR_ENET_STATE_BIT, &dev->enet_state))
1247 			qlnxr_ib_dispatch_event(dev, QLNXR_PORT,
1248 				IB_EVENT_PORT_ACTIVE);
1249                 break;
1250 
1251         case QLNX_ETHDEV_CHANGE_ADDR:
1252                 qlnxr_mac_address_change(dev);
1253                 break;
1254 
1255         case QLNX_ETHDEV_DOWN:
1256 		if (test_and_set_bit(QLNXR_ENET_STATE_BIT, &dev->enet_state))
1257 			qlnxr_ib_dispatch_event(dev, QLNXR_PORT,
1258 				IB_EVENT_PORT_ERR);
1259                 break;
1260         }
1261 
1262 	QL_DPRINT12(ha, "exit (%p, %d)\n", qlnx_rdma_dev, event);
1263 	return;
1264 }
1265 
1266 static int
qlnxr_mod_load(void)1267 qlnxr_mod_load(void)
1268 {
1269 	int ret;
1270 
1271 	qlnxr_drv.add = qlnxr_add;
1272 	qlnxr_drv.remove = qlnxr_remove;
1273 	qlnxr_drv.notify = qlnxr_notify;
1274 
1275 	ret = qlnx_rdma_register_if(&qlnxr_drv);
1276 
1277 	return (ret);
1278 }
1279 
1280 static int
qlnxr_mod_unload(void)1281 qlnxr_mod_unload(void)
1282 {
1283 	int ret;
1284 
1285 	ret = qlnx_rdma_deregister_if(&qlnxr_drv);
1286 	return (ret);
1287 }
1288 
1289 static int
qlnxr_event_handler(module_t mod,int event,void * arg)1290 qlnxr_event_handler(module_t mod, int event, void *arg)
1291 {
1292 
1293 	int ret = 0;
1294 
1295 	switch (event) {
1296 	case MOD_LOAD:
1297 		ret = qlnxr_mod_load();
1298 		break;
1299 
1300 	case MOD_UNLOAD:
1301 		ret = qlnxr_mod_unload();
1302 		break;
1303 
1304 	default:
1305 		break;
1306 	}
1307 
1308         return (ret);
1309 }
1310 
1311 static moduledata_t qlnxr_mod_info = {
1312 	.name = "qlnxr",
1313 	.evhand = qlnxr_event_handler,
1314 };
1315 
1316 MODULE_VERSION(qlnxr, 1);
1317 MODULE_DEPEND(qlnxr, if_qlnxe, 1, 1, 1);
1318 MODULE_DEPEND(qlnxr, ibcore, 1, 1, 1);
1319 MODULE_DEPEND(qlnxr, linuxkpi, 1, 1, 1);
1320 
1321 DECLARE_MODULE(qlnxr, qlnxr_mod_info, SI_SUB_LAST, SI_ORDER_ANY);
1322