1 /*
2 * CDDL HEADER START
3 *
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
7 *
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
12 *
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 *
19 * CDDL HEADER END
20 */
21
22 /*
23 * Copyright (c) 2008, 2010, Oracle and/or its affiliates. All rights reserved.
24 */
25
26 /*
27 * hermon.c
28 * Hermon (InfiniBand) HCA Driver attach/detach Routines
29 *
30 * Implements all the routines necessary for the attach, setup,
31 * initialization (and subsequent possible teardown and detach) of the
32 * Hermon InfiniBand HCA driver.
33 */
34
35 #include <sys/types.h>
36 #include <sys/file.h>
37 #include <sys/open.h>
38 #include <sys/conf.h>
39 #include <sys/ddi.h>
40 #include <sys/sunddi.h>
41 #include <sys/modctl.h>
42 #include <sys/stat.h>
43 #include <sys/pci.h>
44 #include <sys/pci_cap.h>
45 #include <sys/bitmap.h>
46 #include <sys/policy.h>
47
48 #include <sys/ib/adapters/hermon/hermon.h>
49
50 /* /etc/system can tune this down, if that is desirable. */
51 int hermon_msix_max = HERMON_MSIX_MAX;
52
53 /* The following works around a problem in pre-2_7_000 firmware. */
54 #define HERMON_FW_WORKAROUND
55
56 int hermon_verbose = 0;
57
58 /* Hermon HCA State Pointer */
59 void *hermon_statep;
60
61 int debug_vpd = 0;
62
63 /* Disable the internal error-check polling thread */
64 int hermon_no_inter_err_chk = 0;
65
66 /*
67 * The Hermon "userland resource database" is common to instances of the
68 * Hermon HCA driver. This structure "hermon_userland_rsrc_db" contains all
69 * the necessary information to maintain it.
70 */
71 hermon_umap_db_t hermon_userland_rsrc_db;
72
73 static int hermon_attach(dev_info_t *, ddi_attach_cmd_t);
74 static int hermon_detach(dev_info_t *, ddi_detach_cmd_t);
75 static int hermon_open(dev_t *, int, int, cred_t *);
76 static int hermon_close(dev_t, int, int, cred_t *);
77 static int hermon_getinfo(dev_info_t *, ddi_info_cmd_t, void *, void **);
78
79 static int hermon_drv_init(hermon_state_t *state, dev_info_t *dip,
80 int instance);
81 static void hermon_drv_fini(hermon_state_t *state);
82 static void hermon_drv_fini2(hermon_state_t *state);
83 static int hermon_isr_init(hermon_state_t *state);
84 static void hermon_isr_fini(hermon_state_t *state);
85
86 static int hermon_hw_init(hermon_state_t *state);
87
88 static void hermon_hw_fini(hermon_state_t *state,
89 hermon_drv_cleanup_level_t cleanup);
90 static int hermon_soft_state_init(hermon_state_t *state);
91 static void hermon_soft_state_fini(hermon_state_t *state);
92 static int hermon_icm_config_setup(hermon_state_t *state,
93 hermon_hw_initqueryhca_t *inithca);
94 static void hermon_icm_tables_init(hermon_state_t *state);
95 static void hermon_icm_tables_fini(hermon_state_t *state);
96 static int hermon_icm_dma_init(hermon_state_t *state);
97 static void hermon_icm_dma_fini(hermon_state_t *state);
98 static void hermon_inithca_set(hermon_state_t *state,
99 hermon_hw_initqueryhca_t *inithca);
100 static int hermon_hca_port_init(hermon_state_t *state);
101 static int hermon_hca_ports_shutdown(hermon_state_t *state, uint_t num_init);
102 static int hermon_internal_uarpg_init(hermon_state_t *state);
103 static void hermon_internal_uarpg_fini(hermon_state_t *state);
104 static int hermon_special_qp_contexts_reserve(hermon_state_t *state);
105 static void hermon_special_qp_contexts_unreserve(hermon_state_t *state);
106 static int hermon_sw_reset(hermon_state_t *state);
107 static int hermon_mcg_init(hermon_state_t *state);
108 static void hermon_mcg_fini(hermon_state_t *state);
109 static int hermon_fw_version_check(hermon_state_t *state);
110 static void hermon_device_info_report(hermon_state_t *state);
111 static int hermon_pci_capability_list(hermon_state_t *state,
112 ddi_acc_handle_t hdl);
113 static void hermon_pci_capability_vpd(hermon_state_t *state,
114 ddi_acc_handle_t hdl, uint_t offset);
115 static int hermon_pci_read_vpd(ddi_acc_handle_t hdl, uint_t offset,
116 uint32_t addr, uint32_t *data);
117 static int hermon_intr_or_msi_init(hermon_state_t *state);
118 static int hermon_add_intrs(hermon_state_t *state, int intr_type);
119 static int hermon_intr_or_msi_fini(hermon_state_t *state);
120 void hermon_pci_capability_msix(hermon_state_t *state, ddi_acc_handle_t hdl,
121 uint_t offset);
122
123 static uint64_t hermon_size_icm(hermon_state_t *state);
124
125 /* X86 fastreboot support */
126 static ushort_t get_msix_ctrl(dev_info_t *);
127 static size_t get_msix_tbl_size(dev_info_t *);
128 static size_t get_msix_pba_size(dev_info_t *);
129 static void hermon_set_msix_info(hermon_state_t *);
130 static int hermon_intr_disable(hermon_state_t *);
131 static int hermon_quiesce(dev_info_t *);
132
133
134 /* Character/Block Operations */
135 static struct cb_ops hermon_cb_ops = {
136 hermon_open, /* open */
137 hermon_close, /* close */
138 nodev, /* strategy (block) */
139 nodev, /* print (block) */
140 nodev, /* dump (block) */
141 nodev, /* read */
142 nodev, /* write */
143 hermon_ioctl, /* ioctl */
144 hermon_devmap, /* devmap */
145 NULL, /* mmap */
146 nodev, /* segmap */
147 nochpoll, /* chpoll */
148 ddi_prop_op, /* prop_op */
149 NULL, /* streams */
150 D_NEW | D_MP |
151 D_64BIT | D_HOTPLUG |
152 D_DEVMAP, /* flags */
153 CB_REV /* rev */
154 };
155
156 /* Driver Operations */
157 static struct dev_ops hermon_ops = {
158 DEVO_REV, /* struct rev */
159 0, /* refcnt */
160 hermon_getinfo, /* getinfo */
161 nulldev, /* identify */
162 nulldev, /* probe */
163 hermon_attach, /* attach */
164 hermon_detach, /* detach */
165 nodev, /* reset */
166 &hermon_cb_ops, /* cb_ops */
167 NULL, /* bus_ops */
168 nodev, /* power */
169 hermon_quiesce, /* devo_quiesce */
170 };
171
172 /* Module Driver Info */
173 static struct modldrv hermon_modldrv = {
174 &mod_driverops,
175 "ConnectX IB Driver",
176 &hermon_ops
177 };
178
179 /* Module Linkage */
180 static struct modlinkage hermon_modlinkage = {
181 MODREV_1,
182 &hermon_modldrv,
183 NULL
184 };
185
186 /*
187 * This extern refers to the ibc_operations_t function vector that is defined
188 * in the hermon_ci.c file.
189 */
190 extern ibc_operations_t hermon_ibc_ops;
191
192 /*
193 * _init()
194 */
195 int
_init()196 _init()
197 {
198 int status;
199
200 status = ddi_soft_state_init(&hermon_statep, sizeof (hermon_state_t),
201 (size_t)HERMON_INITIAL_STATES);
202 if (status != 0) {
203 return (status);
204 }
205
206 status = ibc_init(&hermon_modlinkage);
207 if (status != 0) {
208 ddi_soft_state_fini(&hermon_statep);
209 return (status);
210 }
211
212 status = mod_install(&hermon_modlinkage);
213 if (status != 0) {
214 ibc_fini(&hermon_modlinkage);
215 ddi_soft_state_fini(&hermon_statep);
216 return (status);
217 }
218
219 /* Initialize the Hermon "userland resources database" */
220 hermon_umap_db_init();
221
222 return (status);
223 }
224
225
226 /*
227 * _info()
228 */
229 int
_info(struct modinfo * modinfop)230 _info(struct modinfo *modinfop)
231 {
232 int status;
233
234 status = mod_info(&hermon_modlinkage, modinfop);
235 return (status);
236 }
237
238
239 /*
240 * _fini()
241 */
242 int
_fini()243 _fini()
244 {
245 int status;
246
247 status = mod_remove(&hermon_modlinkage);
248 if (status != 0) {
249 return (status);
250 }
251
252 /* Destroy the Hermon "userland resources database" */
253 hermon_umap_db_fini();
254
255 ibc_fini(&hermon_modlinkage);
256 ddi_soft_state_fini(&hermon_statep);
257
258 return (status);
259 }
260
261
262 /*
263 * hermon_getinfo()
264 */
265 /* ARGSUSED */
266 static int
hermon_getinfo(dev_info_t * dip,ddi_info_cmd_t cmd,void * arg,void ** result)267 hermon_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg, void **result)
268 {
269 dev_t dev;
270 hermon_state_t *state;
271 minor_t instance;
272
273 switch (cmd) {
274 case DDI_INFO_DEVT2DEVINFO:
275 dev = (dev_t)arg;
276 instance = HERMON_DEV_INSTANCE(dev);
277 state = ddi_get_soft_state(hermon_statep, instance);
278 if (state == NULL) {
279 return (DDI_FAILURE);
280 }
281 *result = (void *)state->hs_dip;
282 return (DDI_SUCCESS);
283
284 case DDI_INFO_DEVT2INSTANCE:
285 dev = (dev_t)arg;
286 instance = HERMON_DEV_INSTANCE(dev);
287 *result = (void *)(uintptr_t)instance;
288 return (DDI_SUCCESS);
289
290 default:
291 break;
292 }
293
294 return (DDI_FAILURE);
295 }
296
297
298 /*
299 * hermon_open()
300 */
301 /* ARGSUSED */
302 static int
hermon_open(dev_t * devp,int flag,int otyp,cred_t * credp)303 hermon_open(dev_t *devp, int flag, int otyp, cred_t *credp)
304 {
305 hermon_state_t *state;
306 hermon_rsrc_t *rsrcp;
307 hermon_umap_db_entry_t *umapdb, *umapdb2;
308 minor_t instance;
309 uint64_t key, value;
310 uint_t hr_indx;
311 dev_t dev;
312 int status;
313
314 instance = HERMON_DEV_INSTANCE(*devp);
315 state = ddi_get_soft_state(hermon_statep, instance);
316 if (state == NULL) {
317 return (ENXIO);
318 }
319
320 /*
321 * Only allow driver to be opened for character access, and verify
322 * whether exclusive access is allowed.
323 */
324 if ((otyp != OTYP_CHR) || ((flag & FEXCL) &&
325 secpolicy_excl_open(credp) != 0)) {
326 return (EINVAL);
327 }
328
329 /*
330 * Search for the current process PID in the "userland resources
331 * database". If it is not found, then attempt to allocate a UAR
332 * page and add the ("key", "value") pair to the database.
333 * Note: As a last step we always return a devp appropriate for
334 * the open. Either we return a new minor number (based on the
335 * instance and the UAR page index) or we return the current minor
336 * number for the given client process.
337 *
338 * We also add an entry to the database to allow for lookup from
339 * "dev_t" to the current process PID. This is necessary because,
340 * under certain circumstance, the process PID that calls the Hermon
341 * close() entry point may not be the same as the one who called
342 * open(). Specifically, this can happen if a child process calls
343 * the Hermon's open() entry point, gets a UAR page, maps it out (using
344 * mmap()), and then exits without calling munmap(). Because mmap()
345 * adds a reference to the file descriptor, at the exit of the child
346 * process the file descriptor is "inherited" by the parent (and will
347 * be close()'d by the parent's PID only when it exits).
348 *
349 * Note: We use the hermon_umap_db_find_nolock() and
350 * hermon_umap_db_add_nolock() database access routines below (with
351 * an explicit mutex_enter of the database lock - "hdl_umapdb_lock")
352 * to ensure that the multiple accesses (in this case searching for,
353 * and then adding _two_ database entries) can be done atomically.
354 */
355 key = ddi_get_pid();
356 mutex_enter(&hermon_userland_rsrc_db.hdl_umapdb_lock);
357 status = hermon_umap_db_find_nolock(instance, key,
358 MLNX_UMAP_UARPG_RSRC, &value, 0, NULL);
359 if (status != DDI_SUCCESS) {
360 /*
361 * If we are in 'maintenance mode', we cannot alloc a UAR page.
362 * But we still need some rsrcp value, and a mostly unique
363 * hr_indx value. So we set rsrcp to NULL for maintenance
364 * mode, and use a rolling count for hr_indx. The field
365 * 'hs_open_hr_indx' is used only in this maintenance mode
366 * condition.
367 *
368 * Otherwise, if we are in operational mode then we allocate
369 * the UAR page as normal, and use the rsrcp value and tr_indx
370 * value from that allocation.
371 */
372 if (!HERMON_IS_OPERATIONAL(state->hs_operational_mode)) {
373 rsrcp = NULL;
374 hr_indx = state->hs_open_ar_indx++;
375 } else {
376 /* Allocate a new UAR page for this process */
377 status = hermon_rsrc_alloc(state, HERMON_UARPG, 1,
378 HERMON_NOSLEEP, &rsrcp);
379 if (status != DDI_SUCCESS) {
380 mutex_exit(
381 &hermon_userland_rsrc_db.hdl_umapdb_lock);
382 return (EAGAIN);
383 }
384
385 hr_indx = rsrcp->hr_indx;
386 }
387
388 /*
389 * Allocate an entry to track the UAR page resource in the
390 * "userland resources database".
391 */
392 umapdb = hermon_umap_db_alloc(instance, key,
393 MLNX_UMAP_UARPG_RSRC, (uint64_t)(uintptr_t)rsrcp);
394 if (umapdb == NULL) {
395 mutex_exit(&hermon_userland_rsrc_db.hdl_umapdb_lock);
396 /* If in "maintenance mode", don't free the rsrc */
397 if (HERMON_IS_OPERATIONAL(state->hs_operational_mode)) {
398 hermon_rsrc_free(state, &rsrcp);
399 }
400 return (EAGAIN);
401 }
402
403 /*
404 * Create a new device number. Minor number is a function of
405 * the UAR page index (15 bits) and the device instance number
406 * (3 bits).
407 */
408 dev = makedevice(getmajor(*devp), (hr_indx <<
409 HERMON_MINORNUM_SHIFT) | instance);
410
411 /*
412 * Allocate another entry in the "userland resources database"
413 * to track the association of the device number (above) to
414 * the current process ID (in "key").
415 */
416 umapdb2 = hermon_umap_db_alloc(instance, dev,
417 MLNX_UMAP_PID_RSRC, (uint64_t)key);
418 if (umapdb2 == NULL) {
419 mutex_exit(&hermon_userland_rsrc_db.hdl_umapdb_lock);
420 hermon_umap_db_free(umapdb);
421 /* If in "maintenance mode", don't free the rsrc */
422 if (HERMON_IS_OPERATIONAL(state->hs_operational_mode)) {
423 hermon_rsrc_free(state, &rsrcp);
424 }
425 return (EAGAIN);
426 }
427
428 /* Add the entries to the database */
429 hermon_umap_db_add_nolock(umapdb);
430 hermon_umap_db_add_nolock(umapdb2);
431
432 } else {
433 /*
434 * Return the same device number as on the original open()
435 * call. This was calculated as a function of the UAR page
436 * index (top 16 bits) and the device instance number
437 */
438 rsrcp = (hermon_rsrc_t *)(uintptr_t)value;
439 dev = makedevice(getmajor(*devp), (rsrcp->hr_indx <<
440 HERMON_MINORNUM_SHIFT) | instance);
441 }
442 mutex_exit(&hermon_userland_rsrc_db.hdl_umapdb_lock);
443
444 *devp = dev;
445
446 return (0);
447 }
448
449
450 /*
451 * hermon_close()
452 */
453 /* ARGSUSED */
454 static int
hermon_close(dev_t dev,int flag,int otyp,cred_t * credp)455 hermon_close(dev_t dev, int flag, int otyp, cred_t *credp)
456 {
457 hermon_state_t *state;
458 hermon_rsrc_t *rsrcp;
459 hermon_umap_db_entry_t *umapdb;
460 hermon_umap_db_priv_t *priv;
461 minor_t instance;
462 uint64_t key, value;
463 int status, reset_status = 0;
464
465 instance = HERMON_DEV_INSTANCE(dev);
466 state = ddi_get_soft_state(hermon_statep, instance);
467 if (state == NULL) {
468 return (ENXIO);
469 }
470
471 /*
472 * Search for "dev_t" in the "userland resources database". As
473 * explained above in hermon_open(), we can't depend on using the
474 * current process ID here to do the lookup because the process
475 * that ultimately closes may not be the same one who opened
476 * (because of inheritance).
477 * So we lookup the "dev_t" (which points to the PID of the process
478 * that opened), and we remove the entry from the database (and free
479 * it up). Then we do another query based on the PID value. And when
480 * we find that database entry, we free it up too and then free the
481 * Hermon UAR page resource.
482 *
483 * Note: We use the hermon_umap_db_find_nolock() database access
484 * routine below (with an explicit mutex_enter of the database lock)
485 * to ensure that the multiple accesses (which attempt to remove the
486 * two database entries) can be done atomically.
487 *
488 * This works the same in both maintenance mode and HCA mode, except
489 * for the call to hermon_rsrc_free(). In the case of maintenance mode,
490 * this call is not needed, as it was not allocated in hermon_open()
491 * above.
492 */
493 key = dev;
494 mutex_enter(&hermon_userland_rsrc_db.hdl_umapdb_lock);
495 status = hermon_umap_db_find_nolock(instance, key, MLNX_UMAP_PID_RSRC,
496 &value, HERMON_UMAP_DB_REMOVE, &umapdb);
497 if (status == DDI_SUCCESS) {
498 /*
499 * If the "hdb_priv" field is non-NULL, it indicates that
500 * some "on close" handling is still necessary. Call
501 * hermon_umap_db_handle_onclose_cb() to do the handling (i.e.
502 * to invoke all the registered callbacks). Then free up
503 * the resources associated with "hdb_priv" and continue
504 * closing.
505 */
506 priv = (hermon_umap_db_priv_t *)umapdb->hdbe_common.hdb_priv;
507 if (priv != NULL) {
508 reset_status = hermon_umap_db_handle_onclose_cb(priv);
509 kmem_free(priv, sizeof (hermon_umap_db_priv_t));
510 umapdb->hdbe_common.hdb_priv = (void *)NULL;
511 }
512
513 hermon_umap_db_free(umapdb);
514
515 /*
516 * Now do another lookup using PID as the key (copy it from
517 * "value"). When this lookup is complete, the "value" field
518 * will contain the hermon_rsrc_t pointer for the UAR page
519 * resource.
520 */
521 key = value;
522 status = hermon_umap_db_find_nolock(instance, key,
523 MLNX_UMAP_UARPG_RSRC, &value, HERMON_UMAP_DB_REMOVE,
524 &umapdb);
525 if (status == DDI_SUCCESS) {
526 hermon_umap_db_free(umapdb);
527 /* If in "maintenance mode", don't free the rsrc */
528 if (HERMON_IS_OPERATIONAL(state->hs_operational_mode)) {
529 rsrcp = (hermon_rsrc_t *)(uintptr_t)value;
530 hermon_rsrc_free(state, &rsrcp);
531 }
532 }
533 }
534 mutex_exit(&hermon_userland_rsrc_db.hdl_umapdb_lock);
535 return (reset_status);
536 }
537
538
539 /*
540 * hermon_attach()
541 * Context: Only called from attach() path context
542 */
543 static int
hermon_attach(dev_info_t * dip,ddi_attach_cmd_t cmd)544 hermon_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
545 {
546 hermon_state_t *state;
547 ibc_clnt_hdl_t tmp_ibtfpriv;
548 ibc_status_t ibc_status;
549 int instance;
550 int status;
551
552 #ifdef __lock_lint
553 (void) hermon_quiesce(dip);
554 #endif
555
556 switch (cmd) {
557 case DDI_ATTACH:
558 instance = ddi_get_instance(dip);
559 status = ddi_soft_state_zalloc(hermon_statep, instance);
560 if (status != DDI_SUCCESS) {
561 cmn_err(CE_NOTE, "hermon%d: driver failed to attach: "
562 "attach_ssz_fail", instance);
563 goto fail_attach_nomsg;
564
565 }
566 state = ddi_get_soft_state(hermon_statep, instance);
567 if (state == NULL) {
568 ddi_soft_state_free(hermon_statep, instance);
569 cmn_err(CE_NOTE, "hermon%d: driver failed to attach: "
570 "attach_gss_fail", instance);
571 goto fail_attach_nomsg;
572 }
573
574 /* clear the attach error buffer */
575 HERMON_ATTACH_MSG_INIT(state->hs_attach_buf);
576
577 /* Save away devinfo and instance before hermon_fm_init() */
578 state->hs_dip = dip;
579 state->hs_instance = instance;
580
581 hermon_fm_init(state);
582
583 /*
584 * Initialize Hermon driver and hardware.
585 *
586 * Note: If this initialization fails we may still wish to
587 * create a device node and remain operational so that Hermon
588 * firmware can be updated/flashed (i.e. "maintenance mode").
589 * If this is the case, then "hs_operational_mode" will be
590 * equal to HERMON_MAINTENANCE_MODE. We will not attempt to
591 * attach to the IBTF or register with the IBMF (i.e. no
592 * InfiniBand interfaces will be enabled).
593 */
594 status = hermon_drv_init(state, dip, instance);
595 if ((status != DDI_SUCCESS) &&
596 (HERMON_IS_OPERATIONAL(state->hs_operational_mode))) {
597 goto fail_attach;
598 }
599
600 /*
601 * Change the Hermon FM mode
602 */
603 if ((hermon_get_state(state) & HCA_PIO_FM) &&
604 HERMON_IS_OPERATIONAL(state->hs_operational_mode)) {
605 /*
606 * Now we wait for 50ms to give an opportunity
607 * to Solaris FMA so that HW errors can be notified.
608 * Then check if there are HW errors or not. If
609 * a HW error is detected, the Hermon attachment
610 * must be failed.
611 */
612 delay(drv_usectohz(50000));
613 if (hermon_init_failure(state)) {
614 hermon_drv_fini(state);
615 HERMON_WARNING(state, "unable to "
616 "attach Hermon due to a HW error");
617 HERMON_ATTACH_MSG(state->hs_attach_buf,
618 "hermon_attach_failure");
619 goto fail_attach;
620 }
621
622 /*
623 * There seems no HW errors during the attachment,
624 * so let's change the Hermon FM state to the
625 * ereport only mode.
626 */
627 if (hermon_fm_ereport_init(state) != DDI_SUCCESS) {
628 /* unwind the resources */
629 hermon_drv_fini(state);
630 HERMON_ATTACH_MSG(state->hs_attach_buf,
631 "hermon_attach_failure");
632 goto fail_attach;
633 }
634 }
635
636 /* Create the minor node for device */
637 status = ddi_create_minor_node(dip, "devctl", S_IFCHR, instance,
638 DDI_PSEUDO, 0);
639 if (status != DDI_SUCCESS) {
640 hermon_drv_fini(state);
641 HERMON_ATTACH_MSG(state->hs_attach_buf,
642 "attach_create_mn_fail");
643 goto fail_attach;
644 }
645
646 /*
647 * If we are in "maintenance mode", then we don't want to
648 * register with the IBTF. All InfiniBand interfaces are
649 * uninitialized, and the device is only capable of handling
650 * requests to update/flash firmware (or test/debug requests).
651 */
652 if (HERMON_IS_OPERATIONAL(state->hs_operational_mode)) {
653 cmn_err(CE_NOTE, "!Hermon is operational\n");
654
655 /* Attach to InfiniBand Transport Framework (IBTF) */
656 ibc_status = ibc_attach(&tmp_ibtfpriv,
657 &state->hs_ibtfinfo);
658 if (ibc_status != IBC_SUCCESS) {
659 cmn_err(CE_CONT, "hermon_attach: ibc_attach "
660 "failed\n");
661 ddi_remove_minor_node(dip, "devctl");
662 hermon_drv_fini(state);
663 HERMON_ATTACH_MSG(state->hs_attach_buf,
664 "attach_ibcattach_fail");
665 goto fail_attach;
666 }
667
668 /*
669 * Now that we've successfully attached to the IBTF,
670 * we enable all appropriate asynch and CQ events to
671 * be forwarded to the IBTF.
672 */
673 HERMON_ENABLE_IBTF_CALLB(state, tmp_ibtfpriv);
674
675 ibc_post_attach(state->hs_ibtfpriv);
676
677 /* Register agents with IB Mgmt Framework (IBMF) */
678 status = hermon_agent_handlers_init(state);
679 if (status != DDI_SUCCESS) {
680 (void) ibc_pre_detach(tmp_ibtfpriv, DDI_DETACH);
681 HERMON_QUIESCE_IBTF_CALLB(state);
682 if (state->hs_in_evcallb != 0) {
683 HERMON_WARNING(state, "unable to "
684 "quiesce Hermon IBTF callbacks");
685 }
686 ibc_detach(tmp_ibtfpriv);
687 ddi_remove_minor_node(dip, "devctl");
688 hermon_drv_fini(state);
689 HERMON_ATTACH_MSG(state->hs_attach_buf,
690 "attach_agentinit_fail");
691 goto fail_attach;
692 }
693 }
694
695 /* Report attach in maintenance mode, if appropriate */
696 if (!(HERMON_IS_OPERATIONAL(state->hs_operational_mode))) {
697 cmn_err(CE_NOTE, "hermon%d: driver attached "
698 "(for maintenance mode only)", state->hs_instance);
699 hermon_fm_ereport(state, HCA_IBA_ERR, HCA_ERR_DEGRADED);
700 }
701
702 /* Report that driver was loaded */
703 ddi_report_dev(dip);
704
705 /* Send device information to log file */
706 hermon_device_info_report(state);
707
708 /* DEBUG PRINT */
709 cmn_err(CE_CONT, "!Hermon attach complete\n");
710 return (DDI_SUCCESS);
711
712 case DDI_RESUME:
713 /* Add code here for DDI_RESUME XXX */
714 return (DDI_FAILURE);
715
716 default:
717 cmn_err(CE_WARN, "hermon_attach: unknown cmd (0x%x)\n", cmd);
718 break;
719 }
720
721 fail_attach:
722 cmn_err(CE_NOTE, "hermon%d: driver failed to attach: %s", instance,
723 state->hs_attach_buf);
724 if (hermon_get_state(state) & HCA_EREPORT_FM) {
725 hermon_fm_ereport(state, HCA_SYS_ERR, HCA_ERR_SRV_LOST);
726 }
727 hermon_drv_fini2(state);
728 hermon_fm_fini(state);
729 ddi_soft_state_free(hermon_statep, instance);
730
731 fail_attach_nomsg:
732 return (DDI_FAILURE);
733 }
734
735
736 /*
737 * hermon_detach()
738 * Context: Only called from detach() path context
739 */
740 static int
hermon_detach(dev_info_t * dip,ddi_detach_cmd_t cmd)741 hermon_detach(dev_info_t *dip, ddi_detach_cmd_t cmd)
742 {
743 hermon_state_t *state;
744 ibc_clnt_hdl_t tmp_ibtfpriv;
745 ibc_status_t ibc_status;
746 int instance, status;
747
748 instance = ddi_get_instance(dip);
749 state = ddi_get_soft_state(hermon_statep, instance);
750 if (state == NULL) {
751 return (DDI_FAILURE);
752 }
753
754 switch (cmd) {
755 case DDI_DETACH:
756 /*
757 * If we are in "maintenance mode", then we do not want to
758 * do teardown for any of the InfiniBand interfaces.
759 * Specifically, this means not detaching from IBTF (we never
760 * attached to begin with) and not deregistering from IBMF.
761 */
762 if (HERMON_IS_OPERATIONAL(state->hs_operational_mode)) {
763 /* Unregister agents from IB Mgmt Framework (IBMF) */
764 status = hermon_agent_handlers_fini(state);
765 if (status != DDI_SUCCESS) {
766 return (DDI_FAILURE);
767 }
768
769 /*
770 * Attempt the "pre-detach" from InfiniBand Transport
771 * Framework (IBTF). At this point the IBTF is still
772 * capable of handling incoming asynch and completion
773 * events. This "pre-detach" is primarily a mechanism
774 * to notify the appropriate IBTF clients that the
775 * HCA is being removed/offlined.
776 */
777 ibc_status = ibc_pre_detach(state->hs_ibtfpriv, cmd);
778 if (ibc_status != IBC_SUCCESS) {
779 status = hermon_agent_handlers_init(state);
780 if (status != DDI_SUCCESS) {
781 HERMON_WARNING(state, "failed to "
782 "restart Hermon agents");
783 }
784 return (DDI_FAILURE);
785 }
786
787 /*
788 * Before we can fully detach from the IBTF we need to
789 * ensure that we have handled all outstanding event
790 * callbacks. This is accomplished by quiescing the
791 * event callback mechanism. Note: if we are unable
792 * to successfully quiesce the callbacks, then this is
793 * an indication that something has probably gone
794 * seriously wrong. We print out a warning, but
795 * continue.
796 */
797 tmp_ibtfpriv = state->hs_ibtfpriv;
798 HERMON_QUIESCE_IBTF_CALLB(state);
799 if (state->hs_in_evcallb != 0) {
800 HERMON_WARNING(state, "unable to quiesce "
801 "Hermon IBTF callbacks");
802 }
803
804 /* Complete the detach from the IBTF */
805 ibc_detach(tmp_ibtfpriv);
806 }
807
808 /* Remove the minor node for device */
809 ddi_remove_minor_node(dip, "devctl");
810
811 /*
812 * Only call hermon_drv_fini() if we are in Hermon HCA mode.
813 * (Because if we are in "maintenance mode", then we never
814 * successfully finished init.) Only report successful
815 * detach for normal HCA mode.
816 */
817 if (HERMON_IS_OPERATIONAL(state->hs_operational_mode)) {
818 /* Cleanup driver resources and shutdown hardware */
819 hermon_drv_fini(state);
820 cmn_err(CE_CONT, "!Hermon driver successfully "
821 "detached\n");
822 }
823
824 hermon_drv_fini2(state);
825 hermon_fm_fini(state);
826 ddi_soft_state_free(hermon_statep, instance);
827
828 return (DDI_SUCCESS);
829
830 case DDI_SUSPEND:
831 /* Add code here for DDI_SUSPEND XXX */
832 return (DDI_FAILURE);
833
834 default:
835 cmn_err(CE_WARN, "hermon_detach: unknown cmd (0x%x)\n", cmd);
836 break;
837 }
838
839 return (DDI_FAILURE);
840 }
841
842 /*
843 * hermon_dma_attr_init()
844 * Context: Can be called from interrupt or base context.
845 */
846
847 /* ARGSUSED */
848 void
hermon_dma_attr_init(hermon_state_t * state,ddi_dma_attr_t * dma_attr)849 hermon_dma_attr_init(hermon_state_t *state, ddi_dma_attr_t *dma_attr)
850 {
851 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*dma_attr))
852
853 dma_attr->dma_attr_version = DMA_ATTR_V0;
854 dma_attr->dma_attr_addr_lo = 0;
855 dma_attr->dma_attr_addr_hi = 0xFFFFFFFFFFFFFFFFull;
856 dma_attr->dma_attr_count_max = 0xFFFFFFFFFFFFFFFFull;
857 dma_attr->dma_attr_align = HERMON_PAGESIZE; /* default 4K */
858 dma_attr->dma_attr_burstsizes = 0x3FF;
859 dma_attr->dma_attr_minxfer = 1;
860 dma_attr->dma_attr_maxxfer = 0xFFFFFFFFFFFFFFFFull;
861 dma_attr->dma_attr_seg = 0xFFFFFFFFFFFFFFFFull;
862 dma_attr->dma_attr_sgllen = 0x7FFFFFFF;
863 dma_attr->dma_attr_granular = 1;
864 dma_attr->dma_attr_flags = 0;
865 }
866
867 /*
868 * hermon_dma_alloc()
869 * Context: Can be called from base context.
870 */
871 int
hermon_dma_alloc(hermon_state_t * state,hermon_dma_info_t * dma_info,uint16_t opcode)872 hermon_dma_alloc(hermon_state_t *state, hermon_dma_info_t *dma_info,
873 uint16_t opcode)
874 {
875 ddi_dma_handle_t dma_hdl;
876 ddi_dma_attr_t dma_attr;
877 ddi_acc_handle_t acc_hdl;
878 ddi_dma_cookie_t cookie;
879 uint64_t kaddr;
880 uint64_t real_len;
881 uint_t ccount;
882 int status;
883
884 hermon_dma_attr_init(state, &dma_attr);
885 #ifdef __sparc
886 if (state->hs_cfg_profile->cp_iommu_bypass == HERMON_BINDMEM_BYPASS)
887 dma_attr.dma_attr_flags = DDI_DMA_FORCE_PHYSICAL;
888 #endif
889
890 /* Allocate a DMA handle */
891 status = ddi_dma_alloc_handle(state->hs_dip, &dma_attr, DDI_DMA_SLEEP,
892 NULL, &dma_hdl);
893 if (status != DDI_SUCCESS) {
894 IBTF_DPRINTF_L2("DMA", "alloc handle failed: %d", status);
895 cmn_err(CE_CONT, "DMA alloc handle failed(status %d)", status);
896 return (DDI_FAILURE);
897 }
898
899 /* Allocate DMA memory */
900 status = ddi_dma_mem_alloc(dma_hdl, dma_info->length,
901 &state->hs_reg_accattr, DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, NULL,
902 (caddr_t *)&kaddr, (size_t *)&real_len, &acc_hdl);
903 if (status != DDI_SUCCESS) {
904 ddi_dma_free_handle(&dma_hdl);
905 IBTF_DPRINTF_L2("DMA", "memory alloc failed: %d", status);
906 cmn_err(CE_CONT, "DMA memory alloc failed(status %d)", status);
907 return (DDI_FAILURE);
908 }
909 bzero((caddr_t)(uintptr_t)kaddr, real_len);
910
911 /* Bind the memory to the handle */
912 status = ddi_dma_addr_bind_handle(dma_hdl, NULL,
913 (caddr_t)(uintptr_t)kaddr, (size_t)real_len, DDI_DMA_RDWR |
914 DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, NULL, &cookie, &ccount);
915 if (status != DDI_SUCCESS) {
916 ddi_dma_mem_free(&acc_hdl);
917 ddi_dma_free_handle(&dma_hdl);
918 IBTF_DPRINTF_L2("DMA", "bind handle failed: %d", status);
919 cmn_err(CE_CONT, "DMA bind handle failed(status %d)", status);
920 return (DDI_FAILURE);
921 }
922
923 /* Package the hermon_dma_info contents and return */
924 dma_info->vaddr = kaddr;
925 dma_info->dma_hdl = dma_hdl;
926 dma_info->acc_hdl = acc_hdl;
927
928 /* Pass the mapping information to the firmware */
929 status = hermon_map_cmd_post(state, dma_info, opcode, cookie, ccount);
930 if (status != DDI_SUCCESS) {
931 char *s;
932 hermon_dma_free(dma_info);
933 switch (opcode) {
934 case MAP_ICM:
935 s = "MAP_ICM";
936 break;
937 case MAP_FA:
938 s = "MAP_FA";
939 break;
940 case MAP_ICM_AUX:
941 s = "MAP_ICM_AUX";
942 break;
943 default:
944 s = "UNKNOWN";
945 }
946 cmn_err(CE_NOTE, "Map cmd '%s' failed, status %08x\n",
947 s, status);
948 return (DDI_FAILURE);
949 }
950
951 return (DDI_SUCCESS);
952 }
953
954 /*
955 * hermon_dma_free()
956 * Context: Can be called from base context.
957 */
958 void
hermon_dma_free(hermon_dma_info_t * info)959 hermon_dma_free(hermon_dma_info_t *info)
960 {
961 /* Unbind the handles and free the memory */
962 (void) ddi_dma_unbind_handle(info->dma_hdl);
963 ddi_dma_mem_free(&info->acc_hdl);
964 ddi_dma_free_handle(&info->dma_hdl);
965 }
966
967 /* These macros are valid for use only in hermon_icm_alloc/hermon_icm_free. */
968 #define HERMON_ICM_ALLOC(rsrc) \
969 hermon_icm_alloc(state, rsrc, index1, index2)
970 #define HERMON_ICM_FREE(rsrc) \
971 hermon_icm_free(state, rsrc, index1, index2)
972
973 /*
974 * hermon_icm_alloc()
975 * Context: Can be called from base context.
976 *
977 * Only one thread can be here for a given hermon_rsrc_type_t "type".
978 *
979 * "num_to_hdl" is set if there is a need for lookups from resource
980 * number/index to resource handle. This is needed for QPs/CQs/SRQs
981 * for the various affiliated events/errors.
982 */
983 int
hermon_icm_alloc(hermon_state_t * state,hermon_rsrc_type_t type,uint32_t index1,uint32_t index2)984 hermon_icm_alloc(hermon_state_t *state, hermon_rsrc_type_t type,
985 uint32_t index1, uint32_t index2)
986 {
987 hermon_icm_table_t *icm;
988 hermon_dma_info_t *dma_info;
989 uint8_t *bitmap;
990 int status;
991 int num_to_hdl = 0;
992
993 if (hermon_verbose) {
994 IBTF_DPRINTF_L2("hermon", "hermon_icm_alloc: rsrc_type (0x%x) "
995 "index1/2 (0x%x/0x%x)", type, index1, index2);
996 }
997
998 icm = &state->hs_icm[type];
999
1000 switch (type) {
1001 case HERMON_QPC:
1002 status = HERMON_ICM_ALLOC(HERMON_CMPT_QPC);
1003 if (status != DDI_SUCCESS) {
1004 return (status);
1005 }
1006 status = HERMON_ICM_ALLOC(HERMON_RDB);
1007 if (status != DDI_SUCCESS) { /* undo icm_alloc's */
1008 HERMON_ICM_FREE(HERMON_CMPT_QPC);
1009 return (status);
1010 }
1011 status = HERMON_ICM_ALLOC(HERMON_ALTC);
1012 if (status != DDI_SUCCESS) { /* undo icm_alloc's */
1013 HERMON_ICM_FREE(HERMON_RDB);
1014 HERMON_ICM_FREE(HERMON_CMPT_QPC);
1015 return (status);
1016 }
1017 status = HERMON_ICM_ALLOC(HERMON_AUXC);
1018 if (status != DDI_SUCCESS) { /* undo icm_alloc's */
1019 HERMON_ICM_FREE(HERMON_ALTC);
1020 HERMON_ICM_FREE(HERMON_RDB);
1021 HERMON_ICM_FREE(HERMON_CMPT_QPC);
1022 return (status);
1023 }
1024 num_to_hdl = 1;
1025 break;
1026 case HERMON_SRQC:
1027 status = HERMON_ICM_ALLOC(HERMON_CMPT_SRQC);
1028 if (status != DDI_SUCCESS) {
1029 return (status);
1030 }
1031 num_to_hdl = 1;
1032 break;
1033 case HERMON_CQC:
1034 status = HERMON_ICM_ALLOC(HERMON_CMPT_CQC);
1035 if (status != DDI_SUCCESS) {
1036 return (status);
1037 }
1038 num_to_hdl = 1;
1039 break;
1040 case HERMON_EQC:
1041 status = HERMON_ICM_ALLOC(HERMON_CMPT_EQC);
1042 if (status != DDI_SUCCESS) { /* undo icm_alloc's */
1043 return (status);
1044 }
1045 break;
1046 }
1047
1048 /* ensure existence of bitmap and dmainfo, sets "dma_info" */
1049 hermon_bitmap(bitmap, dma_info, icm, index1, num_to_hdl);
1050
1051 /* Set up the DMA handle for allocation and mapping */
1052 dma_info += index2;
1053 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*dma_info))
1054 dma_info->length = icm->span << icm->log_object_size;
1055 dma_info->icmaddr = icm->icm_baseaddr +
1056 (((index1 << icm->split_shift) +
1057 (index2 << icm->span_shift)) << icm->log_object_size);
1058
1059 /* Allocate memory for the num_to_qp/cq/srq pointers */
1060 if (num_to_hdl)
1061 icm->num_to_hdl[index1][index2] =
1062 kmem_zalloc(HERMON_ICM_SPAN * sizeof (void *), KM_SLEEP);
1063
1064 if (hermon_verbose) {
1065 IBTF_DPRINTF_L2("hermon", "alloc DMA: "
1066 "rsrc (0x%x) index (%x, %x) "
1067 "icm_addr/len (%llx/%x) bitmap %p", type, index1, index2,
1068 (longlong_t)dma_info->icmaddr, dma_info->length, bitmap);
1069 }
1070
1071 /* Allocate and map memory for this span */
1072 status = hermon_dma_alloc(state, dma_info, MAP_ICM);
1073 if (status != DDI_SUCCESS) {
1074 IBTF_DPRINTF_L2("hermon", "hermon_icm_alloc: DMA "
1075 "allocation failed, status 0x%x", status);
1076 switch (type) {
1077 case HERMON_QPC:
1078 HERMON_ICM_FREE(HERMON_AUXC);
1079 HERMON_ICM_FREE(HERMON_ALTC);
1080 HERMON_ICM_FREE(HERMON_RDB);
1081 HERMON_ICM_FREE(HERMON_CMPT_QPC);
1082 break;
1083 case HERMON_SRQC:
1084 HERMON_ICM_FREE(HERMON_CMPT_SRQC);
1085 break;
1086 case HERMON_CQC:
1087 HERMON_ICM_FREE(HERMON_CMPT_CQC);
1088 break;
1089 case HERMON_EQC:
1090 HERMON_ICM_FREE(HERMON_CMPT_EQC);
1091 break;
1092 }
1093
1094 return (DDI_FAILURE);
1095 }
1096 if (hermon_verbose) {
1097 IBTF_DPRINTF_L2("hermon", "hermon_icm_alloc: mapping ICM: "
1098 "rsrc_type (0x%x) index (0x%x, 0x%x) alloc length (0x%x) "
1099 "icm_addr (0x%lx)", type, index1, index2, dma_info->length,
1100 dma_info->icmaddr);
1101 }
1102
1103 /* Set the bit for this slot in the table bitmap */
1104 HERMON_BMAP_BIT_SET(icm->icm_bitmap[index1], index2);
1105
1106 return (DDI_SUCCESS);
1107 }
1108
1109 /*
1110 * hermon_icm_free()
1111 * Context: Can be called from base context.
1112 *
1113 * ICM resources have been successfully returned from hermon_icm_alloc().
1114 * Associated dma_info is no longer in use. Free the ICM backing memory.
1115 */
1116 void
hermon_icm_free(hermon_state_t * state,hermon_rsrc_type_t type,uint32_t index1,uint32_t index2)1117 hermon_icm_free(hermon_state_t *state, hermon_rsrc_type_t type,
1118 uint32_t index1, uint32_t index2)
1119 {
1120 hermon_icm_table_t *icm;
1121 hermon_dma_info_t *dma_info;
1122 int status;
1123
1124 icm = &state->hs_icm[type];
1125 ASSERT(icm->icm_dma[index1][index2].icm_refcnt == 0);
1126
1127 if (hermon_verbose) {
1128 IBTF_DPRINTF_L2("hermon", "hermon_icm_free: rsrc_type (0x%x) "
1129 "index (0x%x, 0x%x)", type, index1, index2);
1130 }
1131
1132 dma_info = icm->icm_dma[index1] + index2;
1133
1134 /* The following only happens if attach() is failing. */
1135 if (dma_info == NULL)
1136 return;
1137
1138 /* Unmap the ICM allocation, then free the backing DMA memory */
1139 status = hermon_unmap_icm_cmd_post(state, dma_info);
1140 if (status != DDI_SUCCESS) {
1141 HERMON_WARNING(state, "UNMAP_ICM failure");
1142 }
1143 hermon_dma_free(dma_info);
1144
1145 /* Clear the bit in the ICM table bitmap */
1146 HERMON_BMAP_BIT_CLR(icm->icm_bitmap[index1], index2);
1147
1148 switch (type) {
1149 case HERMON_QPC:
1150 HERMON_ICM_FREE(HERMON_AUXC);
1151 HERMON_ICM_FREE(HERMON_ALTC);
1152 HERMON_ICM_FREE(HERMON_RDB);
1153 HERMON_ICM_FREE(HERMON_CMPT_QPC);
1154 break;
1155 case HERMON_SRQC:
1156 HERMON_ICM_FREE(HERMON_CMPT_SRQC);
1157 break;
1158 case HERMON_CQC:
1159 HERMON_ICM_FREE(HERMON_CMPT_CQC);
1160 break;
1161 case HERMON_EQC:
1162 HERMON_ICM_FREE(HERMON_CMPT_EQC);
1163 break;
1164
1165 }
1166 }
1167
1168
1169 /*
1170 * hermon_icm_num_to_hdl()
1171 * Context: Can be called from base or interrupt context.
1172 *
1173 * Given an index of a resource, index through the sparsely allocated
1174 * arrays to find the pointer to its software handle. Return NULL if
1175 * any of the arrays of pointers has been freed (should never happen).
1176 */
1177 void *
hermon_icm_num_to_hdl(hermon_state_t * state,hermon_rsrc_type_t type,uint32_t idx)1178 hermon_icm_num_to_hdl(hermon_state_t *state, hermon_rsrc_type_t type,
1179 uint32_t idx)
1180 {
1181 hermon_icm_table_t *icm;
1182 uint32_t span_offset;
1183 uint32_t index1, index2;
1184 void ***p1, **p2;
1185
1186 icm = &state->hs_icm[type];
1187 hermon_index(index1, index2, idx, icm, span_offset);
1188 p1 = icm->num_to_hdl[index1];
1189 if (p1 == NULL) {
1190 IBTF_DPRINTF_L2("hermon", "icm_num_to_hdl failed at level 1"
1191 ": rsrc_type %d, index 0x%x", type, idx);
1192 return (NULL);
1193 }
1194 p2 = p1[index2];
1195 if (p2 == NULL) {
1196 IBTF_DPRINTF_L2("hermon", "icm_num_to_hdl failed at level 2"
1197 ": rsrc_type %d, index 0x%x", type, idx);
1198 return (NULL);
1199 }
1200 return (p2[span_offset]);
1201 }
1202
1203 /*
1204 * hermon_icm_set_num_to_hdl()
1205 * Context: Can be called from base or interrupt context.
1206 *
1207 * Given an index of a resource, we index through the sparsely allocated
1208 * arrays to store the software handle, used by hermon_icm_num_to_hdl().
1209 * This function is used to both set and reset (set to NULL) the handle.
1210 * This table is allocated during ICM allocation for the given resource,
1211 * so its existence is a given, and the store location does not conflict
1212 * with any other stores to the table (no locking needed).
1213 */
1214 void
hermon_icm_set_num_to_hdl(hermon_state_t * state,hermon_rsrc_type_t type,uint32_t idx,void * hdl)1215 hermon_icm_set_num_to_hdl(hermon_state_t *state, hermon_rsrc_type_t type,
1216 uint32_t idx, void *hdl)
1217 {
1218 hermon_icm_table_t *icm;
1219 uint32_t span_offset;
1220 uint32_t index1, index2;
1221
1222 icm = &state->hs_icm[type];
1223 hermon_index(index1, index2, idx, icm, span_offset);
1224 ASSERT((hdl == NULL) ^
1225 (icm->num_to_hdl[index1][index2][span_offset] == NULL));
1226 icm->num_to_hdl[index1][index2][span_offset] = hdl;
1227 }
1228
1229 /*
1230 * hermon_device_mode()
1231 * Context: Can be called from base or interrupt context.
1232 *
1233 * Return HERMON_HCA_MODE for operational mode
1234 * Return HERMON_MAINTENANCE_MODE for maintenance mode
1235 * Return 0 otherwise
1236 *
1237 * A non-zero return for either operational or maintenance mode simplifies
1238 * one of the 2 uses of this function.
1239 */
1240 int
hermon_device_mode(hermon_state_t * state)1241 hermon_device_mode(hermon_state_t *state)
1242 {
1243 if (state->hs_vendor_id != PCI_VENID_MLX)
1244 return (0);
1245
1246 switch (state->hs_device_id) {
1247 case PCI_DEVID_HERMON_SDR:
1248 case PCI_DEVID_HERMON_DDR:
1249 case PCI_DEVID_HERMON_DDRG2:
1250 case PCI_DEVID_HERMON_QDRG2:
1251 case PCI_DEVID_HERMON_QDRG2V:
1252 return (HERMON_HCA_MODE);
1253 case PCI_DEVID_HERMON_MAINT:
1254 return (HERMON_MAINTENANCE_MODE);
1255 default:
1256 return (0);
1257 }
1258 }
1259
1260 /*
1261 * hermon_drv_init()
1262 * Context: Only called from attach() path context
1263 */
1264 /* ARGSUSED */
1265 static int
hermon_drv_init(hermon_state_t * state,dev_info_t * dip,int instance)1266 hermon_drv_init(hermon_state_t *state, dev_info_t *dip, int instance)
1267 {
1268 int status;
1269
1270 /* Retrieve PCI device, vendor and rev IDs */
1271 state->hs_vendor_id = HERMON_GET_VENDOR_ID(state->hs_dip);
1272 state->hs_device_id = HERMON_GET_DEVICE_ID(state->hs_dip);
1273 state->hs_revision_id = HERMON_GET_REVISION_ID(state->hs_dip);
1274
1275 /*
1276 * Check and set the operational mode of the device. If the driver is
1277 * bound to the Hermon device in "maintenance mode", then this generally
1278 * means that either the device has been specifically jumpered to
1279 * start in this mode or the firmware boot process has failed to
1280 * successfully load either the primary or the secondary firmware
1281 * image.
1282 */
1283 state->hs_operational_mode = hermon_device_mode(state);
1284 switch (state->hs_operational_mode) {
1285 case HERMON_HCA_MODE:
1286 state->hs_cfg_profile_setting = HERMON_CFG_MEMFREE;
1287 break;
1288 case HERMON_MAINTENANCE_MODE:
1289 HERMON_FMANOTE(state, HERMON_FMA_MAINT);
1290 state->hs_fm_degraded_reason = HCA_FW_MISC; /* not fw reason */
1291 return (DDI_FAILURE);
1292 default:
1293 HERMON_FMANOTE(state, HERMON_FMA_PCIID);
1294 HERMON_WARNING(state, "unexpected device type detected");
1295 return (DDI_FAILURE);
1296 }
1297
1298 /*
1299 * Initialize the Hermon hardware.
1300 *
1301 * Note: If this routine returns an error, it is often a reasonably
1302 * good indication that something Hermon firmware-related has caused
1303 * the failure or some HW related errors have caused the failure.
1304 * (also there are few possibilities that SW (e.g. SW resource
1305 * shortage) can cause the failure, but the majority case is due to
1306 * either a firmware related error or a HW related one) In order to
1307 * give the user an opportunity (if desired) to update or reflash
1308 * the Hermon firmware image, we set "hs_operational_mode" flag
1309 * (described above) to indicate that we wish to enter maintenance
1310 * mode in case of the firmware-related issue.
1311 */
1312 status = hermon_hw_init(state);
1313 if (status != DDI_SUCCESS) {
1314 cmn_err(CE_NOTE, "hermon%d: error during attach: %s", instance,
1315 state->hs_attach_buf);
1316 return (DDI_FAILURE);
1317 }
1318
1319 /*
1320 * Now that the ISR has been setup, arm all the EQs for event
1321 * generation.
1322 */
1323
1324 status = hermon_eq_arm_all(state);
1325 if (status != DDI_SUCCESS) {
1326 cmn_err(CE_NOTE, "EQ Arm All failed\n");
1327 hermon_hw_fini(state, HERMON_DRV_CLEANUP_ALL);
1328 return (DDI_FAILURE);
1329 }
1330
1331 /* test interrupts and event queues */
1332 status = hermon_nop_post(state, 0x0, 0x0);
1333 if (status != DDI_SUCCESS) {
1334 cmn_err(CE_NOTE, "Interrupts/EQs failed\n");
1335 hermon_hw_fini(state, HERMON_DRV_CLEANUP_ALL);
1336 return (DDI_FAILURE);
1337 }
1338
1339 /* Initialize Hermon softstate */
1340 status = hermon_soft_state_init(state);
1341 if (status != DDI_SUCCESS) {
1342 cmn_err(CE_NOTE, "Failed to init soft state\n");
1343 hermon_hw_fini(state, HERMON_DRV_CLEANUP_ALL);
1344 return (DDI_FAILURE);
1345 }
1346
1347 return (DDI_SUCCESS);
1348 }
1349
1350
1351 /*
1352 * hermon_drv_fini()
1353 * Context: Only called from attach() and/or detach() path contexts
1354 */
1355 static void
hermon_drv_fini(hermon_state_t * state)1356 hermon_drv_fini(hermon_state_t *state)
1357 {
1358 /* Cleanup Hermon softstate */
1359 hermon_soft_state_fini(state);
1360
1361 /* Cleanup Hermon resources and shutdown hardware */
1362 hermon_hw_fini(state, HERMON_DRV_CLEANUP_ALL);
1363 }
1364
1365
1366 /*
1367 * hermon_drv_fini2()
1368 * Context: Only called from attach() and/or detach() path contexts
1369 */
1370 static void
hermon_drv_fini2(hermon_state_t * state)1371 hermon_drv_fini2(hermon_state_t *state)
1372 {
1373 if (state->hs_fm_poll_thread) {
1374 ddi_periodic_delete(state->hs_fm_poll_thread);
1375 state->hs_fm_poll_thread = NULL;
1376 }
1377
1378 /* HERMON_DRV_CLEANUP_LEVEL1 */
1379 if (state->hs_fm_cmdhdl) {
1380 hermon_regs_map_free(state, &state->hs_fm_cmdhdl);
1381 state->hs_fm_cmdhdl = NULL;
1382 }
1383
1384 if (state->hs_reg_cmdhdl) {
1385 ddi_regs_map_free(&state->hs_reg_cmdhdl);
1386 state->hs_reg_cmdhdl = NULL;
1387 }
1388
1389 /* HERMON_DRV_CLEANUP_LEVEL0 */
1390 if (state->hs_msix_tbl_entries) {
1391 kmem_free(state->hs_msix_tbl_entries,
1392 state->hs_msix_tbl_size);
1393 state->hs_msix_tbl_entries = NULL;
1394 }
1395
1396 if (state->hs_msix_pba_entries) {
1397 kmem_free(state->hs_msix_pba_entries,
1398 state->hs_msix_pba_size);
1399 state->hs_msix_pba_entries = NULL;
1400 }
1401
1402 if (state->hs_fm_msix_tblhdl) {
1403 hermon_regs_map_free(state, &state->hs_fm_msix_tblhdl);
1404 state->hs_fm_msix_tblhdl = NULL;
1405 }
1406
1407 if (state->hs_reg_msix_tblhdl) {
1408 ddi_regs_map_free(&state->hs_reg_msix_tblhdl);
1409 state->hs_reg_msix_tblhdl = NULL;
1410 }
1411
1412 if (state->hs_fm_msix_pbahdl) {
1413 hermon_regs_map_free(state, &state->hs_fm_msix_pbahdl);
1414 state->hs_fm_msix_pbahdl = NULL;
1415 }
1416
1417 if (state->hs_reg_msix_pbahdl) {
1418 ddi_regs_map_free(&state->hs_reg_msix_pbahdl);
1419 state->hs_reg_msix_pbahdl = NULL;
1420 }
1421
1422 if (state->hs_fm_pcihdl) {
1423 hermon_pci_config_teardown(state, &state->hs_fm_pcihdl);
1424 state->hs_fm_pcihdl = NULL;
1425 }
1426
1427 if (state->hs_reg_pcihdl) {
1428 pci_config_teardown(&state->hs_reg_pcihdl);
1429 state->hs_reg_pcihdl = NULL;
1430 }
1431 }
1432
1433
1434 /*
1435 * hermon_isr_init()
1436 * Context: Only called from attach() path context
1437 */
1438 static int
hermon_isr_init(hermon_state_t * state)1439 hermon_isr_init(hermon_state_t *state)
1440 {
1441 int status;
1442 int intr;
1443
1444 for (intr = 0; intr < state->hs_intrmsi_allocd; intr++) {
1445
1446 /*
1447 * Add a handler for the interrupt or MSI
1448 */
1449 status = ddi_intr_add_handler(state->hs_intrmsi_hdl[intr],
1450 hermon_isr, (caddr_t)state, (void *)(uintptr_t)intr);
1451 if (status != DDI_SUCCESS) {
1452 return (DDI_FAILURE);
1453 }
1454
1455 /*
1456 * Enable the software interrupt. Note: depending on the value
1457 * returned in the capability flag, we have to call either
1458 * ddi_intr_block_enable() or ddi_intr_enable().
1459 */
1460 if (state->hs_intrmsi_cap & DDI_INTR_FLAG_BLOCK) {
1461 status = ddi_intr_block_enable(
1462 &state->hs_intrmsi_hdl[intr], 1);
1463 if (status != DDI_SUCCESS) {
1464 return (DDI_FAILURE);
1465 }
1466 } else {
1467 status = ddi_intr_enable(state->hs_intrmsi_hdl[intr]);
1468 if (status != DDI_SUCCESS) {
1469 return (DDI_FAILURE);
1470 }
1471 }
1472 }
1473
1474 /*
1475 * Now that the ISR has been enabled, defer arm_all EQs for event
1476 * generation until later, in case MSIX is enabled
1477 */
1478 return (DDI_SUCCESS);
1479 }
1480
1481
1482 /*
1483 * hermon_isr_fini()
1484 * Context: Only called from attach() and/or detach() path contexts
1485 */
1486 static void
hermon_isr_fini(hermon_state_t * state)1487 hermon_isr_fini(hermon_state_t *state)
1488 {
1489 int intr;
1490
1491 for (intr = 0; intr < state->hs_intrmsi_allocd; intr++) {
1492 /* Disable the software interrupt */
1493 if (state->hs_intrmsi_cap & DDI_INTR_FLAG_BLOCK) {
1494 (void) ddi_intr_block_disable(
1495 &state->hs_intrmsi_hdl[intr], 1);
1496 } else {
1497 (void) ddi_intr_disable(state->hs_intrmsi_hdl[intr]);
1498 }
1499
1500 /*
1501 * Remove the software handler for the interrupt or MSI
1502 */
1503 (void) ddi_intr_remove_handler(state->hs_intrmsi_hdl[intr]);
1504 }
1505 }
1506
1507
1508 /*
1509 * Sum of ICM configured values:
1510 * cMPT, dMPT, MTT, QPC, SRQC, RDB, CQC, ALTC, AUXC, EQC, MCG
1511 *
1512 */
1513 static uint64_t
hermon_size_icm(hermon_state_t * state)1514 hermon_size_icm(hermon_state_t *state)
1515 {
1516 hermon_hw_querydevlim_t *devlim;
1517 hermon_cfg_profile_t *cfg;
1518 uint64_t num_cmpts, num_dmpts, num_mtts;
1519 uint64_t num_qpcs, num_srqc, num_rdbs;
1520 #ifndef HERMON_FW_WORKAROUND
1521 uint64_t num_auxc;
1522 #endif
1523 uint64_t num_cqcs, num_altc;
1524 uint64_t num_eqcs, num_mcgs;
1525 uint64_t size;
1526
1527 devlim = &state->hs_devlim;
1528 cfg = state->hs_cfg_profile;
1529 /* number of respective entries */
1530 num_cmpts = (uint64_t)0x1 << cfg->cp_log_num_cmpt;
1531 num_mtts = (uint64_t)0x1 << cfg->cp_log_num_mtt;
1532 num_dmpts = (uint64_t)0x1 << cfg->cp_log_num_dmpt;
1533 num_qpcs = (uint64_t)0x1 << cfg->cp_log_num_qp;
1534 num_srqc = (uint64_t)0x1 << cfg->cp_log_num_srq;
1535 num_rdbs = (uint64_t)0x1 << cfg->cp_log_num_rdb;
1536 num_cqcs = (uint64_t)0x1 << cfg->cp_log_num_cq;
1537 num_altc = (uint64_t)0x1 << cfg->cp_log_num_qp;
1538 #ifndef HERMON_FW_WORKAROUND
1539 num_auxc = (uint64_t)0x1 << cfg->cp_log_num_qp;
1540 #endif
1541 num_eqcs = (uint64_t)0x1 << cfg->cp_log_num_eq;
1542 num_mcgs = (uint64_t)0x1 << cfg->cp_log_num_mcg;
1543
1544 size =
1545 num_cmpts * devlim->cmpt_entry_sz +
1546 num_dmpts * devlim->dmpt_entry_sz +
1547 num_mtts * devlim->mtt_entry_sz +
1548 num_qpcs * devlim->qpc_entry_sz +
1549 num_srqc * devlim->srq_entry_sz +
1550 num_rdbs * devlim->rdmardc_entry_sz +
1551 num_cqcs * devlim->cqc_entry_sz +
1552 num_altc * devlim->altc_entry_sz +
1553 #ifdef HERMON_FW_WORKAROUND
1554 0x80000000ull +
1555 #else
1556 num_auxc * devlim->aux_entry_sz +
1557 #endif
1558 num_eqcs * devlim->eqc_entry_sz +
1559 num_mcgs * HERMON_MCGMEM_SZ(state);
1560 return (size);
1561 }
1562
1563
1564 /*
1565 * hermon_hw_init()
1566 * Context: Only called from attach() path context
1567 */
1568 static int
hermon_hw_init(hermon_state_t * state)1569 hermon_hw_init(hermon_state_t *state)
1570 {
1571 hermon_drv_cleanup_level_t cleanup;
1572 sm_nodeinfo_t nodeinfo;
1573 uint64_t clr_intr_offset;
1574 int status;
1575 uint32_t fw_size; /* in page */
1576 uint64_t offset;
1577
1578 /* This is where driver initialization begins */
1579 cleanup = HERMON_DRV_CLEANUP_LEVEL0;
1580
1581 /* Setup device access attributes */
1582 state->hs_reg_accattr.devacc_attr_version = DDI_DEVICE_ATTR_V1;
1583 state->hs_reg_accattr.devacc_attr_endian_flags = DDI_STRUCTURE_BE_ACC;
1584 state->hs_reg_accattr.devacc_attr_dataorder = DDI_STRICTORDER_ACC;
1585 state->hs_reg_accattr.devacc_attr_access = DDI_DEFAULT_ACC;
1586
1587 /* Setup fma-protected access attributes */
1588 state->hs_fm_accattr.devacc_attr_version =
1589 hermon_devacc_attr_version(state);
1590 state->hs_fm_accattr.devacc_attr_endian_flags = DDI_STRUCTURE_BE_ACC;
1591 state->hs_fm_accattr.devacc_attr_dataorder = DDI_STRICTORDER_ACC;
1592 /* set acc err protection type */
1593 state->hs_fm_accattr.devacc_attr_access =
1594 hermon_devacc_attr_access(state);
1595
1596 /* Setup for PCI config read/write of HCA device */
1597 status = hermon_pci_config_setup(state, &state->hs_fm_pcihdl);
1598 if (status != DDI_SUCCESS) {
1599 hermon_hw_fini(state, cleanup);
1600 HERMON_ATTACH_MSG(state->hs_attach_buf,
1601 "hw_init_PCI_config_space_regmap_fail");
1602 /* This case is not the degraded one */
1603 return (DDI_FAILURE);
1604 }
1605
1606 /* Map PCI config space and MSI-X tables/pba */
1607 hermon_set_msix_info(state);
1608
1609 /* Map in Hermon registers (CMD, UAR, MSIX) and setup offsets */
1610 status = hermon_regs_map_setup(state, HERMON_CMD_BAR,
1611 &state->hs_reg_cmd_baseaddr, 0, 0, &state->hs_fm_accattr,
1612 &state->hs_fm_cmdhdl);
1613 if (status != DDI_SUCCESS) {
1614 hermon_hw_fini(state, cleanup);
1615 HERMON_ATTACH_MSG(state->hs_attach_buf,
1616 "hw_init_CMD_BAR_regmap_fail");
1617 /* This case is not the degraded one */
1618 return (DDI_FAILURE);
1619 }
1620
1621 cleanup = HERMON_DRV_CLEANUP_LEVEL1;
1622 /*
1623 * We defer UAR-BAR mapping until later. Need to know if
1624 * blueflame mapping is to be done, and don't know that until after
1625 * we get the dev_caps, so do it right after that
1626 */
1627
1628 /*
1629 * There is a third BAR defined for Hermon - it is for MSIX
1630 *
1631 * Will need to explore it's possible need/use w/ Mellanox
1632 * [es] Temporary mapping maybe
1633 */
1634
1635 #ifdef HERMON_SUPPORTS_MSIX_BAR
1636 status = ddi_regs_map_setup(state->hs_dip, HERMON_MSIX_BAR,
1637 &state->hs_reg_msi_baseaddr, 0, 0, &state->hs_reg_accattr,
1638 &state->hs_reg_msihdl);
1639 if (status != DDI_SUCCESS) {
1640 hermon_hw_fini(state, cleanup);
1641 HERMON_ATTACH_MSG(state->hs_attach_buf,
1642 "hw_init_MSIX_BAR_regmap_fail");
1643 /* This case is not the degraded one */
1644 return (DDI_FAILURE);
1645 }
1646 #endif
1647
1648 cleanup = HERMON_DRV_CLEANUP_LEVEL2;
1649
1650 /*
1651 * Save interesting registers away. The offsets of the first two
1652 * here (HCR and sw_reset) are detailed in the PRM, the others are
1653 * derived from values in the QUERY_FW output, so we'll save them
1654 * off later.
1655 */
1656 /* Host Command Register (HCR) */
1657 state->hs_cmd_regs.hcr = (hermon_hw_hcr_t *)
1658 ((uintptr_t)state->hs_reg_cmd_baseaddr + HERMON_CMD_HCR_OFFSET);
1659 state->hs_cmd_toggle = 0; /* initialize it for use */
1660
1661 /* Software Reset register (sw_reset) and semaphore */
1662 state->hs_cmd_regs.sw_reset = (uint32_t *)
1663 ((uintptr_t)state->hs_reg_cmd_baseaddr +
1664 HERMON_CMD_SW_RESET_OFFSET);
1665 state->hs_cmd_regs.sw_semaphore = (uint32_t *)
1666 ((uintptr_t)state->hs_reg_cmd_baseaddr +
1667 HERMON_CMD_SW_SEMAPHORE_OFFSET);
1668
1669 /* make sure init'd before we start filling things in */
1670 bzero(&state->hs_hcaparams, sizeof (struct hermon_hw_initqueryhca_s));
1671
1672 /* Initialize the Phase1 configuration profile */
1673 status = hermon_cfg_profile_init_phase1(state);
1674 if (status != DDI_SUCCESS) {
1675 hermon_hw_fini(state, cleanup);
1676 HERMON_ATTACH_MSG(state->hs_attach_buf,
1677 "hw_init_cfginit1_fail");
1678 /* This case is not the degraded one */
1679 return (DDI_FAILURE);
1680 }
1681 cleanup = HERMON_DRV_CLEANUP_LEVEL3;
1682
1683 /* Do a software reset of the adapter to ensure proper state */
1684 status = hermon_sw_reset(state);
1685 if (status != HERMON_CMD_SUCCESS) {
1686 hermon_hw_fini(state, cleanup);
1687 HERMON_ATTACH_MSG(state->hs_attach_buf,
1688 "hw_init_sw_reset_fail");
1689 /* This case is not the degraded one */
1690 return (DDI_FAILURE);
1691 }
1692
1693 /* Initialize mailboxes */
1694 status = hermon_rsrc_init_phase1(state);
1695 if (status != DDI_SUCCESS) {
1696 hermon_hw_fini(state, cleanup);
1697 HERMON_ATTACH_MSG(state->hs_attach_buf,
1698 "hw_init_rsrcinit1_fail");
1699 /* This case is not the degraded one */
1700 return (DDI_FAILURE);
1701 }
1702 cleanup = HERMON_DRV_CLEANUP_LEVEL4;
1703
1704 /* Post QUERY_FW */
1705 status = hermon_cmn_query_cmd_post(state, QUERY_FW, 0, 0, &state->hs_fw,
1706 sizeof (hermon_hw_queryfw_t), HERMON_CMD_NOSLEEP_SPIN);
1707 if (status != HERMON_CMD_SUCCESS) {
1708 cmn_err(CE_NOTE, "QUERY_FW command failed: %08x\n", status);
1709 hermon_hw_fini(state, cleanup);
1710 HERMON_ATTACH_MSG(state->hs_attach_buf,
1711 "hw_init_query_fw_cmd_fail");
1712 /* This case is not the degraded one */
1713 return (DDI_FAILURE);
1714 }
1715
1716 /* Validate what/that HERMON FW version is appropriate */
1717
1718 status = hermon_fw_version_check(state);
1719 if (status != DDI_SUCCESS) {
1720 HERMON_FMANOTE(state, HERMON_FMA_FWVER);
1721 if (state->hs_operational_mode == HERMON_HCA_MODE) {
1722 cmn_err(CE_CONT, "Unsupported Hermon FW version: "
1723 "expected: %04d.%04d.%04d, "
1724 "actual: %04d.%04d.%04d\n",
1725 HERMON_FW_VER_MAJOR,
1726 HERMON_FW_VER_MINOR,
1727 HERMON_FW_VER_SUBMINOR,
1728 state->hs_fw.fw_rev_major,
1729 state->hs_fw.fw_rev_minor,
1730 state->hs_fw.fw_rev_subminor);
1731 } else {
1732 cmn_err(CE_CONT, "Unsupported FW version: "
1733 "%04d.%04d.%04d\n",
1734 state->hs_fw.fw_rev_major,
1735 state->hs_fw.fw_rev_minor,
1736 state->hs_fw.fw_rev_subminor);
1737 }
1738 state->hs_operational_mode = HERMON_MAINTENANCE_MODE;
1739 state->hs_fm_degraded_reason = HCA_FW_MISMATCH;
1740 hermon_hw_fini(state, cleanup);
1741 HERMON_ATTACH_MSG(state->hs_attach_buf,
1742 "hw_init_checkfwver_fail");
1743 /* This case is the degraded one */
1744 return (HERMON_CMD_BAD_NVMEM);
1745 }
1746
1747 /*
1748 * Save off the rest of the interesting registers that we'll be using.
1749 * Setup the offsets for the other registers.
1750 */
1751
1752 /*
1753 * Hermon does the intr_offset from the BAR - technically should get the
1754 * BAR info from the response, but PRM says it's from BAR0-1, which is
1755 * for us the CMD BAR
1756 */
1757
1758 clr_intr_offset = state->hs_fw.clr_intr_offs & HERMON_CMD_OFFSET_MASK;
1759
1760 /* Save Clear Interrupt address */
1761 state->hs_cmd_regs.clr_intr = (uint64_t *)
1762 (uintptr_t)(state->hs_reg_cmd_baseaddr + clr_intr_offset);
1763
1764 /*
1765 * Set the error buffer also into the structure - used in hermon_event.c
1766 * to check for internal error on the HCA, not reported in eqe or
1767 * (necessarily) by interrupt
1768 */
1769 state->hs_cmd_regs.fw_err_buf = (uint32_t *)(uintptr_t)
1770 (state->hs_reg_cmd_baseaddr + state->hs_fw.error_buf_addr);
1771
1772 /*
1773 * Invoke a polling thread to check the error buffer periodically.
1774 */
1775 if (!hermon_no_inter_err_chk) {
1776 state->hs_fm_poll_thread = ddi_periodic_add(
1777 hermon_inter_err_chk, (void *)state, FM_POLL_INTERVAL,
1778 DDI_IPL_0);
1779 }
1780
1781 cleanup = HERMON_DRV_CLEANUP_LEVEL5;
1782
1783 /*
1784 * Allocate, map, and run the HCA Firmware.
1785 */
1786
1787 /* Allocate memory for the firmware to load into and map it */
1788
1789 /* get next higher power of 2 */
1790 fw_size = 1 << highbit(state->hs_fw.fw_pages);
1791 state->hs_fw_dma.length = fw_size << HERMON_PAGESHIFT;
1792 status = hermon_dma_alloc(state, &state->hs_fw_dma, MAP_FA);
1793 if (status != DDI_SUCCESS) {
1794 cmn_err(CE_NOTE, "FW alloc failed\n");
1795 hermon_hw_fini(state, cleanup);
1796 HERMON_ATTACH_MSG(state->hs_attach_buf,
1797 "hw_init_dma_alloc_fw_fail");
1798 /* This case is not the degraded one */
1799 return (DDI_FAILURE);
1800 }
1801
1802 cleanup = HERMON_DRV_CLEANUP_LEVEL6;
1803
1804 /* Invoke the RUN_FW cmd to run the firmware */
1805 status = hermon_run_fw_cmd_post(state);
1806 if (status != DDI_SUCCESS) {
1807 cmn_err(CE_NOTE, "RUN_FW command failed: 0x%08x\n", status);
1808 if (status == HERMON_CMD_BAD_NVMEM) {
1809 state->hs_operational_mode = HERMON_MAINTENANCE_MODE;
1810 state->hs_fm_degraded_reason = HCA_FW_CORRUPT;
1811 }
1812 hermon_hw_fini(state, cleanup);
1813 HERMON_ATTACH_MSG(state->hs_attach_buf, "hw_init_run_fw_fail");
1814 /*
1815 * If the status is HERMON_CMD_BAD_NVMEM, it's likely the
1816 * firmware is corrupted, so the mode falls into the
1817 * maintenance mode.
1818 */
1819 return (status == HERMON_CMD_BAD_NVMEM ? HERMON_CMD_BAD_NVMEM :
1820 DDI_FAILURE);
1821 }
1822
1823
1824 /*
1825 * QUERY DEVICE LIMITS/CAPABILITIES
1826 * NOTE - in Hermon, the command is changed to QUERY_DEV_CAP,
1827 * but for familiarity we have kept the structure name the
1828 * same as Tavor/Arbel
1829 */
1830
1831 status = hermon_cmn_query_cmd_post(state, QUERY_DEV_CAP, 0, 0,
1832 &state->hs_devlim, sizeof (hermon_hw_querydevlim_t),
1833 HERMON_CMD_NOSLEEP_SPIN);
1834 if (status != HERMON_CMD_SUCCESS) {
1835 cmn_err(CE_NOTE, "QUERY_DEV_CAP command failed: 0x%08x\n",
1836 status);
1837 hermon_hw_fini(state, cleanup);
1838 HERMON_ATTACH_MSG(state->hs_attach_buf, "hw_init_devcap_fail");
1839 /* This case is not the degraded one */
1840 return (DDI_FAILURE);
1841 }
1842
1843 state->hs_rsvd_eqs = max(state->hs_devlim.num_rsvd_eq,
1844 (4 * state->hs_devlim.num_rsvd_uar));
1845
1846 /* now we have enough info to map in the UAR BAR */
1847 /*
1848 * First, we figure out how to map the BAR for UAR - use only half if
1849 * BlueFlame is enabled - in that case the mapped length is 1/2 the
1850 * log_max_uar_sz (max__uar - 1) * 1MB ( +20).
1851 */
1852
1853 if (state->hs_devlim.blu_flm) { /* Blue Flame Enabled */
1854 offset = (uint64_t)1 << (state->hs_devlim.log_max_uar_sz + 20);
1855 } else {
1856 offset = 0; /* a zero length means map the whole thing */
1857 }
1858 status = hermon_regs_map_setup(state, HERMON_UAR_BAR,
1859 &state->hs_reg_uar_baseaddr, 0, offset, &state->hs_fm_accattr,
1860 &state->hs_fm_uarhdl);
1861 if (status != DDI_SUCCESS) {
1862 HERMON_ATTACH_MSG(state->hs_attach_buf, "UAR BAR mapping");
1863 /* This case is not the degraded one */
1864 return (DDI_FAILURE);
1865 }
1866
1867 /* and if BlueFlame is enabled, map the other half there */
1868 if (state->hs_devlim.blu_flm) { /* Blue Flame Enabled */
1869 offset = (uint64_t)1 << (state->hs_devlim.log_max_uar_sz + 20);
1870 status = ddi_regs_map_setup(state->hs_dip, HERMON_UAR_BAR,
1871 &state->hs_reg_bf_baseaddr, offset, offset,
1872 &state->hs_reg_accattr, &state->hs_reg_bfhdl);
1873 if (status != DDI_SUCCESS) {
1874 HERMON_ATTACH_MSG(state->hs_attach_buf,
1875 "BlueFlame BAR mapping");
1876 /* This case is not the degraded one */
1877 return (DDI_FAILURE);
1878 }
1879 /* This will be used in hw_fini if we fail to init. */
1880 state->hs_bf_offset = offset;
1881 }
1882 cleanup = HERMON_DRV_CLEANUP_LEVEL7;
1883
1884 /* Hermon has a couple of things needed for phase 2 in query port */
1885
1886 status = hermon_cmn_query_cmd_post(state, QUERY_PORT, 0, 0x01,
1887 &state->hs_queryport, sizeof (hermon_hw_query_port_t),
1888 HERMON_CMD_NOSLEEP_SPIN);
1889 if (status != HERMON_CMD_SUCCESS) {
1890 cmn_err(CE_NOTE, "QUERY_PORT command failed: 0x%08x\n",
1891 status);
1892 hermon_hw_fini(state, cleanup);
1893 HERMON_ATTACH_MSG(state->hs_attach_buf,
1894 "hw_init_queryport_fail");
1895 /* This case is not the degraded one */
1896 return (DDI_FAILURE);
1897 }
1898
1899 /* Initialize the Phase2 Hermon configuration profile */
1900 status = hermon_cfg_profile_init_phase2(state);
1901 if (status != DDI_SUCCESS) {
1902 cmn_err(CE_NOTE, "CFG phase 2 failed: 0x%08x\n", status);
1903 hermon_hw_fini(state, cleanup);
1904 HERMON_ATTACH_MSG(state->hs_attach_buf,
1905 "hw_init_cfginit2_fail");
1906 /* This case is not the degraded one */
1907 return (DDI_FAILURE);
1908 }
1909
1910 /* Determine and set the ICM size */
1911 state->hs_icm_sz = hermon_size_icm(state);
1912 status = hermon_set_icm_size_cmd_post(state);
1913 if (status != DDI_SUCCESS) {
1914 cmn_err(CE_NOTE, "Hermon: SET_ICM_SIZE cmd failed: 0x%08x\n",
1915 status);
1916 hermon_hw_fini(state, cleanup);
1917 HERMON_ATTACH_MSG(state->hs_attach_buf,
1918 "hw_init_seticmsz_fail");
1919 /* This case is not the degraded one */
1920 return (DDI_FAILURE);
1921 }
1922 /* alloc icm aux physical memory and map it */
1923
1924 state->hs_icma_dma.length = 1 << highbit(state->hs_icma_sz);
1925
1926 status = hermon_dma_alloc(state, &state->hs_icma_dma, MAP_ICM_AUX);
1927 if (status != DDI_SUCCESS) {
1928 cmn_err(CE_NOTE, "failed to alloc (0x%llx) bytes for ICMA\n",
1929 (longlong_t)state->hs_icma_dma.length);
1930 hermon_hw_fini(state, cleanup);
1931 HERMON_ATTACH_MSG(state->hs_attach_buf,
1932 "hw_init_dma_alloc_icm_aux_fail");
1933 /* This case is not the degraded one */
1934 return (DDI_FAILURE);
1935 }
1936 cleanup = HERMON_DRV_CLEANUP_LEVEL8;
1937
1938 cleanup = HERMON_DRV_CLEANUP_LEVEL9;
1939
1940 /* Allocate an array of structures to house the ICM tables */
1941 state->hs_icm = kmem_zalloc(HERMON_NUM_ICM_RESOURCES *
1942 sizeof (hermon_icm_table_t), KM_SLEEP);
1943
1944 /* Set up the ICM address space and the INIT_HCA command input */
1945 status = hermon_icm_config_setup(state, &state->hs_hcaparams);
1946 if (status != HERMON_CMD_SUCCESS) {
1947 cmn_err(CE_NOTE, "ICM configuration failed\n");
1948 hermon_hw_fini(state, cleanup);
1949 HERMON_ATTACH_MSG(state->hs_attach_buf,
1950 "hw_init_icm_config_setup_fail");
1951 /* This case is not the degraded one */
1952 return (DDI_FAILURE);
1953 }
1954 cleanup = HERMON_DRV_CLEANUP_LEVEL10;
1955
1956 /* Initialize the adapter with the INIT_HCA cmd */
1957 status = hermon_init_hca_cmd_post(state, &state->hs_hcaparams,
1958 HERMON_CMD_NOSLEEP_SPIN);
1959 if (status != HERMON_CMD_SUCCESS) {
1960 cmn_err(CE_NOTE, "INIT_HCA command failed: %08x\n", status);
1961 hermon_hw_fini(state, cleanup);
1962 HERMON_ATTACH_MSG(state->hs_attach_buf, "hw_init_hca_fail");
1963 /* This case is not the degraded one */
1964 return (DDI_FAILURE);
1965 }
1966 cleanup = HERMON_DRV_CLEANUP_LEVEL11;
1967
1968 /* Enter the second phase of init for Hermon configuration/resources */
1969 status = hermon_rsrc_init_phase2(state);
1970 if (status != DDI_SUCCESS) {
1971 hermon_hw_fini(state, cleanup);
1972 HERMON_ATTACH_MSG(state->hs_attach_buf,
1973 "hw_init_rsrcinit2_fail");
1974 /* This case is not the degraded one */
1975 return (DDI_FAILURE);
1976 }
1977 cleanup = HERMON_DRV_CLEANUP_LEVEL12;
1978
1979 /* Query the adapter via QUERY_ADAPTER */
1980 status = hermon_cmn_query_cmd_post(state, QUERY_ADAPTER, 0, 0,
1981 &state->hs_adapter, sizeof (hermon_hw_queryadapter_t),
1982 HERMON_CMD_NOSLEEP_SPIN);
1983 if (status != HERMON_CMD_SUCCESS) {
1984 cmn_err(CE_NOTE, "Hermon: QUERY_ADAPTER command failed: %08x\n",
1985 status);
1986 hermon_hw_fini(state, cleanup);
1987 HERMON_ATTACH_MSG(state->hs_attach_buf,
1988 "hw_init_query_adapter_fail");
1989 /* This case is not the degraded one */
1990 return (DDI_FAILURE);
1991 }
1992
1993 /* Allocate protection domain (PD) for Hermon internal use */
1994 status = hermon_pd_alloc(state, &state->hs_pdhdl_internal,
1995 HERMON_SLEEP);
1996 if (status != DDI_SUCCESS) {
1997 cmn_err(CE_NOTE, "failed to alloc internal PD\n");
1998 hermon_hw_fini(state, cleanup);
1999 HERMON_ATTACH_MSG(state->hs_attach_buf,
2000 "hw_init_internal_pd_alloc_fail");
2001 /* This case is not the degraded one */
2002 return (DDI_FAILURE);
2003 }
2004 cleanup = HERMON_DRV_CLEANUP_LEVEL13;
2005
2006 /* Setup UAR page for kernel use */
2007 status = hermon_internal_uarpg_init(state);
2008 if (status != DDI_SUCCESS) {
2009 cmn_err(CE_NOTE, "failed to setup internal UAR\n");
2010 hermon_hw_fini(state, cleanup);
2011 HERMON_ATTACH_MSG(state->hs_attach_buf,
2012 "hw_init_internal_uarpg_alloc_fail");
2013 /* This case is not the degraded one */
2014 return (DDI_FAILURE);
2015 }
2016 cleanup = HERMON_DRV_CLEANUP_LEVEL14;
2017
2018 /* Query and initialize the Hermon interrupt/MSI information */
2019 status = hermon_intr_or_msi_init(state);
2020 if (status != DDI_SUCCESS) {
2021 cmn_err(CE_NOTE, "failed to setup INTR/MSI\n");
2022 hermon_hw_fini(state, cleanup);
2023 HERMON_ATTACH_MSG(state->hs_attach_buf,
2024 "hw_init_intr_or_msi_init_fail");
2025 /* This case is not the degraded one */
2026 return (DDI_FAILURE);
2027 }
2028 cleanup = HERMON_DRV_CLEANUP_LEVEL15;
2029
2030 status = hermon_isr_init(state); /* set up the isr */
2031 if (status != DDI_SUCCESS) {
2032 cmn_err(CE_NOTE, "failed to init isr\n");
2033 hermon_hw_fini(state, cleanup);
2034 HERMON_ATTACH_MSG(state->hs_attach_buf,
2035 "hw_init_isrinit_fail");
2036 /* This case is not the degraded one */
2037 return (DDI_FAILURE);
2038 }
2039 cleanup = HERMON_DRV_CLEANUP_LEVEL16;
2040
2041 /* Setup the event queues */
2042 status = hermon_eq_init_all(state);
2043 if (status != DDI_SUCCESS) {
2044 cmn_err(CE_NOTE, "failed to init EQs\n");
2045 hermon_hw_fini(state, cleanup);
2046 HERMON_ATTACH_MSG(state->hs_attach_buf,
2047 "hw_init_eqinitall_fail");
2048 /* This case is not the degraded one */
2049 return (DDI_FAILURE);
2050 }
2051 cleanup = HERMON_DRV_CLEANUP_LEVEL17;
2052
2053
2054
2055 /* Reserve contexts for QP0 and QP1 */
2056 status = hermon_special_qp_contexts_reserve(state);
2057 if (status != DDI_SUCCESS) {
2058 cmn_err(CE_NOTE, "failed to init special QPs\n");
2059 hermon_hw_fini(state, cleanup);
2060 HERMON_ATTACH_MSG(state->hs_attach_buf,
2061 "hw_init_rsrv_sqp_fail");
2062 /* This case is not the degraded one */
2063 return (DDI_FAILURE);
2064 }
2065 cleanup = HERMON_DRV_CLEANUP_LEVEL18;
2066
2067 /* Initialize for multicast group handling */
2068 status = hermon_mcg_init(state);
2069 if (status != DDI_SUCCESS) {
2070 cmn_err(CE_NOTE, "failed to init multicast\n");
2071 hermon_hw_fini(state, cleanup);
2072 HERMON_ATTACH_MSG(state->hs_attach_buf,
2073 "hw_init_mcg_init_fail");
2074 /* This case is not the degraded one */
2075 return (DDI_FAILURE);
2076 }
2077 cleanup = HERMON_DRV_CLEANUP_LEVEL19;
2078
2079 /* Initialize the Hermon IB port(s) */
2080 status = hermon_hca_port_init(state);
2081 if (status != DDI_SUCCESS) {
2082 cmn_err(CE_NOTE, "failed to init HCA Port\n");
2083 hermon_hw_fini(state, cleanup);
2084 HERMON_ATTACH_MSG(state->hs_attach_buf,
2085 "hw_init_hca_port_init_fail");
2086 /* This case is not the degraded one */
2087 return (DDI_FAILURE);
2088 }
2089
2090 cleanup = HERMON_DRV_CLEANUP_ALL;
2091
2092 /* Determine NodeGUID and SystemImageGUID */
2093 status = hermon_getnodeinfo_cmd_post(state, HERMON_CMD_NOSLEEP_SPIN,
2094 &nodeinfo);
2095 if (status != HERMON_CMD_SUCCESS) {
2096 cmn_err(CE_NOTE, "GetNodeInfo command failed: %08x\n", status);
2097 hermon_hw_fini(state, cleanup);
2098 HERMON_ATTACH_MSG(state->hs_attach_buf,
2099 "hw_init_getnodeinfo_cmd_fail");
2100 /* This case is not the degraded one */
2101 return (DDI_FAILURE);
2102 }
2103
2104 /*
2105 * If the NodeGUID value was set in OBP properties, then we use that
2106 * value. But we still print a message if the value we queried from
2107 * firmware does not match this value.
2108 *
2109 * Otherwise if OBP value is not set then we use the value from
2110 * firmware unconditionally.
2111 */
2112 if (state->hs_cfg_profile->cp_nodeguid) {
2113 state->hs_nodeguid = state->hs_cfg_profile->cp_nodeguid;
2114 } else {
2115 state->hs_nodeguid = nodeinfo.NodeGUID;
2116 }
2117
2118 if (state->hs_nodeguid != nodeinfo.NodeGUID) {
2119 cmn_err(CE_NOTE, "!NodeGUID value queried from firmware "
2120 "does not match value set by device property");
2121 }
2122
2123 /*
2124 * If the SystemImageGUID value was set in OBP properties, then we use
2125 * that value. But we still print a message if the value we queried
2126 * from firmware does not match this value.
2127 *
2128 * Otherwise if OBP value is not set then we use the value from
2129 * firmware unconditionally.
2130 */
2131 if (state->hs_cfg_profile->cp_sysimgguid) {
2132 state->hs_sysimgguid = state->hs_cfg_profile->cp_sysimgguid;
2133 } else {
2134 state->hs_sysimgguid = nodeinfo.SystemImageGUID;
2135 }
2136
2137 if (state->hs_sysimgguid != nodeinfo.SystemImageGUID) {
2138 cmn_err(CE_NOTE, "!SystemImageGUID value queried from firmware "
2139 "does not match value set by device property");
2140 }
2141
2142 /* Get NodeDescription */
2143 status = hermon_getnodedesc_cmd_post(state, HERMON_CMD_NOSLEEP_SPIN,
2144 (sm_nodedesc_t *)&state->hs_nodedesc);
2145 if (status != HERMON_CMD_SUCCESS) {
2146 cmn_err(CE_CONT, "GetNodeDesc command failed: %08x\n", status);
2147 hermon_hw_fini(state, cleanup);
2148 HERMON_ATTACH_MSG(state->hs_attach_buf,
2149 "hw_init_getnodedesc_cmd_fail");
2150 /* This case is not the degraded one */
2151 return (DDI_FAILURE);
2152 }
2153
2154 return (DDI_SUCCESS);
2155 }
2156
2157
2158 /*
2159 * hermon_hw_fini()
2160 * Context: Only called from attach() and/or detach() path contexts
2161 */
2162 static void
hermon_hw_fini(hermon_state_t * state,hermon_drv_cleanup_level_t cleanup)2163 hermon_hw_fini(hermon_state_t *state, hermon_drv_cleanup_level_t cleanup)
2164 {
2165 uint_t num_ports;
2166 int i, status;
2167
2168
2169 /*
2170 * JBDB - We might not want to run these returns in all cases of
2171 * Bad News. We should still attempt to free all of the DMA memory
2172 * resources... This needs to be worked last, after all allocations
2173 * are implemented. For now, and possibly for later, this works.
2174 */
2175
2176 switch (cleanup) {
2177 /*
2178 * If we add more driver initialization steps that should be cleaned
2179 * up here, we need to ensure that HERMON_DRV_CLEANUP_ALL is still the
2180 * first entry (i.e. corresponds to the last init step).
2181 */
2182 case HERMON_DRV_CLEANUP_ALL:
2183 /* Shutdown the Hermon IB port(s) */
2184 num_ports = state->hs_cfg_profile->cp_num_ports;
2185 (void) hermon_hca_ports_shutdown(state, num_ports);
2186 /* FALLTHROUGH */
2187
2188 case HERMON_DRV_CLEANUP_LEVEL19:
2189 /* Teardown resources used for multicast group handling */
2190 hermon_mcg_fini(state);
2191 /* FALLTHROUGH */
2192
2193 case HERMON_DRV_CLEANUP_LEVEL18:
2194 /* Unreserve the special QP contexts */
2195 hermon_special_qp_contexts_unreserve(state);
2196 /* FALLTHROUGH */
2197
2198 case HERMON_DRV_CLEANUP_LEVEL17:
2199 /*
2200 * Attempt to teardown all event queues (EQ). If we fail
2201 * here then print a warning message and return. Something
2202 * (either in HW or SW) has gone seriously wrong.
2203 */
2204 status = hermon_eq_fini_all(state);
2205 if (status != DDI_SUCCESS) {
2206 HERMON_WARNING(state, "failed to teardown EQs");
2207 return;
2208 }
2209 /* FALLTHROUGH */
2210 case HERMON_DRV_CLEANUP_LEVEL16:
2211 /* Teardown Hermon interrupts */
2212 hermon_isr_fini(state);
2213 /* FALLTHROUGH */
2214
2215 case HERMON_DRV_CLEANUP_LEVEL15:
2216 status = hermon_intr_or_msi_fini(state);
2217 if (status != DDI_SUCCESS) {
2218 HERMON_WARNING(state, "failed to free intr/MSI");
2219 return;
2220 }
2221 /* FALLTHROUGH */
2222
2223 case HERMON_DRV_CLEANUP_LEVEL14:
2224 /* Free the resources for the Hermon internal UAR pages */
2225 hermon_internal_uarpg_fini(state);
2226 /* FALLTHROUGH */
2227
2228 case HERMON_DRV_CLEANUP_LEVEL13:
2229 /*
2230 * Free the PD that was used internally by Hermon software. If
2231 * we fail here then print a warning and return. Something
2232 * (probably software-related, but perhaps HW) has gone wrong.
2233 */
2234 status = hermon_pd_free(state, &state->hs_pdhdl_internal);
2235 if (status != DDI_SUCCESS) {
2236 HERMON_WARNING(state, "failed to free internal PD");
2237 return;
2238 }
2239 /* FALLTHROUGH */
2240
2241 case HERMON_DRV_CLEANUP_LEVEL12:
2242 /* Cleanup all the phase2 resources first */
2243 hermon_rsrc_fini(state, HERMON_RSRC_CLEANUP_ALL);
2244 /* FALLTHROUGH */
2245
2246 case HERMON_DRV_CLEANUP_LEVEL11:
2247 /* LEVEL11 is after INIT_HCA */
2248 /* FALLTHROUGH */
2249
2250
2251 case HERMON_DRV_CLEANUP_LEVEL10:
2252 /*
2253 * Unmap the ICM memory area with UNMAP_ICM command.
2254 */
2255 status = hermon_unmap_icm_cmd_post(state, NULL);
2256 if (status != DDI_SUCCESS) {
2257 cmn_err(CE_WARN,
2258 "hermon_hw_fini: failed to unmap ICM\n");
2259 }
2260
2261 /* Free the initial ICM DMA handles */
2262 hermon_icm_dma_fini(state);
2263
2264 /* Free the ICM table structures */
2265 hermon_icm_tables_fini(state);
2266
2267 /* Free the ICM table handles */
2268 kmem_free(state->hs_icm, HERMON_NUM_ICM_RESOURCES *
2269 sizeof (hermon_icm_table_t));
2270
2271 /* FALLTHROUGH */
2272
2273 case HERMON_DRV_CLEANUP_LEVEL9:
2274 /*
2275 * Unmap the ICM Aux memory area with UNMAP_ICM_AUX command.
2276 */
2277 status = hermon_unmap_icm_aux_cmd_post(state);
2278 if (status != HERMON_CMD_SUCCESS) {
2279 cmn_err(CE_NOTE,
2280 "hermon_hw_fini: failed to unmap ICMA\n");
2281 }
2282 /* FALLTHROUGH */
2283
2284 case HERMON_DRV_CLEANUP_LEVEL8:
2285 /*
2286 * Deallocate ICM Aux DMA memory.
2287 */
2288 hermon_dma_free(&state->hs_icma_dma);
2289 /* FALLTHROUGH */
2290
2291 case HERMON_DRV_CLEANUP_LEVEL7:
2292 if (state->hs_fm_uarhdl) {
2293 hermon_regs_map_free(state, &state->hs_fm_uarhdl);
2294 state->hs_fm_uarhdl = NULL;
2295 }
2296
2297 if (state->hs_reg_uarhdl) {
2298 ddi_regs_map_free(&state->hs_reg_uarhdl);
2299 state->hs_reg_uarhdl = NULL;
2300 }
2301
2302 if (state->hs_bf_offset != 0 && state->hs_reg_bfhdl) {
2303 ddi_regs_map_free(&state->hs_reg_bfhdl);
2304 state->hs_reg_bfhdl = NULL;
2305 }
2306
2307 for (i = 0; i < HERMON_MAX_PORTS; i++) {
2308 if (state->hs_pkey[i]) {
2309 kmem_free(state->hs_pkey[i], (1 <<
2310 state->hs_cfg_profile->cp_log_max_pkeytbl) *
2311 sizeof (ib_pkey_t));
2312 state->hs_pkey[i] = NULL;
2313 }
2314 if (state->hs_guid[i]) {
2315 kmem_free(state->hs_guid[i], (1 <<
2316 state->hs_cfg_profile->cp_log_max_gidtbl) *
2317 sizeof (ib_guid_t));
2318 state->hs_guid[i] = NULL;
2319 }
2320 }
2321 /* FALLTHROUGH */
2322
2323 case HERMON_DRV_CLEANUP_LEVEL6:
2324 /*
2325 * Unmap the firmware memory area with UNMAP_FA command.
2326 */
2327 status = hermon_unmap_fa_cmd_post(state);
2328
2329 if (status != HERMON_CMD_SUCCESS) {
2330 cmn_err(CE_NOTE,
2331 "hermon_hw_fini: failed to unmap FW\n");
2332 }
2333
2334 /*
2335 * Deallocate firmware DMA memory.
2336 */
2337 hermon_dma_free(&state->hs_fw_dma);
2338 /* FALLTHROUGH */
2339
2340 case HERMON_DRV_CLEANUP_LEVEL5:
2341 /* stop the poll thread */
2342 if (state->hs_fm_poll_thread) {
2343 ddi_periodic_delete(state->hs_fm_poll_thread);
2344 state->hs_fm_poll_thread = NULL;
2345 }
2346 /* FALLTHROUGH */
2347
2348 case HERMON_DRV_CLEANUP_LEVEL4:
2349 /* Then cleanup the phase1 resources */
2350 hermon_rsrc_fini(state, HERMON_RSRC_CLEANUP_PHASE1_COMPLETE);
2351 /* FALLTHROUGH */
2352
2353 case HERMON_DRV_CLEANUP_LEVEL3:
2354 /* Teardown any resources allocated for the config profile */
2355 hermon_cfg_profile_fini(state);
2356 /* FALLTHROUGH */
2357
2358 case HERMON_DRV_CLEANUP_LEVEL2:
2359 #ifdef HERMON_SUPPORTS_MSIX_BAR
2360 /*
2361 * unmap 3rd BAR, MSIX BAR
2362 */
2363 if (state->hs_reg_msihdl) {
2364 ddi_regs_map_free(&state->hs_reg_msihdl);
2365 state->hs_reg_msihdl = NULL;
2366 }
2367 /* FALLTHROUGH */
2368 #endif
2369 case HERMON_DRV_CLEANUP_LEVEL1:
2370 case HERMON_DRV_CLEANUP_LEVEL0:
2371 /*
2372 * LEVEL1 and LEVEL0 resources are freed in
2373 * hermon_drv_fini2().
2374 */
2375 break;
2376
2377 default:
2378 HERMON_WARNING(state, "unexpected driver cleanup level");
2379 return;
2380 }
2381 }
2382
2383
2384 /*
2385 * hermon_soft_state_init()
2386 * Context: Only called from attach() path context
2387 */
2388 static int
hermon_soft_state_init(hermon_state_t * state)2389 hermon_soft_state_init(hermon_state_t *state)
2390 {
2391 ibt_hca_attr_t *hca_attr;
2392 uint64_t maxval, val;
2393 ibt_hca_flags_t caps = IBT_HCA_NO_FLAGS;
2394 ibt_hca_flags2_t caps2 = IBT_HCA2_NO_FLAGS;
2395 int status;
2396 int max_send_wqe_bytes;
2397 int max_recv_wqe_bytes;
2398
2399 /*
2400 * The ibc_hca_info_t struct is passed to the IBTF. This is the
2401 * routine where we initialize it. Many of the init values come from
2402 * either configuration variables or successful queries of the Hermon
2403 * hardware abilities
2404 */
2405 state->hs_ibtfinfo.hca_ci_vers = IBCI_V4;
2406 state->hs_ibtfinfo.hca_handle = (ibc_hca_hdl_t)state;
2407 state->hs_ibtfinfo.hca_ops = &hermon_ibc_ops;
2408
2409 hca_attr = kmem_zalloc(sizeof (ibt_hca_attr_t), KM_SLEEP);
2410 state->hs_ibtfinfo.hca_attr = hca_attr;
2411
2412 hca_attr->hca_dip = state->hs_dip;
2413 hca_attr->hca_fw_major_version = state->hs_fw.fw_rev_major;
2414 hca_attr->hca_fw_minor_version = state->hs_fw.fw_rev_minor;
2415 hca_attr->hca_fw_micro_version = state->hs_fw.fw_rev_subminor;
2416
2417 /* CQ interrupt moderation maximums - each limited to 16 bits */
2418 hca_attr->hca_max_cq_mod_count = 0xFFFF;
2419 hca_attr->hca_max_cq_mod_usec = 0xFFFF;
2420 hca_attr->hca_max_cq_handlers = state->hs_intrmsi_allocd;
2421
2422
2423 /*
2424 * Determine HCA capabilities:
2425 * No default support for IBT_HCA_RD, IBT_HCA_RAW_MULTICAST,
2426 * IBT_HCA_ATOMICS_GLOBAL, IBT_HCA_RESIZE_CHAN, IBT_HCA_INIT_TYPE,
2427 * or IBT_HCA_SHUTDOWN_PORT
2428 * But IBT_HCA_AH_PORT_CHECK, IBT_HCA_SQD_RTS_PORT, IBT_HCA_SI_GUID,
2429 * IBT_HCA_RNR_NAK, IBT_HCA_CURRENT_QP_STATE, IBT_HCA_PORT_UP,
2430 * IBT_HCA_SRQ, IBT_HCA_RESIZE_SRQ and IBT_HCA_FMR are always
2431 * supported
2432 * All other features are conditionally supported, depending on the
2433 * status return by the Hermon HCA in QUERY_DEV_LIM.
2434 */
2435 if (state->hs_devlim.ud_multi) {
2436 caps |= IBT_HCA_UD_MULTICAST;
2437 }
2438 if (state->hs_devlim.atomic) {
2439 caps |= IBT_HCA_ATOMICS_HCA;
2440 }
2441 if (state->hs_devlim.apm) {
2442 caps |= IBT_HCA_AUTO_PATH_MIG;
2443 }
2444 if (state->hs_devlim.pkey_v) {
2445 caps |= IBT_HCA_PKEY_CNTR;
2446 }
2447 if (state->hs_devlim.qkey_v) {
2448 caps |= IBT_HCA_QKEY_CNTR;
2449 }
2450 if (state->hs_devlim.ipoib_cksm) {
2451 caps |= IBT_HCA_CKSUM_FULL;
2452 caps2 |= IBT_HCA2_IP_CLASS;
2453 }
2454 if (state->hs_devlim.mod_wr_srq) {
2455 caps |= IBT_HCA_RESIZE_SRQ;
2456 }
2457 if (state->hs_devlim.lif) {
2458 caps |= IBT_HCA_LOCAL_INVAL_FENCE;
2459 }
2460 if (state->hs_devlim.reserved_lkey) {
2461 caps2 |= IBT_HCA2_RES_LKEY;
2462 hca_attr->hca_reserved_lkey = state->hs_devlim.rsv_lkey;
2463 }
2464 if (state->hs_devlim.local_inv && state->hs_devlim.remote_inv &&
2465 state->hs_devlim.fast_reg_wr) { /* fw needs to be >= 2.7.000 */
2466 if ((state->hs_fw.fw_rev_major > 2) ||
2467 ((state->hs_fw.fw_rev_major == 2) &&
2468 (state->hs_fw.fw_rev_minor >= 7)))
2469 caps2 |= IBT_HCA2_MEM_MGT_EXT;
2470 }
2471 if (state->hs_devlim.log_max_rss_tbl_sz) {
2472 hca_attr->hca_rss_max_log2_table =
2473 state->hs_devlim.log_max_rss_tbl_sz;
2474 if (state->hs_devlim.rss_xor)
2475 caps2 |= IBT_HCA2_RSS_XOR_ALG;
2476 if (state->hs_devlim.rss_toep)
2477 caps2 |= IBT_HCA2_RSS_TPL_ALG;
2478 }
2479 if (state->hs_devlim.mps) {
2480 caps |= IBT_HCA_ZERO_BASED_VA;
2481 }
2482 if (state->hs_devlim.zb) {
2483 caps |= IBT_HCA_MULT_PAGE_SZ_MR;
2484 }
2485 caps |= (IBT_HCA_AH_PORT_CHECK | IBT_HCA_SQD_SQD_PORT |
2486 IBT_HCA_SI_GUID | IBT_HCA_RNR_NAK | IBT_HCA_CURRENT_QP_STATE |
2487 IBT_HCA_PORT_UP | IBT_HCA_RC_SRQ | IBT_HCA_UD_SRQ | IBT_HCA_FMR);
2488 caps2 |= IBT_HCA2_DMA_MR;
2489
2490 if (state->hs_devlim.log_max_gso_sz) {
2491 hca_attr->hca_max_lso_size =
2492 (1 << state->hs_devlim.log_max_gso_sz);
2493 /* 64 = ctrl & datagram seg, 4 = LSO seg, 16 = 1 SGL */
2494 hca_attr->hca_max_lso_hdr_size =
2495 state->hs_devlim.max_desc_sz_sq - (64 + 4 + 16);
2496 }
2497
2498 caps |= IBT_HCA_WQE_SIZE_INFO;
2499 max_send_wqe_bytes = state->hs_devlim.max_desc_sz_sq;
2500 max_recv_wqe_bytes = state->hs_devlim.max_desc_sz_rq;
2501 hca_attr->hca_ud_send_sgl_sz = (max_send_wqe_bytes / 16) - 4;
2502 hca_attr->hca_conn_send_sgl_sz = (max_send_wqe_bytes / 16) - 1;
2503 hca_attr->hca_conn_rdma_sgl_overhead = 1;
2504 hca_attr->hca_conn_rdma_write_sgl_sz = (max_send_wqe_bytes / 16) - 2;
2505 hca_attr->hca_conn_rdma_read_sgl_sz = (512 / 16) - 2; /* see PRM */
2506 hca_attr->hca_recv_sgl_sz = max_recv_wqe_bytes / 16;
2507
2508 /* We choose not to support "inline" unless it improves performance */
2509 hca_attr->hca_max_inline_size = 0;
2510 hca_attr->hca_ud_send_inline_sz = 0;
2511 hca_attr->hca_conn_send_inline_sz = 0;
2512 hca_attr->hca_conn_rdmaw_inline_overhead = 4;
2513
2514 #if defined(_ELF64)
2515 /* 32-bit kernels are too small for Fibre Channel over IB */
2516 if (state->hs_devlim.fcoib && (caps2 & IBT_HCA2_MEM_MGT_EXT)) {
2517 caps2 |= IBT_HCA2_FC;
2518 hca_attr->hca_rfci_max_log2_qp = 7; /* 128 per port */
2519 hca_attr->hca_fexch_max_log2_qp = 16; /* 64K per port */
2520 hca_attr->hca_fexch_max_log2_mem = 20; /* 1MB per MPT */
2521 }
2522 #endif
2523
2524 hca_attr->hca_flags = caps;
2525 hca_attr->hca_flags2 = caps2;
2526
2527 /*
2528 * Set hca_attr's IDs
2529 */
2530 hca_attr->hca_vendor_id = state->hs_vendor_id;
2531 hca_attr->hca_device_id = state->hs_device_id;
2532 hca_attr->hca_version_id = state->hs_revision_id;
2533
2534 /*
2535 * Determine number of available QPs and max QP size. Number of
2536 * available QPs is determined by subtracting the number of
2537 * "reserved QPs" (i.e. reserved for firmware use) from the
2538 * total number configured.
2539 */
2540 val = ((uint64_t)1 << state->hs_cfg_profile->cp_log_num_qp);
2541 hca_attr->hca_max_qp = val - ((uint64_t)1 <<
2542 state->hs_devlim.log_rsvd_qp);
2543 maxval = ((uint64_t)1 << state->hs_devlim.log_max_qp_sz);
2544 val = ((uint64_t)1 << state->hs_cfg_profile->cp_log_max_qp_sz);
2545 if (val > maxval) {
2546 kmem_free(hca_attr, sizeof (ibt_hca_attr_t));
2547 HERMON_ATTACH_MSG(state->hs_attach_buf,
2548 "soft_state_init_maxqpsz_toobig_fail");
2549 return (DDI_FAILURE);
2550 }
2551 /* we need to reduce this by the max space needed for headroom */
2552 hca_attr->hca_max_qp_sz = (uint_t)val - (HERMON_QP_OH_SIZE >>
2553 HERMON_QP_WQE_LOG_MINIMUM) - 1;
2554
2555 /*
2556 * Determine max scatter-gather size in WQEs. The HCA has split
2557 * the max sgl into rec'v Q and send Q values. Use the least.
2558 *
2559 * This is mainly useful for legacy clients. Smart clients
2560 * such as IPoIB will use the IBT_HCA_WQE_SIZE_INFO sgl info.
2561 */
2562 if (state->hs_devlim.max_sg_rq <= state->hs_devlim.max_sg_sq) {
2563 maxval = state->hs_devlim.max_sg_rq;
2564 } else {
2565 maxval = state->hs_devlim.max_sg_sq;
2566 }
2567 val = state->hs_cfg_profile->cp_wqe_max_sgl;
2568 if (val > maxval) {
2569 kmem_free(hca_attr, sizeof (ibt_hca_attr_t));
2570 HERMON_ATTACH_MSG(state->hs_attach_buf,
2571 "soft_state_init_toomanysgl_fail");
2572 return (DDI_FAILURE);
2573 }
2574 /* If the rounded value for max SGL is too large, cap it */
2575 if (state->hs_cfg_profile->cp_wqe_real_max_sgl > maxval) {
2576 state->hs_cfg_profile->cp_wqe_real_max_sgl = (uint32_t)maxval;
2577 val = maxval;
2578 } else {
2579 val = state->hs_cfg_profile->cp_wqe_real_max_sgl;
2580 }
2581
2582 hca_attr->hca_max_sgl = (uint_t)val;
2583 hca_attr->hca_max_rd_sgl = 0; /* zero because RD is unsupported */
2584
2585 /*
2586 * Determine number of available CQs and max CQ size. Number of
2587 * available CQs is determined by subtracting the number of
2588 * "reserved CQs" (i.e. reserved for firmware use) from the
2589 * total number configured.
2590 */
2591 val = ((uint64_t)1 << state->hs_cfg_profile->cp_log_num_cq);
2592 hca_attr->hca_max_cq = val - ((uint64_t)1 <<
2593 state->hs_devlim.log_rsvd_cq);
2594 maxval = ((uint64_t)1 << state->hs_devlim.log_max_cq_sz);
2595 val = ((uint64_t)1 << state->hs_cfg_profile->cp_log_max_cq_sz) - 1;
2596 if (val > maxval) {
2597 kmem_free(hca_attr, sizeof (ibt_hca_attr_t));
2598 HERMON_ATTACH_MSG(state->hs_attach_buf,
2599 "soft_state_init_maxcqsz_toobig_fail");
2600 return (DDI_FAILURE);
2601 }
2602 hca_attr->hca_max_cq_sz = (uint_t)val;
2603
2604 /*
2605 * Determine number of available SRQs and max SRQ size. Number of
2606 * available SRQs is determined by subtracting the number of
2607 * "reserved SRQs" (i.e. reserved for firmware use) from the
2608 * total number configured.
2609 */
2610 val = ((uint64_t)1 << state->hs_cfg_profile->cp_log_num_srq);
2611 hca_attr->hca_max_srqs = val - ((uint64_t)1 <<
2612 state->hs_devlim.log_rsvd_srq);
2613 maxval = ((uint64_t)1 << state->hs_devlim.log_max_srq_sz);
2614 val = ((uint64_t)1 << state->hs_cfg_profile->cp_log_max_srq_sz);
2615
2616 if (val > maxval) {
2617 kmem_free(hca_attr, sizeof (ibt_hca_attr_t));
2618 HERMON_ATTACH_MSG(state->hs_attach_buf,
2619 "soft_state_init_maxsrqsz_toobig_fail");
2620 return (DDI_FAILURE);
2621 }
2622 hca_attr->hca_max_srqs_sz = (uint_t)val;
2623
2624 val = hca_attr->hca_recv_sgl_sz - 1; /* SRQ has a list link */
2625 maxval = state->hs_devlim.max_sg_rq - 1;
2626 if (val > maxval) {
2627 kmem_free(hca_attr, sizeof (ibt_hca_attr_t));
2628 HERMON_ATTACH_MSG(state->hs_attach_buf,
2629 "soft_state_init_toomanysrqsgl_fail");
2630 return (DDI_FAILURE);
2631 }
2632 hca_attr->hca_max_srq_sgl = (uint_t)val;
2633
2634 /*
2635 * Determine supported HCA page sizes
2636 * XXX
2637 * For now we simply return the system pagesize as the only supported
2638 * pagesize
2639 */
2640 hca_attr->hca_page_sz = ((PAGESIZE == (1 << 13)) ? IBT_PAGE_8K :
2641 IBT_PAGE_4K);
2642
2643 /*
2644 * Determine number of available MemReg, MemWin, and their max size.
2645 * Number of available MRs and MWs is determined by subtracting
2646 * the number of "reserved MPTs" (i.e. reserved for firmware use)
2647 * from the total number configured for each.
2648 */
2649 val = ((uint64_t)1 << state->hs_cfg_profile->cp_log_num_dmpt);
2650 hca_attr->hca_max_memr = val - ((uint64_t)1 <<
2651 state->hs_devlim.log_rsvd_dmpt);
2652 hca_attr->hca_max_mem_win = state->hs_devlim.mem_win ? (val -
2653 ((uint64_t)1 << state->hs_devlim.log_rsvd_dmpt)) : 0;
2654 maxval = state->hs_devlim.log_max_mrw_sz;
2655 val = state->hs_cfg_profile->cp_log_max_mrw_sz;
2656 if (val > maxval) {
2657 kmem_free(hca_attr, sizeof (ibt_hca_attr_t));
2658 HERMON_ATTACH_MSG(state->hs_attach_buf,
2659 "soft_state_init_maxmrwsz_toobig_fail");
2660 return (DDI_FAILURE);
2661 }
2662 hca_attr->hca_max_memr_len = ((uint64_t)1 << val);
2663
2664 /* Determine RDMA/Atomic properties */
2665 val = ((uint64_t)1 << state->hs_cfg_profile->cp_log_num_rdb);
2666 hca_attr->hca_max_rsc = (uint_t)val;
2667 val = state->hs_cfg_profile->cp_hca_max_rdma_in_qp;
2668 hca_attr->hca_max_rdma_in_qp = (uint8_t)val;
2669 val = state->hs_cfg_profile->cp_hca_max_rdma_out_qp;
2670 hca_attr->hca_max_rdma_out_qp = (uint8_t)val;
2671 hca_attr->hca_max_rdma_in_ee = 0;
2672 hca_attr->hca_max_rdma_out_ee = 0;
2673
2674 /*
2675 * Determine maximum number of raw IPv6 and Ether QPs. Set to 0
2676 * because neither type of raw QP is supported
2677 */
2678 hca_attr->hca_max_ipv6_qp = 0;
2679 hca_attr->hca_max_ether_qp = 0;
2680
2681 /* Determine max number of MCGs and max QP-per-MCG */
2682 val = ((uint64_t)1 << state->hs_cfg_profile->cp_log_num_qp);
2683 hca_attr->hca_max_mcg_qps = (uint_t)val;
2684 val = ((uint64_t)1 << state->hs_cfg_profile->cp_log_num_mcg);
2685 hca_attr->hca_max_mcg = (uint_t)val;
2686 val = state->hs_cfg_profile->cp_num_qp_per_mcg;
2687 hca_attr->hca_max_qp_per_mcg = (uint_t)val;
2688
2689 /* Determine max number partitions (i.e. PKeys) */
2690 maxval = ((uint64_t)state->hs_cfg_profile->cp_num_ports <<
2691 state->hs_queryport.log_max_pkey);
2692 val = ((uint64_t)state->hs_cfg_profile->cp_num_ports <<
2693 state->hs_cfg_profile->cp_log_max_pkeytbl);
2694
2695 if (val > maxval) {
2696 kmem_free(hca_attr, sizeof (ibt_hca_attr_t));
2697 HERMON_ATTACH_MSG(state->hs_attach_buf,
2698 "soft_state_init_toomanypkey_fail");
2699 return (DDI_FAILURE);
2700 }
2701 hca_attr->hca_max_partitions = (uint16_t)val;
2702
2703 /* Determine number of ports */
2704 maxval = state->hs_devlim.num_ports;
2705 val = state->hs_cfg_profile->cp_num_ports;
2706 if ((val > maxval) || (val == 0)) {
2707 kmem_free(hca_attr, sizeof (ibt_hca_attr_t));
2708 HERMON_ATTACH_MSG(state->hs_attach_buf,
2709 "soft_state_init_toomanyports_fail");
2710 return (DDI_FAILURE);
2711 }
2712 hca_attr->hca_nports = (uint8_t)val;
2713
2714 /* Copy NodeGUID and SystemImageGUID from softstate */
2715 hca_attr->hca_node_guid = state->hs_nodeguid;
2716 hca_attr->hca_si_guid = state->hs_sysimgguid;
2717
2718 /*
2719 * Determine local ACK delay. Use the value suggested by the Hermon
2720 * hardware (from the QUERY_DEV_CAP command)
2721 */
2722 hca_attr->hca_local_ack_delay = state->hs_devlim.ca_ack_delay;
2723
2724 /* Determine max SGID table and PKey table sizes */
2725 val = ((uint64_t)1 << state->hs_cfg_profile->cp_log_max_gidtbl);
2726 hca_attr->hca_max_port_sgid_tbl_sz = (uint_t)val;
2727 val = ((uint64_t)1 << state->hs_cfg_profile->cp_log_max_pkeytbl);
2728 hca_attr->hca_max_port_pkey_tbl_sz = (uint16_t)val;
2729
2730 /* Determine max number of PDs */
2731 maxval = ((uint64_t)1 << state->hs_devlim.log_max_pd);
2732 val = ((uint64_t)1 << state->hs_cfg_profile->cp_log_num_pd);
2733 if (val > maxval) {
2734 kmem_free(hca_attr, sizeof (ibt_hca_attr_t));
2735 HERMON_ATTACH_MSG(state->hs_attach_buf,
2736 "soft_state_init_toomanypd_fail");
2737 return (DDI_FAILURE);
2738 }
2739 hca_attr->hca_max_pd = (uint_t)val;
2740
2741 /* Determine max number of Address Handles (NOT IN ARBEL or HERMON) */
2742 hca_attr->hca_max_ah = 0;
2743
2744 /* No RDDs or EECs (since Reliable Datagram is not supported) */
2745 hca_attr->hca_max_rdd = 0;
2746 hca_attr->hca_max_eec = 0;
2747
2748 /* Initialize lock for reserved UAR page access */
2749 mutex_init(&state->hs_uar_lock, NULL, MUTEX_DRIVER,
2750 DDI_INTR_PRI(state->hs_intrmsi_pri));
2751
2752 /* Initialize the flash fields */
2753 state->hs_fw_flashstarted = 0;
2754 mutex_init(&state->hs_fw_flashlock, NULL, MUTEX_DRIVER,
2755 DDI_INTR_PRI(state->hs_intrmsi_pri));
2756
2757 /* Initialize the lock for the info ioctl */
2758 mutex_init(&state->hs_info_lock, NULL, MUTEX_DRIVER,
2759 DDI_INTR_PRI(state->hs_intrmsi_pri));
2760
2761 /* Initialize the AVL tree for QP number support */
2762 hermon_qpn_avl_init(state);
2763
2764 /* Initialize the cq_sched info structure */
2765 status = hermon_cq_sched_init(state);
2766 if (status != DDI_SUCCESS) {
2767 hermon_qpn_avl_fini(state);
2768 mutex_destroy(&state->hs_info_lock);
2769 mutex_destroy(&state->hs_fw_flashlock);
2770 mutex_destroy(&state->hs_uar_lock);
2771 kmem_free(hca_attr, sizeof (ibt_hca_attr_t));
2772 HERMON_ATTACH_MSG(state->hs_attach_buf,
2773 "soft_state_init_cqsched_init_fail");
2774 return (DDI_FAILURE);
2775 }
2776
2777 /* Initialize the fcoib info structure */
2778 status = hermon_fcoib_init(state);
2779 if (status != DDI_SUCCESS) {
2780 hermon_cq_sched_fini(state);
2781 hermon_qpn_avl_fini(state);
2782 mutex_destroy(&state->hs_info_lock);
2783 mutex_destroy(&state->hs_fw_flashlock);
2784 mutex_destroy(&state->hs_uar_lock);
2785 kmem_free(hca_attr, sizeof (ibt_hca_attr_t));
2786 HERMON_ATTACH_MSG(state->hs_attach_buf,
2787 "soft_state_init_fcoibinit_fail");
2788 return (DDI_FAILURE);
2789 }
2790
2791 /* Initialize the kstat info structure */
2792 status = hermon_kstat_init(state);
2793 if (status != DDI_SUCCESS) {
2794 hermon_fcoib_fini(state);
2795 hermon_cq_sched_fini(state);
2796 hermon_qpn_avl_fini(state);
2797 mutex_destroy(&state->hs_info_lock);
2798 mutex_destroy(&state->hs_fw_flashlock);
2799 mutex_destroy(&state->hs_uar_lock);
2800 kmem_free(hca_attr, sizeof (ibt_hca_attr_t));
2801 HERMON_ATTACH_MSG(state->hs_attach_buf,
2802 "soft_state_init_kstatinit_fail");
2803 return (DDI_FAILURE);
2804 }
2805
2806 return (DDI_SUCCESS);
2807 }
2808
2809
2810 /*
2811 * hermon_soft_state_fini()
2812 * Context: Called only from detach() path context
2813 */
2814 static void
hermon_soft_state_fini(hermon_state_t * state)2815 hermon_soft_state_fini(hermon_state_t *state)
2816 {
2817
2818 /* Teardown the kstat info */
2819 hermon_kstat_fini(state);
2820
2821 /* Teardown the fcoib info */
2822 hermon_fcoib_fini(state);
2823
2824 /* Teardown the cq_sched info */
2825 hermon_cq_sched_fini(state);
2826
2827 /* Teardown the AVL tree for QP number support */
2828 hermon_qpn_avl_fini(state);
2829
2830 /* Free up info ioctl mutex */
2831 mutex_destroy(&state->hs_info_lock);
2832
2833 /* Free up flash mutex */
2834 mutex_destroy(&state->hs_fw_flashlock);
2835
2836 /* Free up the UAR page access mutex */
2837 mutex_destroy(&state->hs_uar_lock);
2838
2839 /* Free up the hca_attr struct */
2840 kmem_free(state->hs_ibtfinfo.hca_attr, sizeof (ibt_hca_attr_t));
2841
2842 }
2843
2844 /*
2845 * hermon_icm_config_setup()
2846 * Context: Only called from attach() path context
2847 */
2848 static int
hermon_icm_config_setup(hermon_state_t * state,hermon_hw_initqueryhca_t * inithca)2849 hermon_icm_config_setup(hermon_state_t *state,
2850 hermon_hw_initqueryhca_t *inithca)
2851 {
2852 hermon_hw_querydevlim_t *devlim;
2853 hermon_cfg_profile_t *cfg;
2854 hermon_icm_table_t *icm_p[HERMON_NUM_ICM_RESOURCES];
2855 hermon_icm_table_t *icm;
2856 hermon_icm_table_t *tmp;
2857 uint64_t icm_addr;
2858 uint64_t icm_size;
2859 int status, i, j;
2860
2861
2862 /* Bring in local devlims, cfg_profile and hs_icm table list */
2863 devlim = &state->hs_devlim;
2864 cfg = state->hs_cfg_profile;
2865 icm = state->hs_icm;
2866
2867 /*
2868 * Assign each ICM table's entry size from data in the devlims,
2869 * except for RDB and MCG sizes, which are not returned in devlims
2870 * but do have a fixed size, and the UAR context entry size, which
2871 * we determine. For this, we use the "cp_num_pgs_per_uce" value
2872 * from our hs_cfg_profile.
2873 */
2874 icm[HERMON_CMPT].object_size = devlim->cmpt_entry_sz;
2875 icm[HERMON_CMPT_QPC].object_size = devlim->cmpt_entry_sz;
2876 icm[HERMON_CMPT_SRQC].object_size = devlim->cmpt_entry_sz;
2877 icm[HERMON_CMPT_CQC].object_size = devlim->cmpt_entry_sz;
2878 icm[HERMON_CMPT_EQC].object_size = devlim->cmpt_entry_sz;
2879 icm[HERMON_MTT].object_size = devlim->mtt_entry_sz;
2880 icm[HERMON_DMPT].object_size = devlim->dmpt_entry_sz;
2881 icm[HERMON_QPC].object_size = devlim->qpc_entry_sz;
2882 icm[HERMON_CQC].object_size = devlim->cqc_entry_sz;
2883 icm[HERMON_SRQC].object_size = devlim->srq_entry_sz;
2884 icm[HERMON_EQC].object_size = devlim->eqc_entry_sz;
2885 icm[HERMON_RDB].object_size = devlim->rdmardc_entry_sz *
2886 cfg->cp_hca_max_rdma_in_qp;
2887 icm[HERMON_MCG].object_size = HERMON_MCGMEM_SZ(state);
2888 icm[HERMON_ALTC].object_size = devlim->altc_entry_sz;
2889 icm[HERMON_AUXC].object_size = devlim->aux_entry_sz;
2890
2891 /* Assign each ICM table's log2 number of entries */
2892 icm[HERMON_CMPT].log_num_entries = cfg->cp_log_num_cmpt;
2893 icm[HERMON_CMPT_QPC].log_num_entries = cfg->cp_log_num_qp;
2894 icm[HERMON_CMPT_SRQC].log_num_entries = cfg->cp_log_num_srq;
2895 icm[HERMON_CMPT_CQC].log_num_entries = cfg->cp_log_num_cq;
2896 icm[HERMON_CMPT_EQC].log_num_entries = HERMON_NUM_EQ_SHIFT;
2897 icm[HERMON_MTT].log_num_entries = cfg->cp_log_num_mtt;
2898 icm[HERMON_DMPT].log_num_entries = cfg->cp_log_num_dmpt;
2899 icm[HERMON_QPC].log_num_entries = cfg->cp_log_num_qp;
2900 icm[HERMON_SRQC].log_num_entries = cfg->cp_log_num_srq;
2901 icm[HERMON_CQC].log_num_entries = cfg->cp_log_num_cq;
2902 icm[HERMON_EQC].log_num_entries = HERMON_NUM_EQ_SHIFT;
2903 icm[HERMON_RDB].log_num_entries = cfg->cp_log_num_qp;
2904 icm[HERMON_MCG].log_num_entries = cfg->cp_log_num_mcg;
2905 icm[HERMON_ALTC].log_num_entries = cfg->cp_log_num_qp;
2906 icm[HERMON_AUXC].log_num_entries = cfg->cp_log_num_qp;
2907
2908 /* Initialize the ICM tables */
2909 hermon_icm_tables_init(state);
2910
2911 /*
2912 * ICM tables must be aligned on their size in the ICM address
2913 * space. So, here we order the tables from largest total table
2914 * size to the smallest. All tables are a power of 2 in size, so
2915 * this will ensure that all tables are aligned on their own size
2916 * without wasting space in the ICM.
2917 *
2918 * In order to easily set the ICM addresses without needing to
2919 * worry about the ordering of our table indices as relates to
2920 * the hermon_rsrc_type_t enum, we will use a list of pointers
2921 * representing the tables for the sort, then assign ICM addresses
2922 * below using it.
2923 */
2924 for (i = 0; i < HERMON_NUM_ICM_RESOURCES; i++) {
2925 icm_p[i] = &icm[i];
2926 }
2927 for (i = HERMON_NUM_ICM_RESOURCES; i > 0; i--) {
2928 switch (i) {
2929 case HERMON_CMPT_QPC:
2930 case HERMON_CMPT_SRQC:
2931 case HERMON_CMPT_CQC:
2932 case HERMON_CMPT_EQC:
2933 continue;
2934 }
2935 for (j = 1; j < i; j++) {
2936 if (icm_p[j]->table_size > icm_p[j - 1]->table_size) {
2937 tmp = icm_p[j];
2938 icm_p[j] = icm_p[j - 1];
2939 icm_p[j - 1] = tmp;
2940 }
2941 }
2942 }
2943
2944 /* Initialize the ICM address and ICM size */
2945 icm_addr = icm_size = 0;
2946
2947 /*
2948 * Set the ICM base address of each table, using our sorted
2949 * list of pointers from above.
2950 */
2951 for (i = 0; i < HERMON_NUM_ICM_RESOURCES; i++) {
2952 j = icm_p[i]->icm_type;
2953 switch (j) {
2954 case HERMON_CMPT_QPC:
2955 case HERMON_CMPT_SRQC:
2956 case HERMON_CMPT_CQC:
2957 case HERMON_CMPT_EQC:
2958 continue;
2959 }
2960 if (icm[j].table_size) {
2961 /*
2962 * Set the ICM base address in the table, save the
2963 * ICM offset in the rsrc pool and increment the
2964 * total ICM allocation.
2965 */
2966 icm[j].icm_baseaddr = icm_addr;
2967 if (hermon_verbose) {
2968 IBTF_DPRINTF_L2("ICMADDR", "rsrc %x @ %p"
2969 " size %llx", j, icm[j].icm_baseaddr,
2970 icm[j].table_size);
2971 }
2972 icm_size += icm[j].table_size;
2973 }
2974
2975 /* Verify that we don't exceed maximum ICM size */
2976 if (icm_size > devlim->max_icm_size) {
2977 /* free the ICM table memory resources */
2978 hermon_icm_tables_fini(state);
2979 cmn_err(CE_WARN, "ICM configuration exceeds maximum "
2980 "configuration: max (0x%lx) requested (0x%lx)\n",
2981 (ulong_t)devlim->max_icm_size, (ulong_t)icm_size);
2982 HERMON_ATTACH_MSG(state->hs_attach_buf,
2983 "icm_config_toobig_fail");
2984 return (DDI_FAILURE);
2985 }
2986
2987 /* assign address to the 4 pieces of the CMPT */
2988 if (j == HERMON_CMPT) {
2989 uint64_t cmpt_size = icm[j].table_size >> 2;
2990 #define init_cmpt_icm_baseaddr(rsrc, indx) \
2991 icm[rsrc].icm_baseaddr = icm_addr + (indx * cmpt_size);
2992 init_cmpt_icm_baseaddr(HERMON_CMPT_QPC, 0);
2993 init_cmpt_icm_baseaddr(HERMON_CMPT_SRQC, 1);
2994 init_cmpt_icm_baseaddr(HERMON_CMPT_CQC, 2);
2995 init_cmpt_icm_baseaddr(HERMON_CMPT_EQC, 3);
2996 }
2997
2998 /* Increment the ICM address for the next table */
2999 icm_addr += icm[j].table_size;
3000 }
3001
3002 /* Populate the structure for the INIT_HCA command */
3003 hermon_inithca_set(state, inithca);
3004
3005 /*
3006 * Prior to invoking INIT_HCA, we must have ICM memory in place
3007 * for the reserved objects in each table. We will allocate and map
3008 * this initial ICM memory here. Note that given the assignment
3009 * of span_size above, tables that are smaller or equal in total
3010 * size to the default span_size will be mapped in full.
3011 */
3012 status = hermon_icm_dma_init(state);
3013 if (status != DDI_SUCCESS) {
3014 /* free the ICM table memory resources */
3015 hermon_icm_tables_fini(state);
3016 HERMON_WARNING(state, "Failed to allocate initial ICM");
3017 HERMON_ATTACH_MSG(state->hs_attach_buf,
3018 "icm_config_dma_init_fail");
3019 return (DDI_FAILURE);
3020 }
3021
3022 return (DDI_SUCCESS);
3023 }
3024
3025 /*
3026 * hermon_inithca_set()
3027 * Context: Only called from attach() path context
3028 */
3029 static void
hermon_inithca_set(hermon_state_t * state,hermon_hw_initqueryhca_t * inithca)3030 hermon_inithca_set(hermon_state_t *state, hermon_hw_initqueryhca_t *inithca)
3031 {
3032 hermon_cfg_profile_t *cfg;
3033 hermon_icm_table_t *icm;
3034 int i;
3035
3036
3037 /* Populate the INIT_HCA structure */
3038 icm = state->hs_icm;
3039 cfg = state->hs_cfg_profile;
3040
3041 /* set version */
3042 inithca->version = 0x02; /* PRM 0.36 */
3043 /* set cacheline - log2 in 16-byte chunks */
3044 inithca->log2_cacheline = 0x2; /* optimized for 64 byte cache */
3045
3046 /* we need to update the inithca info with thie UAR info too */
3047 inithca->uar.log_max_uars = highbit(cfg->cp_log_num_uar);
3048 inithca->uar.uar_pg_sz = PAGESHIFT - HERMON_PAGESHIFT;
3049
3050 /* Set endianess */
3051 #ifdef _LITTLE_ENDIAN
3052 inithca->big_endian = 0;
3053 #else
3054 inithca->big_endian = 1;
3055 #endif
3056
3057 /* Port Checking is on by default */
3058 inithca->udav_port_chk = HERMON_UDAV_PORTCHK_ENABLED;
3059
3060 /* Enable IPoIB checksum */
3061 if (state->hs_devlim.ipoib_cksm)
3062 inithca->chsum_en = 1;
3063
3064 /* Set each ICM table's attributes */
3065 for (i = 0; i < HERMON_NUM_ICM_RESOURCES; i++) {
3066 switch (icm[i].icm_type) {
3067 case HERMON_CMPT:
3068 inithca->tpt.cmpt_baseaddr = icm[i].icm_baseaddr;
3069 break;
3070
3071 case HERMON_MTT:
3072 inithca->tpt.mtt_baseaddr = icm[i].icm_baseaddr;
3073 break;
3074
3075 case HERMON_DMPT:
3076 inithca->tpt.dmpt_baseaddr = icm[i].icm_baseaddr;
3077 inithca->tpt.log_dmpt_sz = icm[i].log_num_entries;
3078 inithca->tpt.pgfault_rnr_to = 0; /* just in case */
3079 break;
3080
3081 case HERMON_QPC:
3082 inithca->context.log_num_qp = icm[i].log_num_entries;
3083 inithca->context.qpc_baseaddr_h =
3084 icm[i].icm_baseaddr >> 32;
3085 inithca->context.qpc_baseaddr_l =
3086 (icm[i].icm_baseaddr & 0xFFFFFFFF) >> 5;
3087 break;
3088
3089 case HERMON_CQC:
3090 inithca->context.log_num_cq = icm[i].log_num_entries;
3091 inithca->context.cqc_baseaddr_h =
3092 icm[i].icm_baseaddr >> 32;
3093 inithca->context.cqc_baseaddr_l =
3094 (icm[i].icm_baseaddr & 0xFFFFFFFF) >> 5;
3095 break;
3096
3097 case HERMON_SRQC:
3098 inithca->context.log_num_srq = icm[i].log_num_entries;
3099 inithca->context.srqc_baseaddr_h =
3100 icm[i].icm_baseaddr >> 32;
3101 inithca->context.srqc_baseaddr_l =
3102 (icm[i].icm_baseaddr & 0xFFFFFFFF) >> 5;
3103 break;
3104
3105 case HERMON_EQC:
3106 inithca->context.log_num_eq = icm[i].log_num_entries;
3107 inithca->context.eqc_baseaddr_h =
3108 icm[i].icm_baseaddr >> 32;
3109 inithca->context.eqc_baseaddr_l =
3110 (icm[i].icm_baseaddr & 0xFFFFFFFF) >> 5;
3111 break;
3112
3113 case HERMON_RDB:
3114 inithca->context.rdmardc_baseaddr_h =
3115 icm[i].icm_baseaddr >> 32;
3116 inithca->context.rdmardc_baseaddr_l =
3117 (icm[i].icm_baseaddr & 0xFFFFFFFF) >> 5;
3118 inithca->context.log_num_rdmardc =
3119 cfg->cp_log_num_rdb - cfg->cp_log_num_qp;
3120 break;
3121
3122 case HERMON_MCG:
3123 inithca->multi.mc_baseaddr = icm[i].icm_baseaddr;
3124 inithca->multi.log_mc_tbl_sz = icm[i].log_num_entries;
3125 inithca->multi.log_mc_tbl_ent =
3126 highbit(HERMON_MCGMEM_SZ(state)) - 1;
3127 inithca->multi.log_mc_tbl_hash_sz =
3128 cfg->cp_log_num_mcg_hash;
3129 inithca->multi.mc_hash_fn = HERMON_MCG_DEFAULT_HASH_FN;
3130 break;
3131
3132 case HERMON_ALTC:
3133 inithca->context.altc_baseaddr = icm[i].icm_baseaddr;
3134 break;
3135
3136 case HERMON_AUXC:
3137 inithca->context.auxc_baseaddr = icm[i].icm_baseaddr;
3138 break;
3139
3140 default:
3141 break;
3142
3143 }
3144 }
3145
3146 }
3147
3148 /*
3149 * hermon_icm_tables_init()
3150 * Context: Only called from attach() path context
3151 *
3152 * Dynamic ICM breaks the various ICM tables into "span_size" chunks
3153 * to enable allocation of backing memory on demand. Arbel used a
3154 * fixed size ARBEL_ICM_SPAN_SIZE (initially was 512KB) as the
3155 * span_size for all ICM chunks. Hermon has other considerations,
3156 * so the span_size used differs from Arbel.
3157 *
3158 * The basic considerations for why Hermon differs are:
3159 *
3160 * 1) ICM memory is in units of HERMON pages.
3161 *
3162 * 2) The AUXC table is approximately 1 byte per QP.
3163 *
3164 * 3) ICM memory for AUXC, ALTC, and RDB is allocated when
3165 * the ICM memory for the corresponding QPC is allocated.
3166 *
3167 * 4) ICM memory for the CMPT corresponding to the various primary
3168 * resources (QPC, SRQC, CQC, and EQC) is allocated when the ICM
3169 * memory for the primary resource is allocated.
3170 *
3171 * One HERMON page (4KB) would typically map 4K QPs worth of AUXC.
3172 * So, the minimum chunk for the various QPC related ICM memory should
3173 * all be allocated to support the 4K QPs. Currently, this means the
3174 * amount of memory for the various QP chunks is:
3175 *
3176 * QPC 256*4K bytes
3177 * RDB 128*4K bytes
3178 * CMPT 64*4K bytes
3179 * ALTC 64*4K bytes
3180 * AUXC 1*4K bytes
3181 *
3182 * The span_size chosen for the QP resource is 4KB of AUXC entries,
3183 * or 1 HERMON_PAGESIZE worth, which is the minimum ICM mapping size.
3184 *
3185 * Other ICM resources can have their span_size be more arbitrary.
3186 * This is 4K (HERMON_ICM_SPAN), except for MTTs because they are tiny.
3187 */
3188
3189 /* macro to make the code below cleaner */
3190 #define init_dependent(rsrc, dep) \
3191 icm[dep].span = icm[rsrc].span; \
3192 icm[dep].num_spans = icm[rsrc].num_spans; \
3193 icm[dep].split_shift = icm[rsrc].split_shift; \
3194 icm[dep].span_mask = icm[rsrc].span_mask; \
3195 icm[dep].span_shift = icm[rsrc].span_shift; \
3196 icm[dep].rsrc_mask = icm[rsrc].rsrc_mask; \
3197 if (hermon_verbose) { \
3198 IBTF_DPRINTF_L2("hermon", "tables_init: " \
3199 "rsrc (0x%x) size (0x%lx) span (0x%x) " \
3200 "num_spans (0x%x)", dep, icm[dep].table_size, \
3201 icm[dep].span, icm[dep].num_spans); \
3202 IBTF_DPRINTF_L2("hermon", "tables_init: " \
3203 "span_shift (0x%x) split_shift (0x%x)", \
3204 icm[dep].span_shift, icm[dep].split_shift); \
3205 IBTF_DPRINTF_L2("hermon", "tables_init: " \
3206 "span_mask (0x%x) rsrc_mask (0x%x)", \
3207 icm[dep].span_mask, icm[dep].rsrc_mask); \
3208 }
3209
3210 static void
hermon_icm_tables_init(hermon_state_t * state)3211 hermon_icm_tables_init(hermon_state_t *state)
3212 {
3213 hermon_icm_table_t *icm;
3214 int i, k;
3215 uint32_t per_split;
3216
3217
3218 icm = state->hs_icm;
3219
3220 for (i = 0; i < HERMON_NUM_ICM_RESOURCES; i++) {
3221 icm[i].icm_type = i;
3222 icm[i].num_entries = 1 << icm[i].log_num_entries;
3223 icm[i].log_object_size = highbit(icm[i].object_size) - 1;
3224 icm[i].table_size = icm[i].num_entries <<
3225 icm[i].log_object_size;
3226
3227 /* deal with "dependent" resource types */
3228 switch (i) {
3229 case HERMON_AUXC:
3230 #ifdef HERMON_FW_WORKAROUND
3231 icm[i].table_size = 0x80000000ull;
3232 #endif
3233 /* FALLTHROUGH */
3234 case HERMON_CMPT_QPC:
3235 case HERMON_RDB:
3236 case HERMON_ALTC:
3237 init_dependent(HERMON_QPC, i);
3238 continue;
3239 case HERMON_CMPT_SRQC:
3240 init_dependent(HERMON_SRQC, i);
3241 continue;
3242 case HERMON_CMPT_CQC:
3243 init_dependent(HERMON_CQC, i);
3244 continue;
3245 case HERMON_CMPT_EQC:
3246 init_dependent(HERMON_EQC, i);
3247 continue;
3248 }
3249
3250 icm[i].span = HERMON_ICM_SPAN; /* default #rsrc's in 1 span */
3251 if (i == HERMON_MTT) /* Alloc enough MTTs to map 256MB */
3252 icm[i].span = HERMON_ICM_SPAN * 16;
3253 icm[i].num_spans = icm[i].num_entries / icm[i].span;
3254 if (icm[i].num_spans == 0) {
3255 icm[i].span = icm[i].num_entries;
3256 per_split = 1;
3257 icm[i].num_spans = icm[i].num_entries / icm[i].span;
3258 } else {
3259 per_split = icm[i].num_spans / HERMON_ICM_SPLIT;
3260 if (per_split == 0) {
3261 per_split = 1;
3262 }
3263 }
3264 if (hermon_verbose)
3265 IBTF_DPRINTF_L2("ICM", "rsrc %x span %x num_spans %x",
3266 i, icm[i].span, icm[i].num_spans);
3267
3268 /*
3269 * Ensure a minimum table size of an ICM page, and a
3270 * maximum span size of the ICM table size. This ensures
3271 * that we don't have less than an ICM page to map, which is
3272 * impossible, and that we will map an entire table at
3273 * once if it's total size is less than the span size.
3274 */
3275 icm[i].table_size = max(icm[i].table_size, HERMON_PAGESIZE);
3276
3277 icm[i].span_shift = 0;
3278 for (k = icm[i].span; k != 1; k >>= 1)
3279 icm[i].span_shift++;
3280 icm[i].split_shift = icm[i].span_shift;
3281 for (k = per_split; k != 1; k >>= 1)
3282 icm[i].split_shift++;
3283 icm[i].span_mask = (1 << icm[i].split_shift) -
3284 (1 << icm[i].span_shift);
3285 icm[i].rsrc_mask = (1 << icm[i].span_shift) - 1;
3286
3287
3288 /* Initialize the table lock */
3289 mutex_init(&icm[i].icm_table_lock, NULL, MUTEX_DRIVER,
3290 DDI_INTR_PRI(state->hs_intrmsi_pri));
3291 cv_init(&icm[i].icm_table_cv, NULL, CV_DRIVER, NULL);
3292
3293 if (hermon_verbose) {
3294 IBTF_DPRINTF_L2("hermon", "tables_init: "
3295 "rsrc (0x%x) size (0x%lx)", i, icm[i].table_size);
3296 IBTF_DPRINTF_L2("hermon", "tables_init: "
3297 "span (0x%x) num_spans (0x%x)",
3298 icm[i].span, icm[i].num_spans);
3299 IBTF_DPRINTF_L2("hermon", "tables_init: "
3300 "span_shift (0x%x) split_shift (0x%x)",
3301 icm[i].span_shift, icm[i].split_shift);
3302 IBTF_DPRINTF_L2("hermon", "tables_init: "
3303 "span_mask (0x%x) rsrc_mask (0x%x)",
3304 icm[i].span_mask, icm[i].rsrc_mask);
3305 }
3306 }
3307
3308 }
3309
3310 /*
3311 * hermon_icm_tables_fini()
3312 * Context: Only called from attach() path context
3313 *
3314 * Clean up all icm_tables. Free the bitmap and dma_info arrays.
3315 */
3316 static void
hermon_icm_tables_fini(hermon_state_t * state)3317 hermon_icm_tables_fini(hermon_state_t *state)
3318 {
3319 hermon_icm_table_t *icm;
3320 int nspans;
3321 int i, j;
3322
3323
3324 icm = state->hs_icm;
3325
3326 for (i = 0; i < HERMON_NUM_ICM_RESOURCES; i++) {
3327
3328 mutex_enter(&icm[i].icm_table_lock);
3329 nspans = icm[i].num_spans;
3330
3331 for (j = 0; j < HERMON_ICM_SPLIT; j++) {
3332 if (icm[i].icm_dma[j])
3333 /* Free the ICM DMA slots */
3334 kmem_free(icm[i].icm_dma[j],
3335 nspans * sizeof (hermon_dma_info_t));
3336
3337 if (icm[i].icm_bitmap[j])
3338 /* Free the table bitmap */
3339 kmem_free(icm[i].icm_bitmap[j],
3340 (nspans + 7) / 8);
3341 }
3342 /* Destroy the table lock */
3343 cv_destroy(&icm[i].icm_table_cv);
3344 mutex_exit(&icm[i].icm_table_lock);
3345 mutex_destroy(&icm[i].icm_table_lock);
3346 }
3347
3348 }
3349
3350 /*
3351 * hermon_icm_dma_init()
3352 * Context: Only called from attach() path context
3353 */
3354 static int
hermon_icm_dma_init(hermon_state_t * state)3355 hermon_icm_dma_init(hermon_state_t *state)
3356 {
3357 hermon_icm_table_t *icm;
3358 hermon_rsrc_type_t type;
3359 int status;
3360
3361
3362 /*
3363 * This routine will allocate initial ICM DMA resources for ICM
3364 * tables that have reserved ICM objects. This is the only routine
3365 * where we should have to allocate ICM outside of hermon_rsrc_alloc().
3366 * We need to allocate ICM here explicitly, rather than in
3367 * hermon_rsrc_alloc(), because we've not yet completed the resource
3368 * pool initialization. When the resource pools are initialized
3369 * (in hermon_rsrc_init_phase2(), see hermon_rsrc.c for more
3370 * information), resource preallocations will be invoked to match
3371 * the ICM allocations seen here. We will then be able to use the
3372 * normal allocation path. Note we don't need to set a refcnt on
3373 * these initial allocations because that will be done in the calls
3374 * to hermon_rsrc_alloc() from hermon_hw_entries_init() for the
3375 * "prealloc" objects (see hermon_rsrc.c for more information).
3376 */
3377 for (type = 0; type < HERMON_NUM_ICM_RESOURCES; type++) {
3378
3379 /* ICM for these is allocated within hermon_icm_alloc() */
3380 switch (type) {
3381 case HERMON_CMPT:
3382 case HERMON_CMPT_QPC:
3383 case HERMON_CMPT_SRQC:
3384 case HERMON_CMPT_CQC:
3385 case HERMON_CMPT_EQC:
3386 case HERMON_AUXC:
3387 case HERMON_ALTC:
3388 case HERMON_RDB:
3389 continue;
3390 }
3391
3392 icm = &state->hs_icm[type];
3393
3394 mutex_enter(&icm->icm_table_lock);
3395 status = hermon_icm_alloc(state, type, 0, 0);
3396 mutex_exit(&icm->icm_table_lock);
3397 if (status != DDI_SUCCESS) {
3398 while (type--) {
3399 icm = &state->hs_icm[type];
3400 mutex_enter(&icm->icm_table_lock);
3401 hermon_icm_free(state, type, 0, 0);
3402 mutex_exit(&icm->icm_table_lock);
3403 }
3404 return (DDI_FAILURE);
3405 }
3406
3407 if (hermon_verbose) {
3408 IBTF_DPRINTF_L2("hermon", "hermon_icm_dma_init: "
3409 "table (0x%x) index (0x%x) allocated", type, 0);
3410 }
3411 }
3412
3413 return (DDI_SUCCESS);
3414 }
3415
3416 /*
3417 * hermon_icm_dma_fini()
3418 * Context: Only called from attach() path context
3419 *
3420 * ICM has been completely unmapped. We just free the memory here.
3421 */
3422 static void
hermon_icm_dma_fini(hermon_state_t * state)3423 hermon_icm_dma_fini(hermon_state_t *state)
3424 {
3425 hermon_icm_table_t *icm;
3426 hermon_dma_info_t *dma_info;
3427 hermon_rsrc_type_t type;
3428 int index1, index2;
3429
3430
3431 for (type = 0; type < HERMON_NUM_ICM_RESOURCES; type++) {
3432 icm = &state->hs_icm[type];
3433 for (index1 = 0; index1 < HERMON_ICM_SPLIT; index1++) {
3434 dma_info = icm->icm_dma[index1];
3435 if (dma_info == NULL)
3436 continue;
3437 for (index2 = 0; index2 < icm->num_spans; index2++) {
3438 if (dma_info[index2].dma_hdl)
3439 hermon_dma_free(&dma_info[index2]);
3440 dma_info[index2].dma_hdl = NULL;
3441 }
3442 }
3443 }
3444
3445 }
3446
3447 /*
3448 * hermon_hca_port_init()
3449 * Context: Only called from attach() path context
3450 */
3451 static int
hermon_hca_port_init(hermon_state_t * state)3452 hermon_hca_port_init(hermon_state_t *state)
3453 {
3454 hermon_hw_set_port_t *portinits, *initport;
3455 hermon_cfg_profile_t *cfgprof;
3456 uint_t num_ports;
3457 int i = 0, status;
3458 uint64_t maxval, val;
3459 uint64_t sysimgguid, nodeguid, portguid;
3460
3461
3462 cfgprof = state->hs_cfg_profile;
3463
3464 /* Get number of HCA ports */
3465 num_ports = cfgprof->cp_num_ports;
3466
3467 /* Allocate space for Hermon set port struct(s) */
3468 portinits = (hermon_hw_set_port_t *)kmem_zalloc(num_ports *
3469 sizeof (hermon_hw_set_port_t), KM_SLEEP);
3470
3471
3472
3473 /* Post commands to initialize each Hermon HCA port */
3474 /*
3475 * In Hermon, the process is different than in previous HCAs.
3476 * Here, you have to:
3477 * QUERY_PORT - to get basic information from the HCA
3478 * set the fields accordingly
3479 * SET_PORT - to change/set everything as desired
3480 * INIT_PORT - to bring the port up
3481 *
3482 * Needs to be done for each port in turn
3483 */
3484
3485 for (i = 0; i < num_ports; i++) {
3486 bzero(&state->hs_queryport, sizeof (hermon_hw_query_port_t));
3487 status = hermon_cmn_query_cmd_post(state, QUERY_PORT, 0,
3488 (i + 1), &state->hs_queryport,
3489 sizeof (hermon_hw_query_port_t), HERMON_CMD_NOSLEEP_SPIN);
3490 if (status != HERMON_CMD_SUCCESS) {
3491 cmn_err(CE_CONT, "Hermon: QUERY_PORT (port %02d) "
3492 "command failed: %08x\n", i + 1, status);
3493 goto init_ports_fail;
3494 }
3495 initport = &portinits[i];
3496 state->hs_initport = &portinits[i];
3497
3498 bzero(initport, sizeof (hermon_hw_query_port_t));
3499
3500 /*
3501 * Determine whether we need to override the firmware's
3502 * default SystemImageGUID setting.
3503 */
3504 sysimgguid = cfgprof->cp_sysimgguid;
3505 if (sysimgguid != 0) {
3506 initport->sig = 1;
3507 initport->sys_img_guid = sysimgguid;
3508 }
3509
3510 /*
3511 * Determine whether we need to override the firmware's
3512 * default NodeGUID setting.
3513 */
3514 nodeguid = cfgprof->cp_nodeguid;
3515 if (nodeguid != 0) {
3516 initport->ng = 1;
3517 initport->node_guid = nodeguid;
3518 }
3519
3520 /*
3521 * Determine whether we need to override the firmware's
3522 * default PortGUID setting.
3523 */
3524 portguid = cfgprof->cp_portguid[i];
3525 if (portguid != 0) {
3526 initport->g0 = 1;
3527 initport->guid0 = portguid;
3528 }
3529
3530 /* Validate max MTU size */
3531 maxval = state->hs_queryport.ib_mtu;
3532 val = cfgprof->cp_max_mtu;
3533 if (val > maxval) {
3534 goto init_ports_fail;
3535 }
3536
3537 /* Set mtu_cap to 4096 bytes */
3538 initport->mmc = 1; /* set the change bit */
3539 initport->mtu_cap = 5; /* for 4096 bytes */
3540
3541 /* Validate the max port width */
3542 maxval = state->hs_queryport.ib_port_wid;
3543 val = cfgprof->cp_max_port_width;
3544 if (val > maxval) {
3545 goto init_ports_fail;
3546 }
3547
3548 /* Validate max VL cap size */
3549 maxval = state->hs_queryport.max_vl;
3550 val = cfgprof->cp_max_vlcap;
3551 if (val > maxval) {
3552 goto init_ports_fail;
3553 }
3554
3555 /* Since we're doing mtu_cap, cut vl_cap down */
3556 initport->mvc = 1; /* set this change bit */
3557 initport->vl_cap = 3; /* 3 means vl0-vl3, 4 total */
3558
3559 /* Validate max GID table size */
3560 maxval = ((uint64_t)1 << state->hs_queryport.log_max_gid);
3561 val = ((uint64_t)1 << cfgprof->cp_log_max_gidtbl);
3562 if (val > maxval) {
3563 goto init_ports_fail;
3564 }
3565 initport->max_gid = (uint16_t)val;
3566 initport->mg = 1;
3567
3568 /* Validate max PKey table size */
3569 maxval = ((uint64_t)1 << state->hs_queryport.log_max_pkey);
3570 val = ((uint64_t)1 << cfgprof->cp_log_max_pkeytbl);
3571 if (val > maxval) {
3572 goto init_ports_fail;
3573 }
3574 initport->max_pkey = (uint16_t)val;
3575 initport->mp = 1;
3576 /*
3577 * Post the SET_PORT cmd to Hermon firmware. This sets
3578 * the parameters of the port.
3579 */
3580 status = hermon_set_port_cmd_post(state, initport, i + 1,
3581 HERMON_CMD_NOSLEEP_SPIN);
3582 if (status != HERMON_CMD_SUCCESS) {
3583 cmn_err(CE_CONT, "Hermon: SET_PORT (port %02d) command "
3584 "failed: %08x\n", i + 1, status);
3585 goto init_ports_fail;
3586 }
3587 /* issue another SET_PORT cmd - performance fix/workaround */
3588 /* XXX - need to discuss with Mellanox */
3589 bzero(initport, sizeof (hermon_hw_query_port_t));
3590 initport->cap_mask = 0x02500868;
3591 status = hermon_set_port_cmd_post(state, initport, i + 1,
3592 HERMON_CMD_NOSLEEP_SPIN);
3593 if (status != HERMON_CMD_SUCCESS) {
3594 cmn_err(CE_CONT, "Hermon: SET_PORT (port %02d) command "
3595 "failed: %08x\n", i + 1, status);
3596 goto init_ports_fail;
3597 }
3598 }
3599
3600 /*
3601 * Finally, do the INIT_PORT for each port in turn
3602 * When this command completes, the corresponding Hermon port
3603 * will be physically "Up" and initialized.
3604 */
3605 for (i = 0; i < num_ports; i++) {
3606 status = hermon_init_port_cmd_post(state, i + 1,
3607 HERMON_CMD_NOSLEEP_SPIN);
3608 if (status != HERMON_CMD_SUCCESS) {
3609 cmn_err(CE_CONT, "Hermon: INIT_PORT (port %02d) "
3610 "comman failed: %08x\n", i + 1, status);
3611 goto init_ports_fail;
3612 }
3613 }
3614
3615 /* Free up the memory for Hermon port init struct(s), return success */
3616 kmem_free(portinits, num_ports * sizeof (hermon_hw_set_port_t));
3617 return (DDI_SUCCESS);
3618
3619 init_ports_fail:
3620 /*
3621 * Free up the memory for Hermon port init struct(s), shutdown any
3622 * successfully initialized ports, and return failure
3623 */
3624 kmem_free(portinits, num_ports * sizeof (hermon_hw_set_port_t));
3625 (void) hermon_hca_ports_shutdown(state, i);
3626
3627 return (DDI_FAILURE);
3628 }
3629
3630
3631 /*
3632 * hermon_hca_ports_shutdown()
3633 * Context: Only called from attach() and/or detach() path contexts
3634 */
3635 static int
hermon_hca_ports_shutdown(hermon_state_t * state,uint_t num_init)3636 hermon_hca_ports_shutdown(hermon_state_t *state, uint_t num_init)
3637 {
3638 int i, status;
3639
3640 /*
3641 * Post commands to shutdown all init'd Hermon HCA ports. Note: if
3642 * any of these commands fail for any reason, it would be entirely
3643 * unexpected and probably indicative a serious problem (HW or SW).
3644 * Although we do return void from this function, this type of failure
3645 * should not go unreported. That is why we have the warning message.
3646 */
3647 for (i = 0; i < num_init; i++) {
3648 status = hermon_close_port_cmd_post(state, i + 1,
3649 HERMON_CMD_NOSLEEP_SPIN);
3650 if (status != HERMON_CMD_SUCCESS) {
3651 HERMON_WARNING(state, "failed to shutdown HCA port");
3652 return (status);
3653 }
3654 }
3655 return (HERMON_CMD_SUCCESS);
3656 }
3657
3658
3659 /*
3660 * hermon_internal_uarpg_init
3661 * Context: Only called from attach() path context
3662 */
3663 static int
hermon_internal_uarpg_init(hermon_state_t * state)3664 hermon_internal_uarpg_init(hermon_state_t *state)
3665 {
3666 int status;
3667 hermon_dbr_info_t *info;
3668
3669 /*
3670 * Allocate the UAR page for kernel use. This UAR page is
3671 * the privileged UAR page through which all kernel generated
3672 * doorbells will be rung. There are a number of UAR pages
3673 * reserved by hardware at the front of the UAR BAR, indicated
3674 * by DEVCAP.num_rsvd_uar, which we have already allocated. So,
3675 * the kernel page, or UAR page index num_rsvd_uar, will be
3676 * allocated here for kernel use.
3677 */
3678
3679 status = hermon_rsrc_alloc(state, HERMON_UARPG, 1, HERMON_SLEEP,
3680 &state->hs_uarkpg_rsrc);
3681 if (status != DDI_SUCCESS) {
3682 return (DDI_FAILURE);
3683 }
3684
3685 /* Setup pointer to kernel UAR page */
3686 state->hs_uar = (hermon_hw_uar_t *)state->hs_uarkpg_rsrc->hr_addr;
3687
3688 /* need to set up DBr tracking as well */
3689 status = hermon_dbr_page_alloc(state, &info);
3690 if (status != DDI_SUCCESS) {
3691 return (DDI_FAILURE);
3692 }
3693 state->hs_kern_dbr = info;
3694 return (DDI_SUCCESS);
3695 }
3696
3697
3698 /*
3699 * hermon_internal_uarpg_fini
3700 * Context: Only called from attach() and/or detach() path contexts
3701 */
3702 static void
hermon_internal_uarpg_fini(hermon_state_t * state)3703 hermon_internal_uarpg_fini(hermon_state_t *state)
3704 {
3705 /* Free up Hermon UAR page #1 (kernel driver doorbells) */
3706 hermon_rsrc_free(state, &state->hs_uarkpg_rsrc);
3707 }
3708
3709
3710 /*
3711 * hermon_special_qp_contexts_reserve()
3712 * Context: Only called from attach() path context
3713 */
3714 static int
hermon_special_qp_contexts_reserve(hermon_state_t * state)3715 hermon_special_qp_contexts_reserve(hermon_state_t *state)
3716 {
3717 hermon_rsrc_t *qp0_rsrc, *qp1_rsrc, *qp_resvd;
3718 int status;
3719
3720 /* Initialize the lock used for special QP rsrc management */
3721 mutex_init(&state->hs_spec_qplock, NULL, MUTEX_DRIVER,
3722 DDI_INTR_PRI(state->hs_intrmsi_pri));
3723
3724 /*
3725 * Reserve contexts for QP0. These QP contexts will be setup to
3726 * act as aliases for the real QP0. Note: We are required to grab
3727 * two QPs (one per port) even if we are operating in single-port
3728 * mode.
3729 */
3730 status = hermon_rsrc_alloc(state, HERMON_QPC, 2,
3731 HERMON_SLEEP, &qp0_rsrc);
3732 if (status != DDI_SUCCESS) {
3733 mutex_destroy(&state->hs_spec_qplock);
3734 return (DDI_FAILURE);
3735 }
3736 state->hs_spec_qp0 = qp0_rsrc;
3737
3738 /*
3739 * Reserve contexts for QP1. These QP contexts will be setup to
3740 * act as aliases for the real QP1. Note: We are required to grab
3741 * two QPs (one per port) even if we are operating in single-port
3742 * mode.
3743 */
3744 status = hermon_rsrc_alloc(state, HERMON_QPC, 2,
3745 HERMON_SLEEP, &qp1_rsrc);
3746 if (status != DDI_SUCCESS) {
3747 hermon_rsrc_free(state, &qp0_rsrc);
3748 mutex_destroy(&state->hs_spec_qplock);
3749 return (DDI_FAILURE);
3750 }
3751 state->hs_spec_qp1 = qp1_rsrc;
3752
3753 status = hermon_rsrc_alloc(state, HERMON_QPC, 4,
3754 HERMON_SLEEP, &qp_resvd);
3755 if (status != DDI_SUCCESS) {
3756 hermon_rsrc_free(state, &qp1_rsrc);
3757 hermon_rsrc_free(state, &qp0_rsrc);
3758 mutex_destroy(&state->hs_spec_qplock);
3759 return (DDI_FAILURE);
3760 }
3761 state->hs_spec_qp_unused = qp_resvd;
3762
3763 return (DDI_SUCCESS);
3764 }
3765
3766
3767 /*
3768 * hermon_special_qp_contexts_unreserve()
3769 * Context: Only called from attach() and/or detach() path contexts
3770 */
3771 static void
hermon_special_qp_contexts_unreserve(hermon_state_t * state)3772 hermon_special_qp_contexts_unreserve(hermon_state_t *state)
3773 {
3774
3775 /* Unreserve contexts for spec_qp_unused */
3776 hermon_rsrc_free(state, &state->hs_spec_qp_unused);
3777
3778 /* Unreserve contexts for QP1 */
3779 hermon_rsrc_free(state, &state->hs_spec_qp1);
3780
3781 /* Unreserve contexts for QP0 */
3782 hermon_rsrc_free(state, &state->hs_spec_qp0);
3783
3784 /* Destroy the lock used for special QP rsrc management */
3785 mutex_destroy(&state->hs_spec_qplock);
3786
3787 }
3788
3789
3790 /*
3791 * hermon_sw_reset()
3792 * Context: Currently called only from attach() path context
3793 */
3794 static int
hermon_sw_reset(hermon_state_t * state)3795 hermon_sw_reset(hermon_state_t *state)
3796 {
3797 ddi_acc_handle_t hdl = hermon_get_pcihdl(state);
3798 ddi_acc_handle_t cmdhdl = hermon_get_cmdhdl(state);
3799 uint32_t reset_delay;
3800 int status, i;
3801 uint32_t sem;
3802 uint_t offset;
3803 uint32_t data32; /* for devctl & linkctl */
3804 int loopcnt;
3805
3806 /* initialize the FMA retry loop */
3807 hermon_pio_init(fm_loop_cnt, fm_status, fm_test);
3808 hermon_pio_init(fm_loop_cnt2, fm_status2, fm_test2);
3809
3810 /*
3811 * If the configured software reset delay is set to zero, then we
3812 * will not attempt a software reset of the Hermon device.
3813 */
3814 reset_delay = state->hs_cfg_profile->cp_sw_reset_delay;
3815 if (reset_delay == 0) {
3816 return (DDI_SUCCESS);
3817 }
3818
3819 /* the FMA retry loop starts. */
3820 hermon_pio_start(state, cmdhdl, pio_error, fm_loop_cnt, fm_status,
3821 fm_test);
3822 hermon_pio_start(state, hdl, pio_error2, fm_loop_cnt2, fm_status2,
3823 fm_test2);
3824
3825 /* Query the PCI capabilities of the HCA device */
3826 /* but don't process the VPD until after reset */
3827 status = hermon_pci_capability_list(state, hdl);
3828 if (status != DDI_SUCCESS) {
3829 cmn_err(CE_NOTE, "failed to get pci capabilities list(0x%x)\n",
3830 status);
3831 return (DDI_FAILURE);
3832 }
3833
3834 /*
3835 * Read all PCI config info (reg0...reg63). Note: According to the
3836 * Hermon software reset application note, we should not read or
3837 * restore the values in reg22 and reg23.
3838 * NOTE: For Hermon (and Arbel too) it says to restore the command
3839 * register LAST, and technically, you need to restore the
3840 * PCIE Capability "device control" and "link control" (word-sized,
3841 * at offsets 0x08 and 0x10 from the capbility ID respectively).
3842 * We hold off restoring the command register - offset 0x4 - till last
3843 */
3844
3845 /* 1st, wait for the semaphore assure accessibility - per PRM */
3846 status = -1;
3847 for (i = 0; i < NANOSEC/MICROSEC /* 1sec timeout */; i++) {
3848 sem = ddi_get32(cmdhdl, state->hs_cmd_regs.sw_semaphore);
3849 if (sem == 0) {
3850 status = 0;
3851 break;
3852 }
3853 drv_usecwait(1);
3854 }
3855
3856 /* Check if timeout happens */
3857 if (status == -1) {
3858 /*
3859 * Remove this acc handle from Hermon, then log
3860 * the error.
3861 */
3862 hermon_pci_config_teardown(state, &hdl);
3863
3864 cmn_err(CE_WARN, "hermon_sw_reset timeout: "
3865 "failed to get the semaphore(0x%p)\n",
3866 (void *)state->hs_cmd_regs.sw_semaphore);
3867
3868 hermon_fm_ereport(state, HCA_IBA_ERR, HCA_ERR_NON_FATAL);
3869 return (DDI_FAILURE);
3870 }
3871
3872 for (i = 0; i < HERMON_SW_RESET_NUMREGS; i++) {
3873 if ((i != HERMON_SW_RESET_REG22_RSVD) &&
3874 (i != HERMON_SW_RESET_REG23_RSVD)) {
3875 state->hs_cfg_data[i] = pci_config_get32(hdl, i << 2);
3876 }
3877 }
3878
3879 /*
3880 * Perform the software reset (by writing 1 at offset 0xF0010)
3881 */
3882 ddi_put32(cmdhdl, state->hs_cmd_regs.sw_reset, HERMON_SW_RESET_START);
3883
3884 /*
3885 * This delay is required so as not to cause a panic here. If the
3886 * device is accessed too soon after reset it will not respond to
3887 * config cycles, causing a Master Abort and panic.
3888 */
3889 drv_usecwait(reset_delay);
3890
3891 /*
3892 * Poll waiting for the device to finish resetting.
3893 */
3894 loopcnt = 100; /* 100 times @ 100 usec - total delay 10 msec */
3895 while ((pci_config_get32(hdl, 0) & 0x0000FFFF) != PCI_VENID_MLX) {
3896 drv_usecwait(HERMON_SW_RESET_POLL_DELAY);
3897 if (--loopcnt == 0)
3898 break; /* just in case, break and go on */
3899 }
3900 if (loopcnt == 0)
3901 cmn_err(CE_CONT, "!Never see VEND_ID - read == %X",
3902 pci_config_get32(hdl, 0));
3903
3904 /*
3905 * Restore the config info
3906 */
3907 for (i = 0; i < HERMON_SW_RESET_NUMREGS; i++) {
3908 if (i == 1) continue; /* skip the status/ctrl reg */
3909 if ((i != HERMON_SW_RESET_REG22_RSVD) &&
3910 (i != HERMON_SW_RESET_REG23_RSVD)) {
3911 pci_config_put32(hdl, i << 2, state->hs_cfg_data[i]);
3912 }
3913 }
3914
3915 /*
3916 * PCI Express Capability - we saved during capability list, and
3917 * we'll restore them here.
3918 */
3919 offset = state->hs_pci_cap_offset;
3920 data32 = state->hs_pci_cap_devctl;
3921 pci_config_put32(hdl, offset + HERMON_PCI_CAP_DEV_OFFS, data32);
3922 data32 = state->hs_pci_cap_lnkctl;
3923 pci_config_put32(hdl, offset + HERMON_PCI_CAP_LNK_OFFS, data32);
3924
3925 pci_config_put32(hdl, 0x04, (state->hs_cfg_data[1] | 0x0006));
3926
3927 /* the FMA retry loop ends. */
3928 hermon_pio_end(state, hdl, pio_error2, fm_loop_cnt2, fm_status2,
3929 fm_test2);
3930 hermon_pio_end(state, cmdhdl, pio_error, fm_loop_cnt, fm_status,
3931 fm_test);
3932
3933 return (DDI_SUCCESS);
3934
3935 pio_error2:
3936 /* fall through */
3937 pio_error:
3938 hermon_fm_ereport(state, HCA_SYS_ERR, HCA_ERR_NON_FATAL);
3939 return (DDI_FAILURE);
3940 }
3941
3942
3943 /*
3944 * hermon_mcg_init()
3945 * Context: Only called from attach() path context
3946 */
3947 static int
hermon_mcg_init(hermon_state_t * state)3948 hermon_mcg_init(hermon_state_t *state)
3949 {
3950 uint_t mcg_tmp_sz;
3951
3952
3953 /*
3954 * Allocate space for the MCG temporary copy buffer. This is
3955 * used by the Attach/Detach Multicast Group code
3956 */
3957 mcg_tmp_sz = HERMON_MCGMEM_SZ(state);
3958 state->hs_mcgtmp = kmem_zalloc(mcg_tmp_sz, KM_SLEEP);
3959
3960 /*
3961 * Initialize the multicast group mutex. This ensures atomic
3962 * access to add, modify, and remove entries in the multicast
3963 * group hash lists.
3964 */
3965 mutex_init(&state->hs_mcglock, NULL, MUTEX_DRIVER,
3966 DDI_INTR_PRI(state->hs_intrmsi_pri));
3967
3968 return (DDI_SUCCESS);
3969 }
3970
3971
3972 /*
3973 * hermon_mcg_fini()
3974 * Context: Only called from attach() and/or detach() path contexts
3975 */
3976 static void
hermon_mcg_fini(hermon_state_t * state)3977 hermon_mcg_fini(hermon_state_t *state)
3978 {
3979 uint_t mcg_tmp_sz;
3980
3981
3982 /* Free up the space used for the MCG temporary copy buffer */
3983 mcg_tmp_sz = HERMON_MCGMEM_SZ(state);
3984 kmem_free(state->hs_mcgtmp, mcg_tmp_sz);
3985
3986 /* Destroy the multicast group mutex */
3987 mutex_destroy(&state->hs_mcglock);
3988
3989 }
3990
3991
3992 /*
3993 * hermon_fw_version_check()
3994 * Context: Only called from attach() path context
3995 */
3996 static int
hermon_fw_version_check(hermon_state_t * state)3997 hermon_fw_version_check(hermon_state_t *state)
3998 {
3999
4000 uint_t hermon_fw_ver_major;
4001 uint_t hermon_fw_ver_minor;
4002 uint_t hermon_fw_ver_subminor;
4003
4004 #ifdef FMA_TEST
4005 if (hermon_test_num == -1) {
4006 return (DDI_FAILURE);
4007 }
4008 #endif
4009
4010 /*
4011 * Depending on which version of driver we have attached, and which
4012 * HCA we've attached, the firmware version checks will be different.
4013 * We set up the comparison values for both Arbel and Sinai HCAs.
4014 */
4015 switch (state->hs_operational_mode) {
4016 case HERMON_HCA_MODE:
4017 hermon_fw_ver_major = HERMON_FW_VER_MAJOR;
4018 hermon_fw_ver_minor = HERMON_FW_VER_MINOR;
4019 hermon_fw_ver_subminor = HERMON_FW_VER_SUBMINOR;
4020 break;
4021
4022 default:
4023 return (DDI_FAILURE);
4024 }
4025
4026 /*
4027 * If FW revision major number is less than acceptable,
4028 * return failure, else if greater return success. If
4029 * the major numbers are equal than check the minor number
4030 */
4031 if (state->hs_fw.fw_rev_major < hermon_fw_ver_major) {
4032 return (DDI_FAILURE);
4033 } else if (state->hs_fw.fw_rev_major > hermon_fw_ver_major) {
4034 return (DDI_SUCCESS);
4035 }
4036
4037 /*
4038 * Do the same check as above, except for minor revision numbers
4039 * If the minor numbers are equal than check the subminor number
4040 */
4041 if (state->hs_fw.fw_rev_minor < hermon_fw_ver_minor) {
4042 return (DDI_FAILURE);
4043 } else if (state->hs_fw.fw_rev_minor > hermon_fw_ver_minor) {
4044 return (DDI_SUCCESS);
4045 }
4046
4047 /*
4048 * Once again we do the same check as above, except for the subminor
4049 * revision number. If the subminor numbers are equal here, then
4050 * these are the same firmware version, return success
4051 */
4052 if (state->hs_fw.fw_rev_subminor < hermon_fw_ver_subminor) {
4053 return (DDI_FAILURE);
4054 } else if (state->hs_fw.fw_rev_subminor > hermon_fw_ver_subminor) {
4055 return (DDI_SUCCESS);
4056 }
4057
4058 return (DDI_SUCCESS);
4059 }
4060
4061
4062 /*
4063 * hermon_device_info_report()
4064 * Context: Only called from attach() path context
4065 */
4066 static void
hermon_device_info_report(hermon_state_t * state)4067 hermon_device_info_report(hermon_state_t *state)
4068 {
4069
4070 cmn_err(CE_CONT, "?hermon%d: FW ver: %04d.%04d.%04d, "
4071 "HW rev: %02d\n", state->hs_instance, state->hs_fw.fw_rev_major,
4072 state->hs_fw.fw_rev_minor, state->hs_fw.fw_rev_subminor,
4073 state->hs_revision_id);
4074 cmn_err(CE_CONT, "?hermon%d: %64s (0x%016" PRIx64 ")\n",
4075 state->hs_instance, state->hs_nodedesc, state->hs_nodeguid);
4076
4077 }
4078
4079
4080 /*
4081 * hermon_pci_capability_list()
4082 * Context: Only called from attach() path context
4083 */
4084 static int
hermon_pci_capability_list(hermon_state_t * state,ddi_acc_handle_t hdl)4085 hermon_pci_capability_list(hermon_state_t *state, ddi_acc_handle_t hdl)
4086 {
4087 uint_t offset, data;
4088 uint32_t data32;
4089
4090 state->hs_pci_cap_offset = 0; /* make sure it's cleared */
4091
4092 /*
4093 * Check for the "PCI Capabilities" bit in the "Status Register".
4094 * Bit 4 in this register indicates the presence of a "PCI
4095 * Capabilities" list.
4096 *
4097 * PCI-Express requires this bit to be set to 1.
4098 */
4099 data = pci_config_get16(hdl, 0x06);
4100 if ((data & 0x10) == 0) {
4101 return (DDI_FAILURE);
4102 }
4103
4104 /*
4105 * Starting from offset 0x34 in PCI config space, find the
4106 * head of "PCI capabilities" list, and walk the list. If
4107 * capabilities of a known type are encountered (e.g.
4108 * "PCI-X Capability"), then call the appropriate handler
4109 * function.
4110 */
4111 offset = pci_config_get8(hdl, 0x34);
4112 while (offset != 0x0) {
4113 data = pci_config_get8(hdl, offset);
4114 /*
4115 * Check for known capability types. Hermon has the
4116 * following:
4117 * o Power Mgmt (0x02)
4118 * o VPD Capability (0x03)
4119 * o PCI-E Capability (0x10)
4120 * o MSIX Capability (0x11)
4121 */
4122 switch (data) {
4123 case 0x01:
4124 /* power mgmt handling */
4125 break;
4126 case 0x03:
4127
4128 /*
4129 * Reading the PCIe VPD is inconsistent - that is, sometimes causes
4130 * problems on (mostly) X64, though we've also seen problems w/ Sparc
4131 * and Tavor --- so, for now until it's root caused, don't try and
4132 * read it
4133 */
4134 #ifdef HERMON_VPD_WORKS
4135 hermon_pci_capability_vpd(state, hdl, offset);
4136 #else
4137 delay(100);
4138 hermon_pci_capability_vpd(state, hdl, offset);
4139 #endif
4140 break;
4141 case 0x10:
4142 /*
4143 * PCI Express Capability - save offset & contents
4144 * for later in reset
4145 */
4146 state->hs_pci_cap_offset = offset;
4147 data32 = pci_config_get32(hdl,
4148 offset + HERMON_PCI_CAP_DEV_OFFS);
4149 state->hs_pci_cap_devctl = data32;
4150 data32 = pci_config_get32(hdl,
4151 offset + HERMON_PCI_CAP_LNK_OFFS);
4152 state->hs_pci_cap_lnkctl = data32;
4153 break;
4154 case 0x11:
4155 /*
4156 * MSIX support - nothing to do, taken care of in the
4157 * MSI/MSIX interrupt frameworkd
4158 */
4159 break;
4160 default:
4161 /* just go on to the next */
4162 break;
4163 }
4164
4165 /* Get offset of next entry in list */
4166 offset = pci_config_get8(hdl, offset + 1);
4167 }
4168
4169 return (DDI_SUCCESS);
4170 }
4171
4172 /*
4173 * hermon_pci_read_vpd()
4174 * Context: Only called from attach() path context
4175 * utility routine for hermon_pci_capability_vpd()
4176 */
4177 static int
hermon_pci_read_vpd(ddi_acc_handle_t hdl,uint_t offset,uint32_t addr,uint32_t * data)4178 hermon_pci_read_vpd(ddi_acc_handle_t hdl, uint_t offset, uint32_t addr,
4179 uint32_t *data)
4180 {
4181 int retry = 40; /* retry counter for EEPROM poll */
4182 uint32_t val;
4183 int vpd_addr = offset + 2;
4184 int vpd_data = offset + 4;
4185
4186 /*
4187 * In order to read a 32-bit value from VPD, we are to write down
4188 * the address (offset in the VPD itself) to the address register.
4189 * To signal the read, we also clear bit 31. We then poll on bit 31
4190 * and when it is set, we can then read our 4 bytes from the data
4191 * register.
4192 */
4193 (void) pci_config_put32(hdl, offset, addr << 16);
4194 do {
4195 drv_usecwait(1000);
4196 val = pci_config_get16(hdl, vpd_addr);
4197 if (val & 0x8000) { /* flag bit set */
4198 *data = pci_config_get32(hdl, vpd_data);
4199 return (DDI_SUCCESS);
4200 }
4201 } while (--retry);
4202 /* read of flag failed write one message but count the failures */
4203 if (debug_vpd == 0)
4204 cmn_err(CE_NOTE,
4205 "!Failed to see flag bit after VPD addr write\n");
4206 debug_vpd++;
4207
4208
4209 vpd_read_fail:
4210 return (DDI_FAILURE);
4211 }
4212
4213
4214
4215 /*
4216 * hermon_pci_capability_vpd()
4217 * Context: Only called from attach() path context
4218 */
4219 static void
hermon_pci_capability_vpd(hermon_state_t * state,ddi_acc_handle_t hdl,uint_t offset)4220 hermon_pci_capability_vpd(hermon_state_t *state, ddi_acc_handle_t hdl,
4221 uint_t offset)
4222 {
4223 uint8_t name_length;
4224 uint8_t pn_length;
4225 int i, err = 0;
4226 int vpd_str_id = 0;
4227 int vpd_ro_desc;
4228 int vpd_ro_pn_desc;
4229 #ifdef _BIG_ENDIAN
4230 uint32_t data32;
4231 #endif /* _BIG_ENDIAN */
4232 union {
4233 uint32_t vpd_int[HERMON_VPD_HDR_DWSIZE];
4234 uchar_t vpd_char[HERMON_VPD_HDR_BSIZE];
4235 } vpd;
4236
4237
4238 /*
4239 * Read in the Vital Product Data (VPD) to the extend needed
4240 * by the fwflash utility
4241 */
4242 for (i = 0; i < HERMON_VPD_HDR_DWSIZE; i++) {
4243 err = hermon_pci_read_vpd(hdl, offset, i << 2, &vpd.vpd_int[i]);
4244 if (err != DDI_SUCCESS) {
4245 cmn_err(CE_NOTE, "!VPD read failed\n");
4246 goto out;
4247 }
4248 }
4249
4250 #ifdef _BIG_ENDIAN
4251 /* Need to swap bytes for big endian. */
4252 for (i = 0; i < HERMON_VPD_HDR_DWSIZE; i++) {
4253 data32 = vpd.vpd_int[i];
4254 vpd.vpd_char[(i << 2) + 3] =
4255 (uchar_t)((data32 & 0xFF000000) >> 24);
4256 vpd.vpd_char[(i << 2) + 2] =
4257 (uchar_t)((data32 & 0x00FF0000) >> 16);
4258 vpd.vpd_char[(i << 2) + 1] =
4259 (uchar_t)((data32 & 0x0000FF00) >> 8);
4260 vpd.vpd_char[i << 2] = (uchar_t)(data32 & 0x000000FF);
4261 }
4262 #endif /* _BIG_ENDIAN */
4263
4264 /* Check for VPD String ID Tag */
4265 if (vpd.vpd_char[vpd_str_id] == 0x82) {
4266 /* get the product name */
4267 name_length = (uint8_t)vpd.vpd_char[vpd_str_id + 1];
4268 if (name_length > sizeof (state->hs_hca_name)) {
4269 cmn_err(CE_NOTE, "!VPD name too large (0x%x)\n",
4270 name_length);
4271 goto out;
4272 }
4273 (void) memcpy(state->hs_hca_name, &vpd.vpd_char[vpd_str_id + 3],
4274 name_length);
4275 state->hs_hca_name[name_length] = 0;
4276
4277 /* get the part number */
4278 vpd_ro_desc = name_length + 3; /* read-only tag location */
4279 vpd_ro_pn_desc = vpd_ro_desc + 3; /* P/N keyword location */
4280
4281 /* Verify read-only tag and Part Number keyword. */
4282 if (vpd.vpd_char[vpd_ro_desc] != 0x90 ||
4283 (vpd.vpd_char[vpd_ro_pn_desc] != 'P' &&
4284 vpd.vpd_char[vpd_ro_pn_desc + 1] != 'N')) {
4285 cmn_err(CE_NOTE, "!VPD Part Number not found\n");
4286 goto out;
4287 }
4288
4289 pn_length = (uint8_t)vpd.vpd_char[vpd_ro_pn_desc + 2];
4290 if (pn_length > sizeof (state->hs_hca_pn)) {
4291 cmn_err(CE_NOTE, "!VPD part number too large (0x%x)\n",
4292 name_length);
4293 goto out;
4294 }
4295 (void) memcpy(state->hs_hca_pn,
4296 &vpd.vpd_char[vpd_ro_pn_desc + 3],
4297 pn_length);
4298 state->hs_hca_pn[pn_length] = 0;
4299 state->hs_hca_pn_len = pn_length;
4300 cmn_err(CE_CONT, "!vpd %s\n", state->hs_hca_pn);
4301 } else {
4302 /* Wrong VPD String ID Tag */
4303 cmn_err(CE_NOTE, "!VPD String ID Tag not found, tag: %02x\n",
4304 vpd.vpd_char[0]);
4305 goto out;
4306 }
4307 return;
4308 out:
4309 state->hs_hca_pn_len = 0;
4310 }
4311
4312
4313
4314 /*
4315 * hermon_intr_or_msi_init()
4316 * Context: Only called from attach() path context
4317 */
4318 static int
hermon_intr_or_msi_init(hermon_state_t * state)4319 hermon_intr_or_msi_init(hermon_state_t *state)
4320 {
4321 int status;
4322
4323 /* Query for the list of supported interrupt event types */
4324 status = ddi_intr_get_supported_types(state->hs_dip,
4325 &state->hs_intr_types_avail);
4326 if (status != DDI_SUCCESS) {
4327 return (DDI_FAILURE);
4328 }
4329
4330 /*
4331 * If Hermon supports MSI-X in this system (and, if it
4332 * hasn't been overridden by a configuration variable), then
4333 * the default behavior is to use a single MSI-X. Otherwise,
4334 * fallback to using legacy interrupts. Also, if MSI-X is chosen,
4335 * but fails for whatever reasons, then next try MSI
4336 */
4337 if ((state->hs_cfg_profile->cp_use_msi_if_avail != 0) &&
4338 (state->hs_intr_types_avail & DDI_INTR_TYPE_MSIX)) {
4339 status = hermon_add_intrs(state, DDI_INTR_TYPE_MSIX);
4340 if (status == DDI_SUCCESS) {
4341 state->hs_intr_type_chosen = DDI_INTR_TYPE_MSIX;
4342 return (DDI_SUCCESS);
4343 }
4344 }
4345
4346 /*
4347 * If Hermon supports MSI in this system (and, if it
4348 * hasn't been overridden by a configuration variable), then
4349 * the default behavior is to use a single MSIX. Otherwise,
4350 * fallback to using legacy interrupts. Also, if MSI is chosen,
4351 * but fails for whatever reasons, then fallback to using legacy
4352 * interrupts.
4353 */
4354 if ((state->hs_cfg_profile->cp_use_msi_if_avail != 0) &&
4355 (state->hs_intr_types_avail & DDI_INTR_TYPE_MSI)) {
4356 status = hermon_add_intrs(state, DDI_INTR_TYPE_MSI);
4357 if (status == DDI_SUCCESS) {
4358 state->hs_intr_type_chosen = DDI_INTR_TYPE_MSI;
4359 return (DDI_SUCCESS);
4360 }
4361 }
4362
4363 /*
4364 * MSI interrupt allocation failed, or was not available. Fallback to
4365 * legacy interrupt support.
4366 */
4367 if (state->hs_intr_types_avail & DDI_INTR_TYPE_FIXED) {
4368 status = hermon_add_intrs(state, DDI_INTR_TYPE_FIXED);
4369 if (status == DDI_SUCCESS) {
4370 state->hs_intr_type_chosen = DDI_INTR_TYPE_FIXED;
4371 return (DDI_SUCCESS);
4372 }
4373 }
4374
4375 /*
4376 * None of MSI, MSI-X, nor legacy interrupts were successful.
4377 * Return failure.
4378 */
4379 return (DDI_FAILURE);
4380 }
4381
4382 /* ARGSUSED */
4383 static int
hermon_intr_cb_handler(dev_info_t * dip,ddi_cb_action_t action,void * cbarg,void * arg1,void * arg2)4384 hermon_intr_cb_handler(dev_info_t *dip, ddi_cb_action_t action, void *cbarg,
4385 void *arg1, void *arg2)
4386 {
4387 hermon_state_t *state = (hermon_state_t *)arg1;
4388
4389 IBTF_DPRINTF_L2("hermon", "interrupt callback: instance %d, "
4390 "action %d, cbarg %d\n", state->hs_instance, action,
4391 (uint32_t)(uintptr_t)cbarg);
4392 return (DDI_SUCCESS);
4393 }
4394
4395 /*
4396 * hermon_add_intrs()
4397 * Context: Only called from attach() patch context
4398 */
4399 static int
hermon_add_intrs(hermon_state_t * state,int intr_type)4400 hermon_add_intrs(hermon_state_t *state, int intr_type)
4401 {
4402 int status;
4403
4404 if (state->hs_intr_cb_hdl == NULL) {
4405 status = ddi_cb_register(state->hs_dip, DDI_CB_FLAG_INTR,
4406 hermon_intr_cb_handler, state, NULL,
4407 &state->hs_intr_cb_hdl);
4408 if (status != DDI_SUCCESS) {
4409 cmn_err(CE_CONT, "ddi_cb_register failed: 0x%x\n",
4410 status);
4411 state->hs_intr_cb_hdl = NULL;
4412 return (DDI_FAILURE);
4413 }
4414 }
4415
4416 /* Get number of interrupts/MSI supported */
4417 status = ddi_intr_get_nintrs(state->hs_dip, intr_type,
4418 &state->hs_intrmsi_count);
4419 if (status != DDI_SUCCESS) {
4420 (void) ddi_cb_unregister(state->hs_intr_cb_hdl);
4421 state->hs_intr_cb_hdl = NULL;
4422 return (DDI_FAILURE);
4423 }
4424
4425 /* Get number of available interrupts/MSI */
4426 status = ddi_intr_get_navail(state->hs_dip, intr_type,
4427 &state->hs_intrmsi_avail);
4428 if (status != DDI_SUCCESS) {
4429 (void) ddi_cb_unregister(state->hs_intr_cb_hdl);
4430 state->hs_intr_cb_hdl = NULL;
4431 return (DDI_FAILURE);
4432 }
4433
4434 /* Ensure that we have at least one (1) usable MSI or interrupt */
4435 if ((state->hs_intrmsi_avail < 1) || (state->hs_intrmsi_count < 1)) {
4436 (void) ddi_cb_unregister(state->hs_intr_cb_hdl);
4437 state->hs_intr_cb_hdl = NULL;
4438 return (DDI_FAILURE);
4439 }
4440
4441 /*
4442 * Allocate the #interrupt/MSI handles.
4443 * The number we request is the minimum of these three values:
4444 * HERMON_MSIX_MAX driver maximum (array size)
4445 * hermon_msix_max /etc/system override to...
4446 * HERMON_MSIX_MAX
4447 * state->hs_intrmsi_avail Maximum the ddi provides.
4448 */
4449 status = ddi_intr_alloc(state->hs_dip, &state->hs_intrmsi_hdl[0],
4450 intr_type, 0, min(min(HERMON_MSIX_MAX, state->hs_intrmsi_avail),
4451 hermon_msix_max), &state->hs_intrmsi_allocd, DDI_INTR_ALLOC_NORMAL);
4452 if (status != DDI_SUCCESS) {
4453 (void) ddi_cb_unregister(state->hs_intr_cb_hdl);
4454 state->hs_intr_cb_hdl = NULL;
4455 return (DDI_FAILURE);
4456 }
4457
4458 /* Ensure that we have allocated at least one (1) MSI or interrupt */
4459 if (state->hs_intrmsi_allocd < 1) {
4460 (void) ddi_cb_unregister(state->hs_intr_cb_hdl);
4461 state->hs_intr_cb_hdl = NULL;
4462 return (DDI_FAILURE);
4463 }
4464
4465 /*
4466 * Extract the priority for the allocated interrupt/MSI. This
4467 * will be used later when initializing certain mutexes.
4468 */
4469 status = ddi_intr_get_pri(state->hs_intrmsi_hdl[0],
4470 &state->hs_intrmsi_pri);
4471 if (status != DDI_SUCCESS) {
4472 /* Free the allocated interrupt/MSI handle */
4473 (void) ddi_intr_free(state->hs_intrmsi_hdl[0]);
4474
4475 (void) ddi_cb_unregister(state->hs_intr_cb_hdl);
4476 state->hs_intr_cb_hdl = NULL;
4477 return (DDI_FAILURE);
4478 }
4479
4480 /* Make sure the interrupt/MSI priority is below 'high level' */
4481 if (state->hs_intrmsi_pri >= ddi_intr_get_hilevel_pri()) {
4482 /* Free the allocated interrupt/MSI handle */
4483 (void) ddi_intr_free(state->hs_intrmsi_hdl[0]);
4484
4485 return (DDI_FAILURE);
4486 }
4487
4488 /* Get add'l capability information regarding interrupt/MSI */
4489 status = ddi_intr_get_cap(state->hs_intrmsi_hdl[0],
4490 &state->hs_intrmsi_cap);
4491 if (status != DDI_SUCCESS) {
4492 /* Free the allocated interrupt/MSI handle */
4493 (void) ddi_intr_free(state->hs_intrmsi_hdl[0]);
4494
4495 return (DDI_FAILURE);
4496 }
4497
4498 return (DDI_SUCCESS);
4499 }
4500
4501
4502 /*
4503 * hermon_intr_or_msi_fini()
4504 * Context: Only called from attach() and/or detach() path contexts
4505 */
4506 static int
hermon_intr_or_msi_fini(hermon_state_t * state)4507 hermon_intr_or_msi_fini(hermon_state_t *state)
4508 {
4509 int status;
4510 int intr;
4511
4512 for (intr = 0; intr < state->hs_intrmsi_allocd; intr++) {
4513
4514 /* Free the allocated interrupt/MSI handle */
4515 status = ddi_intr_free(state->hs_intrmsi_hdl[intr]);
4516 if (status != DDI_SUCCESS) {
4517 return (DDI_FAILURE);
4518 }
4519 }
4520 if (state->hs_intr_cb_hdl) {
4521 (void) ddi_cb_unregister(state->hs_intr_cb_hdl);
4522 state->hs_intr_cb_hdl = NULL;
4523 }
4524 return (DDI_SUCCESS);
4525 }
4526
4527
4528 /*ARGSUSED*/
4529 void
hermon_pci_capability_msix(hermon_state_t * state,ddi_acc_handle_t hdl,uint_t offset)4530 hermon_pci_capability_msix(hermon_state_t *state, ddi_acc_handle_t hdl,
4531 uint_t offset)
4532 {
4533 uint32_t msix_data;
4534 uint16_t msg_cntr;
4535 uint32_t t_offset; /* table offset */
4536 uint32_t t_bir;
4537 uint32_t p_offset; /* pba */
4538 uint32_t p_bir;
4539 int t_size; /* size in entries - each is 4 dwords */
4540
4541 /* come in with offset pointing at the capability structure */
4542
4543 msix_data = pci_config_get32(hdl, offset);
4544 cmn_err(CE_CONT, "Full cap structure dword = %X\n", msix_data);
4545 msg_cntr = pci_config_get16(hdl, offset+2);
4546 cmn_err(CE_CONT, "MSIX msg_control = %X\n", msg_cntr);
4547 offset += 4;
4548 msix_data = pci_config_get32(hdl, offset); /* table info */
4549 t_offset = (msix_data & 0xFFF8) >> 3;
4550 t_bir = msix_data & 0x07;
4551 offset += 4;
4552 cmn_err(CE_CONT, " table %X --offset = %X, bir(bar) = %X\n",
4553 msix_data, t_offset, t_bir);
4554 msix_data = pci_config_get32(hdl, offset); /* PBA info */
4555 p_offset = (msix_data & 0xFFF8) >> 3;
4556 p_bir = msix_data & 0x07;
4557
4558 cmn_err(CE_CONT, " PBA %X --offset = %X, bir(bar) = %X\n",
4559 msix_data, p_offset, p_bir);
4560 t_size = msg_cntr & 0x7FF; /* low eleven bits */
4561 cmn_err(CE_CONT, " table size = %X entries\n", t_size);
4562
4563 offset = t_offset; /* reuse this for offset from BAR */
4564 #ifdef HERMON_SUPPORTS_MSIX_BAR
4565 cmn_err(CE_CONT, "First 2 table entries behind BAR2 \n");
4566 for (i = 0; i < 2; i++) {
4567 for (j = 0; j < 4; j++, offset += 4) {
4568 msix_data = ddi_get32(state->hs_reg_msihdl,
4569 (uint32_t *)((uintptr_t)state->hs_reg_msi_baseaddr
4570 + offset));
4571 cmn_err(CE_CONT, "MSI table entry %d, dword %d == %X\n",
4572 i, j, msix_data);
4573 }
4574 }
4575 #endif
4576
4577 }
4578
4579 /*
4580 * X86 fastreboot support functions.
4581 * These functions are used to save/restore MSI-X table/PBA and also
4582 * to disable MSI-X interrupts in hermon_quiesce().
4583 */
4584
4585 /* Return the message control for MSI-X */
4586 static ushort_t
get_msix_ctrl(dev_info_t * dip)4587 get_msix_ctrl(dev_info_t *dip)
4588 {
4589 ushort_t msix_ctrl = 0, caps_ctrl = 0;
4590 hermon_state_t *state = ddi_get_soft_state(hermon_statep,
4591 DEVI(dip)->devi_instance);
4592 ddi_acc_handle_t pci_cfg_hdl = hermon_get_pcihdl(state);
4593 ASSERT(pci_cfg_hdl != NULL);
4594
4595 if ((PCI_CAP_LOCATE(pci_cfg_hdl,
4596 PCI_CAP_ID_MSI_X, &caps_ctrl) == DDI_SUCCESS)) {
4597 if ((msix_ctrl = PCI_CAP_GET16(pci_cfg_hdl, 0, caps_ctrl,
4598 PCI_MSIX_CTRL)) == PCI_CAP_EINVAL16)
4599 return (0);
4600 }
4601 ASSERT(msix_ctrl != 0);
4602
4603 return (msix_ctrl);
4604 }
4605
4606 /* Return the MSI-X table size */
4607 static size_t
get_msix_tbl_size(dev_info_t * dip)4608 get_msix_tbl_size(dev_info_t *dip)
4609 {
4610 ushort_t msix_ctrl = get_msix_ctrl(dip);
4611 ASSERT(msix_ctrl != 0);
4612
4613 return (((msix_ctrl & PCI_MSIX_TBL_SIZE_MASK) + 1) *
4614 PCI_MSIX_VECTOR_SIZE);
4615 }
4616
4617 /* Return the MSI-X PBA size */
4618 static size_t
get_msix_pba_size(dev_info_t * dip)4619 get_msix_pba_size(dev_info_t *dip)
4620 {
4621 ushort_t msix_ctrl = get_msix_ctrl(dip);
4622 ASSERT(msix_ctrl != 0);
4623
4624 return (((msix_ctrl & PCI_MSIX_TBL_SIZE_MASK) + 64) / 64 * 8);
4625 }
4626
4627 /* Set up the MSI-X table/PBA save area */
4628 static void
hermon_set_msix_info(hermon_state_t * state)4629 hermon_set_msix_info(hermon_state_t *state)
4630 {
4631 uint_t rnumber, breg, nregs;
4632 ushort_t caps_ctrl, msix_ctrl;
4633 pci_regspec_t *rp;
4634 int reg_size, addr_space, offset, *regs_list, i;
4635
4636 /*
4637 * MSI-X BIR Index Table:
4638 * BAR indicator register (BIR) to Base Address register.
4639 */
4640 uchar_t pci_msix_bir_index[8] = {0x10, 0x14, 0x18, 0x1c,
4641 0x20, 0x24, 0xff, 0xff};
4642
4643 /* Fastreboot data access attribute */
4644 ddi_device_acc_attr_t dev_attr = {
4645 0, /* version */
4646 DDI_STRUCTURE_LE_ACC,
4647 DDI_STRICTORDER_ACC, /* attr access */
4648 0
4649 };
4650
4651 ddi_acc_handle_t pci_cfg_hdl = hermon_get_pcihdl(state);
4652 ASSERT(pci_cfg_hdl != NULL);
4653
4654 if ((PCI_CAP_LOCATE(pci_cfg_hdl,
4655 PCI_CAP_ID_MSI_X, &caps_ctrl) == DDI_SUCCESS)) {
4656 if ((msix_ctrl = PCI_CAP_GET16(pci_cfg_hdl, 0, caps_ctrl,
4657 PCI_MSIX_CTRL)) == PCI_CAP_EINVAL16)
4658 return;
4659 }
4660 ASSERT(msix_ctrl != 0);
4661
4662 state->hs_msix_tbl_offset = PCI_CAP_GET32(pci_cfg_hdl, 0, caps_ctrl,
4663 PCI_MSIX_TBL_OFFSET);
4664
4665 /* Get the BIR for MSI-X table */
4666 breg = pci_msix_bir_index[state->hs_msix_tbl_offset &
4667 PCI_MSIX_TBL_BIR_MASK];
4668 ASSERT(breg != 0xFF);
4669
4670 /* Set the MSI-X table offset */
4671 state->hs_msix_tbl_offset = state->hs_msix_tbl_offset &
4672 ~PCI_MSIX_TBL_BIR_MASK;
4673
4674 /* Set the MSI-X table size */
4675 state->hs_msix_tbl_size = ((msix_ctrl & PCI_MSIX_TBL_SIZE_MASK) + 1) *
4676 PCI_MSIX_VECTOR_SIZE;
4677
4678 if (ddi_prop_lookup_int_array(DDI_DEV_T_ANY, state->hs_dip,
4679 DDI_PROP_DONTPASS, "reg", (int **)®s_list, &nregs) !=
4680 DDI_PROP_SUCCESS) {
4681 return;
4682 }
4683 reg_size = sizeof (pci_regspec_t) / sizeof (int);
4684
4685 /* Check the register number for MSI-X table */
4686 for (i = 1, rnumber = 0; i < nregs/reg_size; i++) {
4687 rp = (pci_regspec_t *)®s_list[i * reg_size];
4688 addr_space = rp->pci_phys_hi & PCI_ADDR_MASK;
4689 offset = PCI_REG_REG_G(rp->pci_phys_hi);
4690
4691 if ((offset == breg) && ((addr_space == PCI_ADDR_MEM32) ||
4692 (addr_space == PCI_ADDR_MEM64))) {
4693 rnumber = i;
4694 break;
4695 }
4696 }
4697 ASSERT(rnumber != 0);
4698 state->hs_msix_tbl_rnumber = rnumber;
4699
4700 /* Set device attribute version and access according to Hermon FM */
4701 dev_attr.devacc_attr_version = hermon_devacc_attr_version(state);
4702 dev_attr.devacc_attr_access = hermon_devacc_attr_access(state);
4703
4704 /* Map the entire MSI-X vector table */
4705 if (hermon_regs_map_setup(state, state->hs_msix_tbl_rnumber,
4706 (caddr_t *)&state->hs_msix_tbl_addr, state->hs_msix_tbl_offset,
4707 state->hs_msix_tbl_size, &dev_attr,
4708 &state->hs_fm_msix_tblhdl) != DDI_SUCCESS) {
4709 return;
4710 }
4711
4712 state->hs_msix_pba_offset = PCI_CAP_GET32(pci_cfg_hdl, 0, caps_ctrl,
4713 PCI_MSIX_PBA_OFFSET);
4714
4715 /* Get the BIR for MSI-X PBA */
4716 breg = pci_msix_bir_index[state->hs_msix_pba_offset &
4717 PCI_MSIX_PBA_BIR_MASK];
4718 ASSERT(breg != 0xFF);
4719
4720 /* Set the MSI-X PBA offset */
4721 state->hs_msix_pba_offset = state->hs_msix_pba_offset &
4722 ~PCI_MSIX_PBA_BIR_MASK;
4723
4724 /* Set the MSI-X PBA size */
4725 state->hs_msix_pba_size =
4726 ((msix_ctrl & PCI_MSIX_TBL_SIZE_MASK) + 64) / 64 * 8;
4727
4728 /* Check the register number for MSI-X PBA */
4729 for (i = 1, rnumber = 0; i < nregs/reg_size; i++) {
4730 rp = (pci_regspec_t *)®s_list[i * reg_size];
4731 addr_space = rp->pci_phys_hi & PCI_ADDR_MASK;
4732 offset = PCI_REG_REG_G(rp->pci_phys_hi);
4733
4734 if ((offset == breg) && ((addr_space == PCI_ADDR_MEM32) ||
4735 (addr_space == PCI_ADDR_MEM64))) {
4736 rnumber = i;
4737 break;
4738 }
4739 }
4740 ASSERT(rnumber != 0);
4741 state->hs_msix_pba_rnumber = rnumber;
4742 ddi_prop_free(regs_list);
4743
4744 /* Map in the MSI-X Pending Bit Array */
4745 if (hermon_regs_map_setup(state, state->hs_msix_pba_rnumber,
4746 (caddr_t *)&state->hs_msix_pba_addr, state->hs_msix_pba_offset,
4747 state->hs_msix_pba_size, &dev_attr,
4748 &state->hs_fm_msix_pbahdl) != DDI_SUCCESS) {
4749 hermon_regs_map_free(state, &state->hs_fm_msix_tblhdl);
4750 state->hs_fm_msix_tblhdl = NULL;
4751 return;
4752 }
4753
4754 /* Set the MSI-X table save area */
4755 state->hs_msix_tbl_entries = kmem_alloc(state->hs_msix_tbl_size,
4756 KM_SLEEP);
4757
4758 /* Set the MSI-X PBA save area */
4759 state->hs_msix_pba_entries = kmem_alloc(state->hs_msix_pba_size,
4760 KM_SLEEP);
4761 }
4762
4763 /* Disable Hermon interrupts */
4764 static int
hermon_intr_disable(hermon_state_t * state)4765 hermon_intr_disable(hermon_state_t *state)
4766 {
4767 ushort_t msix_ctrl = 0, caps_ctrl = 0;
4768 ddi_acc_handle_t pci_cfg_hdl = hermon_get_pcihdl(state);
4769 ddi_acc_handle_t msix_tblhdl = hermon_get_msix_tblhdl(state);
4770 int i, j;
4771 ASSERT(pci_cfg_hdl != NULL && msix_tblhdl != NULL);
4772 ASSERT(state->hs_intr_types_avail &
4773 (DDI_INTR_TYPE_FIXED | DDI_INTR_TYPE_MSI | DDI_INTR_TYPE_MSIX));
4774
4775 /*
4776 * Check if MSI-X interrupts are used. If so, disable MSI-X interupts.
4777 * If not, since Hermon doesn't support MSI interrupts, assuming the
4778 * legacy interrupt is used instead, disable the legacy interrupt.
4779 */
4780 if ((state->hs_cfg_profile->cp_use_msi_if_avail != 0) &&
4781 (state->hs_intr_types_avail & DDI_INTR_TYPE_MSIX)) {
4782
4783 if ((PCI_CAP_LOCATE(pci_cfg_hdl,
4784 PCI_CAP_ID_MSI_X, &caps_ctrl) == DDI_SUCCESS)) {
4785 if ((msix_ctrl = PCI_CAP_GET16(pci_cfg_hdl, 0,
4786 caps_ctrl, PCI_MSIX_CTRL)) == PCI_CAP_EINVAL16)
4787 return (DDI_FAILURE);
4788 }
4789 ASSERT(msix_ctrl != 0);
4790
4791 if (!(msix_ctrl & PCI_MSIX_ENABLE_BIT))
4792 return (DDI_SUCCESS);
4793
4794 /* Clear all inums in MSI-X table */
4795 for (i = 0; i < get_msix_tbl_size(state->hs_dip);
4796 i += PCI_MSIX_VECTOR_SIZE) {
4797 for (j = 0; j < PCI_MSIX_VECTOR_SIZE; j += 4) {
4798 char *addr = state->hs_msix_tbl_addr + i + j;
4799 ddi_put32(msix_tblhdl,
4800 (uint32_t *)(uintptr_t)addr, 0x0);
4801 }
4802 }
4803
4804 /* Disable MSI-X interrupts */
4805 msix_ctrl &= ~PCI_MSIX_ENABLE_BIT;
4806 PCI_CAP_PUT16(pci_cfg_hdl, 0, caps_ctrl, PCI_MSIX_CTRL,
4807 msix_ctrl);
4808
4809 } else {
4810 uint16_t cmdreg = pci_config_get16(pci_cfg_hdl, PCI_CONF_COMM);
4811 ASSERT(state->hs_intr_types_avail & DDI_INTR_TYPE_FIXED);
4812
4813 /* Disable the legacy interrupts */
4814 cmdreg |= PCI_COMM_INTX_DISABLE;
4815 pci_config_put16(pci_cfg_hdl, PCI_CONF_COMM, cmdreg);
4816 }
4817
4818 return (DDI_SUCCESS);
4819 }
4820
4821 /* Hermon quiesce(9F) entry */
4822 static int
hermon_quiesce(dev_info_t * dip)4823 hermon_quiesce(dev_info_t *dip)
4824 {
4825 hermon_state_t *state = ddi_get_soft_state(hermon_statep,
4826 DEVI(dip)->devi_instance);
4827 ddi_acc_handle_t pcihdl = hermon_get_pcihdl(state);
4828 ddi_acc_handle_t cmdhdl = hermon_get_cmdhdl(state);
4829 ddi_acc_handle_t msix_tbl_hdl = hermon_get_msix_tblhdl(state);
4830 ddi_acc_handle_t msix_pba_hdl = hermon_get_msix_pbahdl(state);
4831 uint32_t sem, reset_delay = state->hs_cfg_profile->cp_sw_reset_delay;
4832 uint64_t data64;
4833 uint32_t data32;
4834 int status, i, j, loopcnt;
4835 uint_t offset;
4836
4837 ASSERT(state != NULL);
4838
4839 /* start fastreboot */
4840 state->hs_quiescing = B_TRUE;
4841
4842 /* If it's in maintenance mode, do nothing but return with SUCCESS */
4843 if (!HERMON_IS_OPERATIONAL(state->hs_operational_mode)) {
4844 return (DDI_SUCCESS);
4845 }
4846
4847 /* suppress Hermon FM ereports */
4848 if (hermon_get_state(state) & HCA_EREPORT_FM) {
4849 hermon_clr_state_nolock(state, HCA_EREPORT_FM);
4850 }
4851
4852 /* Shutdown HCA ports */
4853 if (hermon_hca_ports_shutdown(state,
4854 state->hs_cfg_profile->cp_num_ports) != HERMON_CMD_SUCCESS) {
4855 state->hs_quiescing = B_FALSE;
4856 return (DDI_FAILURE);
4857 }
4858
4859 /* Close HCA */
4860 if (hermon_close_hca_cmd_post(state, HERMON_CMD_NOSLEEP_SPIN) !=
4861 HERMON_CMD_SUCCESS) {
4862 state->hs_quiescing = B_FALSE;
4863 return (DDI_FAILURE);
4864 }
4865
4866 /* Disable interrupts */
4867 if (hermon_intr_disable(state) != DDI_SUCCESS) {
4868 state->hs_quiescing = B_FALSE;
4869 return (DDI_FAILURE);
4870 }
4871
4872 /*
4873 * Query the PCI capabilities of the HCA device, but don't process
4874 * the VPD until after reset.
4875 */
4876 if (hermon_pci_capability_list(state, pcihdl) != DDI_SUCCESS) {
4877 state->hs_quiescing = B_FALSE;
4878 return (DDI_FAILURE);
4879 }
4880
4881 /*
4882 * Read all PCI config info (reg0...reg63). Note: According to the
4883 * Hermon software reset application note, we should not read or
4884 * restore the values in reg22 and reg23.
4885 * NOTE: For Hermon (and Arbel too) it says to restore the command
4886 * register LAST, and technically, you need to restore the
4887 * PCIE Capability "device control" and "link control" (word-sized,
4888 * at offsets 0x08 and 0x10 from the capbility ID respectively).
4889 * We hold off restoring the command register - offset 0x4 - till last
4890 */
4891
4892 /* 1st, wait for the semaphore assure accessibility - per PRM */
4893 status = -1;
4894 for (i = 0; i < NANOSEC/MICROSEC /* 1sec timeout */; i++) {
4895 sem = ddi_get32(cmdhdl, state->hs_cmd_regs.sw_semaphore);
4896 if (sem == 0) {
4897 status = 0;
4898 break;
4899 }
4900 drv_usecwait(1);
4901 }
4902
4903 /* Check if timeout happens */
4904 if (status == -1) {
4905 state->hs_quiescing = B_FALSE;
4906 return (DDI_FAILURE);
4907 }
4908
4909 /* MSI-X interrupts are used, save the MSI-X table */
4910 if (msix_tbl_hdl && msix_pba_hdl) {
4911 /* save MSI-X table */
4912 for (i = 0; i < get_msix_tbl_size(state->hs_dip);
4913 i += PCI_MSIX_VECTOR_SIZE) {
4914 for (j = 0; j < PCI_MSIX_VECTOR_SIZE; j += 4) {
4915 char *addr = state->hs_msix_tbl_addr + i + j;
4916 data32 = ddi_get32(msix_tbl_hdl,
4917 (uint32_t *)(uintptr_t)addr);
4918 *(uint32_t *)(uintptr_t)(state->
4919 hs_msix_tbl_entries + i + j) = data32;
4920 }
4921 }
4922 /* save MSI-X PBA */
4923 for (i = 0; i < get_msix_pba_size(state->hs_dip); i += 8) {
4924 char *addr = state->hs_msix_pba_addr + i;
4925 data64 = ddi_get64(msix_pba_hdl,
4926 (uint64_t *)(uintptr_t)addr);
4927 *(uint64_t *)(uintptr_t)(state->
4928 hs_msix_pba_entries + i) = data64;
4929 }
4930 }
4931
4932 /* save PCI config space */
4933 for (i = 0; i < HERMON_SW_RESET_NUMREGS; i++) {
4934 if ((i != HERMON_SW_RESET_REG22_RSVD) &&
4935 (i != HERMON_SW_RESET_REG23_RSVD)) {
4936 state->hs_cfg_data[i] =
4937 pci_config_get32(pcihdl, i << 2);
4938 }
4939 }
4940
4941 /* SW-reset HCA */
4942 ddi_put32(cmdhdl, state->hs_cmd_regs.sw_reset, HERMON_SW_RESET_START);
4943
4944 /*
4945 * This delay is required so as not to cause a panic here. If the
4946 * device is accessed too soon after reset it will not respond to
4947 * config cycles, causing a Master Abort and panic.
4948 */
4949 drv_usecwait(reset_delay);
4950
4951 /* Poll waiting for the device to finish resetting */
4952 loopcnt = 100; /* 100 times @ 100 usec - total delay 10 msec */
4953 while ((pci_config_get32(pcihdl, 0) & 0x0000FFFF) != PCI_VENID_MLX) {
4954 drv_usecwait(HERMON_SW_RESET_POLL_DELAY);
4955 if (--loopcnt == 0)
4956 break; /* just in case, break and go on */
4957 }
4958 if (loopcnt == 0) {
4959 state->hs_quiescing = B_FALSE;
4960 return (DDI_FAILURE);
4961 }
4962
4963 /* Restore the config info */
4964 for (i = 0; i < HERMON_SW_RESET_NUMREGS; i++) {
4965 if (i == 1) continue; /* skip the status/ctrl reg */
4966 if ((i != HERMON_SW_RESET_REG22_RSVD) &&
4967 (i != HERMON_SW_RESET_REG23_RSVD)) {
4968 pci_config_put32(pcihdl, i << 2, state->hs_cfg_data[i]);
4969 }
4970 }
4971
4972 /* If MSI-X interrupts are used, restore the MSI-X table */
4973 if (msix_tbl_hdl && msix_pba_hdl) {
4974 /* restore MSI-X PBA */
4975 for (i = 0; i < get_msix_pba_size(state->hs_dip); i += 8) {
4976 char *addr = state->hs_msix_pba_addr + i;
4977 data64 = *(uint64_t *)(uintptr_t)
4978 (state->hs_msix_pba_entries + i);
4979 ddi_put64(msix_pba_hdl,
4980 (uint64_t *)(uintptr_t)addr, data64);
4981 }
4982 /* restore MSI-X table */
4983 for (i = 0; i < get_msix_tbl_size(state->hs_dip);
4984 i += PCI_MSIX_VECTOR_SIZE) {
4985 for (j = 0; j < PCI_MSIX_VECTOR_SIZE; j += 4) {
4986 char *addr = state->hs_msix_tbl_addr + i + j;
4987 data32 = *(uint32_t *)(uintptr_t)
4988 (state->hs_msix_tbl_entries + i + j);
4989 ddi_put32(msix_tbl_hdl,
4990 (uint32_t *)(uintptr_t)addr, data32);
4991 }
4992 }
4993 }
4994
4995 /*
4996 * PCI Express Capability - we saved during capability list, and
4997 * we'll restore them here.
4998 */
4999 offset = state->hs_pci_cap_offset;
5000 data32 = state->hs_pci_cap_devctl;
5001 pci_config_put32(pcihdl, offset + HERMON_PCI_CAP_DEV_OFFS, data32);
5002 data32 = state->hs_pci_cap_lnkctl;
5003 pci_config_put32(pcihdl, offset + HERMON_PCI_CAP_LNK_OFFS, data32);
5004
5005 /* restore the command register */
5006 pci_config_put32(pcihdl, 0x04, (state->hs_cfg_data[1] | 0x0006));
5007
5008 return (DDI_SUCCESS);
5009 }
5010