1 /*
2 * CDDL HEADER START
3 *
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
7 *
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
12 *
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 *
19 * CDDL HEADER END
20 */
21
22 /*
23 * Copyright (c) 2010, Oracle and/or its affiliates. All rights reserved.
24 */
25
26 #include <sys/types.h>
27 #include <sys/kmem.h>
28 #include <sys/conf.h>
29 #include <sys/ddi.h>
30 #include <sys/sunddi.h>
31 #include <sys/ksynch.h>
32 #include <sys/dlpi.h> /* HCKSUM_INET_FULL_V4 */
33 #include <sys/pattr.h> /* HCK_FULLCKSUM */
34 #include <sys/ib/mgt/sm_attr.h> /* SM_INIT_TYPE_REPLY_... */
35
36 #include <sys/ib/clients/eoib/eib_impl.h>
37
38 /*
39 * Declarations private to this file
40 */
41 static void eib_ibt_reset_partitions(eib_t *);
42 static void eib_ibt_wakeup_sqd_waiters(eib_t *, ibt_channel_hdl_t);
43 static int eib_ibt_chan_pkey(eib_t *, eib_chan_t *, ib_pkey_t, boolean_t,
44 boolean_t *);
45 static boolean_t eib_ibt_has_chan_pkey_changed(eib_t *, eib_chan_t *);
46 static boolean_t eib_ibt_has_any_pkey_changed(eib_t *);
47 static int eib_ibt_fill_avect(eib_t *, eib_avect_t *, ib_lid_t);
48 static void eib_ibt_record_srate(eib_t *);
49
50 /*
51 * Definitions private to this file
52 */
53
54 /*
55 * SM's init type reply flags
56 */
57 #define EIB_PORT_ATTR_LOADED(itr) \
58 (((itr) & SM_INIT_TYPE_REPLY_NO_LOAD_REPLY) == 0)
59 #define EIB_PORT_ATTR_NOT_PRESERVED(itr) \
60 (((itr) & SM_INIT_TYPE_PRESERVE_CONTENT_REPLY) == 0)
61 #define EIB_PORT_PRES_NOT_PRESERVED(itr) \
62 (((itr) & SM_INIT_TYPE_PRESERVE_PRESENCE_REPLY) == 0)
63
64 /*
65 * eib_ibt_hca_init() initialization progress flags
66 */
67 #define EIB_HCAINIT_HCA_OPENED 0x01
68 #define EIB_HCAINIT_ATTRS_ALLOCD 0x02
69 #define EIB_HCAINIT_HCA_PORTS_QUERIED 0x04
70 #define EIB_HCAINIT_PD_ALLOCD 0x08
71 #define EIB_HCAINIT_CAPAB_RECORDED 0x10
72
73 int
eib_ibt_hca_init(eib_t * ss)74 eib_ibt_hca_init(eib_t *ss)
75 {
76 ibt_status_t ret;
77 ibt_hca_portinfo_t *pi;
78 uint_t num_pi;
79 uint_t sz_pi;
80 uint_t progress = 0;
81
82 if (ss->ei_hca_hdl)
83 return (EIB_E_SUCCESS);
84
85 /*
86 * Open the HCA
87 */
88 ret = ibt_open_hca(ss->ei_ibt_hdl, ss->ei_props->ep_hca_guid,
89 &ss->ei_hca_hdl);
90 if (ret != IBT_SUCCESS) {
91 EIB_DPRINTF_ERR(ss->ei_instance,
92 "ibt_open_hca(hca_guid=0x%llx) "
93 "failed, ret=%d", ss->ei_props->ep_hca_guid, ret);
94 goto ibt_hca_init_fail;
95 }
96 progress |= EIB_HCAINIT_HCA_OPENED;
97
98 /*
99 * Query and store HCA attributes
100 */
101 ss->ei_hca_attrs = kmem_zalloc(sizeof (ibt_hca_attr_t), KM_SLEEP);
102 progress |= EIB_HCAINIT_ATTRS_ALLOCD;
103
104 ret = ibt_query_hca(ss->ei_hca_hdl, ss->ei_hca_attrs);
105 if (ret != IBT_SUCCESS) {
106 EIB_DPRINTF_ERR(ss->ei_instance,
107 "ibt_query_hca(hca_hdl=0x%llx, "
108 "hca_guid=0x%llx) failed, ret=%d",
109 ss->ei_hca_hdl, ss->ei_props->ep_hca_guid, ret);
110 goto ibt_hca_init_fail;
111 }
112
113 /*
114 * At this point, we don't even care about the linkstate, we only want
115 * to record our invariant base port guid and mtu
116 */
117 ret = ibt_query_hca_ports(ss->ei_hca_hdl, ss->ei_props->ep_port_num,
118 &pi, &num_pi, &sz_pi);
119 if (ret != IBT_SUCCESS) {
120 EIB_DPRINTF_ERR(ss->ei_instance,
121 "ibt_query_hca_ports(hca_hdl=0x%llx, "
122 "port=0x%x) failed, ret=%d", ss->ei_hca_hdl,
123 ss->ei_props->ep_port_num, ret);
124 goto ibt_hca_init_fail;
125 }
126 if (num_pi != 1) {
127 EIB_DPRINTF_ERR(ss->ei_instance,
128 "ibt_query_hca_ports(hca_hdl=0x%llx, "
129 "port=0x%x) returned num_pi=%d", ss->ei_hca_hdl,
130 ss->ei_props->ep_port_num, num_pi);
131 ibt_free_portinfo(pi, sz_pi);
132 goto ibt_hca_init_fail;
133 }
134
135 ss->ei_props->ep_sgid = pi->p_sgid_tbl[0];
136 ss->ei_props->ep_mtu = (128 << pi->p_mtu);
137 ibt_free_portinfo(pi, sz_pi);
138
139 progress |= EIB_HCAINIT_HCA_PORTS_QUERIED;
140
141 /*
142 * Allocate a protection domain for all our transactions
143 */
144 ret = ibt_alloc_pd(ss->ei_hca_hdl, IBT_PD_NO_FLAGS, &ss->ei_pd_hdl);
145 if (ret != IBT_SUCCESS) {
146 EIB_DPRINTF_ERR(ss->ei_instance,
147 "ibt_alloc_pd(hca_hdl=0x%llx, "
148 "hca_guid=0x%llx) failed, ret=%d",
149 ss->ei_hca_hdl, ss->ei_props->ep_hca_guid, ret);
150 goto ibt_hca_init_fail;
151 }
152 progress |= EIB_HCAINIT_PD_ALLOCD;
153
154 /*
155 * Finally, record the capabilities
156 */
157 ss->ei_caps = kmem_zalloc(sizeof (eib_caps_t), KM_SLEEP);
158 eib_ibt_record_capab(ss, ss->ei_hca_attrs, ss->ei_caps);
159 eib_ibt_record_srate(ss);
160
161 progress |= EIB_HCAINIT_CAPAB_RECORDED;
162
163 return (EIB_E_SUCCESS);
164
165 ibt_hca_init_fail:
166 eib_rb_ibt_hca_init(ss, progress);
167 return (EIB_E_FAILURE);
168 }
169
170 void
eib_ibt_link_mod(eib_t * ss)171 eib_ibt_link_mod(eib_t *ss)
172 {
173 eib_node_state_t *ns = ss->ei_node_state;
174 ibt_hca_portinfo_t *pi;
175 ibt_status_t ret;
176 uint8_t vn0_mac[ETHERADDRL];
177 boolean_t all_zombies = B_FALSE;
178 boolean_t all_need_rejoin = B_FALSE;
179 uint_t num_pi;
180 uint_t sz_pi;
181 uint8_t itr;
182
183 if (ns->ns_link_state == LINK_STATE_UNKNOWN)
184 return;
185
186 /*
187 * See if we can get the port attributes or we're as good as down.
188 */
189 ret = ibt_query_hca_ports(ss->ei_hca_hdl, ss->ei_props->ep_port_num,
190 &pi, &num_pi, &sz_pi);
191 if ((ret != IBT_SUCCESS) || (pi->p_linkstate != IBT_PORT_ACTIVE)) {
192 ibt_free_portinfo(pi, sz_pi);
193 eib_mac_link_down(ss, B_FALSE);
194 return;
195 }
196
197 /*
198 * If the SM re-initialized the port attributes, but did not preserve
199 * the old attributes, we need to check more.
200 */
201 itr = pi->p_init_type_reply;
202 if (EIB_PORT_ATTR_LOADED(itr) && EIB_PORT_ATTR_NOT_PRESERVED(itr)) {
203 /*
204 * We're just coming back up; if we see that our base lid
205 * or sgid table has changed, we'll update these and try to
206 * restart all active vnics. If any of the vnic pkeys have
207 * changed, we'll reset the affected channels to the new pkey.
208 */
209 if (bcmp(pi->p_sgid_tbl, &ss->ei_props->ep_sgid,
210 sizeof (ib_gid_t)) != 0) {
211 EIB_DPRINTF_VERBOSE(ss->ei_instance,
212 "eib_ibt_link_mod: port sgid table changed "
213 "(old %llx.%llx != new %llx.%llx), "
214 "all vnics are zombies now.",
215 ss->ei_props->ep_sgid.gid_prefix,
216 ss->ei_props->ep_sgid.gid_guid,
217 pi->p_sgid_tbl[0].gid_prefix,
218 pi->p_sgid_tbl[0].gid_guid);
219
220 ss->ei_props->ep_sgid = pi->p_sgid_tbl[0];
221 all_zombies = B_TRUE;
222
223 } else if (ss->ei_props->ep_blid != pi->p_base_lid) {
224 EIB_DPRINTF_VERBOSE(ss->ei_instance,
225 "eib_ibt_link_mod: port base lid changed "
226 "(old 0x%x != new 0x%x), "
227 "all vnics are zombies now.",
228 ss->ei_props->ep_blid, pi->p_base_lid);
229
230 ss->ei_props->ep_blid = pi->p_base_lid;
231 all_zombies = B_TRUE;
232
233 } else if (eib_ibt_has_any_pkey_changed(ss)) {
234 EIB_DPRINTF_VERBOSE(ss->ei_instance,
235 "eib_ibt_link_mod: pkey has changed for vnic(s), "
236 "resetting all partitions");
237
238 eib_ibt_reset_partitions(ss);
239 }
240 }
241
242 if (pi) {
243 ibt_free_portinfo(pi, sz_pi);
244 }
245
246 /*
247 * If the SM hasn't preserved our presence in MCGs, we need to
248 * rejoin all of them.
249 */
250 if (EIB_PORT_PRES_NOT_PRESERVED(itr)) {
251 EIB_DPRINTF_VERBOSE(ss->ei_instance, "eib_ibt_link_mod: "
252 "hca_guid=0x%llx, port=0x%x presence not preserved in SM, "
253 "rejoining all mcgs", ss->ei_props->ep_hca_guid,
254 ss->ei_props->ep_port_num);
255
256 all_need_rejoin = B_TRUE;
257 }
258
259 /*
260 * Before we do the actual work of restarting/rejoining, we need to
261 * see if the GW is reachable at this point of time. If not, we
262 * still continue to keep our link "down." Whenever the GW becomes
263 * reachable again, we'll restart/rejoin all the vnics that we've
264 * just marked.
265 */
266 mutex_enter(&ss->ei_vnic_lock);
267 if (all_zombies) {
268 ss->ei_zombie_vnics = ss->ei_active_vnics;
269 }
270 if (all_need_rejoin) {
271 ss->ei_rejoin_vnics = ss->ei_active_vnics;
272 }
273 if (ss->ei_gw_unreachable) {
274 mutex_exit(&ss->ei_vnic_lock);
275
276 EIB_DPRINTF_WARN(ss->ei_instance, "eib_ibt_link_mod: "
277 "gateway (gw_port=0x%x) unreachable for "
278 "hca_guid=0x%llx, port=0x%x, link state down",
279 ss->ei_gw_props->pp_gw_portid, ss->ei_props->ep_hca_guid,
280 ss->ei_props->ep_port_num);
281
282 eib_mac_link_down(ss, B_FALSE);
283 return;
284 }
285 mutex_exit(&ss->ei_vnic_lock);
286
287 /*
288 * Try to awaken the dead if possible
289 */
290 bcopy(eib_zero_mac, vn0_mac, ETHERADDRL);
291 if (all_zombies) {
292 EIB_DPRINTF_VERBOSE(ss->ei_instance, "eib_ibt_link_mod: "
293 "hca_guid=0x%llx, hca_port=0x%x, gw_port=0x%x, "
294 "attempting to resurrect zombies",
295 ss->ei_props->ep_hca_guid, ss->ei_props->ep_port_num,
296 ss->ei_gw_props->pp_gw_portid);
297
298 eib_vnic_resurrect_zombies(ss, vn0_mac);
299 }
300
301 /*
302 * Re-join the mcgs if we need to
303 */
304 if (all_need_rejoin) {
305 EIB_DPRINTF_VERBOSE(ss->ei_instance, "eib_ibt_link_mod: "
306 "hca_guid=0x%llx, hca_port=0x%x, gw_port=0x%x, "
307 "attempting to rejoin mcgs",
308 ss->ei_props->ep_hca_guid, ss->ei_props->ep_port_num,
309 ss->ei_gw_props->pp_gw_portid);
310
311 eib_vnic_rejoin_mcgs(ss);
312 }
313
314 /*
315 * If we've restarted the zombies because the gateway went down and
316 * came back, it is possible our unicast mac address changed from
317 * what it was earlier. If so, we need to update our unicast address
318 * with the mac layer before marking the link up.
319 */
320 if (bcmp(vn0_mac, eib_zero_mac, ETHERADDRL) != 0)
321 mac_unicst_update(ss->ei_mac_hdl, vn0_mac);
322
323 /*
324 * Notify the link state up if required
325 */
326 eib_mac_link_up(ss, B_FALSE);
327 }
328
329 int
eib_ibt_modify_chan_pkey(eib_t * ss,eib_chan_t * chan,ib_pkey_t pkey)330 eib_ibt_modify_chan_pkey(eib_t *ss, eib_chan_t *chan, ib_pkey_t pkey)
331 {
332 /*
333 * Make sure the channel pkey and index are set to what we need
334 */
335 return (eib_ibt_chan_pkey(ss, chan, pkey, B_TRUE, NULL));
336 }
337
338 eib_avect_t *
eib_ibt_hold_avect(eib_t * ss,ib_lid_t dlid,uint8_t sl)339 eib_ibt_hold_avect(eib_t *ss, ib_lid_t dlid, uint8_t sl)
340 {
341 uint_t ndx = dlid % EIB_AV_NBUCKETS; /* simple hashing */
342 eib_avect_t *av;
343 eib_avect_t *prev;
344 int ret;
345
346 mutex_enter(&ss->ei_av_lock);
347
348 /*
349 * See if we have the address vector
350 */
351 prev = NULL;
352 for (av = ss->ei_av[ndx]; av; av = av->av_next) {
353 prev = av;
354 if ((av->av_vect).av_dlid == dlid)
355 break;
356 }
357
358 /*
359 * If we don't have it, create a new one and chain it to
360 * the same bucket
361 */
362 if (av == NULL) {
363 av = kmem_zalloc(sizeof (eib_avect_t), KM_NOSLEEP);
364 if (av == NULL) {
365 mutex_exit(&ss->ei_av_lock);
366 EIB_DPRINTF_WARN(ss->ei_instance, "eib_ibt_hold_avect: "
367 "no memory, could not allocate address vector");
368 return (NULL);
369 }
370
371 ret = EIB_E_FAILURE;
372 if (!eib_wa_no_av_discover)
373 ret = eib_ibt_fill_avect(ss, av, dlid);
374
375 if (ret != EIB_E_SUCCESS) {
376 (av->av_vect).av_srate = IBT_SRATE_10;
377 (av->av_vect).av_srvl = sl;
378 (av->av_vect).av_port_num = ss->ei_props->ep_port_num;
379 (av->av_vect).av_send_grh = B_FALSE;
380 (av->av_vect).av_dlid = dlid;
381 (av->av_vect).av_src_path = 0; /* we use base lid */
382 }
383
384 if (prev)
385 prev->av_next = av;
386 else
387 ss->ei_av[ndx] = av;
388 }
389
390 /*
391 * Increment the address vector reference count before returning
392 */
393 (av->av_ref)++;
394
395 mutex_exit(&ss->ei_av_lock);
396
397 return (av);
398 }
399
400 static int
eib_ibt_fill_avect(eib_t * ss,eib_avect_t * av,ib_lid_t dlid)401 eib_ibt_fill_avect(eib_t *ss, eib_avect_t *av, ib_lid_t dlid)
402 {
403 ibt_node_info_t ni;
404 ibt_path_attr_t attr;
405 ibt_path_info_t path;
406 ibt_status_t ret;
407 ib_gid_t dgid;
408
409 if ((ret = ibt_lid_to_node_info(dlid, &ni)) != IBT_SUCCESS) {
410 EIB_DPRINTF_WARN(ss->ei_instance, "eib_ibt_fill_avect: "
411 "ibt_lid_to_node_info(dlid=0x%x) failed, ret=%d",
412 dlid, ret);
413 return (EIB_E_FAILURE);
414 }
415 dgid.gid_prefix = ss->ei_gw_props->pp_gw_sn_prefix;
416 dgid.gid_guid = ni.n_port_guid;
417
418 /*
419 * Get the reversible path information for this destination
420 */
421 bzero(&attr, sizeof (ibt_path_info_t));
422 attr.pa_sgid = ss->ei_props->ep_sgid;
423 attr.pa_dgids = &dgid;
424 attr.pa_num_dgids = 1;
425
426 bzero(&path, sizeof (ibt_path_info_t));
427 ret = ibt_get_paths(ss->ei_ibt_hdl, IBT_PATH_NO_FLAGS,
428 &attr, 1, &path, NULL);
429 if ((ret != IBT_SUCCESS) || (path.pi_hca_guid == 0)) {
430 EIB_DPRINTF_WARN(ss->ei_instance, "eib_ibt_fill_avect: "
431 "ibt_get_paths(dgid=%llx.%llx) failed, ret=%d",
432 dgid.gid_prefix, dgid.gid_guid);
433 return (EIB_E_FAILURE);
434 }
435
436 /*
437 * Fill in the address vector
438 */
439 bcopy(&path.pi_prim_cep_path.cep_adds_vect, &av->av_vect,
440 sizeof (ibt_adds_vect_t));
441
442 return (EIB_E_SUCCESS);
443 }
444
445 void
eib_ibt_release_avect(eib_t * ss,eib_avect_t * av)446 eib_ibt_release_avect(eib_t *ss, eib_avect_t *av)
447 {
448 mutex_enter(&ss->ei_av_lock);
449
450 ASSERT(av->av_ref > 0);
451 (av->av_ref)--;
452
453 mutex_exit(&ss->ei_av_lock);
454 }
455
456 void
eib_ibt_free_avects(eib_t * ss)457 eib_ibt_free_avects(eib_t *ss)
458 {
459 eib_avect_t *av;
460 eib_avect_t *av_next;
461 int ndx;
462
463 mutex_enter(&ss->ei_av_lock);
464 for (ndx = 0; ndx < EIB_AV_NBUCKETS; ndx++) {
465 for (av = ss->ei_av[ndx]; av; av = av_next) {
466 av_next = av->av_next;
467
468 ASSERT(av->av_ref == 0);
469 kmem_free(av, sizeof (eib_avect_t));
470 }
471 ss->ei_av[ndx] = NULL;
472 }
473 mutex_exit(&ss->ei_av_lock);
474 }
475
476 /*ARGSUSED*/
477 void
eib_ibt_async_handler(void * clnt_private,ibt_hca_hdl_t hca_hdl,ibt_async_code_t code,ibt_async_event_t * event)478 eib_ibt_async_handler(void *clnt_private, ibt_hca_hdl_t hca_hdl,
479 ibt_async_code_t code, ibt_async_event_t *event)
480 {
481 eib_t *ss = (eib_t *)clnt_private;
482 eib_event_t *evi;
483 uint_t ev_code;
484
485 ev_code = EIB_EV_NONE;
486
487 switch (code) {
488 case IBT_EVENT_SQD:
489 EIB_DPRINTF_VERBOSE(ss->ei_instance,
490 "eib_ibt_async_handler: got IBT_EVENT_SQD");
491 eib_ibt_wakeup_sqd_waiters(ss, event->ev_chan_hdl);
492 break;
493
494 case IBT_EVENT_PORT_UP:
495 if (event->ev_port == ss->ei_props->ep_port_num) {
496 EIB_DPRINTF_VERBOSE(ss->ei_instance,
497 "eib_ibt_async_handler: got IBT_EVENT_PORT_UP");
498 ev_code = EIB_EV_PORT_UP;
499 }
500 break;
501
502 case IBT_ERROR_PORT_DOWN:
503 if (event->ev_port == ss->ei_props->ep_port_num) {
504 EIB_DPRINTF_VERBOSE(ss->ei_instance,
505 "eib_ibt_async_handler: got IBT_ERROR_PORT_DOWN");
506 ev_code = EIB_EV_PORT_DOWN;
507 }
508 break;
509
510 case IBT_CLNT_REREG_EVENT:
511 if (event->ev_port == ss->ei_props->ep_port_num) {
512 EIB_DPRINTF_VERBOSE(ss->ei_instance,
513 "eib_ibt_async_handler: got IBT_CLNT_REREG_EVENT");
514 ev_code = EIB_EV_CLNT_REREG;
515 }
516 break;
517
518 case IBT_PORT_CHANGE_EVENT:
519 if ((event->ev_port == ss->ei_props->ep_port_num) &&
520 (event->ev_port_flags & IBT_PORT_CHANGE_PKEY)) {
521 EIB_DPRINTF_VERBOSE(ss->ei_instance,
522 "eib_ibt_async_handler: "
523 "got IBT_PORT_CHANGE_EVENT(PKEY_CHANGE)");
524 ev_code = EIB_EV_PKEY_CHANGE;
525 } else if ((event->ev_port == ss->ei_props->ep_port_num) &&
526 (event->ev_port_flags & IBT_PORT_CHANGE_SGID)) {
527 EIB_DPRINTF_VERBOSE(ss->ei_instance,
528 "eib_ibt_async_handler: "
529 "got IBT_PORT_CHANGE_EVENT(SGID_CHANGE)");
530 ev_code = EIB_EV_SGID_CHANGE;
531 }
532 break;
533
534 case IBT_HCA_ATTACH_EVENT:
535 /*
536 * For HCA attach, after a new HCA is plugged in and
537 * configured using cfgadm, an explicit plumb will need
538 * to be run, so we don't need to do anything here.
539 */
540 EIB_DPRINTF_VERBOSE(ss->ei_instance, "eib_ibt_async_handler: "
541 "got IBT_HCA_ATTACH_EVENT");
542 break;
543
544 case IBT_HCA_DETACH_EVENT:
545 /*
546 * Before an HCA unplug, cfgadm is expected to trigger
547 * any rcm scripts to unplumb the EoIB instances on the
548 * card. If so, we should not be holding any hca resource,
549 * since we don't do ibt_open_hca() until plumb time. However,
550 * if an earlier unplumb hadn't cleaned up the hca resources
551 * properly because the network layer hadn't returned the
552 * buffers at that time, we could be holding hca resources.
553 * We'll try to release them here, and protect the code from
554 * racing with some other plumb/unplumb operation.
555 */
556 EIB_DPRINTF_VERBOSE(ss->ei_instance, "eib_ibt_async_handler: "
557 "got IBT_HCA_DETACH_EVENT");
558
559 eib_mac_set_nic_state(ss, EIB_NIC_STOPPING);
560 eib_rb_rsrc_setup_bufs(ss, B_FALSE);
561 if (ss->ei_tx || ss->ei_rx || ss->ei_lso) {
562 EIB_DPRINTF_WARN(ss->ei_instance,
563 "eib_events_handler: nw layer still holding "
564 "hca resources, could not detach HCA");
565 } else if (ss->ei_hca_hdl) {
566 eib_rb_ibt_hca_init(ss, ~0);
567 }
568 eib_mac_clr_nic_state(ss, EIB_NIC_STOPPING);
569
570 break;
571 }
572
573 if (ev_code != EIB_EV_NONE) {
574 evi = kmem_zalloc(sizeof (eib_event_t), KM_NOSLEEP);
575 if (evi == NULL) {
576 EIB_DPRINTF_WARN(ss->ei_instance,
577 "eib_ibt_async_handler: "
578 "no memory, could not handle event 0x%lx", ev_code);
579 } else {
580 evi->ev_code = ev_code;
581 evi->ev_arg = NULL;
582 eib_svc_enqueue_event(ss, evi);
583 }
584 }
585 }
586
587 /*ARGSUSED*/
588 void
eib_ibt_record_capab(eib_t * ss,ibt_hca_attr_t * hca_attrs,eib_caps_t * caps)589 eib_ibt_record_capab(eib_t *ss, ibt_hca_attr_t *hca_attrs, eib_caps_t *caps)
590 {
591 uint_t max_swqe = EIB_DATA_MAX_SWQE;
592 uint_t max_rwqe = EIB_DATA_MAX_RWQE;
593
594 /*
595 * Checksum
596 */
597 caps->cp_cksum_flags = 0;
598 if ((!eib_wa_no_cksum_offload) &&
599 (hca_attrs->hca_flags & IBT_HCA_CKSUM_FULL)) {
600 caps->cp_cksum_flags =
601 HCK_FULLCKSUM | HCKSUM_INET_FULL_V4;
602 /* HCKSUM_INET_FULL_V4 | HCKSUM_IPHDRCKSUM; */
603 }
604
605 /*
606 * Reserved L-Key
607 */
608 if (hca_attrs->hca_flags2 & IBT_HCA2_RES_LKEY) {
609 caps->cp_resv_lkey_capab = 1;
610 caps->cp_resv_lkey = hca_attrs->hca_reserved_lkey;
611 }
612
613 /*
614 * LSO
615 */
616 caps->cp_lso_maxlen = 0;
617 if (!eib_wa_no_lso) {
618 if (hca_attrs->hca_max_lso_size > EIB_LSO_MAXLEN) {
619 caps->cp_lso_maxlen = EIB_LSO_MAXLEN;
620 } else {
621 caps->cp_lso_maxlen = hca_attrs->hca_max_lso_size;
622 }
623 }
624
625 /*
626 * SGL
627 *
628 * Translating virtual address regions into physical regions
629 * for using the Reserved LKey feature results in a wr sgl that
630 * is a little longer. Since failing ibt_map_mem_iov() is costly,
631 * we'll record a high-water mark (65%) when we should stop
632 * trying to use Reserved LKey
633 */
634 if (hca_attrs->hca_flags & IBT_HCA_WQE_SIZE_INFO) {
635 caps->cp_max_sgl = hca_attrs->hca_ud_send_sgl_sz;
636 } else {
637 caps->cp_max_sgl = hca_attrs->hca_max_sgl;
638 }
639 if (caps->cp_max_sgl > EIB_MAX_SGL) {
640 caps->cp_max_sgl = EIB_MAX_SGL;
641 }
642 caps->cp_hiwm_sgl = (caps->cp_max_sgl * 65) / 100;
643
644 /*
645 * SWQE/RWQE: meet max chan size and max cq size limits (leave room
646 * to avoid cq overflow event)
647 */
648 if (max_swqe > hca_attrs->hca_max_chan_sz)
649 max_swqe = hca_attrs->hca_max_chan_sz;
650 if (max_swqe > (hca_attrs->hca_max_cq_sz - 1))
651 max_swqe = hca_attrs->hca_max_cq_sz - 1;
652 caps->cp_max_swqe = max_swqe;
653
654 if (max_rwqe > hca_attrs->hca_max_chan_sz)
655 max_rwqe = hca_attrs->hca_max_chan_sz;
656 if (max_rwqe > (hca_attrs->hca_max_cq_sz - 1))
657 max_rwqe = hca_attrs->hca_max_cq_sz - 1;
658 caps->cp_max_rwqe = max_rwqe;
659 }
660
661 void
eib_rb_ibt_hca_init(eib_t * ss,uint_t progress)662 eib_rb_ibt_hca_init(eib_t *ss, uint_t progress)
663 {
664 ibt_status_t ret;
665
666 if (progress & EIB_HCAINIT_CAPAB_RECORDED) {
667 if (ss->ei_caps) {
668 kmem_free(ss->ei_caps, sizeof (eib_caps_t));
669 ss->ei_caps = NULL;
670 }
671 }
672
673 if (progress & EIB_HCAINIT_PD_ALLOCD) {
674 if (ss->ei_pd_hdl) {
675 ret = ibt_free_pd(ss->ei_hca_hdl, ss->ei_pd_hdl);
676 if (ret != IBT_SUCCESS) {
677 EIB_DPRINTF_WARN(ss->ei_instance,
678 "eib_rb_ibt_hca_init: "
679 "ibt_free_pd(hca_hdl=0x%lx, pd_hdl=0x%lx) "
680 "failed, ret=%d", ss->ei_hca_hdl,
681 ss->ei_pd_hdl, ret);
682 }
683 ss->ei_pd_hdl = NULL;
684 }
685 }
686
687 if (progress & EIB_HCAINIT_HCA_PORTS_QUERIED) {
688 ss->ei_props->ep_mtu = 0;
689 bzero(&ss->ei_props->ep_sgid, sizeof (ib_gid_t));
690 }
691
692 if (progress & EIB_HCAINIT_ATTRS_ALLOCD) {
693 kmem_free(ss->ei_hca_attrs, sizeof (ibt_hca_attr_t));
694 ss->ei_hca_attrs = NULL;
695 }
696
697 if (progress & EIB_HCAINIT_HCA_OPENED) {
698 ret = ibt_close_hca(ss->ei_hca_hdl);
699 if (ret != IBT_SUCCESS) {
700 EIB_DPRINTF_WARN(ss->ei_instance,
701 "ibt_close_hca(hca_hdl=0x%lx) failed, "
702 "ret=%d", ss->ei_hca_hdl, ret);
703 }
704 ss->ei_hca_hdl = NULL;
705 }
706 }
707
708 static void
eib_ibt_reset_partitions(eib_t * ss)709 eib_ibt_reset_partitions(eib_t *ss)
710 {
711 eib_vnic_t *vnic;
712 eib_chan_t *chan = NULL;
713 uint64_t av;
714 int inst = 0;
715
716 /*
717 * We already have the vhub pkey recorded in our eib_chan_t.
718 * We only need to make sure our pkey index still matches it.
719 * If not, modify the channel appropriately and update our
720 * records.
721 */
722 if ((chan = ss->ei_admin_chan) != NULL)
723 (void) eib_ibt_modify_chan_pkey(ss, chan, chan->ch_pkey);
724
725 mutex_enter(&ss->ei_vnic_lock);
726 av = ss->ei_active_vnics;
727 while ((inst = EIB_FIND_LSB_SET(av)) != -1) {
728 if ((vnic = ss->ei_vnic[inst]) != NULL) {
729 if ((chan = vnic->vn_ctl_chan) != NULL) {
730 (void) eib_ibt_modify_chan_pkey(ss, chan,
731 chan->ch_pkey);
732 }
733 if ((chan = vnic->vn_data_chan) != NULL) {
734 (void) eib_ibt_modify_chan_pkey(ss, chan,
735 chan->ch_pkey);
736 }
737 }
738 av &= (~((uint64_t)1 << inst));
739 }
740 mutex_exit(&ss->ei_vnic_lock);
741 }
742
743 static void
eib_ibt_wakeup_sqd_waiters(eib_t * ss,ibt_channel_hdl_t ev_chan_hdl)744 eib_ibt_wakeup_sqd_waiters(eib_t *ss, ibt_channel_hdl_t ev_chan_hdl)
745 {
746 eib_vnic_t *vnic;
747 eib_chan_t *chan = NULL;
748 uint64_t av;
749 int inst = 0;
750
751 /*
752 * See if this channel has been waiting for its queue to drain.
753 *
754 * Note that since this is especially likely to be called during
755 * logging in to the gateway, we also need to check the vnic
756 * currently being created.
757 */
758 mutex_enter(&ss->ei_vnic_lock);
759
760 if ((vnic = ss->ei_vnic_pending) != NULL) {
761 chan = vnic->vn_ctl_chan;
762 if ((chan) && (chan->ch_chan == ev_chan_hdl))
763 goto wakeup_sqd_waiters;
764
765 chan = vnic->vn_data_chan;
766 if ((chan) && (chan->ch_chan == ev_chan_hdl))
767 goto wakeup_sqd_waiters;
768 }
769
770 av = ss->ei_active_vnics;
771 while ((inst = EIB_FIND_LSB_SET(av)) != -1) {
772 if ((vnic = ss->ei_vnic[inst]) != NULL) {
773 chan = vnic->vn_ctl_chan;
774 if (chan->ch_chan == ev_chan_hdl)
775 break;
776
777 chan = vnic->vn_data_chan;
778 if (chan->ch_chan == ev_chan_hdl)
779 break;
780 }
781 av &= (~((uint64_t)1 << inst));
782 }
783
784 wakeup_sqd_waiters:
785 if (chan) {
786 mutex_enter(&chan->ch_cep_lock);
787 chan->ch_cep_state = IBT_STATE_SQD;
788 cv_broadcast(&chan->ch_cep_cv);
789 mutex_exit(&chan->ch_cep_lock);
790 }
791
792 mutex_exit(&ss->ei_vnic_lock);
793 }
794
795 static int
eib_ibt_chan_pkey(eib_t * ss,eib_chan_t * chan,ib_pkey_t new_pkey,boolean_t set,boolean_t * pkey_changed)796 eib_ibt_chan_pkey(eib_t *ss, eib_chan_t *chan, ib_pkey_t new_pkey,
797 boolean_t set, boolean_t *pkey_changed)
798 {
799 ibt_qp_info_t qp_attr;
800 ibt_status_t ret;
801 uint16_t new_pkey_ix;
802
803 ret = ibt_pkey2index(ss->ei_hca_hdl, ss->ei_props->ep_port_num,
804 new_pkey, &new_pkey_ix);
805 if (ret != IBT_SUCCESS) {
806 EIB_DPRINTF_WARN(ss->ei_instance, "eib_ibt_chan_pkey: "
807 "ibt_pkey2index(hca_hdl=0x%llx, port_num=0x%x, "
808 "pkey=0x%x) failed, ret=%d",
809 ss->ei_hca_hdl, ss->ei_props->ep_port_num, new_pkey, ret);
810 return (EIB_E_FAILURE);
811 }
812
813 /*
814 * If the pkey and the pkey index we have already matches the
815 * new one, nothing to do.
816 */
817 mutex_enter(&chan->ch_pkey_lock);
818 if ((chan->ch_pkey == new_pkey) && (chan->ch_pkey_ix == new_pkey_ix)) {
819 if (pkey_changed) {
820 *pkey_changed = B_FALSE;
821 }
822 mutex_exit(&chan->ch_pkey_lock);
823 return (EIB_E_SUCCESS);
824 }
825 if (pkey_changed) {
826 *pkey_changed = B_TRUE;
827 }
828 mutex_exit(&chan->ch_pkey_lock);
829
830 /*
831 * Otherwise, if we're asked only to test if the pkey index
832 * supplied matches the one recorded in the channel, return
833 * success, but don't set the pkey.
834 */
835 if (!set) {
836 return (EIB_E_SUCCESS);
837 }
838
839 /*
840 * Otherwise, we need to change channel pkey. Pause the
841 * channel sendq first.
842 */
843 ret = ibt_pause_sendq(chan->ch_chan, IBT_CEP_SET_SQD_EVENT);
844 if (ret != IBT_SUCCESS) {
845 EIB_DPRINTF_WARN(ss->ei_instance, "eib_ibt_chan_pkey: "
846 "ibt_pause_sendq(chan_hdl=0x%llx) failed, ret=%d",
847 chan->ch_chan, ret);
848 return (EIB_E_FAILURE);
849 }
850
851 /*
852 * Wait for the channel to enter the IBT_STATE_SQD state
853 */
854 mutex_enter(&chan->ch_cep_lock);
855 while (chan->ch_cep_state != IBT_STATE_SQD)
856 cv_wait(&chan->ch_cep_cv, &chan->ch_cep_lock);
857 mutex_exit(&chan->ch_cep_lock);
858
859 /*
860 * Modify the qp with the supplied pkey index and unpause the channel
861 * If either of these operations fail, we'll leave the channel in
862 * the paused state and fail.
863 */
864 bzero(&qp_attr, sizeof (ibt_qp_info_t));
865
866 qp_attr.qp_trans = IBT_UD_SRV;
867 qp_attr.qp_current_state = IBT_STATE_SQD;
868 qp_attr.qp_state = IBT_STATE_SQD;
869 qp_attr.qp_transport.ud.ud_pkey_ix = new_pkey_ix;
870
871 /*
872 * Modify the qp to set the new pkey index, then unpause the
873 * channel and put it in RTS state and update the new values
874 * in our records
875 */
876 mutex_enter(&chan->ch_pkey_lock);
877
878 ret = ibt_modify_qp(chan->ch_chan,
879 IBT_CEP_SET_STATE | IBT_CEP_SET_PKEY_IX, &qp_attr, NULL);
880 if (ret != IBT_SUCCESS) {
881 mutex_exit(&chan->ch_pkey_lock);
882 EIB_DPRINTF_WARN(ss->ei_instance, "eib_ibt_chan_pkey: "
883 "ibt_modify_qp(chan_hdl=0x%llx, IBT_CEP_SET_PKEY_IX) "
884 "failed for new_pkey_ix=0x%x, ret=%d",
885 chan->ch_chan, new_pkey_ix, ret);
886 return (EIB_E_FAILURE);
887 }
888
889 if ((ret = ibt_unpause_sendq(chan->ch_chan)) != IBT_SUCCESS) {
890 mutex_exit(&chan->ch_pkey_lock);
891 EIB_DPRINTF_WARN(ss->ei_instance, "eib_ibt_chan_pkey: "
892 "ibt_unpause_sendq(chan_hdl=0x%llx) failed, ret=%d",
893 chan->ch_chan, ret);
894 return (EIB_E_FAILURE);
895 }
896
897 chan->ch_pkey = new_pkey;
898 chan->ch_pkey_ix = new_pkey_ix;
899 mutex_exit(&chan->ch_pkey_lock);
900
901 return (EIB_E_SUCCESS);
902 }
903
904 static boolean_t
eib_ibt_has_chan_pkey_changed(eib_t * ss,eib_chan_t * chan)905 eib_ibt_has_chan_pkey_changed(eib_t *ss, eib_chan_t *chan)
906 {
907 boolean_t changed;
908 int ret;
909
910 /*
911 * Don't modify the pkey, just ask if the pkey index for the channel's
912 * pkey has changed for any reason. If we fail, assume that the pkey
913 * has changed.
914 */
915 ret = eib_ibt_chan_pkey(ss, chan, chan->ch_pkey, B_FALSE, &changed);
916 if (ret != EIB_E_SUCCESS)
917 changed = B_TRUE;
918
919 return (changed);
920 }
921
922 static boolean_t
eib_ibt_has_any_pkey_changed(eib_t * ss)923 eib_ibt_has_any_pkey_changed(eib_t *ss)
924 {
925 eib_vnic_t *vnic;
926 eib_chan_t *chan = NULL;
927 uint64_t av;
928 int inst = 0;
929
930 /*
931 * Return true if the pkey index of any our pkeys (of the channels
932 * of all active vnics) has changed.
933 */
934
935 chan = ss->ei_admin_chan;
936 if ((chan) && (eib_ibt_has_chan_pkey_changed(ss, chan)))
937 return (B_TRUE);
938
939 mutex_enter(&ss->ei_vnic_lock);
940 av = ss->ei_active_vnics;
941 while ((inst = EIB_FIND_LSB_SET(av)) != -1) {
942 if ((vnic = ss->ei_vnic[inst]) != NULL) {
943 chan = vnic->vn_ctl_chan;
944 if ((chan) && (eib_ibt_has_chan_pkey_changed(ss, chan)))
945 return (B_TRUE);
946
947 chan = vnic->vn_data_chan;
948 if ((chan) && (eib_ibt_has_chan_pkey_changed(ss, chan)))
949 return (B_TRUE);
950 }
951 av &= (~((uint64_t)1 << inst));
952 }
953 mutex_exit(&ss->ei_vnic_lock);
954
955 return (B_FALSE);
956 }
957
958 /*
959 * This routine is currently used simply to derive and record the port
960 * speed from the loopback path information (for debug purposes). For
961 * EoIB, currently the srate used in address vectors to IB neighbors
962 * and the gateway is fixed at IBT_SRATE_10. Eventually though, this
963 * information (and sl) has to come from the gateway for all destinations
964 * in the vhub table.
965 */
966 static void
eib_ibt_record_srate(eib_t * ss)967 eib_ibt_record_srate(eib_t *ss)
968 {
969 ib_gid_t sgid = ss->ei_props->ep_sgid;
970 ibt_srate_t srate = IBT_SRATE_10;
971 ibt_path_info_t path;
972 ibt_path_attr_t path_attr;
973 ibt_status_t ret;
974 uint8_t num_paths;
975
976 bzero(&path_attr, sizeof (path_attr));
977 path_attr.pa_dgids = &sgid;
978 path_attr.pa_num_dgids = 1;
979 path_attr.pa_sgid = sgid;
980
981 ret = ibt_get_paths(ss->ei_ibt_hdl, IBT_PATH_NO_FLAGS,
982 &path_attr, 1, &path, &num_paths);
983 if (ret == IBT_SUCCESS && num_paths >= 1) {
984 switch (srate = path.pi_prim_cep_path.cep_adds_vect.av_srate) {
985 case IBT_SRATE_2:
986 case IBT_SRATE_10:
987 case IBT_SRATE_30:
988 case IBT_SRATE_5:
989 case IBT_SRATE_20:
990 case IBT_SRATE_40:
991 case IBT_SRATE_60:
992 case IBT_SRATE_80:
993 case IBT_SRATE_120:
994 break;
995 default:
996 srate = IBT_SRATE_10;
997 }
998 }
999
1000 ss->ei_props->ep_srate = srate;
1001
1002 EIB_DPRINTF_DEBUG(ss->ei_instance, "eib_ibt_record_srate: "
1003 "srate = %d", srate);
1004 }
1005