1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright 2009 Sun Microsystems, Inc. All rights reserved. 24 * Use is subject to license terms. 25 * 26 * UDAPL kernel agent 27 */ 28 29 #include <sys/types.h> 30 #include <sys/errno.h> 31 #include <sys/debug.h> 32 #include <sys/stropts.h> 33 #include <sys/stream.h> 34 #include <sys/strlog.h> 35 #include <sys/cmn_err.h> 36 #include <sys/kmem.h> 37 #include <sys/conf.h> 38 #include <sys/stat.h> 39 #include <sys/modctl.h> 40 #include <sys/kstat.h> 41 #include <sys/ddi.h> 42 #include <sys/sunddi.h> 43 #include <sys/strsun.h> 44 #include <sys/taskq.h> 45 #include <sys/open.h> 46 #include <sys/uio.h> 47 #include <sys/cpuvar.h> 48 #include <sys/atomic.h> 49 #include <sys/sysmacros.h> 50 #include <sys/esunddi.h> 51 #include <sys/avl.h> 52 #include <sys/cred.h> 53 #include <sys/note.h> 54 #include <sys/ib/ibtl/ibti.h> 55 #include <sys/socket.h> 56 #include <netinet/in.h> 57 #include <daplt_if.h> 58 #include <daplt.h> 59 60 /* 61 * The following variables support the debug log buffer scheme. 62 */ 63 #ifdef DEBUG 64 static char daplka_dbgbuf[0x80000]; 65 #else /* DEBUG */ 66 static char daplka_dbgbuf[0x4000]; 67 #endif /* DEBUG */ 68 static int daplka_dbgsize = sizeof (daplka_dbgbuf); 69 static size_t daplka_dbgnext; 70 static int daplka_dbginit = 0; 71 static kmutex_t daplka_dbglock; 72 _NOTE(MUTEX_PROTECTS_DATA(daplka_dbglock, 73 daplka_dbgbuf 74 daplka_dbgnext)) 75 76 static int daplka_dbg = 0x0103; 77 static void daplka_console(const char *, ...); 78 static void daplka_debug(const char *, ...); 79 static int daplka_apm = 0x1; /* default enable */ 80 static int daplka_failback = 0x1; /* default enable */ 81 static int daplka_query_aft_setaltpath = 10; 82 83 #define DERR \ 84 if (daplka_dbg & 0x100) \ 85 daplka_debug 86 87 #ifdef DEBUG 88 89 #define DINFO \ 90 daplka_console 91 92 #define D1 \ 93 if (daplka_dbg & 0x01) \ 94 daplka_debug 95 #define D2 \ 96 if (daplka_dbg & 0x02) \ 97 daplka_debug 98 #define D3 \ 99 if (daplka_dbg & 0x04) \ 100 daplka_debug 101 #define D4 \ 102 if (daplka_dbg & 0x08) \ 103 daplka_debug 104 105 #else /* DEBUG */ 106 107 #define DINFO if (0) printf 108 #define D1 if (0) printf 109 #define D2 if (0) printf 110 #define D3 if (0) printf 111 #define D4 if (0) printf 112 113 #endif /* DEBUG */ 114 115 /* 116 * driver entry points 117 */ 118 static int daplka_open(dev_t *, int, int, struct cred *); 119 static int daplka_close(dev_t, int, int, struct cred *); 120 static int daplka_attach(dev_info_t *, ddi_attach_cmd_t); 121 static int daplka_detach(dev_info_t *, ddi_detach_cmd_t); 122 static int daplka_info(dev_info_t *, ddi_info_cmd_t, void *, void **); 123 static int daplka_ioctl(dev_t, int, intptr_t, int, cred_t *, int *); 124 125 /* 126 * types of ioctls 127 */ 128 static int daplka_common_ioctl(int, minor_t, intptr_t, int, cred_t *, int *); 129 static int daplka_misc_ioctl(int, daplka_ia_resource_t *, intptr_t, int, 130 cred_t *, int *); 131 static int daplka_ep_ioctl(int, daplka_ia_resource_t *, intptr_t, int, 132 cred_t *, int *); 133 static int daplka_evd_ioctl(int, daplka_ia_resource_t *, intptr_t, int, 134 cred_t *, int *); 135 static int daplka_mr_ioctl(int, daplka_ia_resource_t *, intptr_t, int, 136 cred_t *, int *); 137 static int daplka_cno_ioctl(int, daplka_ia_resource_t *, intptr_t, int, 138 cred_t *, int *); 139 static int daplka_pd_ioctl(int, daplka_ia_resource_t *, intptr_t, int, 140 cred_t *, int *); 141 static int daplka_sp_ioctl(int, daplka_ia_resource_t *, intptr_t, int, 142 cred_t *, int *); 143 static int daplka_srq_ioctl(int, daplka_ia_resource_t *, intptr_t, int, 144 cred_t *, int *); 145 146 /* 147 * common ioctls and supporting functions 148 */ 149 static int daplka_ia_create(minor_t, intptr_t, int, cred_t *, int *); 150 static int daplka_ia_destroy(daplka_resource_t *); 151 152 /* 153 * EP ioctls and supporting functions 154 */ 155 static int daplka_ep_create(daplka_ia_resource_t *, intptr_t, int, 156 cred_t *, int *); 157 static int daplka_ep_modify(daplka_ia_resource_t *, intptr_t, int, 158 cred_t *, int *); 159 static int daplka_ep_free(daplka_ia_resource_t *, intptr_t, int, 160 cred_t *, int *); 161 static int daplka_ep_connect(daplka_ia_resource_t *, intptr_t, int, 162 cred_t *, int *); 163 static int daplka_ep_disconnect(daplka_ia_resource_t *, intptr_t, int, 164 cred_t *, int *); 165 static int daplka_ep_reinit(daplka_ia_resource_t *, intptr_t, int, 166 cred_t *, int *); 167 static int daplka_ep_destroy(daplka_resource_t *); 168 static void daplka_hash_ep_free(void *); 169 static int daplka_ep_failback(void *objp, void *arg); 170 static int daplka_ep_altpath(daplka_ep_resource_t *, ib_gid_t *); 171 172 static uint32_t daplka_ep_get_state(daplka_ep_resource_t *); 173 static void daplka_ep_set_state(daplka_ep_resource_t *, uint32_t, uint32_t); 174 static boolean_t daplka_ep_transition_is_valid(uint32_t, uint32_t); 175 static daplka_timer_info_t *daplka_timer_info_alloc(daplka_ep_resource_t *); 176 static void daplka_timer_info_free(daplka_timer_info_t *); 177 static void daplka_timer_handler(void *); 178 static void daplka_timer_dispatch(void *); 179 static void daplka_timer_thread(void *); 180 static int daplka_cancel_timer(daplka_ep_resource_t *); 181 static void daplka_hash_timer_free(void *); 182 183 /* 184 * EVD ioctls and supporting functions 185 */ 186 static int daplka_evd_create(daplka_ia_resource_t *, intptr_t, int, 187 cred_t *, int *); 188 static int daplka_cq_resize(daplka_ia_resource_t *, intptr_t, int, 189 cred_t *, int *); 190 static int daplka_evd_free(daplka_ia_resource_t *, intptr_t, int, 191 cred_t *, int *); 192 static int daplka_event_poll(daplka_ia_resource_t *, intptr_t, int, 193 cred_t *, int *); 194 static int daplka_evd_destroy(daplka_resource_t *); 195 static void daplka_cq_handler(ibt_cq_hdl_t, void *); 196 static void daplka_evd_wakeup(daplka_evd_resource_t *, 197 daplka_evd_event_list_t *, daplka_evd_event_t *); 198 static void daplka_evd_event_enqueue(daplka_evd_event_list_t *, 199 daplka_evd_event_t *); 200 static daplka_evd_event_t *daplka_evd_event_dequeue(daplka_evd_event_list_t *); 201 static void daplka_hash_evd_free(void *); 202 203 204 /* 205 * SRQ ioctls and supporting functions 206 */ 207 static int daplka_srq_create(daplka_ia_resource_t *, intptr_t, int, 208 cred_t *, int *); 209 static int daplka_srq_resize(daplka_ia_resource_t *, intptr_t, int, 210 cred_t *, int *); 211 static int daplka_srq_free(daplka_ia_resource_t *, intptr_t, int, 212 cred_t *, int *); 213 static int daplka_srq_destroy(daplka_resource_t *); 214 static void daplka_hash_srq_free(void *); 215 216 /* 217 * Miscellaneous ioctls 218 */ 219 static int daplka_cr_accept(daplka_ia_resource_t *, intptr_t, int, 220 cred_t *, int *); 221 static int daplka_cr_reject(daplka_ia_resource_t *, intptr_t, int, 222 cred_t *, int *); 223 static int daplka_cr_handoff(daplka_ia_resource_t *, intptr_t, int, 224 cred_t *, int *); 225 static int daplka_ia_query(daplka_ia_resource_t *, intptr_t, int, 226 cred_t *, int *); 227 228 /* 229 * PD ioctls and supporting functions 230 */ 231 static int daplka_pd_alloc(daplka_ia_resource_t *, intptr_t, int, 232 cred_t *, int *); 233 static int daplka_pd_free(daplka_ia_resource_t *, intptr_t, int, 234 cred_t *, int *); 235 static int daplka_pd_destroy(daplka_resource_t *); 236 static void daplka_hash_pd_free(void *); 237 238 /* 239 * SP ioctls and supporting functions 240 */ 241 static int daplka_service_register(daplka_ia_resource_t *, intptr_t, int, 242 cred_t *, int *); 243 static int daplka_service_deregister(daplka_ia_resource_t *, intptr_t, int, 244 cred_t *, int *); 245 static int daplka_sp_destroy(daplka_resource_t *); 246 static void daplka_hash_sp_free(void *); 247 static void daplka_hash_sp_unref(void *); 248 249 /* 250 * MR ioctls and supporting functions 251 */ 252 static int daplka_mr_register(daplka_ia_resource_t *, intptr_t, int, 253 cred_t *, int *); 254 static int daplka_mr_register_lmr(daplka_ia_resource_t *, intptr_t, int, 255 cred_t *, int *); 256 static int daplka_mr_register_shared(daplka_ia_resource_t *, intptr_t, int, 257 cred_t *, int *); 258 static int daplka_mr_deregister(daplka_ia_resource_t *, intptr_t, int, 259 cred_t *, int *); 260 static int daplka_mr_sync(daplka_ia_resource_t *, intptr_t, int, 261 cred_t *, int *); 262 static int daplka_mr_destroy(daplka_resource_t *); 263 static void daplka_hash_mr_free(void *); 264 static void daplka_shared_mr_free(daplka_mr_resource_t *); 265 266 /* 267 * MW ioctls and supporting functions 268 */ 269 static int daplka_mw_alloc(daplka_ia_resource_t *, intptr_t, int, 270 cred_t *, int *); 271 static int daplka_mw_free(daplka_ia_resource_t *, intptr_t, int, 272 cred_t *, int *); 273 static int daplka_mw_destroy(daplka_resource_t *); 274 static void daplka_hash_mw_free(void *); 275 276 /* 277 * CNO ioctls and supporting functions 278 */ 279 static int daplka_cno_alloc(daplka_ia_resource_t *, intptr_t, int, 280 cred_t *, int *); 281 static int daplka_cno_free(daplka_ia_resource_t *, intptr_t, int, 282 cred_t *, int *); 283 static int daplka_cno_wait(daplka_ia_resource_t *, intptr_t, int, 284 cred_t *, int *); 285 static int daplka_cno_destroy(daplka_resource_t *); 286 static void daplka_hash_cno_free(void *); 287 288 /* 289 * CM handlers 290 */ 291 static ibt_cm_status_t daplka_cm_rc_handler(void *, ibt_cm_event_t *, 292 ibt_cm_return_args_t *, void *, ibt_priv_data_len_t); 293 294 static ibt_cm_status_t daplka_cm_service_handler(void *, ibt_cm_event_t *, 295 ibt_cm_return_args_t *, void *, ibt_priv_data_len_t); 296 297 static ibt_cm_status_t daplka_cm_service_req(daplka_sp_resource_t *, 298 ibt_cm_event_t *, ibt_cm_return_args_t *, void *, ibt_priv_data_len_t); 299 300 /* 301 * resource management routines 302 */ 303 static int daplka_resource_reserve(minor_t *); 304 static int daplka_resource_insert(minor_t, daplka_resource_t *); 305 static daplka_resource_t *daplka_resource_remove(minor_t rnum); 306 static daplka_resource_t *daplka_resource_lookup(minor_t); 307 static void daplka_resource_init(void); 308 static void daplka_resource_fini(void); 309 static struct daplka_resource_table daplka_resource; 310 311 /* 312 * hash table routines 313 */ 314 static int daplka_hash_insert(daplka_hash_table_t *, uint64_t *, void *); 315 static int daplka_hash_remove(daplka_hash_table_t *, uint64_t, void **); 316 static void daplka_hash_walk(daplka_hash_table_t *, int (*)(void *, void *), 317 void *, krw_t); 318 static void *daplka_hash_lookup(daplka_hash_table_t *, uint64_t); 319 static int daplka_hash_create(daplka_hash_table_t *, uint_t, 320 void (*)(void *), void (*)(void *)); 321 static void daplka_hash_destroy(daplka_hash_table_t *); 322 static uint32_t daplka_hash_getsize(daplka_hash_table_t *); 323 static void daplka_hash_generic_lookup(void *); 324 325 static uint32_t daplka_timer_hkey_gen(); 326 327 /* 328 * async event handlers 329 */ 330 static void daplka_async_event_create(ibt_async_code_t, ibt_async_event_t *, 331 uint64_t, daplka_ia_resource_t *); 332 static void daplka_rc_async_handler(void *, ibt_hca_hdl_t, ibt_async_code_t, 333 ibt_async_event_t *); 334 static void daplka_cq_async_handler(void *, ibt_hca_hdl_t, ibt_async_code_t, 335 ibt_async_event_t *); 336 static void daplka_un_async_handler(void *, ibt_hca_hdl_t, ibt_async_code_t, 337 ibt_async_event_t *); 338 static void daplka_async_handler(void *, ibt_hca_hdl_t, ibt_async_code_t, 339 ibt_async_event_t *); 340 static void daplka_sm_notice_handler(void *, ib_gid_t, ibt_subnet_event_code_t, 341 ibt_subnet_event_t *event); 342 static void daplka_sm_gid_avail(ib_gid_t *, ib_gid_t *); 343 344 /* 345 * IBTF wrappers and default limits used for resource accounting 346 */ 347 static boolean_t daplka_accounting_enabled = B_TRUE; 348 static uint32_t daplka_max_qp_percent = 100; 349 static uint32_t daplka_max_cq_percent = 100; 350 static uint32_t daplka_max_pd_percent = 100; 351 static uint32_t daplka_max_mw_percent = 100; 352 static uint32_t daplka_max_mr_percent = 100; 353 static uint32_t daplka_max_srq_percent = 100; 354 355 static ibt_status_t 356 daplka_ibt_alloc_rc_channel(daplka_ep_resource_t *, ibt_hca_hdl_t, 357 ibt_chan_alloc_flags_t, ibt_rc_chan_alloc_args_t *, 358 ibt_channel_hdl_t *, ibt_chan_sizes_t *); 359 360 static ibt_status_t 361 daplka_ibt_free_channel(daplka_ep_resource_t *, ibt_channel_hdl_t); 362 363 static ibt_status_t 364 daplka_ibt_alloc_cq(daplka_evd_resource_t *, ibt_hca_hdl_t, 365 ibt_cq_attr_t *, ibt_cq_hdl_t *, uint_t *); 366 367 static ibt_status_t 368 daplka_ibt_free_cq(daplka_evd_resource_t *, ibt_cq_hdl_t); 369 370 static ibt_status_t 371 daplka_ibt_alloc_pd(daplka_pd_resource_t *, ibt_hca_hdl_t, 372 ibt_pd_flags_t, ibt_pd_hdl_t *); 373 374 static ibt_status_t 375 daplka_ibt_free_pd(daplka_pd_resource_t *, ibt_hca_hdl_t, ibt_pd_hdl_t); 376 377 static ibt_status_t 378 daplka_ibt_alloc_mw(daplka_mw_resource_t *, ibt_hca_hdl_t, ibt_pd_hdl_t, 379 ibt_mw_flags_t, ibt_mw_hdl_t *, ibt_rkey_t *); 380 381 static ibt_status_t 382 daplka_ibt_free_mw(daplka_mw_resource_t *, ibt_hca_hdl_t, ibt_mw_hdl_t); 383 384 static ibt_status_t 385 daplka_ibt_register_mr(daplka_mr_resource_t *, ibt_hca_hdl_t, ibt_pd_hdl_t, 386 ibt_mr_attr_t *, ibt_mr_hdl_t *, ibt_mr_desc_t *); 387 388 static ibt_status_t 389 daplka_ibt_register_shared_mr(daplka_mr_resource_t *, ibt_hca_hdl_t, 390 ibt_mr_hdl_t, ibt_pd_hdl_t, ibt_smr_attr_t *, ibt_mr_hdl_t *, 391 ibt_mr_desc_t *); 392 393 static ibt_status_t 394 daplka_ibt_deregister_mr(daplka_mr_resource_t *, ibt_hca_hdl_t, ibt_mr_hdl_t); 395 396 static ibt_status_t 397 daplka_ibt_alloc_srq(daplka_srq_resource_t *, ibt_hca_hdl_t, ibt_srq_flags_t, 398 ibt_pd_hdl_t, ibt_srq_sizes_t *, ibt_srq_hdl_t *, ibt_srq_sizes_t *); 399 400 static ibt_status_t 401 daplka_ibt_free_srq(daplka_srq_resource_t *, ibt_srq_hdl_t); 402 403 /* 404 * macros for manipulating resource objects. 405 * these macros can be used on objects that begin with a 406 * daplka_resource_t header. 407 */ 408 #define DAPLKA_RS_REFCNT(rp) ((rp)->header.rs_refcnt) 409 410 #define DAPLKA_RS_REF(rp) { \ 411 mutex_enter(&(rp)->header.rs_reflock); \ 412 (rp)->header.rs_refcnt++; \ 413 ASSERT((rp)->header.rs_refcnt != 0); \ 414 mutex_exit(&(rp)->header.rs_reflock); \ 415 } 416 417 #define DAPLKA_RS_UNREF(rp) { \ 418 mutex_enter(&(rp)->header.rs_reflock); \ 419 ASSERT((rp)->header.rs_refcnt != 0); \ 420 if (--(rp)->header.rs_refcnt == 0) { \ 421 ASSERT((rp)->header.rs_free != NULL); \ 422 mutex_exit(&(rp)->header.rs_reflock); \ 423 (rp)->header.rs_free((daplka_resource_t *)rp); \ 424 } else { \ 425 mutex_exit(&(rp)->header.rs_reflock); \ 426 } \ 427 } 428 429 #define DAPLKA_RS_INIT(rp, type, rnum, free_func) { \ 430 (rp)->header.rs_refcnt = 1; \ 431 (rp)->header.rs_type = (type); \ 432 (rp)->header.rs_rnum = (rnum); \ 433 (rp)->header.rs_charged = 0; \ 434 (rp)->header.rs_free = (free_func); \ 435 mutex_init(&(rp)->header.rs_reflock, NULL, \ 436 MUTEX_DRIVER, NULL); \ 437 } 438 439 #define DAPLKA_RS_FINI(rp) { \ 440 mutex_destroy(&(rp)->header.rs_reflock); \ 441 } 442 443 #define DAPLKA_RS_ACCT_INC(rp, cnt) { \ 444 atomic_add_32(&(rp)->header.rs_charged, (cnt)); \ 445 } 446 #define DAPLKA_RS_ACCT_DEC(rp, cnt) { \ 447 atomic_add_32(&(rp)->header.rs_charged, -(cnt)); \ 448 } 449 #define DAPLKA_RS_ACCT_CHARGED(rp) ((rp)->header.rs_charged) 450 451 #define DAPLKA_RS_RNUM(rp) ((rp)->header.rs_rnum) 452 #define DAPLKA_RS_TYPE(rp) ((rp)->header.rs_type) 453 #define DAPLKA_RS_RESERVED(rp) ((intptr_t)(rp) == DAPLKA_RC_RESERVED) 454 455 /* 456 * depending on the timeout value does a cv_wait_sig or cv_timedwait_sig 457 */ 458 #define DAPLKA_EVD_WAIT(cvp, mp, timeout) \ 459 ((timeout) == LONG_MAX) ? cv_wait_sig((cvp), (mp)) : \ 460 cv_timedwait_sig((cvp), (mp), (timeout)) 461 462 #define DAPLKA_HOLD_HCA_WITHOUT_LOCK(hca) ((hca)->hca_ref_cnt++) 463 #define DAPLKA_RELE_HCA_WITHOUT_LOCK(hca) ((hca)->hca_ref_cnt--) 464 465 #define DAPLKA_HOLD_HCA(dp, hca) { \ 466 mutex_enter(&(dp)->daplka_mutex); \ 467 DAPLKA_HOLD_HCA_WITHOUT_LOCK(hca); \ 468 mutex_exit(&(dp)->daplka_mutex); \ 469 } 470 471 #define DAPLKA_RELE_HCA(dp, hca) { \ 472 mutex_enter(&(dp)->daplka_mutex); \ 473 DAPLKA_RELE_HCA_WITHOUT_LOCK(hca); \ 474 mutex_exit(&(dp)->daplka_mutex); \ 475 } 476 477 #define DAPLKA_HCA_BUSY(hca) \ 478 ((hca)->hca_ref_cnt != 0 || \ 479 (hca)->hca_qp_count != 0 || \ 480 (hca)->hca_cq_count != 0 || \ 481 (hca)->hca_pd_count != 0 || \ 482 (hca)->hca_mw_count != 0 || \ 483 (hca)->hca_mr_count != 0) 484 485 486 static struct cb_ops daplka_cb_ops = { 487 daplka_open, /* cb_open */ 488 daplka_close, /* cb_close */ 489 nodev, /* cb_strategy */ 490 nodev, /* cb_print */ 491 nodev, /* cb_dump */ 492 nodev, /* cb_read */ 493 nodev, /* cb_write */ 494 daplka_ioctl, /* cb_ioctl */ 495 nodev, /* cb_devmap */ 496 nodev, /* cb_mmap */ 497 nodev, /* cb_segmap */ 498 nochpoll, /* cb_chpoll */ 499 ddi_prop_op, /* cb_prop_op */ 500 NULL, /* cb_stream */ 501 D_NEW | D_MP, /* cb_flag */ 502 CB_REV, /* rev */ 503 nodev, /* int (*cb_aread)() */ 504 nodev /* int (*cb_awrite)() */ 505 }; 506 507 static struct dev_ops daplka_ops = { 508 DEVO_REV, /* devo_rev */ 509 0, /* devo_refcnt */ 510 daplka_info, /* devo_getinfo */ 511 nulldev, /* devo_identify */ 512 nulldev, /* devo_probe */ 513 daplka_attach, /* devo_attach */ 514 daplka_detach, /* devo_detach */ 515 nodev, /* devo_reset */ 516 &daplka_cb_ops, /* devo_cb_ops */ 517 (struct bus_ops *)NULL, /* devo_bus_ops */ 518 nulldev, /* power */ 519 ddi_quiesce_not_needed, /* devo_quiesce */ 520 }; 521 522 /* 523 * Module linkage information for the kernel. 524 */ 525 static struct modldrv modldrv = { 526 &mod_driverops, 527 "uDAPL Service Driver", 528 &daplka_ops, 529 }; 530 531 static struct modlinkage modlinkage = { 532 #ifdef _LP64 533 MODREV_1, { (void *) &modldrv, NULL, NULL, NULL, NULL, NULL, NULL } 534 #else 535 MODREV_1, { (void *) &modldrv, NULL, NULL, NULL } 536 #endif 537 }; 538 539 /* 540 * daplka_dev holds global driver state and a list of HCAs 541 */ 542 static daplka_t *daplka_dev = NULL; 543 static void *daplka_state = NULL; 544 545 /* 546 * global SP hash table 547 */ 548 static daplka_hash_table_t daplka_global_sp_htbl; 549 550 /* 551 * timer_info hash table 552 */ 553 static daplka_hash_table_t daplka_timer_info_htbl; 554 static uint32_t daplka_timer_hkey = 0; 555 556 /* 557 * shared MR avl tree 558 */ 559 static avl_tree_t daplka_shared_mr_tree; 560 static kmutex_t daplka_shared_mr_lock; 561 static int daplka_shared_mr_cmp(const void *, const void *); 562 _NOTE(MUTEX_PROTECTS_DATA(daplka_shared_mr_lock, 563 daplka_shared_mr_tree)) 564 565 /* 566 * default kmem flags used by this driver 567 */ 568 static int daplka_km_flags = KM_SLEEP; 569 570 /* 571 * taskq used for handling background tasks 572 */ 573 static taskq_t *daplka_taskq = NULL; 574 575 /* 576 * daplka_cm_delay is the length of time the active 577 * side needs to wait before timing out on the REP message. 578 */ 579 static clock_t daplka_cm_delay = 60000000; 580 581 /* 582 * modunload will fail if pending_close is non-zero 583 */ 584 static uint32_t daplka_pending_close = 0; 585 586 static struct ibt_clnt_modinfo_s daplka_clnt_modinfo = { 587 IBTI_V_CURR, 588 IBT_USER, 589 daplka_async_handler, 590 NULL, 591 DAPLKA_DRV_NAME 592 }; 593 594 /* 595 * Module Installation 596 */ 597 int 598 _init(void) 599 { 600 int status; 601 602 status = ddi_soft_state_init(&daplka_state, sizeof (daplka_t), 1); 603 if (status != 0) { 604 return (status); 605 } 606 607 mutex_init(&daplka_dbglock, NULL, MUTEX_DRIVER, NULL); 608 bzero(daplka_dbgbuf, sizeof (daplka_dbgbuf)); 609 daplka_dbgnext = 0; 610 daplka_dbginit = 1; 611 612 daplka_resource_init(); 613 614 status = mod_install(&modlinkage); 615 if (status != DDI_SUCCESS) { 616 /* undo inits done before mod_install */ 617 daplka_resource_fini(); 618 mutex_destroy(&daplka_dbglock); 619 ddi_soft_state_fini(&daplka_state); 620 } 621 return (status); 622 } 623 624 /* 625 * Module Removal 626 */ 627 int 628 _fini(void) 629 { 630 int status; 631 632 /* 633 * mod_remove causes detach to be called 634 */ 635 if ((status = mod_remove(&modlinkage)) != 0) { 636 DERR("fini: mod_remove failed: 0x%x\n", status); 637 return (status); 638 } 639 640 daplka_resource_fini(); 641 mutex_destroy(&daplka_dbglock); 642 ddi_soft_state_fini(&daplka_state); 643 644 return (status); 645 } 646 647 /* 648 * Return Module Info. 649 */ 650 int 651 _info(struct modinfo *modinfop) 652 { 653 return (mod_info(&modlinkage, modinfop)); 654 } 655 656 static void 657 daplka_enqueue_hca(daplka_t *dp, daplka_hca_t *hca) 658 { 659 daplka_hca_t *h; 660 661 ASSERT(mutex_owned(&dp->daplka_mutex)); 662 663 if (dp->daplka_hca_list_head == NULL) { 664 dp->daplka_hca_list_head = hca; 665 } else { 666 h = dp->daplka_hca_list_head; 667 while (h->hca_next != NULL) 668 h = h->hca_next; 669 670 h->hca_next = hca; 671 } 672 } 673 674 static void 675 daplka_dequeue_hca(daplka_t *dp, daplka_hca_t *hca) 676 { 677 daplka_hca_t *h; 678 679 ASSERT(mutex_owned(&dp->daplka_mutex)); 680 681 if (dp->daplka_hca_list_head == hca) 682 dp->daplka_hca_list_head = hca->hca_next; 683 else { 684 h = dp->daplka_hca_list_head; 685 while (h->hca_next != hca) 686 h = h->hca_next; 687 h->hca_next = hca->hca_next; 688 } 689 } 690 691 static int 692 daplka_init_hca(daplka_t *dp, ib_guid_t hca_guid) 693 { 694 daplka_hca_t *hca; 695 ibt_hca_portinfo_t *pinfop; 696 uint_t size; 697 int j; 698 ibt_status_t status; 699 700 hca = kmem_zalloc(sizeof (daplka_hca_t), KM_SLEEP); 701 702 hca->hca_guid = hca_guid; 703 704 /* 705 * open the HCA for use 706 */ 707 status = ibt_open_hca(dp->daplka_clnt_hdl, hca_guid, &hca->hca_hdl); 708 if (status != IBT_SUCCESS) { 709 if (status == IBT_HCA_IN_USE) { 710 DERR("ibt_open_hca() returned IBT_HCA_IN_USE\n"); 711 } else { 712 DERR("ibt_open_hca() returned %d\n", status); 713 } 714 kmem_free(hca, sizeof (daplka_hca_t)); 715 return (status); 716 } 717 718 /* 719 * query HCA to get its info 720 */ 721 status = ibt_query_hca(hca->hca_hdl, &hca->hca_attr); 722 if (status != IBT_SUCCESS) { 723 DERR("ibt_query_hca returned %d (hca_guid 0x%llx)\n", 724 status, (longlong_t)hca_guid); 725 goto out; 726 } 727 728 /* 729 * query HCA to get info of all ports 730 */ 731 status = ibt_query_hca_ports(hca->hca_hdl, 732 0, &pinfop, &hca->hca_nports, &size); 733 if (status != IBT_SUCCESS) { 734 DERR("ibt_query_all_ports returned %d " 735 "(hca_guid 0x%llx)\n", status, 736 (longlong_t)hca_guid); 737 goto out; 738 } 739 hca->hca_ports = pinfop; 740 hca->hca_pinfosz = size; 741 742 DERR("hca guid 0x%llx, nports %d\n", 743 (longlong_t)hca_guid, hca->hca_nports); 744 for (j = 0; j < hca->hca_nports; j++) { 745 DERR("port %d: state %d prefix 0x%016llx " 746 "guid %016llx\n", 747 pinfop[j].p_port_num, pinfop[j].p_linkstate, 748 (longlong_t)pinfop[j].p_sgid_tbl[0].gid_prefix, 749 (longlong_t)pinfop[j].p_sgid_tbl[0].gid_guid); 750 } 751 752 mutex_enter(&dp->daplka_mutex); 753 daplka_enqueue_hca(dp, hca); 754 mutex_exit(&dp->daplka_mutex); 755 756 return (IBT_SUCCESS); 757 758 out: 759 (void) ibt_close_hca(hca->hca_hdl); 760 kmem_free(hca, sizeof (daplka_hca_t)); 761 return (status); 762 } 763 764 /* 765 * this function obtains the list of HCAs from IBTF. 766 * the HCAs are then opened and the returned handles 767 * and attributes are stored into the global daplka_dev 768 * structure. 769 */ 770 static int 771 daplka_init_hcas(daplka_t *dp) 772 { 773 int i; 774 ib_guid_t *hca_guids; 775 uint32_t hca_count; 776 777 /* 778 * get the num & list of HCAs present 779 */ 780 hca_count = ibt_get_hca_list(&hca_guids); 781 DERR("No. of HCAs present %d\n", hca_count); 782 783 if (hca_count != 0) { 784 /* 785 * get the info for each available HCA 786 */ 787 for (i = 0; i < hca_count; i++) 788 (void) daplka_init_hca(dp, hca_guids[i]); 789 790 ibt_free_hca_list(hca_guids, hca_count); 791 } 792 793 if (dp->daplka_hca_list_head != NULL) 794 return (IBT_SUCCESS); 795 else 796 return (IBT_FAILURE); 797 } 798 799 static int 800 daplka_fini_hca(daplka_t *dp, daplka_hca_t *hca) 801 { 802 ibt_status_t status; 803 804 if (hca->hca_hdl != NULL) { 805 status = ibt_close_hca(hca->hca_hdl); 806 if (status != IBT_SUCCESS) { 807 DERR("ibt_close_hca returned %d" 808 " (hca_guid 0x%llx)\n", status, 809 (longlong_t)hca->hca_guid); 810 811 mutex_enter(&dp->daplka_mutex); 812 daplka_enqueue_hca(dp, hca); 813 mutex_exit(&dp->daplka_mutex); 814 815 return (status); 816 } 817 } 818 819 if (hca->hca_ports != NULL) 820 ibt_free_portinfo(hca->hca_ports, hca->hca_pinfosz); 821 822 kmem_free(hca, sizeof (daplka_hca_t)); 823 return (IBT_SUCCESS); 824 } 825 826 /* 827 * closes all HCAs and frees up the HCA list 828 */ 829 static int 830 daplka_fini_hcas(daplka_t *dp) 831 { 832 ibt_status_t status; 833 daplka_hca_t *hca; 834 835 mutex_enter(&daplka_dev->daplka_mutex); 836 while ((hca = dp->daplka_hca_list_head) != NULL) { 837 if (DAPLKA_HCA_BUSY(hca)) { 838 mutex_exit(&daplka_dev->daplka_mutex); 839 return (IBT_HCA_RESOURCES_NOT_FREED); 840 } 841 daplka_dequeue_hca(daplka_dev, hca); 842 mutex_exit(&daplka_dev->daplka_mutex); 843 844 if ((status = daplka_fini_hca(dp, hca)) != IBT_SUCCESS) 845 return (status); 846 847 mutex_enter(&daplka_dev->daplka_mutex); 848 } 849 mutex_exit(&daplka_dev->daplka_mutex); 850 851 DERR("dapl kernel agent unloaded\n"); 852 return (IBT_SUCCESS); 853 } 854 855 856 /* 857 * Attach the device, create and fill in daplka_dev 858 */ 859 static int 860 daplka_attach(dev_info_t *dip, ddi_attach_cmd_t cmd) 861 { 862 daplka_t *dp; 863 int instance, retval, err; 864 boolean_t sp_htbl_allocated = B_FALSE; 865 boolean_t timer_htbl_allocated = B_FALSE; 866 boolean_t shared_mr_tree_allocated = B_FALSE; 867 868 switch (cmd) { 869 case DDI_ATTACH: 870 break; 871 case DDI_RESUME: 872 return (DDI_SUCCESS); 873 default: 874 return (DDI_FAILURE); 875 } 876 877 /* 878 * Allocate soft data structure 879 */ 880 instance = ddi_get_instance(dip); 881 if (ddi_soft_state_zalloc(daplka_state, instance) != DDI_SUCCESS) { 882 DERR("attach: bad state zalloc\n"); 883 return (DDI_FAILURE); 884 } 885 886 dp = ddi_get_soft_state(daplka_state, instance); 887 if (dp == NULL) { 888 ddi_soft_state_free(daplka_state, instance); 889 DERR("attach: cannot get soft state\n"); 890 return (DDI_FAILURE); 891 } 892 /* 893 * Stuff private info into dip. 894 */ 895 dp->daplka_dip = dip; 896 ddi_set_driver_private(dip, dp); 897 daplka_dev = dp; 898 mutex_init(&dp->daplka_mutex, NULL, MUTEX_DRIVER, NULL); 899 900 /* 901 * Register driver with IBTF 902 */ 903 retval = ibt_attach(&daplka_clnt_modinfo, dip, dp, 904 &dp->daplka_clnt_hdl); 905 if (retval != IBT_SUCCESS) { 906 DERR("attach: ibt_attach failed: error = %d\n", retval); 907 retval = DDI_FAILURE; 908 goto error; 909 } 910 /* Register to receive SM events */ 911 ibt_register_subnet_notices(dp->daplka_clnt_hdl, 912 daplka_sm_notice_handler, NULL); 913 914 retval = daplka_init_hcas(dp); 915 if (retval != IBT_SUCCESS) { 916 DERR("attach: hca_init failed: error = %d\n", retval); 917 retval = DDI_FAILURE; 918 goto error; 919 } 920 /* 921 * this table is used by cr_handoff 922 */ 923 retval = daplka_hash_create(&daplka_global_sp_htbl, 924 DAPLKA_G_SP_HTBL_SZ, daplka_hash_sp_unref, 925 daplka_hash_generic_lookup); 926 if (retval != 0) { 927 DERR("attach: cannot create sp hash table\n"); 928 retval = DDI_FAILURE; 929 goto error; 930 } 931 sp_htbl_allocated = B_TRUE; 932 933 /* 934 * this table stores per EP timer information. 935 * timer_info_t objects are inserted into this table whenever 936 * a EP timer is set. timers get removed when they expire 937 * or when they get cancelled. 938 */ 939 retval = daplka_hash_create(&daplka_timer_info_htbl, 940 DAPLKA_TIMER_HTBL_SZ, daplka_hash_timer_free, NULL); 941 if (retval != 0) { 942 DERR("attach: cannot create timer hash table\n"); 943 retval = DDI_FAILURE; 944 goto error; 945 } 946 timer_htbl_allocated = B_TRUE; 947 948 /* 949 * this taskq is currently only used for processing timers. 950 * other processing may also use this taskq in the future. 951 */ 952 daplka_taskq = taskq_create(DAPLKA_DRV_NAME, DAPLKA_TQ_NTHREADS, 953 maxclsyspri, 1, DAPLKA_TQ_NTHREADS, TASKQ_DYNAMIC); 954 if (daplka_taskq == NULL) { 955 DERR("attach: cannot create daplka_taskq\n"); 956 retval = DDI_FAILURE; 957 goto error; 958 } 959 960 /* 961 * daplka_shared_mr_tree holds daplka_shared_mr_t objects that 962 * gets retrieved or created when daplka_mr_register_shared is 963 * called. 964 */ 965 mutex_init(&daplka_shared_mr_lock, NULL, MUTEX_DRIVER, NULL); 966 967 avl_create(&daplka_shared_mr_tree, daplka_shared_mr_cmp, 968 sizeof (daplka_shared_mr_t), 969 offsetof(daplka_shared_mr_t, smr_node)); 970 shared_mr_tree_allocated = B_TRUE; 971 972 /* 973 * Create the filesystem device node. 974 */ 975 if (ddi_create_minor_node(dip, DAPLKA_MINOR_NAME, S_IFCHR, 976 0, DDI_PSEUDO, NULL) != DDI_SUCCESS) { 977 DERR("attach: bad create_minor_node\n"); 978 retval = DDI_FAILURE; 979 goto error; 980 } 981 dp->daplka_status = DAPLKA_STATE_ATTACHED; 982 ddi_report_dev(dip); 983 return (DDI_SUCCESS); 984 985 error: 986 if (shared_mr_tree_allocated) { 987 avl_destroy(&daplka_shared_mr_tree); 988 mutex_destroy(&daplka_shared_mr_lock); 989 } 990 991 if (daplka_taskq) { 992 taskq_destroy(daplka_taskq); 993 daplka_taskq = NULL; 994 } 995 996 if (timer_htbl_allocated) { 997 daplka_hash_destroy(&daplka_timer_info_htbl); 998 } 999 1000 if (sp_htbl_allocated) { 1001 daplka_hash_destroy(&daplka_global_sp_htbl); 1002 } 1003 1004 err = daplka_fini_hcas(dp); 1005 if (err != IBT_SUCCESS) { 1006 DERR("attach: hca_fini returned %d\n", err); 1007 } 1008 1009 if (dp->daplka_clnt_hdl != NULL) { 1010 /* unregister SM event notification */ 1011 ibt_register_subnet_notices(dp->daplka_clnt_hdl, 1012 (ibt_sm_notice_handler_t)NULL, NULL); 1013 err = ibt_detach(dp->daplka_clnt_hdl); 1014 1015 if (err != IBT_SUCCESS) { 1016 DERR("attach: ibt_detach returned %d\n", err); 1017 } 1018 } 1019 mutex_destroy(&dp->daplka_mutex); 1020 1021 if (dp->daplka_status == DAPLKA_STATE_ATTACHED) { 1022 ddi_remove_minor_node(dip, NULL); 1023 } 1024 ddi_soft_state_free(daplka_state, instance); 1025 return (retval); 1026 } 1027 1028 /* 1029 * Detach - Free resources allocated in attach 1030 */ 1031 /* ARGSUSED */ 1032 static int 1033 daplka_detach(dev_info_t *dip, ddi_detach_cmd_t cmd) 1034 { 1035 int instance, err; 1036 void *cookie = NULL; 1037 daplka_t *dp; 1038 1039 if (cmd != DDI_DETACH) { 1040 return (DDI_FAILURE); 1041 } 1042 if (daplka_resource.daplka_rc_cnt > 0 || 1043 daplka_pending_close > 0) { 1044 DERR("detach: driver in use\n"); 1045 return (DDI_FAILURE); 1046 } 1047 1048 instance = ddi_get_instance(dip); 1049 dp = ddi_get_soft_state(daplka_state, instance); 1050 if (dp == NULL) { 1051 DERR("detach: cannot get soft state\n"); 1052 return (DDI_FAILURE); 1053 } 1054 err = daplka_fini_hcas(dp); 1055 if (err != IBT_SUCCESS) { 1056 DERR("detach: hca_fini returned %d\n", err); 1057 return (DDI_FAILURE); 1058 } 1059 if (dp->daplka_clnt_hdl != NULL) { 1060 /* unregister SM event notification */ 1061 ibt_register_subnet_notices(dp->daplka_clnt_hdl, 1062 (ibt_sm_notice_handler_t)NULL, NULL); 1063 err = ibt_detach(dp->daplka_clnt_hdl); 1064 if (err != IBT_SUCCESS) { 1065 DERR("detach: ibt_detach returned %d\n", err); 1066 return (DDI_FAILURE); 1067 } 1068 dp->daplka_clnt_hdl = NULL; 1069 } 1070 mutex_destroy(&dp->daplka_mutex); 1071 if (dp->daplka_status == DAPLKA_STATE_ATTACHED) { 1072 ddi_remove_minor_node(dip, NULL); 1073 } 1074 dp->daplka_status = DAPLKA_STATE_DETACHED; 1075 ddi_soft_state_free(daplka_state, instance); 1076 daplka_dev = NULL; 1077 1078 /* 1079 * by the time we get here, all clients of dapl should 1080 * have exited and completed their cleanup properly. 1081 * we can assert that all global data structures are now 1082 * empty. 1083 */ 1084 ASSERT(avl_destroy_nodes(&daplka_shared_mr_tree, &cookie) == NULL); 1085 avl_destroy(&daplka_shared_mr_tree); 1086 mutex_destroy(&daplka_shared_mr_lock); 1087 1088 ASSERT(daplka_hash_getsize(&daplka_timer_info_htbl) == 0); 1089 daplka_hash_destroy(&daplka_timer_info_htbl); 1090 1091 ASSERT(daplka_hash_getsize(&daplka_global_sp_htbl) == 0); 1092 daplka_hash_destroy(&daplka_global_sp_htbl); 1093 1094 taskq_destroy(daplka_taskq); 1095 1096 return (DDI_SUCCESS); 1097 } 1098 1099 /* ARGSUSED */ 1100 static int 1101 daplka_info(dev_info_t *dip, ddi_info_cmd_t infocmd, void *arg, void **result) 1102 { 1103 switch (infocmd) { 1104 case DDI_INFO_DEVT2DEVINFO: 1105 if (daplka_dev != NULL) { 1106 *result = daplka_dev->daplka_dip; 1107 return (DDI_SUCCESS); 1108 } else { 1109 return (DDI_FAILURE); 1110 } 1111 1112 case DDI_INFO_DEVT2INSTANCE: 1113 *result = 0; 1114 return (DDI_SUCCESS); 1115 1116 default: 1117 return (DDI_FAILURE); 1118 } 1119 } 1120 1121 /* 1122 * creates a EP resource. 1123 * A EP resource contains a RC channel. A EP resource holds a 1124 * reference to a send_evd (for the send CQ), recv_evd (for the 1125 * recv CQ), a connection evd and a PD. These references ensure 1126 * that the referenced resources are not freed until the EP itself 1127 * gets freed. 1128 */ 1129 /* ARGSUSED */ 1130 static int 1131 daplka_ep_create(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 1132 cred_t *cred, int *rvalp) 1133 { 1134 daplka_ep_resource_t *ep_rp; 1135 daplka_pd_resource_t *pd_rp; 1136 dapl_ep_create_t args; 1137 ibt_rc_chan_alloc_args_t chan_args; 1138 ibt_chan_alloc_flags_t achan_flags; 1139 ibt_chan_sizes_t chan_real_sizes; 1140 ibt_hca_attr_t *hca_attrp; 1141 uint64_t ep_hkey = 0; 1142 boolean_t inserted = B_FALSE; 1143 uint32_t old_state, new_state; 1144 int retval; 1145 ibt_status_t status; 1146 1147 D3("ep_create: enter\n"); 1148 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_ep_create_t), 1149 mode); 1150 if (retval != 0) { 1151 DERR("ep_create: copyin error %d\n", retval); 1152 return (EFAULT); 1153 } 1154 ep_rp = kmem_zalloc(sizeof (daplka_ep_resource_t), daplka_km_flags); 1155 if (ep_rp == NULL) { 1156 DERR("ep_create: cannot allocate ep_rp\n"); 1157 return (ENOMEM); 1158 } 1159 DAPLKA_RS_INIT(ep_rp, DAPL_TYPE_EP, 1160 DAPLKA_RS_RNUM(ia_rp), daplka_ep_destroy); 1161 1162 mutex_init(&ep_rp->ep_lock, NULL, MUTEX_DRIVER, NULL); 1163 cv_init(&ep_rp->ep_cv, NULL, CV_DRIVER, NULL); 1164 ep_rp->ep_hca = ia_rp->ia_hca; 1165 ep_rp->ep_cookie = args.ep_cookie; 1166 ep_rp->ep_timer_hkey = 0; 1167 1168 /* 1169 * we don't have to use ep_get_state here because ep_rp is not in 1170 * ep_htbl yet. refer to the description of daplka_ep_set_state 1171 * for details about the EP state machine. 1172 */ 1173 ep_rp->ep_state = DAPLKA_EP_STATE_TRANSITIONING; 1174 new_state = old_state = DAPLKA_EP_STATE_CLOSED; 1175 1176 /* get reference to send evd and get cq handle */ 1177 ep_rp->ep_snd_evd = (daplka_evd_resource_t *) 1178 daplka_hash_lookup(&ia_rp->ia_evd_htbl, args.ep_snd_evd_hkey); 1179 if (ep_rp->ep_snd_evd == NULL) { 1180 DERR("ep_create: ep_snd_evd %llx not found\n", 1181 args.ep_snd_evd_hkey); 1182 retval = EINVAL; 1183 goto cleanup; 1184 } 1185 chan_args.rc_scq = ep_rp->ep_snd_evd->evd_cq_hdl; 1186 if (chan_args.rc_scq == NULL) { 1187 DERR("ep_create: ep_snd_evd cq invalid\n"); 1188 retval = EINVAL; 1189 goto cleanup; 1190 } 1191 1192 /* get reference to recv evd and get cq handle */ 1193 ep_rp->ep_rcv_evd = (daplka_evd_resource_t *) 1194 daplka_hash_lookup(&ia_rp->ia_evd_htbl, args.ep_rcv_evd_hkey); 1195 if (ep_rp->ep_rcv_evd == NULL) { 1196 DERR("ep_create: ep_rcv_evd %llx not found\n", 1197 args.ep_rcv_evd_hkey); 1198 retval = EINVAL; 1199 goto cleanup; 1200 } 1201 chan_args.rc_rcq = ep_rp->ep_rcv_evd->evd_cq_hdl; 1202 if (chan_args.rc_rcq == NULL) { 1203 DERR("ep_create: ep_rcv_evd cq invalid\n"); 1204 retval = EINVAL; 1205 goto cleanup; 1206 } 1207 1208 /* get reference to conn evd */ 1209 ep_rp->ep_conn_evd = (daplka_evd_resource_t *) 1210 daplka_hash_lookup(&ia_rp->ia_evd_htbl, args.ep_conn_evd_hkey); 1211 if (ep_rp->ep_conn_evd == NULL) { 1212 DERR("ep_create: ep_conn_evd %llx not found\n", 1213 args.ep_conn_evd_hkey); 1214 retval = EINVAL; 1215 goto cleanup; 1216 } 1217 1218 /* get reference to SRQ if needed */ 1219 if (args.ep_srq_attached) { 1220 ep_rp->ep_srq_res = (daplka_srq_resource_t *)daplka_hash_lookup( 1221 &ia_rp->ia_srq_htbl, args.ep_srq_hkey); 1222 if (ep_rp->ep_srq_res == NULL) { 1223 DERR("ep_create: ep_srq %llx not found\n", 1224 (longlong_t)args.ep_srq_hkey); 1225 retval = EINVAL; 1226 goto cleanup; 1227 } 1228 ASSERT(DAPLKA_RS_TYPE(ep_rp->ep_srq_res) == DAPL_TYPE_SRQ); 1229 D3("ep_create: ep_srq %p %llx\n", ep_rp->ep_srq_res, 1230 (longlong_t)args.ep_srq_hkey); 1231 } else { 1232 ep_rp->ep_srq_res = NULL; 1233 } 1234 1235 /* get pd handle */ 1236 pd_rp = (daplka_pd_resource_t *) 1237 daplka_hash_lookup(&ia_rp->ia_pd_htbl, args.ep_pd_hkey); 1238 if (pd_rp == NULL) { 1239 DERR("ep_create: cannot find pd resource\n"); 1240 retval = EINVAL; 1241 goto cleanup; 1242 } 1243 ASSERT(DAPLKA_RS_TYPE(pd_rp) == DAPL_TYPE_PD); 1244 ep_rp->ep_pd_res = pd_rp; 1245 chan_args.rc_pd = pd_rp->pd_hdl; 1246 1247 1248 /* 1249 * these checks ensure that the requested channel sizes 1250 * are within the limits supported by the chosen HCA. 1251 */ 1252 hca_attrp = &ia_rp->ia_hca->hca_attr; 1253 if (args.ep_ch_sizes.dcs_sq_sgl > hca_attrp->hca_max_sgl) { 1254 DERR("ep_create: invalid cs_sq_sgl %d\n", 1255 args.ep_ch_sizes.dcs_sq_sgl); 1256 retval = EINVAL; 1257 goto cleanup; 1258 } 1259 if (args.ep_ch_sizes.dcs_rq_sgl > hca_attrp->hca_max_sgl) { 1260 DERR("ep_create: invalid cs_rq_sgl %d\n", 1261 args.ep_ch_sizes.dcs_rq_sgl); 1262 retval = EINVAL; 1263 goto cleanup; 1264 } 1265 if (args.ep_ch_sizes.dcs_sq > hca_attrp->hca_max_chan_sz) { 1266 DERR("ep_create: invalid cs_sq %d\n", 1267 args.ep_ch_sizes.dcs_sq); 1268 retval = EINVAL; 1269 goto cleanup; 1270 } 1271 if (args.ep_ch_sizes.dcs_rq > hca_attrp->hca_max_chan_sz) { 1272 DERR("ep_create: invalid cs_rq %d\n", 1273 args.ep_ch_sizes.dcs_rq); 1274 retval = EINVAL; 1275 goto cleanup; 1276 } 1277 1278 chan_args.rc_sizes.cs_sq_sgl = args.ep_ch_sizes.dcs_sq_sgl; 1279 chan_args.rc_sizes.cs_rq_sgl = args.ep_ch_sizes.dcs_rq_sgl; 1280 chan_args.rc_sizes.cs_sq = args.ep_ch_sizes.dcs_sq; 1281 chan_args.rc_sizes.cs_rq = args.ep_ch_sizes.dcs_rq; 1282 chan_args.rc_flags = IBT_WR_SIGNALED; 1283 chan_args.rc_control = IBT_CEP_RDMA_RD | IBT_CEP_RDMA_WR; 1284 chan_args.rc_hca_port_num = ia_rp->ia_port_num; 1285 chan_args.rc_clone_chan = NULL; 1286 if (args.ep_srq_attached) { 1287 chan_args.rc_srq = ep_rp->ep_srq_res->srq_hdl; 1288 } else { 1289 chan_args.rc_srq = NULL; 1290 } 1291 1292 D3("ep_create: sq_sgl %d, rq_sgl %d, sq %d, rq %d, " 1293 "sig_type 0x%x, control 0x%x, portnum %d, clone_chan 0x%p\n", 1294 args.ep_ch_sizes.dcs_sq_sgl, args.ep_ch_sizes.dcs_rq_sgl, 1295 args.ep_ch_sizes.dcs_sq, args.ep_ch_sizes.dcs_rq, 1296 chan_args.rc_flags, chan_args.rc_control, 1297 chan_args.rc_hca_port_num, chan_args.rc_clone_chan); 1298 1299 if (args.ep_srq_attached) { 1300 achan_flags = IBT_ACHAN_USER_MAP | IBT_ACHAN_USES_SRQ; 1301 } else { 1302 achan_flags = IBT_ACHAN_USER_MAP; 1303 } 1304 /* create rc channel */ 1305 status = daplka_ibt_alloc_rc_channel(ep_rp, ia_rp->ia_hca_hdl, 1306 achan_flags, &chan_args, &ep_rp->ep_chan_hdl, 1307 &chan_real_sizes); 1308 if (status != IBT_SUCCESS) { 1309 DERR("ep_create: alloc_rc_channel returned %d\n", status); 1310 *rvalp = (int)status; 1311 retval = 0; 1312 goto cleanup; 1313 } 1314 1315 args.ep_ch_real_sizes.dcs_sq = chan_real_sizes.cs_sq; 1316 args.ep_ch_real_sizes.dcs_rq = chan_real_sizes.cs_rq; 1317 args.ep_ch_real_sizes.dcs_sq_sgl = chan_real_sizes.cs_sq_sgl; 1318 args.ep_ch_real_sizes.dcs_rq_sgl = chan_real_sizes.cs_rq_sgl; 1319 1320 /* 1321 * store ep ptr with chan_hdl. 1322 * this ep_ptr is used by the CM handlers (both active and 1323 * passive) 1324 * mutex is only needed for race of "destroy" and "async" 1325 */ 1326 mutex_enter(&daplka_dev->daplka_mutex); 1327 ibt_set_chan_private(ep_rp->ep_chan_hdl, (void *)ep_rp); 1328 mutex_exit(&daplka_dev->daplka_mutex); 1329 1330 /* Get HCA-specific data_out info */ 1331 status = ibt_ci_data_out(ia_rp->ia_hca_hdl, 1332 IBT_CI_NO_FLAGS, IBT_HDL_CHANNEL, (void *)ep_rp->ep_chan_hdl, 1333 &args.ep_qp_data_out, sizeof (args.ep_qp_data_out)); 1334 1335 if (status != IBT_SUCCESS) { 1336 DERR("ep_create: ibt_ci_data_out error(%d)\n", 1337 status); 1338 *rvalp = (int)status; 1339 retval = 0; 1340 goto cleanup; 1341 } 1342 1343 /* insert into ep hash table */ 1344 retval = daplka_hash_insert(&ia_rp->ia_ep_htbl, 1345 &ep_hkey, (void *)ep_rp); 1346 if (retval != 0) { 1347 DERR("ep_create: cannot insert ep resource into ep_htbl\n"); 1348 goto cleanup; 1349 } 1350 inserted = B_TRUE; 1351 1352 /* 1353 * at this point, the ep_rp can be looked up by other threads 1354 * if they manage to guess the correct hkey. but they are not 1355 * permitted to operate on ep_rp until we transition to the 1356 * CLOSED state. 1357 */ 1358 1359 /* return hkey to library */ 1360 args.ep_hkey = ep_hkey; 1361 1362 retval = ddi_copyout(&args, (void *)arg, sizeof (dapl_ep_create_t), 1363 mode); 1364 if (retval != 0) { 1365 DERR("ep_create: copyout error %d\n", retval); 1366 retval = EFAULT; 1367 goto cleanup; 1368 } 1369 1370 daplka_ep_set_state(ep_rp, old_state, new_state); 1371 D3("ep_create: exit\n"); 1372 return (0); 1373 1374 cleanup: 1375 if (inserted) { 1376 daplka_ep_resource_t *free_rp = NULL; 1377 1378 (void) daplka_hash_remove(&ia_rp->ia_ep_htbl, ep_hkey, 1379 (void **)&free_rp); 1380 if (free_rp != ep_rp) { 1381 /* 1382 * this case is impossible because ep_free will 1383 * wait until our state transition is complete. 1384 */ 1385 DERR("ep_create: cannot remove ep from hash table\n"); 1386 ASSERT(B_FALSE); 1387 return (retval); 1388 } 1389 } 1390 new_state = DAPLKA_EP_STATE_FREED; 1391 daplka_ep_set_state(ep_rp, old_state, new_state); 1392 DAPLKA_RS_UNREF(ep_rp); 1393 return (retval); 1394 } 1395 1396 /* 1397 * daplka_ep_get_state retrieves the current state of the EP and 1398 * sets the state to TRANSITIONING. if the current state is already 1399 * TRANSITIONING, this function will wait until the state becomes one 1400 * of the other EP states. Most of the EP related ioctls follow the 1401 * call sequence: 1402 * 1403 * new_state = old_state = daplka_ep_get_state(ep_rp); 1404 * ... 1405 * ...some code that affects the EP 1406 * ... 1407 * new_state = <NEW_STATE>; 1408 * daplka_ep_set_state(ep_rp, old_state, new_state); 1409 * 1410 * this call sequence ensures that only one thread may access the EP 1411 * during the time ep_state is in TRANSITIONING. daplka_ep_set_state 1412 * transitions ep_state to new_state and wakes up any waiters blocking 1413 * on ep_cv. 1414 * 1415 */ 1416 static uint32_t 1417 daplka_ep_get_state(daplka_ep_resource_t *ep_rp) 1418 { 1419 uint32_t old_state = 0; 1420 1421 mutex_enter(&ep_rp->ep_lock); 1422 while (ep_rp->ep_state == DAPLKA_EP_STATE_TRANSITIONING) { 1423 D2("get_state: wait for state transition to complete\n"); 1424 cv_wait(&ep_rp->ep_cv, &ep_rp->ep_lock); 1425 D2("get_state: done, curr state = %d\n", ep_rp->ep_state); 1426 } 1427 ASSERT(ep_rp->ep_state != DAPLKA_EP_STATE_TRANSITIONING); 1428 old_state = ep_rp->ep_state; 1429 1430 /* 1431 * an ep that is in the FREED state cannot transition 1432 * back to any of the regular states 1433 */ 1434 if (old_state != DAPLKA_EP_STATE_FREED) { 1435 ep_rp->ep_state = DAPLKA_EP_STATE_TRANSITIONING; 1436 } 1437 mutex_exit(&ep_rp->ep_lock); 1438 return (old_state); 1439 } 1440 1441 /* 1442 * EP state transition diagram 1443 * 1444 * CLOSED<------------------- 1445 * | | 1446 * | | 1447 * ------------------------ | 1448 * | | | 1449 * | | | 1450 * v v | 1451 * CONNECTING ACCEPTING | 1452 * | | | | | | 1453 * | | | | | | 1454 * | | | | | | 1455 * | | |_______|_______| | 1456 * | | | | | | 1457 * | |___________| | | | 1458 * | | | | | 1459 * | v | |---->DISCONNECTED 1460 * | CONNECTED | ^ 1461 * v | | | 1462 * ABORTING |---------|--------------| 1463 * | | | | 1464 * | | v | 1465 * | |-------->DISCONNECTING--| 1466 * | | 1467 * |---------------------------------| 1468 * 1469 * *not shown in this diagram: 1470 * -loopback transitions 1471 * -transitions to the FREED state 1472 */ 1473 static boolean_t 1474 daplka_ep_transition_is_valid(uint32_t old_state, uint32_t new_state) 1475 { 1476 boolean_t valid = B_FALSE; 1477 1478 /* 1479 * reseting to the same state is a no-op and is always 1480 * permitted. transitioning to the FREED state indicates 1481 * that the ep is about to be freed and no further operation 1482 * is allowed on it. to support abrupt close, the ep is 1483 * permitted to transition to the FREED state from any state. 1484 */ 1485 if (old_state == new_state || 1486 new_state == DAPLKA_EP_STATE_FREED) { 1487 return (B_TRUE); 1488 } 1489 1490 switch (old_state) { 1491 case DAPLKA_EP_STATE_CLOSED: 1492 /* 1493 * this is the initial ep_state. 1494 * a transition to CONNECTING or ACCEPTING may occur 1495 * upon calling daplka_ep_connect or daplka_cr_accept, 1496 * respectively. 1497 */ 1498 if (new_state == DAPLKA_EP_STATE_CONNECTING || 1499 new_state == DAPLKA_EP_STATE_ACCEPTING) { 1500 valid = B_TRUE; 1501 } 1502 break; 1503 case DAPLKA_EP_STATE_CONNECTING: 1504 /* 1505 * we transition to this state if daplka_ep_connect 1506 * is successful. from this state, we can transition 1507 * to CONNECTED if daplka_cm_rc_conn_est gets called; 1508 * or to DISCONNECTED if daplka_cm_rc_conn_closed or 1509 * daplka_cm_rc_event_failure gets called. If the 1510 * client calls daplka_ep_disconnect, we transition 1511 * to DISCONNECTING. If a timer was set at ep_connect 1512 * time and if the timer expires prior to any of the 1513 * CM callbacks, we transition to ABORTING and then 1514 * to DISCONNECTED. 1515 */ 1516 if (new_state == DAPLKA_EP_STATE_CONNECTED || 1517 new_state == DAPLKA_EP_STATE_DISCONNECTING || 1518 new_state == DAPLKA_EP_STATE_DISCONNECTED || 1519 new_state == DAPLKA_EP_STATE_ABORTING) { 1520 valid = B_TRUE; 1521 } 1522 break; 1523 case DAPLKA_EP_STATE_ACCEPTING: 1524 /* 1525 * we transition to this state if daplka_cr_accept 1526 * is successful. from this state, we can transition 1527 * to CONNECTED if daplka_cm_service_conn_est gets called; 1528 * or to DISCONNECTED if daplka_cm_service_conn_closed or 1529 * daplka_cm_service_event_failure gets called. If the 1530 * client calls daplka_ep_disconnect, we transition to 1531 * DISCONNECTING. 1532 */ 1533 if (new_state == DAPLKA_EP_STATE_CONNECTED || 1534 new_state == DAPLKA_EP_STATE_DISCONNECTING || 1535 new_state == DAPLKA_EP_STATE_DISCONNECTED) { 1536 valid = B_TRUE; 1537 } 1538 break; 1539 case DAPLKA_EP_STATE_CONNECTED: 1540 /* 1541 * we transition to this state if a active or passive 1542 * connection gets established. if the client calls 1543 * daplka_ep_disconnect, we transition to the 1544 * DISCONNECTING state. subsequent CM callbacks will 1545 * cause ep_state to be set to DISCONNECTED. If the 1546 * remote peer terminates the connection before we do, 1547 * it is possible for us to transition directly from 1548 * CONNECTED to DISCONNECTED. 1549 */ 1550 if (new_state == DAPLKA_EP_STATE_DISCONNECTING || 1551 new_state == DAPLKA_EP_STATE_DISCONNECTED) { 1552 valid = B_TRUE; 1553 } 1554 break; 1555 case DAPLKA_EP_STATE_DISCONNECTING: 1556 /* 1557 * we transition to this state if the client calls 1558 * daplka_ep_disconnect. 1559 */ 1560 if (new_state == DAPLKA_EP_STATE_DISCONNECTED) { 1561 valid = B_TRUE; 1562 } 1563 break; 1564 case DAPLKA_EP_STATE_ABORTING: 1565 /* 1566 * we transition to this state if the active side 1567 * EP timer has expired. this is only a transient 1568 * state that is set during timer processing. when 1569 * timer processing completes, ep_state will become 1570 * DISCONNECTED. 1571 */ 1572 if (new_state == DAPLKA_EP_STATE_DISCONNECTED) { 1573 valid = B_TRUE; 1574 } 1575 break; 1576 case DAPLKA_EP_STATE_DISCONNECTED: 1577 /* 1578 * we transition to this state if we get a closed 1579 * or event_failure CM callback. an expired timer 1580 * can also cause us to be in this state. this 1581 * is the only state in which we permit the 1582 * ep_reinit operation. 1583 */ 1584 if (new_state == DAPLKA_EP_STATE_CLOSED) { 1585 valid = B_TRUE; 1586 } 1587 break; 1588 default: 1589 break; 1590 } 1591 1592 if (!valid) { 1593 DERR("ep_transition: invalid state change %d -> %d\n", 1594 old_state, new_state); 1595 } 1596 return (valid); 1597 } 1598 1599 /* 1600 * first check if the transition is valid. then set ep_state 1601 * to new_state and wake up all waiters. 1602 */ 1603 static void 1604 daplka_ep_set_state(daplka_ep_resource_t *ep_rp, uint32_t old_state, 1605 uint32_t new_state) 1606 { 1607 boolean_t valid; 1608 1609 ASSERT(new_state != DAPLKA_EP_STATE_TRANSITIONING); 1610 1611 valid = daplka_ep_transition_is_valid(old_state, new_state); 1612 mutex_enter(&ep_rp->ep_lock); 1613 if (ep_rp->ep_state != DAPLKA_EP_STATE_FREED) { 1614 if (valid) { 1615 ep_rp->ep_state = new_state; 1616 } else { 1617 /* 1618 * this case is impossible. 1619 * we have a serious problem if we get here. 1620 * instead of panicing, we reset the state to 1621 * old_state. doing this would at least prevent 1622 * threads from hanging due to ep_state being 1623 * stuck in TRANSITIONING. 1624 */ 1625 ep_rp->ep_state = old_state; 1626 ASSERT(B_FALSE); 1627 } 1628 } 1629 cv_broadcast(&ep_rp->ep_cv); 1630 mutex_exit(&ep_rp->ep_lock); 1631 } 1632 1633 /* 1634 * modifies RC channel attributes. 1635 * currently, only the rdma_in and rdma_out attributes may 1636 * be modified. the channel must be in quiescent state when 1637 * this function is called. 1638 */ 1639 /* ARGSUSED */ 1640 static int 1641 daplka_ep_modify(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 1642 cred_t *cred, int *rvalp) 1643 { 1644 daplka_ep_resource_t *ep_rp = NULL; 1645 ibt_cep_modify_flags_t good_flags; 1646 ibt_rc_chan_modify_attr_t rcm_attr; 1647 ibt_hca_attr_t *hca_attrp; 1648 dapl_ep_modify_t args; 1649 ibt_status_t status; 1650 uint32_t old_state, new_state; 1651 int retval = 0; 1652 1653 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_ep_modify_t), 1654 mode); 1655 if (retval != 0) { 1656 DERR("ep_modify: copyin error %d\n", retval); 1657 return (EFAULT); 1658 } 1659 ep_rp = (daplka_ep_resource_t *) 1660 daplka_hash_lookup(&ia_rp->ia_ep_htbl, args.epm_hkey); 1661 if (ep_rp == NULL) { 1662 DERR("ep_modify: cannot find ep resource\n"); 1663 return (EINVAL); 1664 } 1665 ASSERT(DAPLKA_RS_TYPE(ep_rp) == DAPL_TYPE_EP); 1666 new_state = old_state = daplka_ep_get_state(ep_rp); 1667 1668 if (old_state != DAPLKA_EP_STATE_CLOSED && 1669 old_state != DAPLKA_EP_STATE_DISCONNECTED) { 1670 DERR("ep_modify: invalid state %d\n", old_state); 1671 retval = EINVAL; 1672 goto cleanup; 1673 } 1674 1675 good_flags = IBT_CEP_SET_RDMARA_OUT | IBT_CEP_SET_RDMARA_IN; 1676 if ((args.epm_flags & ~good_flags) != 0) { 1677 DERR("ep_modify: invalid flags 0x%x\n", args.epm_flags); 1678 retval = EINVAL; 1679 goto cleanup; 1680 } 1681 1682 hca_attrp = &ia_rp->ia_hca->hca_attr; 1683 1684 bzero(&rcm_attr, sizeof (ibt_rc_chan_modify_attr_t)); 1685 if ((args.epm_flags & IBT_CEP_SET_RDMARA_OUT) != 0) { 1686 if (args.epm_rdma_ra_out > hca_attrp->hca_max_rdma_out_chan) { 1687 DERR("ep_modify: invalid epm_rdma_ra_out %d\n", 1688 args.epm_rdma_ra_out); 1689 retval = EINVAL; 1690 goto cleanup; 1691 } 1692 rcm_attr.rc_rdma_ra_out = args.epm_rdma_ra_out; 1693 } 1694 if ((args.epm_flags & IBT_CEP_SET_RDMARA_IN) != 0) { 1695 if (args.epm_rdma_ra_in > hca_attrp->hca_max_rdma_in_chan) { 1696 DERR("ep_modify: epm_rdma_ra_in %d\n", 1697 args.epm_rdma_ra_in); 1698 retval = EINVAL; 1699 goto cleanup; 1700 } 1701 rcm_attr.rc_rdma_ra_in = args.epm_rdma_ra_in; 1702 } 1703 status = ibt_modify_rc_channel(ep_rp->ep_chan_hdl, args.epm_flags, 1704 &rcm_attr, NULL); 1705 if (status != IBT_SUCCESS) { 1706 DERR("ep_modify: modify_rc_channel returned %d\n", status); 1707 *rvalp = (int)status; 1708 retval = 0; 1709 goto cleanup; 1710 } 1711 1712 /* 1713 * ep_modify does not change ep_state 1714 */ 1715 cleanup:; 1716 daplka_ep_set_state(ep_rp, old_state, new_state); 1717 DAPLKA_RS_UNREF(ep_rp); 1718 return (retval); 1719 } 1720 1721 /* 1722 * Frees a EP resource. 1723 * a EP may only be freed when it is in the CLOSED or 1724 * DISCONNECTED state. 1725 */ 1726 /* ARGSUSED */ 1727 static int 1728 daplka_ep_free(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 1729 cred_t *cred, int *rvalp) 1730 { 1731 daplka_ep_resource_t *ep_rp = NULL; 1732 dapl_ep_free_t args; 1733 uint32_t old_state, new_state; 1734 int retval; 1735 1736 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_ep_free_t), mode); 1737 if (retval != 0) { 1738 DERR("ep_free: copyin error %d\n", retval); 1739 return (EFAULT); 1740 } 1741 ep_rp = (daplka_ep_resource_t *) 1742 daplka_hash_lookup(&ia_rp->ia_ep_htbl, args.epf_hkey); 1743 if (ep_rp == NULL) { 1744 DERR("ep_free: cannot find ep resource\n"); 1745 return (EINVAL); 1746 } 1747 ASSERT(DAPLKA_RS_TYPE(ep_rp) == DAPL_TYPE_EP); 1748 new_state = old_state = daplka_ep_get_state(ep_rp); 1749 1750 /* 1751 * ep cannot be freed if it is in an invalid state. 1752 */ 1753 if (old_state != DAPLKA_EP_STATE_CLOSED && 1754 old_state != DAPLKA_EP_STATE_DISCONNECTED) { 1755 DERR("ep_free: invalid state %d\n", old_state); 1756 retval = EINVAL; 1757 goto cleanup; 1758 } 1759 ep_rp = NULL; 1760 retval = daplka_hash_remove(&ia_rp->ia_ep_htbl, 1761 args.epf_hkey, (void **)&ep_rp); 1762 if (retval != 0 || ep_rp == NULL) { 1763 /* 1764 * this is only possible if we have two threads 1765 * calling ep_free in parallel. 1766 */ 1767 DERR("ep_free: cannot find ep resource\n"); 1768 goto cleanup; 1769 } 1770 /* there should not be any outstanding timers */ 1771 ASSERT(ep_rp->ep_timer_hkey == 0); 1772 1773 new_state = DAPLKA_EP_STATE_FREED; 1774 daplka_ep_set_state(ep_rp, old_state, new_state); 1775 1776 /* remove reference obtained by lookup */ 1777 DAPLKA_RS_UNREF(ep_rp); 1778 1779 /* UNREF calls the actual free function when refcnt is zero */ 1780 DAPLKA_RS_UNREF(ep_rp); 1781 return (0); 1782 1783 cleanup:; 1784 daplka_ep_set_state(ep_rp, old_state, new_state); 1785 1786 /* remove reference obtained by lookup */ 1787 DAPLKA_RS_UNREF(ep_rp); 1788 return (retval); 1789 } 1790 1791 /* 1792 * The following routines supports the timeout feature of ep_connect. 1793 * Refer to the description of ep_connect for details. 1794 */ 1795 1796 /* 1797 * this is the timer processing thread. 1798 */ 1799 static void 1800 daplka_timer_thread(void *arg) 1801 { 1802 daplka_timer_info_t *timerp = (daplka_timer_info_t *)arg; 1803 daplka_ep_resource_t *ep_rp; 1804 daplka_evd_event_t *disc_ev = NULL; 1805 ibt_status_t status; 1806 int old_state, new_state; 1807 1808 ep_rp = timerp->ti_ep_res; 1809 ASSERT(ep_rp != NULL); 1810 ASSERT(timerp->ti_tmo_id != 0); 1811 timerp->ti_tmo_id = 0; 1812 1813 new_state = old_state = daplka_ep_get_state(ep_rp); 1814 if (old_state != DAPLKA_EP_STATE_CONNECTING) { 1815 /* unblock hash_ep_free */ 1816 mutex_enter(&ep_rp->ep_lock); 1817 ASSERT(ep_rp->ep_timer_hkey != 0); 1818 ep_rp->ep_timer_hkey = 0; 1819 cv_broadcast(&ep_rp->ep_cv); 1820 mutex_exit(&ep_rp->ep_lock); 1821 1822 /* reset state to original state */ 1823 daplka_ep_set_state(ep_rp, old_state, new_state); 1824 1825 /* this function will also unref ep_rp */ 1826 daplka_timer_info_free(timerp); 1827 return; 1828 } 1829 1830 ASSERT(ep_rp->ep_timer_hkey != 0); 1831 ep_rp->ep_timer_hkey = 0; 1832 1833 /* 1834 * we cannot keep ep_state in TRANSITIONING if we call 1835 * ibt_close_rc_channel in blocking mode. this would cause 1836 * a deadlock because the cm callbacks will be blocked and 1837 * will not be able to wake us up. 1838 */ 1839 new_state = DAPLKA_EP_STATE_ABORTING; 1840 daplka_ep_set_state(ep_rp, old_state, new_state); 1841 1842 /* 1843 * when we return from close_rc_channel, all callbacks should have 1844 * completed. we can also be certain that these callbacks did not 1845 * enqueue any events to conn_evd. 1846 */ 1847 status = ibt_close_rc_channel(ep_rp->ep_chan_hdl, IBT_BLOCKING, 1848 NULL, 0, NULL, NULL, NULL); 1849 if (status != IBT_SUCCESS) { 1850 DERR("timer_thread: ibt_close_rc_channel returned %d\n", 1851 status); 1852 } 1853 old_state = daplka_ep_get_state(ep_rp); 1854 1855 /* 1856 * this is the only thread that can transition ep_state out 1857 * of ABORTING. all other ep operations would fail when 1858 * ep_state is in ABORTING. 1859 */ 1860 ASSERT(old_state == DAPLKA_EP_STATE_ABORTING); 1861 1862 disc_ev = kmem_zalloc(sizeof (daplka_evd_event_t), KM_SLEEP); 1863 ASSERT(disc_ev != NULL); 1864 1865 disc_ev->ee_cmev.ec_cm_ev_type = DAPL_IB_CME_TIMED_OUT; 1866 disc_ev->ee_cmev.ec_cm_cookie = ep_rp->ep_cookie; 1867 disc_ev->ee_cmev.ec_cm_is_passive = B_FALSE; 1868 disc_ev->ee_cmev.ec_cm_psep_cookie = 0; 1869 disc_ev->ee_cmev.ec_cm_ev_priv_data = NULL; 1870 disc_ev->ee_cmev.ec_cm_ev_priv_data_len = 0; 1871 1872 D2("timer_thread: enqueue event(%p) evdp(%p)\n", 1873 disc_ev, ep_rp->ep_conn_evd); 1874 1875 new_state = DAPLKA_EP_STATE_DISCONNECTED; 1876 daplka_ep_set_state(ep_rp, old_state, new_state); 1877 1878 daplka_evd_wakeup(ep_rp->ep_conn_evd, 1879 &ep_rp->ep_conn_evd->evd_conn_events, disc_ev); 1880 1881 /* this function will also unref ep_rp */ 1882 daplka_timer_info_free(timerp); 1883 } 1884 1885 /* 1886 * dispatches a thread to continue with timer processing. 1887 */ 1888 static void 1889 daplka_timer_dispatch(void *arg) 1890 { 1891 /* 1892 * keep rescheduling this function until 1893 * taskq_dispatch succeeds. 1894 */ 1895 if (taskq_dispatch(daplka_taskq, 1896 daplka_timer_thread, arg, TQ_NOSLEEP) == 0) { 1897 DERR("timer_dispatch: taskq_dispatch failed, retrying...\n"); 1898 (void) timeout(daplka_timer_dispatch, arg, 10); 1899 } 1900 } 1901 1902 /* 1903 * this function is called by the kernel's callout thread. 1904 * we first attempt to remove the timer object from the 1905 * global timer table. if it is found, we dispatch a thread 1906 * to continue processing the timer object. if it is not 1907 * found, that means the timer has been cancelled by someone 1908 * else. 1909 */ 1910 static void 1911 daplka_timer_handler(void *arg) 1912 { 1913 uint64_t timer_hkey = (uintptr_t)arg; 1914 daplka_timer_info_t *timerp = NULL; 1915 1916 D2("timer_handler: timer_hkey 0x%llx\n", (longlong_t)timer_hkey); 1917 1918 (void) daplka_hash_remove(&daplka_timer_info_htbl, 1919 timer_hkey, (void **)&timerp); 1920 if (timerp == NULL) { 1921 D2("timer_handler: timer already cancelled\n"); 1922 return; 1923 } 1924 daplka_timer_dispatch((void *)timerp); 1925 } 1926 1927 /* 1928 * allocates a timer_info object. 1929 * a reference to a EP is held by this object. this ensures 1930 * that the EP stays valid when a timer is outstanding. 1931 */ 1932 static daplka_timer_info_t * 1933 daplka_timer_info_alloc(daplka_ep_resource_t *ep_rp) 1934 { 1935 daplka_timer_info_t *timerp; 1936 1937 timerp = kmem_zalloc(sizeof (*timerp), daplka_km_flags); 1938 if (timerp == NULL) { 1939 DERR("timer_info_alloc: cannot allocate timer info\n"); 1940 return (NULL); 1941 } 1942 timerp->ti_ep_res = ep_rp; 1943 timerp->ti_tmo_id = 0; 1944 1945 return (timerp); 1946 } 1947 1948 /* 1949 * Frees the timer_info object. 1950 * we release the EP reference before freeing the object. 1951 */ 1952 static void 1953 daplka_timer_info_free(daplka_timer_info_t *timerp) 1954 { 1955 ASSERT(timerp->ti_ep_res != NULL); 1956 DAPLKA_RS_UNREF(timerp->ti_ep_res); 1957 timerp->ti_ep_res = NULL; 1958 ASSERT(timerp->ti_tmo_id == 0); 1959 kmem_free(timerp, sizeof (*timerp)); 1960 } 1961 1962 /* 1963 * cancels the timer set by ep_connect. 1964 * returns -1 if timer handling is in progress 1965 * and 0 otherwise. 1966 */ 1967 static int 1968 daplka_cancel_timer(daplka_ep_resource_t *ep_rp) 1969 { 1970 /* 1971 * this function can only be called when ep_state 1972 * is frozen. 1973 */ 1974 ASSERT(ep_rp->ep_state == DAPLKA_EP_STATE_TRANSITIONING); 1975 if (ep_rp->ep_timer_hkey != 0) { 1976 daplka_timer_info_t *timerp = NULL; 1977 1978 (void) daplka_hash_remove(&daplka_timer_info_htbl, 1979 ep_rp->ep_timer_hkey, (void **)&timerp); 1980 if (timerp == NULL) { 1981 /* 1982 * this is possible if the timer_handler has 1983 * removed the timerp but the taskq thread has 1984 * not transitioned the ep_state to DISCONNECTED. 1985 * we need to reset the ep_state to allow the 1986 * taskq thread to continue with its work. the 1987 * taskq thread will set the ep_timer_hkey to 0 1988 * so we don't have to do it here. 1989 */ 1990 DERR("cancel_timer: timer is being processed\n"); 1991 return (-1); 1992 } 1993 /* 1994 * we got the timer object. if the handler fires at 1995 * this point, it will not be able to find the object 1996 * and will return immediately. normally, ti_tmo_id gets 1997 * cleared when the handler fires. 1998 */ 1999 ASSERT(timerp->ti_tmo_id != 0); 2000 2001 /* 2002 * note that untimeout can possibly call the handler. 2003 * we are safe because the handler will be a no-op. 2004 */ 2005 (void) untimeout(timerp->ti_tmo_id); 2006 timerp->ti_tmo_id = 0; 2007 daplka_timer_info_free(timerp); 2008 ep_rp->ep_timer_hkey = 0; 2009 } 2010 return (0); 2011 } 2012 2013 /* 2014 * this function is called by daplka_hash_destroy for 2015 * freeing timer_info objects 2016 */ 2017 static void 2018 daplka_hash_timer_free(void *obj) 2019 { 2020 daplka_timer_info_free((daplka_timer_info_t *)obj); 2021 } 2022 2023 /* ARGSUSED */ 2024 static uint16_t 2025 daplka_hellomsg_cksum(DAPL_PRIVATE *dp) 2026 { 2027 uint8_t *bp; 2028 int i; 2029 uint16_t cksum = 0; 2030 2031 bp = (uint8_t *)dp; 2032 for (i = 0; i < sizeof (DAPL_PRIVATE); i++) { 2033 cksum += bp[i]; 2034 } 2035 return (cksum); 2036 } 2037 2038 /* 2039 * ep_connect is called by the client to initiate a connection to a 2040 * remote service point. It is a non-blocking call. If a non-zero 2041 * timeout is specified by the client, a timer will be set just before 2042 * returning from ep_connect. Upon a successful return from ep_connect, 2043 * the client will call evd_wait to wait for the connection to complete. 2044 * If the connection is rejected or has failed due to an error, the 2045 * client will be notified with an event containing the appropriate error 2046 * code. If the connection is accepted, the client will be notified with 2047 * the CONN_ESTABLISHED event. If the timer expires before either of the 2048 * above events (error or established), a TIMED_OUT event will be delivered 2049 * to the client. 2050 * 2051 * the complicated part of the timer logic is the handling of race 2052 * conditions with CM callbacks. we need to ensure that either the CM or 2053 * the timer thread gets to deliver an event, but not both. when the 2054 * CM callback is about to deliver an event, it always tries to cancel 2055 * the outstanding timer. if cancel_timer indicates a that the timer is 2056 * already being processed, the CM callback will simply return without 2057 * delivering an event. when the timer thread executes, it tries to check 2058 * if the EP is still in CONNECTING state (timers only work on the active 2059 * side). if the EP is not in this state, the timer thread will return 2060 * without delivering an event. 2061 */ 2062 /* ARGSUSED */ 2063 static int 2064 daplka_ep_connect(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 2065 cred_t *cred, int *rvalp) 2066 { 2067 daplka_ep_resource_t *ep_rp = NULL; 2068 dapl_ep_connect_t args; 2069 daplka_timer_info_t *timerp = NULL; 2070 uint32_t old_state, new_state; 2071 boolean_t timer_inserted = B_FALSE; 2072 uint64_t timer_hkey = 0; 2073 ibt_path_info_t path_info; 2074 ibt_path_attr_t path_attr; 2075 ibt_hca_attr_t *hca_attrp; 2076 ibt_chan_open_args_t chan_args; 2077 ibt_status_t status = IBT_SUCCESS; 2078 uint8_t num_paths; 2079 void *priv_data; 2080 DAPL_PRIVATE *dp; 2081 int retval = 0; 2082 ib_gid_t *sgid; 2083 ib_gid_t *dgid; 2084 uint64_t dgid_ored; 2085 ibt_ar_t ar_query_s; 2086 ibt_ar_t ar_result_s; 2087 ibt_path_flags_t pathflags; 2088 2089 D3("ep_connect: enter\n"); 2090 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_ep_connect_t), 2091 mode); 2092 if (retval != 0) { 2093 DERR("ep_connect: copyin error %d\n", retval); 2094 return (EFAULT); 2095 } 2096 ep_rp = (daplka_ep_resource_t *) 2097 daplka_hash_lookup(&ia_rp->ia_ep_htbl, args.epc_hkey); 2098 if (ep_rp == NULL) { 2099 DERR("ep_connect: cannot find ep resource\n"); 2100 return (EINVAL); 2101 } 2102 ASSERT(DAPLKA_RS_TYPE(ep_rp) == DAPL_TYPE_EP); 2103 2104 new_state = old_state = daplka_ep_get_state(ep_rp); 2105 if (old_state != DAPLKA_EP_STATE_CLOSED) { 2106 DERR("ep_connect: invalid state %d\n", old_state); 2107 retval = EINVAL; 2108 goto cleanup; 2109 } 2110 if (args.epc_priv_sz > DAPL_MAX_PRIVATE_DATA_SIZE) { 2111 DERR("ep_connect: private data len (%d) exceeded " 2112 "max size %d\n", args.epc_priv_sz, 2113 DAPL_MAX_PRIVATE_DATA_SIZE); 2114 retval = EINVAL; 2115 goto cleanup; 2116 } 2117 2118 /* 2119 * check for remote ipaddress to dgid resolution needs ATS 2120 */ 2121 dgid = &args.epc_dgid; 2122 dgid_ored = dgid->gid_guid | dgid->gid_prefix; 2123 #if defined(DAPLKA_DEBUG_FORCE_ATS) 2124 dgid_ored = 0ULL; 2125 #endif /* DAPLKA_DEBUG_FORCE_ATS */ 2126 /* check for unidentified dgid */ 2127 if (dgid_ored == 0ULL) { 2128 /* 2129 * setup for ibt_query_ar() 2130 */ 2131 sgid = &ia_rp->ia_hca_sgid; 2132 ar_query_s.ar_gid.gid_guid = 0ULL; 2133 ar_query_s.ar_gid.gid_prefix = 0ULL; 2134 ar_query_s.ar_pkey = 0; 2135 bcopy(args.epc_raddr_sadata.iad_sadata, 2136 ar_query_s.ar_data, DAPL_ATS_NBYTES); 2137 #define UR(b) ar_query_s.ar_data[(b)] 2138 D3("daplka_ep_connect: SA[8] %d.%d.%d.%d\n", 2139 UR(8), UR(9), UR(10), UR(11)); 2140 D3("daplka_ep_connect: SA[12] %d.%d.%d.%d\n", 2141 UR(12), UR(13), UR(14), UR(15)); 2142 status = ibt_query_ar(sgid, &ar_query_s, &ar_result_s); 2143 if (status != IBT_SUCCESS) { 2144 DERR("ep_connect: ibt_query_ar returned %d\n", status); 2145 *rvalp = (int)status; 2146 retval = 0; 2147 goto cleanup; 2148 } 2149 /* 2150 * dgid identified from SA record 2151 */ 2152 dgid = &ar_result_s.ar_gid; 2153 D2("daplka_ep_connect: ATS dgid=%llx:%llx\n", 2154 (longlong_t)dgid->gid_prefix, (longlong_t)dgid->gid_guid); 2155 } 2156 2157 bzero(&path_info, sizeof (ibt_path_info_t)); 2158 bzero(&path_attr, sizeof (ibt_path_attr_t)); 2159 bzero(&chan_args, sizeof (ibt_chan_open_args_t)); 2160 2161 path_attr.pa_dgids = dgid; 2162 path_attr.pa_num_dgids = 1; 2163 /* 2164 * don't set sid in path_attr saves 1 SA query 2165 * Also makes server side not to write the service record 2166 */ 2167 path_attr.pa_sgid = ia_rp->ia_hca_sgid; 2168 path_attr.pa_pkey = ia_rp->ia_port_pkey; 2169 2170 /* save the connection ep - struct copy */ 2171 ep_rp->ep_sgid = ia_rp->ia_hca_sgid; 2172 ep_rp->ep_dgid = *dgid; 2173 2174 num_paths = 0; 2175 pathflags = IBT_PATH_PKEY; 2176 /* enable APM on remote port but not on loopback case */ 2177 if (daplka_apm && ((dgid->gid_prefix != path_attr.pa_sgid.gid_prefix) || 2178 (dgid->gid_guid != path_attr.pa_sgid.gid_guid))) { 2179 pathflags |= IBT_PATH_APM; 2180 } 2181 status = ibt_get_paths(daplka_dev->daplka_clnt_hdl, 2182 pathflags, &path_attr, 1, &path_info, &num_paths); 2183 2184 if (status != IBT_SUCCESS && status != IBT_INSUFF_DATA) { 2185 DERR("ep_connect: ibt_get_paths returned %d paths %d\n", 2186 status, num_paths); 2187 *rvalp = (int)status; 2188 retval = 0; 2189 goto cleanup; 2190 } 2191 /* fill in the sid directly to path_info */ 2192 path_info.pi_sid = args.epc_sid; 2193 hca_attrp = &ia_rp->ia_hca->hca_attr; 2194 2195 /* fill in open channel args */ 2196 chan_args.oc_path = &path_info; 2197 chan_args.oc_cm_handler = daplka_cm_rc_handler; 2198 chan_args.oc_cm_clnt_private = (void *)ep_rp; 2199 chan_args.oc_rdma_ra_out = hca_attrp->hca_max_rdma_out_chan; 2200 chan_args.oc_rdma_ra_in = hca_attrp->hca_max_rdma_in_chan; 2201 chan_args.oc_path_retry_cnt = 7; /* 3-bit field */ 2202 chan_args.oc_path_rnr_retry_cnt = IBT_RNR_INFINITE_RETRY; 2203 2204 ASSERT(args.epc_priv_sz > 0); 2205 priv_data = (void *)args.epc_priv; 2206 2207 chan_args.oc_priv_data_len = args.epc_priv_sz; 2208 chan_args.oc_priv_data = priv_data; 2209 2210 /* 2211 * calculate checksum value of hello message and 2212 * put hello message in networking byte order 2213 */ 2214 dp = (DAPL_PRIVATE *)priv_data; 2215 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*dp)) 2216 dp->hello_msg.hi_port = htons(dp->hello_msg.hi_port); 2217 dp->hello_msg.hi_checksum = 0; 2218 dp->hello_msg.hi_checksum = htons(daplka_hellomsg_cksum(dp)); 2219 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(*dp)) 2220 2221 if (args.epc_timeout > 0) { 2222 /* 2223 * increment refcnt before passing reference to 2224 * timer_info_alloc. 2225 */ 2226 DAPLKA_RS_REF(ep_rp); 2227 timerp = daplka_timer_info_alloc(ep_rp); 2228 if (timerp == NULL) { 2229 DERR("ep_connect: cannot allocate timer\n"); 2230 /* 2231 * we need to remove the reference if 2232 * allocation failed. 2233 */ 2234 DAPLKA_RS_UNREF(ep_rp); 2235 retval = ENOMEM; 2236 goto cleanup; 2237 } 2238 /* 2239 * We generate our own hkeys so that timer_hkey can fit 2240 * into a pointer and passed as an arg to timeout() 2241 */ 2242 timer_hkey = (uint64_t)daplka_timer_hkey_gen(); 2243 retval = daplka_hash_insert(&daplka_timer_info_htbl, 2244 &timer_hkey, (void *)timerp); 2245 if (retval != 0) { 2246 DERR("ep_connect: cannot insert timer info\n"); 2247 goto cleanup; 2248 } 2249 ASSERT(ep_rp->ep_timer_hkey == 0); 2250 ep_rp->ep_timer_hkey = timer_hkey; 2251 timer_inserted = B_TRUE; 2252 D2("ep_connect: timer_hkey = 0x%llx\n", 2253 (longlong_t)timer_hkey); 2254 } 2255 status = ibt_open_rc_channel(ep_rp->ep_chan_hdl, IBT_OCHAN_NO_FLAGS, 2256 IBT_NONBLOCKING, &chan_args, NULL); 2257 2258 if (status != IBT_SUCCESS) { 2259 DERR("ep_connect: ibt_open_rc_channel returned %d\n", status); 2260 *rvalp = (int)status; 2261 retval = 0; 2262 goto cleanup; 2263 } 2264 /* 2265 * if a cm callback gets called at this point, it'll have to wait until 2266 * ep_state becomes connecting (or some other state if another thread 2267 * manages to get ahead of the callback). this guarantees that the 2268 * callback will not touch the timer until it gets set. 2269 */ 2270 if (timerp != NULL) { 2271 clock_t tmo; 2272 2273 tmo = drv_usectohz((clock_t)args.epc_timeout); 2274 /* 2275 * We generate our own 32 bit timer_hkey so that it can fit 2276 * into a pointer 2277 */ 2278 ASSERT(timer_hkey != 0); 2279 timerp->ti_tmo_id = timeout(daplka_timer_handler, 2280 (void *)(uintptr_t)timer_hkey, tmo); 2281 } 2282 new_state = DAPLKA_EP_STATE_CONNECTING; 2283 2284 cleanup:; 2285 if (timerp != NULL && (retval != 0 || status != IBT_SUCCESS)) { 2286 /* 2287 * if ibt_open_rc_channel failed, the timerp must still 2288 * be in daplka_timer_info_htbl because neither the cm 2289 * callback nor the timer_handler will be called. 2290 */ 2291 if (timer_inserted) { 2292 daplka_timer_info_t *new_timerp = NULL; 2293 2294 ASSERT(timer_hkey != 0); 2295 (void) daplka_hash_remove(&daplka_timer_info_htbl, 2296 timer_hkey, (void **)&new_timerp); 2297 ASSERT(new_timerp == timerp); 2298 ep_rp->ep_timer_hkey = 0; 2299 } 2300 daplka_timer_info_free(timerp); 2301 } 2302 daplka_ep_set_state(ep_rp, old_state, new_state); 2303 DAPLKA_RS_UNREF(ep_rp); 2304 D3("ep_connect: exit\n"); 2305 return (retval); 2306 } 2307 2308 /* 2309 * ep_disconnect closes a connection with a remote peer. 2310 * if a connection has not been established, ep_disconnect 2311 * will instead flush all recv bufs posted to this channel. 2312 * if the EP state is CONNECTED, CONNECTING or ACCEPTING upon 2313 * entry to ep_disconnect, the EP state will transition to 2314 * DISCONNECTING upon exit. the CM callbacks triggered by 2315 * ibt_close_rc_channel will cause EP state to become 2316 * DISCONNECTED. This function is a no-op if EP state is 2317 * DISCONNECTED. 2318 */ 2319 /* ARGSUSED */ 2320 static int 2321 daplka_ep_disconnect(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 2322 cred_t *cred, int *rvalp) 2323 { 2324 daplka_ep_resource_t *ep_rp = NULL; 2325 dapl_ep_disconnect_t args; 2326 ibt_status_t status; 2327 uint32_t old_state, new_state; 2328 int retval = 0; 2329 2330 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_ep_disconnect_t), 2331 mode); 2332 if (retval != 0) { 2333 DERR("ep_disconnect: copyin error %d\n", retval); 2334 return (EFAULT); 2335 } 2336 ep_rp = (daplka_ep_resource_t *) 2337 daplka_hash_lookup(&ia_rp->ia_ep_htbl, args.epd_hkey); 2338 if (ep_rp == NULL) { 2339 DERR("ep_disconnect: cannot find ep resource\n"); 2340 return (EINVAL); 2341 } 2342 ASSERT(DAPLKA_RS_TYPE(ep_rp) == DAPL_TYPE_EP); 2343 2344 new_state = old_state = daplka_ep_get_state(ep_rp); 2345 if (old_state != DAPLKA_EP_STATE_CONNECTED && 2346 old_state != DAPLKA_EP_STATE_CONNECTING && 2347 old_state != DAPLKA_EP_STATE_ACCEPTING && 2348 old_state != DAPLKA_EP_STATE_DISCONNECTED && 2349 old_state != DAPLKA_EP_STATE_DISCONNECTING && 2350 old_state != DAPLKA_EP_STATE_CLOSED) { 2351 DERR("ep_disconnect: invalid state %d\n", old_state); 2352 retval = EINVAL; 2353 goto cleanup; 2354 } 2355 2356 if ((old_state == DAPLKA_EP_STATE_DISCONNECTED) || 2357 (old_state == DAPLKA_EP_STATE_DISCONNECTING)) { 2358 D2("ep_disconnect: ep already disconnected\n"); 2359 retval = 0; 2360 /* we leave the state as DISCONNECTED */ 2361 goto cleanup; 2362 } 2363 if (old_state == DAPLKA_EP_STATE_CONNECTING || 2364 old_state == DAPLKA_EP_STATE_ACCEPTING) { 2365 D2("ep_disconnect: aborting, old_state = %d\n", old_state); 2366 } 2367 2368 /* 2369 * according to the udapl spec, ep_disconnect should 2370 * flush the channel if the channel is not CONNECTED. 2371 */ 2372 if (old_state == DAPLKA_EP_STATE_CLOSED) { 2373 status = ibt_flush_channel(ep_rp->ep_chan_hdl); 2374 if (status != IBT_SUCCESS) { 2375 DERR("ep_disconnect: ibt_flush_channel failed %d\n", 2376 status); 2377 *rvalp = (int)status; 2378 } 2379 retval = 0; 2380 /* we leave the state as CLOSED */ 2381 goto cleanup; 2382 } 2383 2384 new_state = DAPLKA_EP_STATE_DISCONNECTING; 2385 daplka_ep_set_state(ep_rp, old_state, new_state); 2386 status = ibt_close_rc_channel(ep_rp->ep_chan_hdl, IBT_NONBLOCKING, 2387 NULL, 0, NULL, NULL, NULL); 2388 2389 if (status == IBT_SUCCESS) { 2390 DAPLKA_RS_UNREF(ep_rp); 2391 return (retval); 2392 } else { 2393 DERR("ep_disconnect: ibt_close_rc_channel returned %d\n", 2394 status); 2395 *rvalp = (int)status; 2396 retval = 0; 2397 new_state = old_state; 2398 } 2399 2400 cleanup:; 2401 daplka_ep_set_state(ep_rp, old_state, new_state); 2402 DAPLKA_RS_UNREF(ep_rp); 2403 return (retval); 2404 } 2405 2406 /* 2407 * this function resets the EP to a usable state (ie. from 2408 * DISCONNECTED to CLOSED). this function is best implemented using 2409 * the ibt_recycle_channel interface. until that is available, we will 2410 * instead clone and tear down the existing channel and replace the 2411 * existing channel with the cloned one. 2412 */ 2413 /* ARGSUSED */ 2414 static int 2415 daplka_ep_reinit(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 2416 cred_t *cred, int *rvalp) 2417 { 2418 daplka_ep_resource_t *ep_rp = NULL; 2419 dapl_ep_reinit_t args; 2420 ibt_status_t status; 2421 uint32_t old_state, new_state; 2422 int retval = 0; 2423 2424 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_ep_reinit_t), 2425 mode); 2426 if (retval != 0) { 2427 DERR("reinit: copyin error %d\n", retval); 2428 return (EFAULT); 2429 } 2430 ep_rp = (daplka_ep_resource_t *) 2431 daplka_hash_lookup(&ia_rp->ia_ep_htbl, args.epri_hkey); 2432 if (ep_rp == NULL) { 2433 DERR("reinit: cannot find ep resource\n"); 2434 return (EINVAL); 2435 } 2436 ASSERT(DAPLKA_RS_TYPE(ep_rp) == DAPL_TYPE_EP); 2437 new_state = old_state = daplka_ep_get_state(ep_rp); 2438 if ((old_state != DAPLKA_EP_STATE_CLOSED) && 2439 (old_state != DAPLKA_EP_STATE_DISCONNECTED)) { 2440 DERR("reinit: invalid state %d\n", old_state); 2441 retval = EINVAL; 2442 goto cleanup; 2443 } 2444 2445 status = ibt_recycle_rc(ep_rp->ep_chan_hdl, 2446 IBT_CEP_RDMA_RD|IBT_CEP_RDMA_WR, 2447 ia_rp->ia_port_num, NULL, NULL); 2448 if (status != IBT_SUCCESS) { 2449 DERR("reinit: unable to clone channel\n"); 2450 *rvalp = (int)status; 2451 retval = 0; 2452 goto cleanup; 2453 } 2454 new_state = DAPLKA_EP_STATE_CLOSED; 2455 2456 cleanup:; 2457 daplka_ep_set_state(ep_rp, old_state, new_state); 2458 DAPLKA_RS_UNREF(ep_rp); 2459 return (retval); 2460 } 2461 2462 /* 2463 * destroys a EP resource. 2464 * called when refcnt drops to zero. 2465 */ 2466 static int 2467 daplka_ep_destroy(daplka_resource_t *gen_rp) 2468 { 2469 daplka_ep_resource_t *ep_rp = (daplka_ep_resource_t *)gen_rp; 2470 ibt_status_t status; 2471 2472 ASSERT(DAPLKA_RS_REFCNT(ep_rp) == 0); 2473 ASSERT(ep_rp->ep_state == DAPLKA_EP_STATE_FREED); 2474 2475 /* 2476 * by the time we get here, we can be sure that 2477 * there is no outstanding timer. 2478 */ 2479 ASSERT(ep_rp->ep_timer_hkey == 0); 2480 2481 D3("ep_destroy: entering, ep_rp 0x%p, rnum %d\n", 2482 ep_rp, DAPLKA_RS_RNUM(ep_rp)); 2483 /* 2484 * free rc channel 2485 */ 2486 if (ep_rp->ep_chan_hdl != NULL) { 2487 mutex_enter(&daplka_dev->daplka_mutex); 2488 ibt_set_chan_private(ep_rp->ep_chan_hdl, NULL); 2489 mutex_exit(&daplka_dev->daplka_mutex); 2490 status = daplka_ibt_free_channel(ep_rp, ep_rp->ep_chan_hdl); 2491 if (status != IBT_SUCCESS) { 2492 DERR("ep_free: ibt_free_channel returned %d\n", 2493 status); 2494 } 2495 ep_rp->ep_chan_hdl = NULL; 2496 D3("ep_destroy: qp freed, rnum %d\n", DAPLKA_RS_RNUM(ep_rp)); 2497 } 2498 /* 2499 * release all references 2500 */ 2501 if (ep_rp->ep_snd_evd != NULL) { 2502 DAPLKA_RS_UNREF(ep_rp->ep_snd_evd); 2503 ep_rp->ep_snd_evd = NULL; 2504 } 2505 if (ep_rp->ep_rcv_evd != NULL) { 2506 DAPLKA_RS_UNREF(ep_rp->ep_rcv_evd); 2507 ep_rp->ep_rcv_evd = NULL; 2508 } 2509 if (ep_rp->ep_conn_evd != NULL) { 2510 DAPLKA_RS_UNREF(ep_rp->ep_conn_evd); 2511 ep_rp->ep_conn_evd = NULL; 2512 } 2513 if (ep_rp->ep_srq_res != NULL) { 2514 DAPLKA_RS_UNREF(ep_rp->ep_srq_res); 2515 ep_rp->ep_srq_res = NULL; 2516 } 2517 if (ep_rp->ep_pd_res != NULL) { 2518 DAPLKA_RS_UNREF(ep_rp->ep_pd_res); 2519 ep_rp->ep_pd_res = NULL; 2520 } 2521 cv_destroy(&ep_rp->ep_cv); 2522 mutex_destroy(&ep_rp->ep_lock); 2523 2524 DAPLKA_RS_FINI(ep_rp); 2525 kmem_free(ep_rp, sizeof (daplka_ep_resource_t)); 2526 D3("ep_destroy: exiting, ep_rp 0x%p\n", ep_rp); 2527 return (0); 2528 } 2529 2530 /* 2531 * this function is called by daplka_hash_destroy for 2532 * freeing EP resource objects 2533 */ 2534 static void 2535 daplka_hash_ep_free(void *obj) 2536 { 2537 daplka_ep_resource_t *ep_rp = (daplka_ep_resource_t *)obj; 2538 ibt_status_t status; 2539 uint32_t old_state, new_state; 2540 int retval; 2541 2542 old_state = daplka_ep_get_state(ep_rp); 2543 retval = daplka_cancel_timer(ep_rp); 2544 new_state = DAPLKA_EP_STATE_FREED; 2545 daplka_ep_set_state(ep_rp, old_state, new_state); 2546 2547 if (retval != 0) { 2548 D2("hash_ep_free: ep_rp 0x%p " 2549 "timer is still being processed\n", ep_rp); 2550 mutex_enter(&ep_rp->ep_lock); 2551 if (ep_rp->ep_timer_hkey != 0) { 2552 D2("hash_ep_free: ep_rp 0x%p " 2553 "waiting for timer_hkey to be 0\n", ep_rp); 2554 cv_wait(&ep_rp->ep_cv, &ep_rp->ep_lock); 2555 } 2556 mutex_exit(&ep_rp->ep_lock); 2557 } 2558 2559 /* call ibt_close_rc_channel regardless of what state we are in */ 2560 status = ibt_close_rc_channel(ep_rp->ep_chan_hdl, IBT_BLOCKING, 2561 NULL, 0, NULL, NULL, NULL); 2562 if (status != IBT_SUCCESS) { 2563 if (old_state == DAPLKA_EP_STATE_CONNECTED || 2564 old_state == DAPLKA_EP_STATE_CONNECTING || 2565 old_state == DAPLKA_EP_STATE_ACCEPTING) { 2566 DERR("hash_ep_free: ep_rp 0x%p state %d " 2567 "unexpected error %d from close_rc_channel\n", 2568 ep_rp, old_state, status); 2569 } 2570 D2("hash_ep_free: close_rc_channel, status %d\n", status); 2571 } 2572 2573 DAPLKA_RS_UNREF(ep_rp); 2574 } 2575 2576 /* 2577 * creates a EVD resource. 2578 * a EVD is used by the client to wait for events from one 2579 * or more sources. 2580 */ 2581 /* ARGSUSED */ 2582 static int 2583 daplka_evd_create(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 2584 cred_t *cred, int *rvalp) 2585 { 2586 daplka_evd_resource_t *evd_rp = NULL; 2587 daplka_async_evd_hkey_t *async_evd; 2588 ibt_hca_attr_t *hca_attrp; 2589 ibt_cq_attr_t cq_attr; 2590 dapl_evd_create_t args; 2591 uint64_t evd_hkey = 0; 2592 boolean_t inserted = B_FALSE; 2593 int retval = 0; 2594 ibt_status_t status; 2595 2596 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_evd_create_t), 2597 mode); 2598 if (retval != 0) { 2599 DERR("evd_create: copyin error %d", retval); 2600 return (EFAULT); 2601 } 2602 if ((args.evd_flags & 2603 ~(DAT_EVD_DEFAULT_FLAG | DAT_EVD_SOFTWARE_FLAG)) != 0) { 2604 DERR("evd_create: invalid flags 0x%x\n", args.evd_flags); 2605 return (EINVAL); 2606 } 2607 2608 evd_rp = kmem_zalloc(sizeof (daplka_evd_resource_t), daplka_km_flags); 2609 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*evd_rp)) 2610 DAPLKA_RS_INIT(evd_rp, DAPL_TYPE_EVD, 2611 DAPLKA_RS_RNUM(ia_rp), daplka_evd_destroy); 2612 2613 mutex_init(&evd_rp->evd_lock, NULL, MUTEX_DRIVER, NULL); 2614 cv_init(&evd_rp->evd_cv, NULL, CV_DRIVER, NULL); 2615 evd_rp->evd_hca = ia_rp->ia_hca; 2616 evd_rp->evd_flags = args.evd_flags; 2617 evd_rp->evd_hca_hdl = ia_rp->ia_hca_hdl; 2618 evd_rp->evd_cookie = args.evd_cookie; 2619 evd_rp->evd_cno_res = NULL; 2620 evd_rp->evd_cr_events.eel_event_type = DAPLKA_EVD_CM_EVENTS; 2621 evd_rp->evd_conn_events.eel_event_type = DAPLKA_EVD_CM_EVENTS; 2622 evd_rp->evd_async_events.eel_event_type = DAPLKA_EVD_ASYNC_EVENTS; 2623 2624 /* 2625 * if the client specified a non-zero cno_hkey, we 2626 * lookup the cno and save the reference for later use. 2627 */ 2628 if (args.evd_cno_hkey > 0) { 2629 daplka_cno_resource_t *cno_rp; 2630 2631 cno_rp = (daplka_cno_resource_t *) 2632 daplka_hash_lookup(&ia_rp->ia_cno_htbl, 2633 args.evd_cno_hkey); 2634 if (cno_rp == NULL) { 2635 DERR("evd_create: cannot find cno resource\n"); 2636 goto cleanup; 2637 } 2638 ASSERT(DAPLKA_RS_TYPE(cno_rp) == DAPL_TYPE_CNO); 2639 evd_rp->evd_cno_res = cno_rp; 2640 } 2641 hca_attrp = &ia_rp->ia_hca->hca_attr; 2642 if ((evd_rp->evd_flags & 2643 (DAT_EVD_DTO_FLAG | DAT_EVD_RMR_BIND_FLAG)) != 0) { 2644 if (args.evd_cq_size > hca_attrp->hca_max_cq_sz) { 2645 DERR("evd_create: invalid cq size %d", 2646 args.evd_cq_size); 2647 retval = EINVAL; 2648 goto cleanup; 2649 } 2650 cq_attr.cq_size = args.evd_cq_size; 2651 cq_attr.cq_sched = NULL; 2652 cq_attr.cq_flags = IBT_CQ_USER_MAP; 2653 2654 status = daplka_ibt_alloc_cq(evd_rp, evd_rp->evd_hca_hdl, 2655 &cq_attr, &evd_rp->evd_cq_hdl, &evd_rp->evd_cq_real_size); 2656 2657 if (status != IBT_SUCCESS) { 2658 DERR("evd_create: ibt_alloc_cq returned %d", status); 2659 *rvalp = (int)status; 2660 retval = 0; 2661 goto cleanup; 2662 } 2663 2664 /* 2665 * store evd ptr with cq_hdl 2666 * mutex is only needed for race of "destroy" and "async" 2667 */ 2668 mutex_enter(&daplka_dev->daplka_mutex); 2669 ibt_set_cq_private(evd_rp->evd_cq_hdl, (void *)evd_rp); 2670 mutex_exit(&daplka_dev->daplka_mutex); 2671 2672 /* Get HCA-specific data_out info */ 2673 status = ibt_ci_data_out(evd_rp->evd_hca_hdl, 2674 IBT_CI_NO_FLAGS, IBT_HDL_CQ, (void *)evd_rp->evd_cq_hdl, 2675 &args.evd_cq_data_out, sizeof (args.evd_cq_data_out)); 2676 2677 if (status != IBT_SUCCESS) { 2678 DERR("evd_create: ibt_ci_data_out error(%d)", status); 2679 *rvalp = (int)status; 2680 retval = 0; 2681 goto cleanup; 2682 } 2683 2684 args.evd_cq_real_size = evd_rp->evd_cq_real_size; 2685 2686 ibt_set_cq_handler(evd_rp->evd_cq_hdl, daplka_cq_handler, 2687 (void *)evd_rp); 2688 } 2689 2690 retval = daplka_hash_insert(&ia_rp->ia_evd_htbl, 2691 &evd_hkey, (void *)evd_rp); 2692 if (retval != 0) { 2693 DERR("evd_ceate: cannot insert evd %d\n", retval); 2694 goto cleanup; 2695 } 2696 inserted = B_TRUE; 2697 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(*evd_rp)) 2698 2699 /* 2700 * If this evd handles async events need to add to the IA resource 2701 * async evd list 2702 */ 2703 if (evd_rp->evd_flags & DAT_EVD_ASYNC_FLAG) { 2704 async_evd = kmem_zalloc(sizeof (daplka_async_evd_hkey_t), 2705 daplka_km_flags); 2706 /* add the evd to the head of the list */ 2707 mutex_enter(&ia_rp->ia_lock); 2708 async_evd->aeh_evd_hkey = evd_hkey; 2709 async_evd->aeh_next = ia_rp->ia_async_evd_hkeys; 2710 ia_rp->ia_async_evd_hkeys = async_evd; 2711 mutex_exit(&ia_rp->ia_lock); 2712 } 2713 2714 args.evd_hkey = evd_hkey; 2715 retval = copyout(&args, (void *)arg, sizeof (dapl_evd_create_t)); 2716 if (retval != 0) { 2717 DERR("evd_create: copyout error %d\n", retval); 2718 retval = EFAULT; 2719 goto cleanup; 2720 } 2721 return (0); 2722 2723 cleanup:; 2724 if (inserted) { 2725 daplka_evd_resource_t *free_rp = NULL; 2726 2727 (void) daplka_hash_remove(&ia_rp->ia_evd_htbl, evd_hkey, 2728 (void **)&free_rp); 2729 if (free_rp != evd_rp) { 2730 DERR("evd_create: cannot remove evd\n"); 2731 /* 2732 * we can only get here if another thread 2733 * has completed the cleanup in evd_free 2734 */ 2735 return (retval); 2736 } 2737 } 2738 DAPLKA_RS_UNREF(evd_rp); 2739 return (retval); 2740 } 2741 2742 /* 2743 * resizes CQ and returns new mapping info to library. 2744 */ 2745 /* ARGSUSED */ 2746 static int 2747 daplka_cq_resize(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 2748 cred_t *cred, int *rvalp) 2749 { 2750 daplka_evd_resource_t *evd_rp = NULL; 2751 ibt_hca_attr_t *hca_attrp; 2752 dapl_cq_resize_t args; 2753 ibt_status_t status; 2754 int retval = 0; 2755 2756 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_cq_resize_t), 2757 mode); 2758 if (retval != 0) { 2759 DERR("cq_resize: copyin error %d\n", retval); 2760 return (EFAULT); 2761 } 2762 2763 /* get evd resource */ 2764 evd_rp = (daplka_evd_resource_t *) 2765 daplka_hash_lookup(&ia_rp->ia_evd_htbl, args.cqr_evd_hkey); 2766 if (evd_rp == NULL) { 2767 DERR("cq_resize: cannot find evd resource\n"); 2768 return (EINVAL); 2769 } 2770 ASSERT(DAPLKA_RS_TYPE(evd_rp) == DAPL_TYPE_EVD); 2771 2772 hca_attrp = &ia_rp->ia_hca->hca_attr; 2773 if (args.cqr_cq_new_size > hca_attrp->hca_max_cq_sz) { 2774 DERR("cq_resize: invalid cq size %d", args.cqr_cq_new_size); 2775 retval = EINVAL; 2776 goto cleanup; 2777 } 2778 /* 2779 * If ibt_resize_cq fails that it is primarily due to resource 2780 * shortage. Per IB spec resize will never loose events and 2781 * a resize error leaves the CQ intact. Therefore even if the 2782 * resize request fails we proceed and get the mapping data 2783 * from the CQ so that the library can mmap it. 2784 */ 2785 status = ibt_resize_cq(evd_rp->evd_cq_hdl, args.cqr_cq_new_size, 2786 &args.cqr_cq_real_size); 2787 if (status != IBT_SUCCESS) { 2788 /* we return the size of the old CQ if resize fails */ 2789 args.cqr_cq_real_size = evd_rp->evd_cq_real_size; 2790 ASSERT(status != IBT_CQ_HDL_INVALID); 2791 DERR("cq_resize: ibt_resize_cq failed:%d\n", status); 2792 } else { 2793 mutex_enter(&evd_rp->evd_lock); 2794 evd_rp->evd_cq_real_size = args.cqr_cq_real_size; 2795 mutex_exit(&evd_rp->evd_lock); 2796 } 2797 2798 D2("cq_resize(%d): done new_sz(%u) real_sz(%u)\n", 2799 DAPLKA_RS_RNUM(evd_rp), 2800 args.cqr_cq_new_size, args.cqr_cq_real_size); 2801 2802 /* Get HCA-specific data_out info */ 2803 status = ibt_ci_data_out(evd_rp->evd_hca_hdl, 2804 IBT_CI_NO_FLAGS, IBT_HDL_CQ, (void *)evd_rp->evd_cq_hdl, 2805 &args.cqr_cq_data_out, sizeof (args.cqr_cq_data_out)); 2806 if (status != IBT_SUCCESS) { 2807 DERR("cq_resize: ibt_ci_data_out error(%d)\n", status); 2808 /* return ibt_ci_data_out status */ 2809 *rvalp = (int)status; 2810 retval = 0; 2811 goto cleanup; 2812 } 2813 2814 retval = ddi_copyout(&args, (void *)arg, sizeof (dapl_cq_resize_t), 2815 mode); 2816 if (retval != 0) { 2817 DERR("cq_resize: copyout error %d\n", retval); 2818 retval = EFAULT; 2819 goto cleanup; 2820 } 2821 2822 cleanup:; 2823 if (evd_rp != NULL) { 2824 DAPLKA_RS_UNREF(evd_rp); 2825 } 2826 return (retval); 2827 } 2828 2829 /* 2830 * Routine to copyin the event poll message so that 32 bit libraries 2831 * can be safely supported 2832 */ 2833 int 2834 daplka_event_poll_copyin(intptr_t inarg, dapl_event_poll_t *outarg, int mode) 2835 { 2836 int retval; 2837 2838 #ifdef _MULTI_DATAMODEL 2839 if ((mode & DATAMODEL_MASK) == DATAMODEL_ILP32) { 2840 dapl_event_poll32_t args32; 2841 2842 retval = ddi_copyin((void *)inarg, &args32, 2843 sizeof (dapl_event_poll32_t), mode); 2844 if (retval != 0) { 2845 DERR("event_poll_copyin: 32bit error %d\n", retval); 2846 return (EFAULT); 2847 } 2848 2849 outarg->evp_evd_hkey = args32.evp_evd_hkey; 2850 outarg->evp_threshold = args32.evp_threshold; 2851 outarg->evp_timeout = args32.evp_timeout; 2852 outarg->evp_ep = (dapl_ib_event_t *)(uintptr_t)args32.evp_ep; 2853 outarg->evp_num_ev = args32.evp_num_ev; 2854 outarg->evp_num_polled = args32.evp_num_polled; 2855 return (0); 2856 } 2857 #endif 2858 retval = ddi_copyin((void *)inarg, outarg, sizeof (dapl_event_poll_t), 2859 mode); 2860 if (retval != 0) { 2861 DERR("event_poll: copyin error %d\n", retval); 2862 return (EFAULT); 2863 } 2864 2865 return (0); 2866 } 2867 2868 /* 2869 * Routine to copyout the event poll message so that 32 bit libraries 2870 * can be safely supported 2871 */ 2872 int 2873 daplka_event_poll_copyout(dapl_event_poll_t *inarg, intptr_t outarg, int mode) 2874 { 2875 int retval; 2876 2877 #ifdef _MULTI_DATAMODEL 2878 if ((mode & DATAMODEL_MASK) == DATAMODEL_ILP32) { 2879 dapl_event_poll32_t args32; 2880 2881 args32.evp_evd_hkey = inarg->evp_evd_hkey; 2882 args32.evp_threshold = inarg->evp_threshold; 2883 args32.evp_timeout = inarg->evp_timeout; 2884 args32.evp_ep = (caddr32_t)(uintptr_t)inarg->evp_ep; 2885 args32.evp_num_ev = inarg->evp_num_ev; 2886 args32.evp_num_polled = inarg->evp_num_polled; 2887 2888 retval = ddi_copyout((void *)&args32, (void *)outarg, 2889 sizeof (dapl_event_poll32_t), mode); 2890 if (retval != 0) { 2891 DERR("event_poll_copyout: 32bit error %d\n", retval); 2892 return (EFAULT); 2893 } 2894 return (0); 2895 } 2896 #endif 2897 retval = ddi_copyout((void *)inarg, (void *)outarg, 2898 sizeof (dapl_event_poll_t), mode); 2899 if (retval != 0) { 2900 DERR("event_poll_copyout: error %d\n", retval); 2901 return (EFAULT); 2902 } 2903 2904 return (0); 2905 } 2906 2907 /* 2908 * fucntion to handle CM REQ RCV private data from Solaris or third parties 2909 */ 2910 /* ARGSUSED */ 2911 static void 2912 daplka_crevent_privdata_post(daplka_ia_resource_t *ia_rp, 2913 dapl_ib_event_t *evd_rp, daplka_evd_event_t *cr_ev) 2914 { 2915 DAPL_PRIVATE *dp; 2916 ib_gid_t *lgid; 2917 ibt_ar_t ar_query_s; 2918 ibt_ar_t ar_result_s; 2919 DAPL_HELLO_MSG *hip; 2920 uint32_t ipaddr_ord; 2921 ibt_priv_data_len_t clen; 2922 ibt_priv_data_len_t olen; 2923 ibt_status_t status; 2924 uint16_t cksum; 2925 2926 /* 2927 * get private data and len 2928 */ 2929 dp = (DAPL_PRIVATE *)cr_ev->ee_cmev.ec_cm_ev_priv_data; 2930 clen = cr_ev->ee_cmev.ec_cm_ev_priv_data_len; 2931 #if defined(DAPLKA_DEBUG_FORCE_ATS) 2932 /* skip the DAPL_PRIVATE chekcsum check */ 2933 #else 2934 /* for remote connects */ 2935 /* look up hello message in the CM private data area */ 2936 if (clen >= sizeof (DAPL_PRIVATE) && 2937 (dp->hello_msg.hi_vers == DAPL_HELLO_MSG_VERS)) { 2938 cksum = ntohs(dp->hello_msg.hi_checksum); 2939 dp->hello_msg.hi_checksum = 0; 2940 if (daplka_hellomsg_cksum(dp) == cksum) { 2941 D2("daplka_crevent_privdata_post: Solaris msg\n"); 2942 evd_rp->ibe_ce.ibce_priv_data_size = clen; 2943 dp->hello_msg.hi_checksum = DAPL_CHECKSUM; 2944 dp->hello_msg.hi_port = ntohs(dp->hello_msg.hi_port); 2945 bcopy(dp, evd_rp->ibe_ce.ibce_priv_data_ptr, clen); 2946 kmem_free(dp, clen); 2947 return; 2948 } 2949 } 2950 #endif /* DAPLKA_DEBUG_FORCE_ATS */ 2951 2952 D2("daplka_crevent_privdata_post: 3rd party msg\n"); 2953 /* transpose CM private data into hello message */ 2954 if (clen) { 2955 olen = clen; 2956 if (clen > DAPL_CONSUMER_MAX_PRIVATE_DATA_SIZE) { 2957 clen = DAPL_CONSUMER_MAX_PRIVATE_DATA_SIZE; 2958 } 2959 bcopy(dp, evd_rp->ibe_ce.ibce_priv_data_ptr, clen); 2960 kmem_free(dp, olen); 2961 } else { 2962 bzero(evd_rp->ibe_ce.ibce_priv_data_ptr, 2963 DAPL_CONSUMER_MAX_PRIVATE_DATA_SIZE); 2964 } 2965 evd_rp->ibe_ce.ibce_priv_data_size = sizeof (DAPL_PRIVATE); 2966 dp = (DAPL_PRIVATE *)evd_rp->ibe_ce.ibce_priv_data_ptr; 2967 /* 2968 * fill in hello message 2969 */ 2970 hip = &dp->hello_msg; 2971 hip->hi_checksum = DAPL_CHECKSUM; 2972 hip->hi_clen = clen; 2973 hip->hi_mid = 0; 2974 hip->hi_vers = DAPL_HELLO_MSG_VERS; 2975 hip->hi_port = 0; 2976 2977 /* assign sgid and dgid */ 2978 lgid = &ia_rp->ia_hca_sgid; 2979 ar_query_s.ar_gid.gid_prefix = 2980 cr_ev->ee_cmev.ec_cm_req_prim_addr.gid_prefix; 2981 ar_query_s.ar_gid.gid_guid = 2982 cr_ev->ee_cmev.ec_cm_req_prim_addr.gid_guid; 2983 ar_query_s.ar_pkey = ia_rp->ia_port_pkey; 2984 bzero(ar_query_s.ar_data, DAPL_ATS_NBYTES); 2985 2986 /* reverse ip address lookup through ATS */ 2987 status = ibt_query_ar(lgid, &ar_query_s, &ar_result_s); 2988 if (status == IBT_SUCCESS) { 2989 bcopy(ar_result_s.ar_data, hip->hi_saaddr, DAPL_ATS_NBYTES); 2990 /* determine the address families */ 2991 ipaddr_ord = hip->hi_v4pad[0] | hip->hi_v4pad[1] | 2992 hip->hi_v4pad[2]; 2993 if (ipaddr_ord == 0) { 2994 hip->hi_ipv = AF_INET; 2995 } else { 2996 hip->hi_ipv = AF_INET6; 2997 } 2998 2999 #define UL(b) ar_result_s.ar_data[(b)] 3000 D3("daplka_privdata_post: family=%d :SA[8] %d.%d.%d.%d\n", 3001 hip->hi_ipv, UL(8), UL(9), UL(10), UL(11)); 3002 D3("daplka_privdata_post: SA[12] %d.%d.%d.%d\n", 3003 UL(12), UL(13), UL(14), UL(15)); 3004 } else { 3005 /* non-conformed third parties */ 3006 hip->hi_ipv = AF_UNSPEC; 3007 bzero(hip->hi_saaddr, DAPL_ATS_NBYTES); 3008 } 3009 } 3010 3011 /* 3012 * this function is called by evd_wait and evd_dequeue to wait for 3013 * connection events and CQ notifications. typically this function 3014 * is called when the userland CQ is empty and the client has 3015 * specified a non-zero timeout to evd_wait. if the client is 3016 * interested in CQ events, the CQ must be armed in userland prior 3017 * to calling this function. 3018 */ 3019 /* ARGSUSED */ 3020 static int 3021 daplka_event_poll(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 3022 cred_t *cred, int *rvalp) 3023 { 3024 daplka_evd_resource_t *evd_rp = NULL; 3025 dapl_event_poll_t args; 3026 daplka_evd_event_t *head; 3027 dapl_ib_event_t evp_arr[NUM_EVENTS_PER_POLL]; 3028 dapl_ib_event_t *evp; 3029 dapl_ib_event_t *evp_start; 3030 size_t evp_size; 3031 int threshold; 3032 clock_t timeout; 3033 uint32_t max_events; 3034 uint32_t num_events = 0; 3035 void *pd; 3036 ibt_priv_data_len_t n; 3037 int retval = 0; 3038 int rc; 3039 3040 retval = daplka_event_poll_copyin(arg, &args, mode); 3041 if (retval != 0) { 3042 return (EFAULT); 3043 } 3044 3045 if ((args.evp_num_ev > 0) && (args.evp_ep == NULL)) { 3046 DERR("event_poll: evp_ep cannot be NULL if num_wc=%d", 3047 args.evp_num_ev); 3048 return (EINVAL); 3049 } 3050 /* 3051 * Note: dequeue requests have a threshold = 0, timeout = 0 3052 */ 3053 threshold = args.evp_threshold; 3054 3055 max_events = args.evp_num_ev; 3056 /* ensure library is passing sensible values */ 3057 if (max_events < threshold) { 3058 DERR("event_poll: max_events(%d) < threshold(%d)\n", 3059 max_events, threshold); 3060 return (EINVAL); 3061 } 3062 /* Do a sanity check to avoid excessive memory allocation */ 3063 if (max_events > DAPL_EVD_MAX_EVENTS) { 3064 DERR("event_poll: max_events(%d) > %d", 3065 max_events, DAPL_EVD_MAX_EVENTS); 3066 return (EINVAL); 3067 } 3068 D4("event_poll: threshold(%d) timeout(0x%llx) max_events(%d)\n", 3069 threshold, (longlong_t)args.evp_timeout, max_events); 3070 3071 /* get evd resource */ 3072 evd_rp = (daplka_evd_resource_t *) 3073 daplka_hash_lookup(&ia_rp->ia_evd_htbl, args.evp_evd_hkey); 3074 if (evd_rp == NULL) { 3075 DERR("event_poll: cannot find evd resource\n"); 3076 return (EINVAL); 3077 } 3078 ASSERT(DAPLKA_RS_TYPE(evd_rp) == DAPL_TYPE_EVD); 3079 3080 /* 3081 * Use event array on the stack if possible 3082 */ 3083 if (max_events <= NUM_EVENTS_PER_POLL) { 3084 evp_start = evp = &evp_arr[0]; 3085 } else { 3086 evp_size = max_events * sizeof (dapl_ib_event_t); 3087 evp_start = evp = kmem_zalloc(evp_size, daplka_km_flags); 3088 if (evp == NULL) { 3089 DERR("event_poll: kmem_zalloc failed, evp_size %d", 3090 evp_size); 3091 retval = ENOMEM; 3092 goto cleanup; 3093 } 3094 } 3095 3096 /* 3097 * The Event poll algorithm is as follows - 3098 * The library passes a buffer big enough to hold "max_events" 3099 * events. max_events is >= threshold. If at any stage we get 3100 * max_events no. of events we bail. The events are polled in 3101 * the following order - 3102 * 1) Check for CR events in the evd_cr_events list 3103 * 2) Check for Connection events in the evd_connection_events list 3104 * 3105 * If after the above 2 steps we don't have enough(>= threshold) events 3106 * we block for CQ notification and sleep. Upon being woken up we start 3107 * at step 1 again. 3108 */ 3109 3110 /* 3111 * Note: this could be 0 or INFINITE or anyother value in microsec 3112 */ 3113 if (args.evp_timeout > 0) { 3114 if (args.evp_timeout >= LONG_MAX) { 3115 timeout = LONG_MAX; 3116 } else { 3117 clock_t curr_time = ddi_get_lbolt(); 3118 3119 timeout = curr_time + 3120 drv_usectohz((clock_t)args.evp_timeout); 3121 /* 3122 * use the max value if we wrapped around 3123 */ 3124 if (timeout <= curr_time) { 3125 timeout = LONG_MAX; 3126 } 3127 } 3128 } else { 3129 timeout = 0; 3130 } 3131 3132 mutex_enter(&evd_rp->evd_lock); 3133 for (;;) { 3134 /* 3135 * If this evd is waiting for CM events check that now. 3136 */ 3137 if ((evd_rp->evd_flags & DAT_EVD_CR_FLAG) && 3138 (evd_rp->evd_cr_events.eel_num_elements > 0)) { 3139 /* dequeue events from evd_cr_events list */ 3140 while (head = daplka_evd_event_dequeue( 3141 &evd_rp->evd_cr_events)) { 3142 /* 3143 * populate the evp array 3144 */ 3145 evp[num_events].ibe_ev_family = DAPL_CR_EVENTS; 3146 evp[num_events].ibe_ce.ibce_event = 3147 head->ee_cmev.ec_cm_ev_type; 3148 evp[num_events].ibe_ce.ibce_cookie = 3149 (uint64_t)head->ee_cmev.ec_cm_cookie; 3150 evp[num_events].ibe_ce.ibce_psep_cookie = 3151 head->ee_cmev.ec_cm_psep_cookie; 3152 daplka_crevent_privdata_post(ia_rp, 3153 &evp[num_events], head); 3154 kmem_free(head, sizeof (daplka_evd_event_t)); 3155 3156 if (++num_events == max_events) { 3157 mutex_exit(&evd_rp->evd_lock); 3158 goto maxevent_reached; 3159 } 3160 } 3161 } 3162 3163 if ((evd_rp->evd_flags & DAT_EVD_CONNECTION_FLAG) && 3164 (evd_rp->evd_conn_events.eel_num_elements > 0)) { 3165 /* dequeue events from evd_connection_events list */ 3166 while ((head = daplka_evd_event_dequeue 3167 (&evd_rp->evd_conn_events))) { 3168 /* 3169 * populate the evp array - 3170 * 3171 */ 3172 if (head->ee_cmev.ec_cm_is_passive) { 3173 evp[num_events].ibe_ev_family = 3174 DAPL_PASSIVE_CONNECTION_EVENTS; 3175 } else { 3176 evp[num_events].ibe_ev_family = 3177 DAPL_ACTIVE_CONNECTION_EVENTS; 3178 } 3179 evp[num_events].ibe_ce.ibce_event = 3180 head->ee_cmev.ec_cm_ev_type; 3181 evp[num_events].ibe_ce.ibce_cookie = 3182 (uint64_t)head->ee_cmev.ec_cm_cookie; 3183 evp[num_events].ibe_ce.ibce_psep_cookie = 3184 head->ee_cmev.ec_cm_psep_cookie; 3185 3186 if (head->ee_cmev.ec_cm_ev_priv_data_len > 0) { 3187 pd = head->ee_cmev.ec_cm_ev_priv_data; 3188 n = head-> 3189 ee_cmev.ec_cm_ev_priv_data_len; 3190 bcopy(pd, (void *)evp[num_events]. 3191 ibe_ce.ibce_priv_data_ptr, n); 3192 evp[num_events].ibe_ce. 3193 ibce_priv_data_size = n; 3194 kmem_free(pd, n); 3195 } 3196 3197 kmem_free(head, sizeof (daplka_evd_event_t)); 3198 3199 if (++num_events == max_events) { 3200 mutex_exit(&evd_rp->evd_lock); 3201 goto maxevent_reached; 3202 } 3203 } 3204 } 3205 3206 if ((evd_rp->evd_flags & DAT_EVD_ASYNC_FLAG) && 3207 (evd_rp->evd_async_events.eel_num_elements > 0)) { 3208 /* dequeue events from evd_async_events list */ 3209 while (head = daplka_evd_event_dequeue( 3210 &evd_rp->evd_async_events)) { 3211 /* 3212 * populate the evp array 3213 */ 3214 evp[num_events].ibe_ev_family = 3215 DAPL_ASYNC_EVENTS; 3216 evp[num_events].ibe_async.ibae_type = 3217 head->ee_aev.ibae_type; 3218 evp[num_events].ibe_async.ibae_hca_guid = 3219 head->ee_aev.ibae_hca_guid; 3220 evp[num_events].ibe_async.ibae_cookie = 3221 head->ee_aev.ibae_cookie; 3222 evp[num_events].ibe_async.ibae_port = 3223 head->ee_aev.ibae_port; 3224 3225 kmem_free(head, sizeof (daplka_evd_event_t)); 3226 3227 if (++num_events == max_events) { 3228 break; 3229 } 3230 } 3231 } 3232 3233 /* 3234 * We have sufficient events for this call so no need to wait 3235 */ 3236 if ((threshold > 0) && (num_events >= threshold)) { 3237 mutex_exit(&evd_rp->evd_lock); 3238 break; 3239 } 3240 3241 evd_rp->evd_waiters++; 3242 /* 3243 * There are no new events and a timeout was specified. 3244 * Note: for CQ events threshold is 0 but timeout is 3245 * not necessarily 0. 3246 */ 3247 while ((evd_rp->evd_newevents == DAPLKA_EVD_NO_EVENTS) && 3248 timeout) { 3249 retval = DAPLKA_EVD_WAIT(&evd_rp->evd_cv, 3250 &evd_rp->evd_lock, timeout); 3251 if (retval == 0) { 3252 retval = EINTR; 3253 break; 3254 } else if (retval == -1) { 3255 retval = ETIME; 3256 break; 3257 } else { 3258 retval = 0; 3259 continue; 3260 } 3261 } 3262 evd_rp->evd_waiters--; 3263 if (evd_rp->evd_newevents != DAPLKA_EVD_NO_EVENTS) { 3264 /* 3265 * If we got woken up by the CQ handler due to events 3266 * in the CQ. Need to go to userland to check for 3267 * CQ events. Or if we were woken up due to S/W events 3268 */ 3269 3270 /* check for userland events only */ 3271 if (!(evd_rp->evd_newevents & 3272 ~DAPLKA_EVD_ULAND_EVENTS)) { 3273 evd_rp->evd_newevents = DAPLKA_EVD_NO_EVENTS; 3274 mutex_exit(&evd_rp->evd_lock); 3275 break; 3276 } 3277 /* 3278 * Clear newevents since we are going to loopback 3279 * back and check for both CM and CQ events 3280 */ 3281 evd_rp->evd_newevents = DAPLKA_EVD_NO_EVENTS; 3282 } else { /* error */ 3283 mutex_exit(&evd_rp->evd_lock); 3284 break; 3285 } 3286 } 3287 3288 maxevent_reached: 3289 args.evp_num_polled = num_events; 3290 3291 /* 3292 * At this point retval might have a value that we want to return 3293 * back to the user. So the copyouts shouldn't tamper retval. 3294 */ 3295 if (args.evp_num_polled > 0) { /* copyout the events */ 3296 rc = ddi_copyout(evp, args.evp_ep, args.evp_num_polled * 3297 sizeof (dapl_ib_event_t), mode); 3298 if (rc != 0) { /* XXX: we are losing events here */ 3299 DERR("event_poll: event array copyout error %d", rc); 3300 retval = EFAULT; 3301 goto cleanup; 3302 } 3303 rc = daplka_event_poll_copyout(&args, arg, mode); 3304 if (rc != 0) { /* XXX: we are losing events here */ 3305 DERR("event_poll: copyout error %d\n", rc); 3306 retval = EFAULT; 3307 goto cleanup; 3308 } 3309 } 3310 3311 cleanup:; 3312 if ((max_events > NUM_EVENTS_PER_POLL) && (evp_start != NULL)) { 3313 kmem_free(evp_start, evp_size); 3314 } 3315 3316 if (evd_rp != NULL) { 3317 DAPLKA_RS_UNREF(evd_rp); 3318 } 3319 return (retval); 3320 } 3321 3322 /* ARGSUSED */ 3323 static int 3324 daplka_event_wakeup(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 3325 cred_t *cred, int *rvalp) 3326 { 3327 dapl_event_wakeup_t args; 3328 daplka_evd_resource_t *evd_rp; 3329 int retval; 3330 3331 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_event_wakeup_t), 3332 mode); 3333 if (retval != 0) { 3334 DERR("event_wakeup: copyin error %d\n", retval); 3335 return (EFAULT); 3336 } 3337 3338 /* get evd resource */ 3339 evd_rp = (daplka_evd_resource_t *) 3340 daplka_hash_lookup(&ia_rp->ia_evd_htbl, args.evw_hkey); 3341 if (evd_rp == NULL) { 3342 DERR("event_wakeup: cannot find evd resource\n"); 3343 return (EINVAL); 3344 } 3345 ASSERT(DAPLKA_RS_TYPE(evd_rp) == DAPL_TYPE_EVD); 3346 3347 daplka_evd_wakeup(evd_rp, NULL, NULL); 3348 3349 DAPLKA_RS_UNREF(evd_rp); 3350 3351 return (retval); 3352 } 3353 3354 /* ARGSUSED */ 3355 static int 3356 daplka_evd_modify_cno(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 3357 cred_t *cred, int *rvalp) 3358 { 3359 dapl_evd_modify_cno_t args; 3360 daplka_evd_resource_t *evd_rp; 3361 daplka_cno_resource_t *cno_rp; 3362 daplka_cno_resource_t *old_cno_rp; 3363 int retval; 3364 3365 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_evd_modify_cno_t), 3366 mode); 3367 if (retval != 0) { 3368 DERR("evd_modify_cno: copyin error %d\n", retval); 3369 return (EFAULT); 3370 } 3371 3372 /* get evd resource */ 3373 evd_rp = (daplka_evd_resource_t *) 3374 daplka_hash_lookup(&ia_rp->ia_evd_htbl, args.evmc_hkey); 3375 if (evd_rp == NULL) { 3376 DERR("evd_modify_cno: cannot find evd resource\n"); 3377 retval = EINVAL; 3378 goto cleanup; 3379 } 3380 ASSERT(DAPLKA_RS_TYPE(evd_rp) == DAPL_TYPE_EVD); 3381 3382 if (args.evmc_cno_hkey > 0) { 3383 /* get cno resource corresponding to the new CNO */ 3384 cno_rp = (daplka_cno_resource_t *) 3385 daplka_hash_lookup(&ia_rp->ia_cno_htbl, 3386 args.evmc_cno_hkey); 3387 if (cno_rp == NULL) { 3388 DERR("evd_modify_cno: cannot find CNO resource\n"); 3389 retval = EINVAL; 3390 goto cleanup; 3391 } 3392 ASSERT(DAPLKA_RS_TYPE(cno_rp) == DAPL_TYPE_CNO); 3393 } else { 3394 cno_rp = NULL; 3395 } 3396 3397 mutex_enter(&evd_rp->evd_lock); 3398 old_cno_rp = evd_rp->evd_cno_res; 3399 evd_rp->evd_cno_res = cno_rp; 3400 mutex_exit(&evd_rp->evd_lock); 3401 3402 /* 3403 * drop the refcnt on the old CNO, the refcnt on the new CNO is 3404 * retained since the evd holds a reference to it. 3405 */ 3406 if (old_cno_rp) { 3407 DAPLKA_RS_UNREF(old_cno_rp); 3408 } 3409 3410 cleanup: 3411 if (evd_rp) { 3412 DAPLKA_RS_UNREF(evd_rp); 3413 } 3414 3415 return (retval); 3416 } 3417 3418 /* 3419 * Frees the EVD and associated resources. 3420 * If there are other threads still using this EVD, the destruction 3421 * will defer until the EVD's refcnt drops to zero. 3422 */ 3423 /* ARGSUSED */ 3424 static int 3425 daplka_evd_free(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 3426 cred_t *cred, int *rvalp) 3427 { 3428 daplka_evd_resource_t *evd_rp = NULL; 3429 daplka_async_evd_hkey_t *curr; 3430 daplka_async_evd_hkey_t *prev; 3431 dapl_evd_free_t args; 3432 int retval = 0; 3433 3434 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_evd_free_t), mode); 3435 if (retval != 0) { 3436 DERR("evd_free: copyin error %d\n", retval); 3437 return (EFAULT); 3438 } 3439 retval = daplka_hash_remove(&ia_rp->ia_evd_htbl, args.evf_hkey, 3440 (void **)&evd_rp); 3441 if (retval != 0 || evd_rp == NULL) { 3442 DERR("evd_free: cannot find evd resource\n"); 3443 return (EINVAL); 3444 } 3445 ASSERT(DAPLKA_RS_TYPE(evd_rp) == DAPL_TYPE_EVD); 3446 3447 /* If this is an async evd remove it from the IA's async evd list */ 3448 if (evd_rp->evd_flags & DAT_EVD_ASYNC_FLAG) { 3449 mutex_enter(&ia_rp->ia_lock); 3450 curr = prev = ia_rp->ia_async_evd_hkeys; 3451 while (curr != NULL) { 3452 if (curr->aeh_evd_hkey == args.evf_hkey) { 3453 /* unlink curr from the list */ 3454 if (curr == prev) { 3455 /* 3456 * if first element in the list update 3457 * the list head 3458 */ 3459 ia_rp->ia_async_evd_hkeys = 3460 curr->aeh_next; 3461 } else { 3462 prev->aeh_next = curr->aeh_next; 3463 } 3464 break; 3465 } 3466 prev = curr; 3467 curr = curr->aeh_next; 3468 } 3469 mutex_exit(&ia_rp->ia_lock); 3470 /* free the curr entry */ 3471 kmem_free(curr, sizeof (daplka_async_evd_hkey_t)); 3472 } 3473 3474 /* UNREF calls the actual free function when refcnt is zero */ 3475 DAPLKA_RS_UNREF(evd_rp); 3476 return (0); 3477 } 3478 3479 /* 3480 * destroys EVD resource. 3481 * called when refcnt drops to zero. 3482 */ 3483 static int 3484 daplka_evd_destroy(daplka_resource_t *gen_rp) 3485 { 3486 daplka_evd_resource_t *evd_rp = (daplka_evd_resource_t *)gen_rp; 3487 ibt_status_t status; 3488 daplka_evd_event_t *evt; 3489 ibt_priv_data_len_t len; 3490 3491 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*evd_rp)) 3492 D3("evd_destroy: entering, evd_rp 0x%p, rnum %d\n", 3493 evd_rp, DAPLKA_RS_RNUM(evd_rp)); 3494 /* 3495 * free CQ 3496 */ 3497 if (evd_rp->evd_cq_hdl) { 3498 ibt_set_cq_handler(evd_rp->evd_cq_hdl, NULL, NULL); 3499 mutex_enter(&daplka_dev->daplka_mutex); 3500 ibt_set_cq_private(evd_rp->evd_cq_hdl, NULL); 3501 mutex_exit(&daplka_dev->daplka_mutex); 3502 3503 status = daplka_ibt_free_cq(evd_rp, evd_rp->evd_cq_hdl); 3504 if (status != IBT_SUCCESS) { 3505 DERR("evd_destroy: ibt_free_cq returned %d\n", status); 3506 } 3507 evd_rp->evd_cq_hdl = NULL; 3508 D2("evd_destroy: cq freed, rnum %d\n", DAPLKA_RS_RNUM(evd_rp)); 3509 } 3510 3511 /* 3512 * release reference on CNO 3513 */ 3514 if (evd_rp->evd_cno_res != NULL) { 3515 mutex_enter(&evd_rp->evd_cno_res->cno_lock); 3516 if (evd_rp->evd_cno_res->cno_evd_cookie == 3517 evd_rp->evd_cookie) { 3518 evd_rp->evd_cno_res->cno_evd_cookie = 0; 3519 } 3520 mutex_exit(&evd_rp->evd_cno_res->cno_lock); 3521 DAPLKA_RS_UNREF(evd_rp->evd_cno_res); 3522 evd_rp->evd_cno_res = NULL; 3523 } 3524 3525 /* 3526 * discard all remaining events 3527 */ 3528 mutex_enter(&evd_rp->evd_lock); 3529 while ((evt = daplka_evd_event_dequeue(&evd_rp->evd_cr_events))) { 3530 D2("evd_destroy: discarding CR event: %d\n", 3531 evt->ee_cmev.ec_cm_ev_type); 3532 len = evt->ee_cmev.ec_cm_ev_priv_data_len; 3533 if (len > 0) { 3534 kmem_free(evt->ee_cmev.ec_cm_ev_priv_data, len); 3535 evt->ee_cmev.ec_cm_ev_priv_data = NULL; 3536 evt->ee_cmev.ec_cm_ev_priv_data_len = 0; 3537 } 3538 kmem_free(evt, sizeof (*evt)); 3539 } 3540 ASSERT(evd_rp->evd_cr_events.eel_num_elements == 0); 3541 3542 while ((evt = daplka_evd_event_dequeue(&evd_rp->evd_conn_events))) { 3543 D2("evd_destroy: discarding CONN event: %d\n", 3544 evt->ee_cmev.ec_cm_ev_type); 3545 len = evt->ee_cmev.ec_cm_ev_priv_data_len; 3546 if (len > 0) { 3547 kmem_free(evt->ee_cmev.ec_cm_ev_priv_data, len); 3548 evt->ee_cmev.ec_cm_ev_priv_data = NULL; 3549 evt->ee_cmev.ec_cm_ev_priv_data_len = 0; 3550 } 3551 kmem_free(evt, sizeof (*evt)); 3552 } 3553 ASSERT(evd_rp->evd_conn_events.eel_num_elements == 0); 3554 3555 while ((evt = daplka_evd_event_dequeue(&evd_rp->evd_async_events))) { 3556 DERR("evd_destroy: discarding ASYNC event: %d\n", 3557 evt->ee_aev.ibae_type); 3558 kmem_free(evt, sizeof (*evt)); 3559 } 3560 ASSERT(evd_rp->evd_async_events.eel_num_elements == 0); 3561 mutex_exit(&evd_rp->evd_lock); 3562 3563 mutex_destroy(&evd_rp->evd_lock); 3564 DAPLKA_RS_FINI(evd_rp); 3565 kmem_free(evd_rp, sizeof (daplka_evd_resource_t)); 3566 D3("evd_destroy: exiting, evd_rp 0x%p\n", evd_rp); 3567 return (0); 3568 } 3569 3570 static void 3571 daplka_hash_evd_free(void *obj) 3572 { 3573 daplka_evd_resource_t *evd_rp = (daplka_evd_resource_t *)obj; 3574 3575 ASSERT(DAPLKA_RS_TYPE(evd_rp) == DAPL_TYPE_EVD); 3576 DAPLKA_RS_UNREF(evd_rp); 3577 } 3578 3579 /* 3580 * this handler fires when new completions arrive. 3581 */ 3582 /* ARGSUSED */ 3583 static void 3584 daplka_cq_handler(ibt_cq_hdl_t ibt_cq, void *arg) 3585 { 3586 D3("cq_handler: fired setting evd_newevents\n"); 3587 daplka_evd_wakeup((daplka_evd_resource_t *)arg, NULL, NULL); 3588 } 3589 3590 /* 3591 * this routine wakes up a client from evd_wait. if evtq and evt 3592 * are non-null, the event evt will be enqueued prior to waking 3593 * up the client. if the evd is associated with a CNO and if there 3594 * are no waiters on the evd, the CNO will be notified. 3595 */ 3596 static void 3597 daplka_evd_wakeup(daplka_evd_resource_t *evd_rp, daplka_evd_event_list_t *evtq, 3598 daplka_evd_event_t *evt) 3599 { 3600 uint32_t waiters = 0; 3601 3602 mutex_enter(&evd_rp->evd_lock); 3603 if (evtq != NULL && evt != NULL) { 3604 ASSERT(evtq == &evd_rp->evd_cr_events || 3605 evtq == &evd_rp->evd_conn_events || 3606 evtq == &evd_rp->evd_async_events); 3607 daplka_evd_event_enqueue(evtq, evt); 3608 ASSERT((evtq->eel_event_type == DAPLKA_EVD_CM_EVENTS) || 3609 (evtq->eel_event_type == DAPLKA_EVD_ASYNC_EVENTS)); 3610 evd_rp->evd_newevents |= evtq->eel_event_type; 3611 } else { 3612 evd_rp->evd_newevents |= DAPLKA_EVD_ULAND_EVENTS; 3613 } 3614 waiters = evd_rp->evd_waiters; 3615 cv_broadcast(&evd_rp->evd_cv); 3616 mutex_exit(&evd_rp->evd_lock); 3617 3618 /* 3619 * only wakeup the CNO if there are no waiters on this evd. 3620 */ 3621 if (evd_rp->evd_cno_res != NULL && waiters == 0) { 3622 mutex_enter(&evd_rp->evd_cno_res->cno_lock); 3623 evd_rp->evd_cno_res->cno_evd_cookie = evd_rp->evd_cookie; 3624 cv_broadcast(&evd_rp->evd_cno_res->cno_cv); 3625 mutex_exit(&evd_rp->evd_cno_res->cno_lock); 3626 } 3627 } 3628 3629 /* 3630 * daplka_evd_event_enqueue adds elem to the end of the event list 3631 * The caller is expected to acquire appropriate locks before 3632 * calling enqueue 3633 */ 3634 static void 3635 daplka_evd_event_enqueue(daplka_evd_event_list_t *evlist, 3636 daplka_evd_event_t *elem) 3637 { 3638 if (evlist->eel_tail) { 3639 evlist->eel_tail->ee_next = elem; 3640 evlist->eel_tail = elem; 3641 } else { 3642 /* list is empty */ 3643 ASSERT(evlist->eel_head == NULL); 3644 evlist->eel_head = elem; 3645 evlist->eel_tail = elem; 3646 } 3647 evlist->eel_num_elements++; 3648 } 3649 3650 /* 3651 * daplka_evd_event_dequeue removes and returns the first element of event 3652 * list. NULL is returned if the list is empty. The caller is expected to 3653 * acquire appropriate locks before calling enqueue. 3654 */ 3655 static daplka_evd_event_t * 3656 daplka_evd_event_dequeue(daplka_evd_event_list_t *evlist) 3657 { 3658 daplka_evd_event_t *head; 3659 3660 head = evlist->eel_head; 3661 if (head == NULL) { 3662 return (NULL); 3663 } 3664 3665 evlist->eel_head = head->ee_next; 3666 evlist->eel_num_elements--; 3667 /* if it was the last element update the tail pointer too */ 3668 if (evlist->eel_head == NULL) { 3669 ASSERT(evlist->eel_num_elements == 0); 3670 evlist->eel_tail = NULL; 3671 } 3672 return (head); 3673 } 3674 3675 /* 3676 * A CNO allows the client to wait for notifications from multiple EVDs. 3677 * To use a CNO, the client needs to follow the procedure below: 3678 * 1. allocate a CNO. this returns a cno_hkey that identifies the CNO. 3679 * 2. create one or more EVDs using the returned cno_hkey. 3680 * 3. call cno_wait. when one of the associated EVDs get notified, the 3681 * CNO will also get notified. cno_wait will then return with a 3682 * evd_cookie identifying the EVD that triggered the event. 3683 * 3684 * A note about cno_wait: 3685 * -unlike a EVD, a CNO does not maintain a queue of notifications. For 3686 * example, suppose multiple EVDs triggered a CNO before the client calls 3687 * cno_wait; when the client calls cno_wait, it will return with the 3688 * evd_cookie that identifies the *last* EVD that triggered the CNO. It 3689 * is the responsibility of the client, upon returning from cno_wait, to 3690 * check on all EVDs that can potentially trigger the CNO. the returned 3691 * evd_cookie is only meant to be a hint. there is no guarantee that the 3692 * EVD identified by the evd_cookie still contains an event or still 3693 * exists by the time cno_wait returns. 3694 */ 3695 3696 /* 3697 * allocates a CNO. 3698 * the returned cno_hkey may subsequently be used in evd_create. 3699 */ 3700 /* ARGSUSED */ 3701 static int 3702 daplka_cno_alloc(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 3703 cred_t *cred, int *rvalp) 3704 { 3705 dapl_cno_alloc_t args; 3706 daplka_cno_resource_t *cno_rp = NULL; 3707 uint64_t cno_hkey = 0; 3708 boolean_t inserted = B_FALSE; 3709 int retval = 0; 3710 3711 cno_rp = kmem_zalloc(sizeof (*cno_rp), daplka_km_flags); 3712 if (cno_rp == NULL) { 3713 DERR("cno_alloc: cannot allocate cno resource\n"); 3714 return (ENOMEM); 3715 } 3716 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*cno_rp)) 3717 DAPLKA_RS_INIT(cno_rp, DAPL_TYPE_CNO, 3718 DAPLKA_RS_RNUM(ia_rp), daplka_cno_destroy); 3719 3720 mutex_init(&cno_rp->cno_lock, NULL, MUTEX_DRIVER, NULL); 3721 cv_init(&cno_rp->cno_cv, NULL, CV_DRIVER, NULL); 3722 cno_rp->cno_evd_cookie = 0; 3723 3724 /* insert into cno hash table */ 3725 retval = daplka_hash_insert(&ia_rp->ia_cno_htbl, 3726 &cno_hkey, (void *)cno_rp); 3727 if (retval != 0) { 3728 DERR("cno_alloc: cannot insert cno resource\n"); 3729 goto cleanup; 3730 } 3731 inserted = B_TRUE; 3732 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(*cno_rp)) 3733 3734 /* return hkey to library */ 3735 args.cno_hkey = cno_hkey; 3736 3737 retval = ddi_copyout(&args, (void *)arg, sizeof (dapl_cno_alloc_t), 3738 mode); 3739 if (retval != 0) { 3740 DERR("cno_alloc: copyout error %d\n", retval); 3741 retval = EFAULT; 3742 goto cleanup; 3743 } 3744 return (0); 3745 3746 cleanup:; 3747 if (inserted) { 3748 daplka_cno_resource_t *free_rp = NULL; 3749 3750 (void) daplka_hash_remove(&ia_rp->ia_cno_htbl, cno_hkey, 3751 (void **)&free_rp); 3752 if (free_rp != cno_rp) { 3753 DERR("cno_alloc: cannot remove cno\n"); 3754 /* 3755 * we can only get here if another thread 3756 * has completed the cleanup in cno_free 3757 */ 3758 return (retval); 3759 } 3760 } 3761 DAPLKA_RS_UNREF(cno_rp); 3762 return (retval); 3763 } 3764 3765 /* 3766 * destroys a CNO. 3767 * this gets called when a CNO resource's refcnt drops to zero. 3768 */ 3769 static int 3770 daplka_cno_destroy(daplka_resource_t *gen_rp) 3771 { 3772 daplka_cno_resource_t *cno_rp = (daplka_cno_resource_t *)gen_rp; 3773 3774 ASSERT(DAPLKA_RS_REFCNT(cno_rp) == 0); 3775 D2("cno_destroy: entering, cno_rp %p, rnum %d\n", 3776 cno_rp, DAPLKA_RS_RNUM(cno_rp)); 3777 3778 ASSERT(DAPLKA_RS_TYPE(cno_rp) == DAPL_TYPE_CNO); 3779 cv_destroy(&cno_rp->cno_cv); 3780 mutex_destroy(&cno_rp->cno_lock); 3781 3782 DAPLKA_RS_FINI(cno_rp); 3783 kmem_free(cno_rp, sizeof (daplka_cno_resource_t)); 3784 D2("cno_destroy: exiting, cno_rp %p\n", cno_rp); 3785 return (0); 3786 } 3787 3788 static void 3789 daplka_hash_cno_free(void *obj) 3790 { 3791 daplka_cno_resource_t *cno_rp = (daplka_cno_resource_t *)obj; 3792 3793 ASSERT(DAPLKA_RS_TYPE(cno_rp) == DAPL_TYPE_CNO); 3794 DAPLKA_RS_UNREF(cno_rp); 3795 } 3796 3797 /* 3798 * removes the CNO from the cno hash table and frees the CNO 3799 * if there are no references to it. if there are references to 3800 * it, the CNO will be destroyed when the last of the references 3801 * is released. once the CNO is removed from the cno hash table, 3802 * the client will no longer be able to call cno_wait on the CNO. 3803 */ 3804 /* ARGSUSED */ 3805 static int 3806 daplka_cno_free(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 3807 cred_t *cred, int *rvalp) 3808 { 3809 daplka_cno_resource_t *cno_rp = NULL; 3810 dapl_cno_free_t args; 3811 int retval = 0; 3812 3813 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_cno_free_t), mode); 3814 if (retval != 0) { 3815 DERR("cno_free: copyin error %d\n", retval); 3816 return (EINVAL); 3817 } 3818 3819 retval = daplka_hash_remove(&ia_rp->ia_cno_htbl, 3820 args.cnf_hkey, (void **)&cno_rp); 3821 if (retval != 0 || cno_rp == NULL) { 3822 DERR("cno_free: cannot find cno resource\n"); 3823 return (EINVAL); 3824 } 3825 ASSERT(DAPLKA_RS_TYPE(cno_rp) == DAPL_TYPE_CNO); 3826 3827 /* UNREF calls the actual free function when refcnt is zero */ 3828 DAPLKA_RS_UNREF(cno_rp); 3829 return (0); 3830 } 3831 3832 /* 3833 * wait for a notification from one of the associated EVDs. 3834 */ 3835 /* ARGSUSED */ 3836 static int 3837 daplka_cno_wait(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 3838 cred_t *cred, int *rvalp) 3839 { 3840 daplka_cno_resource_t *cno_rp = NULL; 3841 dapl_cno_wait_t args; 3842 int retval = 0; 3843 uint64_t evd_cookie = 0; 3844 clock_t timeout, curr_time; 3845 3846 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_cno_wait_t), mode); 3847 if (retval != 0) { 3848 DERR("cno_wait: copyin error %d\n", retval); 3849 return (EINVAL); 3850 } 3851 /* get cno resource */ 3852 cno_rp = (daplka_cno_resource_t *) 3853 daplka_hash_lookup(&ia_rp->ia_cno_htbl, args.cnw_hkey); 3854 if (cno_rp == NULL) { 3855 DERR("cno_wait: cannot find cno resource\n"); 3856 return (EINVAL); 3857 } 3858 ASSERT(DAPLKA_RS_TYPE(cno_rp) == DAPL_TYPE_CNO); 3859 3860 curr_time = ddi_get_lbolt(); 3861 timeout = curr_time + drv_usectohz(args.cnw_timeout); 3862 3863 /* 3864 * use the max value if we wrapped around 3865 */ 3866 if (args.cnw_timeout > 0 && timeout <= curr_time) { 3867 /* 3868 * clock_t (size long) changes between 32 and 64-bit kernels 3869 */ 3870 timeout = LONG_MAX >> 4; 3871 } 3872 mutex_enter(&cno_rp->cno_lock); 3873 while (cno_rp->cno_evd_cookie == 0) { 3874 int rval = 0; 3875 3876 rval = cv_timedwait_sig(&cno_rp->cno_cv, 3877 &cno_rp->cno_lock, timeout); 3878 if (rval == 0) { 3879 DERR("cno_wait: interrupted\n"); 3880 mutex_exit(&cno_rp->cno_lock); 3881 retval = EINTR; 3882 goto cleanup; 3883 } else if (rval == -1) { 3884 DERR("cno_wait: timed out\n"); 3885 mutex_exit(&cno_rp->cno_lock); 3886 retval = ETIME; 3887 goto cleanup; 3888 } 3889 } 3890 evd_cookie = cno_rp->cno_evd_cookie; 3891 cno_rp->cno_evd_cookie = 0; 3892 mutex_exit(&cno_rp->cno_lock); 3893 3894 ASSERT(evd_cookie != 0); 3895 D2("cno_wait: returning evd_cookie 0x%p\n", 3896 (void *)(uintptr_t)evd_cookie); 3897 args.cnw_evd_cookie = evd_cookie; 3898 retval = ddi_copyout((void *)&args, (void *)arg, 3899 sizeof (dapl_cno_wait_t), mode); 3900 if (retval != 0) { 3901 DERR("cno_wait: copyout error %d\n", retval); 3902 retval = EFAULT; 3903 goto cleanup; 3904 } 3905 3906 cleanup:; 3907 if (cno_rp != NULL) { 3908 DAPLKA_RS_UNREF(cno_rp); 3909 } 3910 return (retval); 3911 } 3912 3913 /* 3914 * this function is called by the client when it decides to 3915 * accept a connection request. a connection request is generated 3916 * when the active side generates REQ MAD to a service point on 3917 * the destination node. this causes the CM service handler 3918 * (daplka_cm_service_req) on the passive side to be callee. This 3919 * handler will then enqueue this connection request to the backlog 3920 * array of the service point. A connection event containing the 3921 * backlog array index and connection request private data is passed 3922 * to the client's service point EVD (sp_evd_res). once the event 3923 * is passed up to the userland, the client may examine the request 3924 * to decide whether to call daplka_cr_accept or dapka_cr_reject. 3925 */ 3926 /* ARGSUSED */ 3927 static int 3928 daplka_cr_accept(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 3929 cred_t *cred, int *rvalp) 3930 { 3931 daplka_ep_resource_t *ep_rp = NULL; 3932 daplka_sp_resource_t *sp_rp = NULL; 3933 dapl_cr_accept_t args; 3934 daplka_sp_conn_pend_t *conn; 3935 ibt_cm_proceed_reply_t proc_reply; 3936 ibt_status_t status; 3937 uint16_t bkl_index; 3938 uint32_t old_state, new_state; 3939 int retval = 0; 3940 void *priv_data = NULL, *sid; 3941 3942 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_cr_accept_t), 3943 mode); 3944 if (retval != 0) { 3945 DERR("cr_accept: copyin error %d\n", retval); 3946 return (EFAULT); 3947 } 3948 if (args.cra_priv_sz > DAPL_MAX_PRIVATE_DATA_SIZE) { 3949 DERR("cr_accept: private data len (%d) exceeded " 3950 "max size %d\n", args.cra_priv_sz, 3951 DAPL_MAX_PRIVATE_DATA_SIZE); 3952 return (EINVAL); 3953 } 3954 priv_data = (args.cra_priv_sz > 0) ? (void *)args.cra_priv : NULL; 3955 3956 D2("cr_accept: priv(0x%p) priv_len(%u) psep(0x%llx)\n", priv_data, 3957 args.cra_priv_sz, (longlong_t)args.cra_bkl_cookie); 3958 3959 /* get sp resource */ 3960 sp_rp = (daplka_sp_resource_t *)daplka_hash_lookup(&ia_rp->ia_sp_htbl, 3961 args.cra_sp_hkey); 3962 if (sp_rp == NULL) { 3963 DERR("cr_accept: cannot find sp resource\n"); 3964 return (EINVAL); 3965 } 3966 ASSERT(DAPLKA_RS_TYPE(sp_rp) == DAPL_TYPE_SP); 3967 3968 /* get ep resource */ 3969 ep_rp = (daplka_ep_resource_t *)daplka_hash_lookup(&ia_rp->ia_ep_htbl, 3970 args.cra_ep_hkey); 3971 if (ep_rp == NULL) { 3972 DERR("cr_accept: cannot find ep resource\n"); 3973 retval = EINVAL; 3974 goto cleanup; 3975 } 3976 ASSERT(DAPLKA_RS_TYPE(ep_rp) == DAPL_TYPE_EP); 3977 3978 /* 3979 * accept is only allowed if ep_state is CLOSED. 3980 * note that after this point, the ep_state is frozen 3981 * (i.e. TRANSITIONING) until we transition ep_state 3982 * to ACCEPTING or back to CLOSED if we get an error. 3983 */ 3984 new_state = old_state = daplka_ep_get_state(ep_rp); 3985 if (old_state != DAPLKA_EP_STATE_CLOSED) { 3986 DERR("cr_accept: invalid ep state %d\n", old_state); 3987 retval = EINVAL; 3988 goto cleanup; 3989 } 3990 3991 mutex_enter(&sp_rp->sp_lock); 3992 bkl_index = DAPLKA_GET_PSEP_INDEX(args.cra_bkl_cookie); 3993 /* 3994 * make sure the backlog index is not bogus. 3995 */ 3996 if (bkl_index >= sp_rp->sp_backlog_size) { 3997 DERR("cr_accept: invalid backlog index 0x%llx %d\n", 3998 (longlong_t)args.cra_bkl_cookie, bkl_index); 3999 mutex_exit(&sp_rp->sp_lock); 4000 retval = EINVAL; 4001 goto cleanup; 4002 } 4003 /* 4004 * make sure the backlog index indeed refers 4005 * to a pending connection. 4006 */ 4007 conn = &sp_rp->sp_backlog[bkl_index]; 4008 if (conn->spcp_state != DAPLKA_SPCP_PENDING) { 4009 DERR("cr_accept: invalid conn state %d\n", 4010 conn->spcp_state); 4011 mutex_exit(&sp_rp->sp_lock); 4012 retval = EINVAL; 4013 goto cleanup; 4014 } 4015 if (conn->spcp_sid == NULL) { 4016 DERR("cr_accept: sid == NULL\n"); 4017 mutex_exit(&sp_rp->sp_lock); 4018 retval = EINVAL; 4019 goto cleanup; 4020 } 4021 if (ep_rp->ep_chan_hdl == NULL) { 4022 /* 4023 * a ep_rp with a NULL chan_hdl is impossible. 4024 */ 4025 DERR("cr_accept: ep_chan_hdl == NULL\n"); 4026 mutex_exit(&sp_rp->sp_lock); 4027 ASSERT(B_FALSE); 4028 retval = EINVAL; 4029 goto cleanup; 4030 } 4031 proc_reply.rep.cm_channel = ep_rp->ep_chan_hdl; 4032 proc_reply.rep.cm_rdma_ra_out = conn->spcp_rdma_ra_out; 4033 proc_reply.rep.cm_rdma_ra_in = conn->spcp_rdma_ra_in; 4034 proc_reply.rep.cm_rnr_retry_cnt = IBT_RNR_INFINITE_RETRY; 4035 sid = conn->spcp_sid; 4036 4037 /* 4038 * this clears our slot in the backlog array. 4039 * this slot may now be used by other pending connections. 4040 */ 4041 conn->spcp_sid = NULL; 4042 conn->spcp_state = DAPLKA_SPCP_INIT; 4043 conn->spcp_req_len = 0; 4044 mutex_exit(&sp_rp->sp_lock); 4045 4046 /* 4047 * Set the unique cookie corresponding to the CR to this EP 4048 * so that is can be used in passive side CM callbacks 4049 */ 4050 ep_rp->ep_psep_cookie = args.cra_bkl_cookie; 4051 4052 status = ibt_cm_proceed(IBT_CM_EVENT_REQ_RCV, sid, IBT_CM_ACCEPT, 4053 &proc_reply, priv_data, (ibt_priv_data_len_t)args.cra_priv_sz); 4054 4055 if (status != IBT_SUCCESS) { 4056 DERR("cr_accept: ibt_cm_proceed returned %d\n", status); 4057 *rvalp = (int)status; 4058 retval = 0; 4059 } 4060 /* 4061 * note that the CM handler may actually be called at this 4062 * point. but since ep_state is still in TRANSITIONING, the 4063 * handler will wait until we transition to ACCEPTING. this 4064 * prevents the case where we set ep_state to ACCEPTING after 4065 * daplka_service_conn_est sets ep_state to CONNECTED. 4066 */ 4067 new_state = DAPLKA_EP_STATE_ACCEPTING; 4068 4069 cleanup:; 4070 if (sp_rp != NULL) { 4071 DAPLKA_RS_UNREF(sp_rp); 4072 } 4073 if (ep_rp != NULL) { 4074 daplka_ep_set_state(ep_rp, old_state, new_state); 4075 DAPLKA_RS_UNREF(ep_rp); 4076 } 4077 return (retval); 4078 } 4079 4080 /* 4081 * this function is called by the client to reject a 4082 * connection request. 4083 */ 4084 /* ARGSUSED */ 4085 static int 4086 daplka_cr_reject(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 4087 cred_t *cred, int *rvalp) 4088 { 4089 dapl_cr_reject_t args; 4090 daplka_sp_resource_t *sp_rp = NULL; 4091 daplka_sp_conn_pend_t *conn; 4092 ibt_cm_proceed_reply_t proc_reply; 4093 ibt_cm_status_t proc_status; 4094 ibt_status_t status; 4095 uint16_t bkl_index; 4096 int retval = 0; 4097 void *sid; 4098 4099 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_cr_reject_t), 4100 mode); 4101 if (retval != 0) { 4102 DERR("cr_reject: copyin error %d\n", retval); 4103 return (EFAULT); 4104 } 4105 /* get sp resource */ 4106 sp_rp = (daplka_sp_resource_t *)daplka_hash_lookup(&ia_rp->ia_sp_htbl, 4107 args.crr_sp_hkey); 4108 if (sp_rp == NULL) { 4109 DERR("cr_reject: cannot find sp resource\n"); 4110 return (EINVAL); 4111 } 4112 ASSERT(DAPLKA_RS_TYPE(sp_rp) == DAPL_TYPE_SP); 4113 4114 D2("cr_reject: psep(0x%llx)\n", (longlong_t)args.crr_bkl_cookie); 4115 4116 mutex_enter(&sp_rp->sp_lock); 4117 bkl_index = DAPLKA_GET_PSEP_INDEX(args.crr_bkl_cookie); 4118 /* 4119 * make sure the backlog index is not bogus. 4120 */ 4121 if (bkl_index >= sp_rp->sp_backlog_size) { 4122 DERR("cr_reject: invalid backlog index 0x%llx %d\n", 4123 (longlong_t)args.crr_bkl_cookie, bkl_index); 4124 mutex_exit(&sp_rp->sp_lock); 4125 retval = EINVAL; 4126 goto cleanup; 4127 } 4128 /* 4129 * make sure the backlog index indeed refers 4130 * to a pending connection. 4131 */ 4132 conn = &sp_rp->sp_backlog[bkl_index]; 4133 if (conn->spcp_state != DAPLKA_SPCP_PENDING) { 4134 DERR("cr_reject: invalid conn state %d\n", 4135 conn->spcp_state); 4136 mutex_exit(&sp_rp->sp_lock); 4137 retval = EINVAL; 4138 goto cleanup; 4139 } 4140 if (conn->spcp_sid == NULL) { 4141 DERR("cr_reject: sid == NULL\n"); 4142 mutex_exit(&sp_rp->sp_lock); 4143 retval = EINVAL; 4144 goto cleanup; 4145 } 4146 bzero(&proc_reply, sizeof (proc_reply)); 4147 sid = conn->spcp_sid; 4148 4149 /* 4150 * this clears our slot in the backlog array. 4151 * this slot may now be used by other pending connections. 4152 */ 4153 conn->spcp_sid = NULL; 4154 conn->spcp_state = DAPLKA_SPCP_INIT; 4155 conn->spcp_req_len = 0; 4156 4157 switch (args.crr_reason) { 4158 case DAPL_IB_CM_REJ_REASON_CONSUMER_REJ: 4159 /* results in IBT_CM_CONSUMER as the reason for reject */ 4160 proc_status = IBT_CM_REJECT; 4161 break; 4162 case DAPL_IB_CME_LOCAL_FAILURE: 4163 /*FALLTHRU*/ 4164 case DAPL_IB_CME_DESTINATION_UNREACHABLE: 4165 /* results in IBT_CM_NO_RESC as the reason for reject */ 4166 proc_status = IBT_CM_NO_RESOURCE; 4167 break; 4168 default: 4169 /* unexpect reason code */ 4170 ASSERT(!"unexpected reject reason code"); 4171 proc_status = IBT_CM_NO_RESOURCE; 4172 break; 4173 } 4174 4175 mutex_exit(&sp_rp->sp_lock); 4176 4177 status = ibt_cm_proceed(IBT_CM_EVENT_REQ_RCV, sid, proc_status, 4178 &proc_reply, NULL, 0); 4179 4180 if (status != IBT_SUCCESS) { 4181 DERR("cr_reject: ibt_cm_proceed returned %d\n", status); 4182 *rvalp = (int)status; 4183 retval = 0; 4184 } 4185 4186 cleanup:; 4187 if (sp_rp != NULL) { 4188 DAPLKA_RS_UNREF(sp_rp); 4189 } 4190 return (retval); 4191 } 4192 4193 4194 /* 4195 * daplka_sp_match is used by daplka_hash_walk for finding SPs 4196 */ 4197 typedef struct daplka_sp_match_s { 4198 uint64_t spm_conn_qual; 4199 daplka_sp_resource_t *spm_sp_rp; 4200 } daplka_sp_match_t; 4201 _NOTE(SCHEME_PROTECTS_DATA("daplka", daplka_sp_match_s::spm_sp_rp)) 4202 4203 static int 4204 daplka_sp_match(void *objp, void *arg) 4205 { 4206 daplka_sp_resource_t *sp_rp = (daplka_sp_resource_t *)objp; 4207 4208 ASSERT(DAPLKA_RS_TYPE(sp_rp) == DAPL_TYPE_SP); 4209 if (sp_rp->sp_conn_qual == 4210 ((daplka_sp_match_t *)arg)->spm_conn_qual) { 4211 ((daplka_sp_match_t *)arg)->spm_sp_rp = sp_rp; 4212 D2("daplka_sp_match: found sp, conn_qual %016llu\n", 4213 (longlong_t)((daplka_sp_match_t *)arg)->spm_conn_qual); 4214 DAPLKA_RS_REF(sp_rp); 4215 return (1); 4216 } 4217 return (0); 4218 } 4219 4220 /* 4221 * cr_handoff allows the client to handoff a connection request from 4222 * one service point to another. 4223 */ 4224 /* ARGSUSED */ 4225 static int 4226 daplka_cr_handoff(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 4227 cred_t *cred, int *rvalp) 4228 { 4229 dapl_cr_handoff_t args; 4230 daplka_sp_resource_t *sp_rp = NULL, *new_sp_rp = NULL; 4231 daplka_sp_conn_pend_t *conn; 4232 daplka_sp_match_t sp_match; 4233 ibt_cm_event_t fake_event; 4234 ibt_cm_status_t cm_status; 4235 ibt_status_t status; 4236 uint16_t bkl_index; 4237 void *sid, *priv = NULL; 4238 int retval = 0, priv_len = 0; 4239 4240 D3("cr_handoff: entering\n"); 4241 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_cr_handoff_t), 4242 mode); 4243 if (retval != 0) { 4244 DERR("cr_handoff: copyin error %d\n", retval); 4245 return (EFAULT); 4246 } 4247 /* get sp resource */ 4248 sp_rp = (daplka_sp_resource_t *)daplka_hash_lookup(&ia_rp->ia_sp_htbl, 4249 args.crh_sp_hkey); 4250 if (sp_rp == NULL) { 4251 DERR("cr_handoff: cannot find sp resource\n"); 4252 return (EINVAL); 4253 } 4254 ASSERT(DAPLKA_RS_TYPE(sp_rp) == DAPL_TYPE_SP); 4255 4256 /* 4257 * find the destination service point. 4258 */ 4259 sp_match.spm_conn_qual = args.crh_conn_qual; 4260 sp_match.spm_sp_rp = NULL; 4261 daplka_hash_walk(&daplka_global_sp_htbl, daplka_sp_match, 4262 (void *)&sp_match, RW_READER); 4263 4264 /* 4265 * return if we cannot find the service point 4266 */ 4267 if (sp_match.spm_sp_rp == NULL) { 4268 DERR("cr_handoff: new sp not found, conn qual = %llu\n", 4269 (longlong_t)args.crh_conn_qual); 4270 retval = EINVAL; 4271 goto cleanup; 4272 } 4273 new_sp_rp = sp_match.spm_sp_rp; 4274 4275 /* 4276 * the spec does not discuss the security implications of this 4277 * function. to be safe, we currently only allow processes 4278 * owned by the same user to handoff connection requests 4279 * to each other. 4280 */ 4281 if (crgetruid(cred) != new_sp_rp->sp_ruid) { 4282 DERR("cr_handoff: permission denied\n"); 4283 retval = EPERM; 4284 goto cleanup; 4285 } 4286 4287 D2("cr_handoff: psep(0x%llx)\n", (longlong_t)args.crh_bkl_cookie); 4288 4289 mutex_enter(&sp_rp->sp_lock); 4290 bkl_index = DAPLKA_GET_PSEP_INDEX(args.crh_bkl_cookie); 4291 /* 4292 * make sure the backlog index is not bogus. 4293 */ 4294 if (bkl_index >= sp_rp->sp_backlog_size) { 4295 DERR("cr_handoff: invalid backlog index 0x%llx %d\n", 4296 (longlong_t)args.crh_bkl_cookie, bkl_index); 4297 mutex_exit(&sp_rp->sp_lock); 4298 retval = EINVAL; 4299 goto cleanup; 4300 } 4301 /* 4302 * make sure the backlog index indeed refers 4303 * to a pending connection. 4304 */ 4305 conn = &sp_rp->sp_backlog[bkl_index]; 4306 if (conn->spcp_state != DAPLKA_SPCP_PENDING) { 4307 DERR("cr_handoff: invalid conn state %d\n", 4308 conn->spcp_state); 4309 mutex_exit(&sp_rp->sp_lock); 4310 retval = EINVAL; 4311 goto cleanup; 4312 } 4313 if (conn->spcp_sid == NULL) { 4314 DERR("cr_handoff: sid == NULL\n"); 4315 mutex_exit(&sp_rp->sp_lock); 4316 retval = EINVAL; 4317 goto cleanup; 4318 } 4319 sid = conn->spcp_sid; 4320 priv = NULL; 4321 priv_len = conn->spcp_req_len; 4322 if (priv_len > 0) { 4323 priv = kmem_zalloc(priv_len, daplka_km_flags); 4324 if (priv == NULL) { 4325 mutex_exit(&sp_rp->sp_lock); 4326 retval = ENOMEM; 4327 goto cleanup; 4328 } 4329 bcopy(conn->spcp_req_data, priv, priv_len); 4330 } 4331 /* 4332 * this clears our slot in the backlog array. 4333 * this slot may now be used by other pending connections. 4334 */ 4335 conn->spcp_sid = NULL; 4336 conn->spcp_state = DAPLKA_SPCP_INIT; 4337 conn->spcp_req_len = 0; 4338 mutex_exit(&sp_rp->sp_lock); 4339 4340 /* fill fake_event and call service_req handler */ 4341 bzero(&fake_event, sizeof (fake_event)); 4342 fake_event.cm_type = IBT_CM_EVENT_REQ_RCV; 4343 fake_event.cm_session_id = sid; 4344 fake_event.cm_priv_data_len = priv_len; 4345 fake_event.cm_priv_data = priv; 4346 4347 cm_status = daplka_cm_service_req(new_sp_rp, 4348 &fake_event, NULL, priv, (ibt_priv_data_len_t)priv_len); 4349 if (cm_status != IBT_CM_DEFER) { 4350 ibt_cm_proceed_reply_t proc_reply; 4351 4352 DERR("cr_handoff: service_req returned %d\n", cm_status); 4353 /* 4354 * if for some reason cm_service_req failed, we 4355 * reject the connection. 4356 */ 4357 bzero(&proc_reply, sizeof (proc_reply)); 4358 4359 status = ibt_cm_proceed(IBT_CM_EVENT_REQ_RCV, sid, 4360 IBT_CM_NO_RESOURCE, &proc_reply, NULL, 0); 4361 if (status != IBT_SUCCESS) { 4362 DERR("cr_handoff: ibt_cm_proceed returned %d\n", 4363 status); 4364 } 4365 *rvalp = (int)status; 4366 retval = 0; 4367 } 4368 4369 cleanup:; 4370 if (priv_len > 0 && priv != NULL) { 4371 kmem_free(priv, priv_len); 4372 } 4373 if (new_sp_rp != NULL) { 4374 DAPLKA_RS_UNREF(new_sp_rp); 4375 } 4376 if (sp_rp != NULL) { 4377 DAPLKA_RS_UNREF(sp_rp); 4378 } 4379 D3("cr_handoff: exiting\n"); 4380 return (retval); 4381 } 4382 4383 /* 4384 * returns a list of hca attributes 4385 */ 4386 /* ARGSUSED */ 4387 static int 4388 daplka_ia_query(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 4389 cred_t *cred, int *rvalp) 4390 { 4391 dapl_ia_query_t args; 4392 int retval; 4393 ibt_hca_attr_t *hcap; 4394 4395 hcap = &ia_rp->ia_hca->hca_attr; 4396 4397 /* 4398 * Take the ibt_hca_attr_t and stuff them into dapl_hca_attr_t 4399 */ 4400 args.hca_attr.dhca_vendor_id = hcap->hca_vendor_id; 4401 args.hca_attr.dhca_device_id = hcap->hca_device_id; 4402 args.hca_attr.dhca_version_id = hcap->hca_version_id; 4403 args.hca_attr.dhca_max_chans = hcap->hca_max_chans; 4404 args.hca_attr.dhca_max_chan_sz = hcap->hca_max_chan_sz; 4405 args.hca_attr.dhca_max_sgl = hcap->hca_max_sgl; 4406 args.hca_attr.dhca_max_cq = hcap->hca_max_cq; 4407 args.hca_attr.dhca_max_cq_sz = hcap->hca_max_cq_sz; 4408 args.hca_attr.dhca_max_memr = hcap->hca_max_memr; 4409 args.hca_attr.dhca_max_memr_len = hcap->hca_max_memr_len; 4410 args.hca_attr.dhca_max_mem_win = hcap->hca_max_mem_win; 4411 args.hca_attr.dhca_max_rdma_in_chan = hcap->hca_max_rdma_in_chan; 4412 args.hca_attr.dhca_max_rdma_out_chan = hcap->hca_max_rdma_out_chan; 4413 args.hca_attr.dhca_max_partitions = hcap->hca_max_partitions; 4414 args.hca_attr.dhca_nports = hcap->hca_nports; 4415 args.hca_attr.dhca_node_guid = hcap->hca_node_guid; 4416 args.hca_attr.dhca_max_pd = hcap->hca_max_pd; 4417 args.hca_attr.dhca_max_srqs = hcap->hca_max_srqs; 4418 args.hca_attr.dhca_max_srqs_sz = hcap->hca_max_srqs_sz; 4419 args.hca_attr.dhca_max_srq_sgl = hcap->hca_max_srq_sgl; 4420 4421 retval = ddi_copyout(&args, (void *)arg, sizeof (dapl_ia_query_t), 4422 mode); 4423 if (retval != 0) { 4424 DERR("ia_query: copyout error %d\n", retval); 4425 return (EFAULT); 4426 } 4427 return (0); 4428 } 4429 4430 /* 4431 * This routine is passed to hash walk in the daplka_pre_mr_cleanup_callback, 4432 * it frees the mw embedded in the mw resource object. 4433 */ 4434 4435 /* ARGSUSED */ 4436 static int 4437 daplka_mr_cb_freemw(void *objp, void *arg) 4438 { 4439 daplka_mw_resource_t *mw_rp = (daplka_mw_resource_t *)objp; 4440 ibt_mw_hdl_t mw_hdl; 4441 ibt_status_t status; 4442 4443 D3("mr_cb_freemw: entering, mw_rp 0x%p\n", mw_rp); 4444 DAPLKA_RS_REF(mw_rp); 4445 4446 mutex_enter(&mw_rp->mw_lock); 4447 mw_hdl = mw_rp->mw_hdl; 4448 /* 4449 * we set mw_hdl to NULL so it won't get freed again 4450 */ 4451 mw_rp->mw_hdl = NULL; 4452 mutex_exit(&mw_rp->mw_lock); 4453 4454 if (mw_hdl != NULL) { 4455 status = daplka_ibt_free_mw(mw_rp, mw_rp->mw_hca_hdl, mw_hdl); 4456 if (status != IBT_SUCCESS) { 4457 DERR("mr_cb_freemw: ibt_free_mw returned %d\n", status); 4458 } 4459 D3("mr_cb_freemw: mw freed\n"); 4460 } 4461 4462 DAPLKA_RS_UNREF(mw_rp); 4463 return (0); 4464 } 4465 4466 /* 4467 * This routine is called from HCA driver's umem lock undo callback 4468 * when the memory associated with an MR is being unmapped. In this callback 4469 * we free all the MW associated with the IA and post an unaffiliated 4470 * async event to tell the app that there was a catastrophic event. 4471 * This allows the HCA to deregister the MR in its callback processing. 4472 */ 4473 static void 4474 daplka_pre_mr_cleanup_callback(void *arg1, void *arg2 /*ARGSUSED*/) 4475 { 4476 daplka_mr_resource_t *mr_rp; 4477 daplka_ia_resource_t *ia_rp; 4478 #ifdef _THROW_ASYNC_EVENT_FROM_MRUNLOCKCB 4479 ibt_async_event_t event; 4480 ibt_hca_attr_t *hca_attrp; 4481 #endif 4482 minor_t rnum; 4483 4484 mr_rp = (daplka_mr_resource_t *)arg1; 4485 rnum = DAPLKA_RS_RNUM(mr_rp); 4486 daplka_shared_mr_free(mr_rp); 4487 4488 ia_rp = (daplka_ia_resource_t *)daplka_resource_lookup(rnum); 4489 if (ia_rp == NULL) { 4490 DERR("daplka_mr_unlock_callback: resource not found, rnum %d\n", 4491 rnum); 4492 return; 4493 } 4494 4495 DERR("daplka_mr_unlock_callback: resource(%p) rnum(%d)\n", ia_rp, rnum); 4496 4497 mutex_enter(&ia_rp->ia_lock); 4498 /* 4499 * MW is being alloced OR MW freeze has already begun. In 4500 * both these cases we wait for that to complete before 4501 * continuing. 4502 */ 4503 while ((ia_rp->ia_state == DAPLKA_IA_MW_ALLOC_IN_PROGRESS) || 4504 (ia_rp->ia_state == DAPLKA_IA_MW_FREEZE_IN_PROGRESS)) { 4505 cv_wait(&ia_rp->ia_cv, &ia_rp->ia_lock); 4506 } 4507 4508 switch (ia_rp->ia_state) { 4509 case DAPLKA_IA_INIT: 4510 ia_rp->ia_state = DAPLKA_IA_MW_FREEZE_IN_PROGRESS; 4511 mutex_exit(&ia_rp->ia_lock); 4512 break; 4513 case DAPLKA_IA_MW_FROZEN: 4514 /* the mw on this ia have been freed */ 4515 D2("daplka_mr_unlock_callback: ia_state %d nothing to do\n", 4516 ia_rp->ia_state); 4517 mutex_exit(&ia_rp->ia_lock); 4518 goto cleanup; 4519 default: 4520 ASSERT(!"daplka_mr_unlock_callback: IA state invalid"); 4521 DERR("daplka_mr_unlock_callback: invalid ia_state %d\n", 4522 ia_rp->ia_state); 4523 mutex_exit(&ia_rp->ia_lock); 4524 goto cleanup; 4525 } 4526 4527 /* 4528 * Walk the mw hash table and free the mws. Acquire a writer 4529 * lock since we don't want anyone else traversing this tree 4530 * while we are freeing the MW. 4531 */ 4532 daplka_hash_walk(&ia_rp->ia_mw_htbl, daplka_mr_cb_freemw, NULL, 4533 RW_WRITER); 4534 4535 mutex_enter(&ia_rp->ia_lock); 4536 ASSERT(ia_rp->ia_state == DAPLKA_IA_MW_FREEZE_IN_PROGRESS); 4537 ia_rp->ia_state = DAPLKA_IA_MW_FROZEN; 4538 cv_broadcast(&ia_rp->ia_cv); 4539 mutex_exit(&ia_rp->ia_lock); 4540 4541 /* 4542 * Currently commented out because Oracle skgxp is incapable 4543 * of handling async events correctly. 4544 */ 4545 #ifdef _THROW_ASYNC_EVENT_FROM_MRUNLOCKCB 4546 /* 4547 * Enqueue an unaffiliated async error event to indicate this 4548 * IA has encountered a problem that caused the MW to freed up 4549 */ 4550 4551 /* Create a fake event, only relevant field is the hca_guid */ 4552 bzero(&event, sizeof (ibt_async_event_t)); 4553 hca_attrp = &ia_rp->ia_hca->hca_attr; 4554 event.ev_hca_guid = hca_attrp->hca_node_guid; 4555 4556 daplka_async_event_create(IBT_ERROR_LOCAL_CATASTROPHIC, &event, 0, 4557 ia_rp); 4558 #endif /* _THROW_ASYNC_EVENT_FROM_MRUNLOCKCB */ 4559 4560 cleanup:; 4561 D2("daplka_mr_unlock_callback: resource(%p) done\n", ia_rp); 4562 DAPLKA_RS_UNREF(ia_rp); 4563 } 4564 4565 /* 4566 * registers a memory region. 4567 * memory locking will be done by the HCA driver. 4568 */ 4569 /* ARGSUSED */ 4570 static int 4571 daplka_mr_register(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 4572 cred_t *cred, int *rvalp) 4573 { 4574 boolean_t inserted = B_FALSE; 4575 daplka_mr_resource_t *mr_rp; 4576 daplka_pd_resource_t *pd_rp; 4577 dapl_mr_register_t args; 4578 ibt_mr_data_in_t mr_cb_data_in; 4579 uint64_t mr_hkey = 0; 4580 ibt_status_t status; 4581 int retval; 4582 4583 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_mr_register_t), 4584 mode); 4585 if (retval != 0) { 4586 DERR("mr_register: copyin error %d\n", retval); 4587 return (EINVAL); 4588 } 4589 mr_rp = kmem_zalloc(sizeof (daplka_mr_resource_t), daplka_km_flags); 4590 if (mr_rp == NULL) { 4591 DERR("mr_register: cannot allocate mr resource\n"); 4592 return (ENOMEM); 4593 } 4594 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*mr_rp)) 4595 DAPLKA_RS_INIT(mr_rp, DAPL_TYPE_MR, 4596 DAPLKA_RS_RNUM(ia_rp), daplka_mr_destroy); 4597 4598 mutex_init(&mr_rp->mr_lock, NULL, MUTEX_DRIVER, NULL); 4599 mr_rp->mr_hca = ia_rp->ia_hca; 4600 mr_rp->mr_hca_hdl = ia_rp->ia_hca_hdl; 4601 mr_rp->mr_next = NULL; 4602 mr_rp->mr_shared_mr = NULL; 4603 4604 /* get pd handle */ 4605 pd_rp = (daplka_pd_resource_t *) 4606 daplka_hash_lookup(&ia_rp->ia_pd_htbl, args.mr_pd_hkey); 4607 if (pd_rp == NULL) { 4608 DERR("mr_register: cannot find pd resource\n"); 4609 retval = EINVAL; 4610 goto cleanup; 4611 } 4612 ASSERT(DAPLKA_RS_TYPE(pd_rp) == DAPL_TYPE_PD); 4613 mr_rp->mr_pd_res = pd_rp; 4614 4615 mr_rp->mr_attr.mr_vaddr = args.mr_vaddr; 4616 mr_rp->mr_attr.mr_len = args.mr_len; 4617 mr_rp->mr_attr.mr_as = curproc->p_as; 4618 mr_rp->mr_attr.mr_flags = args.mr_flags | IBT_MR_NOSLEEP; 4619 4620 D3("mr_register: mr_vaddr %p, mr_len %llu, mr_flags 0x%x\n", 4621 (void *)(uintptr_t)mr_rp->mr_attr.mr_vaddr, 4622 (longlong_t)mr_rp->mr_attr.mr_len, 4623 mr_rp->mr_attr.mr_flags); 4624 4625 status = daplka_ibt_register_mr(mr_rp, ia_rp->ia_hca_hdl, 4626 mr_rp->mr_pd_res->pd_hdl, &mr_rp->mr_attr, &mr_rp->mr_hdl, 4627 &mr_rp->mr_desc); 4628 4629 if (status != IBT_SUCCESS) { 4630 DERR("mr_register: ibt_register_mr error %d\n", status); 4631 *rvalp = (int)status; 4632 retval = 0; 4633 goto cleanup; 4634 } 4635 4636 mr_cb_data_in.mr_rev = IBT_MR_DATA_IN_IF_VERSION; 4637 mr_cb_data_in.mr_func = daplka_pre_mr_cleanup_callback; 4638 mr_cb_data_in.mr_arg1 = (void *)mr_rp; 4639 mr_cb_data_in.mr_arg2 = NULL; 4640 4641 /* Pass the service driver mr cleanup handler to the hca driver */ 4642 status = ibt_ci_data_in(ia_rp->ia_hca_hdl, 4643 IBT_CI_NO_FLAGS, IBT_HDL_MR, (void *)mr_rp->mr_hdl, 4644 &mr_cb_data_in, sizeof (mr_cb_data_in)); 4645 4646 if (status != IBT_SUCCESS) { 4647 DERR("mr_register: ibt_ci_data_in error(%d) ver(%d)", 4648 status, mr_cb_data_in.mr_rev); 4649 *rvalp = (int)status; 4650 retval = 0; 4651 goto cleanup; 4652 } 4653 4654 /* insert into mr hash table */ 4655 retval = daplka_hash_insert(&ia_rp->ia_mr_htbl, 4656 &mr_hkey, (void *)mr_rp); 4657 if (retval != 0) { 4658 DERR("mr_register: cannot insert mr resource into mr_htbl\n"); 4659 goto cleanup; 4660 } 4661 inserted = B_TRUE; 4662 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(*mr_rp)) 4663 4664 args.mr_lkey = mr_rp->mr_desc.md_lkey; 4665 args.mr_rkey = mr_rp->mr_desc.md_rkey; 4666 args.mr_hkey = mr_hkey; 4667 4668 retval = ddi_copyout((void *)&args, (void *)arg, 4669 sizeof (dapl_mr_register_t), mode); 4670 if (retval != 0) { 4671 DERR("mr_register: copyout error %d\n", retval); 4672 retval = EFAULT; 4673 goto cleanup; 4674 } 4675 return (0); 4676 4677 cleanup:; 4678 if (inserted) { 4679 daplka_mr_resource_t *free_rp = NULL; 4680 4681 (void) daplka_hash_remove(&ia_rp->ia_mr_htbl, mr_hkey, 4682 (void **)&free_rp); 4683 if (free_rp != mr_rp) { 4684 DERR("mr_register: cannot remove mr from hash table\n"); 4685 /* 4686 * we can only get here if another thread 4687 * has completed the cleanup in mr_deregister 4688 */ 4689 return (retval); 4690 } 4691 } 4692 DAPLKA_RS_UNREF(mr_rp); 4693 return (retval); 4694 } 4695 4696 /* 4697 * registers a shared memory region. 4698 * the client calls this function with the intention to share the memory 4699 * region with other clients. it is assumed that, prior to calling this 4700 * function, the client(s) are already sharing parts of their address 4701 * space using a mechanism such as SYSV shared memory. the first client 4702 * that calls this function will create and insert a daplka_shared_mr_t 4703 * object into the global daplka_shared_mr_tree. this shared mr object 4704 * will be identified by a unique 40-byte key and will maintain a list 4705 * of mr resources. every time this function gets called with the same 4706 * 40-byte key, a new mr resource (containing a new mr handle generated 4707 * by ibt_register_mr or ibt_register_shared_mr) is created and inserted 4708 * into this list. similarly, every time a shared mr gets deregistered 4709 * or invalidated by a callback, the mr resource gets removed from this 4710 * list. the shared mr object has a reference count. when it drops to 4711 * zero, the shared mr object will be removed from the global avl tree 4712 * and be freed. 4713 */ 4714 /* ARGSUSED */ 4715 static int 4716 daplka_mr_register_shared(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 4717 cred_t *cred, int *rvalp) 4718 { 4719 dapl_mr_register_shared_t args; 4720 daplka_shared_mr_t *smrp = NULL; 4721 daplka_shared_mr_t tmp_smr; 4722 ibt_mr_data_in_t mr_cb_data_in; 4723 avl_index_t where; 4724 boolean_t inserted = B_FALSE; 4725 daplka_mr_resource_t *mr_rp = NULL; 4726 daplka_pd_resource_t *pd_rp; 4727 uint64_t mr_hkey = 0; 4728 ibt_status_t status; 4729 int retval; 4730 4731 retval = ddi_copyin((void *)arg, &args, 4732 sizeof (dapl_mr_register_shared_t), mode); 4733 if (retval != 0) { 4734 DERR("mr_register_shared: copyin error %d\n", retval); 4735 return (EINVAL); 4736 } 4737 4738 mutex_enter(&daplka_shared_mr_lock); 4739 /* 4740 * find smrp from the global avl tree. 4741 * the 40-byte key is used as the lookup key. 4742 */ 4743 tmp_smr.smr_cookie = args.mrs_shm_cookie; 4744 smrp = (daplka_shared_mr_t *) 4745 avl_find(&daplka_shared_mr_tree, &tmp_smr, &where); 4746 if (smrp != NULL) { 4747 D2("mr_register_shared: smrp 0x%p, found cookie:\n" 4748 "0x%016llx%016llx%016llx%016llx%016llx\n", smrp, 4749 (longlong_t)tmp_smr.smr_cookie.mc_uint_arr[4], 4750 (longlong_t)tmp_smr.smr_cookie.mc_uint_arr[3], 4751 (longlong_t)tmp_smr.smr_cookie.mc_uint_arr[2], 4752 (longlong_t)tmp_smr.smr_cookie.mc_uint_arr[1], 4753 (longlong_t)tmp_smr.smr_cookie.mc_uint_arr[0]); 4754 4755 /* 4756 * if the smrp exists, other threads could still be 4757 * accessing it. we wait until they are done before 4758 * we continue. 4759 */ 4760 smrp->smr_refcnt++; 4761 while (smrp->smr_state == DAPLKA_SMR_TRANSITIONING) { 4762 D2("mr_register_shared: smrp 0x%p, " 4763 "waiting in transitioning state, refcnt %d\n", 4764 smrp, smrp->smr_refcnt); 4765 cv_wait(&smrp->smr_cv, &daplka_shared_mr_lock); 4766 } 4767 ASSERT(smrp->smr_state == DAPLKA_SMR_READY); 4768 D2("mr_register_shared: smrp 0x%p, refcnt %d, ready\n", 4769 smrp, smrp->smr_refcnt); 4770 4771 /* 4772 * we set smr_state to TRANSITIONING to temporarily 4773 * prevent other threads from trying to access smrp. 4774 */ 4775 smrp->smr_state = DAPLKA_SMR_TRANSITIONING; 4776 } else { 4777 D2("mr_register_shared: cannot find cookie:\n" 4778 "0x%016llx%016llx%016llx%016llx%016llx\n", 4779 (longlong_t)tmp_smr.smr_cookie.mc_uint_arr[4], 4780 (longlong_t)tmp_smr.smr_cookie.mc_uint_arr[3], 4781 (longlong_t)tmp_smr.smr_cookie.mc_uint_arr[2], 4782 (longlong_t)tmp_smr.smr_cookie.mc_uint_arr[1], 4783 (longlong_t)tmp_smr.smr_cookie.mc_uint_arr[0]); 4784 4785 /* 4786 * if we cannot find smrp, we need to create and 4787 * insert one into daplka_shared_mr_tree 4788 */ 4789 smrp = kmem_zalloc(sizeof (daplka_shared_mr_t), 4790 daplka_km_flags); 4791 if (smrp == NULL) { 4792 retval = ENOMEM; 4793 mutex_exit(&daplka_shared_mr_lock); 4794 goto cleanup; 4795 } 4796 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*smrp)) 4797 smrp->smr_refcnt = 1; 4798 smrp->smr_cookie = args.mrs_shm_cookie; 4799 smrp->smr_state = DAPLKA_SMR_TRANSITIONING; 4800 smrp->smr_mr_list = NULL; 4801 cv_init(&smrp->smr_cv, NULL, CV_DRIVER, NULL); 4802 avl_insert(&daplka_shared_mr_tree, smrp, where); 4803 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(*smrp)) 4804 } 4805 mutex_exit(&daplka_shared_mr_lock); 4806 4807 mr_rp = kmem_zalloc(sizeof (daplka_mr_resource_t), daplka_km_flags); 4808 if (mr_rp == NULL) { 4809 DERR("mr_register_shared: cannot allocate mr resource\n"); 4810 goto cleanup; 4811 } 4812 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*mr_rp)) 4813 DAPLKA_RS_INIT(mr_rp, DAPL_TYPE_MR, 4814 DAPLKA_RS_RNUM(ia_rp), daplka_mr_destroy); 4815 4816 mutex_init(&mr_rp->mr_lock, NULL, MUTEX_DRIVER, NULL); 4817 mr_rp->mr_hca = ia_rp->ia_hca; 4818 mr_rp->mr_hca_hdl = ia_rp->ia_hca_hdl; 4819 mr_rp->mr_next = NULL; 4820 mr_rp->mr_shared_mr = NULL; 4821 4822 /* get pd handle */ 4823 pd_rp = (daplka_pd_resource_t *) 4824 daplka_hash_lookup(&ia_rp->ia_pd_htbl, args.mrs_pd_hkey); 4825 if (pd_rp == NULL) { 4826 DERR("mr_register_shared: cannot find pd resource\n"); 4827 retval = EINVAL; 4828 goto cleanup; 4829 } 4830 ASSERT(DAPLKA_RS_TYPE(pd_rp) == DAPL_TYPE_PD); 4831 mr_rp->mr_pd_res = pd_rp; 4832 4833 mr_rp->mr_attr.mr_vaddr = args.mrs_vaddr; 4834 mr_rp->mr_attr.mr_len = args.mrs_len; 4835 mr_rp->mr_attr.mr_flags = args.mrs_flags | IBT_MR_NOSLEEP; 4836 mr_rp->mr_attr.mr_as = curproc->p_as; 4837 4838 D2("mr_register_shared: mr_vaddr 0x%p, mr_len %llu, " 4839 "mr_flags 0x%x, mr_as 0x%p, mr_exists %d, smrp 0x%p\n", 4840 (void *)(uintptr_t)mr_rp->mr_attr.mr_vaddr, 4841 (longlong_t)mr_rp->mr_attr.mr_len, 4842 mr_rp->mr_attr.mr_flags, mr_rp->mr_attr.mr_as, 4843 (int)(smrp->smr_mr_list != NULL), smrp); 4844 4845 /* 4846 * since we are in TRANSITIONING state, we are guaranteed 4847 * that we have exclusive access to smr_mr_list. 4848 */ 4849 if (smrp->smr_mr_list != NULL) { 4850 ibt_smr_attr_t mem_sattr; 4851 4852 /* 4853 * a non-null smr_mr_list indicates that someone 4854 * else has already inserted an mr_resource into 4855 * smr_mr_list. we use the mr_handle from the first 4856 * element as an arg to ibt_register_shared_mr. 4857 */ 4858 mem_sattr.mr_vaddr = smrp->smr_mr_list->mr_desc.md_vaddr; 4859 mem_sattr.mr_flags = mr_rp->mr_attr.mr_flags; 4860 4861 D2("mr_register_shared: mem_sattr vaddr 0x%p flags 0x%x\n", 4862 (void *)(uintptr_t)mem_sattr.mr_vaddr, mem_sattr.mr_flags); 4863 status = daplka_ibt_register_shared_mr(mr_rp, ia_rp->ia_hca_hdl, 4864 smrp->smr_mr_list->mr_hdl, mr_rp->mr_pd_res->pd_hdl, 4865 &mem_sattr, &mr_rp->mr_hdl, &mr_rp->mr_desc); 4866 4867 if (status != IBT_SUCCESS) { 4868 DERR("mr_register_shared: " 4869 "ibt_register_shared_mr error %d\n", status); 4870 *rvalp = (int)status; 4871 retval = 0; 4872 goto cleanup; 4873 } 4874 } else { 4875 /* 4876 * an mr does not exist yet. we need to create one 4877 * using ibt_register_mr. 4878 */ 4879 status = daplka_ibt_register_mr(mr_rp, ia_rp->ia_hca_hdl, 4880 mr_rp->mr_pd_res->pd_hdl, &mr_rp->mr_attr, 4881 &mr_rp->mr_hdl, &mr_rp->mr_desc); 4882 4883 if (status != IBT_SUCCESS) { 4884 DERR("mr_register_shared: " 4885 "ibt_register_mr error %d\n", status); 4886 *rvalp = (int)status; 4887 retval = 0; 4888 goto cleanup; 4889 } 4890 } 4891 4892 mr_cb_data_in.mr_rev = IBT_MR_DATA_IN_IF_VERSION; 4893 mr_cb_data_in.mr_func = daplka_pre_mr_cleanup_callback; 4894 mr_cb_data_in.mr_arg1 = (void *)mr_rp; 4895 mr_cb_data_in.mr_arg2 = NULL; 4896 4897 /* Pass the service driver mr cleanup handler to the hca driver */ 4898 status = ibt_ci_data_in(ia_rp->ia_hca_hdl, 4899 IBT_CI_NO_FLAGS, IBT_HDL_MR, (void *)mr_rp->mr_hdl, 4900 &mr_cb_data_in, sizeof (mr_cb_data_in)); 4901 4902 if (status != IBT_SUCCESS) { 4903 DERR("mr_register_shared: ibt_ci_data_in error(%d) ver(%d)", 4904 status, mr_cb_data_in.mr_rev); 4905 *rvalp = (int)status; 4906 retval = 0; 4907 goto cleanup; 4908 } 4909 4910 /* 4911 * we bump reference of mr_rp and enqueue it onto smrp. 4912 */ 4913 DAPLKA_RS_REF(mr_rp); 4914 mr_rp->mr_next = smrp->smr_mr_list; 4915 smrp->smr_mr_list = mr_rp; 4916 mr_rp->mr_shared_mr = smrp; 4917 4918 /* insert into mr hash table */ 4919 retval = daplka_hash_insert(&ia_rp->ia_mr_htbl, 4920 &mr_hkey, (void *)mr_rp); 4921 if (retval != 0) { 4922 DERR("mr_register_shared: cannot insert mr resource\n"); 4923 goto cleanup; 4924 } 4925 inserted = B_TRUE; 4926 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(*mr_rp)) 4927 4928 /* 4929 * at this point, there are two references to our mr resource. 4930 * one is kept in ia_mr_htbl. the other is kept in the list 4931 * within this shared mr object (smrp). when we deregister this 4932 * mr or when a callback invalidates this mr, the reference kept 4933 * by this shared mr object will be removed. 4934 */ 4935 4936 args.mrs_lkey = mr_rp->mr_desc.md_lkey; 4937 args.mrs_rkey = mr_rp->mr_desc.md_rkey; 4938 args.mrs_hkey = mr_hkey; 4939 4940 retval = ddi_copyout((void *)&args, (void *)arg, 4941 sizeof (dapl_mr_register_shared_t), mode); 4942 if (retval != 0) { 4943 DERR("mr_register_shared: copyout error %d\n", retval); 4944 retval = EFAULT; 4945 goto cleanup; 4946 } 4947 4948 /* 4949 * set the state to READY to allow others to continue 4950 */ 4951 mutex_enter(&daplka_shared_mr_lock); 4952 smrp->smr_state = DAPLKA_SMR_READY; 4953 cv_broadcast(&smrp->smr_cv); 4954 mutex_exit(&daplka_shared_mr_lock); 4955 return (0); 4956 4957 cleanup:; 4958 if (inserted) { 4959 daplka_mr_resource_t *free_rp = NULL; 4960 4961 (void) daplka_hash_remove(&ia_rp->ia_mr_htbl, mr_hkey, 4962 (void **)&free_rp); 4963 if (free_rp != mr_rp) { 4964 DERR("mr_register_shared: " 4965 "cannot remove mr from hash table\n"); 4966 /* 4967 * we can only get here if another thread 4968 * has completed the cleanup in mr_deregister 4969 */ 4970 return (retval); 4971 } 4972 } 4973 if (smrp != NULL) { 4974 mutex_enter(&daplka_shared_mr_lock); 4975 ASSERT(smrp->smr_refcnt > 0); 4976 smrp->smr_refcnt--; 4977 4978 if (smrp->smr_refcnt == 0) { 4979 DERR("mr_register_shared: freeing smrp 0x%p\n", smrp); 4980 avl_remove(&daplka_shared_mr_tree, smrp); 4981 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*smrp)) 4982 if (smrp->smr_mr_list != NULL) { 4983 /* 4984 * the refcnt is 0. if there is anything 4985 * left on the list, it must be ours. 4986 */ 4987 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*mr_rp)) 4988 ASSERT(smrp->smr_mr_list == mr_rp); 4989 DAPLKA_RS_UNREF(mr_rp); 4990 smrp->smr_mr_list = NULL; 4991 ASSERT(mr_rp->mr_shared_mr == smrp); 4992 mr_rp->mr_shared_mr = NULL; 4993 ASSERT(mr_rp->mr_next == NULL); 4994 } 4995 smrp->smr_state = DAPLKA_SMR_FREED; 4996 cv_destroy(&smrp->smr_cv); 4997 kmem_free(smrp, sizeof (daplka_shared_mr_t)); 4998 } else { 4999 DERR("mr_register_shared: resetting smr_state " 5000 "smrp 0x%p, %d waiters remain\n", smrp, 5001 smrp->smr_refcnt); 5002 ASSERT(smrp->smr_state == DAPLKA_SMR_TRANSITIONING); 5003 if (smrp->smr_mr_list != NULL && mr_rp != NULL) { 5004 daplka_mr_resource_t **mpp; 5005 5006 /* 5007 * search and remove mr_rp from smr_mr_list 5008 */ 5009 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*mr_rp)) 5010 mpp = &smrp->smr_mr_list; 5011 while (*mpp != NULL) { 5012 if (*mpp == mr_rp) { 5013 *mpp = (*mpp)->mr_next; 5014 DAPLKA_RS_UNREF(mr_rp); 5015 ASSERT(mr_rp->mr_shared_mr == 5016 smrp); 5017 mr_rp->mr_shared_mr = NULL; 5018 mr_rp->mr_next = NULL; 5019 break; 5020 } 5021 mpp = &(*mpp)->mr_next; 5022 } 5023 } 5024 /* 5025 * note that smr_state == READY does not necessarily 5026 * mean that smr_mr_list is non empty. for this case, 5027 * we are doing cleanup because of a failure. we set 5028 * the state to READY to allow other threads to 5029 * continue. 5030 */ 5031 smrp->smr_state = DAPLKA_SMR_READY; 5032 cv_broadcast(&smrp->smr_cv); 5033 } 5034 mutex_exit(&daplka_shared_mr_lock); 5035 } 5036 if (mr_rp != NULL) { 5037 DAPLKA_RS_UNREF(mr_rp); 5038 } 5039 return (retval); 5040 } 5041 5042 /* 5043 * registers a memory region using the attributes of an 5044 * existing region. 5045 */ 5046 /* ARGSUSED */ 5047 static int 5048 daplka_mr_register_lmr(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 5049 cred_t *cred, int *rvalp) 5050 { 5051 boolean_t inserted = B_FALSE; 5052 dapl_mr_register_lmr_t args; 5053 ibt_mr_data_in_t mr_cb_data_in; 5054 daplka_mr_resource_t *orig_mr_rp = NULL; 5055 daplka_mr_resource_t *mr_rp; 5056 ibt_smr_attr_t mem_sattr; 5057 uint64_t mr_hkey = 0; 5058 ibt_status_t status; 5059 int retval; 5060 5061 retval = ddi_copyin((void *)arg, &args, 5062 sizeof (dapl_mr_register_lmr_t), mode); 5063 if (retval != 0) { 5064 DERR("mr_register_lmr: copyin error %d\n", retval); 5065 return (EINVAL); 5066 } 5067 orig_mr_rp = (daplka_mr_resource_t *) 5068 daplka_hash_lookup(&ia_rp->ia_mr_htbl, args.mrl_orig_hkey); 5069 if (orig_mr_rp == NULL) { 5070 DERR("mr_register_lmr: cannot find mr resource\n"); 5071 return (EINVAL); 5072 } 5073 ASSERT(DAPLKA_RS_TYPE(orig_mr_rp) == DAPL_TYPE_MR); 5074 5075 mr_rp = kmem_zalloc(sizeof (daplka_mr_resource_t), daplka_km_flags); 5076 if (mr_rp == NULL) { 5077 DERR("mr_register_lmr: cannot allocate mr resource\n"); 5078 retval = ENOMEM; 5079 goto cleanup; 5080 } 5081 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*mr_rp)) 5082 DAPLKA_RS_INIT(mr_rp, DAPL_TYPE_MR, 5083 DAPLKA_RS_RNUM(ia_rp), daplka_mr_destroy); 5084 5085 mutex_init(&mr_rp->mr_lock, NULL, MUTEX_DRIVER, NULL); 5086 mr_rp->mr_hca = ia_rp->ia_hca; 5087 mr_rp->mr_hca_hdl = ia_rp->ia_hca_hdl; 5088 mr_rp->mr_next = NULL; 5089 mr_rp->mr_shared_mr = NULL; 5090 5091 DAPLKA_RS_REF(orig_mr_rp->mr_pd_res); 5092 mr_rp->mr_pd_res = orig_mr_rp->mr_pd_res; 5093 mr_rp->mr_attr = orig_mr_rp->mr_attr; 5094 5095 /* Pass the IO addr that was returned while allocating the orig MR */ 5096 mem_sattr.mr_vaddr = orig_mr_rp->mr_desc.md_vaddr; 5097 mem_sattr.mr_flags = args.mrl_flags | IBT_MR_NOSLEEP; 5098 5099 status = daplka_ibt_register_shared_mr(mr_rp, ia_rp->ia_hca_hdl, 5100 orig_mr_rp->mr_hdl, mr_rp->mr_pd_res->pd_hdl, &mem_sattr, 5101 &mr_rp->mr_hdl, &mr_rp->mr_desc); 5102 5103 if (status != IBT_SUCCESS) { 5104 DERR("mr_register_lmr: ibt_register_shared_mr error %d\n", 5105 status); 5106 *rvalp = (int)status; 5107 retval = 0; 5108 goto cleanup; 5109 } 5110 5111 mr_cb_data_in.mr_rev = IBT_MR_DATA_IN_IF_VERSION; 5112 mr_cb_data_in.mr_func = daplka_pre_mr_cleanup_callback; 5113 mr_cb_data_in.mr_arg1 = (void *)mr_rp; 5114 mr_cb_data_in.mr_arg2 = NULL; 5115 5116 /* Pass the service driver mr cleanup handler to the hca driver */ 5117 status = ibt_ci_data_in(ia_rp->ia_hca_hdl, 5118 IBT_CI_NO_FLAGS, IBT_HDL_MR, (void *)mr_rp->mr_hdl, 5119 &mr_cb_data_in, sizeof (mr_cb_data_in)); 5120 5121 if (status != IBT_SUCCESS) { 5122 DERR("mr_register_lmr: ibt_ci_data_in error(%d) ver(%d)", 5123 status, mr_cb_data_in.mr_rev); 5124 *rvalp = (int)status; 5125 retval = 0; 5126 goto cleanup; 5127 } 5128 mr_rp->mr_attr.mr_len = orig_mr_rp->mr_attr.mr_len; 5129 mr_rp->mr_attr.mr_flags = mem_sattr.mr_flags; 5130 5131 /* insert into mr hash table */ 5132 retval = daplka_hash_insert(&ia_rp->ia_mr_htbl, &mr_hkey, 5133 (void *)mr_rp); 5134 if (retval != 0) { 5135 DERR("mr_register: cannot insert mr resource into mr_htbl\n"); 5136 goto cleanup; 5137 } 5138 inserted = B_TRUE; 5139 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(*mr_rp)) 5140 5141 args.mrl_lkey = mr_rp->mr_desc.md_lkey; 5142 args.mrl_rkey = mr_rp->mr_desc.md_rkey; 5143 args.mrl_hkey = mr_hkey; 5144 5145 retval = ddi_copyout((void *)&args, (void *)arg, 5146 sizeof (dapl_mr_register_lmr_t), mode); 5147 if (retval != 0) { 5148 DERR("mr_register_lmr: copyout error %d\n", retval); 5149 retval = EFAULT; 5150 goto cleanup; 5151 } 5152 if (orig_mr_rp != NULL) { 5153 DAPLKA_RS_UNREF(orig_mr_rp); 5154 } 5155 return (0); 5156 5157 cleanup:; 5158 if (inserted) { 5159 daplka_mr_resource_t *free_rp = NULL; 5160 5161 (void) daplka_hash_remove(&ia_rp->ia_mr_htbl, mr_hkey, 5162 (void **)&free_rp); 5163 if (free_rp != mr_rp) { 5164 DERR("mr_register: cannot remove mr from hash table\n"); 5165 /* 5166 * we can only get here if another thread 5167 * has completed the cleanup in mr_deregister 5168 */ 5169 return (retval); 5170 } 5171 } 5172 if (orig_mr_rp != NULL) { 5173 DAPLKA_RS_UNREF(orig_mr_rp); 5174 } 5175 if (mr_rp != NULL) { 5176 DAPLKA_RS_UNREF(mr_rp); 5177 } 5178 return (retval); 5179 } 5180 5181 /* 5182 * this function is called by mr_deregister and mr_cleanup_callback to 5183 * remove a mr resource from the shared mr object mr_rp->mr_shared_mr. 5184 * if mr_shared_mr is already NULL, that means the region being 5185 * deregistered or invalidated is not a shared mr region and we can 5186 * return immediately. 5187 */ 5188 static void 5189 daplka_shared_mr_free(daplka_mr_resource_t *mr_rp) 5190 { 5191 daplka_shared_mr_t *smrp; 5192 5193 /* 5194 * we need a lock because mr_callback also checks this field. 5195 * for the rare case that mr_deregister and mr_cleanup_callback 5196 * gets called simultaneously, we are guaranteed that smrp won't 5197 * be dereferenced twice because either function will find 5198 * mr_shared_mr to be NULL. 5199 */ 5200 mutex_enter(&mr_rp->mr_lock); 5201 smrp = mr_rp->mr_shared_mr; 5202 mr_rp->mr_shared_mr = NULL; 5203 mutex_exit(&mr_rp->mr_lock); 5204 5205 if (smrp != NULL) { 5206 daplka_mr_resource_t **mpp; 5207 boolean_t mr_found = B_FALSE; 5208 5209 mutex_enter(&daplka_shared_mr_lock); 5210 ASSERT(smrp->smr_refcnt > 0); 5211 while (smrp->smr_state == DAPLKA_SMR_TRANSITIONING) { 5212 cv_wait(&smrp->smr_cv, &daplka_shared_mr_lock); 5213 } 5214 ASSERT(smrp->smr_state == DAPLKA_SMR_READY); 5215 smrp->smr_state = DAPLKA_SMR_TRANSITIONING; 5216 smrp->smr_refcnt--; 5217 5218 /* 5219 * search and remove mr_rp from smr_mr_list. 5220 * also UNREF mr_rp because it is no longer 5221 * on the list. 5222 */ 5223 mpp = &smrp->smr_mr_list; 5224 while (*mpp != NULL) { 5225 if (*mpp == mr_rp) { 5226 *mpp = (*mpp)->mr_next; 5227 DAPLKA_RS_UNREF(mr_rp); 5228 mr_rp->mr_next = NULL; 5229 mr_found = B_TRUE; 5230 break; 5231 } 5232 mpp = &(*mpp)->mr_next; 5233 } 5234 /* 5235 * since mr_clean_callback may not touch smr_mr_list 5236 * at this time (due to smr_state), we can be sure 5237 * that we can find and remove mr_rp from smr_mr_list 5238 */ 5239 ASSERT(mr_found); 5240 if (smrp->smr_refcnt == 0) { 5241 D3("shared_mr_free: freeing smrp 0x%p\n", smrp); 5242 avl_remove(&daplka_shared_mr_tree, smrp); 5243 ASSERT(smrp->smr_mr_list == NULL); 5244 smrp->smr_state = DAPLKA_SMR_FREED; 5245 cv_destroy(&smrp->smr_cv); 5246 kmem_free(smrp, sizeof (daplka_shared_mr_t)); 5247 } else { 5248 D3("shared_mr_free: smrp 0x%p, refcnt %d\n", 5249 smrp, smrp->smr_refcnt); 5250 smrp->smr_state = DAPLKA_SMR_READY; 5251 cv_broadcast(&smrp->smr_cv); 5252 } 5253 mutex_exit(&daplka_shared_mr_lock); 5254 } 5255 } 5256 5257 /* 5258 * deregisters a memory region. 5259 * if mr is shared, remove reference from global shared mr object. 5260 * release the initial reference to the mr. if the mr's refcnt is 5261 * zero, call mr_destroy to free mr. 5262 */ 5263 /* ARGSUSED */ 5264 static int 5265 daplka_mr_deregister(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 5266 cred_t *cred, int *rvalp) 5267 { 5268 daplka_mr_resource_t *mr_rp; 5269 dapl_mr_deregister_t args; 5270 int retval; 5271 5272 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_mr_deregister_t), 5273 mode); 5274 if (retval != 0) { 5275 DERR("mr_deregister: copyin error %d\n", retval); 5276 return (EINVAL); 5277 } 5278 retval = daplka_hash_remove(&ia_rp->ia_mr_htbl, 5279 args.mrd_hkey, (void **)&mr_rp); 5280 if (retval != 0 || mr_rp == NULL) { 5281 DERR("mr_deregister: cannot find mr resource\n"); 5282 return (EINVAL); 5283 } 5284 ASSERT(DAPLKA_RS_TYPE(mr_rp) == DAPL_TYPE_MR); 5285 5286 daplka_shared_mr_free(mr_rp); 5287 DAPLKA_RS_UNREF(mr_rp); 5288 return (0); 5289 } 5290 5291 /* 5292 * sync local memory regions on RDMA read or write. 5293 */ 5294 /* ARGSUSED */ 5295 static int 5296 daplka_mr_sync(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 5297 cred_t *cred, int *rvalp) 5298 { 5299 dapl_mr_sync_t args; 5300 daplka_mr_resource_t *mr_rp[DAPL_MR_PER_SYNC]; 5301 ibt_mr_sync_t mrs[DAPL_MR_PER_SYNC]; 5302 uint32_t sync_direction_flags; 5303 ibt_status_t status; 5304 int i, j; 5305 int retval; 5306 5307 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_mr_sync_t), mode); 5308 if (retval != 0) { 5309 DERR("mr_sync: copyin error %d\n", retval); 5310 return (EFAULT); 5311 } 5312 5313 /* number of segments bound check */ 5314 if (args.mrs_numseg > DAPL_MR_PER_SYNC) { 5315 DERR("mr_sync: number of segments too large\n"); 5316 return (EINVAL); 5317 } 5318 5319 /* translate MR sync direction flag */ 5320 if (args.mrs_flags == DAPL_MR_SYNC_RDMA_RD) { 5321 sync_direction_flags = IBT_SYNC_READ; 5322 } else if (args.mrs_flags == DAPL_MR_SYNC_RDMA_WR) { 5323 sync_direction_flags = IBT_SYNC_WRITE; 5324 } else { 5325 DERR("mr_sync: unknown flags\n"); 5326 return (EINVAL); 5327 } 5328 5329 /* 5330 * all the segments are going to be sync'd by ibtl together 5331 */ 5332 for (i = 0; i < args.mrs_numseg; i++) { 5333 mr_rp[i] = (daplka_mr_resource_t *)daplka_hash_lookup( 5334 &ia_rp->ia_mr_htbl, args.mrs_vec[i].mrsv_hkey); 5335 if (mr_rp[i] == NULL) { 5336 for (j = 0; j < i; j++) { 5337 DAPLKA_RS_UNREF(mr_rp[j]); 5338 } 5339 DERR("mr_sync: lookup error\n"); 5340 return (EINVAL); 5341 } 5342 ASSERT(DAPLKA_RS_TYPE(mr_rp[i]) == DAPL_TYPE_MR); 5343 mrs[i].ms_handle = mr_rp[i]->mr_hdl; 5344 mrs[i].ms_vaddr = args.mrs_vec[i].mrsv_va; 5345 mrs[i].ms_len = args.mrs_vec[i].mrsv_len; 5346 mrs[i].ms_flags = sync_direction_flags; 5347 } 5348 5349 status = ibt_sync_mr(ia_rp->ia_hca_hdl, mrs, args.mrs_numseg); 5350 if (status != IBT_SUCCESS) { 5351 DERR("mr_sync: ibt_sync_mr error %d\n", status); 5352 *rvalp = (int)status; 5353 } 5354 for (i = 0; i < args.mrs_numseg; i++) { 5355 DAPLKA_RS_UNREF(mr_rp[i]); 5356 } 5357 return (0); 5358 } 5359 5360 /* 5361 * destroys a memory region. 5362 * called when refcnt drops to zero. 5363 */ 5364 static int 5365 daplka_mr_destroy(daplka_resource_t *gen_rp) 5366 { 5367 daplka_mr_resource_t *mr_rp = (daplka_mr_resource_t *)gen_rp; 5368 ibt_status_t status; 5369 5370 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*mr_rp)) 5371 ASSERT(DAPLKA_RS_REFCNT(mr_rp) == 0); 5372 ASSERT(mr_rp->mr_shared_mr == NULL); 5373 D3("mr_destroy: entering, mr_rp 0x%p, rnum %d\n", 5374 mr_rp, DAPLKA_RS_RNUM(mr_rp)); 5375 5376 /* 5377 * deregister mr 5378 */ 5379 if (mr_rp->mr_hdl) { 5380 status = daplka_ibt_deregister_mr(mr_rp, mr_rp->mr_hca_hdl, 5381 mr_rp->mr_hdl); 5382 if (status != IBT_SUCCESS) { 5383 DERR("mr_destroy: ibt_deregister_mr returned %d\n", 5384 status); 5385 } 5386 mr_rp->mr_hdl = NULL; 5387 D3("mr_destroy: mr deregistered\n"); 5388 } 5389 mr_rp->mr_attr.mr_vaddr = NULL; 5390 5391 /* 5392 * release reference on PD 5393 */ 5394 if (mr_rp->mr_pd_res != NULL) { 5395 DAPLKA_RS_UNREF(mr_rp->mr_pd_res); 5396 mr_rp->mr_pd_res = NULL; 5397 } 5398 mutex_destroy(&mr_rp->mr_lock); 5399 DAPLKA_RS_FINI(mr_rp); 5400 kmem_free(mr_rp, sizeof (daplka_mr_resource_t)); 5401 D3("mr_destroy: exiting, mr_rp 0x%p\n", mr_rp); 5402 return (0); 5403 } 5404 5405 /* 5406 * this function is called by daplka_hash_destroy for 5407 * freeing MR resource objects 5408 */ 5409 static void 5410 daplka_hash_mr_free(void *obj) 5411 { 5412 daplka_mr_resource_t *mr_rp = (daplka_mr_resource_t *)obj; 5413 5414 daplka_shared_mr_free(mr_rp); 5415 DAPLKA_RS_UNREF(mr_rp); 5416 } 5417 5418 /* 5419 * comparison function used for finding a shared mr object 5420 * from the global shared mr avl tree. 5421 */ 5422 static int 5423 daplka_shared_mr_cmp(const void *smr1, const void *smr2) 5424 { 5425 daplka_shared_mr_t *s1 = (daplka_shared_mr_t *)smr1; 5426 daplka_shared_mr_t *s2 = (daplka_shared_mr_t *)smr2; 5427 int i; 5428 5429 for (i = 4; i >= 0; i--) { 5430 if (s1->smr_cookie.mc_uint_arr[i] < 5431 s2->smr_cookie.mc_uint_arr[i]) { 5432 return (-1); 5433 } 5434 if (s1->smr_cookie.mc_uint_arr[i] > 5435 s2->smr_cookie.mc_uint_arr[i]) { 5436 return (1); 5437 } 5438 } 5439 return (0); 5440 } 5441 5442 /* 5443 * allocates a protection domain. 5444 */ 5445 /* ARGSUSED */ 5446 static int 5447 daplka_pd_alloc(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 5448 cred_t *cred, int *rvalp) 5449 { 5450 dapl_pd_alloc_t args; 5451 daplka_pd_resource_t *pd_rp; 5452 ibt_status_t status; 5453 uint64_t pd_hkey = 0; 5454 boolean_t inserted = B_FALSE; 5455 int retval; 5456 5457 pd_rp = kmem_zalloc(sizeof (*pd_rp), daplka_km_flags); 5458 if (pd_rp == NULL) { 5459 DERR("pd_alloc: cannot allocate pd resource\n"); 5460 return (ENOMEM); 5461 } 5462 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*pd_rp)) 5463 DAPLKA_RS_INIT(pd_rp, DAPL_TYPE_PD, 5464 DAPLKA_RS_RNUM(ia_rp), daplka_pd_destroy); 5465 5466 pd_rp->pd_hca = ia_rp->ia_hca; 5467 pd_rp->pd_hca_hdl = ia_rp->ia_hca_hdl; 5468 status = daplka_ibt_alloc_pd(pd_rp, pd_rp->pd_hca_hdl, 5469 IBT_PD_NO_FLAGS, &pd_rp->pd_hdl); 5470 if (status != IBT_SUCCESS) { 5471 DERR("pd_alloc: ibt_alloc_pd returned %d\n", status); 5472 *rvalp = (int)status; 5473 retval = 0; 5474 goto cleanup; 5475 } 5476 5477 /* insert into pd hash table */ 5478 retval = daplka_hash_insert(&ia_rp->ia_pd_htbl, 5479 &pd_hkey, (void *)pd_rp); 5480 if (retval != 0) { 5481 DERR("pd_alloc: cannot insert pd resource into pd_htbl\n"); 5482 goto cleanup; 5483 } 5484 inserted = B_TRUE; 5485 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(*pd_rp)) 5486 5487 /* return hkey to library */ 5488 args.pda_hkey = pd_hkey; 5489 5490 retval = ddi_copyout(&args, (void *)arg, sizeof (dapl_pd_alloc_t), 5491 mode); 5492 if (retval != 0) { 5493 DERR("pd_alloc: copyout error %d\n", retval); 5494 retval = EFAULT; 5495 goto cleanup; 5496 } 5497 return (0); 5498 5499 cleanup:; 5500 if (inserted) { 5501 daplka_pd_resource_t *free_rp = NULL; 5502 5503 (void) daplka_hash_remove(&ia_rp->ia_pd_htbl, pd_hkey, 5504 (void **)&free_rp); 5505 if (free_rp != pd_rp) { 5506 DERR("pd_alloc: cannot remove pd from hash table\n"); 5507 /* 5508 * we can only get here if another thread 5509 * has completed the cleanup in pd_free 5510 */ 5511 return (retval); 5512 } 5513 } 5514 DAPLKA_RS_UNREF(pd_rp); 5515 return (retval); 5516 } 5517 5518 /* 5519 * destroys a protection domain. 5520 * called when refcnt drops to zero. 5521 */ 5522 static int 5523 daplka_pd_destroy(daplka_resource_t *gen_rp) 5524 { 5525 daplka_pd_resource_t *pd_rp = (daplka_pd_resource_t *)gen_rp; 5526 ibt_status_t status; 5527 5528 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*pd_rp)) 5529 ASSERT(DAPLKA_RS_REFCNT(pd_rp) == 0); 5530 D3("pd_destroy: entering, pd_rp %p, rnum %d\n", 5531 pd_rp, DAPLKA_RS_RNUM(pd_rp)); 5532 5533 ASSERT(DAPLKA_RS_TYPE(pd_rp) == DAPL_TYPE_PD); 5534 if (pd_rp->pd_hdl != NULL) { 5535 status = daplka_ibt_free_pd(pd_rp, pd_rp->pd_hca_hdl, 5536 pd_rp->pd_hdl); 5537 if (status != IBT_SUCCESS) { 5538 DERR("pd_destroy: ibt_free_pd returned %d\n", status); 5539 } 5540 } 5541 DAPLKA_RS_FINI(pd_rp); 5542 kmem_free(pd_rp, sizeof (daplka_pd_resource_t)); 5543 D3("pd_destroy: exiting, pd_rp %p\n", pd_rp); 5544 return (0); 5545 } 5546 5547 static void 5548 daplka_hash_pd_free(void *obj) 5549 { 5550 daplka_pd_resource_t *pd_rp = (daplka_pd_resource_t *)obj; 5551 5552 ASSERT(DAPLKA_RS_TYPE(pd_rp) == DAPL_TYPE_PD); 5553 DAPLKA_RS_UNREF(pd_rp); 5554 } 5555 5556 /* 5557 * removes the pd reference from ia_pd_htbl and releases the 5558 * initial reference to the pd. also destroys the pd if the refcnt 5559 * is zero. 5560 */ 5561 /* ARGSUSED */ 5562 static int 5563 daplka_pd_free(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 5564 cred_t *cred, int *rvalp) 5565 { 5566 daplka_pd_resource_t *pd_rp; 5567 dapl_pd_free_t args; 5568 int retval; 5569 5570 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_pd_free_t), mode); 5571 if (retval != 0) { 5572 DERR("pd_free: copyin error %d\n", retval); 5573 return (EINVAL); 5574 } 5575 5576 retval = daplka_hash_remove(&ia_rp->ia_pd_htbl, 5577 args.pdf_hkey, (void **)&pd_rp); 5578 if (retval != 0 || pd_rp == NULL) { 5579 DERR("pd_free: cannot find pd resource\n"); 5580 return (EINVAL); 5581 } 5582 ASSERT(DAPLKA_RS_TYPE(pd_rp) == DAPL_TYPE_PD); 5583 5584 /* UNREF calls the actual free function when refcnt is zero */ 5585 DAPLKA_RS_UNREF(pd_rp); 5586 return (0); 5587 } 5588 5589 /* 5590 * allocates a memory window 5591 */ 5592 /* ARGSUSED */ 5593 static int 5594 daplka_mw_alloc(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 5595 cred_t *cred, int *rvalp) 5596 { 5597 daplka_pd_resource_t *pd_rp; 5598 daplka_mw_resource_t *mw_rp; 5599 dapl_mw_alloc_t args; 5600 ibt_status_t status; 5601 boolean_t inserted = B_FALSE; 5602 uint64_t mw_hkey; 5603 ibt_rkey_t mw_rkey; 5604 int retval; 5605 5606 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_mw_alloc_t), mode); 5607 if (retval != 0) { 5608 DERR("mw_alloc: copyin error %d\n", retval); 5609 return (EFAULT); 5610 } 5611 5612 /* 5613 * Allocate and initialize a MW resource 5614 */ 5615 mw_rp = kmem_zalloc(sizeof (daplka_mw_resource_t), daplka_km_flags); 5616 if (mw_rp == NULL) { 5617 DERR("mw_alloc: cannot allocate mw resource\n"); 5618 return (ENOMEM); 5619 } 5620 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*mw_rp)) 5621 DAPLKA_RS_INIT(mw_rp, DAPL_TYPE_MW, 5622 DAPLKA_RS_RNUM(ia_rp), daplka_mw_destroy); 5623 5624 mutex_init(&mw_rp->mw_lock, NULL, MUTEX_DRIVER, NULL); 5625 mw_rp->mw_hca = ia_rp->ia_hca; 5626 mw_rp->mw_hca_hdl = ia_rp->ia_hca_hdl; 5627 5628 /* get pd handle */ 5629 pd_rp = (daplka_pd_resource_t *) 5630 daplka_hash_lookup(&ia_rp->ia_pd_htbl, args.mw_pd_hkey); 5631 if (pd_rp == NULL) { 5632 DERR("mw_alloc: cannot find pd resource\n"); 5633 goto cleanup; 5634 } 5635 ASSERT(DAPLKA_RS_TYPE(pd_rp) == DAPL_TYPE_PD); 5636 5637 mw_rp->mw_pd_res = pd_rp; 5638 5639 status = daplka_ibt_alloc_mw(mw_rp, mw_rp->mw_hca_hdl, 5640 pd_rp->pd_hdl, IBT_MW_NOSLEEP, &mw_rp->mw_hdl, &mw_rkey); 5641 5642 if (status != IBT_SUCCESS) { 5643 DERR("mw_alloc: ibt_alloc_mw returned %d\n", status); 5644 *rvalp = (int)status; 5645 retval = 0; 5646 goto cleanup; 5647 } 5648 5649 mutex_enter(&ia_rp->ia_lock); 5650 switch (ia_rp->ia_state) { 5651 case DAPLKA_IA_INIT: 5652 ia_rp->ia_state = DAPLKA_IA_MW_ALLOC_IN_PROGRESS; 5653 ia_rp->ia_mw_alloccnt++; 5654 retval = 0; 5655 break; 5656 case DAPLKA_IA_MW_ALLOC_IN_PROGRESS: 5657 /* another mw_alloc is already in progress increase cnt */ 5658 ia_rp->ia_mw_alloccnt++; 5659 retval = 0; 5660 break; 5661 case DAPLKA_IA_MW_FREEZE_IN_PROGRESS: 5662 /* FALLTHRU */ 5663 case DAPLKA_IA_MW_FROZEN: 5664 /* 5665 * IA is being or already frozen don't allow more MWs to be 5666 * allocated. 5667 */ 5668 DERR("mw_alloc: IA is freezing MWs (state=%d)\n", 5669 ia_rp->ia_state); 5670 retval = EINVAL; 5671 break; 5672 default: 5673 ASSERT(!"Invalid IA state in mw_alloc"); 5674 DERR("mw_alloc: IA state=%d invalid\n", ia_rp->ia_state); 5675 retval = EINVAL; 5676 break; 5677 } 5678 mutex_exit(&ia_rp->ia_lock); 5679 /* retval is 0 when ia_mw_alloccnt is incremented */ 5680 if (retval != 0) { 5681 goto cleanup; 5682 } 5683 5684 /* insert into mw hash table */ 5685 mw_hkey = 0; 5686 retval = daplka_hash_insert(&ia_rp->ia_mw_htbl, &mw_hkey, 5687 (void *)mw_rp); 5688 if (retval != 0) { 5689 DERR("mw_alloc: cannot insert mw resource into mw_htbl\n"); 5690 mutex_enter(&ia_rp->ia_lock); 5691 ASSERT(ia_rp->ia_state == DAPLKA_IA_MW_ALLOC_IN_PROGRESS); 5692 ia_rp->ia_mw_alloccnt--; 5693 if (ia_rp->ia_mw_alloccnt == 0) { 5694 ia_rp->ia_state = DAPLKA_IA_INIT; 5695 cv_broadcast(&ia_rp->ia_cv); 5696 } 5697 mutex_exit(&ia_rp->ia_lock); 5698 goto cleanup; 5699 } 5700 inserted = B_TRUE; 5701 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(*mw_rp)) 5702 5703 D3("mw_alloc: ibt_alloc_mw mw_hdl(%p) mw_rkey(0x%llx)\n", 5704 mw_rp->mw_hdl, (longlong_t)mw_rkey); 5705 5706 mutex_enter(&ia_rp->ia_lock); 5707 /* 5708 * We are done with mw_alloc if this was the last mw_alloc 5709 * change state back to DAPLKA_IA_INIT and wake up waiters 5710 * specifically the unlock callback. 5711 */ 5712 ASSERT(ia_rp->ia_state == DAPLKA_IA_MW_ALLOC_IN_PROGRESS); 5713 ia_rp->ia_mw_alloccnt--; 5714 if (ia_rp->ia_mw_alloccnt == 0) { 5715 ia_rp->ia_state = DAPLKA_IA_INIT; 5716 cv_broadcast(&ia_rp->ia_cv); 5717 } 5718 mutex_exit(&ia_rp->ia_lock); 5719 5720 args.mw_hkey = mw_hkey; 5721 args.mw_rkey = mw_rkey; 5722 5723 retval = ddi_copyout(&args, (void *)arg, sizeof (dapl_mw_alloc_t), 5724 mode); 5725 if (retval != 0) { 5726 DERR("mw_alloc: copyout error %d\n", retval); 5727 retval = EFAULT; 5728 goto cleanup; 5729 } 5730 return (0); 5731 5732 cleanup:; 5733 if (inserted) { 5734 daplka_mw_resource_t *free_rp = NULL; 5735 5736 (void) daplka_hash_remove(&ia_rp->ia_mw_htbl, mw_hkey, 5737 (void **)&free_rp); 5738 if (free_rp != mw_rp) { 5739 DERR("mw_alloc: cannot remove mw from hash table\n"); 5740 /* 5741 * we can only get here if another thread 5742 * has completed the cleanup in mw_free 5743 */ 5744 return (retval); 5745 } 5746 } 5747 DAPLKA_RS_UNREF(mw_rp); 5748 return (retval); 5749 } 5750 5751 /* 5752 * removes the mw reference from ia_mw_htbl and releases the 5753 * initial reference to the mw. also destroys the mw if the refcnt 5754 * is zero. 5755 */ 5756 /* ARGSUSED */ 5757 static int 5758 daplka_mw_free(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 5759 cred_t *cred, int *rvalp) 5760 { 5761 daplka_mw_resource_t *mw_rp = NULL; 5762 dapl_mw_free_t args; 5763 int retval = 0; 5764 5765 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_mw_free_t), mode); 5766 if (retval != 0) { 5767 DERR("mw_free: copyin error %d\n", retval); 5768 return (EFAULT); 5769 } 5770 5771 retval = daplka_hash_remove(&ia_rp->ia_mw_htbl, args.mw_hkey, 5772 (void **)&mw_rp); 5773 if (retval != 0 || mw_rp == NULL) { 5774 DERR("mw_free: cannot find mw resrc (0x%llx)\n", 5775 (longlong_t)args.mw_hkey); 5776 return (EINVAL); 5777 } 5778 5779 ASSERT(DAPLKA_RS_TYPE(mw_rp) == DAPL_TYPE_MW); 5780 5781 /* UNREF calls the actual free function when refcnt is zero */ 5782 DAPLKA_RS_UNREF(mw_rp); 5783 return (retval); 5784 } 5785 5786 /* 5787 * destroys the memory window. 5788 * called when refcnt drops to zero. 5789 */ 5790 static int 5791 daplka_mw_destroy(daplka_resource_t *gen_rp) 5792 { 5793 daplka_mw_resource_t *mw_rp = (daplka_mw_resource_t *)gen_rp; 5794 ibt_status_t status; 5795 5796 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*mw_rp)) 5797 ASSERT(DAPLKA_RS_REFCNT(mw_rp) == 0); 5798 D3("mw_destroy: entering, mw_rp 0x%p, rnum %d\n", 5799 mw_rp, DAPLKA_RS_RNUM(mw_rp)); 5800 5801 /* 5802 * free memory window 5803 */ 5804 if (mw_rp->mw_hdl) { 5805 status = daplka_ibt_free_mw(mw_rp, mw_rp->mw_hca_hdl, 5806 mw_rp->mw_hdl); 5807 if (status != IBT_SUCCESS) { 5808 DERR("mw_destroy: ibt_free_mw returned %d\n", status); 5809 } 5810 mw_rp->mw_hdl = NULL; 5811 D3("mw_destroy: mw freed\n"); 5812 } 5813 5814 /* 5815 * release reference on PD 5816 */ 5817 if (mw_rp->mw_pd_res != NULL) { 5818 DAPLKA_RS_UNREF(mw_rp->mw_pd_res); 5819 mw_rp->mw_pd_res = NULL; 5820 } 5821 mutex_destroy(&mw_rp->mw_lock); 5822 DAPLKA_RS_FINI(mw_rp); 5823 kmem_free(mw_rp, sizeof (daplka_mw_resource_t)); 5824 D3("mw_destroy: exiting, mw_rp 0x%p\n", mw_rp); 5825 return (0); 5826 } 5827 5828 static void 5829 daplka_hash_mw_free(void *obj) 5830 { 5831 daplka_mw_resource_t *mw_rp = (daplka_mw_resource_t *)obj; 5832 5833 ASSERT(DAPLKA_RS_TYPE(mw_rp) == DAPL_TYPE_MW); 5834 DAPLKA_RS_UNREF(mw_rp); 5835 } 5836 5837 /* 5838 * SRQ ioctls and supporting functions 5839 */ 5840 /* ARGSUSED */ 5841 static int 5842 daplka_srq_create(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 5843 cred_t *cred, int *rvalp) 5844 { 5845 daplka_srq_resource_t *srq_rp; 5846 daplka_pd_resource_t *pd_rp; 5847 dapl_srq_create_t args; 5848 ibt_srq_sizes_t srq_sizes; 5849 ibt_srq_sizes_t srq_real_sizes; 5850 ibt_hca_attr_t *hca_attrp; 5851 uint64_t srq_hkey = 0; 5852 boolean_t inserted = B_FALSE; 5853 int retval; 5854 ibt_status_t status; 5855 5856 D3("srq_create: enter\n"); 5857 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_srq_create_t), 5858 mode); 5859 if (retval != 0) { 5860 DERR("srq_create: copyin error %d\n", retval); 5861 return (EFAULT); 5862 } 5863 srq_rp = kmem_zalloc(sizeof (daplka_srq_resource_t), daplka_km_flags); 5864 if (srq_rp == NULL) { 5865 DERR("srq_create: cannot allocate ep_rp\n"); 5866 return (ENOMEM); 5867 } 5868 DAPLKA_RS_INIT(srq_rp, DAPL_TYPE_SRQ, 5869 DAPLKA_RS_RNUM(ia_rp), daplka_srq_destroy); 5870 5871 srq_rp->srq_hca = ia_rp->ia_hca; 5872 srq_rp->srq_hca_hdl = ia_rp->ia_hca_hdl; 5873 mutex_init(&srq_rp->srq_lock, NULL, MUTEX_DRIVER, NULL); 5874 5875 /* get pd handle */ 5876 pd_rp = (daplka_pd_resource_t *) 5877 daplka_hash_lookup(&ia_rp->ia_pd_htbl, args.srqc_pd_hkey); 5878 if (pd_rp == NULL) { 5879 DERR("srq_create: cannot find pd resource\n"); 5880 retval = EINVAL; 5881 goto cleanup; 5882 } 5883 ASSERT(DAPLKA_RS_TYPE(pd_rp) == DAPL_TYPE_PD); 5884 srq_rp->srq_pd_res = pd_rp; 5885 5886 /* 5887 * these checks ensure that the requested SRQ sizes 5888 * are within the limits supported by the chosen HCA. 5889 */ 5890 hca_attrp = &ia_rp->ia_hca->hca_attr; 5891 if (args.srqc_sizes.srqs_sz > hca_attrp->hca_max_srqs_sz) { 5892 DERR("srq_create: invalid srqs_sz %d\n", 5893 args.srqc_sizes.srqs_sz); 5894 retval = EINVAL; 5895 goto cleanup; 5896 } 5897 if (args.srqc_sizes.srqs_sgl > hca_attrp->hca_max_srq_sgl) { 5898 DERR("srq_create: invalid srqs_sgl %d\n", 5899 args.srqc_sizes.srqs_sgl); 5900 retval = EINVAL; 5901 goto cleanup; 5902 } 5903 5904 D3("srq_create: srq_sgl %d, srq_sz %d\n", 5905 args.srqc_sizes.srqs_sgl, args.srqc_sizes.srqs_sz); 5906 5907 srq_sizes.srq_wr_sz = args.srqc_sizes.srqs_sz; 5908 srq_sizes.srq_sgl_sz = args.srqc_sizes.srqs_sgl; 5909 5910 /* create srq */ 5911 status = daplka_ibt_alloc_srq(srq_rp, ia_rp->ia_hca_hdl, 5912 IBT_SRQ_USER_MAP, pd_rp->pd_hdl, &srq_sizes, &srq_rp->srq_hdl, 5913 &srq_real_sizes); 5914 if (status != IBT_SUCCESS) { 5915 DERR("srq_create: alloc_srq returned %d\n", status); 5916 *rvalp = (int)status; 5917 retval = 0; 5918 goto cleanup; 5919 } 5920 5921 args.srqc_real_sizes.srqs_sz = srq_real_sizes.srq_wr_sz; 5922 args.srqc_real_sizes.srqs_sgl = srq_real_sizes.srq_sgl_sz; 5923 5924 /* Get HCA-specific data_out info */ 5925 status = ibt_ci_data_out(ia_rp->ia_hca_hdl, 5926 IBT_CI_NO_FLAGS, IBT_HDL_SRQ, (void *)srq_rp->srq_hdl, 5927 &args.srqc_data_out, sizeof (args.srqc_data_out)); 5928 5929 if (status != IBT_SUCCESS) { 5930 DERR("srq_create: ibt_ci_data_out error(%d)\n", status); 5931 *rvalp = (int)status; 5932 retval = 0; 5933 goto cleanup; 5934 } 5935 5936 srq_rp->srq_real_size = srq_real_sizes.srq_wr_sz; 5937 5938 /* preparing to copyout map_data back to the library */ 5939 args.srqc_real_sizes.srqs_sz = srq_real_sizes.srq_wr_sz; 5940 args.srqc_real_sizes.srqs_sgl = srq_real_sizes.srq_sgl_sz; 5941 5942 /* insert into srq hash table */ 5943 retval = daplka_hash_insert(&ia_rp->ia_srq_htbl, 5944 &srq_hkey, (void *)srq_rp); 5945 if (retval != 0) { 5946 DERR("srq_create: cannot insert srq resource into srq_htbl\n"); 5947 goto cleanup; 5948 } 5949 inserted = B_TRUE; 5950 5951 /* return hkey to library */ 5952 args.srqc_hkey = srq_hkey; 5953 5954 retval = ddi_copyout(&args, (void *)arg, sizeof (dapl_srq_create_t), 5955 mode); 5956 if (retval != 0) { 5957 DERR("srq_create: copyout error %d\n", retval); 5958 retval = EFAULT; 5959 goto cleanup; 5960 } 5961 5962 D3("srq_create: %p, 0x%llx\n", srq_rp->srq_hdl, (longlong_t)srq_hkey); 5963 D3(" sz(%d) sgl(%d)\n", 5964 args.srqc_real_sizes.srqs_sz, args.srqc_real_sizes.srqs_sgl); 5965 D3("srq_create: exit\n"); 5966 return (0); 5967 5968 cleanup: 5969 if (inserted) { 5970 daplka_srq_resource_t *free_rp = NULL; 5971 5972 (void) daplka_hash_remove(&ia_rp->ia_srq_htbl, srq_hkey, 5973 (void **)&free_rp); 5974 if (free_rp != srq_rp) { 5975 /* 5976 * this case is impossible because ep_free will 5977 * wait until our state transition is complete. 5978 */ 5979 DERR("srq_create: cannot remove srq from hash table\n"); 5980 ASSERT(B_FALSE); 5981 return (retval); 5982 } 5983 } 5984 DAPLKA_RS_UNREF(srq_rp); 5985 return (retval); 5986 } 5987 5988 /* 5989 * Resize an existing SRQ 5990 */ 5991 /* ARGSUSED */ 5992 static int 5993 daplka_srq_resize(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 5994 cred_t *cred, int *rvalp) 5995 { 5996 daplka_srq_resource_t *srq_rp = NULL; 5997 ibt_hca_attr_t *hca_attrp; 5998 dapl_srq_resize_t args; 5999 ibt_status_t status; 6000 int retval = 0; 6001 6002 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_srq_resize_t), 6003 mode); 6004 if (retval != 0) { 6005 DERR("srq_resize: copyin error %d\n", retval); 6006 return (EFAULT); 6007 } 6008 6009 /* get srq resource */ 6010 srq_rp = (daplka_srq_resource_t *) 6011 daplka_hash_lookup(&ia_rp->ia_srq_htbl, args.srqr_hkey); 6012 if (srq_rp == NULL) { 6013 DERR("srq_resize: cannot find srq resource\n"); 6014 return (EINVAL); 6015 } 6016 ASSERT(DAPLKA_RS_TYPE(srq_rp) == DAPL_TYPE_SRQ); 6017 6018 hca_attrp = &ia_rp->ia_hca->hca_attr; 6019 if (args.srqr_new_size > hca_attrp->hca_max_srqs_sz) { 6020 DERR("srq_resize: invalid srq size %d", args.srqr_new_size); 6021 retval = EINVAL; 6022 goto cleanup; 6023 } 6024 6025 mutex_enter(&srq_rp->srq_lock); 6026 /* 6027 * If ibt_resize_srq fails that it is primarily due to resource 6028 * shortage. Per IB spec resize will never loose events and 6029 * a resize error leaves the SRQ intact. Therefore even if the 6030 * resize request fails we proceed and get the mapping data 6031 * from the SRQ so that the library can mmap it. 6032 */ 6033 status = ibt_modify_srq(srq_rp->srq_hdl, IBT_SRQ_SET_SIZE, 6034 args.srqr_new_size, 0, &args.srqr_real_size); 6035 if (status != IBT_SUCCESS) { 6036 /* we return the size of the old CQ if resize fails */ 6037 args.srqr_real_size = srq_rp->srq_real_size; 6038 ASSERT(status != IBT_SRQ_HDL_INVALID); 6039 DERR("srq_resize: ibt_modify_srq failed:%d\n", status); 6040 } else { 6041 srq_rp->srq_real_size = args.srqr_real_size; 6042 } 6043 mutex_exit(&srq_rp->srq_lock); 6044 6045 6046 D2("srq_resize(%d): done new_sz(%u) real_sz(%u)\n", 6047 DAPLKA_RS_RNUM(srq_rp), args.srqr_new_size, args.srqr_real_size); 6048 6049 /* Get HCA-specific data_out info */ 6050 status = ibt_ci_data_out(srq_rp->srq_hca_hdl, 6051 IBT_CI_NO_FLAGS, IBT_HDL_SRQ, (void *)srq_rp->srq_hdl, 6052 &args.srqr_data_out, sizeof (args.srqr_data_out)); 6053 if (status != IBT_SUCCESS) { 6054 DERR("srq_resize: ibt_ci_data_out error(%d)\n", status); 6055 /* return ibt_ci_data_out status */ 6056 *rvalp = (int)status; 6057 retval = 0; 6058 goto cleanup; 6059 } 6060 6061 retval = ddi_copyout(&args, (void *)arg, sizeof (dapl_srq_resize_t), 6062 mode); 6063 if (retval != 0) { 6064 DERR("srq_resize: copyout error %d\n", retval); 6065 retval = EFAULT; 6066 goto cleanup; 6067 } 6068 6069 cleanup:; 6070 if (srq_rp != NULL) { 6071 DAPLKA_RS_UNREF(srq_rp); 6072 } 6073 return (retval); 6074 } 6075 6076 /* 6077 * Frees an SRQ resource. 6078 */ 6079 /* ARGSUSED */ 6080 static int 6081 daplka_srq_free(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 6082 cred_t *cred, int *rvalp) 6083 { 6084 daplka_srq_resource_t *srq_rp = NULL; 6085 dapl_srq_free_t args; 6086 int retval; 6087 6088 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_srq_free_t), mode); 6089 if (retval != 0) { 6090 DERR("srq_free: copyin error %d\n", retval); 6091 return (EFAULT); 6092 } 6093 6094 retval = daplka_hash_remove(&ia_rp->ia_srq_htbl, 6095 args.srqf_hkey, (void **)&srq_rp); 6096 if (retval != 0 || srq_rp == NULL) { 6097 /* 6098 * this is only possible if we have two threads 6099 * calling ep_free in parallel. 6100 */ 6101 DERR("srq_free: cannot find resource retval(%d) 0x%llx\n", 6102 retval, args.srqf_hkey); 6103 return (EINVAL); 6104 } 6105 6106 /* UNREF calls the actual free function when refcnt is zero */ 6107 DAPLKA_RS_UNREF(srq_rp); 6108 return (0); 6109 } 6110 6111 /* 6112 * destroys a SRQ resource. 6113 * called when refcnt drops to zero. 6114 */ 6115 static int 6116 daplka_srq_destroy(daplka_resource_t *gen_rp) 6117 { 6118 daplka_srq_resource_t *srq_rp = (daplka_srq_resource_t *)gen_rp; 6119 ibt_status_t status; 6120 6121 ASSERT(DAPLKA_RS_REFCNT(srq_rp) == 0); 6122 6123 D3("srq_destroy: entering, srq_rp 0x%p, rnum %d\n", 6124 srq_rp, DAPLKA_RS_RNUM(srq_rp)); 6125 /* 6126 * destroy the srq 6127 */ 6128 if (srq_rp->srq_hdl != NULL) { 6129 status = daplka_ibt_free_srq(srq_rp, srq_rp->srq_hdl); 6130 if (status != IBT_SUCCESS) { 6131 DERR("srq_destroy: ibt_free_srq returned %d\n", 6132 status); 6133 } 6134 srq_rp->srq_hdl = NULL; 6135 D3("srq_destroy: srq freed, rnum %d\n", DAPLKA_RS_RNUM(srq_rp)); 6136 } 6137 /* 6138 * release all references 6139 */ 6140 if (srq_rp->srq_pd_res != NULL) { 6141 DAPLKA_RS_UNREF(srq_rp->srq_pd_res); 6142 srq_rp->srq_pd_res = NULL; 6143 } 6144 6145 mutex_destroy(&srq_rp->srq_lock); 6146 DAPLKA_RS_FINI(srq_rp); 6147 kmem_free(srq_rp, sizeof (daplka_srq_resource_t)); 6148 D3("srq_destroy: exiting, srq_rp 0x%p\n", srq_rp); 6149 return (0); 6150 } 6151 6152 static void 6153 daplka_hash_srq_free(void *obj) 6154 { 6155 daplka_srq_resource_t *srq_rp = (daplka_srq_resource_t *)obj; 6156 6157 ASSERT(DAPLKA_RS_TYPE(srq_rp) == DAPL_TYPE_SRQ); 6158 DAPLKA_RS_UNREF(srq_rp); 6159 } 6160 6161 /* 6162 * This function tells the CM to start listening on a service id. 6163 * It must be called by the passive side client before the client 6164 * can receive connection requests from remote endpoints. If the 6165 * client specifies a non-zero service id (connection qualifier in 6166 * dapl terms), this function will attempt to bind to this service 6167 * id and return an error if the id is already in use. If the client 6168 * specifies zero as the service id, this function will try to find 6169 * the next available service id and return it back to the client. 6170 * To support the cr_handoff function, this function will, in addition 6171 * to creating and inserting an SP resource into the per-IA SP hash 6172 * table, insert the SP resource into a global SP table. This table 6173 * maintains all active service points created by all dapl clients. 6174 * CR handoff locates the target SP by iterating through this global 6175 * table. 6176 */ 6177 /* ARGSUSED */ 6178 static int 6179 daplka_service_register(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 6180 cred_t *cred, int *rvalp) 6181 { 6182 daplka_evd_resource_t *evd_rp = NULL; 6183 daplka_sp_resource_t *sp_rp = NULL; 6184 dapl_service_register_t args; 6185 ibt_srv_desc_t sd_args; 6186 ibt_srv_bind_t sb_args; 6187 ibt_status_t status; 6188 ib_svc_id_t retsid = 0; 6189 uint64_t sp_hkey = 0; 6190 boolean_t bumped = B_FALSE; 6191 int backlog_size; 6192 int retval = 0; 6193 6194 retval = ddi_copyin((void *)arg, &args, 6195 sizeof (dapl_service_register_t), mode); 6196 if (retval != 0) { 6197 DERR("service_register: copyin error %d\n", retval); 6198 return (EINVAL); 6199 } 6200 6201 sp_rp = kmem_zalloc(sizeof (*sp_rp), daplka_km_flags); 6202 if (sp_rp == NULL) { 6203 DERR("service_register: cannot allocate sp resource\n"); 6204 return (ENOMEM); 6205 } 6206 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*sp_rp)) 6207 DAPLKA_RS_INIT(sp_rp, DAPL_TYPE_SP, 6208 DAPLKA_RS_RNUM(ia_rp), daplka_sp_destroy); 6209 6210 /* check if evd exists */ 6211 evd_rp = (daplka_evd_resource_t *) 6212 daplka_hash_lookup(&ia_rp->ia_evd_htbl, args.sr_evd_hkey); 6213 if (evd_rp == NULL) { 6214 DERR("service_register: evd resource not found\n"); 6215 retval = EINVAL; 6216 goto cleanup; 6217 } 6218 /* 6219 * initialize backlog size 6220 */ 6221 if (evd_rp && evd_rp->evd_cq_real_size > 0) { 6222 backlog_size = evd_rp->evd_cq_real_size + 1; 6223 } else { 6224 backlog_size = DAPLKA_DEFAULT_SP_BACKLOG; 6225 } 6226 D2("service_register: args.sr_sid = %llu\n", (longlong_t)args.sr_sid); 6227 6228 /* save the userland sp ptr */ 6229 sp_rp->sp_cookie = args.sr_sp_cookie; 6230 sp_rp->sp_backlog_size = backlog_size; 6231 D3("service_register: backlog set to %d\n", sp_rp->sp_backlog_size); 6232 sp_rp->sp_backlog = kmem_zalloc(sp_rp->sp_backlog_size * 6233 sizeof (daplka_sp_conn_pend_t), daplka_km_flags); 6234 6235 /* save evd resource pointer */ 6236 sp_rp->sp_evd_res = evd_rp; 6237 6238 /* 6239 * save ruid here so that we can do a comparison later 6240 * when someone does cr_handoff. the check will prevent 6241 * a malicious app from passing a CR to us. 6242 */ 6243 sp_rp->sp_ruid = crgetruid(cred); 6244 6245 /* fill in args for register_service */ 6246 sd_args.sd_ud_handler = NULL; 6247 sd_args.sd_handler = daplka_cm_service_handler; 6248 sd_args.sd_flags = IBT_SRV_NO_FLAGS; 6249 6250 status = ibt_register_service(daplka_dev->daplka_clnt_hdl, 6251 &sd_args, args.sr_sid, 1, &sp_rp->sp_srv_hdl, &retsid); 6252 6253 if (status != IBT_SUCCESS) { 6254 DERR("service_register: ibt_register_service returned %d\n", 6255 status); 6256 *rvalp = (int)status; 6257 retval = 0; 6258 goto cleanup; 6259 } 6260 /* save returned sid */ 6261 sp_rp->sp_conn_qual = retsid; 6262 args.sr_retsid = retsid; 6263 6264 /* fill in args for bind_service */ 6265 sb_args.sb_pkey = ia_rp->ia_port_pkey; 6266 sb_args.sb_lease = 0xffffffff; 6267 sb_args.sb_key[0] = 0x1234; 6268 sb_args.sb_key[1] = 0x5678; 6269 sb_args.sb_name = DAPLKA_DRV_NAME; 6270 6271 D2("service_register: bind(0x%llx:0x%llx)\n", 6272 (longlong_t)ia_rp->ia_hca_sgid.gid_prefix, 6273 (longlong_t)ia_rp->ia_hca_sgid.gid_guid); 6274 6275 status = ibt_bind_service(sp_rp->sp_srv_hdl, ia_rp->ia_hca_sgid, 6276 &sb_args, (void *)sp_rp, &sp_rp->sp_bind_hdl); 6277 if (status != IBT_SUCCESS) { 6278 DERR("service_register: ibt_bind_service returned %d\n", 6279 status); 6280 *rvalp = (int)status; 6281 retval = 0; 6282 goto cleanup; 6283 } 6284 6285 /* 6286 * need to bump refcnt because the global hash table will 6287 * have a reference to sp_rp 6288 */ 6289 DAPLKA_RS_REF(sp_rp); 6290 bumped = B_TRUE; 6291 6292 /* insert into global sp hash table */ 6293 sp_rp->sp_global_hkey = 0; 6294 retval = daplka_hash_insert(&daplka_global_sp_htbl, 6295 &sp_rp->sp_global_hkey, (void *)sp_rp); 6296 if (retval != 0) { 6297 DERR("service_register: cannot insert sp resource\n"); 6298 goto cleanup; 6299 } 6300 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(*sp_rp)) 6301 6302 /* insert into per-IA sp hash table */ 6303 retval = daplka_hash_insert(&ia_rp->ia_sp_htbl, 6304 &sp_hkey, (void *)sp_rp); 6305 if (retval != 0) { 6306 DERR("service_register: cannot insert sp resource\n"); 6307 goto cleanup; 6308 } 6309 6310 /* pass index to application */ 6311 args.sr_sp_hkey = sp_hkey; 6312 retval = ddi_copyout(&args, (void *)arg, 6313 sizeof (dapl_service_register_t), mode); 6314 if (retval != 0) { 6315 DERR("service_register: copyout error %d\n", retval); 6316 retval = EFAULT; 6317 goto cleanup; 6318 } 6319 return (0); 6320 6321 cleanup:; 6322 ASSERT(sp_rp != NULL); 6323 /* remove from ia table */ 6324 if (sp_hkey != 0) { 6325 daplka_sp_resource_t *free_rp = NULL; 6326 6327 (void) daplka_hash_remove(&ia_rp->ia_sp_htbl, 6328 sp_hkey, (void **)&free_rp); 6329 if (free_rp != sp_rp) { 6330 DERR("service_register: cannot remove sp\n"); 6331 /* 6332 * we can only get here if another thread 6333 * has completed the cleanup in svc_deregister 6334 */ 6335 return (retval); 6336 } 6337 } 6338 6339 /* remove from global table */ 6340 if (sp_rp->sp_global_hkey != 0) { 6341 daplka_sp_resource_t *free_rp = NULL; 6342 6343 /* 6344 * we get here if either the hash_insert into 6345 * ia_sp_htbl failed or the ddi_copyout failed. 6346 * hash_insert failure implies that we are the 6347 * only thread with a reference to sp. ddi_copyout 6348 * failure implies that svc_deregister could have 6349 * picked up the sp and destroyed it. but since 6350 * we got to this point, we must have removed 6351 * the sp ourselves in hash_remove above and 6352 * that the sp can be destroyed by us. 6353 */ 6354 (void) daplka_hash_remove(&daplka_global_sp_htbl, 6355 sp_rp->sp_global_hkey, (void **)&free_rp); 6356 if (free_rp != sp_rp) { 6357 DERR("service_register: cannot remove sp\n"); 6358 /* 6359 * this case is impossible. see explanation above. 6360 */ 6361 ASSERT(B_FALSE); 6362 return (retval); 6363 } 6364 sp_rp->sp_global_hkey = 0; 6365 } 6366 /* unreference sp */ 6367 if (bumped) { 6368 DAPLKA_RS_UNREF(sp_rp); 6369 } 6370 6371 /* destroy sp resource */ 6372 DAPLKA_RS_UNREF(sp_rp); 6373 return (retval); 6374 } 6375 6376 /* 6377 * deregisters the service and removes SP from the global table. 6378 */ 6379 /* ARGSUSED */ 6380 static int 6381 daplka_service_deregister(daplka_ia_resource_t *ia_rp, intptr_t arg, int mode, 6382 cred_t *cred, int *rvalp) 6383 { 6384 dapl_service_deregister_t args; 6385 daplka_sp_resource_t *sp_rp = NULL, *g_sp_rp = NULL; 6386 int retval; 6387 6388 retval = ddi_copyin((void *)arg, &args, 6389 sizeof (dapl_service_deregister_t), mode); 6390 6391 if (retval != 0) { 6392 DERR("service_deregister: copyin error %d\n", retval); 6393 return (EINVAL); 6394 } 6395 6396 retval = daplka_hash_remove(&ia_rp->ia_sp_htbl, 6397 args.sdr_sp_hkey, (void **)&sp_rp); 6398 if (retval != 0 || sp_rp == NULL) { 6399 DERR("service_deregister: cannot find sp resource\n"); 6400 return (EINVAL); 6401 } 6402 6403 retval = daplka_hash_remove(&daplka_global_sp_htbl, 6404 sp_rp->sp_global_hkey, (void **)&g_sp_rp); 6405 if (retval != 0 || g_sp_rp == NULL) { 6406 DERR("service_deregister: cannot find sp resource\n"); 6407 } 6408 6409 /* remove the global reference */ 6410 if (g_sp_rp == sp_rp) { 6411 DAPLKA_RS_UNREF(g_sp_rp); 6412 } 6413 6414 DAPLKA_RS_UNREF(sp_rp); 6415 return (0); 6416 } 6417 6418 /* 6419 * destroys a service point. 6420 * called when the refcnt drops to zero. 6421 */ 6422 static int 6423 daplka_sp_destroy(daplka_resource_t *gen_rp) 6424 { 6425 daplka_sp_resource_t *sp_rp = (daplka_sp_resource_t *)gen_rp; 6426 ibt_status_t status; 6427 6428 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*sp_rp)) 6429 ASSERT(DAPLKA_RS_REFCNT(sp_rp) == 0); 6430 D3("sp_destroy: entering, sp_rp %p, rnum %d\n", 6431 sp_rp, DAPLKA_RS_RNUM(sp_rp)); 6432 6433 /* 6434 * it is possible for pending connections to remain 6435 * on an SP. We need to clean them up here. 6436 */ 6437 if (sp_rp->sp_backlog != NULL) { 6438 ibt_cm_proceed_reply_t proc_reply; 6439 int i, cnt = 0; 6440 void *spcp_sidp; 6441 6442 for (i = 0; i < sp_rp->sp_backlog_size; i++) { 6443 if (sp_rp->sp_backlog[i].spcp_state == 6444 DAPLKA_SPCP_PENDING) { 6445 cnt++; 6446 if (sp_rp->sp_backlog[i].spcp_sid == NULL) { 6447 DERR("sp_destroy: " 6448 "spcp_sid == NULL!\n"); 6449 continue; 6450 } 6451 mutex_enter(&sp_rp->sp_lock); 6452 spcp_sidp = sp_rp->sp_backlog[i].spcp_sid; 6453 sp_rp->sp_backlog[i].spcp_state = 6454 DAPLKA_SPCP_INIT; 6455 sp_rp->sp_backlog[i].spcp_sid = NULL; 6456 sp_rp->sp_backlog[i].spcp_req_len = 0; 6457 mutex_exit(&sp_rp->sp_lock); 6458 status = ibt_cm_proceed(IBT_CM_EVENT_REQ_RCV, 6459 spcp_sidp, 6460 IBT_CM_NO_RESOURCE, &proc_reply, NULL, 0); 6461 if (status != IBT_SUCCESS) { 6462 DERR("sp_destroy: proceed failed %d\n", 6463 status); 6464 } 6465 } 6466 } 6467 if (cnt > 0) { 6468 DERR("sp_destroy: found %d pending " 6469 "connections\n", cnt); 6470 } 6471 } 6472 6473 if (sp_rp->sp_srv_hdl != NULL && sp_rp->sp_bind_hdl != NULL) { 6474 status = ibt_unbind_service(sp_rp->sp_srv_hdl, 6475 sp_rp->sp_bind_hdl); 6476 if (status != IBT_SUCCESS) { 6477 DERR("sp_destroy: ibt_unbind_service " 6478 "failed: %d\n", status); 6479 } 6480 } 6481 6482 if (sp_rp->sp_srv_hdl != NULL) { 6483 status = ibt_deregister_service(daplka_dev->daplka_clnt_hdl, 6484 sp_rp->sp_srv_hdl); 6485 if (status != IBT_SUCCESS) { 6486 DERR("sp_destroy: ibt_deregister_service " 6487 "failed: %d\n", status); 6488 } 6489 } 6490 if (sp_rp->sp_backlog != NULL) { 6491 kmem_free(sp_rp->sp_backlog, 6492 sp_rp->sp_backlog_size * sizeof (daplka_sp_conn_pend_t)); 6493 sp_rp->sp_backlog = NULL; 6494 sp_rp->sp_backlog_size = 0; 6495 } 6496 6497 /* 6498 * release reference to evd 6499 */ 6500 if (sp_rp->sp_evd_res != NULL) { 6501 DAPLKA_RS_UNREF(sp_rp->sp_evd_res); 6502 } 6503 sp_rp->sp_bind_hdl = NULL; 6504 sp_rp->sp_srv_hdl = NULL; 6505 DAPLKA_RS_FINI(sp_rp); 6506 kmem_free(sp_rp, sizeof (*sp_rp)); 6507 D3("sp_destroy: exiting, sp_rp %p\n", sp_rp); 6508 return (0); 6509 } 6510 6511 /* 6512 * this function is called by daplka_hash_destroy for 6513 * freeing SP resource objects 6514 */ 6515 static void 6516 daplka_hash_sp_free(void *obj) 6517 { 6518 daplka_sp_resource_t *sp_rp = (daplka_sp_resource_t *)obj; 6519 daplka_sp_resource_t *g_sp_rp; 6520 int retval; 6521 6522 ASSERT(DAPLKA_RS_TYPE(sp_rp) == DAPL_TYPE_SP); 6523 6524 retval = daplka_hash_remove(&daplka_global_sp_htbl, 6525 sp_rp->sp_global_hkey, (void **)&g_sp_rp); 6526 if (retval != 0 || g_sp_rp == NULL) { 6527 DERR("sp_free: cannot find sp resource\n"); 6528 } 6529 if (g_sp_rp == sp_rp) { 6530 DAPLKA_RS_UNREF(g_sp_rp); 6531 } 6532 6533 DAPLKA_RS_UNREF(sp_rp); 6534 } 6535 6536 static void 6537 daplka_hash_sp_unref(void *obj) 6538 { 6539 daplka_sp_resource_t *sp_rp = (daplka_sp_resource_t *)obj; 6540 6541 ASSERT(DAPLKA_RS_TYPE(sp_rp) == DAPL_TYPE_SP); 6542 DAPLKA_RS_UNREF(sp_rp); 6543 } 6544 6545 /* 6546 * Passive side CM handlers 6547 */ 6548 6549 /* 6550 * processes the REQ_RCV event 6551 */ 6552 /* ARGSUSED */ 6553 static ibt_cm_status_t 6554 daplka_cm_service_req(daplka_sp_resource_t *spp, ibt_cm_event_t *event, 6555 ibt_cm_return_args_t *ret_args, void *pr_data, ibt_priv_data_len_t pr_len) 6556 { 6557 daplka_sp_conn_pend_t *conn = NULL; 6558 daplka_evd_event_t *cr_ev = NULL; 6559 ibt_cm_status_t cm_status = IBT_CM_DEFAULT; 6560 uint16_t bkl_index; 6561 ibt_status_t status; 6562 6563 /* 6564 * acquire a slot in the connection backlog of this service point 6565 */ 6566 mutex_enter(&spp->sp_lock); 6567 for (bkl_index = 0; bkl_index < spp->sp_backlog_size; bkl_index++) { 6568 if (spp->sp_backlog[bkl_index].spcp_state == DAPLKA_SPCP_INIT) { 6569 conn = &spp->sp_backlog[bkl_index]; 6570 ASSERT(conn->spcp_sid == NULL); 6571 conn->spcp_state = DAPLKA_SPCP_PENDING; 6572 conn->spcp_sid = event->cm_session_id; 6573 break; 6574 } 6575 } 6576 mutex_exit(&spp->sp_lock); 6577 6578 /* 6579 * too many pending connections 6580 */ 6581 if (bkl_index == spp->sp_backlog_size) { 6582 DERR("service_req: connection pending exceeded %d limit\n", 6583 spp->sp_backlog_size); 6584 return (IBT_CM_NO_RESOURCE); 6585 } 6586 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*conn)) 6587 6588 /* 6589 * save data for cr_handoff 6590 */ 6591 if (pr_data != NULL && pr_len > 0) { 6592 int trunc_len = pr_len; 6593 6594 if (trunc_len > DAPL_MAX_PRIVATE_DATA_SIZE) { 6595 DERR("service_req: private data truncated\n"); 6596 trunc_len = DAPL_MAX_PRIVATE_DATA_SIZE; 6597 } 6598 conn->spcp_req_len = trunc_len; 6599 bcopy(pr_data, conn->spcp_req_data, trunc_len); 6600 } else { 6601 conn->spcp_req_len = 0; 6602 } 6603 conn->spcp_rdma_ra_in = event->cm_event.req.req_rdma_ra_in; 6604 conn->spcp_rdma_ra_out = event->cm_event.req.req_rdma_ra_out; 6605 6606 /* 6607 * create a CR event 6608 */ 6609 cr_ev = kmem_zalloc(sizeof (daplka_evd_event_t), KM_NOSLEEP); 6610 if (cr_ev == NULL) { 6611 DERR("service_req: could not alloc cr_ev\n"); 6612 cm_status = IBT_CM_NO_RESOURCE; 6613 goto cleanup; 6614 } 6615 6616 cr_ev->ee_next = NULL; 6617 cr_ev->ee_cmev.ec_cm_cookie = spp->sp_cookie; 6618 cr_ev->ee_cmev.ec_cm_is_passive = B_TRUE; 6619 cr_ev->ee_cmev.ec_cm_psep_cookie = DAPLKA_CREATE_PSEP_COOKIE(bkl_index); 6620 /* 6621 * save the requestor gid 6622 * daplka_event_poll needs this if this is a third party REQ_RCV 6623 */ 6624 cr_ev->ee_cmev.ec_cm_req_prim_addr.gid_prefix = 6625 event->cm_event.req.req_prim_addr.av_dgid.gid_prefix; 6626 cr_ev->ee_cmev.ec_cm_req_prim_addr.gid_guid = 6627 event->cm_event.req.req_prim_addr.av_dgid.gid_guid; 6628 6629 /* 6630 * set event type 6631 */ 6632 if (pr_len == 0) { 6633 cr_ev->ee_cmev.ec_cm_ev_type = 6634 DAPL_IB_CME_CONNECTION_REQUEST_PENDING; 6635 } else { 6636 cr_ev->ee_cmev.ec_cm_ev_priv_data = 6637 kmem_zalloc(pr_len, KM_NOSLEEP); 6638 if (cr_ev->ee_cmev.ec_cm_ev_priv_data == NULL) { 6639 DERR("service_req: could not alloc priv\n"); 6640 cm_status = IBT_CM_NO_RESOURCE; 6641 goto cleanup; 6642 } 6643 bcopy(pr_data, cr_ev->ee_cmev.ec_cm_ev_priv_data, pr_len); 6644 cr_ev->ee_cmev.ec_cm_ev_type = 6645 DAPL_IB_CME_CONNECTION_REQUEST_PENDING_PRIVATE_DATA; 6646 } 6647 cr_ev->ee_cmev.ec_cm_ev_priv_data_len = pr_len; 6648 6649 /* 6650 * tell the active side to expect the processing time to be 6651 * at most equal to daplka_cm_delay 6652 */ 6653 status = ibt_cm_delay(IBT_CM_DELAY_REQ, event->cm_session_id, 6654 daplka_cm_delay, NULL, 0); 6655 if (status != IBT_SUCCESS) { 6656 DERR("service_req: ibt_cm_delay failed %d\n", status); 6657 cm_status = IBT_CM_NO_RESOURCE; 6658 goto cleanup; 6659 } 6660 6661 /* 6662 * enqueue cr_ev onto the cr_events list of the EVD 6663 * corresponding to the SP 6664 */ 6665 D2("service_req: enqueue event(%p) evdp(%p) priv_data(%p) " 6666 "priv_len(%d) psep(0x%llx)\n", cr_ev, spp->sp_evd_res, 6667 cr_ev->ee_cmev.ec_cm_ev_priv_data, 6668 (int)cr_ev->ee_cmev.ec_cm_ev_priv_data_len, 6669 (longlong_t)cr_ev->ee_cmev.ec_cm_psep_cookie); 6670 6671 daplka_evd_wakeup(spp->sp_evd_res, 6672 &spp->sp_evd_res->evd_cr_events, cr_ev); 6673 6674 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(*conn)) 6675 return (IBT_CM_DEFER); 6676 6677 cleanup:; 6678 /* 6679 * free the cr event 6680 */ 6681 if (cr_ev != NULL) { 6682 if (cr_ev->ee_cmev.ec_cm_ev_priv_data != NULL) { 6683 kmem_free(cr_ev->ee_cmev.ec_cm_ev_priv_data, pr_len); 6684 cr_ev->ee_cmev.ec_cm_ev_priv_data = NULL; 6685 cr_ev->ee_cmev.ec_cm_ev_priv_data_len = 0; 6686 } 6687 kmem_free(cr_ev, sizeof (daplka_evd_event_t)); 6688 } 6689 /* 6690 * release our slot in the backlog array 6691 */ 6692 if (conn != NULL) { 6693 mutex_enter(&spp->sp_lock); 6694 ASSERT(conn->spcp_state == DAPLKA_SPCP_PENDING); 6695 ASSERT(conn->spcp_sid == event->cm_session_id); 6696 conn->spcp_state = DAPLKA_SPCP_INIT; 6697 conn->spcp_req_len = 0; 6698 conn->spcp_sid = NULL; 6699 mutex_exit(&spp->sp_lock); 6700 } 6701 return (cm_status); 6702 } 6703 6704 /* 6705 * processes the CONN_CLOSED event 6706 */ 6707 /* ARGSUSED */ 6708 static ibt_cm_status_t 6709 daplka_cm_service_conn_closed(daplka_sp_resource_t *sp_rp, 6710 ibt_cm_event_t *event, ibt_cm_return_args_t *ret_args, 6711 void *priv_data, ibt_priv_data_len_t len) 6712 { 6713 daplka_ep_resource_t *ep_rp; 6714 daplka_evd_event_t *disc_ev; 6715 uint32_t old_state, new_state; 6716 6717 ep_rp = (daplka_ep_resource_t *) 6718 ibt_get_chan_private(event->cm_channel); 6719 if (ep_rp == NULL) { 6720 DERR("service_conn_closed: ep_rp == NULL\n"); 6721 return (IBT_CM_ACCEPT); 6722 } 6723 6724 /* 6725 * verify that the ep_state is either CONNECTED or 6726 * DISCONNECTING. if it is not in either states return 6727 * without generating an event. 6728 */ 6729 new_state = old_state = daplka_ep_get_state(ep_rp); 6730 if (old_state != DAPLKA_EP_STATE_CONNECTED && 6731 old_state != DAPLKA_EP_STATE_DISCONNECTING) { 6732 /* 6733 * we can get here if the connection is being aborted 6734 */ 6735 D2("service_conn_closed: conn aborted, state = %d, " 6736 "closed = %d\n", old_state, (int)event->cm_event.closed); 6737 daplka_ep_set_state(ep_rp, old_state, new_state); 6738 return (IBT_CM_ACCEPT); 6739 } 6740 6741 /* 6742 * create a DAPL_IB_CME_DISCONNECTED event 6743 */ 6744 disc_ev = kmem_zalloc(sizeof (daplka_evd_event_t), KM_NOSLEEP); 6745 if (disc_ev == NULL) { 6746 DERR("service_conn_closed: cannot alloc disc_ev\n"); 6747 daplka_ep_set_state(ep_rp, old_state, new_state); 6748 return (IBT_CM_ACCEPT); 6749 } 6750 6751 disc_ev->ee_cmev.ec_cm_ev_type = DAPL_IB_CME_DISCONNECTED; 6752 disc_ev->ee_cmev.ec_cm_cookie = sp_rp->sp_cookie; 6753 disc_ev->ee_cmev.ec_cm_is_passive = B_TRUE; 6754 disc_ev->ee_cmev.ec_cm_psep_cookie = ep_rp->ep_psep_cookie; 6755 disc_ev->ee_cmev.ec_cm_ev_priv_data = NULL; 6756 disc_ev->ee_cmev.ec_cm_ev_priv_data_len = 0; 6757 6758 D2("service_conn_closed: enqueue event(%p) evdp(%p) psep(0x%llx)\n", 6759 disc_ev, sp_rp->sp_evd_res, (longlong_t)ep_rp->ep_psep_cookie); 6760 6761 /* 6762 * transition ep_state to DISCONNECTED 6763 */ 6764 new_state = DAPLKA_EP_STATE_DISCONNECTED; 6765 daplka_ep_set_state(ep_rp, old_state, new_state); 6766 6767 /* 6768 * enqueue event onto the conn_evd owned by ep_rp 6769 */ 6770 daplka_evd_wakeup(ep_rp->ep_conn_evd, 6771 &ep_rp->ep_conn_evd->evd_conn_events, disc_ev); 6772 6773 return (IBT_CM_ACCEPT); 6774 } 6775 6776 /* 6777 * processes the CONN_EST event 6778 */ 6779 /* ARGSUSED */ 6780 static ibt_cm_status_t 6781 daplka_cm_service_conn_est(daplka_sp_resource_t *sp_rp, ibt_cm_event_t *event, 6782 ibt_cm_return_args_t *ret_args, void *priv_data, ibt_priv_data_len_t len) 6783 { 6784 daplka_ep_resource_t *ep_rp; 6785 daplka_evd_event_t *conn_ev; 6786 void *pr_data = event->cm_priv_data; 6787 ibt_priv_data_len_t pr_len = event->cm_priv_data_len; 6788 uint32_t old_state, new_state; 6789 6790 ep_rp = (daplka_ep_resource_t *) 6791 ibt_get_chan_private(event->cm_channel); 6792 if (ep_rp == NULL) { 6793 DERR("service_conn_est: ep_rp == NULL\n"); 6794 return (IBT_CM_ACCEPT); 6795 } 6796 6797 /* 6798 * verify that ep_state is ACCEPTING. if it is not in this 6799 * state, return without generating an event. 6800 */ 6801 new_state = old_state = daplka_ep_get_state(ep_rp); 6802 if (old_state != DAPLKA_EP_STATE_ACCEPTING) { 6803 /* 6804 * we can get here if the connection is being aborted 6805 */ 6806 DERR("service_conn_est: conn aborted, state = %d\n", 6807 old_state); 6808 daplka_ep_set_state(ep_rp, old_state, new_state); 6809 return (IBT_CM_ACCEPT); 6810 } 6811 6812 /* 6813 * create a DAPL_IB_CME_CONNECTED event 6814 */ 6815 conn_ev = kmem_zalloc(sizeof (daplka_evd_event_t), KM_NOSLEEP); 6816 if (conn_ev == NULL) { 6817 DERR("service_conn_est: conn_ev alloc failed\n"); 6818 daplka_ep_set_state(ep_rp, old_state, new_state); 6819 return (IBT_CM_ACCEPT); 6820 } 6821 6822 conn_ev->ee_cmev.ec_cm_ev_type = DAPL_IB_CME_CONNECTED; 6823 conn_ev->ee_cmev.ec_cm_cookie = sp_rp->sp_cookie; 6824 conn_ev->ee_cmev.ec_cm_is_passive = B_TRUE; 6825 conn_ev->ee_cmev.ec_cm_psep_cookie = ep_rp->ep_psep_cookie; 6826 6827 /* 6828 * copy private data into event 6829 */ 6830 if (pr_len > 0) { 6831 conn_ev->ee_cmev.ec_cm_ev_priv_data = 6832 kmem_zalloc(pr_len, KM_NOSLEEP); 6833 if (conn_ev->ee_cmev.ec_cm_ev_priv_data == NULL) { 6834 DERR("service_conn_est: pr_data alloc failed\n"); 6835 daplka_ep_set_state(ep_rp, old_state, new_state); 6836 kmem_free(conn_ev, sizeof (daplka_evd_event_t)); 6837 return (IBT_CM_ACCEPT); 6838 } 6839 bcopy(pr_data, conn_ev->ee_cmev.ec_cm_ev_priv_data, pr_len); 6840 } 6841 conn_ev->ee_cmev.ec_cm_ev_priv_data_len = pr_len; 6842 6843 D2("service_conn_est: enqueue event(%p) evdp(%p)\n", 6844 conn_ev, ep_rp->ep_conn_evd); 6845 6846 /* 6847 * transition ep_state to CONNECTED 6848 */ 6849 new_state = DAPLKA_EP_STATE_CONNECTED; 6850 daplka_ep_set_state(ep_rp, old_state, new_state); 6851 6852 /* 6853 * enqueue event onto the conn_evd owned by ep_rp 6854 */ 6855 daplka_evd_wakeup(ep_rp->ep_conn_evd, 6856 &ep_rp->ep_conn_evd->evd_conn_events, conn_ev); 6857 6858 return (IBT_CM_ACCEPT); 6859 } 6860 6861 /* 6862 * processes the FAILURE event 6863 */ 6864 /* ARGSUSED */ 6865 static ibt_cm_status_t 6866 daplka_cm_service_event_failure(daplka_sp_resource_t *sp_rp, 6867 ibt_cm_event_t *event, ibt_cm_return_args_t *ret_args, void *priv_data, 6868 ibt_priv_data_len_t len) 6869 { 6870 daplka_evd_event_t *disc_ev; 6871 daplka_ep_resource_t *ep_rp; 6872 uint32_t old_state, new_state; 6873 ibt_rc_chan_query_attr_t chan_attrs; 6874 ibt_status_t status; 6875 6876 /* 6877 * check that we still have a valid cm_channel before continuing 6878 */ 6879 if (event->cm_channel == NULL) { 6880 DERR("serice_event_failure: event->cm_channel == NULL\n"); 6881 return (IBT_CM_ACCEPT); 6882 } 6883 ep_rp = (daplka_ep_resource_t *) 6884 ibt_get_chan_private(event->cm_channel); 6885 if (ep_rp == NULL) { 6886 DERR("service_event_failure: ep_rp == NULL\n"); 6887 return (IBT_CM_ACCEPT); 6888 } 6889 6890 /* 6891 * verify that ep_state is ACCEPTING or DISCONNECTING. if it 6892 * is not in either state, return without generating an event. 6893 */ 6894 new_state = old_state = daplka_ep_get_state(ep_rp); 6895 if (old_state != DAPLKA_EP_STATE_ACCEPTING && 6896 old_state != DAPLKA_EP_STATE_DISCONNECTING) { 6897 /* 6898 * we can get here if the connection is being aborted 6899 */ 6900 DERR("service_event_failure: conn aborted, state = %d, " 6901 "cf_code = %d, cf_msg = %d, cf_reason = %d\n", old_state, 6902 (int)event->cm_event.failed.cf_code, 6903 (int)event->cm_event.failed.cf_msg, 6904 (int)event->cm_event.failed.cf_reason); 6905 6906 daplka_ep_set_state(ep_rp, old_state, new_state); 6907 return (IBT_CM_ACCEPT); 6908 } 6909 6910 bzero(&chan_attrs, sizeof (ibt_rc_chan_query_attr_t)); 6911 status = ibt_query_rc_channel(ep_rp->ep_chan_hdl, &chan_attrs); 6912 6913 if ((status == IBT_SUCCESS) && 6914 (chan_attrs.rc_state != IBT_STATE_ERROR)) { 6915 DERR("service_event_failure: conn abort qpn %d state %d\n", 6916 chan_attrs.rc_qpn, chan_attrs.rc_state); 6917 6918 /* explicit transition the QP to ERROR state */ 6919 status = ibt_flush_channel(ep_rp->ep_chan_hdl); 6920 } 6921 6922 /* 6923 * create an event 6924 */ 6925 disc_ev = kmem_zalloc(sizeof (daplka_evd_event_t), KM_NOSLEEP); 6926 if (disc_ev == NULL) { 6927 DERR("service_event_failure: cannot alloc disc_ev\n"); 6928 daplka_ep_set_state(ep_rp, old_state, new_state); 6929 return (IBT_CM_ACCEPT); 6930 } 6931 6932 /* 6933 * fill in the appropriate event type 6934 */ 6935 if (event->cm_event.failed.cf_code == IBT_CM_FAILURE_TIMEOUT) { 6936 disc_ev->ee_cmev.ec_cm_ev_type = DAPL_IB_CME_TIMED_OUT; 6937 } else if (event->cm_event.failed.cf_code == IBT_CM_FAILURE_REJ_RCV) { 6938 switch (event->cm_event.failed.cf_reason) { 6939 case IBT_CM_INVALID_CID: 6940 disc_ev->ee_cmev.ec_cm_ev_type = 6941 DAPL_IB_CME_DESTINATION_REJECT; 6942 break; 6943 default: 6944 disc_ev->ee_cmev.ec_cm_ev_type = 6945 DAPL_IB_CME_LOCAL_FAILURE; 6946 break; 6947 } 6948 } else { 6949 disc_ev->ee_cmev.ec_cm_ev_type = DAPL_IB_CME_LOCAL_FAILURE; 6950 } 6951 disc_ev->ee_cmev.ec_cm_cookie = sp_rp->sp_cookie; 6952 disc_ev->ee_cmev.ec_cm_is_passive = B_TRUE; 6953 disc_ev->ee_cmev.ec_cm_psep_cookie = ep_rp->ep_psep_cookie; 6954 disc_ev->ee_cmev.ec_cm_ev_priv_data_len = 0; 6955 disc_ev->ee_cmev.ec_cm_ev_priv_data = NULL; 6956 6957 D2("service_event_failure: enqueue event(%p) evdp(%p) cf_code(%d) " 6958 "cf_msg(%d) cf_reason(%d) psep(0x%llx)\n", disc_ev, 6959 ep_rp->ep_conn_evd, (int)event->cm_event.failed.cf_code, 6960 (int)event->cm_event.failed.cf_msg, 6961 (int)event->cm_event.failed.cf_reason, 6962 (longlong_t)ep_rp->ep_psep_cookie); 6963 6964 /* 6965 * transition ep_state to DISCONNECTED 6966 */ 6967 new_state = DAPLKA_EP_STATE_DISCONNECTED; 6968 daplka_ep_set_state(ep_rp, old_state, new_state); 6969 6970 /* 6971 * enqueue event onto the conn_evd owned by ep_rp 6972 */ 6973 daplka_evd_wakeup(ep_rp->ep_conn_evd, 6974 &ep_rp->ep_conn_evd->evd_conn_events, disc_ev); 6975 6976 return (IBT_CM_ACCEPT); 6977 } 6978 6979 /* 6980 * this is the passive side CM handler. it gets registered 6981 * when an SP resource is created in daplka_service_register. 6982 */ 6983 static ibt_cm_status_t 6984 daplka_cm_service_handler(void *cm_private, ibt_cm_event_t *event, 6985 ibt_cm_return_args_t *ret_args, void *priv_data, ibt_priv_data_len_t len) 6986 { 6987 daplka_sp_resource_t *sp_rp = (daplka_sp_resource_t *)cm_private; 6988 6989 if (sp_rp == NULL) { 6990 DERR("service_handler: sp_rp == NULL\n"); 6991 return (IBT_CM_NO_RESOURCE); 6992 } 6993 /* 6994 * default is not to return priv data 6995 */ 6996 if (ret_args != NULL) { 6997 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*ret_args)) 6998 ret_args->cm_ret_len = 0; 6999 } 7000 7001 switch (event->cm_type) { 7002 case IBT_CM_EVENT_REQ_RCV: 7003 D2("service_handler: IBT_CM_EVENT_REQ_RCV\n"); 7004 return (daplka_cm_service_req(sp_rp, event, ret_args, 7005 event->cm_priv_data, event->cm_priv_data_len)); 7006 7007 case IBT_CM_EVENT_REP_RCV: 7008 /* passive side should not receive this event */ 7009 D2("service_handler: IBT_CM_EVENT_REP_RCV\n"); 7010 return (IBT_CM_DEFAULT); 7011 7012 case IBT_CM_EVENT_CONN_CLOSED: 7013 D2("service_handler: IBT_CM_EVENT_CONN_CLOSED %d\n", 7014 event->cm_event.closed); 7015 return (daplka_cm_service_conn_closed(sp_rp, event, ret_args, 7016 priv_data, len)); 7017 7018 case IBT_CM_EVENT_MRA_RCV: 7019 /* passive side does default processing MRA event */ 7020 D2("service_handler: IBT_CM_EVENT_MRA_RCV\n"); 7021 return (IBT_CM_DEFAULT); 7022 7023 case IBT_CM_EVENT_CONN_EST: 7024 D2("service_handler: IBT_CM_EVENT_CONN_EST\n"); 7025 return (daplka_cm_service_conn_est(sp_rp, event, ret_args, 7026 priv_data, len)); 7027 7028 case IBT_CM_EVENT_FAILURE: 7029 D2("service_handler: IBT_CM_EVENT_FAILURE\n"); 7030 return (daplka_cm_service_event_failure(sp_rp, event, ret_args, 7031 priv_data, len)); 7032 case IBT_CM_EVENT_LAP_RCV: 7033 /* active side had initiated a path migration operation */ 7034 D2("service_handler: IBT_CM_EVENT_LAP_RCV\n"); 7035 return (IBT_CM_ACCEPT); 7036 default: 7037 DERR("service_handler: invalid event %d\n", event->cm_type); 7038 break; 7039 } 7040 return (IBT_CM_DEFAULT); 7041 } 7042 7043 /* 7044 * Active side CM handlers 7045 */ 7046 7047 /* 7048 * Processes the REP_RCV event. When the passive side accepts the 7049 * connection, this handler is called. We make a copy of the private 7050 * data into the ep so that it can be passed back to userland in when 7051 * the CONN_EST event occurs. 7052 */ 7053 /* ARGSUSED */ 7054 static ibt_cm_status_t 7055 daplka_cm_rc_rep_rcv(daplka_ep_resource_t *ep_rp, ibt_cm_event_t *event, 7056 ibt_cm_return_args_t *ret_args, void *priv_data, ibt_priv_data_len_t len) 7057 { 7058 void *pr_data = event->cm_priv_data; 7059 ibt_priv_data_len_t pr_len = event->cm_priv_data_len; 7060 uint32_t old_state, new_state; 7061 7062 D2("rc_rep_rcv: pr_data(0x%p), pr_len(%d)\n", pr_data, 7063 (int)pr_len); 7064 7065 ASSERT(ep_rp != NULL); 7066 new_state = old_state = daplka_ep_get_state(ep_rp); 7067 if (old_state != DAPLKA_EP_STATE_CONNECTING) { 7068 /* 7069 * we can get here if the connection is being aborted 7070 */ 7071 DERR("rc_rep_rcv: conn aborted, state = %d\n", old_state); 7072 daplka_ep_set_state(ep_rp, old_state, new_state); 7073 return (IBT_CM_NO_CHANNEL); 7074 } 7075 7076 /* 7077 * we do not cancel the timer here because the connection 7078 * handshake is still in progress. 7079 */ 7080 7081 /* 7082 * save the private data. it will be passed up when 7083 * the connection is established. 7084 */ 7085 if (pr_len > 0) { 7086 ep_rp->ep_priv_len = pr_len; 7087 bcopy(pr_data, ep_rp->ep_priv_data, (size_t)pr_len); 7088 } 7089 7090 /* 7091 * we do not actually transition to a different state. 7092 * the state will change when we get a conn_est, failure, 7093 * closed, or timeout event. 7094 */ 7095 daplka_ep_set_state(ep_rp, old_state, new_state); 7096 return (IBT_CM_ACCEPT); 7097 } 7098 7099 /* 7100 * Processes the CONN_CLOSED event. This gets called when either 7101 * the active or passive side closes the rc channel. 7102 */ 7103 /* ARGSUSED */ 7104 static ibt_cm_status_t 7105 daplka_cm_rc_conn_closed(daplka_ep_resource_t *ep_rp, ibt_cm_event_t *event, 7106 ibt_cm_return_args_t *ret_args, void *priv_data, ibt_priv_data_len_t len) 7107 { 7108 daplka_evd_event_t *disc_ev; 7109 uint32_t old_state, new_state; 7110 7111 ASSERT(ep_rp != NULL); 7112 old_state = new_state = daplka_ep_get_state(ep_rp); 7113 if (old_state != DAPLKA_EP_STATE_CONNECTED && 7114 old_state != DAPLKA_EP_STATE_DISCONNECTING) { 7115 /* 7116 * we can get here if the connection is being aborted 7117 */ 7118 D2("rc_conn_closed: conn aborted, state = %d, " 7119 "closed = %d\n", old_state, (int)event->cm_event.closed); 7120 daplka_ep_set_state(ep_rp, old_state, new_state); 7121 return (IBT_CM_ACCEPT); 7122 } 7123 7124 /* 7125 * it's ok for the timer to fire at this point. the 7126 * taskq thread that processes the timer will just wait 7127 * until we are done with our state transition. 7128 */ 7129 if (daplka_cancel_timer(ep_rp) != 0) { 7130 /* 7131 * daplka_cancel_timer returns -1 if the timer is 7132 * being processed and 0 for all other cases. 7133 * we need to reset ep_state to allow timer processing 7134 * to continue. 7135 */ 7136 DERR("rc_conn_closed: timer is being processed\n"); 7137 daplka_ep_set_state(ep_rp, old_state, new_state); 7138 return (IBT_CM_ACCEPT); 7139 } 7140 7141 /* 7142 * create a DAPL_IB_CME_DISCONNECTED event 7143 */ 7144 disc_ev = kmem_zalloc(sizeof (daplka_evd_event_t), KM_NOSLEEP); 7145 if (disc_ev == NULL) { 7146 DERR("rc_conn_closed: could not alloc ev\n"); 7147 daplka_ep_set_state(ep_rp, old_state, new_state); 7148 return (IBT_CM_ACCEPT); 7149 } 7150 7151 disc_ev->ee_cmev.ec_cm_ev_type = DAPL_IB_CME_DISCONNECTED; 7152 disc_ev->ee_cmev.ec_cm_cookie = ep_rp->ep_cookie; 7153 disc_ev->ee_cmev.ec_cm_is_passive = B_FALSE; 7154 disc_ev->ee_cmev.ec_cm_psep_cookie = 0; 7155 disc_ev->ee_cmev.ec_cm_ev_priv_data = NULL; 7156 disc_ev->ee_cmev.ec_cm_ev_priv_data_len = 0; 7157 7158 D2("rc_conn_closed: enqueue event(%p) evdp(%p) closed(%d)\n", 7159 disc_ev, ep_rp->ep_conn_evd, (int)event->cm_event.closed); 7160 7161 /* 7162 * transition ep_state to DISCONNECTED 7163 */ 7164 new_state = DAPLKA_EP_STATE_DISCONNECTED; 7165 daplka_ep_set_state(ep_rp, old_state, new_state); 7166 7167 /* 7168 * enqueue event onto the conn_evd owned by ep_rp 7169 */ 7170 daplka_evd_wakeup(ep_rp->ep_conn_evd, 7171 &ep_rp->ep_conn_evd->evd_conn_events, disc_ev); 7172 7173 return (IBT_CM_ACCEPT); 7174 } 7175 7176 /* 7177 * processes the CONN_EST event 7178 */ 7179 /* ARGSUSED */ 7180 static ibt_cm_status_t 7181 daplka_cm_rc_conn_est(daplka_ep_resource_t *ep_rp, ibt_cm_event_t *event, 7182 ibt_cm_return_args_t *ret_args, void *priv_data, ibt_priv_data_len_t len) 7183 { 7184 daplka_evd_event_t *conn_ev; 7185 uint32_t old_state, new_state; 7186 7187 ASSERT(ep_rp != NULL); 7188 old_state = new_state = daplka_ep_get_state(ep_rp); 7189 if (old_state != DAPLKA_EP_STATE_CONNECTING) { 7190 /* 7191 * we can get here if the connection is being aborted 7192 */ 7193 DERR("rc_conn_est: conn aborted, state = %d\n", old_state); 7194 daplka_ep_set_state(ep_rp, old_state, new_state); 7195 return (IBT_CM_ACCEPT); 7196 } 7197 7198 /* 7199 * it's ok for the timer to fire at this point. the 7200 * taskq thread that processes the timer will just wait 7201 * until we are done with our state transition. 7202 */ 7203 if (daplka_cancel_timer(ep_rp) != 0) { 7204 /* 7205 * daplka_cancel_timer returns -1 if the timer is 7206 * being processed and 0 for all other cases. 7207 * we need to reset ep_state to allow timer processing 7208 * to continue. 7209 */ 7210 DERR("rc_conn_est: timer is being processed\n"); 7211 daplka_ep_set_state(ep_rp, old_state, new_state); 7212 return (IBT_CM_ACCEPT); 7213 } 7214 7215 /* 7216 * create a DAPL_IB_CME_CONNECTED event 7217 */ 7218 conn_ev = kmem_zalloc(sizeof (daplka_evd_event_t), KM_NOSLEEP); 7219 if (conn_ev == NULL) { 7220 DERR("rc_conn_est: could not alloc ev\n"); 7221 daplka_ep_set_state(ep_rp, old_state, new_state); 7222 return (IBT_CM_ACCEPT); 7223 } 7224 7225 conn_ev->ee_cmev.ec_cm_ev_type = DAPL_IB_CME_CONNECTED; 7226 conn_ev->ee_cmev.ec_cm_cookie = ep_rp->ep_cookie; 7227 conn_ev->ee_cmev.ec_cm_is_passive = B_FALSE; 7228 conn_ev->ee_cmev.ec_cm_psep_cookie = 0; 7229 7230 /* 7231 * The private data passed back in the connection established 7232 * event is what was recvd in the daplka_cm_rc_rep_rcv handler and 7233 * saved in ep resource structure. 7234 */ 7235 if (ep_rp->ep_priv_len > 0) { 7236 conn_ev->ee_cmev.ec_cm_ev_priv_data = 7237 kmem_zalloc(ep_rp->ep_priv_len, KM_NOSLEEP); 7238 7239 if (conn_ev->ee_cmev.ec_cm_ev_priv_data == NULL) { 7240 DERR("rc_conn_est: could not alloc pr_data\n"); 7241 kmem_free(conn_ev, sizeof (daplka_evd_event_t)); 7242 daplka_ep_set_state(ep_rp, old_state, new_state); 7243 return (IBT_CM_ACCEPT); 7244 } 7245 bcopy(ep_rp->ep_priv_data, conn_ev->ee_cmev.ec_cm_ev_priv_data, 7246 ep_rp->ep_priv_len); 7247 } 7248 conn_ev->ee_cmev.ec_cm_ev_priv_data_len = ep_rp->ep_priv_len; 7249 7250 D2("rc_conn_est: enqueue event(%p) evdp(%p) pr_data(0x%p), " 7251 "pr_len(%d)\n", conn_ev, ep_rp->ep_conn_evd, 7252 conn_ev->ee_cmev.ec_cm_ev_priv_data, 7253 (int)conn_ev->ee_cmev.ec_cm_ev_priv_data_len); 7254 7255 /* 7256 * transition ep_state to CONNECTED 7257 */ 7258 new_state = DAPLKA_EP_STATE_CONNECTED; 7259 daplka_ep_set_state(ep_rp, old_state, new_state); 7260 7261 /* 7262 * enqueue event onto the conn_evd owned by ep_rp 7263 */ 7264 daplka_evd_wakeup(ep_rp->ep_conn_evd, 7265 &ep_rp->ep_conn_evd->evd_conn_events, conn_ev); 7266 7267 return (IBT_CM_ACCEPT); 7268 } 7269 7270 /* 7271 * processes the FAILURE event 7272 */ 7273 /* ARGSUSED */ 7274 static ibt_cm_status_t 7275 daplka_cm_rc_event_failure(daplka_ep_resource_t *ep_rp, ibt_cm_event_t *event, 7276 ibt_cm_return_args_t *ret_args, void *priv_data, ibt_priv_data_len_t len) 7277 { 7278 daplka_evd_event_t *disc_ev; 7279 ibt_priv_data_len_t pr_len = event->cm_priv_data_len; 7280 void *pr_data = event->cm_priv_data; 7281 uint32_t old_state, new_state; 7282 ibt_rc_chan_query_attr_t chan_attrs; 7283 ibt_status_t status; 7284 7285 ASSERT(ep_rp != NULL); 7286 old_state = new_state = daplka_ep_get_state(ep_rp); 7287 if (old_state != DAPLKA_EP_STATE_CONNECTING && 7288 old_state != DAPLKA_EP_STATE_DISCONNECTING) { 7289 /* 7290 * we can get here if the connection is being aborted 7291 */ 7292 DERR("rc_event_failure: conn aborted, state = %d, " 7293 "cf_code = %d, cf_msg = %d, cf_reason = %d\n", old_state, 7294 (int)event->cm_event.failed.cf_code, 7295 (int)event->cm_event.failed.cf_msg, 7296 (int)event->cm_event.failed.cf_reason); 7297 7298 daplka_ep_set_state(ep_rp, old_state, new_state); 7299 return (IBT_CM_ACCEPT); 7300 } 7301 7302 /* 7303 * it's ok for the timer to fire at this point. the 7304 * taskq thread that processes the timer will just wait 7305 * until we are done with our state transition. 7306 */ 7307 if (daplka_cancel_timer(ep_rp) != 0) { 7308 /* 7309 * daplka_cancel_timer returns -1 if the timer is 7310 * being processed and 0 for all other cases. 7311 * we need to reset ep_state to allow timer processing 7312 * to continue. 7313 */ 7314 DERR("rc_event_failure: timer is being processed\n"); 7315 daplka_ep_set_state(ep_rp, old_state, new_state); 7316 return (IBT_CM_ACCEPT); 7317 } 7318 7319 bzero(&chan_attrs, sizeof (ibt_rc_chan_query_attr_t)); 7320 status = ibt_query_rc_channel(ep_rp->ep_chan_hdl, &chan_attrs); 7321 7322 if ((status == IBT_SUCCESS) && 7323 (chan_attrs.rc_state != IBT_STATE_ERROR)) { 7324 DERR("rc_event_failure: conn abort qpn %d state %d\n", 7325 chan_attrs.rc_qpn, chan_attrs.rc_state); 7326 7327 /* explicit transition the QP to ERROR state */ 7328 status = ibt_flush_channel(ep_rp->ep_chan_hdl); 7329 } 7330 7331 /* 7332 * create an event 7333 */ 7334 disc_ev = kmem_zalloc(sizeof (daplka_evd_event_t), KM_NOSLEEP); 7335 if (disc_ev == NULL) { 7336 DERR("rc_event_failure: cannot alloc disc_ev\n"); 7337 daplka_ep_set_state(ep_rp, old_state, new_state); 7338 return (IBT_CM_ACCEPT); 7339 } 7340 7341 /* 7342 * copy private data into event 7343 */ 7344 if (pr_len > 0) { 7345 disc_ev->ee_cmev.ec_cm_ev_priv_data = 7346 kmem_zalloc(pr_len, KM_NOSLEEP); 7347 7348 if (disc_ev->ee_cmev.ec_cm_ev_priv_data == NULL) { 7349 DERR("rc_event_failure: cannot alloc pr data\n"); 7350 kmem_free(disc_ev, sizeof (daplka_evd_event_t)); 7351 daplka_ep_set_state(ep_rp, old_state, new_state); 7352 return (IBT_CM_ACCEPT); 7353 } 7354 bcopy(pr_data, disc_ev->ee_cmev.ec_cm_ev_priv_data, pr_len); 7355 } 7356 disc_ev->ee_cmev.ec_cm_ev_priv_data_len = pr_len; 7357 7358 /* 7359 * fill in the appropriate event type 7360 */ 7361 if (event->cm_event.failed.cf_code == IBT_CM_FAILURE_REJ_RCV) { 7362 switch (event->cm_event.failed.cf_reason) { 7363 case IBT_CM_CONSUMER: 7364 disc_ev->ee_cmev.ec_cm_ev_type = 7365 DAPL_IB_CME_DESTINATION_REJECT_PRIVATE_DATA; 7366 break; 7367 case IBT_CM_NO_CHAN: 7368 case IBT_CM_NO_RESC: 7369 disc_ev->ee_cmev.ec_cm_ev_type = 7370 DAPL_IB_CME_DESTINATION_REJECT; 7371 break; 7372 default: 7373 disc_ev->ee_cmev.ec_cm_ev_type = 7374 DAPL_IB_CME_DESTINATION_REJECT; 7375 break; 7376 } 7377 } else if (event->cm_event.failed.cf_code == IBT_CM_FAILURE_TIMEOUT) { 7378 disc_ev->ee_cmev.ec_cm_ev_type = DAPL_IB_CME_TIMED_OUT; 7379 } else { 7380 /* others we'll mark as local failure */ 7381 disc_ev->ee_cmev.ec_cm_ev_type = DAPL_IB_CME_LOCAL_FAILURE; 7382 } 7383 disc_ev->ee_cmev.ec_cm_cookie = ep_rp->ep_cookie; 7384 disc_ev->ee_cmev.ec_cm_is_passive = B_FALSE; 7385 disc_ev->ee_cmev.ec_cm_psep_cookie = 0; 7386 7387 D2("rc_event_failure: enqueue event(%p) evdp(%p) cf_code(%d) " 7388 "cf_msg(%d) cf_reason(%d)\n", disc_ev, ep_rp->ep_conn_evd, 7389 (int)event->cm_event.failed.cf_code, 7390 (int)event->cm_event.failed.cf_msg, 7391 (int)event->cm_event.failed.cf_reason); 7392 7393 /* 7394 * transition ep_state to DISCONNECTED 7395 */ 7396 new_state = DAPLKA_EP_STATE_DISCONNECTED; 7397 daplka_ep_set_state(ep_rp, old_state, new_state); 7398 7399 /* 7400 * enqueue event onto the conn_evd owned by ep_rp 7401 */ 7402 daplka_evd_wakeup(ep_rp->ep_conn_evd, 7403 &ep_rp->ep_conn_evd->evd_conn_events, disc_ev); 7404 7405 return (IBT_CM_ACCEPT); 7406 } 7407 7408 /* 7409 * This is the active side CM handler. It gets registered when 7410 * ibt_open_rc_channel is called. 7411 */ 7412 static ibt_cm_status_t 7413 daplka_cm_rc_handler(void *cm_private, ibt_cm_event_t *event, 7414 ibt_cm_return_args_t *ret_args, void *priv_data, ibt_priv_data_len_t len) 7415 { 7416 daplka_ep_resource_t *ep_rp = (daplka_ep_resource_t *)cm_private; 7417 7418 if (ep_rp == NULL) { 7419 DERR("rc_handler: ep_rp == NULL\n"); 7420 return (IBT_CM_NO_CHANNEL); 7421 } 7422 /* 7423 * default is not to return priv data 7424 */ 7425 if (ret_args != NULL) { 7426 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*ret_args)) 7427 ret_args->cm_ret_len = 0; 7428 } 7429 7430 switch (event->cm_type) { 7431 case IBT_CM_EVENT_REQ_RCV: 7432 /* active side should not receive this event */ 7433 D2("rc_handler: IBT_CM_EVENT_REQ_RCV\n"); 7434 break; 7435 7436 case IBT_CM_EVENT_REP_RCV: 7437 /* connection accepted by passive side */ 7438 D2("rc_handler: IBT_CM_EVENT_REP_RCV\n"); 7439 return (daplka_cm_rc_rep_rcv(ep_rp, event, ret_args, 7440 priv_data, len)); 7441 7442 case IBT_CM_EVENT_CONN_CLOSED: 7443 D2("rc_handler: IBT_CM_EVENT_CONN_CLOSED %d\n", 7444 event->cm_event.closed); 7445 return (daplka_cm_rc_conn_closed(ep_rp, event, ret_args, 7446 priv_data, len)); 7447 7448 case IBT_CM_EVENT_MRA_RCV: 7449 /* passive side does default processing MRA event */ 7450 D2("rc_handler: IBT_CM_EVENT_MRA_RCV\n"); 7451 return (IBT_CM_DEFAULT); 7452 7453 case IBT_CM_EVENT_CONN_EST: 7454 D2("rc_handler: IBT_CM_EVENT_CONN_EST\n"); 7455 return (daplka_cm_rc_conn_est(ep_rp, event, ret_args, 7456 priv_data, len)); 7457 7458 case IBT_CM_EVENT_FAILURE: 7459 D2("rc_handler: IBT_CM_EVENT_FAILURE\n"); 7460 return (daplka_cm_rc_event_failure(ep_rp, event, ret_args, 7461 priv_data, len)); 7462 7463 default: 7464 D2("rc_handler: invalid event %d\n", event->cm_type); 7465 break; 7466 } 7467 return (IBT_CM_DEFAULT); 7468 } 7469 7470 /* 7471 * creates an IA resource and inserts it into the global resource table. 7472 */ 7473 /* ARGSUSED */ 7474 static int 7475 daplka_ia_create(minor_t rnum, intptr_t arg, int mode, 7476 cred_t *cred, int *rvalp) 7477 { 7478 daplka_ia_resource_t *ia_rp, *tmp_rp; 7479 boolean_t inserted = B_FALSE; 7480 dapl_ia_create_t args; 7481 ibt_hca_hdl_t hca_hdl; 7482 ibt_status_t status; 7483 ib_gid_t sgid; 7484 int retval; 7485 ibt_hca_portinfo_t *pinfop; 7486 uint_t pinfon; 7487 uint_t size; 7488 ibt_ar_t ar_s; 7489 daplka_hca_t *hca; 7490 7491 retval = ddi_copyin((void *)arg, &args, sizeof (dapl_ia_create_t), 7492 mode); 7493 if (retval != 0) { 7494 DERR("ia_create: copyin error %d\n", retval); 7495 return (EFAULT); 7496 } 7497 if (args.ia_version != DAPL_IF_VERSION) { 7498 DERR("ia_create: invalid version %d, expected version %d\n", 7499 args.ia_version, DAPL_IF_VERSION); 7500 return (EINVAL); 7501 } 7502 7503 /* 7504 * find the hca with the matching guid 7505 */ 7506 mutex_enter(&daplka_dev->daplka_mutex); 7507 for (hca = daplka_dev->daplka_hca_list_head; hca != NULL; 7508 hca = hca->hca_next) { 7509 if (hca->hca_guid == args.ia_guid) { 7510 DAPLKA_HOLD_HCA_WITHOUT_LOCK(hca); 7511 break; 7512 } 7513 } 7514 mutex_exit(&daplka_dev->daplka_mutex); 7515 7516 if (hca == NULL) { 7517 DERR("ia_create: guid 0x%016llx not found\n", 7518 (longlong_t)args.ia_guid); 7519 return (EINVAL); 7520 } 7521 7522 /* 7523 * check whether port number is valid and whether it is up 7524 */ 7525 if (args.ia_port > hca->hca_nports) { 7526 DERR("ia_create: invalid hca_port %d\n", args.ia_port); 7527 DAPLKA_RELE_HCA(daplka_dev, hca); 7528 return (EINVAL); 7529 } 7530 hca_hdl = hca->hca_hdl; 7531 if (hca_hdl == NULL) { 7532 DERR("ia_create: hca_hdl == NULL\n"); 7533 DAPLKA_RELE_HCA(daplka_dev, hca); 7534 return (EINVAL); 7535 } 7536 status = ibt_query_hca_ports(hca_hdl, (uint8_t)args.ia_port, 7537 &pinfop, &pinfon, &size); 7538 if (status != IBT_SUCCESS) { 7539 DERR("ia_create: ibt_query_hca_ports returned %d\n", status); 7540 *rvalp = (int)status; 7541 DAPLKA_RELE_HCA(daplka_dev, hca); 7542 return (0); 7543 } 7544 sgid = pinfop->p_sgid_tbl[0]; 7545 ibt_free_portinfo(pinfop, size); 7546 7547 ia_rp = kmem_zalloc(sizeof (daplka_ia_resource_t), daplka_km_flags); 7548 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*ia_rp)) 7549 DAPLKA_RS_INIT(ia_rp, DAPL_TYPE_IA, rnum, daplka_ia_destroy); 7550 7551 mutex_init(&ia_rp->ia_lock, NULL, MUTEX_DRIVER, NULL); 7552 cv_init(&ia_rp->ia_cv, NULL, CV_DRIVER, NULL); 7553 ia_rp->ia_hca_hdl = hca_hdl; 7554 ia_rp->ia_hca_sgid = sgid; 7555 ia_rp->ia_hca = hca; 7556 ia_rp->ia_port_num = args.ia_port; 7557 ia_rp->ia_port_pkey = args.ia_pkey; 7558 ia_rp->ia_pid = ddi_get_pid(); 7559 ia_rp->ia_async_evd_hkeys = NULL; 7560 ia_rp->ia_ar_registered = B_FALSE; 7561 bcopy(args.ia_sadata, ia_rp->ia_sadata, DAPL_ATS_NBYTES); 7562 7563 /* register Address Record */ 7564 ar_s.ar_gid = ia_rp->ia_hca_sgid; 7565 ar_s.ar_pkey = ia_rp->ia_port_pkey; 7566 bcopy(ia_rp->ia_sadata, ar_s.ar_data, DAPL_ATS_NBYTES); 7567 #define UC(b) ar_s.ar_data[(b)] 7568 D3("daplka_ia_create: SA[8] %d.%d.%d.%d\n", 7569 UC(8), UC(9), UC(10), UC(11)); 7570 D3("daplka_ia_create: SA[12] %d.%d.%d.%d\n", 7571 UC(12), UC(13), UC(14), UC(15)); 7572 retval = ibt_register_ar(daplka_dev->daplka_clnt_hdl, &ar_s); 7573 if (retval != IBT_SUCCESS) { 7574 DERR("ia_create: failed to register Address Record.\n"); 7575 retval = EINVAL; 7576 goto cleanup; 7577 } 7578 ia_rp->ia_ar_registered = B_TRUE; 7579 7580 /* 7581 * create hash tables for all object types 7582 */ 7583 retval = daplka_hash_create(&ia_rp->ia_ep_htbl, DAPLKA_EP_HTBL_SZ, 7584 daplka_hash_ep_free, daplka_hash_generic_lookup); 7585 if (retval != 0) { 7586 DERR("ia_create: cannot create ep hash table\n"); 7587 goto cleanup; 7588 } 7589 retval = daplka_hash_create(&ia_rp->ia_mr_htbl, DAPLKA_MR_HTBL_SZ, 7590 daplka_hash_mr_free, daplka_hash_generic_lookup); 7591 if (retval != 0) { 7592 DERR("ia_create: cannot create mr hash table\n"); 7593 goto cleanup; 7594 } 7595 retval = daplka_hash_create(&ia_rp->ia_mw_htbl, DAPLKA_MW_HTBL_SZ, 7596 daplka_hash_mw_free, daplka_hash_generic_lookup); 7597 if (retval != 0) { 7598 DERR("ia_create: cannot create mw hash table\n"); 7599 goto cleanup; 7600 } 7601 retval = daplka_hash_create(&ia_rp->ia_pd_htbl, DAPLKA_PD_HTBL_SZ, 7602 daplka_hash_pd_free, daplka_hash_generic_lookup); 7603 if (retval != 0) { 7604 DERR("ia_create: cannot create pd hash table\n"); 7605 goto cleanup; 7606 } 7607 retval = daplka_hash_create(&ia_rp->ia_evd_htbl, DAPLKA_EVD_HTBL_SZ, 7608 daplka_hash_evd_free, daplka_hash_generic_lookup); 7609 if (retval != 0) { 7610 DERR("ia_create: cannot create evd hash table\n"); 7611 goto cleanup; 7612 } 7613 retval = daplka_hash_create(&ia_rp->ia_cno_htbl, DAPLKA_CNO_HTBL_SZ, 7614 daplka_hash_cno_free, daplka_hash_generic_lookup); 7615 if (retval != 0) { 7616 DERR("ia_create: cannot create cno hash table\n"); 7617 goto cleanup; 7618 } 7619 retval = daplka_hash_create(&ia_rp->ia_sp_htbl, DAPLKA_SP_HTBL_SZ, 7620 daplka_hash_sp_free, daplka_hash_generic_lookup); 7621 if (retval != 0) { 7622 DERR("ia_create: cannot create sp hash table\n"); 7623 goto cleanup; 7624 } 7625 retval = daplka_hash_create(&ia_rp->ia_srq_htbl, DAPLKA_SRQ_HTBL_SZ, 7626 daplka_hash_srq_free, daplka_hash_generic_lookup); 7627 if (retval != 0) { 7628 DERR("ia_create: cannot create srq hash table\n"); 7629 goto cleanup; 7630 } 7631 /* 7632 * insert ia_rp into the global resource table 7633 */ 7634 retval = daplka_resource_insert(rnum, (daplka_resource_t *)ia_rp); 7635 if (retval != 0) { 7636 DERR("ia_create: cannot insert resource\n"); 7637 goto cleanup; 7638 } 7639 inserted = B_TRUE; 7640 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(*ia_rp)) 7641 7642 args.ia_resnum = rnum; 7643 retval = copyout(&args, (void *)arg, sizeof (dapl_ia_create_t)); 7644 if (retval != 0) { 7645 DERR("ia_create: copyout error %d\n", retval); 7646 retval = EFAULT; 7647 goto cleanup; 7648 } 7649 return (0); 7650 7651 cleanup:; 7652 if (inserted) { 7653 tmp_rp = (daplka_ia_resource_t *)daplka_resource_remove(rnum); 7654 if (tmp_rp != ia_rp) { 7655 /* 7656 * we can return here because another thread must 7657 * have freed up the resource 7658 */ 7659 DERR("ia_create: cannot remove resource\n"); 7660 return (retval); 7661 } 7662 } 7663 DAPLKA_RS_UNREF(ia_rp); 7664 return (retval); 7665 } 7666 7667 /* 7668 * destroys an IA resource 7669 */ 7670 static int 7671 daplka_ia_destroy(daplka_resource_t *gen_rp) 7672 { 7673 daplka_ia_resource_t *ia_rp = (daplka_ia_resource_t *)gen_rp; 7674 daplka_async_evd_hkey_t *hkp; 7675 int cnt; 7676 ibt_ar_t ar_s; 7677 7678 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*ia_rp)) 7679 D3("ia_destroy: entering, ia_rp 0x%p\n", ia_rp); 7680 7681 /* deregister Address Record */ 7682 if (ia_rp->ia_ar_registered) { 7683 ar_s.ar_gid = ia_rp->ia_hca_sgid; 7684 ar_s.ar_pkey = ia_rp->ia_port_pkey; 7685 bcopy(ia_rp->ia_sadata, ar_s.ar_data, DAPL_ATS_NBYTES); 7686 (void) ibt_deregister_ar(daplka_dev->daplka_clnt_hdl, &ar_s); 7687 ia_rp->ia_ar_registered = B_FALSE; 7688 } 7689 7690 /* 7691 * destroy hash tables. make sure resources are 7692 * destroyed in the correct order. 7693 */ 7694 daplka_hash_destroy(&ia_rp->ia_mw_htbl); 7695 daplka_hash_destroy(&ia_rp->ia_mr_htbl); 7696 daplka_hash_destroy(&ia_rp->ia_ep_htbl); 7697 daplka_hash_destroy(&ia_rp->ia_srq_htbl); 7698 daplka_hash_destroy(&ia_rp->ia_evd_htbl); 7699 daplka_hash_destroy(&ia_rp->ia_cno_htbl); 7700 daplka_hash_destroy(&ia_rp->ia_pd_htbl); 7701 daplka_hash_destroy(&ia_rp->ia_sp_htbl); 7702 7703 /* 7704 * free the async evd list 7705 */ 7706 cnt = 0; 7707 hkp = ia_rp->ia_async_evd_hkeys; 7708 while (hkp != NULL) { 7709 daplka_async_evd_hkey_t *free_hkp; 7710 7711 cnt++; 7712 free_hkp = hkp; 7713 hkp = hkp->aeh_next; 7714 kmem_free(free_hkp, sizeof (*free_hkp)); 7715 } 7716 if (cnt > 0) { 7717 D3("ia_destroy: freed %d hkeys\n", cnt); 7718 } 7719 mutex_destroy(&ia_rp->ia_lock); 7720 cv_destroy(&ia_rp->ia_cv); 7721 ia_rp->ia_hca_hdl = NULL; 7722 7723 DAPLKA_RS_FINI(ia_rp); 7724 7725 if (ia_rp->ia_hca) 7726 DAPLKA_RELE_HCA(daplka_dev, ia_rp->ia_hca); 7727 7728 kmem_free(ia_rp, sizeof (daplka_ia_resource_t)); 7729 D3("ia_destroy: exiting, ia_rp 0x%p\n", ia_rp); 7730 return (0); 7731 } 7732 7733 static void 7734 daplka_async_event_create(ibt_async_code_t code, ibt_async_event_t *event, 7735 uint64_t cookie, daplka_ia_resource_t *ia_rp) 7736 { 7737 daplka_evd_event_t *evp; 7738 daplka_evd_resource_t *async_evd; 7739 daplka_async_evd_hkey_t *curr; 7740 7741 mutex_enter(&ia_rp->ia_lock); 7742 curr = ia_rp->ia_async_evd_hkeys; 7743 while (curr != NULL) { 7744 /* 7745 * Note: this allocation does not zero out the buffer 7746 * since we init all the fields. 7747 */ 7748 evp = kmem_alloc(sizeof (daplka_evd_event_t), KM_NOSLEEP); 7749 if (evp == NULL) { 7750 DERR("async_event_enqueue: event alloc failed" 7751 "!found\n", ia_rp, curr->aeh_evd_hkey); 7752 curr = curr->aeh_next; 7753 continue; 7754 } 7755 evp->ee_next = NULL; 7756 evp->ee_aev.ibae_type = code; 7757 evp->ee_aev.ibae_hca_guid = event->ev_hca_guid; 7758 evp->ee_aev.ibae_cookie = cookie; 7759 evp->ee_aev.ibae_port = event->ev_port; 7760 7761 /* 7762 * Lookup the async evd corresponding to this ia and enqueue 7763 * evp and wakeup any waiter. 7764 */ 7765 async_evd = (daplka_evd_resource_t *) 7766 daplka_hash_lookup(&ia_rp->ia_evd_htbl, curr->aeh_evd_hkey); 7767 if (async_evd == NULL) { /* async evd is being freed */ 7768 DERR("async_event_enqueue: ia_rp(%p) asycn_evd %llx " 7769 "!found\n", ia_rp, (longlong_t)curr->aeh_evd_hkey); 7770 kmem_free(evp, sizeof (daplka_evd_event_t)); 7771 curr = curr->aeh_next; 7772 continue; 7773 } 7774 daplka_evd_wakeup(async_evd, &async_evd->evd_async_events, evp); 7775 7776 /* decrement refcnt on async_evd */ 7777 DAPLKA_RS_UNREF(async_evd); 7778 curr = curr->aeh_next; 7779 } 7780 mutex_exit(&ia_rp->ia_lock); 7781 } 7782 /* 7783 * This routine is called in kernel context 7784 */ 7785 7786 /* ARGSUSED */ 7787 static void 7788 daplka_rc_async_handler(void *clnt_private, ibt_hca_hdl_t hca_hdl, 7789 ibt_async_code_t code, ibt_async_event_t *event) 7790 { 7791 daplka_ep_resource_t *epp; 7792 daplka_ia_resource_t *ia_rp; 7793 minor_t ia_rnum; 7794 7795 if (event->ev_chan_hdl == NULL) { 7796 DERR("daplka_rc_async_handler: ev_chan_hdl is NULL\n"); 7797 return; 7798 } 7799 7800 mutex_enter(&daplka_dev->daplka_mutex); 7801 epp = ibt_get_chan_private(event->ev_chan_hdl); 7802 if (epp == NULL) { 7803 mutex_exit(&daplka_dev->daplka_mutex); 7804 DERR("daplka_rc_async_handler: chan_private is NULL\n"); 7805 return; 7806 } 7807 7808 /* grab a reference to this ep */ 7809 DAPLKA_RS_REF(epp); 7810 mutex_exit(&daplka_dev->daplka_mutex); 7811 7812 /* 7813 * The endpoint resource has the resource number corresponding to 7814 * the IA resource. Use that to lookup the ia resource entry 7815 */ 7816 ia_rnum = DAPLKA_RS_RNUM(epp); 7817 ia_rp = (daplka_ia_resource_t *)daplka_resource_lookup(ia_rnum); 7818 if ((ia_rp == NULL) || DAPLKA_RS_RESERVED(ia_rp)) { 7819 D2("daplka_rc_async_handler: resource (%d) not found\n", 7820 ia_rnum); 7821 DAPLKA_RS_UNREF(epp); 7822 return; 7823 } 7824 7825 /* 7826 * Create an async event and chain it to the async evd 7827 */ 7828 daplka_async_event_create(code, event, epp->ep_cookie, ia_rp); 7829 7830 DAPLKA_RS_UNREF(ia_rp); 7831 DAPLKA_RS_UNREF(epp); 7832 } 7833 7834 /* 7835 * This routine is called in kernel context 7836 */ 7837 7838 /* ARGSUSED */ 7839 static void 7840 daplka_cq_async_handler(void *clnt_private, ibt_hca_hdl_t hca_hdl, 7841 ibt_async_code_t code, ibt_async_event_t *event) 7842 { 7843 daplka_evd_resource_t *evdp; 7844 daplka_ia_resource_t *ia_rp; 7845 minor_t ia_rnum; 7846 7847 if (event->ev_cq_hdl == NULL) 7848 return; 7849 7850 mutex_enter(&daplka_dev->daplka_mutex); 7851 evdp = ibt_get_cq_private(event->ev_cq_hdl); 7852 if (evdp == NULL) { 7853 mutex_exit(&daplka_dev->daplka_mutex); 7854 DERR("daplka_cq_async_handler: get cq private(%p) failed\n", 7855 event->ev_cq_hdl); 7856 return; 7857 } 7858 /* grab a reference to this evd resource */ 7859 DAPLKA_RS_REF(evdp); 7860 mutex_exit(&daplka_dev->daplka_mutex); 7861 7862 /* 7863 * The endpoint resource has the resource number corresponding to 7864 * the IA resource. Use that to lookup the ia resource entry 7865 */ 7866 ia_rnum = DAPLKA_RS_RNUM(evdp); 7867 ia_rp = (daplka_ia_resource_t *)daplka_resource_lookup(ia_rnum); 7868 if ((ia_rp == NULL) || DAPLKA_RS_RESERVED(ia_rp)) { 7869 DERR("daplka_cq_async_handler: resource (%d) not found\n", 7870 ia_rnum); 7871 DAPLKA_RS_UNREF(evdp); 7872 return; 7873 } 7874 7875 /* 7876 * Create an async event and chain it to the async evd 7877 */ 7878 daplka_async_event_create(code, event, evdp->evd_cookie, ia_rp); 7879 7880 /* release all the refcount that were acquired */ 7881 DAPLKA_RS_UNREF(ia_rp); 7882 DAPLKA_RS_UNREF(evdp); 7883 } 7884 7885 /* 7886 * This routine is called in kernel context, handles unaffiliated async errors 7887 */ 7888 7889 /* ARGSUSED */ 7890 static void 7891 daplka_un_async_handler(void *clnt_private, ibt_hca_hdl_t hca_hdl, 7892 ibt_async_code_t code, ibt_async_event_t *event) 7893 { 7894 int i, j; 7895 daplka_resource_blk_t *blk; 7896 daplka_resource_t *rp; 7897 daplka_ia_resource_t *ia_rp; 7898 7899 /* 7900 * Walk the resource table looking for an ia that matches the 7901 * hca_hdl. 7902 */ 7903 rw_enter(&daplka_resource.daplka_rct_lock, RW_READER); 7904 for (i = 0; i < daplka_resource.daplka_rc_len; i++) { 7905 blk = daplka_resource.daplka_rc_root[i]; 7906 if (blk == NULL) 7907 continue; 7908 for (j = 0; j < DAPLKA_RC_BLKSZ; j++) { 7909 rp = blk->daplka_rcblk_blks[j]; 7910 if ((rp == NULL) || 7911 ((intptr_t)rp == DAPLKA_RC_RESERVED) || 7912 (rp->rs_type != DAPL_TYPE_IA)) { 7913 continue; 7914 } 7915 /* 7916 * rp is an IA resource check if it belongs 7917 * to the hca/port for which we got the event 7918 */ 7919 ia_rp = (daplka_ia_resource_t *)rp; 7920 DAPLKA_RS_REF(ia_rp); 7921 if ((hca_hdl == ia_rp->ia_hca_hdl) && 7922 (event->ev_port == ia_rp->ia_port_num)) { 7923 /* 7924 * walk the ep hash table. Acquire a 7925 * reader lock. NULL dgid indicates 7926 * local port up event. 7927 */ 7928 daplka_hash_walk(&ia_rp->ia_ep_htbl, 7929 daplka_ep_failback, NULL, RW_READER); 7930 } 7931 DAPLKA_RS_UNREF(ia_rp); 7932 } 7933 } 7934 rw_exit(&daplka_resource.daplka_rct_lock); 7935 } 7936 7937 static int 7938 daplka_handle_hca_detach_event(ibt_async_event_t *event) 7939 { 7940 daplka_hca_t *hca; 7941 7942 /* 7943 * find the hca with the matching guid 7944 */ 7945 mutex_enter(&daplka_dev->daplka_mutex); 7946 for (hca = daplka_dev->daplka_hca_list_head; hca != NULL; 7947 hca = hca->hca_next) { 7948 if (hca->hca_guid == event->ev_hca_guid) { 7949 if (DAPLKA_HCA_BUSY(hca)) { 7950 mutex_exit(&daplka_dev->daplka_mutex); 7951 return (IBT_HCA_RESOURCES_NOT_FREED); 7952 } 7953 daplka_dequeue_hca(daplka_dev, hca); 7954 break; 7955 } 7956 } 7957 mutex_exit(&daplka_dev->daplka_mutex); 7958 7959 if (hca == NULL) 7960 return (IBT_FAILURE); 7961 7962 return (daplka_fini_hca(daplka_dev, hca)); 7963 } 7964 7965 /* 7966 * This routine is called in kernel context 7967 */ 7968 static void 7969 daplka_async_handler(void *clnt_private, ibt_hca_hdl_t hca_hdl, 7970 ibt_async_code_t code, ibt_async_event_t *event) 7971 { 7972 switch (code) { 7973 case IBT_ERROR_CATASTROPHIC_CHAN: 7974 case IBT_ERROR_INVALID_REQUEST_CHAN: 7975 case IBT_ERROR_ACCESS_VIOLATION_CHAN: 7976 case IBT_ERROR_PATH_MIGRATE_REQ: 7977 D2("daplka_async_handler(): Channel affiliated=0x%x\n", code); 7978 /* These events are affiliated with a the RC channel */ 7979 daplka_rc_async_handler(clnt_private, hca_hdl, code, event); 7980 break; 7981 case IBT_ERROR_CQ: 7982 /* This event is affiliated with a the CQ */ 7983 D2("daplka_async_handler(): IBT_ERROR_CQ\n"); 7984 daplka_cq_async_handler(clnt_private, hca_hdl, code, event); 7985 break; 7986 case IBT_ERROR_PORT_DOWN: 7987 D2("daplka_async_handler(): IBT_PORT_DOWN\n"); 7988 break; 7989 case IBT_EVENT_PORT_UP: 7990 D2("daplka_async_handler(): IBT_PORT_UP\n"); 7991 if (daplka_apm) { 7992 daplka_un_async_handler(clnt_private, hca_hdl, code, 7993 event); 7994 } 7995 break; 7996 case IBT_HCA_ATTACH_EVENT: 7997 /* 7998 * NOTE: In some error recovery paths, it is possible to 7999 * receive IBT_HCA_ATTACH_EVENTs on already known HCAs. 8000 */ 8001 D2("daplka_async_handler(): IBT_HCA_ATTACH\n"); 8002 (void) daplka_init_hca(daplka_dev, event->ev_hca_guid); 8003 break; 8004 case IBT_HCA_DETACH_EVENT: 8005 D2("daplka_async_handler(): IBT_HCA_DETACH\n"); 8006 /* Free all hca resources and close the HCA. */ 8007 (void) daplka_handle_hca_detach_event(event); 8008 break; 8009 case IBT_EVENT_PATH_MIGRATED: 8010 /* This event is affiliated with APM */ 8011 D2("daplka_async_handler(): IBT_PATH_MIGRATED.\n"); 8012 break; 8013 default: 8014 D2("daplka_async_handler(): unhandled code = 0x%x\n", code); 8015 break; 8016 } 8017 } 8018 8019 /* 8020 * This routine is called in kernel context related to Subnet events 8021 */ 8022 /*ARGSUSED*/ 8023 static void 8024 daplka_sm_notice_handler(void *arg, ib_gid_t gid, ibt_subnet_event_code_t code, 8025 ibt_subnet_event_t *event) 8026 { 8027 ib_gid_t *sgid = &gid; 8028 ib_gid_t *dgid; 8029 8030 dgid = &event->sm_notice_gid; 8031 switch (code) { 8032 case IBT_SM_EVENT_GID_AVAIL: 8033 /* This event is affiliated with remote port up */ 8034 D2("daplka_sm_notice_handler(): IBT_SM_EVENT_GID_AVAIL\n"); 8035 if (daplka_apm) 8036 daplka_sm_gid_avail(sgid, dgid); 8037 return; 8038 case IBT_SM_EVENT_GID_UNAVAIL: 8039 /* This event is affiliated with remote port down */ 8040 D2("daplka_sm_notice_handler(): IBT_SM_EVENT_GID_UNAVAIL\n"); 8041 return; 8042 default: 8043 D2("daplka_sm_notice_handler(): unhandled IBT_SM_EVENT_[%d]\n", 8044 code); 8045 return; 8046 } 8047 } 8048 8049 /* 8050 * This routine is called in kernel context, handles Subnet GID avail events 8051 * which correspond to remote port up. Setting up alternate path or path 8052 * migration (failback) has to be initiated from the active side of the 8053 * original connect. 8054 */ 8055 static void 8056 daplka_sm_gid_avail(ib_gid_t *sgid, ib_gid_t *dgid) 8057 { 8058 int i, j; 8059 daplka_resource_blk_t *blk; 8060 daplka_resource_t *rp; 8061 daplka_ia_resource_t *ia_rp; 8062 8063 D2("daplka_sm_gid_avail: sgid=%llx:%llx dgid=%llx:%llx\n", 8064 (longlong_t)sgid->gid_prefix, (longlong_t)sgid->gid_guid, 8065 (longlong_t)dgid->gid_prefix, (longlong_t)dgid->gid_guid); 8066 8067 /* 8068 * Walk the resource table looking for an ia that matches the sgid 8069 */ 8070 rw_enter(&daplka_resource.daplka_rct_lock, RW_READER); 8071 for (i = 0; i < daplka_resource.daplka_rc_len; i++) { 8072 blk = daplka_resource.daplka_rc_root[i]; 8073 if (blk == NULL) 8074 continue; 8075 for (j = 0; j < DAPLKA_RC_BLKSZ; j++) { 8076 rp = blk->daplka_rcblk_blks[j]; 8077 if ((rp == NULL) || 8078 ((intptr_t)rp == DAPLKA_RC_RESERVED) || 8079 (rp->rs_type != DAPL_TYPE_IA)) { 8080 continue; 8081 } 8082 /* 8083 * rp is an IA resource check if its gid 8084 * matches with the calling sgid 8085 */ 8086 ia_rp = (daplka_ia_resource_t *)rp; 8087 DAPLKA_RS_REF(ia_rp); 8088 if ((sgid->gid_prefix == 8089 ia_rp->ia_hca_sgid.gid_prefix) && 8090 (sgid->gid_guid == ia_rp->ia_hca_sgid.gid_guid)) { 8091 /* 8092 * walk the ep hash table. Acquire a 8093 * reader lock. 8094 */ 8095 daplka_hash_walk(&ia_rp->ia_ep_htbl, 8096 daplka_ep_failback, 8097 (void *)dgid, RW_READER); 8098 } 8099 DAPLKA_RS_UNREF(ia_rp); 8100 } 8101 } 8102 rw_exit(&daplka_resource.daplka_rct_lock); 8103 } 8104 8105 /* 8106 * This routine is called in kernel context to get and set an alternate path 8107 */ 8108 static int 8109 daplka_ep_altpath(daplka_ep_resource_t *ep_rp, ib_gid_t *dgid) 8110 { 8111 ibt_alt_path_info_t path_info; 8112 ibt_alt_path_attr_t path_attr; 8113 ibt_ap_returns_t ap_rets; 8114 ibt_status_t status; 8115 8116 D2("daplka_ep_altpath : ibt_get_alt_path()\n"); 8117 bzero(&path_info, sizeof (ibt_alt_path_info_t)); 8118 bzero(&path_attr, sizeof (ibt_alt_path_attr_t)); 8119 if (dgid != NULL) { 8120 path_attr.apa_sgid = ep_rp->ep_sgid; 8121 path_attr.apa_dgid = *dgid; 8122 } 8123 status = ibt_get_alt_path(ep_rp->ep_chan_hdl, IBT_PATH_AVAIL, 8124 &path_attr, &path_info); 8125 if (status != IBT_SUCCESS) { 8126 DERR("daplka_ep_altpath : ibt_get_alt_path failed %d\n", 8127 status); 8128 return (1); 8129 } 8130 8131 D2("daplka_ep_altpath : ibt_set_alt_path()\n"); 8132 bzero(&ap_rets, sizeof (ibt_ap_returns_t)); 8133 status = ibt_set_alt_path(ep_rp->ep_chan_hdl, IBT_BLOCKING, 8134 &path_info, NULL, 0, &ap_rets); 8135 if ((status != IBT_SUCCESS) || 8136 (ap_rets.ap_status != IBT_CM_AP_LOADED)) { 8137 DERR("daplka_ep_altpath : ibt_set_alt_path failed " 8138 "status %d ap_status %d\n", status, ap_rets.ap_status); 8139 return (1); 8140 } 8141 return (0); 8142 } 8143 8144 /* 8145 * This routine is called in kernel context to failback to the original path 8146 */ 8147 static int 8148 daplka_ep_failback(void *objp, void *arg) 8149 { 8150 daplka_ep_resource_t *ep_rp = (daplka_ep_resource_t *)objp; 8151 ib_gid_t *dgid; 8152 ibt_status_t status; 8153 ibt_rc_chan_query_attr_t chan_attrs; 8154 int i; 8155 8156 ASSERT(DAPLKA_RS_TYPE(ep_rp) == DAPL_TYPE_EP); 8157 D2("daplka_ep_failback ep : sgid=%llx:%llx dgid=%llx:%llx\n", 8158 (longlong_t)ep_rp->ep_sgid.gid_prefix, 8159 (longlong_t)ep_rp->ep_sgid.gid_guid, 8160 (longlong_t)ep_rp->ep_dgid.gid_prefix, 8161 (longlong_t)ep_rp->ep_dgid.gid_guid); 8162 8163 /* 8164 * daplka_ep_failback is called from daplka_hash_walk 8165 * which holds the read lock on hash table to protect 8166 * the endpoint resource from removal 8167 */ 8168 mutex_enter(&ep_rp->ep_lock); 8169 /* check for unconnected endpoints */ 8170 /* first check for ep state */ 8171 if (ep_rp->ep_state != DAPLKA_EP_STATE_CONNECTED) { 8172 mutex_exit(&ep_rp->ep_lock); 8173 D2("daplka_ep_failback : endpoints not connected\n"); 8174 return (0); 8175 } 8176 8177 /* second check for gids */ 8178 if (((ep_rp->ep_sgid.gid_prefix == 0) && 8179 (ep_rp->ep_sgid.gid_guid == 0)) || 8180 ((ep_rp->ep_dgid.gid_prefix == 0) && 8181 (ep_rp->ep_dgid.gid_guid == 0))) { 8182 mutex_exit(&ep_rp->ep_lock); 8183 D2("daplka_ep_failback : skip unconnected endpoints\n"); 8184 return (0); 8185 } 8186 8187 /* 8188 * matching destination ep 8189 * when dgid is NULL, the async event is a local port up. 8190 * dgid becomes wild card, i.e. all endpoints match 8191 */ 8192 dgid = (ib_gid_t *)arg; 8193 if (dgid == NULL) { 8194 /* ignore loopback ep */ 8195 if ((ep_rp->ep_sgid.gid_prefix == ep_rp->ep_dgid.gid_prefix) && 8196 (ep_rp->ep_sgid.gid_guid == ep_rp->ep_dgid.gid_guid)) { 8197 mutex_exit(&ep_rp->ep_lock); 8198 D2("daplka_ep_failback : skip loopback endpoints\n"); 8199 return (0); 8200 } 8201 } else { 8202 /* matching remote ep */ 8203 if ((ep_rp->ep_dgid.gid_prefix != dgid->gid_prefix) || 8204 (ep_rp->ep_dgid.gid_guid != dgid->gid_guid)) { 8205 mutex_exit(&ep_rp->ep_lock); 8206 D2("daplka_ep_failback : unrelated endpoints\n"); 8207 return (0); 8208 } 8209 } 8210 8211 /* call get and set altpath with original dgid used in ep_connect */ 8212 if (daplka_ep_altpath(ep_rp, &ep_rp->ep_dgid)) { 8213 mutex_exit(&ep_rp->ep_lock); 8214 return (0); 8215 } 8216 8217 /* 8218 * wait for migration state to be ARMed 8219 * e.g. a post_send msg will transit mig_state from REARM to ARM 8220 */ 8221 for (i = 0; i < daplka_query_aft_setaltpath; i++) { 8222 bzero(&chan_attrs, sizeof (ibt_rc_chan_query_attr_t)); 8223 status = ibt_query_rc_channel(ep_rp->ep_chan_hdl, &chan_attrs); 8224 if (status != IBT_SUCCESS) { 8225 mutex_exit(&ep_rp->ep_lock); 8226 DERR("daplka_ep_altpath : ibt_query_rc_channel err\n"); 8227 return (0); 8228 } 8229 if (chan_attrs.rc_mig_state == IBT_STATE_ARMED) 8230 break; 8231 } 8232 8233 D2("daplka_ep_altpath : query[%d] mig_st=%d\n", 8234 i, chan_attrs.rc_mig_state); 8235 D2("daplka_ep_altpath : P sgid=%llx:%llx dgid=%llx:%llx\n", 8236 (longlong_t) 8237 chan_attrs.rc_prim_path.cep_adds_vect.av_sgid.gid_prefix, 8238 (longlong_t)chan_attrs.rc_prim_path.cep_adds_vect.av_sgid.gid_guid, 8239 (longlong_t) 8240 chan_attrs.rc_prim_path.cep_adds_vect.av_dgid.gid_prefix, 8241 (longlong_t)chan_attrs.rc_prim_path.cep_adds_vect.av_dgid.gid_guid); 8242 D2("daplka_ep_altpath : A sgid=%llx:%llx dgid=%llx:%llx\n", 8243 (longlong_t)chan_attrs.rc_alt_path.cep_adds_vect.av_sgid.gid_prefix, 8244 (longlong_t)chan_attrs.rc_alt_path.cep_adds_vect.av_sgid.gid_guid, 8245 (longlong_t)chan_attrs.rc_alt_path.cep_adds_vect.av_dgid.gid_prefix, 8246 (longlong_t)chan_attrs.rc_alt_path.cep_adds_vect.av_dgid.gid_guid); 8247 8248 /* skip failback on ARMed state not reached or env override */ 8249 if ((i >= daplka_query_aft_setaltpath) || (daplka_failback == 0)) { 8250 mutex_exit(&ep_rp->ep_lock); 8251 DERR("daplka_ep_altpath : ARMed state not reached\n"); 8252 return (0); 8253 } 8254 8255 D2("daplka_ep_failback : ibt_migrate_path() to original ep\n"); 8256 status = ibt_migrate_path(ep_rp->ep_chan_hdl); 8257 if (status != IBT_SUCCESS) { 8258 mutex_exit(&ep_rp->ep_lock); 8259 DERR("daplka_ep_failback : migration failed " 8260 "status %d\n", status); 8261 return (0); 8262 } 8263 8264 /* call get and altpath with NULL dgid to indicate unspecified dgid */ 8265 (void) daplka_ep_altpath(ep_rp, NULL); 8266 mutex_exit(&ep_rp->ep_lock); 8267 return (0); 8268 } 8269 8270 /* 8271 * IBTF wrappers used for resource accounting 8272 */ 8273 static ibt_status_t 8274 daplka_ibt_alloc_rc_channel(daplka_ep_resource_t *ep_rp, ibt_hca_hdl_t hca_hdl, 8275 ibt_chan_alloc_flags_t flags, ibt_rc_chan_alloc_args_t *args, 8276 ibt_channel_hdl_t *chan_hdl_p, ibt_chan_sizes_t *sizes) 8277 { 8278 daplka_hca_t *hca_p; 8279 uint32_t max_qps; 8280 boolean_t acct_enabled; 8281 ibt_status_t status; 8282 8283 acct_enabled = daplka_accounting_enabled; 8284 hca_p = ep_rp->ep_hca; 8285 max_qps = daplka_max_qp_percent * hca_p->hca_attr.hca_max_chans / 100; 8286 8287 if (acct_enabled) { 8288 if (daplka_max_qp_percent != 0 && 8289 max_qps <= hca_p->hca_qp_count) { 8290 DERR("ibt_alloc_rc_channel: resource limit exceeded " 8291 "(limit %d, count %d)\n", max_qps, 8292 hca_p->hca_qp_count); 8293 return (IBT_INSUFF_RESOURCE); 8294 } 8295 DAPLKA_RS_ACCT_INC(ep_rp, 1); 8296 atomic_add_32(&hca_p->hca_qp_count, 1); 8297 } 8298 status = ibt_alloc_rc_channel(hca_hdl, flags, args, chan_hdl_p, sizes); 8299 8300 if (status != IBT_SUCCESS && acct_enabled) { 8301 DAPLKA_RS_ACCT_DEC(ep_rp, 1); 8302 atomic_add_32(&hca_p->hca_qp_count, -1); 8303 } 8304 return (status); 8305 } 8306 8307 static ibt_status_t 8308 daplka_ibt_free_channel(daplka_ep_resource_t *ep_rp, ibt_channel_hdl_t chan_hdl) 8309 { 8310 daplka_hca_t *hca_p; 8311 ibt_status_t status; 8312 8313 hca_p = ep_rp->ep_hca; 8314 8315 status = ibt_free_channel(chan_hdl); 8316 if (status != IBT_SUCCESS) { 8317 return (status); 8318 } 8319 if (DAPLKA_RS_ACCT_CHARGED(ep_rp) > 0) { 8320 DAPLKA_RS_ACCT_DEC(ep_rp, 1); 8321 atomic_add_32(&hca_p->hca_qp_count, -1); 8322 } 8323 return (status); 8324 } 8325 8326 static ibt_status_t 8327 daplka_ibt_alloc_cq(daplka_evd_resource_t *evd_rp, ibt_hca_hdl_t hca_hdl, 8328 ibt_cq_attr_t *cq_attr, ibt_cq_hdl_t *ibt_cq_p, uint32_t *real_size) 8329 { 8330 daplka_hca_t *hca_p; 8331 uint32_t max_cqs; 8332 boolean_t acct_enabled; 8333 ibt_status_t status; 8334 8335 acct_enabled = daplka_accounting_enabled; 8336 hca_p = evd_rp->evd_hca; 8337 max_cqs = daplka_max_cq_percent * hca_p->hca_attr.hca_max_cq / 100; 8338 8339 if (acct_enabled) { 8340 if (daplka_max_cq_percent != 0 && 8341 max_cqs <= hca_p->hca_cq_count) { 8342 DERR("ibt_alloc_cq: resource limit exceeded " 8343 "(limit %d, count %d)\n", max_cqs, 8344 hca_p->hca_cq_count); 8345 return (IBT_INSUFF_RESOURCE); 8346 } 8347 DAPLKA_RS_ACCT_INC(evd_rp, 1); 8348 atomic_add_32(&hca_p->hca_cq_count, 1); 8349 } 8350 status = ibt_alloc_cq(hca_hdl, cq_attr, ibt_cq_p, real_size); 8351 8352 if (status != IBT_SUCCESS && acct_enabled) { 8353 DAPLKA_RS_ACCT_DEC(evd_rp, 1); 8354 atomic_add_32(&hca_p->hca_cq_count, -1); 8355 } 8356 return (status); 8357 } 8358 8359 static ibt_status_t 8360 daplka_ibt_free_cq(daplka_evd_resource_t *evd_rp, ibt_cq_hdl_t cq_hdl) 8361 { 8362 daplka_hca_t *hca_p; 8363 ibt_status_t status; 8364 8365 hca_p = evd_rp->evd_hca; 8366 8367 status = ibt_free_cq(cq_hdl); 8368 if (status != IBT_SUCCESS) { 8369 return (status); 8370 } 8371 if (DAPLKA_RS_ACCT_CHARGED(evd_rp) > 0) { 8372 DAPLKA_RS_ACCT_DEC(evd_rp, 1); 8373 atomic_add_32(&hca_p->hca_cq_count, -1); 8374 } 8375 return (status); 8376 } 8377 8378 static ibt_status_t 8379 daplka_ibt_alloc_pd(daplka_pd_resource_t *pd_rp, ibt_hca_hdl_t hca_hdl, 8380 ibt_pd_flags_t flags, ibt_pd_hdl_t *pd_hdl_p) 8381 { 8382 daplka_hca_t *hca_p; 8383 uint32_t max_pds; 8384 boolean_t acct_enabled; 8385 ibt_status_t status; 8386 8387 acct_enabled = daplka_accounting_enabled; 8388 hca_p = pd_rp->pd_hca; 8389 max_pds = daplka_max_pd_percent * hca_p->hca_attr.hca_max_pd / 100; 8390 8391 if (acct_enabled) { 8392 if (daplka_max_pd_percent != 0 && 8393 max_pds <= hca_p->hca_pd_count) { 8394 DERR("ibt_alloc_pd: resource limit exceeded " 8395 "(limit %d, count %d)\n", max_pds, 8396 hca_p->hca_pd_count); 8397 return (IBT_INSUFF_RESOURCE); 8398 } 8399 DAPLKA_RS_ACCT_INC(pd_rp, 1); 8400 atomic_add_32(&hca_p->hca_pd_count, 1); 8401 } 8402 status = ibt_alloc_pd(hca_hdl, flags, pd_hdl_p); 8403 8404 if (status != IBT_SUCCESS && acct_enabled) { 8405 DAPLKA_RS_ACCT_DEC(pd_rp, 1); 8406 atomic_add_32(&hca_p->hca_pd_count, -1); 8407 } 8408 return (status); 8409 } 8410 8411 static ibt_status_t 8412 daplka_ibt_free_pd(daplka_pd_resource_t *pd_rp, ibt_hca_hdl_t hca_hdl, 8413 ibt_pd_hdl_t pd_hdl) 8414 { 8415 daplka_hca_t *hca_p; 8416 ibt_status_t status; 8417 8418 hca_p = pd_rp->pd_hca; 8419 8420 status = ibt_free_pd(hca_hdl, pd_hdl); 8421 if (status != IBT_SUCCESS) { 8422 return (status); 8423 } 8424 if (DAPLKA_RS_ACCT_CHARGED(pd_rp) > 0) { 8425 DAPLKA_RS_ACCT_DEC(pd_rp, 1); 8426 atomic_add_32(&hca_p->hca_pd_count, -1); 8427 } 8428 return (status); 8429 } 8430 8431 static ibt_status_t 8432 daplka_ibt_alloc_mw(daplka_mw_resource_t *mw_rp, ibt_hca_hdl_t hca_hdl, 8433 ibt_pd_hdl_t pd_hdl, ibt_mw_flags_t flags, ibt_mw_hdl_t *mw_hdl_p, 8434 ibt_rkey_t *rkey_p) 8435 { 8436 daplka_hca_t *hca_p; 8437 uint32_t max_mws; 8438 boolean_t acct_enabled; 8439 ibt_status_t status; 8440 8441 acct_enabled = daplka_accounting_enabled; 8442 hca_p = mw_rp->mw_hca; 8443 max_mws = daplka_max_mw_percent * hca_p->hca_attr.hca_max_mem_win / 100; 8444 8445 if (acct_enabled) { 8446 if (daplka_max_mw_percent != 0 && 8447 max_mws <= hca_p->hca_mw_count) { 8448 DERR("ibt_alloc_mw: resource limit exceeded " 8449 "(limit %d, count %d)\n", max_mws, 8450 hca_p->hca_mw_count); 8451 return (IBT_INSUFF_RESOURCE); 8452 } 8453 DAPLKA_RS_ACCT_INC(mw_rp, 1); 8454 atomic_add_32(&hca_p->hca_mw_count, 1); 8455 } 8456 status = ibt_alloc_mw(hca_hdl, pd_hdl, flags, mw_hdl_p, rkey_p); 8457 8458 if (status != IBT_SUCCESS && acct_enabled) { 8459 DAPLKA_RS_ACCT_DEC(mw_rp, 1); 8460 atomic_add_32(&hca_p->hca_mw_count, -1); 8461 } 8462 return (status); 8463 } 8464 8465 static ibt_status_t 8466 daplka_ibt_free_mw(daplka_mw_resource_t *mw_rp, ibt_hca_hdl_t hca_hdl, 8467 ibt_mw_hdl_t mw_hdl) 8468 { 8469 daplka_hca_t *hca_p; 8470 ibt_status_t status; 8471 8472 hca_p = mw_rp->mw_hca; 8473 8474 status = ibt_free_mw(hca_hdl, mw_hdl); 8475 if (status != IBT_SUCCESS) { 8476 return (status); 8477 } 8478 if (DAPLKA_RS_ACCT_CHARGED(mw_rp) > 0) { 8479 DAPLKA_RS_ACCT_DEC(mw_rp, 1); 8480 atomic_add_32(&hca_p->hca_mw_count, -1); 8481 } 8482 return (status); 8483 } 8484 8485 static ibt_status_t 8486 daplka_ibt_register_mr(daplka_mr_resource_t *mr_rp, ibt_hca_hdl_t hca_hdl, 8487 ibt_pd_hdl_t pd_hdl, ibt_mr_attr_t *mr_attr, ibt_mr_hdl_t *mr_hdl_p, 8488 ibt_mr_desc_t *mr_desc_p) 8489 { 8490 daplka_hca_t *hca_p; 8491 uint32_t max_mrs; 8492 boolean_t acct_enabled; 8493 ibt_status_t status; 8494 8495 acct_enabled = daplka_accounting_enabled; 8496 hca_p = mr_rp->mr_hca; 8497 max_mrs = daplka_max_mr_percent * hca_p->hca_attr.hca_max_memr / 100; 8498 8499 if (acct_enabled) { 8500 if (daplka_max_mr_percent != 0 && 8501 max_mrs <= hca_p->hca_mr_count) { 8502 DERR("ibt_register_mr: resource limit exceeded " 8503 "(limit %d, count %d)\n", max_mrs, 8504 hca_p->hca_mr_count); 8505 return (IBT_INSUFF_RESOURCE); 8506 } 8507 DAPLKA_RS_ACCT_INC(mr_rp, 1); 8508 atomic_add_32(&hca_p->hca_mr_count, 1); 8509 } 8510 status = ibt_register_mr(hca_hdl, pd_hdl, mr_attr, mr_hdl_p, mr_desc_p); 8511 8512 if (status != IBT_SUCCESS && acct_enabled) { 8513 DAPLKA_RS_ACCT_DEC(mr_rp, 1); 8514 atomic_add_32(&hca_p->hca_mr_count, -1); 8515 } 8516 return (status); 8517 } 8518 8519 static ibt_status_t 8520 daplka_ibt_register_shared_mr(daplka_mr_resource_t *mr_rp, 8521 ibt_hca_hdl_t hca_hdl, ibt_mr_hdl_t mr_hdl, ibt_pd_hdl_t pd_hdl, 8522 ibt_smr_attr_t *smr_attr_p, ibt_mr_hdl_t *mr_hdl_p, 8523 ibt_mr_desc_t *mr_desc_p) 8524 { 8525 daplka_hca_t *hca_p; 8526 uint32_t max_mrs; 8527 boolean_t acct_enabled; 8528 ibt_status_t status; 8529 8530 acct_enabled = daplka_accounting_enabled; 8531 hca_p = mr_rp->mr_hca; 8532 max_mrs = daplka_max_mr_percent * hca_p->hca_attr.hca_max_memr / 100; 8533 8534 if (acct_enabled) { 8535 if (daplka_max_mr_percent != 0 && 8536 max_mrs <= hca_p->hca_mr_count) { 8537 DERR("ibt_register_shared_mr: resource limit exceeded " 8538 "(limit %d, count %d)\n", max_mrs, 8539 hca_p->hca_mr_count); 8540 return (IBT_INSUFF_RESOURCE); 8541 } 8542 DAPLKA_RS_ACCT_INC(mr_rp, 1); 8543 atomic_add_32(&hca_p->hca_mr_count, 1); 8544 } 8545 status = ibt_register_shared_mr(hca_hdl, mr_hdl, pd_hdl, 8546 smr_attr_p, mr_hdl_p, mr_desc_p); 8547 8548 if (status != IBT_SUCCESS && acct_enabled) { 8549 DAPLKA_RS_ACCT_DEC(mr_rp, 1); 8550 atomic_add_32(&hca_p->hca_mr_count, -1); 8551 } 8552 return (status); 8553 } 8554 8555 static ibt_status_t 8556 daplka_ibt_deregister_mr(daplka_mr_resource_t *mr_rp, ibt_hca_hdl_t hca_hdl, 8557 ibt_mr_hdl_t mr_hdl) 8558 { 8559 daplka_hca_t *hca_p; 8560 ibt_status_t status; 8561 8562 hca_p = mr_rp->mr_hca; 8563 8564 status = ibt_deregister_mr(hca_hdl, mr_hdl); 8565 if (status != IBT_SUCCESS) { 8566 return (status); 8567 } 8568 if (DAPLKA_RS_ACCT_CHARGED(mr_rp) > 0) { 8569 DAPLKA_RS_ACCT_DEC(mr_rp, 1); 8570 atomic_add_32(&hca_p->hca_mr_count, -1); 8571 } 8572 return (status); 8573 } 8574 8575 static ibt_status_t 8576 daplka_ibt_alloc_srq(daplka_srq_resource_t *srq_rp, ibt_hca_hdl_t hca_hdl, 8577 ibt_srq_flags_t flags, ibt_pd_hdl_t pd, ibt_srq_sizes_t *reqsz, 8578 ibt_srq_hdl_t *srq_hdl_p, ibt_srq_sizes_t *realsz) 8579 { 8580 daplka_hca_t *hca_p; 8581 uint32_t max_srqs; 8582 boolean_t acct_enabled; 8583 ibt_status_t status; 8584 8585 acct_enabled = daplka_accounting_enabled; 8586 hca_p = srq_rp->srq_hca; 8587 max_srqs = daplka_max_srq_percent * hca_p->hca_attr.hca_max_srqs / 100; 8588 8589 if (acct_enabled) { 8590 if (daplka_max_srq_percent != 0 && 8591 max_srqs <= hca_p->hca_srq_count) { 8592 DERR("ibt_alloc_srq: resource limit exceeded " 8593 "(limit %d, count %d)\n", max_srqs, 8594 hca_p->hca_srq_count); 8595 return (IBT_INSUFF_RESOURCE); 8596 } 8597 DAPLKA_RS_ACCT_INC(srq_rp, 1); 8598 atomic_add_32(&hca_p->hca_srq_count, 1); 8599 } 8600 status = ibt_alloc_srq(hca_hdl, flags, pd, reqsz, srq_hdl_p, realsz); 8601 8602 if (status != IBT_SUCCESS && acct_enabled) { 8603 DAPLKA_RS_ACCT_DEC(srq_rp, 1); 8604 atomic_add_32(&hca_p->hca_srq_count, -1); 8605 } 8606 return (status); 8607 } 8608 8609 static ibt_status_t 8610 daplka_ibt_free_srq(daplka_srq_resource_t *srq_rp, ibt_srq_hdl_t srq_hdl) 8611 { 8612 daplka_hca_t *hca_p; 8613 ibt_status_t status; 8614 8615 hca_p = srq_rp->srq_hca; 8616 8617 D3("ibt_free_srq: %p %p\n", srq_rp, srq_hdl); 8618 8619 status = ibt_free_srq(srq_hdl); 8620 if (status != IBT_SUCCESS) { 8621 return (status); 8622 } 8623 if (DAPLKA_RS_ACCT_CHARGED(srq_rp) > 0) { 8624 DAPLKA_RS_ACCT_DEC(srq_rp, 1); 8625 atomic_add_32(&hca_p->hca_srq_count, -1); 8626 } 8627 return (status); 8628 } 8629 8630 8631 static int 8632 daplka_common_ioctl(int cmd, minor_t rnum, intptr_t arg, int mode, 8633 cred_t *cred, int *rvalp) 8634 { 8635 int error; 8636 8637 switch (cmd) { 8638 case DAPL_IA_CREATE: 8639 error = daplka_ia_create(rnum, arg, mode, cred, rvalp); 8640 break; 8641 8642 /* can potentially add other commands here */ 8643 8644 default: 8645 DERR("daplka_common_ioctl: cmd not supported\n"); 8646 error = DDI_FAILURE; 8647 } 8648 return (error); 8649 } 8650 8651 static int 8652 daplka_evd_ioctl(int cmd, daplka_ia_resource_t *rp, intptr_t arg, int mode, 8653 cred_t *cred, int *rvalp) 8654 { 8655 int error; 8656 8657 switch (cmd) { 8658 case DAPL_EVD_CREATE: 8659 error = daplka_evd_create(rp, arg, mode, cred, rvalp); 8660 break; 8661 8662 case DAPL_CQ_RESIZE: 8663 error = daplka_cq_resize(rp, arg, mode, cred, rvalp); 8664 break; 8665 8666 case DAPL_EVENT_POLL: 8667 error = daplka_event_poll(rp, arg, mode, cred, rvalp); 8668 break; 8669 8670 case DAPL_EVENT_WAKEUP: 8671 error = daplka_event_wakeup(rp, arg, mode, cred, rvalp); 8672 break; 8673 8674 case DAPL_EVD_MODIFY_CNO: 8675 error = daplka_evd_modify_cno(rp, arg, mode, cred, rvalp); 8676 break; 8677 8678 case DAPL_EVD_FREE: 8679 error = daplka_evd_free(rp, arg, mode, cred, rvalp); 8680 break; 8681 8682 default: 8683 DERR("daplka_evd_ioctl: cmd not supported\n"); 8684 error = DDI_FAILURE; 8685 } 8686 return (error); 8687 } 8688 8689 static int 8690 daplka_ep_ioctl(int cmd, daplka_ia_resource_t *rp, intptr_t arg, int mode, 8691 cred_t *cred, int *rvalp) 8692 { 8693 int error; 8694 8695 switch (cmd) { 8696 case DAPL_EP_MODIFY: 8697 error = daplka_ep_modify(rp, arg, mode, cred, rvalp); 8698 break; 8699 8700 case DAPL_EP_FREE: 8701 error = daplka_ep_free(rp, arg, mode, cred, rvalp); 8702 break; 8703 8704 case DAPL_EP_CONNECT: 8705 error = daplka_ep_connect(rp, arg, mode, cred, rvalp); 8706 break; 8707 8708 case DAPL_EP_DISCONNECT: 8709 error = daplka_ep_disconnect(rp, arg, mode, cred, rvalp); 8710 break; 8711 8712 case DAPL_EP_REINIT: 8713 error = daplka_ep_reinit(rp, arg, mode, cred, rvalp); 8714 break; 8715 8716 case DAPL_EP_CREATE: 8717 error = daplka_ep_create(rp, arg, mode, cred, rvalp); 8718 break; 8719 8720 default: 8721 DERR("daplka_ep_ioctl: cmd not supported\n"); 8722 error = DDI_FAILURE; 8723 } 8724 return (error); 8725 } 8726 8727 static int 8728 daplka_mr_ioctl(int cmd, daplka_ia_resource_t *rp, intptr_t arg, int mode, 8729 cred_t *cred, int *rvalp) 8730 { 8731 int error; 8732 8733 switch (cmd) { 8734 case DAPL_MR_REGISTER: 8735 error = daplka_mr_register(rp, arg, mode, cred, rvalp); 8736 break; 8737 8738 case DAPL_MR_REGISTER_LMR: 8739 error = daplka_mr_register_lmr(rp, arg, mode, cred, rvalp); 8740 break; 8741 8742 case DAPL_MR_REGISTER_SHARED: 8743 error = daplka_mr_register_shared(rp, arg, mode, cred, rvalp); 8744 break; 8745 8746 case DAPL_MR_DEREGISTER: 8747 error = daplka_mr_deregister(rp, arg, mode, cred, rvalp); 8748 break; 8749 8750 case DAPL_MR_SYNC: 8751 error = daplka_mr_sync(rp, arg, mode, cred, rvalp); 8752 break; 8753 8754 default: 8755 DERR("daplka_mr_ioctl: cmd not supported\n"); 8756 error = DDI_FAILURE; 8757 } 8758 return (error); 8759 } 8760 8761 static int 8762 daplka_mw_ioctl(int cmd, daplka_ia_resource_t *rp, intptr_t arg, int mode, 8763 cred_t *cred, int *rvalp) 8764 { 8765 int error; 8766 8767 switch (cmd) { 8768 case DAPL_MW_ALLOC: 8769 error = daplka_mw_alloc(rp, arg, mode, cred, rvalp); 8770 break; 8771 8772 case DAPL_MW_FREE: 8773 error = daplka_mw_free(rp, arg, mode, cred, rvalp); 8774 break; 8775 8776 default: 8777 DERR("daplka_mw_ioctl: cmd not supported\n"); 8778 error = DDI_FAILURE; 8779 } 8780 return (error); 8781 } 8782 8783 static int 8784 daplka_cno_ioctl(int cmd, daplka_ia_resource_t *rp, intptr_t arg, int mode, 8785 cred_t *cred, int *rvalp) 8786 { 8787 int error; 8788 8789 switch (cmd) { 8790 case DAPL_CNO_ALLOC: 8791 error = daplka_cno_alloc(rp, arg, mode, cred, rvalp); 8792 break; 8793 8794 case DAPL_CNO_FREE: 8795 error = daplka_cno_free(rp, arg, mode, cred, rvalp); 8796 break; 8797 8798 case DAPL_CNO_WAIT: 8799 error = daplka_cno_wait(rp, arg, mode, cred, rvalp); 8800 break; 8801 8802 default: 8803 DERR("daplka_cno_ioctl: cmd not supported\n"); 8804 error = DDI_FAILURE; 8805 } 8806 return (error); 8807 } 8808 8809 static int 8810 daplka_pd_ioctl(int cmd, daplka_ia_resource_t *rp, intptr_t arg, int mode, 8811 cred_t *cred, int *rvalp) 8812 { 8813 int error; 8814 8815 switch (cmd) { 8816 case DAPL_PD_ALLOC: 8817 error = daplka_pd_alloc(rp, arg, mode, cred, rvalp); 8818 break; 8819 8820 case DAPL_PD_FREE: 8821 error = daplka_pd_free(rp, arg, mode, cred, rvalp); 8822 break; 8823 8824 default: 8825 DERR("daplka_pd_ioctl: cmd not supported\n"); 8826 error = DDI_FAILURE; 8827 } 8828 return (error); 8829 } 8830 8831 static int 8832 daplka_sp_ioctl(int cmd, daplka_ia_resource_t *rp, intptr_t arg, int mode, 8833 cred_t *cred, int *rvalp) 8834 { 8835 int error; 8836 8837 switch (cmd) { 8838 case DAPL_SERVICE_REGISTER: 8839 error = daplka_service_register(rp, arg, mode, cred, rvalp); 8840 break; 8841 8842 case DAPL_SERVICE_DEREGISTER: 8843 error = daplka_service_deregister(rp, arg, mode, cred, rvalp); 8844 break; 8845 8846 default: 8847 DERR("daplka_sp_ioctl: cmd not supported\n"); 8848 error = DDI_FAILURE; 8849 } 8850 return (error); 8851 } 8852 8853 static int 8854 daplka_srq_ioctl(int cmd, daplka_ia_resource_t *rp, intptr_t arg, int mode, 8855 cred_t *cred, int *rvalp) 8856 { 8857 int error; 8858 8859 switch (cmd) { 8860 case DAPL_SRQ_CREATE: 8861 error = daplka_srq_create(rp, arg, mode, cred, rvalp); 8862 break; 8863 8864 case DAPL_SRQ_RESIZE: 8865 error = daplka_srq_resize(rp, arg, mode, cred, rvalp); 8866 break; 8867 8868 case DAPL_SRQ_FREE: 8869 error = daplka_srq_free(rp, arg, mode, cred, rvalp); 8870 break; 8871 8872 default: 8873 DERR("daplka_srq_ioctl: cmd(%d) not supported\n", cmd); 8874 error = DDI_FAILURE; 8875 break; 8876 } 8877 return (error); 8878 } 8879 8880 static int 8881 daplka_misc_ioctl(int cmd, daplka_ia_resource_t *rp, intptr_t arg, int mode, 8882 cred_t *cred, int *rvalp) 8883 { 8884 int error; 8885 8886 switch (cmd) { 8887 case DAPL_CR_ACCEPT: 8888 error = daplka_cr_accept(rp, arg, mode, cred, rvalp); 8889 break; 8890 8891 case DAPL_CR_REJECT: 8892 error = daplka_cr_reject(rp, arg, mode, cred, rvalp); 8893 break; 8894 8895 case DAPL_IA_QUERY: 8896 error = daplka_ia_query(rp, arg, mode, cred, rvalp); 8897 break; 8898 8899 case DAPL_CR_HANDOFF: 8900 error = daplka_cr_handoff(rp, arg, mode, cred, rvalp); 8901 break; 8902 8903 default: 8904 DERR("daplka_misc_ioctl: cmd not supported\n"); 8905 error = DDI_FAILURE; 8906 } 8907 return (error); 8908 } 8909 8910 /*ARGSUSED*/ 8911 static int 8912 daplka_ioctl(dev_t dev, int cmd, intptr_t arg, int mode, cred_t *cred, 8913 int *rvalp) 8914 { 8915 daplka_ia_resource_t *ia_rp; 8916 minor_t rnum; 8917 int error = 0; 8918 8919 rnum = getminor(dev); 8920 ia_rp = (daplka_ia_resource_t *)daplka_resource_lookup(rnum); 8921 if (ia_rp == NULL) { 8922 DERR("ioctl: resource not found, rnum %d\n", rnum); 8923 return (ENXIO); 8924 } 8925 8926 D4("ioctl: rnum = %d, cmd = 0x%x\n", rnum, cmd); 8927 if (DAPLKA_RS_RESERVED(ia_rp)) { 8928 error = daplka_common_ioctl(cmd, rnum, arg, mode, cred, rvalp); 8929 return (error); 8930 } 8931 if (DAPLKA_RS_TYPE(ia_rp) != DAPL_TYPE_IA) { 8932 DERR("ioctl: invalid type %d\n", DAPLKA_RS_TYPE(ia_rp)); 8933 error = EINVAL; 8934 goto cleanup; 8935 } 8936 if (ia_rp->ia_pid != ddi_get_pid()) { 8937 DERR("ioctl: ia_pid %d != pid %d\n", 8938 ia_rp->ia_pid, ddi_get_pid()); 8939 error = EINVAL; 8940 goto cleanup; 8941 } 8942 8943 switch (cmd & DAPL_TYPE_MASK) { 8944 case DAPL_TYPE_EVD: 8945 error = daplka_evd_ioctl(cmd, ia_rp, arg, mode, cred, rvalp); 8946 break; 8947 8948 case DAPL_TYPE_EP: 8949 error = daplka_ep_ioctl(cmd, ia_rp, arg, mode, cred, rvalp); 8950 break; 8951 8952 case DAPL_TYPE_MR: 8953 error = daplka_mr_ioctl(cmd, ia_rp, arg, mode, cred, rvalp); 8954 break; 8955 8956 case DAPL_TYPE_MW: 8957 error = daplka_mw_ioctl(cmd, ia_rp, arg, mode, cred, rvalp); 8958 break; 8959 8960 case DAPL_TYPE_PD: 8961 error = daplka_pd_ioctl(cmd, ia_rp, arg, mode, cred, rvalp); 8962 break; 8963 8964 case DAPL_TYPE_SP: 8965 error = daplka_sp_ioctl(cmd, ia_rp, arg, mode, cred, rvalp); 8966 break; 8967 8968 case DAPL_TYPE_CNO: 8969 error = daplka_cno_ioctl(cmd, ia_rp, arg, mode, cred, rvalp); 8970 break; 8971 8972 case DAPL_TYPE_MISC: 8973 error = daplka_misc_ioctl(cmd, ia_rp, arg, mode, cred, rvalp); 8974 break; 8975 8976 case DAPL_TYPE_SRQ: 8977 error = daplka_srq_ioctl(cmd, ia_rp, arg, mode, cred, rvalp); 8978 break; 8979 8980 default: 8981 DERR("ioctl: invalid dapl type = %d\n", DAPLKA_RS_TYPE(ia_rp)); 8982 error = DDI_FAILURE; 8983 } 8984 8985 cleanup:; 8986 DAPLKA_RS_UNREF(ia_rp); 8987 return (error); 8988 } 8989 8990 /* ARGSUSED */ 8991 static int 8992 daplka_open(dev_t *devp, int flag, int otyp, struct cred *cred) 8993 { 8994 minor_t rnum; 8995 8996 /* 8997 * Char only 8998 */ 8999 if (otyp != OTYP_CHR) { 9000 return (EINVAL); 9001 } 9002 9003 /* 9004 * Only zero can be opened, clones are used for resources. 9005 */ 9006 if (getminor(*devp) != DAPLKA_DRIVER_MINOR) { 9007 DERR("daplka_open: bad minor %d\n", getminor(*devp)); 9008 return (ENODEV); 9009 } 9010 9011 /* 9012 * - allocate new minor number 9013 * - update devp argument to new device 9014 */ 9015 if (daplka_resource_reserve(&rnum) == 0) { 9016 *devp = makedevice(getmajor(*devp), rnum); 9017 } else { 9018 return (ENOMEM); 9019 } 9020 9021 return (DDI_SUCCESS); 9022 } 9023 9024 /* ARGSUSED */ 9025 static int 9026 daplka_close(dev_t dev, int flag, int otyp, struct cred *cred) 9027 { 9028 daplka_ia_resource_t *ia_rp; 9029 minor_t rnum = getminor(dev); 9030 9031 /* 9032 * Char only 9033 */ 9034 if (otyp != OTYP_CHR) { 9035 return (EINVAL); 9036 } 9037 D2("daplka_close: closing rnum = %d\n", rnum); 9038 atomic_add_32(&daplka_pending_close, 1); 9039 9040 /* 9041 * remove from resource table. 9042 */ 9043 ia_rp = (daplka_ia_resource_t *)daplka_resource_remove(rnum); 9044 9045 /* 9046 * remove the initial reference 9047 */ 9048 if (ia_rp != NULL) { 9049 DAPLKA_RS_UNREF(ia_rp); 9050 } 9051 atomic_add_32(&daplka_pending_close, -1); 9052 return (DDI_SUCCESS); 9053 } 9054 9055 9056 /* 9057 * Resource management routines 9058 * 9059 * We start with no resource array. Each time we run out of slots, we 9060 * reallocate a new larger array and copy the pointer to the new array and 9061 * a new resource blk is allocated and added to the hash table. 9062 * 9063 * The resource control block contains: 9064 * root - array of pointer of resource blks 9065 * sz - current size of array. 9066 * len - last valid entry in array. 9067 * 9068 * A search operation based on a resource number is as follows: 9069 * index = rnum / RESOURCE_BLKSZ; 9070 * ASSERT(index < resource_block.len); 9071 * ASSERT(index < resource_block.sz); 9072 * offset = rnum % RESOURCE_BLKSZ; 9073 * ASSERT(offset >= resource_block.root[index]->base); 9074 * ASSERT(offset < resource_block.root[index]->base + RESOURCE_BLKSZ); 9075 * return resource_block.root[index]->blks[offset]; 9076 * 9077 * A resource blk is freed when its used count reaches zero. 9078 */ 9079 9080 /* 9081 * initializes the global resource table 9082 */ 9083 static void 9084 daplka_resource_init(void) 9085 { 9086 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(daplka_resource)) 9087 rw_init(&daplka_resource.daplka_rct_lock, NULL, RW_DRIVER, NULL); 9088 daplka_resource.daplka_rc_len = 0; 9089 daplka_resource.daplka_rc_sz = 0; 9090 daplka_resource.daplka_rc_cnt = 0; 9091 daplka_resource.daplka_rc_flag = 0; 9092 daplka_resource.daplka_rc_root = NULL; 9093 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(daplka_resource)) 9094 } 9095 9096 /* 9097 * destroys the global resource table 9098 */ 9099 static void 9100 daplka_resource_fini(void) 9101 { 9102 int i; 9103 9104 rw_enter(&daplka_resource.daplka_rct_lock, RW_WRITER); 9105 for (i = 0; i < daplka_resource.daplka_rc_len; i++) { 9106 daplka_resource_blk_t *blk; 9107 int j; 9108 9109 blk = daplka_resource.daplka_rc_root[i]; 9110 if (blk == NULL) { 9111 continue; 9112 } 9113 for (j = 0; j < DAPLKA_RC_BLKSZ; j++) { 9114 if (blk->daplka_rcblk_blks[j] != NULL) { 9115 DERR("resource_fini: non-null slot %d, %p\n", 9116 j, blk->daplka_rcblk_blks[j]); 9117 } 9118 } 9119 kmem_free(blk, sizeof (*blk)); 9120 daplka_resource.daplka_rc_root[i] = NULL; 9121 } 9122 if (daplka_resource.daplka_rc_root != NULL) { 9123 uint_t sz; 9124 9125 sz = daplka_resource.daplka_rc_sz * 9126 sizeof (daplka_resource_blk_t *); 9127 kmem_free(daplka_resource.daplka_rc_root, (uint_t)sz); 9128 daplka_resource.daplka_rc_root = NULL; 9129 daplka_resource.daplka_rc_len = 0; 9130 daplka_resource.daplka_rc_sz = 0; 9131 } 9132 rw_exit(&daplka_resource.daplka_rct_lock); 9133 rw_destroy(&daplka_resource.daplka_rct_lock); 9134 } 9135 9136 /* 9137 * reserves a slot in the global resource table. 9138 * this is called by the open() syscall. it is needed because 9139 * at open() time, we do not have sufficient information to 9140 * create an IA resource. the library needs to subsequently 9141 * call daplka_ia_create to insert an IA resource into this 9142 * reserved slot. 9143 */ 9144 static int 9145 daplka_resource_reserve(minor_t *rnum) 9146 { 9147 int i, j, empty = -1; 9148 daplka_resource_blk_t *blk; 9149 9150 rw_enter(&daplka_resource.daplka_rct_lock, RW_WRITER); 9151 /* 9152 * Try to find an empty slot 9153 */ 9154 for (i = 0; i < daplka_resource.daplka_rc_len; i++) { 9155 blk = daplka_resource.daplka_rc_root[i]; 9156 if (blk != NULL && blk->daplka_rcblk_avail > 0) { 9157 9158 D3("resource_alloc: available blks %d\n", 9159 blk->daplka_rcblk_avail); 9160 9161 /* 9162 * found an empty slot in this blk 9163 */ 9164 for (j = 0; j < DAPLKA_RC_BLKSZ; j++) { 9165 if (blk->daplka_rcblk_blks[j] == NULL) { 9166 *rnum = (minor_t) 9167 (j + (i * DAPLKA_RC_BLKSZ)); 9168 blk->daplka_rcblk_blks[j] = 9169 (daplka_resource_t *) 9170 DAPLKA_RC_RESERVED; 9171 blk->daplka_rcblk_avail--; 9172 daplka_resource.daplka_rc_cnt++; 9173 rw_exit(&daplka_resource. 9174 daplka_rct_lock); 9175 return (0); 9176 } 9177 } 9178 } else if (blk == NULL && empty < 0) { 9179 /* 9180 * remember first empty slot 9181 */ 9182 empty = i; 9183 } 9184 } 9185 9186 /* 9187 * Couldn't find anything, allocate a new blk 9188 * Do we need to reallocate the root array 9189 */ 9190 if (empty < 0) { 9191 if (daplka_resource.daplka_rc_len == 9192 daplka_resource.daplka_rc_sz) { 9193 /* 9194 * Allocate new array and copy current stuff into it 9195 */ 9196 daplka_resource_blk_t **p; 9197 uint_t newsz = (uint_t)daplka_resource.daplka_rc_sz + 9198 DAPLKA_RC_BLKSZ; 9199 9200 D3("resource_alloc: increasing no. of buckets to %d\n", 9201 newsz); 9202 9203 p = kmem_zalloc(newsz * sizeof (*p), daplka_km_flags); 9204 9205 if (daplka_resource.daplka_rc_root) { 9206 uint_t oldsz; 9207 9208 oldsz = (uint_t)(daplka_resource.daplka_rc_sz * 9209 (int)sizeof (*p)); 9210 9211 /* 9212 * Copy old data into new space and 9213 * free old stuff 9214 */ 9215 bcopy(daplka_resource.daplka_rc_root, p, oldsz); 9216 kmem_free(daplka_resource.daplka_rc_root, 9217 oldsz); 9218 } 9219 9220 daplka_resource.daplka_rc_root = p; 9221 daplka_resource.daplka_rc_sz = (int)newsz; 9222 } 9223 9224 empty = daplka_resource.daplka_rc_len; 9225 daplka_resource.daplka_rc_len++; 9226 9227 D3("resource_alloc: daplka_rc_len %d\n", 9228 daplka_resource.daplka_rc_len); 9229 } 9230 9231 /* 9232 * Allocate a new blk 9233 */ 9234 blk = kmem_zalloc(sizeof (*blk), daplka_km_flags); 9235 ASSERT(daplka_resource.daplka_rc_root[empty] == NULL); 9236 daplka_resource.daplka_rc_root[empty] = blk; 9237 blk->daplka_rcblk_avail = DAPLKA_RC_BLKSZ - 1; 9238 9239 /* 9240 * Allocate slot 9241 */ 9242 *rnum = (minor_t)(empty * DAPLKA_RC_BLKSZ); 9243 blk->daplka_rcblk_blks[0] = (daplka_resource_t *)DAPLKA_RC_RESERVED; 9244 daplka_resource.daplka_rc_cnt++; 9245 rw_exit(&daplka_resource.daplka_rct_lock); 9246 9247 return (0); 9248 } 9249 9250 /* 9251 * removes resource from global resource table 9252 */ 9253 static daplka_resource_t * 9254 daplka_resource_remove(minor_t rnum) 9255 { 9256 int i, j; 9257 daplka_resource_blk_t *blk; 9258 daplka_resource_t *p; 9259 9260 i = (int)(rnum / DAPLKA_RC_BLKSZ); 9261 j = (int)(rnum % DAPLKA_RC_BLKSZ); 9262 9263 rw_enter(&daplka_resource.daplka_rct_lock, RW_WRITER); 9264 if (i >= daplka_resource.daplka_rc_len) { 9265 rw_exit(&daplka_resource.daplka_rct_lock); 9266 DERR("resource_remove: invalid rnum %d\n", rnum); 9267 return (NULL); 9268 } 9269 9270 ASSERT(daplka_resource.daplka_rc_root); 9271 ASSERT(i < daplka_resource.daplka_rc_len); 9272 ASSERT(i < daplka_resource.daplka_rc_sz); 9273 blk = daplka_resource.daplka_rc_root[i]; 9274 if (blk == NULL) { 9275 rw_exit(&daplka_resource.daplka_rct_lock); 9276 DERR("resource_remove: invalid rnum %d\n", rnum); 9277 return (NULL); 9278 } 9279 9280 if (blk->daplka_rcblk_blks[j] == NULL) { 9281 rw_exit(&daplka_resource.daplka_rct_lock); 9282 DERR("resource_remove: blk->daplka_rcblk_blks[j] == NULL\n"); 9283 return (NULL); 9284 } 9285 p = blk->daplka_rcblk_blks[j]; 9286 blk->daplka_rcblk_blks[j] = NULL; 9287 blk->daplka_rcblk_avail++; 9288 if (blk->daplka_rcblk_avail == DAPLKA_RC_BLKSZ) { 9289 /* 9290 * free this blk 9291 */ 9292 kmem_free(blk, sizeof (*blk)); 9293 daplka_resource.daplka_rc_root[i] = NULL; 9294 } 9295 daplka_resource.daplka_rc_cnt--; 9296 rw_exit(&daplka_resource.daplka_rct_lock); 9297 9298 if ((intptr_t)p == DAPLKA_RC_RESERVED) { 9299 return (NULL); 9300 } else { 9301 return (p); 9302 } 9303 } 9304 9305 /* 9306 * inserts resource into the slot designated by rnum 9307 */ 9308 static int 9309 daplka_resource_insert(minor_t rnum, daplka_resource_t *rp) 9310 { 9311 int i, j, error = -1; 9312 daplka_resource_blk_t *blk; 9313 9314 /* 9315 * Find resource and lock it in WRITER mode 9316 * search for available resource slot 9317 */ 9318 9319 i = (int)(rnum / DAPLKA_RC_BLKSZ); 9320 j = (int)(rnum % DAPLKA_RC_BLKSZ); 9321 9322 rw_enter(&daplka_resource.daplka_rct_lock, RW_WRITER); 9323 if (i >= daplka_resource.daplka_rc_len) { 9324 rw_exit(&daplka_resource.daplka_rct_lock); 9325 DERR("resource_insert: resource %d not found\n", rnum); 9326 return (-1); 9327 } 9328 9329 blk = daplka_resource.daplka_rc_root[i]; 9330 if (blk != NULL) { 9331 ASSERT(i < daplka_resource.daplka_rc_len); 9332 ASSERT(i < daplka_resource.daplka_rc_sz); 9333 9334 if ((intptr_t)blk->daplka_rcblk_blks[j] == DAPLKA_RC_RESERVED) { 9335 blk->daplka_rcblk_blks[j] = rp; 9336 error = 0; 9337 } else { 9338 DERR("resource_insert: %d not reserved, blk = %p\n", 9339 rnum, blk->daplka_rcblk_blks[j]); 9340 } 9341 } else { 9342 DERR("resource_insert: resource %d not found\n", rnum); 9343 } 9344 rw_exit(&daplka_resource.daplka_rct_lock); 9345 return (error); 9346 } 9347 9348 /* 9349 * finds resource using minor device number 9350 */ 9351 static daplka_resource_t * 9352 daplka_resource_lookup(minor_t rnum) 9353 { 9354 int i, j; 9355 daplka_resource_blk_t *blk; 9356 daplka_resource_t *rp; 9357 9358 /* 9359 * Find resource and lock it in READER mode 9360 * search for available resource slot 9361 */ 9362 9363 i = (int)(rnum / DAPLKA_RC_BLKSZ); 9364 j = (int)(rnum % DAPLKA_RC_BLKSZ); 9365 9366 rw_enter(&daplka_resource.daplka_rct_lock, RW_READER); 9367 if (i >= daplka_resource.daplka_rc_len) { 9368 rw_exit(&daplka_resource.daplka_rct_lock); 9369 DERR("resource_lookup: resource %d not found\n", rnum); 9370 return (NULL); 9371 } 9372 9373 blk = daplka_resource.daplka_rc_root[i]; 9374 if (blk != NULL) { 9375 ASSERT(i < daplka_resource.daplka_rc_len); 9376 ASSERT(i < daplka_resource.daplka_rc_sz); 9377 9378 rp = blk->daplka_rcblk_blks[j]; 9379 if (rp == NULL || (intptr_t)rp == DAPLKA_RC_RESERVED) { 9380 D3("resource_lookup: %d not found, blk = %p\n", 9381 rnum, blk->daplka_rcblk_blks[j]); 9382 } else { 9383 DAPLKA_RS_REF((daplka_ia_resource_t *)rp); 9384 } 9385 } else { 9386 DERR("resource_lookup: resource %d not found\n", rnum); 9387 rp = NULL; 9388 } 9389 rw_exit(&daplka_resource.daplka_rct_lock); 9390 return (rp); 9391 } 9392 9393 /* 9394 * generic hash table implementation 9395 */ 9396 9397 /* 9398 * daplka_hash_create: 9399 * initializes a hash table with the specified parameters 9400 * 9401 * input: 9402 * htblp pointer to hash table 9403 * 9404 * nbuckets number of buckets (must be power of 2) 9405 * 9406 * free_func this function is called on each hash 9407 * table element when daplka_hash_destroy 9408 * is called 9409 * 9410 * lookup_func if daplka_hash_lookup is able to find 9411 * the desired object, this function is 9412 * applied on the object before 9413 * daplka_hash_lookup returns 9414 * output: 9415 * none 9416 * 9417 * return value(s): 9418 * EINVAL nbuckets is not a power of 2 9419 * ENOMEM cannot allocate buckets 9420 * 0 success 9421 */ 9422 static int 9423 daplka_hash_create(daplka_hash_table_t *htblp, uint_t nbuckets, 9424 void (*free_func)(void *), void (*lookup_func)(void *)) 9425 { 9426 int i; 9427 9428 if ((nbuckets & ~(nbuckets - 1)) != nbuckets) { 9429 DERR("hash_create: nbuckets not power of 2\n"); 9430 return (EINVAL); 9431 } 9432 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*htblp)) 9433 9434 htblp->ht_buckets = 9435 kmem_zalloc(sizeof (daplka_hash_bucket_t) * nbuckets, 9436 daplka_km_flags); 9437 if (htblp->ht_buckets == NULL) { 9438 DERR("hash_create: cannot allocate buckets\n"); 9439 return (ENOMEM); 9440 } 9441 for (i = 0; i < nbuckets; i++) { 9442 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(htblp->ht_buckets[i])) 9443 htblp->ht_buckets[i].hb_count = 0; 9444 htblp->ht_buckets[i].hb_entries = NULL; 9445 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(htblp->ht_buckets[i])) 9446 } 9447 rw_init(&htblp->ht_table_lock, NULL, RW_DRIVER, NULL); 9448 mutex_init(&htblp->ht_key_lock, NULL, MUTEX_DRIVER, NULL); 9449 9450 htblp->ht_count = 0; 9451 htblp->ht_next_hkey = (uint64_t)gethrtime(); 9452 htblp->ht_nbuckets = nbuckets; 9453 htblp->ht_free_func = free_func; 9454 htblp->ht_lookup_func = lookup_func; 9455 htblp->ht_initialized = B_TRUE; 9456 D3("hash_create: done, buckets = %d\n", nbuckets); 9457 _NOTE(NOW_VISIBLE_TO_OTHER_THREADS(*htblp)) 9458 return (0); 9459 } 9460 9461 /* 9462 * daplka_hash_insert: 9463 * inserts an object into a hash table 9464 * 9465 * input: 9466 * htblp pointer to hash table 9467 * 9468 * hkeyp pointer to hash key. 9469 * *hkeyp being non-zero means that the caller 9470 * has generated its own hkey. if *hkeyp is zero, 9471 * this function will generate an hkey for the 9472 * caller. it is recommended that the caller 9473 * leave the hkey generation to this function 9474 * because the hkey is more likely to be evenly 9475 * distributed. 9476 * 9477 * objp pointer to object to be inserted into 9478 * hash table 9479 * 9480 * output: 9481 * hkeyp the generated hkey is returned via this pointer 9482 * 9483 * return value(s): 9484 * EINVAL invalid parameter 9485 * ENOMEM cannot allocate hash entry 9486 * 0 successful 9487 */ 9488 static int 9489 daplka_hash_insert(daplka_hash_table_t *htblp, uint64_t *hkeyp, void *objp) 9490 { 9491 daplka_hash_entry_t *hep, *curr_hep; 9492 daplka_hash_bucket_t *hbp; 9493 uint32_t bucket; 9494 uint64_t hkey; 9495 9496 if (hkeyp == NULL) { 9497 DERR("hash_insert: hkeyp == NULL\n"); 9498 return (EINVAL); 9499 } 9500 hep = kmem_zalloc(sizeof (*hep), daplka_km_flags); 9501 if (hep == NULL) { 9502 DERR("hash_insert: cannot alloc hash_entry\n"); 9503 return (ENOMEM); 9504 } 9505 if (*hkeyp == 0) { 9506 /* generate a new key */ 9507 mutex_enter(&htblp->ht_key_lock); 9508 hkey = ++htblp->ht_next_hkey; 9509 if (hkey == 0) { 9510 hkey = htblp->ht_next_hkey = (uint64_t)gethrtime(); 9511 } 9512 mutex_exit(&htblp->ht_key_lock); 9513 } else { 9514 /* use user generated key */ 9515 hkey = *hkeyp; 9516 } 9517 9518 /* only works if ht_nbuckets is a power of 2 */ 9519 bucket = (uint32_t)(hkey & (htblp->ht_nbuckets - 1)); 9520 ASSERT(objp != NULL); 9521 ASSERT(bucket < htblp->ht_nbuckets); 9522 9523 rw_enter(&htblp->ht_table_lock, RW_WRITER); 9524 hep->he_hkey = hkey; 9525 hep->he_objp = objp; 9526 9527 /* look for duplicate entries */ 9528 hbp = &htblp->ht_buckets[bucket]; 9529 curr_hep = hbp->hb_entries; 9530 while (curr_hep != NULL) { 9531 if (curr_hep->he_hkey == hep->he_hkey) { 9532 break; 9533 } 9534 curr_hep = curr_hep->he_next; 9535 } 9536 if (curr_hep != NULL) { 9537 DERR("hash_insert: found duplicate hash entry: " 9538 "bucket %d, hkey 0x%016llx\n", 9539 bucket, (longlong_t)hep->he_hkey); 9540 kmem_free(hep, sizeof (*hep)); 9541 rw_exit(&htblp->ht_table_lock); 9542 return (EINVAL); 9543 } 9544 hep->he_next = hbp->hb_entries; 9545 hbp->hb_entries = hep; 9546 hbp->hb_count++; 9547 htblp->ht_count++; 9548 rw_exit(&htblp->ht_table_lock); 9549 9550 if (*hkeyp == 0) { 9551 *hkeyp = hkey; 9552 ASSERT(*hkeyp != 0); 9553 } 9554 D3("hash_insert: htblp 0x%p, hkey = 0x%016llx, bucket = %d\n", 9555 htblp, (longlong_t)*hkeyp, bucket); 9556 return (0); 9557 } 9558 9559 /* 9560 * daplka_hash_remove: 9561 * removes object identified by hkey from hash table 9562 * 9563 * input: 9564 * htblp pointer to hash table 9565 * 9566 * hkey hkey that identifies the object to be removed 9567 * 9568 * output: 9569 * objpp pointer to pointer to object. 9570 * if remove is successful, the removed object 9571 * will be returned via *objpp. 9572 * 9573 * return value(s): 9574 * EINVAL cannot find hash entry 9575 * 0 successful 9576 */ 9577 static int 9578 daplka_hash_remove(daplka_hash_table_t *htblp, uint64_t hkey, void **objpp) 9579 { 9580 daplka_hash_entry_t *free_hep, **curr_hepp; 9581 daplka_hash_bucket_t *hbp; 9582 uint32_t bucket; 9583 9584 bucket = (uint32_t)(hkey & (htblp->ht_nbuckets - 1)); 9585 9586 rw_enter(&htblp->ht_table_lock, RW_WRITER); 9587 hbp = &htblp->ht_buckets[bucket]; 9588 9589 curr_hepp = &hbp->hb_entries; 9590 while (*curr_hepp != NULL) { 9591 if ((*curr_hepp)->he_hkey == hkey) { 9592 break; 9593 } 9594 curr_hepp = &(*curr_hepp)->he_next; 9595 } 9596 if (*curr_hepp == NULL) { 9597 DERR("hash_remove: cannot find hash entry: " 9598 "bucket %d, hkey 0x%016llx\n", bucket, (longlong_t)hkey); 9599 rw_exit(&htblp->ht_table_lock); 9600 return (EINVAL); 9601 } else { 9602 if (objpp != NULL) { 9603 *objpp = (*curr_hepp)->he_objp; 9604 } 9605 free_hep = *curr_hepp; 9606 *curr_hepp = (*curr_hepp)->he_next; 9607 kmem_free(free_hep, sizeof (*free_hep)); 9608 } 9609 hbp->hb_count--; 9610 htblp->ht_count--; 9611 D3("hash_remove: removed entry, hkey 0x%016llx, bucket %d, " 9612 "hb_count %d, hb_count %d\n", 9613 (longlong_t)hkey, bucket, hbp->hb_count, htblp->ht_count); 9614 rw_exit(&htblp->ht_table_lock); 9615 return (0); 9616 } 9617 9618 /* 9619 * daplka_hash_walk: 9620 * walks through the entire hash table. applying func on each of 9621 * the inserted objects. stops walking if func returns non-zero. 9622 * 9623 * input: 9624 * htblp pointer to hash table 9625 * 9626 * func function to be applied on each object 9627 * 9628 * farg second argument to func 9629 * 9630 * lockmode can be RW_WRITER or RW_READER. this 9631 * allows the caller to choose what type 9632 * of lock to acquire before walking the 9633 * table. 9634 * 9635 * output: 9636 * none 9637 * 9638 * return value(s): 9639 * none 9640 */ 9641 static void 9642 daplka_hash_walk(daplka_hash_table_t *htblp, int (*func)(void *, void *), 9643 void *farg, krw_t lockmode) 9644 { 9645 daplka_hash_entry_t *curr_hep; 9646 daplka_hash_bucket_t *hbp; 9647 uint32_t bucket, retval = 0; 9648 9649 ASSERT(lockmode == RW_WRITER || lockmode == RW_READER); 9650 9651 /* needed for warlock */ 9652 if (lockmode == RW_WRITER) { 9653 rw_enter(&htblp->ht_table_lock, RW_WRITER); 9654 } else { 9655 rw_enter(&htblp->ht_table_lock, RW_READER); 9656 } 9657 for (bucket = 0; bucket < htblp->ht_nbuckets && retval == 0; bucket++) { 9658 hbp = &htblp->ht_buckets[bucket]; 9659 curr_hep = hbp->hb_entries; 9660 while (curr_hep != NULL) { 9661 retval = (*func)(curr_hep->he_objp, farg); 9662 if (retval != 0) { 9663 break; 9664 } 9665 curr_hep = curr_hep->he_next; 9666 } 9667 } 9668 rw_exit(&htblp->ht_table_lock); 9669 } 9670 9671 /* 9672 * daplka_hash_lookup: 9673 * finds object from hkey 9674 * 9675 * input: 9676 * htblp pointer to hash table 9677 * 9678 * hkey hkey that identifies the object to be looked up 9679 * 9680 * output: 9681 * none 9682 * 9683 * return value(s): 9684 * NULL if not found 9685 * object pointer if found 9686 */ 9687 static void * 9688 daplka_hash_lookup(daplka_hash_table_t *htblp, uint64_t hkey) 9689 { 9690 daplka_hash_entry_t *curr_hep; 9691 uint32_t bucket; 9692 void *objp; 9693 9694 bucket = (uint32_t)(hkey & (htblp->ht_nbuckets - 1)); 9695 9696 rw_enter(&htblp->ht_table_lock, RW_READER); 9697 curr_hep = htblp->ht_buckets[bucket].hb_entries; 9698 while (curr_hep != NULL) { 9699 if (curr_hep->he_hkey == hkey) { 9700 break; 9701 } 9702 curr_hep = curr_hep->he_next; 9703 } 9704 if (curr_hep == NULL) { 9705 DERR("hash_lookup: cannot find hash entry: " 9706 "bucket %d, hkey 0x%016llx\n", bucket, (longlong_t)hkey); 9707 rw_exit(&htblp->ht_table_lock); 9708 return (NULL); 9709 } 9710 objp = curr_hep->he_objp; 9711 ASSERT(objp != NULL); 9712 if (htblp->ht_lookup_func != NULL) { 9713 (*htblp->ht_lookup_func)(objp); 9714 } 9715 rw_exit(&htblp->ht_table_lock); 9716 return (objp); 9717 } 9718 9719 /* 9720 * daplka_hash_destroy: 9721 * destroys hash table. applies free_func on all inserted objects. 9722 * 9723 * input: 9724 * htblp pointer to hash table 9725 * 9726 * output: 9727 * none 9728 * 9729 * return value(s): 9730 * none 9731 */ 9732 static void 9733 daplka_hash_destroy(daplka_hash_table_t *htblp) 9734 { 9735 daplka_hash_entry_t *curr_hep, *free_hep; 9736 daplka_hash_entry_t *free_list = NULL; 9737 daplka_hash_bucket_t *hbp; 9738 uint32_t bucket, cnt, total = 0; 9739 9740 if (!htblp->ht_initialized) { 9741 DERR("hash_destroy: not initialized\n"); 9742 return; 9743 } 9744 /* free all elements from hash table */ 9745 rw_enter(&htblp->ht_table_lock, RW_WRITER); 9746 for (bucket = 0; bucket < htblp->ht_nbuckets; bucket++) { 9747 hbp = &htblp->ht_buckets[bucket]; 9748 9749 /* build list of elements to be freed */ 9750 curr_hep = hbp->hb_entries; 9751 cnt = 0; 9752 while (curr_hep != NULL) { 9753 cnt++; 9754 free_hep = curr_hep; 9755 curr_hep = curr_hep->he_next; 9756 9757 free_hep->he_next = free_list; 9758 free_list = free_hep; 9759 } 9760 ASSERT(cnt == hbp->hb_count); 9761 total += cnt; 9762 hbp->hb_count = 0; 9763 hbp->hb_entries = NULL; 9764 } 9765 ASSERT(total == htblp->ht_count); 9766 D3("hash_destroy: htblp 0x%p, nbuckets %d, freed %d hash entries\n", 9767 htblp, htblp->ht_nbuckets, total); 9768 rw_exit(&htblp->ht_table_lock); 9769 9770 /* free all objects, now without holding the hash table lock */ 9771 cnt = 0; 9772 while (free_list != NULL) { 9773 cnt++; 9774 free_hep = free_list; 9775 free_list = free_list->he_next; 9776 if (htblp->ht_free_func != NULL) { 9777 (*htblp->ht_free_func)(free_hep->he_objp); 9778 } 9779 kmem_free(free_hep, sizeof (*free_hep)); 9780 } 9781 ASSERT(total == cnt); 9782 9783 /* free hash buckets and destroy locks */ 9784 kmem_free(htblp->ht_buckets, 9785 sizeof (daplka_hash_bucket_t) * htblp->ht_nbuckets); 9786 9787 rw_enter(&htblp->ht_table_lock, RW_WRITER); 9788 htblp->ht_buckets = NULL; 9789 htblp->ht_count = 0; 9790 htblp->ht_nbuckets = 0; 9791 htblp->ht_free_func = NULL; 9792 htblp->ht_lookup_func = NULL; 9793 htblp->ht_initialized = B_FALSE; 9794 rw_exit(&htblp->ht_table_lock); 9795 9796 mutex_destroy(&htblp->ht_key_lock); 9797 rw_destroy(&htblp->ht_table_lock); 9798 } 9799 9800 /* 9801 * daplka_hash_getsize: 9802 * return the number of objects in hash table 9803 * 9804 * input: 9805 * htblp pointer to hash table 9806 * 9807 * output: 9808 * none 9809 * 9810 * return value(s): 9811 * number of objects in hash table 9812 */ 9813 static uint32_t 9814 daplka_hash_getsize(daplka_hash_table_t *htblp) 9815 { 9816 uint32_t sz; 9817 9818 rw_enter(&htblp->ht_table_lock, RW_READER); 9819 sz = htblp->ht_count; 9820 rw_exit(&htblp->ht_table_lock); 9821 9822 return (sz); 9823 } 9824 9825 /* 9826 * this function is used as ht_lookup_func above when lookup is called. 9827 * other types of objs may use a more elaborate lookup_func. 9828 */ 9829 static void 9830 daplka_hash_generic_lookup(void *obj) 9831 { 9832 daplka_resource_t *rp = (daplka_resource_t *)obj; 9833 9834 mutex_enter(&rp->rs_reflock); 9835 rp->rs_refcnt++; 9836 ASSERT(rp->rs_refcnt != 0); 9837 mutex_exit(&rp->rs_reflock); 9838 } 9839 9840 /* 9841 * Generates a non-zero 32 bit hash key used for the timer hash table. 9842 */ 9843 static uint32_t 9844 daplka_timer_hkey_gen() 9845 { 9846 uint32_t new_hkey; 9847 9848 do { 9849 new_hkey = atomic_add_32_nv(&daplka_timer_hkey, 1); 9850 } while (new_hkey == 0); 9851 9852 return (new_hkey); 9853 } 9854 9855 9856 /* 9857 * The DAPL KA debug logging routines 9858 */ 9859 9860 /* 9861 * Add the string str to the end of the debug log, followed by a newline. 9862 */ 9863 static void 9864 daplka_dbglog(char *str) 9865 { 9866 size_t length; 9867 size_t remlen; 9868 9869 /* 9870 * If this is the first time we've written to the log, initialize it. 9871 */ 9872 if (!daplka_dbginit) { 9873 return; 9874 } 9875 mutex_enter(&daplka_dbglock); 9876 /* 9877 * Note the log is circular; if this string would run over the end, 9878 * we copy the first piece to the end and then the last piece to 9879 * the beginning of the log. 9880 */ 9881 length = strlen(str); 9882 9883 remlen = (size_t)sizeof (daplka_dbgbuf) - daplka_dbgnext - 1; 9884 9885 if (length > remlen) { 9886 if (remlen) 9887 bcopy(str, daplka_dbgbuf + daplka_dbgnext, remlen); 9888 daplka_dbgbuf[sizeof (daplka_dbgbuf) - 1] = (char)NULL; 9889 str += remlen; 9890 length -= remlen; 9891 daplka_dbgnext = 0; 9892 } 9893 bcopy(str, daplka_dbgbuf + daplka_dbgnext, length); 9894 daplka_dbgnext += length; 9895 9896 if (daplka_dbgnext >= sizeof (daplka_dbgbuf)) 9897 daplka_dbgnext = 0; 9898 mutex_exit(&daplka_dbglock); 9899 } 9900 9901 9902 /* 9903 * Add a printf-style message to whichever debug logs we're currently using. 9904 */ 9905 static void 9906 daplka_debug(const char *fmt, ...) 9907 { 9908 char buff[512]; 9909 va_list ap; 9910 /* 9911 * The system prepends the thread id and high resolution time 9912 * (nanoseconds are dropped and so are the upper digits) 9913 * to the specified string. 9914 * The unit for timestamp is 10 microseconds. 9915 * It wraps around every 10000 seconds. 9916 * Ex: gethrtime() = X ns = X/1000 us = X/10000 10 micro sec. 9917 */ 9918 int micro_time = (int)((gethrtime() / 10000) % 1000000000); 9919 (void) sprintf(buff, "th %p tm %9d: ", (void *)curthread, micro_time); 9920 9921 va_start(ap, fmt); 9922 (void) vsprintf(buff+strlen(buff), fmt, ap); 9923 va_end(ap); 9924 9925 daplka_dbglog(buff); 9926 } 9927 9928 static void 9929 daplka_console(const char *fmt, ...) 9930 { 9931 char buff[512]; 9932 va_list ap; 9933 9934 va_start(ap, fmt); 9935 (void) vsprintf(buff, fmt, ap); 9936 va_end(ap); 9937 9938 cmn_err(CE_CONT, "%s", buff); 9939 } 9940