1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 24 * Use is subject to license terms. 25 */ 26 27 #include <sys/conf.h> 28 #include <sys/file.h> 29 #include <sys/ddi.h> 30 #include <sys/sunddi.h> 31 #include <sys/cpuvar.h> 32 33 #include <sys/socket.h> 34 #include <sys/strsubr.h> 35 #include <sys/socketvar.h> 36 #include <sys/sysmacros.h> 37 38 #include <sys/idm/idm.h> 39 #include <sys/idm/idm_so.h> 40 41 extern idm_transport_t idm_transport_list[]; 42 43 void 44 idm_pdu_rx(idm_conn_t *ic, idm_pdu_t *pdu) 45 { 46 iscsi_async_evt_hdr_t *async_evt; 47 48 /* 49 * If we are in full-featured mode then route SCSI-related 50 * commands to the appropriate function vector 51 */ 52 ic->ic_timestamp = ddi_get_lbolt(); 53 mutex_enter(&ic->ic_state_mutex); 54 if (ic->ic_ffp && ic->ic_pdu_events == 0) { 55 mutex_exit(&ic->ic_state_mutex); 56 57 if (idm_pdu_rx_forward_ffp(ic, pdu) == B_TRUE) { 58 /* Forwarded SCSI-related commands */ 59 return; 60 } 61 mutex_enter(&ic->ic_state_mutex); 62 } 63 64 /* 65 * If we get here with a SCSI-related PDU then we are not in 66 * full-feature mode and the PDU is a protocol error (SCSI command 67 * PDU's may sometimes be an exception, see below). All 68 * non-SCSI PDU's get treated them the same regardless of whether 69 * we are in full-feature mode. 70 * 71 * Look at the opcode and in some cases the PDU status and 72 * determine the appropriate event to send to the connection 73 * state machine. Generate the event, passing the PDU as data. 74 * If the current connection state allows reception of the event 75 * the PDU will be submitted to the IDM client for processing, 76 * otherwise the PDU will be dropped. 77 */ 78 switch (IDM_PDU_OPCODE(pdu)) { 79 case ISCSI_OP_LOGIN_CMD: 80 idm_conn_rx_pdu_event(ic, CE_LOGIN_RCV, (uintptr_t)pdu); 81 break; 82 case ISCSI_OP_LOGIN_RSP: 83 idm_parse_login_rsp(ic, pdu, /* RX */ B_TRUE); 84 break; 85 case ISCSI_OP_LOGOUT_CMD: 86 idm_parse_logout_req(ic, pdu, /* RX */ B_TRUE); 87 break; 88 case ISCSI_OP_LOGOUT_RSP: 89 idm_parse_logout_rsp(ic, pdu, /* RX */ B_TRUE); 90 break; 91 case ISCSI_OP_ASYNC_EVENT: 92 async_evt = (iscsi_async_evt_hdr_t *)pdu->isp_hdr; 93 switch (async_evt->opcode) { 94 case ISCSI_ASYNC_EVENT_REQUEST_LOGOUT: 95 idm_conn_rx_pdu_event(ic, CE_ASYNC_LOGOUT_RCV, 96 (uintptr_t)pdu); 97 break; 98 case ISCSI_ASYNC_EVENT_DROPPING_CONNECTION: 99 idm_conn_rx_pdu_event(ic, CE_ASYNC_DROP_CONN_RCV, 100 (uintptr_t)pdu); 101 break; 102 case ISCSI_ASYNC_EVENT_DROPPING_ALL_CONNECTIONS: 103 idm_conn_rx_pdu_event(ic, CE_ASYNC_DROP_ALL_CONN_RCV, 104 (uintptr_t)pdu); 105 break; 106 case ISCSI_ASYNC_EVENT_SCSI_EVENT: 107 case ISCSI_ASYNC_EVENT_PARAM_NEGOTIATION: 108 default: 109 idm_conn_rx_pdu_event(ic, CE_MISC_RX, 110 (uintptr_t)pdu); 111 break; 112 } 113 break; 114 case ISCSI_OP_SCSI_CMD: 115 /* 116 * Consider this scenario: We are a target connection 117 * in "in login" state and a "login success sent" event has 118 * been generated but not yet handled. Since we've sent 119 * the login response but we haven't actually transitioned 120 * to FFP mode we might conceivably receive a SCSI command 121 * from the initiator before we are ready. We are actually 122 * in FFP we just don't know it yet -- to address this we 123 * can generate an event corresponding to the SCSI command. 124 * At the point when the event is handled by the state 125 * machine the login request will have been handled and we 126 * should be in FFP. If we are not in FFP by that time 127 * we can reject the SCSI command with a protocol error. 128 * 129 * This scenario only applies to the target. 130 */ 131 case ISCSI_OP_SCSI_DATA: 132 case ISCSI_OP_SCSI_DATA_RSP: 133 case ISCSI_OP_RTT_RSP: 134 case ISCSI_OP_SNACK_CMD: 135 case ISCSI_OP_NOOP_IN: 136 case ISCSI_OP_NOOP_OUT: 137 case ISCSI_OP_TEXT_CMD: 138 case ISCSI_OP_TEXT_RSP: 139 case ISCSI_OP_REJECT_MSG: 140 case ISCSI_OP_SCSI_TASK_MGT_MSG: 141 case ISCSI_OP_SCSI_TASK_MGT_RSP: 142 /* Validate received PDU against current state */ 143 idm_conn_rx_pdu_event(ic, CE_MISC_RX, 144 (uintptr_t)pdu); 145 break; 146 } 147 mutex_exit(&ic->ic_state_mutex); 148 } 149 150 void 151 idm_pdu_tx_forward(idm_conn_t *ic, idm_pdu_t *pdu) 152 { 153 (*ic->ic_transport_ops->it_tx_pdu)(ic, pdu); 154 } 155 156 boolean_t 157 idm_pdu_rx_forward_ffp(idm_conn_t *ic, idm_pdu_t *pdu) 158 { 159 /* 160 * If this is an FFP request, call the appropriate handler 161 * and return B_TRUE, otherwise return B_FALSE. 162 */ 163 switch (IDM_PDU_OPCODE(pdu)) { 164 case ISCSI_OP_SCSI_CMD: 165 (*ic->ic_conn_ops.icb_rx_scsi_cmd)(ic, pdu); 166 return (B_TRUE); 167 case ISCSI_OP_SCSI_RSP: 168 (*ic->ic_conn_ops.icb_rx_scsi_rsp)(ic, pdu); 169 return (B_TRUE); 170 case ISCSI_OP_SCSI_DATA: 171 (*ic->ic_transport_ops->it_rx_dataout)(ic, pdu); 172 return (B_TRUE); 173 case ISCSI_OP_SCSI_DATA_RSP: 174 (*ic->ic_transport_ops->it_rx_datain)(ic, pdu); 175 return (B_TRUE); 176 case ISCSI_OP_RTT_RSP: 177 (*ic->ic_transport_ops->it_rx_rtt)(ic, pdu); 178 return (B_TRUE); 179 case ISCSI_OP_SCSI_TASK_MGT_MSG: 180 case ISCSI_OP_SCSI_TASK_MGT_RSP: 181 case ISCSI_OP_TEXT_CMD: 182 case ISCSI_OP_TEXT_RSP: 183 case ISCSI_OP_NOOP_OUT: 184 case ISCSI_OP_NOOP_IN: 185 (*ic->ic_conn_ops.icb_rx_misc)(ic, pdu); 186 return (B_TRUE); 187 default: 188 return (B_FALSE); 189 } 190 /*NOTREACHED*/ 191 } 192 193 void 194 idm_pdu_rx_forward(idm_conn_t *ic, idm_pdu_t *pdu) 195 { 196 /* 197 * Some PDU's specific to FFP get special handling. This function 198 * will normally never be called in FFP with an FFP PDU since this 199 * is a slow path but in can happen on the target side during 200 * the transition to FFP. We primarily call 201 * idm_pdu_rx_forward_ffp here to avoid code duplication. 202 */ 203 if (idm_pdu_rx_forward_ffp(ic, pdu) == B_FALSE) { 204 /* 205 * Non-FFP PDU, use generic RC handler 206 */ 207 (*ic->ic_conn_ops.icb_rx_misc)(ic, pdu); 208 } 209 } 210 211 void 212 idm_parse_login_rsp(idm_conn_t *ic, idm_pdu_t *login_rsp_pdu, boolean_t rx) 213 { 214 iscsi_login_rsp_hdr_t *login_rsp = 215 (iscsi_login_rsp_hdr_t *)login_rsp_pdu->isp_hdr; 216 idm_conn_event_t new_event; 217 218 if (login_rsp->status_class == ISCSI_STATUS_CLASS_SUCCESS) { 219 if (!(login_rsp->flags & ISCSI_FLAG_LOGIN_CONTINUE) && 220 (login_rsp->flags & ISCSI_FLAG_LOGIN_TRANSIT) && 221 (ISCSI_LOGIN_NEXT_STAGE(login_rsp->flags) == 222 ISCSI_FULL_FEATURE_PHASE)) { 223 new_event = (rx ? CE_LOGIN_SUCCESS_RCV : 224 CE_LOGIN_SUCCESS_SND); 225 } else { 226 new_event = (rx ? CE_MISC_RX : CE_MISC_TX); 227 } 228 } else { 229 new_event = (rx ? CE_LOGIN_FAIL_RCV : CE_LOGIN_FAIL_SND); 230 } 231 232 if (rx) { 233 idm_conn_rx_pdu_event(ic, new_event, (uintptr_t)login_rsp_pdu); 234 } else { 235 idm_conn_tx_pdu_event(ic, new_event, (uintptr_t)login_rsp_pdu); 236 } 237 } 238 239 240 void 241 idm_parse_logout_req(idm_conn_t *ic, idm_pdu_t *logout_req_pdu, boolean_t rx) 242 { 243 iscsi_logout_hdr_t *logout_req = 244 (iscsi_logout_hdr_t *)logout_req_pdu->isp_hdr; 245 idm_conn_event_t new_event; 246 uint8_t reason = 247 (logout_req->flags & ISCSI_FLAG_LOGOUT_REASON_MASK); 248 249 /* 250 * For a normal logout (close connection or close session) IDM 251 * will terminate processing of all tasks completing the tasks 252 * back to the client with a status indicating the connection 253 * was logged out. These tasks do not get completed. 254 * 255 * For a "close connection for recovery logout) IDM suspends 256 * processing of all tasks and completes them back to the client 257 * with a status indicating connection was logged out for 258 * recovery. Both initiator and target hang onto these tasks. 259 * When we add ERL2 support IDM will need to provide mechanisms 260 * to change the task and buffer associations to a new connection. 261 * 262 * This code doesn't address the possibility of MC/S. We'll 263 * need to decide how the separate connections get handled 264 * in that case. One simple option is to make the client 265 * generate the events for the other connections. 266 */ 267 if (reason == ISCSI_LOGOUT_REASON_CLOSE_SESSION) { 268 new_event = 269 (rx ? CE_LOGOUT_SESSION_RCV : CE_LOGOUT_SESSION_SND); 270 } else if ((reason == ISCSI_LOGOUT_REASON_CLOSE_CONNECTION) || 271 (reason == ISCSI_LOGOUT_REASON_RECOVERY)) { 272 /* Check logout CID against this connection's CID */ 273 if (ntohs(logout_req->cid) == ic->ic_login_cid) { 274 /* Logout is for this connection */ 275 new_event = (rx ? CE_LOGOUT_THIS_CONN_RCV : 276 CE_LOGOUT_THIS_CONN_SND); 277 } else { 278 /* 279 * Logout affects another connection. This is not 280 * a relevant event for this connection so we'll 281 * just treat it as a normal PDU event. Client 282 * will need to lookup the other connection and 283 * generate the event. 284 */ 285 new_event = (rx ? CE_MISC_RX : CE_MISC_TX); 286 } 287 } else { 288 /* Invalid reason code */ 289 new_event = (rx ? CE_RX_PROTOCOL_ERROR : CE_TX_PROTOCOL_ERROR); 290 } 291 292 if (rx) { 293 idm_conn_rx_pdu_event(ic, new_event, (uintptr_t)logout_req_pdu); 294 } else { 295 idm_conn_tx_pdu_event(ic, new_event, (uintptr_t)logout_req_pdu); 296 } 297 } 298 299 300 301 void 302 idm_parse_logout_rsp(idm_conn_t *ic, idm_pdu_t *logout_rsp_pdu, boolean_t rx) 303 { 304 idm_conn_event_t new_event; 305 iscsi_logout_rsp_hdr_t *logout_rsp = 306 (iscsi_logout_rsp_hdr_t *)logout_rsp_pdu->isp_hdr; 307 308 if (logout_rsp->response == ISCSI_STATUS_CLASS_SUCCESS) { 309 new_event = rx ? CE_LOGOUT_SUCCESS_RCV : CE_LOGOUT_SUCCESS_SND; 310 } else { 311 new_event = rx ? CE_LOGOUT_FAIL_RCV : CE_LOGOUT_FAIL_SND; 312 } 313 314 if (rx) { 315 idm_conn_rx_pdu_event(ic, new_event, (uintptr_t)logout_rsp_pdu); 316 } else { 317 idm_conn_tx_pdu_event(ic, new_event, (uintptr_t)logout_rsp_pdu); 318 } 319 } 320 321 /* 322 * idm_svc_conn_create() 323 * Transport-agnostic service connection creation, invoked from the transport 324 * layer. 325 */ 326 idm_status_t 327 idm_svc_conn_create(idm_svc_t *is, idm_transport_type_t tt, 328 idm_conn_t **ic_result) 329 { 330 idm_conn_t *ic; 331 idm_status_t rc; 332 333 ic = idm_conn_create_common(CONN_TYPE_TGT, tt, 334 &is->is_svc_req.sr_conn_ops); 335 ic->ic_svc_binding = is; 336 337 /* 338 * Prepare connection state machine 339 */ 340 if ((rc = idm_conn_sm_init(ic)) != 0) { 341 idm_conn_destroy_common(ic); 342 return (rc); 343 } 344 345 346 *ic_result = ic; 347 348 mutex_enter(&idm.idm_global_mutex); 349 list_insert_tail(&idm.idm_tgt_conn_list, ic); 350 idm.idm_tgt_conn_count++; 351 mutex_exit(&idm.idm_global_mutex); 352 353 return (0); 354 } 355 356 void 357 idm_svc_conn_destroy(idm_conn_t *ic) 358 { 359 mutex_enter(&idm.idm_global_mutex); 360 list_remove(&idm.idm_tgt_conn_list, ic); 361 idm.idm_tgt_conn_count--; 362 mutex_exit(&idm.idm_global_mutex); 363 364 idm_conn_sm_fini(ic); 365 366 if (ic->ic_transport_private != NULL) { 367 ic->ic_transport_ops->it_tgt_conn_destroy(ic); 368 } 369 idm_conn_destroy_common(ic); 370 } 371 372 /* 373 * idm_conn_create_common() 374 * 375 * Allocate and initialize IDM connection context 376 */ 377 idm_conn_t * 378 idm_conn_create_common(idm_conn_type_t conn_type, idm_transport_type_t tt, 379 idm_conn_ops_t *conn_ops) 380 { 381 idm_conn_t *ic; 382 idm_transport_t *it; 383 idm_transport_type_t type; 384 385 for (type = 0; type < IDM_TRANSPORT_NUM_TYPES; type++) { 386 it = &idm_transport_list[type]; 387 388 if ((it->it_ops != NULL) && (it->it_type == tt)) 389 break; 390 } 391 ASSERT(it->it_type == tt); 392 if (it->it_type != tt) 393 return (NULL); 394 395 ic = kmem_zalloc(sizeof (idm_conn_t), KM_SLEEP); 396 397 /* Initialize data */ 398 ic->ic_conn_type = conn_type; 399 ic->ic_conn_ops = *conn_ops; 400 ic->ic_transport_ops = it->it_ops; 401 ic->ic_transport_type = tt; 402 ic->ic_transport_private = NULL; /* Set by transport service */ 403 ic->ic_internal_cid = idm_cid_alloc(); 404 if (ic->ic_internal_cid == 0) { 405 kmem_free(ic, sizeof (idm_conn_t)); 406 return (NULL); 407 } 408 mutex_init(&ic->ic_mutex, NULL, MUTEX_DEFAULT, NULL); 409 cv_init(&ic->ic_cv, NULL, CV_DEFAULT, NULL); 410 idm_refcnt_init(&ic->ic_refcnt, ic); 411 412 return (ic); 413 } 414 415 void 416 idm_conn_destroy_common(idm_conn_t *ic) 417 { 418 idm_refcnt_destroy(&ic->ic_refcnt); 419 cv_destroy(&ic->ic_cv); 420 mutex_destroy(&ic->ic_mutex); 421 idm_cid_free(ic->ic_internal_cid); 422 423 kmem_free(ic, sizeof (idm_conn_t)); 424 } 425 426 /* 427 * Invoked from the SM as a result of client's invocation of 428 * idm_ini_conn_connect() 429 */ 430 idm_status_t 431 idm_ini_conn_finish(idm_conn_t *ic) 432 { 433 /* invoke transport-specific connection */ 434 return (ic->ic_transport_ops->it_ini_conn_connect(ic)); 435 } 436 437 idm_status_t 438 idm_tgt_conn_finish(idm_conn_t *ic) 439 { 440 idm_status_t rc; 441 442 rc = idm_notify_client(ic, CN_CONNECT_ACCEPT, NULL); 443 if (rc != IDM_STATUS_SUCCESS) { 444 return (IDM_STATUS_REJECT); 445 } 446 447 /* Target client is ready to receive a login, start connection */ 448 return (ic->ic_transport_ops->it_tgt_conn_connect(ic)); 449 } 450 451 idm_transport_t * 452 idm_transport_lookup(idm_conn_req_t *cr) 453 { 454 idm_transport_type_t type; 455 idm_transport_t *it; 456 idm_transport_caps_t caps; 457 458 /* 459 * Make sure all available transports are setup. We call this now 460 * instead of at initialization time in case IB has become available 461 * since we started (hotplug, etc). 462 */ 463 idm_transport_setup(cr->cr_li); 464 465 /* Determine the transport for this connection */ 466 for (type = 0; type < IDM_TRANSPORT_NUM_TYPES; type++) { 467 it = &idm_transport_list[type]; 468 469 if (it->it_ops == NULL) { 470 /* transport is not registered */ 471 continue; 472 } 473 474 if (it->it_ops->it_conn_is_capable(cr, &caps)) { 475 return (it); 476 } 477 } 478 479 ASSERT(0); 480 return (NULL); /* Make gcc happy */ 481 } 482 483 void 484 idm_transport_setup(ldi_ident_t li) 485 { 486 idm_transport_type_t type; 487 idm_transport_t *it; 488 int rc; 489 490 for (type = 0; type < IDM_TRANSPORT_NUM_TYPES; type++) { 491 it = &idm_transport_list[type]; 492 /* 493 * We may want to store the LDI handle in the idm_svc_t 494 * and then allow multiple calls to ldi_open_by_name. This 495 * would enable the LDI code to track who has the device open 496 * which could be useful in the case where we have multiple 497 * services and perhaps also have initiator and target opening 498 * the transport simultaneously. For now we stick with the 499 * plan. 500 */ 501 if (it->it_ops == NULL) { 502 /* transport is not ready, try to initialize it */ 503 if (it->it_type == IDM_TRANSPORT_TYPE_SOCKETS) { 504 idm_so_init(it); 505 } else { 506 rc = ldi_open_by_name(it->it_device_path, 507 FREAD | FWRITE, kcred, &it->it_ldi_hdl, li); 508 /* 509 * If the open is successful we will have 510 * filled in the LDI handle in the transport 511 * table and we expect that the transport 512 * registered itself. 513 */ 514 if (rc != 0) { 515 it->it_ldi_hdl = NULL; 516 } 517 } 518 } 519 } 520 } 521 522 /* 523 * ID pool code. We use this to generate unique structure identifiers without 524 * searching the existing structures. This avoids the need to lock entire 525 * sets of structures at inopportune times. Adapted from the CIFS server code. 526 * 527 * A pool of IDs is a pool of 16 bit numbers. It is implemented as a bitmap. 528 * A bit set to '1' indicates that that particular value has been allocated. 529 * The allocation process is done shifting a bit through the whole bitmap. 530 * The current position of that index bit is kept in the idm_idpool_t 531 * structure and represented by a byte index (0 to buffer size minus 1) and 532 * a bit index (0 to 7). 533 * 534 * The pools start with a size of 8 bytes or 64 IDs. Each time the pool runs 535 * out of IDs its current size is doubled until it reaches its maximum size 536 * (8192 bytes or 65536 IDs). The IDs 0 and 65535 are never given out which 537 * means that a pool can have a maximum number of 65534 IDs available. 538 */ 539 540 static int 541 idm_idpool_increment( 542 idm_idpool_t *pool) 543 { 544 uint8_t *new_pool; 545 uint32_t new_size; 546 547 ASSERT(pool->id_magic == IDM_IDPOOL_MAGIC); 548 549 new_size = pool->id_size * 2; 550 if (new_size <= IDM_IDPOOL_MAX_SIZE) { 551 new_pool = kmem_alloc(new_size / 8, KM_NOSLEEP); 552 if (new_pool) { 553 bzero(new_pool, new_size / 8); 554 bcopy(pool->id_pool, new_pool, pool->id_size / 8); 555 kmem_free(pool->id_pool, pool->id_size / 8); 556 pool->id_pool = new_pool; 557 pool->id_free_counter += new_size - pool->id_size; 558 pool->id_max_free_counter += new_size - pool->id_size; 559 pool->id_size = new_size; 560 pool->id_idx_msk = (new_size / 8) - 1; 561 if (new_size >= IDM_IDPOOL_MAX_SIZE) { 562 /* id -1 made unavailable */ 563 pool->id_pool[pool->id_idx_msk] = 0x80; 564 pool->id_free_counter--; 565 pool->id_max_free_counter--; 566 } 567 return (0); 568 } 569 } 570 return (-1); 571 } 572 573 /* 574 * idm_idpool_constructor 575 * 576 * This function initializes the pool structure provided. 577 */ 578 579 int 580 idm_idpool_create(idm_idpool_t *pool) 581 { 582 583 ASSERT(pool->id_magic != IDM_IDPOOL_MAGIC); 584 585 pool->id_size = IDM_IDPOOL_MIN_SIZE; 586 pool->id_idx_msk = (IDM_IDPOOL_MIN_SIZE / 8) - 1; 587 pool->id_free_counter = IDM_IDPOOL_MIN_SIZE - 1; 588 pool->id_max_free_counter = IDM_IDPOOL_MIN_SIZE - 1; 589 pool->id_bit = 0x02; 590 pool->id_bit_idx = 1; 591 pool->id_idx = 0; 592 pool->id_pool = (uint8_t *)kmem_alloc((IDM_IDPOOL_MIN_SIZE / 8), 593 KM_SLEEP); 594 bzero(pool->id_pool, (IDM_IDPOOL_MIN_SIZE / 8)); 595 /* -1 id made unavailable */ 596 pool->id_pool[0] = 0x01; /* id 0 made unavailable */ 597 mutex_init(&pool->id_mutex, NULL, MUTEX_DEFAULT, NULL); 598 pool->id_magic = IDM_IDPOOL_MAGIC; 599 return (0); 600 } 601 602 /* 603 * idm_idpool_destructor 604 * 605 * This function tears down and frees the resources associated with the 606 * pool provided. 607 */ 608 609 void 610 idm_idpool_destroy(idm_idpool_t *pool) 611 { 612 ASSERT(pool->id_magic == IDM_IDPOOL_MAGIC); 613 ASSERT(pool->id_free_counter == pool->id_max_free_counter); 614 pool->id_magic = (uint32_t)~IDM_IDPOOL_MAGIC; 615 mutex_destroy(&pool->id_mutex); 616 kmem_free(pool->id_pool, (size_t)(pool->id_size / 8)); 617 } 618 619 /* 620 * idm_idpool_alloc 621 * 622 * This function allocates an ID from the pool provided. 623 */ 624 int 625 idm_idpool_alloc(idm_idpool_t *pool, uint16_t *id) 626 { 627 uint32_t i; 628 uint8_t bit; 629 uint8_t bit_idx; 630 uint8_t byte; 631 632 ASSERT(pool->id_magic == IDM_IDPOOL_MAGIC); 633 634 mutex_enter(&pool->id_mutex); 635 if ((pool->id_free_counter == 0) && idm_idpool_increment(pool)) { 636 mutex_exit(&pool->id_mutex); 637 return (-1); 638 } 639 640 i = pool->id_size; 641 while (i) { 642 bit = pool->id_bit; 643 bit_idx = pool->id_bit_idx; 644 byte = pool->id_pool[pool->id_idx]; 645 while (bit) { 646 if (byte & bit) { 647 bit = bit << 1; 648 bit_idx++; 649 continue; 650 } 651 pool->id_pool[pool->id_idx] |= bit; 652 *id = (uint16_t)(pool->id_idx * 8 + (uint32_t)bit_idx); 653 pool->id_free_counter--; 654 pool->id_bit = bit; 655 pool->id_bit_idx = bit_idx; 656 mutex_exit(&pool->id_mutex); 657 return (0); 658 } 659 pool->id_bit = 1; 660 pool->id_bit_idx = 0; 661 pool->id_idx++; 662 pool->id_idx &= pool->id_idx_msk; 663 --i; 664 } 665 /* 666 * This section of code shouldn't be reached. If there are IDs 667 * available and none could be found there's a problem. 668 */ 669 ASSERT(0); 670 mutex_exit(&pool->id_mutex); 671 return (-1); 672 } 673 674 /* 675 * idm_idpool_free 676 * 677 * This function frees the ID provided. 678 */ 679 void 680 idm_idpool_free(idm_idpool_t *pool, uint16_t id) 681 { 682 ASSERT(pool->id_magic == IDM_IDPOOL_MAGIC); 683 ASSERT(id != 0); 684 ASSERT(id != 0xFFFF); 685 686 mutex_enter(&pool->id_mutex); 687 if (pool->id_pool[id >> 3] & (1 << (id & 7))) { 688 pool->id_pool[id >> 3] &= ~(1 << (id & 7)); 689 pool->id_free_counter++; 690 ASSERT(pool->id_free_counter <= pool->id_max_free_counter); 691 mutex_exit(&pool->id_mutex); 692 return; 693 } 694 /* Freeing a free ID. */ 695 ASSERT(0); 696 mutex_exit(&pool->id_mutex); 697 } 698 699 uint32_t 700 idm_cid_alloc(void) 701 { 702 /* 703 * ID pool works with 16-bit identifiers right now. That should 704 * be plenty since we will probably never have more than 2^16 705 * connections simultaneously. 706 */ 707 uint16_t cid16; 708 709 if (idm_idpool_alloc(&idm.idm_conn_id_pool, &cid16) == -1) { 710 return (0); /* Fail */ 711 } 712 713 return ((uint32_t)cid16); 714 } 715 716 void 717 idm_cid_free(uint32_t cid) 718 { 719 idm_idpool_free(&idm.idm_conn_id_pool, (uint16_t)cid); 720 } 721 722 723 /* 724 * Code for generating the header and data digests 725 * 726 * This is the CRC-32C table 727 * Generated with: 728 * width = 32 bits 729 * poly = 0x1EDC6F41 730 * reflect input bytes = true 731 * reflect output bytes = true 732 */ 733 734 uint32_t idm_crc32c_table[256] = 735 { 736 0x00000000, 0xF26B8303, 0xE13B70F7, 0x1350F3F4, 737 0xC79A971F, 0x35F1141C, 0x26A1E7E8, 0xD4CA64EB, 738 0x8AD958CF, 0x78B2DBCC, 0x6BE22838, 0x9989AB3B, 739 0x4D43CFD0, 0xBF284CD3, 0xAC78BF27, 0x5E133C24, 740 0x105EC76F, 0xE235446C, 0xF165B798, 0x030E349B, 741 0xD7C45070, 0x25AFD373, 0x36FF2087, 0xC494A384, 742 0x9A879FA0, 0x68EC1CA3, 0x7BBCEF57, 0x89D76C54, 743 0x5D1D08BF, 0xAF768BBC, 0xBC267848, 0x4E4DFB4B, 744 0x20BD8EDE, 0xD2D60DDD, 0xC186FE29, 0x33ED7D2A, 745 0xE72719C1, 0x154C9AC2, 0x061C6936, 0xF477EA35, 746 0xAA64D611, 0x580F5512, 0x4B5FA6E6, 0xB93425E5, 747 0x6DFE410E, 0x9F95C20D, 0x8CC531F9, 0x7EAEB2FA, 748 0x30E349B1, 0xC288CAB2, 0xD1D83946, 0x23B3BA45, 749 0xF779DEAE, 0x05125DAD, 0x1642AE59, 0xE4292D5A, 750 0xBA3A117E, 0x4851927D, 0x5B016189, 0xA96AE28A, 751 0x7DA08661, 0x8FCB0562, 0x9C9BF696, 0x6EF07595, 752 0x417B1DBC, 0xB3109EBF, 0xA0406D4B, 0x522BEE48, 753 0x86E18AA3, 0x748A09A0, 0x67DAFA54, 0x95B17957, 754 0xCBA24573, 0x39C9C670, 0x2A993584, 0xD8F2B687, 755 0x0C38D26C, 0xFE53516F, 0xED03A29B, 0x1F682198, 756 0x5125DAD3, 0xA34E59D0, 0xB01EAA24, 0x42752927, 757 0x96BF4DCC, 0x64D4CECF, 0x77843D3B, 0x85EFBE38, 758 0xDBFC821C, 0x2997011F, 0x3AC7F2EB, 0xC8AC71E8, 759 0x1C661503, 0xEE0D9600, 0xFD5D65F4, 0x0F36E6F7, 760 0x61C69362, 0x93AD1061, 0x80FDE395, 0x72966096, 761 0xA65C047D, 0x5437877E, 0x4767748A, 0xB50CF789, 762 0xEB1FCBAD, 0x197448AE, 0x0A24BB5A, 0xF84F3859, 763 0x2C855CB2, 0xDEEEDFB1, 0xCDBE2C45, 0x3FD5AF46, 764 0x7198540D, 0x83F3D70E, 0x90A324FA, 0x62C8A7F9, 765 0xB602C312, 0x44694011, 0x5739B3E5, 0xA55230E6, 766 0xFB410CC2, 0x092A8FC1, 0x1A7A7C35, 0xE811FF36, 767 0x3CDB9BDD, 0xCEB018DE, 0xDDE0EB2A, 0x2F8B6829, 768 0x82F63B78, 0x709DB87B, 0x63CD4B8F, 0x91A6C88C, 769 0x456CAC67, 0xB7072F64, 0xA457DC90, 0x563C5F93, 770 0x082F63B7, 0xFA44E0B4, 0xE9141340, 0x1B7F9043, 771 0xCFB5F4A8, 0x3DDE77AB, 0x2E8E845F, 0xDCE5075C, 772 0x92A8FC17, 0x60C37F14, 0x73938CE0, 0x81F80FE3, 773 0x55326B08, 0xA759E80B, 0xB4091BFF, 0x466298FC, 774 0x1871A4D8, 0xEA1A27DB, 0xF94AD42F, 0x0B21572C, 775 0xDFEB33C7, 0x2D80B0C4, 0x3ED04330, 0xCCBBC033, 776 0xA24BB5A6, 0x502036A5, 0x4370C551, 0xB11B4652, 777 0x65D122B9, 0x97BAA1BA, 0x84EA524E, 0x7681D14D, 778 0x2892ED69, 0xDAF96E6A, 0xC9A99D9E, 0x3BC21E9D, 779 0xEF087A76, 0x1D63F975, 0x0E330A81, 0xFC588982, 780 0xB21572C9, 0x407EF1CA, 0x532E023E, 0xA145813D, 781 0x758FE5D6, 0x87E466D5, 0x94B49521, 0x66DF1622, 782 0x38CC2A06, 0xCAA7A905, 0xD9F75AF1, 0x2B9CD9F2, 783 0xFF56BD19, 0x0D3D3E1A, 0x1E6DCDEE, 0xEC064EED, 784 0xC38D26C4, 0x31E6A5C7, 0x22B65633, 0xD0DDD530, 785 0x0417B1DB, 0xF67C32D8, 0xE52CC12C, 0x1747422F, 786 0x49547E0B, 0xBB3FFD08, 0xA86F0EFC, 0x5A048DFF, 787 0x8ECEE914, 0x7CA56A17, 0x6FF599E3, 0x9D9E1AE0, 788 0xD3D3E1AB, 0x21B862A8, 0x32E8915C, 0xC083125F, 789 0x144976B4, 0xE622F5B7, 0xF5720643, 0x07198540, 790 0x590AB964, 0xAB613A67, 0xB831C993, 0x4A5A4A90, 791 0x9E902E7B, 0x6CFBAD78, 0x7FAB5E8C, 0x8DC0DD8F, 792 0xE330A81A, 0x115B2B19, 0x020BD8ED, 0xF0605BEE, 793 0x24AA3F05, 0xD6C1BC06, 0xC5914FF2, 0x37FACCF1, 794 0x69E9F0D5, 0x9B8273D6, 0x88D28022, 0x7AB90321, 795 0xAE7367CA, 0x5C18E4C9, 0x4F48173D, 0xBD23943E, 796 0xF36E6F75, 0x0105EC76, 0x12551F82, 0xE03E9C81, 797 0x34F4F86A, 0xC69F7B69, 0xD5CF889D, 0x27A40B9E, 798 0x79B737BA, 0x8BDCB4B9, 0x988C474D, 0x6AE7C44E, 799 0xBE2DA0A5, 0x4C4623A6, 0x5F16D052, 0xAD7D5351 800 }; 801 802 /* 803 * iscsi_crc32c - Steps through buffer one byte at at time, calculates 804 * reflected crc using table. 805 */ 806 uint32_t 807 idm_crc32c(void *address, unsigned long length) 808 { 809 uint8_t *buffer = address; 810 uint32_t crc = 0xffffffff, result; 811 #ifdef _BIG_ENDIAN 812 uint8_t byte0, byte1, byte2, byte3; 813 #endif 814 815 ASSERT(address != NULL); 816 817 while (length--) { 818 crc = idm_crc32c_table[(crc ^ *buffer++) & 0xFFL] ^ 819 (crc >> 8); 820 } 821 result = crc ^ 0xffffffff; 822 823 #ifdef _BIG_ENDIAN 824 byte0 = (uint8_t)(result & 0xFF); 825 byte1 = (uint8_t)((result >> 8) & 0xFF); 826 byte2 = (uint8_t)((result >> 16) & 0xFF); 827 byte3 = (uint8_t)((result >> 24) & 0xFF); 828 result = ((byte0 << 24) | (byte1 << 16) | (byte2 << 8) | byte3); 829 #endif /* _BIG_ENDIAN */ 830 831 return (result); 832 } 833 834 835 /* 836 * idm_crc32c_continued - Continues stepping through buffer one 837 * byte at at time, calculates reflected crc using table. 838 */ 839 uint32_t 840 idm_crc32c_continued(void *address, unsigned long length, uint32_t crc) 841 { 842 uint8_t *buffer = address; 843 uint32_t result; 844 #ifdef _BIG_ENDIAN 845 uint8_t byte0, byte1, byte2, byte3; 846 #endif 847 848 ASSERT(address != NULL); 849 850 #ifdef _BIG_ENDIAN 851 byte0 = (uint8_t)((crc >> 24) & 0xFF); 852 byte1 = (uint8_t)((crc >> 16) & 0xFF); 853 byte2 = (uint8_t)((crc >> 8) & 0xFF); 854 byte3 = (uint8_t)(crc & 0xFF); 855 crc = ((byte3 << 24) | (byte2 << 16) | (byte1 << 8) | byte0); 856 #endif 857 858 crc = crc ^ 0xffffffff; 859 while (length--) { 860 crc = idm_crc32c_table[(crc ^ *buffer++) & 0xFFL] ^ 861 (crc >> 8); 862 } 863 result = crc ^ 0xffffffff; 864 865 #ifdef _BIG_ENDIAN 866 byte0 = (uint8_t)(result & 0xFF); 867 byte1 = (uint8_t)((result >> 8) & 0xFF); 868 byte2 = (uint8_t)((result >> 16) & 0xFF); 869 byte3 = (uint8_t)((result >> 24) & 0xFF); 870 result = ((byte0 << 24) | (byte1 << 16) | (byte2 << 8) | byte3); 871 #endif 872 return (result); 873 } 874 875 /* ARGSUSED */ 876 int 877 idm_task_constructor(void *hdl, void *arg, int flags) 878 { 879 idm_task_t *idt = (idm_task_t *)hdl; 880 uint32_t next_task; 881 882 mutex_init(&idt->idt_mutex, NULL, MUTEX_DEFAULT, NULL); 883 884 /* Find the next free task ID */ 885 rw_enter(&idm.idm_taskid_table_lock, RW_WRITER); 886 next_task = idm.idm_taskid_next; 887 while (idm.idm_taskid_table[next_task]) { 888 next_task++; 889 if (next_task == idm.idm_taskid_max) 890 next_task = 0; 891 if (next_task == idm.idm_taskid_next) { 892 rw_exit(&idm.idm_taskid_table_lock); 893 return (-1); 894 } 895 } 896 897 idm.idm_taskid_table[next_task] = idt; 898 idm.idm_taskid_next = (next_task + 1) % idm.idm_taskid_max; 899 rw_exit(&idm.idm_taskid_table_lock); 900 901 idt->idt_tt = next_task; 902 903 list_create(&idt->idt_inbufv, sizeof (idm_buf_t), 904 offsetof(idm_buf_t, idb_buflink)); 905 list_create(&idt->idt_outbufv, sizeof (idm_buf_t), 906 offsetof(idm_buf_t, idb_buflink)); 907 idm_refcnt_init(&idt->idt_refcnt, idt); 908 909 /* 910 * Set the transport header pointer explicitly. This removes the 911 * need for per-transport header allocation, which simplifies cache 912 * init considerably. If at a later date we have an additional IDM 913 * transport that requires a different size, we'll revisit this. 914 */ 915 idt->idt_transport_hdr = (void *)(idt + 1); /* pointer arithmetic */ 916 917 return (0); 918 } 919 920 /* ARGSUSED */ 921 void 922 idm_task_destructor(void *hdl, void *arg) 923 { 924 idm_task_t *idt = (idm_task_t *)hdl; 925 926 /* Remove the task from the ID table */ 927 rw_enter(&idm.idm_taskid_table_lock, RW_WRITER); 928 idm.idm_taskid_table[idt->idt_tt] = NULL; 929 rw_exit(&idm.idm_taskid_table_lock); 930 931 /* free the inbuf and outbuf */ 932 idm_refcnt_destroy(&idt->idt_refcnt); 933 list_destroy(&idt->idt_inbufv); 934 list_destroy(&idt->idt_outbufv); 935 936 mutex_destroy(&idt->idt_mutex); 937 } 938 939 /* 940 * idm_listbuf_insert searches from the back of the list looking for the 941 * insertion point. 942 */ 943 void 944 idm_listbuf_insert(list_t *lst, idm_buf_t *buf) 945 { 946 idm_buf_t *idb; 947 948 /* iterate through the list to find the insertion point */ 949 for (idb = list_tail(lst); idb != NULL; idb = list_prev(lst, idb)) { 950 951 if (idb->idb_bufoffset < buf->idb_bufoffset) { 952 953 list_insert_after(lst, idb, buf); 954 return; 955 } 956 } 957 958 /* add the buf to the head of the list */ 959 list_insert_head(lst, buf); 960 961 } 962 963 /*ARGSUSED*/ 964 void 965 idm_wd_thread(void *arg) 966 { 967 idm_conn_t *ic; 968 clock_t wake_time; 969 clock_t idle_time; 970 971 mutex_enter(&idm.idm_global_mutex); 972 idm.idm_wd_thread_running = B_TRUE; 973 idm.idm_wd_thread_did = idm.idm_wd_thread->t_did; 974 975 cv_signal(&idm.idm_wd_cv); 976 977 while (idm.idm_wd_thread_running) { 978 for (ic = list_head(&idm.idm_tgt_conn_list); 979 ic != NULL; 980 ic = list_next(&idm.idm_tgt_conn_list, ic)) { 981 idle_time = ddi_get_lbolt() - ic->ic_timestamp; 982 983 /* 984 * If there hasn't been any activity on this 985 * connection for the specified period then 986 * drop the connection. We expect the initiator 987 * to keep the connection alive if it wants the 988 * connection to stay open. 989 * 990 * If it turns out to be desireable to take a 991 * more active role in maintaining the connect 992 * we could add a client callback to send 993 * a "keepalive" kind of message (no doubt a nop) 994 * and fire that on a shorter timer. 995 */ 996 if (TICK_TO_SEC(idle_time) > 997 IDM_TRANSPORT_FAIL_IDLE_TIMEOUT) { 998 /* 999 * Only send the transport fail if we're in 1000 * FFP. State machine timers should handle 1001 * problems in non-ffp states. 1002 */ 1003 if (ic->ic_ffp) { 1004 mutex_exit(&idm.idm_global_mutex); 1005 IDM_SM_LOG(CE_WARN, "idm_wd_thread: " 1006 "conn %p idle for %d seconds, " 1007 "sending CE_TRANSPORT_FAIL", 1008 (void *)ic, (int)idle_time); 1009 idm_conn_event(ic, CE_TRANSPORT_FAIL, 1010 NULL); 1011 mutex_enter(&idm.idm_global_mutex); 1012 } 1013 } 1014 } 1015 1016 wake_time = lbolt + SEC_TO_TICK(IDM_WD_INTERVAL); 1017 (void) cv_timedwait(&idm.idm_wd_cv, &idm.idm_global_mutex, 1018 wake_time); 1019 } 1020 mutex_exit(&idm.idm_global_mutex); 1021 1022 thread_exit(); 1023 } 1024