1 /* $FreeBSD$ */ 2 /*- 3 * Copyright (c) 2015, Mellanox Technologies, Inc. All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 24 * SUCH DAMAGE. 25 */ 26 27 #include "icl_iser.h" 28 29 static MALLOC_DEFINE(M_ISER_INITIATOR, "iser_initiator", "iser initiator backend"); 30 31 /* Register user buffer memory and initialize passive rdma 32 * dto descriptor. Data size is stored in 33 * task->data[ISER_DIR_IN].data_len, Protection size 34 * os stored in task->prot[ISER_DIR_IN].data_len 35 */ 36 static int 37 iser_prepare_read_cmd(struct icl_iser_pdu *iser_pdu) 38 { 39 struct iser_hdr *hdr = &iser_pdu->desc.iser_header; 40 struct iser_data_buf *buf_in = &iser_pdu->data[ISER_DIR_IN]; 41 struct iser_mem_reg *mem_reg; 42 int err; 43 44 err = iser_dma_map_task_data(iser_pdu, 45 buf_in, 46 ISER_DIR_IN, 47 DMA_FROM_DEVICE); 48 if (err) 49 return (err); 50 51 err = iser_reg_rdma_mem(iser_pdu, ISER_DIR_IN); 52 if (err) { 53 ISER_ERR("Failed to set up Data-IN RDMA"); 54 return (err); 55 } 56 57 mem_reg = &iser_pdu->rdma_reg[ISER_DIR_IN]; 58 59 hdr->flags |= ISER_RSV; 60 hdr->read_stag = cpu_to_be32(mem_reg->rkey); 61 hdr->read_va = cpu_to_be64(mem_reg->sge.addr); 62 63 return (0); 64 } 65 66 /* Register user buffer memory and initialize passive rdma 67 * dto descriptor. Data size is stored in 68 * task->data[ISER_DIR_OUT].data_len, Protection size 69 * is stored at task->prot[ISER_DIR_OUT].data_len 70 */ 71 static int 72 iser_prepare_write_cmd(struct icl_iser_pdu *iser_pdu) 73 { 74 struct iser_hdr *hdr = &iser_pdu->desc.iser_header; 75 struct iser_data_buf *buf_out = &iser_pdu->data[ISER_DIR_OUT]; 76 struct iser_mem_reg *mem_reg; 77 int err; 78 79 err = iser_dma_map_task_data(iser_pdu, 80 buf_out, 81 ISER_DIR_OUT, 82 DMA_TO_DEVICE); 83 if (err) 84 return (err); 85 86 err = iser_reg_rdma_mem(iser_pdu, ISER_DIR_OUT); 87 if (err) { 88 ISER_ERR("Failed to set up Data-out RDMA"); 89 return (err); 90 } 91 92 mem_reg = &iser_pdu->rdma_reg[ISER_DIR_OUT]; 93 94 hdr->flags |= ISER_WSV; 95 hdr->write_stag = cpu_to_be32(mem_reg->rkey); 96 hdr->write_va = cpu_to_be64(mem_reg->sge.addr); 97 98 return (0); 99 } 100 101 /* creates a new tx descriptor and adds header regd buffer */ 102 void 103 iser_create_send_desc(struct iser_conn *iser_conn, 104 struct iser_tx_desc *tx_desc) 105 { 106 struct iser_device *device = iser_conn->ib_conn.device; 107 108 ib_dma_sync_single_for_cpu(device->ib_device, 109 tx_desc->dma_addr, ISER_HEADERS_LEN, DMA_TO_DEVICE); 110 111 memset(&tx_desc->iser_header, 0, sizeof(struct iser_hdr)); 112 tx_desc->iser_header.flags = ISER_VER; 113 114 tx_desc->num_sge = 1; 115 116 if (tx_desc->tx_sg[0].lkey != device->mr->lkey) { 117 tx_desc->tx_sg[0].lkey = device->mr->lkey; 118 ISER_DBG("sdesc %p lkey mismatch, fixing", tx_desc); 119 } 120 } 121 122 void 123 iser_free_login_buf(struct iser_conn *iser_conn) 124 { 125 struct iser_device *device = iser_conn->ib_conn.device; 126 127 if (!iser_conn->login_buf) 128 return; 129 130 if (iser_conn->login_req_dma) 131 ib_dma_unmap_single(device->ib_device, 132 iser_conn->login_req_dma, 133 ISCSI_DEF_MAX_RECV_SEG_LEN, DMA_TO_DEVICE); 134 135 if (iser_conn->login_resp_dma) 136 ib_dma_unmap_single(device->ib_device, 137 iser_conn->login_resp_dma, 138 ISER_RX_LOGIN_SIZE, DMA_FROM_DEVICE); 139 140 free(iser_conn->login_buf, M_ISER_INITIATOR); 141 142 /* make sure we never redo any unmapping */ 143 iser_conn->login_req_dma = 0; 144 iser_conn->login_resp_dma = 0; 145 iser_conn->login_buf = NULL; 146 } 147 148 int 149 iser_alloc_login_buf(struct iser_conn *iser_conn) 150 { 151 struct iser_device *device = iser_conn->ib_conn.device; 152 int req_err, resp_err; 153 154 BUG_ON(device == NULL); 155 156 iser_conn->login_buf = malloc(ISCSI_DEF_MAX_RECV_SEG_LEN + ISER_RX_LOGIN_SIZE, 157 M_ISER_INITIATOR, M_WAITOK | M_ZERO); 158 159 if (!iser_conn->login_buf) 160 goto out_err; 161 162 iser_conn->login_req_buf = iser_conn->login_buf; 163 iser_conn->login_resp_buf = iser_conn->login_buf + 164 ISCSI_DEF_MAX_RECV_SEG_LEN; 165 166 iser_conn->login_req_dma = ib_dma_map_single(device->ib_device, 167 iser_conn->login_req_buf, 168 ISCSI_DEF_MAX_RECV_SEG_LEN, 169 DMA_TO_DEVICE); 170 171 iser_conn->login_resp_dma = ib_dma_map_single(device->ib_device, 172 iser_conn->login_resp_buf, 173 ISER_RX_LOGIN_SIZE, 174 DMA_FROM_DEVICE); 175 176 req_err = ib_dma_mapping_error(device->ib_device, 177 iser_conn->login_req_dma); 178 resp_err = ib_dma_mapping_error(device->ib_device, 179 iser_conn->login_resp_dma); 180 181 if (req_err || resp_err) { 182 if (req_err) 183 iser_conn->login_req_dma = 0; 184 if (resp_err) 185 iser_conn->login_resp_dma = 0; 186 goto free_login_buf; 187 } 188 189 return (0); 190 191 free_login_buf: 192 iser_free_login_buf(iser_conn); 193 194 out_err: 195 ISER_DBG("unable to alloc or map login buf"); 196 return (ENOMEM); 197 } 198 199 int iser_alloc_rx_descriptors(struct iser_conn *iser_conn, int cmds_max) 200 { 201 int i, j; 202 u64 dma_addr; 203 struct iser_rx_desc *rx_desc; 204 struct ib_sge *rx_sg; 205 struct ib_conn *ib_conn = &iser_conn->ib_conn; 206 struct iser_device *device = ib_conn->device; 207 208 iser_conn->qp_max_recv_dtos = cmds_max; 209 iser_conn->min_posted_rx = iser_conn->qp_max_recv_dtos >> 2; 210 211 if (iser_create_fastreg_pool(ib_conn, cmds_max)) 212 goto create_rdma_reg_res_failed; 213 214 215 iser_conn->num_rx_descs = cmds_max; 216 iser_conn->rx_descs = malloc(iser_conn->num_rx_descs * 217 sizeof(struct iser_rx_desc), M_ISER_INITIATOR, 218 M_WAITOK | M_ZERO); 219 if (!iser_conn->rx_descs) 220 goto rx_desc_alloc_fail; 221 222 rx_desc = iser_conn->rx_descs; 223 224 for (i = 0; i < iser_conn->qp_max_recv_dtos; i++, rx_desc++) { 225 dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc, 226 ISER_RX_PAYLOAD_SIZE, DMA_FROM_DEVICE); 227 if (ib_dma_mapping_error(device->ib_device, dma_addr)) 228 goto rx_desc_dma_map_failed; 229 230 rx_desc->dma_addr = dma_addr; 231 232 rx_sg = &rx_desc->rx_sg; 233 rx_sg->addr = rx_desc->dma_addr; 234 rx_sg->length = ISER_RX_PAYLOAD_SIZE; 235 rx_sg->lkey = device->mr->lkey; 236 } 237 238 iser_conn->rx_desc_head = 0; 239 240 return (0); 241 242 rx_desc_dma_map_failed: 243 rx_desc = iser_conn->rx_descs; 244 for (j = 0; j < i; j++, rx_desc++) 245 ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr, 246 ISER_RX_PAYLOAD_SIZE, DMA_FROM_DEVICE); 247 free(iser_conn->rx_descs, M_ISER_INITIATOR); 248 iser_conn->rx_descs = NULL; 249 rx_desc_alloc_fail: 250 iser_free_fastreg_pool(ib_conn); 251 create_rdma_reg_res_failed: 252 ISER_ERR("failed allocating rx descriptors / data buffers"); 253 254 return (ENOMEM); 255 } 256 257 void 258 iser_free_rx_descriptors(struct iser_conn *iser_conn) 259 { 260 int i; 261 struct iser_rx_desc *rx_desc; 262 struct ib_conn *ib_conn = &iser_conn->ib_conn; 263 struct iser_device *device = ib_conn->device; 264 265 iser_free_fastreg_pool(ib_conn); 266 267 rx_desc = iser_conn->rx_descs; 268 for (i = 0; i < iser_conn->qp_max_recv_dtos; i++, rx_desc++) 269 ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr, 270 ISER_RX_PAYLOAD_SIZE, DMA_FROM_DEVICE); 271 272 free(iser_conn->rx_descs, M_ISER_INITIATOR); 273 274 /* make sure we never redo any unmapping */ 275 iser_conn->rx_descs = NULL; 276 } 277 278 static void 279 iser_buf_to_sg(void *buf, struct iser_data_buf *data_buf) 280 { 281 struct scatterlist *sg; 282 int i; 283 size_t len, tlen; 284 int offset; 285 286 tlen = data_buf->data_len; 287 288 for (i = 0; 0 < tlen; i++, tlen -= len) { 289 sg = &data_buf->sgl[i]; 290 offset = ((uintptr_t)buf) & ~PAGE_MASK; 291 len = min(PAGE_SIZE - offset, tlen); 292 sg_set_buf(sg, buf, len); 293 buf = (void *)((uintptr_t)buf + len); 294 } 295 296 data_buf->size = i; 297 sg_mark_end(sg); 298 } 299 300 301 static void 302 iser_bio_to_sg(struct bio *bp, struct iser_data_buf *data_buf) 303 { 304 struct scatterlist *sg; 305 int i; 306 size_t len, tlen; 307 int offset; 308 309 tlen = bp->bio_bcount; 310 offset = bp->bio_ma_offset; 311 312 for (i = 0; 0 < tlen; i++, tlen -= len) { 313 sg = &data_buf->sgl[i]; 314 len = min(PAGE_SIZE - offset, tlen); 315 sg_set_page(sg, bp->bio_ma[i], len, offset); 316 offset = 0; 317 } 318 319 data_buf->size = i; 320 sg_mark_end(sg); 321 } 322 323 static int 324 iser_csio_to_sg(struct ccb_scsiio *csio, struct iser_data_buf *data_buf) 325 { 326 struct ccb_hdr *ccbh; 327 int err = 0; 328 329 ccbh = &csio->ccb_h; 330 switch ((ccbh->flags & CAM_DATA_MASK)) { 331 case CAM_DATA_BIO: 332 iser_bio_to_sg((struct bio *) csio->data_ptr, data_buf); 333 break; 334 case CAM_DATA_VADDR: 335 /* 336 * Support KVA buffers for various scsi commands such as: 337 * - REPORT_LUNS 338 * - MODE_SENSE_6 339 * - INQUIRY 340 * - SERVICE_ACTION_IN. 341 * The data of these commands always mapped into KVA. 342 */ 343 iser_buf_to_sg(csio->data_ptr, data_buf); 344 break; 345 default: 346 ISER_ERR("flags 0x%X unimplemented", ccbh->flags); 347 err = EINVAL; 348 } 349 return (err); 350 } 351 352 static inline bool 353 iser_signal_comp(u8 sig_count) 354 { 355 return ((sig_count % ISER_SIGNAL_CMD_COUNT) == 0); 356 } 357 358 int 359 iser_send_command(struct iser_conn *iser_conn, 360 struct icl_iser_pdu *iser_pdu) 361 { 362 struct iser_data_buf *data_buf; 363 struct iser_tx_desc *tx_desc = &iser_pdu->desc; 364 struct iscsi_bhs_scsi_command *hdr = (struct iscsi_bhs_scsi_command *) &(iser_pdu->desc.iscsi_header); 365 struct ccb_scsiio *csio = iser_pdu->csio; 366 int err = 0; 367 u8 sig_count = ++iser_conn->ib_conn.sig_count; 368 369 /* build the tx desc regd header and add it to the tx desc dto */ 370 tx_desc->type = ISCSI_TX_SCSI_COMMAND; 371 iser_create_send_desc(iser_conn, tx_desc); 372 373 if (hdr->bhssc_flags & BHSSC_FLAGS_R) { 374 data_buf = &iser_pdu->data[ISER_DIR_IN]; 375 } else { 376 data_buf = &iser_pdu->data[ISER_DIR_OUT]; 377 } 378 379 data_buf->sg = csio->data_ptr; 380 data_buf->data_len = csio->dxfer_len; 381 382 if (likely(csio->dxfer_len)) { 383 err = iser_csio_to_sg(csio, data_buf); 384 if (unlikely(err)) 385 goto send_command_error; 386 } 387 388 if (hdr->bhssc_flags & BHSSC_FLAGS_R) { 389 err = iser_prepare_read_cmd(iser_pdu); 390 if (err) 391 goto send_command_error; 392 } else if (hdr->bhssc_flags & BHSSC_FLAGS_W) { 393 err = iser_prepare_write_cmd(iser_pdu); 394 if (err) 395 goto send_command_error; 396 } 397 398 err = iser_post_send(&iser_conn->ib_conn, tx_desc, 399 iser_signal_comp(sig_count)); 400 if (!err) 401 return (0); 402 403 send_command_error: 404 ISER_ERR("iser_conn %p itt %u len %u err %d", iser_conn, 405 hdr->bhssc_initiator_task_tag, 406 hdr->bhssc_expected_data_transfer_length, 407 err); 408 return (err); 409 } 410 411 int 412 iser_send_control(struct iser_conn *iser_conn, 413 struct icl_iser_pdu *iser_pdu) 414 { 415 struct iser_tx_desc *mdesc; 416 struct iser_device *device; 417 size_t datalen = iser_pdu->icl_pdu.ip_data_len; 418 int err; 419 420 mdesc = &iser_pdu->desc; 421 422 /* build the tx desc regd header and add it to the tx desc dto */ 423 mdesc->type = ISCSI_TX_CONTROL; 424 iser_create_send_desc(iser_conn, mdesc); 425 426 device = iser_conn->ib_conn.device; 427 428 if (datalen > 0) { 429 struct ib_sge *tx_dsg = &mdesc->tx_sg[1]; 430 ib_dma_sync_single_for_cpu(device->ib_device, 431 iser_conn->login_req_dma, datalen, 432 DMA_TO_DEVICE); 433 434 ib_dma_sync_single_for_device(device->ib_device, 435 iser_conn->login_req_dma, datalen, 436 DMA_TO_DEVICE); 437 438 tx_dsg->addr = iser_conn->login_req_dma; 439 tx_dsg->length = datalen; 440 tx_dsg->lkey = device->mr->lkey; 441 mdesc->num_sge = 2; 442 } 443 444 /* For login phase and discovery session we re-use the login buffer */ 445 if (!iser_conn->handoff_done) { 446 err = iser_post_recvl(iser_conn); 447 if (err) 448 goto send_control_error; 449 } 450 451 err = iser_post_send(&iser_conn->ib_conn, mdesc, true); 452 if (!err) 453 return (0); 454 455 send_control_error: 456 ISER_ERR("conn %p failed err %d", iser_conn, err); 457 458 return (err); 459 460 } 461 462 /** 463 * iser_rcv_dto_completion - recv DTO completion 464 */ 465 void 466 iser_rcv_completion(struct iser_rx_desc *rx_desc, 467 unsigned long rx_xfer_len, 468 struct ib_conn *ib_conn) 469 { 470 struct iser_conn *iser_conn = container_of(ib_conn, struct iser_conn, 471 ib_conn); 472 struct icl_conn *ic = &iser_conn->icl_conn; 473 struct icl_pdu *response; 474 struct iscsi_bhs *hdr; 475 u64 rx_dma; 476 int rx_buflen; 477 int outstanding, count, err; 478 479 /* differentiate between login to all other PDUs */ 480 if ((char *)rx_desc == iser_conn->login_resp_buf) { 481 rx_dma = iser_conn->login_resp_dma; 482 rx_buflen = ISER_RX_LOGIN_SIZE; 483 } else { 484 rx_dma = rx_desc->dma_addr; 485 rx_buflen = ISER_RX_PAYLOAD_SIZE; 486 } 487 488 ib_dma_sync_single_for_cpu(ib_conn->device->ib_device, rx_dma, 489 rx_buflen, DMA_FROM_DEVICE); 490 491 hdr = &rx_desc->iscsi_header; 492 493 response = iser_new_pdu(ic, M_NOWAIT); 494 response->ip_bhs = hdr; 495 response->ip_data_len = rx_xfer_len - ISER_HEADERS_LEN; 496 497 /* 498 * In case we got data in the receive buffer, assign the ip_data_mbuf 499 * to the rx_buffer - later we'll copy it to upper layer buffers 500 */ 501 if (response->ip_data_len) 502 response->ip_data_mbuf = (struct mbuf *)(rx_desc->data); 503 504 ib_dma_sync_single_for_device(ib_conn->device->ib_device, rx_dma, 505 rx_buflen, DMA_FROM_DEVICE); 506 507 /* decrementing conn->post_recv_buf_count only --after-- freeing the * 508 * task eliminates the need to worry on tasks which are completed in * 509 * parallel to the execution of iser_conn_term. So the code that waits * 510 * for the posted rx bufs refcount to become zero handles everything */ 511 ib_conn->post_recv_buf_count--; 512 513 if (rx_dma == iser_conn->login_resp_dma) 514 goto receive; 515 516 outstanding = ib_conn->post_recv_buf_count; 517 if (outstanding + iser_conn->min_posted_rx <= iser_conn->qp_max_recv_dtos) { 518 count = min(iser_conn->qp_max_recv_dtos - outstanding, 519 iser_conn->min_posted_rx); 520 err = iser_post_recvm(iser_conn, count); 521 if (err) 522 ISER_ERR("posting %d rx bufs err %d", count, err); 523 } 524 525 receive: 526 (ic->ic_receive)(response); 527 } 528 529 void 530 iser_snd_completion(struct iser_tx_desc *tx_desc, 531 struct ib_conn *ib_conn) 532 { 533 struct icl_iser_pdu *iser_pdu = container_of(tx_desc, struct icl_iser_pdu, desc); 534 struct iser_conn *iser_conn = iser_pdu->iser_conn; 535 536 if (tx_desc && tx_desc->type == ISCSI_TX_CONTROL) 537 iser_pdu_free(&iser_conn->icl_conn, &iser_pdu->icl_pdu); 538 } 539