1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /******************************************************************************* 3 * Filename: target_core_pscsi.c 4 * 5 * This file contains the generic target mode <-> Linux SCSI subsystem plugin. 6 * 7 * (c) Copyright 2003-2013 Datera, Inc. 8 * 9 * Nicholas A. Bellinger <nab@kernel.org> 10 * 11 ******************************************************************************/ 12 13 #include <linux/string.h> 14 #include <linux/parser.h> 15 #include <linux/timer.h> 16 #include <linux/blkdev.h> 17 #include <linux/blk_types.h> 18 #include <linux/slab.h> 19 #include <linux/spinlock.h> 20 #include <linux/cdrom.h> 21 #include <linux/ratelimit.h> 22 #include <linux/module.h> 23 #include <asm/unaligned.h> 24 25 #include <scsi/scsi_device.h> 26 #include <scsi/scsi_host.h> 27 #include <scsi/scsi_tcq.h> 28 29 #include <target/target_core_base.h> 30 #include <target/target_core_backend.h> 31 32 #include "target_core_alua.h" 33 #include "target_core_internal.h" 34 #include "target_core_pscsi.h" 35 36 static inline struct pscsi_dev_virt *PSCSI_DEV(struct se_device *dev) 37 { 38 return container_of(dev, struct pscsi_dev_virt, dev); 39 } 40 41 static sense_reason_t pscsi_execute_cmd(struct se_cmd *cmd); 42 static enum rq_end_io_ret pscsi_req_done(struct request *, blk_status_t); 43 44 /* pscsi_attach_hba(): 45 * 46 * pscsi_get_sh() used scsi_host_lookup() to locate struct Scsi_Host. 47 * from the passed SCSI Host ID. 48 */ 49 static int pscsi_attach_hba(struct se_hba *hba, u32 host_id) 50 { 51 struct pscsi_hba_virt *phv; 52 53 phv = kzalloc(sizeof(struct pscsi_hba_virt), GFP_KERNEL); 54 if (!phv) { 55 pr_err("Unable to allocate struct pscsi_hba_virt\n"); 56 return -ENOMEM; 57 } 58 phv->phv_host_id = host_id; 59 phv->phv_mode = PHV_VIRTUAL_HOST_ID; 60 61 hba->hba_ptr = phv; 62 63 pr_debug("CORE_HBA[%d] - TCM SCSI HBA Driver %s on" 64 " Generic Target Core Stack %s\n", hba->hba_id, 65 PSCSI_VERSION, TARGET_CORE_VERSION); 66 pr_debug("CORE_HBA[%d] - Attached SCSI HBA to Generic\n", 67 hba->hba_id); 68 69 return 0; 70 } 71 72 static void pscsi_detach_hba(struct se_hba *hba) 73 { 74 struct pscsi_hba_virt *phv = hba->hba_ptr; 75 struct Scsi_Host *scsi_host = phv->phv_lld_host; 76 77 if (scsi_host) { 78 scsi_host_put(scsi_host); 79 80 pr_debug("CORE_HBA[%d] - Detached SCSI HBA: %s from" 81 " Generic Target Core\n", hba->hba_id, 82 (scsi_host->hostt->name) ? (scsi_host->hostt->name) : 83 "Unknown"); 84 } else 85 pr_debug("CORE_HBA[%d] - Detached Virtual SCSI HBA" 86 " from Generic Target Core\n", hba->hba_id); 87 88 kfree(phv); 89 hba->hba_ptr = NULL; 90 } 91 92 static int pscsi_pmode_enable_hba(struct se_hba *hba, unsigned long mode_flag) 93 { 94 struct pscsi_hba_virt *phv = hba->hba_ptr; 95 struct Scsi_Host *sh = phv->phv_lld_host; 96 /* 97 * Release the struct Scsi_Host 98 */ 99 if (!mode_flag) { 100 if (!sh) 101 return 0; 102 103 phv->phv_lld_host = NULL; 104 phv->phv_mode = PHV_VIRTUAL_HOST_ID; 105 106 pr_debug("CORE_HBA[%d] - Disabled pSCSI HBA Passthrough" 107 " %s\n", hba->hba_id, (sh->hostt->name) ? 108 (sh->hostt->name) : "Unknown"); 109 110 scsi_host_put(sh); 111 return 0; 112 } 113 /* 114 * Otherwise, locate struct Scsi_Host from the original passed 115 * pSCSI Host ID and enable for phba mode 116 */ 117 sh = scsi_host_lookup(phv->phv_host_id); 118 if (!sh) { 119 pr_err("pSCSI: Unable to locate SCSI Host for" 120 " phv_host_id: %d\n", phv->phv_host_id); 121 return -EINVAL; 122 } 123 124 phv->phv_lld_host = sh; 125 phv->phv_mode = PHV_LLD_SCSI_HOST_NO; 126 127 pr_debug("CORE_HBA[%d] - Enabled pSCSI HBA Passthrough %s\n", 128 hba->hba_id, (sh->hostt->name) ? (sh->hostt->name) : "Unknown"); 129 130 return 1; 131 } 132 133 static void pscsi_tape_read_blocksize(struct se_device *dev, 134 struct scsi_device *sdev) 135 { 136 unsigned char cdb[MAX_COMMAND_SIZE], *buf; 137 int ret; 138 139 buf = kzalloc(12, GFP_KERNEL); 140 if (!buf) 141 goto out_free; 142 143 memset(cdb, 0, MAX_COMMAND_SIZE); 144 cdb[0] = MODE_SENSE; 145 cdb[4] = 0x0c; /* 12 bytes */ 146 147 ret = scsi_execute_req(sdev, cdb, DMA_FROM_DEVICE, buf, 12, NULL, 148 HZ, 1, NULL); 149 if (ret) 150 goto out_free; 151 152 /* 153 * If MODE_SENSE still returns zero, set the default value to 1024. 154 */ 155 sdev->sector_size = get_unaligned_be24(&buf[9]); 156 out_free: 157 if (!sdev->sector_size) 158 sdev->sector_size = 1024; 159 160 kfree(buf); 161 } 162 163 static void 164 pscsi_set_inquiry_info(struct scsi_device *sdev, struct t10_wwn *wwn) 165 { 166 if (sdev->inquiry_len < INQUIRY_LEN) 167 return; 168 /* 169 * Use sdev->inquiry data from drivers/scsi/scsi_scan.c:scsi_add_lun() 170 */ 171 BUILD_BUG_ON(sizeof(wwn->vendor) != INQUIRY_VENDOR_LEN + 1); 172 snprintf(wwn->vendor, sizeof(wwn->vendor), 173 "%." __stringify(INQUIRY_VENDOR_LEN) "s", sdev->vendor); 174 BUILD_BUG_ON(sizeof(wwn->model) != INQUIRY_MODEL_LEN + 1); 175 snprintf(wwn->model, sizeof(wwn->model), 176 "%." __stringify(INQUIRY_MODEL_LEN) "s", sdev->model); 177 BUILD_BUG_ON(sizeof(wwn->revision) != INQUIRY_REVISION_LEN + 1); 178 snprintf(wwn->revision, sizeof(wwn->revision), 179 "%." __stringify(INQUIRY_REVISION_LEN) "s", sdev->rev); 180 } 181 182 static int 183 pscsi_get_inquiry_vpd_serial(struct scsi_device *sdev, struct t10_wwn *wwn) 184 { 185 unsigned char cdb[MAX_COMMAND_SIZE], *buf; 186 int ret; 187 188 buf = kzalloc(INQUIRY_VPD_SERIAL_LEN, GFP_KERNEL); 189 if (!buf) 190 return -ENOMEM; 191 192 memset(cdb, 0, MAX_COMMAND_SIZE); 193 cdb[0] = INQUIRY; 194 cdb[1] = 0x01; /* Query VPD */ 195 cdb[2] = 0x80; /* Unit Serial Number */ 196 put_unaligned_be16(INQUIRY_VPD_SERIAL_LEN, &cdb[3]); 197 198 ret = scsi_execute_req(sdev, cdb, DMA_FROM_DEVICE, buf, 199 INQUIRY_VPD_SERIAL_LEN, NULL, HZ, 1, NULL); 200 if (ret) 201 goto out_free; 202 203 snprintf(&wwn->unit_serial[0], INQUIRY_VPD_SERIAL_LEN, "%s", &buf[4]); 204 205 wwn->t10_dev->dev_flags |= DF_FIRMWARE_VPD_UNIT_SERIAL; 206 207 kfree(buf); 208 return 0; 209 210 out_free: 211 kfree(buf); 212 return -EPERM; 213 } 214 215 static void 216 pscsi_get_inquiry_vpd_device_ident(struct scsi_device *sdev, 217 struct t10_wwn *wwn) 218 { 219 unsigned char cdb[MAX_COMMAND_SIZE], *buf, *page_83; 220 int ident_len, page_len, off = 4, ret; 221 struct t10_vpd *vpd; 222 223 buf = kzalloc(INQUIRY_VPD_SERIAL_LEN, GFP_KERNEL); 224 if (!buf) 225 return; 226 227 memset(cdb, 0, MAX_COMMAND_SIZE); 228 cdb[0] = INQUIRY; 229 cdb[1] = 0x01; /* Query VPD */ 230 cdb[2] = 0x83; /* Device Identifier */ 231 put_unaligned_be16(INQUIRY_VPD_DEVICE_IDENTIFIER_LEN, &cdb[3]); 232 233 ret = scsi_execute_req(sdev, cdb, DMA_FROM_DEVICE, buf, 234 INQUIRY_VPD_DEVICE_IDENTIFIER_LEN, 235 NULL, HZ, 1, NULL); 236 if (ret) 237 goto out; 238 239 page_len = get_unaligned_be16(&buf[2]); 240 while (page_len > 0) { 241 /* Grab a pointer to the Identification descriptor */ 242 page_83 = &buf[off]; 243 ident_len = page_83[3]; 244 if (!ident_len) { 245 pr_err("page_83[3]: identifier" 246 " length zero!\n"); 247 break; 248 } 249 pr_debug("T10 VPD Identifier Length: %d\n", ident_len); 250 251 vpd = kzalloc(sizeof(struct t10_vpd), GFP_KERNEL); 252 if (!vpd) { 253 pr_err("Unable to allocate memory for" 254 " struct t10_vpd\n"); 255 goto out; 256 } 257 INIT_LIST_HEAD(&vpd->vpd_list); 258 259 transport_set_vpd_proto_id(vpd, page_83); 260 transport_set_vpd_assoc(vpd, page_83); 261 262 if (transport_set_vpd_ident_type(vpd, page_83) < 0) { 263 off += (ident_len + 4); 264 page_len -= (ident_len + 4); 265 kfree(vpd); 266 continue; 267 } 268 if (transport_set_vpd_ident(vpd, page_83) < 0) { 269 off += (ident_len + 4); 270 page_len -= (ident_len + 4); 271 kfree(vpd); 272 continue; 273 } 274 275 list_add_tail(&vpd->vpd_list, &wwn->t10_vpd_list); 276 off += (ident_len + 4); 277 page_len -= (ident_len + 4); 278 } 279 280 out: 281 kfree(buf); 282 } 283 284 static int pscsi_add_device_to_list(struct se_device *dev, 285 struct scsi_device *sd) 286 { 287 struct pscsi_dev_virt *pdv = PSCSI_DEV(dev); 288 struct request_queue *q = sd->request_queue; 289 290 pdv->pdv_sd = sd; 291 292 if (!sd->queue_depth) { 293 sd->queue_depth = PSCSI_DEFAULT_QUEUEDEPTH; 294 295 pr_err("Set broken SCSI Device %d:%d:%llu" 296 " queue_depth to %d\n", sd->channel, sd->id, 297 sd->lun, sd->queue_depth); 298 } 299 300 dev->dev_attrib.hw_block_size = 301 min_not_zero((int)sd->sector_size, 512); 302 dev->dev_attrib.hw_max_sectors = 303 min_not_zero(sd->host->max_sectors, queue_max_hw_sectors(q)); 304 dev->dev_attrib.hw_queue_depth = sd->queue_depth; 305 306 /* 307 * Setup our standard INQUIRY info into se_dev->t10_wwn 308 */ 309 pscsi_set_inquiry_info(sd, &dev->t10_wwn); 310 311 /* 312 * Locate VPD WWN Information used for various purposes within 313 * the Storage Engine. 314 */ 315 if (!pscsi_get_inquiry_vpd_serial(sd, &dev->t10_wwn)) { 316 /* 317 * If VPD Unit Serial returned GOOD status, try 318 * VPD Device Identification page (0x83). 319 */ 320 pscsi_get_inquiry_vpd_device_ident(sd, &dev->t10_wwn); 321 } 322 323 /* 324 * For TYPE_TAPE, attempt to determine blocksize with MODE_SENSE. 325 */ 326 if (sd->type == TYPE_TAPE) { 327 pscsi_tape_read_blocksize(dev, sd); 328 dev->dev_attrib.hw_block_size = sd->sector_size; 329 } 330 return 0; 331 } 332 333 static struct se_device *pscsi_alloc_device(struct se_hba *hba, 334 const char *name) 335 { 336 struct pscsi_dev_virt *pdv; 337 338 pdv = kzalloc(sizeof(struct pscsi_dev_virt), GFP_KERNEL); 339 if (!pdv) { 340 pr_err("Unable to allocate memory for struct pscsi_dev_virt\n"); 341 return NULL; 342 } 343 344 pr_debug("PSCSI: Allocated pdv: %p for %s\n", pdv, name); 345 return &pdv->dev; 346 } 347 348 /* 349 * Called with struct Scsi_Host->host_lock called. 350 */ 351 static int pscsi_create_type_disk(struct se_device *dev, struct scsi_device *sd) 352 __releases(sh->host_lock) 353 { 354 struct pscsi_hba_virt *phv = dev->se_hba->hba_ptr; 355 struct pscsi_dev_virt *pdv = PSCSI_DEV(dev); 356 struct Scsi_Host *sh = sd->host; 357 struct block_device *bd; 358 int ret; 359 360 if (scsi_device_get(sd)) { 361 pr_err("scsi_device_get() failed for %d:%d:%d:%llu\n", 362 sh->host_no, sd->channel, sd->id, sd->lun); 363 spin_unlock_irq(sh->host_lock); 364 return -EIO; 365 } 366 spin_unlock_irq(sh->host_lock); 367 /* 368 * Claim exclusive struct block_device access to struct scsi_device 369 * for TYPE_DISK and TYPE_ZBC using supplied udev_path 370 */ 371 bd = blkdev_get_by_path(dev->udev_path, 372 FMODE_WRITE|FMODE_READ|FMODE_EXCL, pdv); 373 if (IS_ERR(bd)) { 374 pr_err("pSCSI: blkdev_get_by_path() failed\n"); 375 scsi_device_put(sd); 376 return PTR_ERR(bd); 377 } 378 pdv->pdv_bd = bd; 379 380 ret = pscsi_add_device_to_list(dev, sd); 381 if (ret) { 382 blkdev_put(pdv->pdv_bd, FMODE_WRITE|FMODE_READ|FMODE_EXCL); 383 scsi_device_put(sd); 384 return ret; 385 } 386 387 pr_debug("CORE_PSCSI[%d] - Added TYPE_%s for %d:%d:%d:%llu\n", 388 phv->phv_host_id, sd->type == TYPE_DISK ? "DISK" : "ZBC", 389 sh->host_no, sd->channel, sd->id, sd->lun); 390 return 0; 391 } 392 393 /* 394 * Called with struct Scsi_Host->host_lock called. 395 */ 396 static int pscsi_create_type_nondisk(struct se_device *dev, struct scsi_device *sd) 397 __releases(sh->host_lock) 398 { 399 struct pscsi_hba_virt *phv = dev->se_hba->hba_ptr; 400 struct Scsi_Host *sh = sd->host; 401 int ret; 402 403 if (scsi_device_get(sd)) { 404 pr_err("scsi_device_get() failed for %d:%d:%d:%llu\n", 405 sh->host_no, sd->channel, sd->id, sd->lun); 406 spin_unlock_irq(sh->host_lock); 407 return -EIO; 408 } 409 spin_unlock_irq(sh->host_lock); 410 411 ret = pscsi_add_device_to_list(dev, sd); 412 if (ret) { 413 scsi_device_put(sd); 414 return ret; 415 } 416 pr_debug("CORE_PSCSI[%d] - Added Type: %s for %d:%d:%d:%llu\n", 417 phv->phv_host_id, scsi_device_type(sd->type), sh->host_no, 418 sd->channel, sd->id, sd->lun); 419 420 return 0; 421 } 422 423 static int pscsi_configure_device(struct se_device *dev) 424 { 425 struct se_hba *hba = dev->se_hba; 426 struct pscsi_dev_virt *pdv = PSCSI_DEV(dev); 427 struct scsi_device *sd; 428 struct pscsi_hba_virt *phv = dev->se_hba->hba_ptr; 429 struct Scsi_Host *sh = phv->phv_lld_host; 430 int legacy_mode_enable = 0; 431 int ret; 432 433 if (!(pdv->pdv_flags & PDF_HAS_CHANNEL_ID) || 434 !(pdv->pdv_flags & PDF_HAS_TARGET_ID) || 435 !(pdv->pdv_flags & PDF_HAS_LUN_ID)) { 436 pr_err("Missing scsi_channel_id=, scsi_target_id= and" 437 " scsi_lun_id= parameters\n"); 438 return -EINVAL; 439 } 440 441 /* 442 * If not running in PHV_LLD_SCSI_HOST_NO mode, locate the 443 * struct Scsi_Host we will need to bring the TCM/pSCSI object online 444 */ 445 if (!sh) { 446 if (phv->phv_mode == PHV_LLD_SCSI_HOST_NO) { 447 pr_err("pSCSI: Unable to locate struct" 448 " Scsi_Host for PHV_LLD_SCSI_HOST_NO\n"); 449 return -ENODEV; 450 } 451 /* 452 * For the newer PHV_VIRTUAL_HOST_ID struct scsi_device 453 * reference, we enforce that udev_path has been set 454 */ 455 if (!(dev->dev_flags & DF_USING_UDEV_PATH)) { 456 pr_err("pSCSI: udev_path attribute has not" 457 " been set before ENABLE=1\n"); 458 return -EINVAL; 459 } 460 /* 461 * If no scsi_host_id= was passed for PHV_VIRTUAL_HOST_ID, 462 * use the original TCM hba ID to reference Linux/SCSI Host No 463 * and enable for PHV_LLD_SCSI_HOST_NO mode. 464 */ 465 if (!(pdv->pdv_flags & PDF_HAS_VIRT_HOST_ID)) { 466 if (hba->dev_count) { 467 pr_err("pSCSI: Unable to set hba_mode" 468 " with active devices\n"); 469 return -EEXIST; 470 } 471 472 if (pscsi_pmode_enable_hba(hba, 1) != 1) 473 return -ENODEV; 474 475 legacy_mode_enable = 1; 476 hba->hba_flags |= HBA_FLAGS_PSCSI_MODE; 477 sh = phv->phv_lld_host; 478 } else { 479 sh = scsi_host_lookup(pdv->pdv_host_id); 480 if (!sh) { 481 pr_err("pSCSI: Unable to locate" 482 " pdv_host_id: %d\n", pdv->pdv_host_id); 483 return -EINVAL; 484 } 485 pdv->pdv_lld_host = sh; 486 } 487 } else { 488 if (phv->phv_mode == PHV_VIRTUAL_HOST_ID) { 489 pr_err("pSCSI: PHV_VIRTUAL_HOST_ID set while" 490 " struct Scsi_Host exists\n"); 491 return -EEXIST; 492 } 493 } 494 495 spin_lock_irq(sh->host_lock); 496 list_for_each_entry(sd, &sh->__devices, siblings) { 497 if ((pdv->pdv_channel_id != sd->channel) || 498 (pdv->pdv_target_id != sd->id) || 499 (pdv->pdv_lun_id != sd->lun)) 500 continue; 501 /* 502 * Functions will release the held struct scsi_host->host_lock 503 * before calling pscsi_add_device_to_list() to register 504 * struct scsi_device with target_core_mod. 505 */ 506 switch (sd->type) { 507 case TYPE_DISK: 508 case TYPE_ZBC: 509 ret = pscsi_create_type_disk(dev, sd); 510 break; 511 default: 512 ret = pscsi_create_type_nondisk(dev, sd); 513 break; 514 } 515 516 if (ret) { 517 if (phv->phv_mode == PHV_VIRTUAL_HOST_ID) 518 scsi_host_put(sh); 519 else if (legacy_mode_enable) { 520 pscsi_pmode_enable_hba(hba, 0); 521 hba->hba_flags &= ~HBA_FLAGS_PSCSI_MODE; 522 } 523 pdv->pdv_sd = NULL; 524 return ret; 525 } 526 return 0; 527 } 528 spin_unlock_irq(sh->host_lock); 529 530 pr_err("pSCSI: Unable to locate %d:%d:%d:%d\n", sh->host_no, 531 pdv->pdv_channel_id, pdv->pdv_target_id, pdv->pdv_lun_id); 532 533 if (phv->phv_mode == PHV_VIRTUAL_HOST_ID) 534 scsi_host_put(sh); 535 else if (legacy_mode_enable) { 536 pscsi_pmode_enable_hba(hba, 0); 537 hba->hba_flags &= ~HBA_FLAGS_PSCSI_MODE; 538 } 539 540 return -ENODEV; 541 } 542 543 static void pscsi_dev_call_rcu(struct rcu_head *p) 544 { 545 struct se_device *dev = container_of(p, struct se_device, rcu_head); 546 struct pscsi_dev_virt *pdv = PSCSI_DEV(dev); 547 548 kfree(pdv); 549 } 550 551 static void pscsi_free_device(struct se_device *dev) 552 { 553 call_rcu(&dev->rcu_head, pscsi_dev_call_rcu); 554 } 555 556 static void pscsi_destroy_device(struct se_device *dev) 557 { 558 struct pscsi_dev_virt *pdv = PSCSI_DEV(dev); 559 struct pscsi_hba_virt *phv = dev->se_hba->hba_ptr; 560 struct scsi_device *sd = pdv->pdv_sd; 561 562 if (sd) { 563 /* 564 * Release exclusive pSCSI internal struct block_device claim for 565 * struct scsi_device with TYPE_DISK or TYPE_ZBC 566 * from pscsi_create_type_disk() 567 */ 568 if ((sd->type == TYPE_DISK || sd->type == TYPE_ZBC) && 569 pdv->pdv_bd) { 570 blkdev_put(pdv->pdv_bd, 571 FMODE_WRITE|FMODE_READ|FMODE_EXCL); 572 pdv->pdv_bd = NULL; 573 } 574 /* 575 * For HBA mode PHV_LLD_SCSI_HOST_NO, release the reference 576 * to struct Scsi_Host now. 577 */ 578 if ((phv->phv_mode == PHV_LLD_SCSI_HOST_NO) && 579 (phv->phv_lld_host != NULL)) 580 scsi_host_put(phv->phv_lld_host); 581 else if (pdv->pdv_lld_host) 582 scsi_host_put(pdv->pdv_lld_host); 583 584 scsi_device_put(sd); 585 586 pdv->pdv_sd = NULL; 587 } 588 } 589 590 static void pscsi_complete_cmd(struct se_cmd *cmd, u8 scsi_status, 591 unsigned char *req_sense, int valid_data) 592 { 593 struct pscsi_dev_virt *pdv = PSCSI_DEV(cmd->se_dev); 594 struct scsi_device *sd = pdv->pdv_sd; 595 unsigned char *cdb = cmd->priv; 596 597 /* 598 * Special case for REPORT_LUNs which is emulated and not passed on. 599 */ 600 if (!cdb) 601 return; 602 603 /* 604 * Hack to make sure that Write-Protect modepage is set if R/O mode is 605 * forced. 606 */ 607 if (!cmd->data_length) 608 goto after_mode_sense; 609 610 if (((cdb[0] == MODE_SENSE) || (cdb[0] == MODE_SENSE_10)) && 611 scsi_status == SAM_STAT_GOOD) { 612 bool read_only = target_lun_is_rdonly(cmd); 613 614 if (read_only) { 615 unsigned char *buf; 616 617 buf = transport_kmap_data_sg(cmd); 618 if (!buf) { 619 ; /* XXX: TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE */ 620 } else { 621 if (cdb[0] == MODE_SENSE_10) { 622 if (!(buf[3] & 0x80)) 623 buf[3] |= 0x80; 624 } else { 625 if (!(buf[2] & 0x80)) 626 buf[2] |= 0x80; 627 } 628 629 transport_kunmap_data_sg(cmd); 630 } 631 } 632 } 633 after_mode_sense: 634 635 if (sd->type != TYPE_TAPE || !cmd->data_length) 636 goto after_mode_select; 637 638 /* 639 * Hack to correctly obtain the initiator requested blocksize for 640 * TYPE_TAPE. Since this value is dependent upon each tape media, 641 * struct scsi_device->sector_size will not contain the correct value 642 * by default, so we go ahead and set it so 643 * TRANSPORT(dev)->get_blockdev() returns the correct value to the 644 * storage engine. 645 */ 646 if (((cdb[0] == MODE_SELECT) || (cdb[0] == MODE_SELECT_10)) && 647 scsi_status == SAM_STAT_GOOD) { 648 unsigned char *buf; 649 u16 bdl; 650 u32 blocksize; 651 652 buf = sg_virt(&cmd->t_data_sg[0]); 653 if (!buf) { 654 pr_err("Unable to get buf for scatterlist\n"); 655 goto after_mode_select; 656 } 657 658 if (cdb[0] == MODE_SELECT) 659 bdl = buf[3]; 660 else 661 bdl = get_unaligned_be16(&buf[6]); 662 663 if (!bdl) 664 goto after_mode_select; 665 666 if (cdb[0] == MODE_SELECT) 667 blocksize = get_unaligned_be24(&buf[9]); 668 else 669 blocksize = get_unaligned_be24(&buf[13]); 670 671 sd->sector_size = blocksize; 672 } 673 after_mode_select: 674 675 if (scsi_status == SAM_STAT_CHECK_CONDITION) { 676 transport_copy_sense_to_cmd(cmd, req_sense); 677 678 /* 679 * check for TAPE device reads with 680 * FM/EOM/ILI set, so that we can get data 681 * back despite framework assumption that a 682 * check condition means there is no data 683 */ 684 if (sd->type == TYPE_TAPE && valid_data && 685 cmd->data_direction == DMA_FROM_DEVICE) { 686 /* 687 * is sense data valid, fixed format, 688 * and have FM, EOM, or ILI set? 689 */ 690 if (req_sense[0] == 0xf0 && /* valid, fixed format */ 691 req_sense[2] & 0xe0 && /* FM, EOM, or ILI */ 692 (req_sense[2] & 0xf) == 0) { /* key==NO_SENSE */ 693 pr_debug("Tape FM/EOM/ILI status detected. Treat as normal read.\n"); 694 cmd->se_cmd_flags |= SCF_TREAT_READ_AS_NORMAL; 695 } 696 } 697 } 698 } 699 700 enum { 701 Opt_scsi_host_id, Opt_scsi_channel_id, Opt_scsi_target_id, 702 Opt_scsi_lun_id, Opt_err 703 }; 704 705 static match_table_t tokens = { 706 {Opt_scsi_host_id, "scsi_host_id=%d"}, 707 {Opt_scsi_channel_id, "scsi_channel_id=%d"}, 708 {Opt_scsi_target_id, "scsi_target_id=%d"}, 709 {Opt_scsi_lun_id, "scsi_lun_id=%d"}, 710 {Opt_err, NULL} 711 }; 712 713 static ssize_t pscsi_set_configfs_dev_params(struct se_device *dev, 714 const char *page, ssize_t count) 715 { 716 struct pscsi_dev_virt *pdv = PSCSI_DEV(dev); 717 struct pscsi_hba_virt *phv = dev->se_hba->hba_ptr; 718 char *orig, *ptr, *opts; 719 substring_t args[MAX_OPT_ARGS]; 720 int ret = 0, arg, token; 721 722 opts = kstrdup(page, GFP_KERNEL); 723 if (!opts) 724 return -ENOMEM; 725 726 orig = opts; 727 728 while ((ptr = strsep(&opts, ",\n")) != NULL) { 729 if (!*ptr) 730 continue; 731 732 token = match_token(ptr, tokens, args); 733 switch (token) { 734 case Opt_scsi_host_id: 735 if (phv->phv_mode == PHV_LLD_SCSI_HOST_NO) { 736 pr_err("PSCSI[%d]: Unable to accept" 737 " scsi_host_id while phv_mode ==" 738 " PHV_LLD_SCSI_HOST_NO\n", 739 phv->phv_host_id); 740 ret = -EINVAL; 741 goto out; 742 } 743 ret = match_int(args, &arg); 744 if (ret) 745 goto out; 746 pdv->pdv_host_id = arg; 747 pr_debug("PSCSI[%d]: Referencing SCSI Host ID:" 748 " %d\n", phv->phv_host_id, pdv->pdv_host_id); 749 pdv->pdv_flags |= PDF_HAS_VIRT_HOST_ID; 750 break; 751 case Opt_scsi_channel_id: 752 ret = match_int(args, &arg); 753 if (ret) 754 goto out; 755 pdv->pdv_channel_id = arg; 756 pr_debug("PSCSI[%d]: Referencing SCSI Channel" 757 " ID: %d\n", phv->phv_host_id, 758 pdv->pdv_channel_id); 759 pdv->pdv_flags |= PDF_HAS_CHANNEL_ID; 760 break; 761 case Opt_scsi_target_id: 762 ret = match_int(args, &arg); 763 if (ret) 764 goto out; 765 pdv->pdv_target_id = arg; 766 pr_debug("PSCSI[%d]: Referencing SCSI Target" 767 " ID: %d\n", phv->phv_host_id, 768 pdv->pdv_target_id); 769 pdv->pdv_flags |= PDF_HAS_TARGET_ID; 770 break; 771 case Opt_scsi_lun_id: 772 ret = match_int(args, &arg); 773 if (ret) 774 goto out; 775 pdv->pdv_lun_id = arg; 776 pr_debug("PSCSI[%d]: Referencing SCSI LUN ID:" 777 " %d\n", phv->phv_host_id, pdv->pdv_lun_id); 778 pdv->pdv_flags |= PDF_HAS_LUN_ID; 779 break; 780 default: 781 break; 782 } 783 } 784 785 out: 786 kfree(orig); 787 return (!ret) ? count : ret; 788 } 789 790 static ssize_t pscsi_show_configfs_dev_params(struct se_device *dev, char *b) 791 { 792 struct pscsi_hba_virt *phv = dev->se_hba->hba_ptr; 793 struct pscsi_dev_virt *pdv = PSCSI_DEV(dev); 794 struct scsi_device *sd = pdv->pdv_sd; 795 unsigned char host_id[16]; 796 ssize_t bl; 797 798 if (phv->phv_mode == PHV_VIRTUAL_HOST_ID) 799 snprintf(host_id, 16, "%d", pdv->pdv_host_id); 800 else 801 snprintf(host_id, 16, "PHBA Mode"); 802 803 bl = sprintf(b, "SCSI Device Bus Location:" 804 " Channel ID: %d Target ID: %d LUN: %d Host ID: %s\n", 805 pdv->pdv_channel_id, pdv->pdv_target_id, pdv->pdv_lun_id, 806 host_id); 807 808 if (sd) { 809 bl += sprintf(b + bl, " Vendor: %." 810 __stringify(INQUIRY_VENDOR_LEN) "s", sd->vendor); 811 bl += sprintf(b + bl, " Model: %." 812 __stringify(INQUIRY_MODEL_LEN) "s", sd->model); 813 bl += sprintf(b + bl, " Rev: %." 814 __stringify(INQUIRY_REVISION_LEN) "s\n", sd->rev); 815 } 816 return bl; 817 } 818 819 static void pscsi_bi_endio(struct bio *bio) 820 { 821 bio_uninit(bio); 822 kfree(bio); 823 } 824 825 static sense_reason_t 826 pscsi_map_sg(struct se_cmd *cmd, struct scatterlist *sgl, u32 sgl_nents, 827 struct request *req) 828 { 829 struct pscsi_dev_virt *pdv = PSCSI_DEV(cmd->se_dev); 830 struct bio *bio = NULL; 831 struct page *page; 832 struct scatterlist *sg; 833 u32 data_len = cmd->data_length, i, len, bytes, off; 834 int nr_pages = (cmd->data_length + sgl[0].offset + 835 PAGE_SIZE - 1) >> PAGE_SHIFT; 836 int nr_vecs = 0, rc; 837 int rw = (cmd->data_direction == DMA_TO_DEVICE); 838 839 BUG_ON(!cmd->data_length); 840 841 pr_debug("PSCSI: nr_pages: %d\n", nr_pages); 842 843 for_each_sg(sgl, sg, sgl_nents, i) { 844 page = sg_page(sg); 845 off = sg->offset; 846 len = sg->length; 847 848 pr_debug("PSCSI: i: %d page: %p len: %d off: %d\n", i, 849 page, len, off); 850 851 /* 852 * We only have one page of data in each sg element, 853 * we can not cross a page boundary. 854 */ 855 if (off + len > PAGE_SIZE) 856 goto fail; 857 858 if (len > 0 && data_len > 0) { 859 bytes = min_t(unsigned int, len, PAGE_SIZE - off); 860 bytes = min(bytes, data_len); 861 862 if (!bio) { 863 new_bio: 864 nr_vecs = bio_max_segs(nr_pages); 865 bio = bio_kmalloc(nr_vecs, GFP_KERNEL); 866 if (!bio) 867 goto fail; 868 bio_init(bio, NULL, bio->bi_inline_vecs, nr_vecs, 869 rw ? REQ_OP_WRITE : REQ_OP_READ); 870 bio->bi_end_io = pscsi_bi_endio; 871 872 pr_debug("PSCSI: Allocated bio: %p," 873 " dir: %s nr_vecs: %d\n", bio, 874 (rw) ? "rw" : "r", nr_vecs); 875 } 876 877 pr_debug("PSCSI: Calling bio_add_pc_page() i: %d" 878 " bio: %p page: %p len: %d off: %d\n", i, bio, 879 page, len, off); 880 881 rc = bio_add_pc_page(pdv->pdv_sd->request_queue, 882 bio, page, bytes, off); 883 pr_debug("PSCSI: bio->bi_vcnt: %d nr_vecs: %d\n", 884 bio_segments(bio), nr_vecs); 885 if (rc != bytes) { 886 pr_debug("PSCSI: Reached bio->bi_vcnt max:" 887 " %d i: %d bio: %p, allocating another" 888 " bio\n", bio->bi_vcnt, i, bio); 889 890 rc = blk_rq_append_bio(req, bio); 891 if (rc) { 892 pr_err("pSCSI: failed to append bio\n"); 893 goto fail; 894 } 895 896 goto new_bio; 897 } 898 899 data_len -= bytes; 900 } 901 } 902 903 if (bio) { 904 rc = blk_rq_append_bio(req, bio); 905 if (rc) { 906 pr_err("pSCSI: failed to append bio\n"); 907 goto fail; 908 } 909 } 910 911 return 0; 912 fail: 913 if (bio) 914 bio_put(bio); 915 while (req->bio) { 916 bio = req->bio; 917 req->bio = bio->bi_next; 918 bio_put(bio); 919 } 920 req->biotail = NULL; 921 return TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE; 922 } 923 924 static sense_reason_t 925 pscsi_parse_cdb(struct se_cmd *cmd) 926 { 927 if (cmd->se_cmd_flags & SCF_BIDI) 928 return TCM_UNSUPPORTED_SCSI_OPCODE; 929 930 return passthrough_parse_cdb(cmd, pscsi_execute_cmd); 931 } 932 933 static sense_reason_t 934 pscsi_execute_cmd(struct se_cmd *cmd) 935 { 936 struct scatterlist *sgl = cmd->t_data_sg; 937 u32 sgl_nents = cmd->t_data_nents; 938 struct pscsi_dev_virt *pdv = PSCSI_DEV(cmd->se_dev); 939 struct scsi_cmnd *scmd; 940 struct request *req; 941 sense_reason_t ret; 942 943 req = scsi_alloc_request(pdv->pdv_sd->request_queue, 944 cmd->data_direction == DMA_TO_DEVICE ? 945 REQ_OP_DRV_OUT : REQ_OP_DRV_IN, 0); 946 if (IS_ERR(req)) 947 return TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE; 948 949 if (sgl) { 950 ret = pscsi_map_sg(cmd, sgl, sgl_nents, req); 951 if (ret) 952 goto fail_put_request; 953 } 954 955 req->end_io = pscsi_req_done; 956 req->end_io_data = cmd; 957 958 scmd = blk_mq_rq_to_pdu(req); 959 scmd->cmd_len = scsi_command_size(cmd->t_task_cdb); 960 if (scmd->cmd_len > sizeof(scmd->cmnd)) { 961 ret = TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE; 962 goto fail_put_request; 963 } 964 memcpy(scmd->cmnd, cmd->t_task_cdb, scmd->cmd_len); 965 966 if (pdv->pdv_sd->type == TYPE_DISK || 967 pdv->pdv_sd->type == TYPE_ZBC) 968 req->timeout = PS_TIMEOUT_DISK; 969 else 970 req->timeout = PS_TIMEOUT_OTHER; 971 scmd->allowed = PS_RETRY; 972 973 cmd->priv = scmd->cmnd; 974 975 blk_execute_rq_nowait(req, cmd->sam_task_attr == TCM_HEAD_TAG); 976 977 return 0; 978 979 fail_put_request: 980 blk_mq_free_request(req); 981 return ret; 982 } 983 984 /* pscsi_get_device_type(): 985 * 986 * 987 */ 988 static u32 pscsi_get_device_type(struct se_device *dev) 989 { 990 struct pscsi_dev_virt *pdv = PSCSI_DEV(dev); 991 struct scsi_device *sd = pdv->pdv_sd; 992 993 return (sd) ? sd->type : TYPE_NO_LUN; 994 } 995 996 static sector_t pscsi_get_blocks(struct se_device *dev) 997 { 998 struct pscsi_dev_virt *pdv = PSCSI_DEV(dev); 999 1000 if (pdv->pdv_bd) 1001 return bdev_nr_sectors(pdv->pdv_bd); 1002 return 0; 1003 } 1004 1005 static enum rq_end_io_ret pscsi_req_done(struct request *req, 1006 blk_status_t status) 1007 { 1008 struct se_cmd *cmd = req->end_io_data; 1009 struct scsi_cmnd *scmd = blk_mq_rq_to_pdu(req); 1010 enum sam_status scsi_status = scmd->result & 0xff; 1011 int valid_data = cmd->data_length - scmd->resid_len; 1012 u8 *cdb = cmd->priv; 1013 1014 if (scsi_status != SAM_STAT_GOOD) { 1015 pr_debug("PSCSI Status Byte exception at cmd: %p CDB:" 1016 " 0x%02x Result: 0x%08x\n", cmd, cdb[0], scmd->result); 1017 } 1018 1019 pscsi_complete_cmd(cmd, scsi_status, scmd->sense_buffer, valid_data); 1020 1021 switch (host_byte(scmd->result)) { 1022 case DID_OK: 1023 target_complete_cmd_with_length(cmd, scsi_status, valid_data); 1024 break; 1025 default: 1026 pr_debug("PSCSI Host Byte exception at cmd: %p CDB:" 1027 " 0x%02x Result: 0x%08x\n", cmd, cdb[0], scmd->result); 1028 target_complete_cmd(cmd, SAM_STAT_CHECK_CONDITION); 1029 break; 1030 } 1031 1032 blk_mq_free_request(req); 1033 return RQ_END_IO_NONE; 1034 } 1035 1036 static const struct target_backend_ops pscsi_ops = { 1037 .name = "pscsi", 1038 .owner = THIS_MODULE, 1039 .transport_flags_default = TRANSPORT_FLAG_PASSTHROUGH | 1040 TRANSPORT_FLAG_PASSTHROUGH_ALUA | 1041 TRANSPORT_FLAG_PASSTHROUGH_PGR, 1042 .attach_hba = pscsi_attach_hba, 1043 .detach_hba = pscsi_detach_hba, 1044 .pmode_enable_hba = pscsi_pmode_enable_hba, 1045 .alloc_device = pscsi_alloc_device, 1046 .configure_device = pscsi_configure_device, 1047 .destroy_device = pscsi_destroy_device, 1048 .free_device = pscsi_free_device, 1049 .parse_cdb = pscsi_parse_cdb, 1050 .set_configfs_dev_params = pscsi_set_configfs_dev_params, 1051 .show_configfs_dev_params = pscsi_show_configfs_dev_params, 1052 .get_device_type = pscsi_get_device_type, 1053 .get_blocks = pscsi_get_blocks, 1054 .tb_dev_attrib_attrs = passthrough_attrib_attrs, 1055 }; 1056 1057 static int __init pscsi_module_init(void) 1058 { 1059 return transport_backend_register(&pscsi_ops); 1060 } 1061 1062 static void __exit pscsi_module_exit(void) 1063 { 1064 target_backend_unregister(&pscsi_ops); 1065 } 1066 1067 MODULE_DESCRIPTION("TCM PSCSI subsystem plugin"); 1068 MODULE_AUTHOR("nab@Linux-iSCSI.org"); 1069 MODULE_LICENSE("GPL"); 1070 1071 module_init(pscsi_module_init); 1072 module_exit(pscsi_module_exit); 1073