1 /* 2 * Scsi Host Layer for MPT (Message Passing Technology) based controllers 3 * 4 * This code is based on drivers/scsi/mpt3sas/mpt3sas_scsih.c 5 * Copyright (C) 2012-2014 LSI Corporation 6 * Copyright (C) 2013-2014 Avago Technologies 7 * (mailto: MPT-FusionLinux.pdl@avagotech.com) 8 * 9 * This program is free software; you can redistribute it and/or 10 * modify it under the terms of the GNU General Public License 11 * as published by the Free Software Foundation; either version 2 12 * of the License, or (at your option) any later version. 13 * 14 * This program is distributed in the hope that it will be useful, 15 * but WITHOUT ANY WARRANTY; without even the implied warranty of 16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 17 * GNU General Public License for more details. 18 * 19 * NO WARRANTY 20 * THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR 21 * CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT 22 * LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT, 23 * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is 24 * solely responsible for determining the appropriateness of using and 25 * distributing the Program and assumes all risks associated with its 26 * exercise of rights under this Agreement, including but not limited to 27 * the risks and costs of program errors, damage to or loss of data, 28 * programs or equipment, and unavailability or interruption of operations. 29 30 * DISCLAIMER OF LIABILITY 31 * NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY 32 * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 33 * DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND 34 * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR 35 * TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE 36 * USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED 37 * HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES 38 39 * You should have received a copy of the GNU General Public License 40 * along with this program; if not, write to the Free Software 41 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, 42 * USA. 43 */ 44 45 #include <linux/module.h> 46 #include <linux/kernel.h> 47 #include <linux/init.h> 48 #include <linux/errno.h> 49 #include <linux/blkdev.h> 50 #include <linux/sched.h> 51 #include <linux/workqueue.h> 52 #include <linux/delay.h> 53 #include <linux/pci.h> 54 #include <linux/interrupt.h> 55 #include <linux/raid_class.h> 56 #include <linux/blk-mq-pci.h> 57 #include <linux/unaligned.h> 58 59 #include "mpt3sas_base.h" 60 61 #define RAID_CHANNEL 1 62 63 #define PCIE_CHANNEL 2 64 65 /* forward proto's */ 66 static void _scsih_expander_node_remove(struct MPT3SAS_ADAPTER *ioc, 67 struct _sas_node *sas_expander); 68 static void _firmware_event_work(struct work_struct *work); 69 70 static void _scsih_remove_device(struct MPT3SAS_ADAPTER *ioc, 71 struct _sas_device *sas_device); 72 static int _scsih_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle, 73 u8 retry_count, u8 is_pd); 74 static int _scsih_pcie_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle); 75 static void _scsih_pcie_device_remove_from_sml(struct MPT3SAS_ADAPTER *ioc, 76 struct _pcie_device *pcie_device); 77 static void 78 _scsih_pcie_check_device(struct MPT3SAS_ADAPTER *ioc, u16 handle); 79 static u8 _scsih_check_for_pending_tm(struct MPT3SAS_ADAPTER *ioc, u16 smid); 80 static void _scsih_complete_devices_scanning(struct MPT3SAS_ADAPTER *ioc); 81 82 /* global parameters */ 83 LIST_HEAD(mpt3sas_ioc_list); 84 /* global ioc lock for list operations */ 85 DEFINE_SPINLOCK(gioc_lock); 86 87 MODULE_AUTHOR(MPT3SAS_AUTHOR); 88 MODULE_DESCRIPTION(MPT3SAS_DESCRIPTION); 89 MODULE_LICENSE("GPL"); 90 MODULE_VERSION(MPT3SAS_DRIVER_VERSION); 91 MODULE_ALIAS("mpt2sas"); 92 93 /* local parameters */ 94 static u8 scsi_io_cb_idx = -1; 95 static u8 tm_cb_idx = -1; 96 static u8 ctl_cb_idx = -1; 97 static u8 base_cb_idx = -1; 98 static u8 port_enable_cb_idx = -1; 99 static u8 transport_cb_idx = -1; 100 static u8 scsih_cb_idx = -1; 101 static u8 config_cb_idx = -1; 102 static int mpt2_ids; 103 static int mpt3_ids; 104 105 static u8 tm_tr_cb_idx = -1 ; 106 static u8 tm_tr_volume_cb_idx = -1 ; 107 static u8 tm_sas_control_cb_idx = -1; 108 109 /* command line options */ 110 static u32 logging_level; 111 MODULE_PARM_DESC(logging_level, 112 " bits for enabling additional logging info (default=0)"); 113 114 115 static ushort max_sectors = 0xFFFF; 116 module_param(max_sectors, ushort, 0444); 117 MODULE_PARM_DESC(max_sectors, "max sectors, range 64 to 32767 default=32767"); 118 119 120 static int missing_delay[2] = {-1, -1}; 121 module_param_array(missing_delay, int, NULL, 0444); 122 MODULE_PARM_DESC(missing_delay, " device missing delay , io missing delay"); 123 124 /* scsi-mid layer global parmeter is max_report_luns, which is 511 */ 125 #define MPT3SAS_MAX_LUN (16895) 126 static u64 max_lun = MPT3SAS_MAX_LUN; 127 module_param(max_lun, ullong, 0444); 128 MODULE_PARM_DESC(max_lun, " max lun, default=16895 "); 129 130 static ushort hbas_to_enumerate; 131 module_param(hbas_to_enumerate, ushort, 0444); 132 MODULE_PARM_DESC(hbas_to_enumerate, 133 " 0 - enumerates both SAS 2.0 & SAS 3.0 generation HBAs\n \ 134 1 - enumerates only SAS 2.0 generation HBAs\n \ 135 2 - enumerates only SAS 3.0 generation HBAs (default=0)"); 136 137 /* diag_buffer_enable is bitwise 138 * bit 0 set = TRACE 139 * bit 1 set = SNAPSHOT 140 * bit 2 set = EXTENDED 141 * 142 * Either bit can be set, or both 143 */ 144 static int diag_buffer_enable = -1; 145 module_param(diag_buffer_enable, int, 0444); 146 MODULE_PARM_DESC(diag_buffer_enable, 147 " post diag buffers (TRACE=1/SNAPSHOT=2/EXTENDED=4/default=0)"); 148 static int disable_discovery = -1; 149 module_param(disable_discovery, int, 0444); 150 MODULE_PARM_DESC(disable_discovery, " disable discovery "); 151 152 153 /* permit overriding the host protection capabilities mask (EEDP/T10 PI) */ 154 static int prot_mask = -1; 155 module_param(prot_mask, int, 0444); 156 MODULE_PARM_DESC(prot_mask, " host protection capabilities mask, def=7 "); 157 158 static bool enable_sdev_max_qd; 159 module_param(enable_sdev_max_qd, bool, 0444); 160 MODULE_PARM_DESC(enable_sdev_max_qd, 161 "Enable sdev max qd as can_queue, def=disabled(0)"); 162 163 static int multipath_on_hba = -1; 164 module_param(multipath_on_hba, int, 0); 165 MODULE_PARM_DESC(multipath_on_hba, 166 "Multipath support to add same target device\n\t\t" 167 "as many times as it is visible to HBA from various paths\n\t\t" 168 "(by default:\n\t\t" 169 "\t SAS 2.0 & SAS 3.0 HBA - This will be disabled,\n\t\t" 170 "\t SAS 3.5 HBA - This will be enabled)"); 171 172 static int host_tagset_enable = 1; 173 module_param(host_tagset_enable, int, 0444); 174 MODULE_PARM_DESC(host_tagset_enable, 175 "Shared host tagset enable/disable Default: enable(1)"); 176 177 /* raid transport support */ 178 static struct raid_template *mpt3sas_raid_template; 179 static struct raid_template *mpt2sas_raid_template; 180 181 182 /** 183 * struct sense_info - common structure for obtaining sense keys 184 * @skey: sense key 185 * @asc: additional sense code 186 * @ascq: additional sense code qualifier 187 */ 188 struct sense_info { 189 u8 skey; 190 u8 asc; 191 u8 ascq; 192 }; 193 194 #define MPT3SAS_PROCESS_TRIGGER_DIAG (0xFFFB) 195 #define MPT3SAS_TURN_ON_PFA_LED (0xFFFC) 196 #define MPT3SAS_PORT_ENABLE_COMPLETE (0xFFFD) 197 #define MPT3SAS_ABRT_TASK_SET (0xFFFE) 198 #define MPT3SAS_REMOVE_UNRESPONDING_DEVICES (0xFFFF) 199 /** 200 * struct fw_event_work - firmware event struct 201 * @list: link list framework 202 * @work: work object (ioc->fault_reset_work_q) 203 * @ioc: per adapter object 204 * @device_handle: device handle 205 * @VF_ID: virtual function id 206 * @VP_ID: virtual port id 207 * @ignore: flag meaning this event has been marked to ignore 208 * @event: firmware event MPI2_EVENT_XXX defined in mpi2_ioc.h 209 * @refcount: kref for this event 210 * @event_data: reply event data payload follows 211 * 212 * This object stored on ioc->fw_event_list. 213 */ 214 struct fw_event_work { 215 struct list_head list; 216 struct work_struct work; 217 218 struct MPT3SAS_ADAPTER *ioc; 219 u16 device_handle; 220 u8 VF_ID; 221 u8 VP_ID; 222 u8 ignore; 223 u16 event; 224 struct kref refcount; 225 char event_data[] __aligned(4); 226 }; 227 228 static void fw_event_work_free(struct kref *r) 229 { 230 kfree(container_of(r, struct fw_event_work, refcount)); 231 } 232 233 static void fw_event_work_get(struct fw_event_work *fw_work) 234 { 235 kref_get(&fw_work->refcount); 236 } 237 238 static void fw_event_work_put(struct fw_event_work *fw_work) 239 { 240 kref_put(&fw_work->refcount, fw_event_work_free); 241 } 242 243 static struct fw_event_work *alloc_fw_event_work(int len) 244 { 245 struct fw_event_work *fw_event; 246 247 fw_event = kzalloc(sizeof(*fw_event) + len, GFP_ATOMIC); 248 if (!fw_event) 249 return NULL; 250 251 kref_init(&fw_event->refcount); 252 return fw_event; 253 } 254 255 /** 256 * struct _scsi_io_transfer - scsi io transfer 257 * @handle: sas device handle (assigned by firmware) 258 * @is_raid: flag set for hidden raid components 259 * @dir: DMA_TO_DEVICE, DMA_FROM_DEVICE, 260 * @data_length: data transfer length 261 * @data_dma: dma pointer to data 262 * @sense: sense data 263 * @lun: lun number 264 * @cdb_length: cdb length 265 * @cdb: cdb contents 266 * @timeout: timeout for this command 267 * @VF_ID: virtual function id 268 * @VP_ID: virtual port id 269 * @valid_reply: flag set for reply message 270 * @sense_length: sense length 271 * @ioc_status: ioc status 272 * @scsi_state: scsi state 273 * @scsi_status: scsi staus 274 * @log_info: log information 275 * @transfer_length: data length transfer when there is a reply message 276 * 277 * Used for sending internal scsi commands to devices within this module. 278 * Refer to _scsi_send_scsi_io(). 279 */ 280 struct _scsi_io_transfer { 281 u16 handle; 282 u8 is_raid; 283 enum dma_data_direction dir; 284 u32 data_length; 285 dma_addr_t data_dma; 286 u8 sense[SCSI_SENSE_BUFFERSIZE]; 287 u32 lun; 288 u8 cdb_length; 289 u8 cdb[32]; 290 u8 timeout; 291 u8 VF_ID; 292 u8 VP_ID; 293 u8 valid_reply; 294 /* the following bits are only valid when 'valid_reply = 1' */ 295 u32 sense_length; 296 u16 ioc_status; 297 u8 scsi_state; 298 u8 scsi_status; 299 u32 log_info; 300 u32 transfer_length; 301 }; 302 303 /** 304 * _scsih_set_debug_level - global setting of ioc->logging_level. 305 * @val: value of the parameter to be set 306 * @kp: pointer to kernel_param structure 307 * 308 * Note: The logging levels are defined in mpt3sas_debug.h. 309 */ 310 static int 311 _scsih_set_debug_level(const char *val, const struct kernel_param *kp) 312 { 313 int ret = param_set_int(val, kp); 314 struct MPT3SAS_ADAPTER *ioc; 315 316 if (ret) 317 return ret; 318 319 pr_info("setting logging_level(0x%08x)\n", logging_level); 320 spin_lock(&gioc_lock); 321 list_for_each_entry(ioc, &mpt3sas_ioc_list, list) 322 ioc->logging_level = logging_level; 323 spin_unlock(&gioc_lock); 324 return 0; 325 } 326 module_param_call(logging_level, _scsih_set_debug_level, param_get_int, 327 &logging_level, 0644); 328 329 /** 330 * _scsih_srch_boot_sas_address - search based on sas_address 331 * @sas_address: sas address 332 * @boot_device: boot device object from bios page 2 333 * 334 * Return: 1 when there's a match, 0 means no match. 335 */ 336 static inline int 337 _scsih_srch_boot_sas_address(u64 sas_address, 338 Mpi2BootDeviceSasWwid_t *boot_device) 339 { 340 return (sas_address == le64_to_cpu(boot_device->SASAddress)) ? 1 : 0; 341 } 342 343 /** 344 * _scsih_srch_boot_device_name - search based on device name 345 * @device_name: device name specified in INDENTIFY fram 346 * @boot_device: boot device object from bios page 2 347 * 348 * Return: 1 when there's a match, 0 means no match. 349 */ 350 static inline int 351 _scsih_srch_boot_device_name(u64 device_name, 352 Mpi2BootDeviceDeviceName_t *boot_device) 353 { 354 return (device_name == le64_to_cpu(boot_device->DeviceName)) ? 1 : 0; 355 } 356 357 /** 358 * _scsih_srch_boot_encl_slot - search based on enclosure_logical_id/slot 359 * @enclosure_logical_id: enclosure logical id 360 * @slot_number: slot number 361 * @boot_device: boot device object from bios page 2 362 * 363 * Return: 1 when there's a match, 0 means no match. 364 */ 365 static inline int 366 _scsih_srch_boot_encl_slot(u64 enclosure_logical_id, u16 slot_number, 367 Mpi2BootDeviceEnclosureSlot_t *boot_device) 368 { 369 return (enclosure_logical_id == le64_to_cpu(boot_device-> 370 EnclosureLogicalID) && slot_number == le16_to_cpu(boot_device-> 371 SlotNumber)) ? 1 : 0; 372 } 373 374 /** 375 * mpt3sas_get_port_by_id - get hba port entry corresponding to provided 376 * port number from port list 377 * @ioc: per adapter object 378 * @port_id: port number 379 * @bypass_dirty_port_flag: when set look the matching hba port entry even 380 * if hba port entry is marked as dirty. 381 * 382 * Search for hba port entry corresponding to provided port number, 383 * if available return port object otherwise return NULL. 384 */ 385 struct hba_port * 386 mpt3sas_get_port_by_id(struct MPT3SAS_ADAPTER *ioc, 387 u8 port_id, u8 bypass_dirty_port_flag) 388 { 389 struct hba_port *port, *port_next; 390 391 /* 392 * When multipath_on_hba is disabled then 393 * search the hba_port entry using default 394 * port id i.e. 255 395 */ 396 if (!ioc->multipath_on_hba) 397 port_id = MULTIPATH_DISABLED_PORT_ID; 398 399 list_for_each_entry_safe(port, port_next, 400 &ioc->port_table_list, list) { 401 if (port->port_id != port_id) 402 continue; 403 if (bypass_dirty_port_flag) 404 return port; 405 if (port->flags & HBA_PORT_FLAG_DIRTY_PORT) 406 continue; 407 return port; 408 } 409 410 /* 411 * Allocate hba_port object for default port id (i.e. 255) 412 * when multipath_on_hba is disabled for the HBA. 413 * And add this object to port_table_list. 414 */ 415 if (!ioc->multipath_on_hba) { 416 port = kzalloc(sizeof(struct hba_port), GFP_ATOMIC); 417 if (!port) 418 return NULL; 419 420 port->port_id = port_id; 421 ioc_info(ioc, 422 "hba_port entry: %p, port: %d is added to hba_port list\n", 423 port, port->port_id); 424 list_add_tail(&port->list, 425 &ioc->port_table_list); 426 return port; 427 } 428 return NULL; 429 } 430 431 /** 432 * mpt3sas_get_vphy_by_phy - get virtual_phy object corresponding to phy number 433 * @ioc: per adapter object 434 * @port: hba_port object 435 * @phy: phy number 436 * 437 * Return virtual_phy object corresponding to phy number. 438 */ 439 struct virtual_phy * 440 mpt3sas_get_vphy_by_phy(struct MPT3SAS_ADAPTER *ioc, 441 struct hba_port *port, u32 phy) 442 { 443 struct virtual_phy *vphy, *vphy_next; 444 445 if (!port->vphys_mask) 446 return NULL; 447 448 list_for_each_entry_safe(vphy, vphy_next, &port->vphys_list, list) { 449 if (vphy->phy_mask & (1 << phy)) 450 return vphy; 451 } 452 return NULL; 453 } 454 455 /** 456 * _scsih_is_boot_device - search for matching boot device. 457 * @sas_address: sas address 458 * @device_name: device name specified in INDENTIFY fram 459 * @enclosure_logical_id: enclosure logical id 460 * @slot: slot number 461 * @form: specifies boot device form 462 * @boot_device: boot device object from bios page 2 463 * 464 * Return: 1 when there's a match, 0 means no match. 465 */ 466 static int 467 _scsih_is_boot_device(u64 sas_address, u64 device_name, 468 u64 enclosure_logical_id, u16 slot, u8 form, 469 Mpi2BiosPage2BootDevice_t *boot_device) 470 { 471 int rc = 0; 472 473 switch (form) { 474 case MPI2_BIOSPAGE2_FORM_SAS_WWID: 475 if (!sas_address) 476 break; 477 rc = _scsih_srch_boot_sas_address( 478 sas_address, &boot_device->SasWwid); 479 break; 480 case MPI2_BIOSPAGE2_FORM_ENCLOSURE_SLOT: 481 if (!enclosure_logical_id) 482 break; 483 rc = _scsih_srch_boot_encl_slot( 484 enclosure_logical_id, 485 slot, &boot_device->EnclosureSlot); 486 break; 487 case MPI2_BIOSPAGE2_FORM_DEVICE_NAME: 488 if (!device_name) 489 break; 490 rc = _scsih_srch_boot_device_name( 491 device_name, &boot_device->DeviceName); 492 break; 493 case MPI2_BIOSPAGE2_FORM_NO_DEVICE_SPECIFIED: 494 break; 495 } 496 497 return rc; 498 } 499 500 /** 501 * _scsih_get_sas_address - set the sas_address for given device handle 502 * @ioc: ? 503 * @handle: device handle 504 * @sas_address: sas address 505 * 506 * Return: 0 success, non-zero when failure 507 */ 508 static int 509 _scsih_get_sas_address(struct MPT3SAS_ADAPTER *ioc, u16 handle, 510 u64 *sas_address) 511 { 512 Mpi2SasDevicePage0_t sas_device_pg0; 513 Mpi2ConfigReply_t mpi_reply; 514 u32 ioc_status; 515 516 *sas_address = 0; 517 518 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, 519 MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { 520 ioc_err(ioc, "failure at %s:%d/%s()!\n", 521 __FILE__, __LINE__, __func__); 522 return -ENXIO; 523 } 524 525 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; 526 if (ioc_status == MPI2_IOCSTATUS_SUCCESS) { 527 /* For HBA, vSES doesn't return HBA SAS address. Instead return 528 * vSES's sas address. 529 */ 530 if ((handle <= ioc->sas_hba.num_phys) && 531 (!(le32_to_cpu(sas_device_pg0.DeviceInfo) & 532 MPI2_SAS_DEVICE_INFO_SEP))) 533 *sas_address = ioc->sas_hba.sas_address; 534 else 535 *sas_address = le64_to_cpu(sas_device_pg0.SASAddress); 536 return 0; 537 } 538 539 /* we hit this because the given parent handle doesn't exist */ 540 if (ioc_status == MPI2_IOCSTATUS_CONFIG_INVALID_PAGE) 541 return -ENXIO; 542 543 /* else error case */ 544 ioc_err(ioc, "handle(0x%04x), ioc_status(0x%04x), failure at %s:%d/%s()!\n", 545 handle, ioc_status, __FILE__, __LINE__, __func__); 546 return -EIO; 547 } 548 549 /** 550 * _scsih_determine_boot_device - determine boot device. 551 * @ioc: per adapter object 552 * @device: sas_device or pcie_device object 553 * @channel: SAS or PCIe channel 554 * 555 * Determines whether this device should be first reported device to 556 * to scsi-ml or sas transport, this purpose is for persistent boot device. 557 * There are primary, alternate, and current entries in bios page 2. The order 558 * priority is primary, alternate, then current. This routine saves 559 * the corresponding device object. 560 * The saved data to be used later in _scsih_probe_boot_devices(). 561 */ 562 static void 563 _scsih_determine_boot_device(struct MPT3SAS_ADAPTER *ioc, void *device, 564 u32 channel) 565 { 566 struct _sas_device *sas_device; 567 struct _pcie_device *pcie_device; 568 struct _raid_device *raid_device; 569 u64 sas_address; 570 u64 device_name; 571 u64 enclosure_logical_id; 572 u16 slot; 573 574 /* only process this function when driver loads */ 575 if (!ioc->is_driver_loading) 576 return; 577 578 /* no Bios, return immediately */ 579 if (!ioc->bios_pg3.BiosVersion) 580 return; 581 582 if (channel == RAID_CHANNEL) { 583 raid_device = device; 584 sas_address = raid_device->wwid; 585 device_name = 0; 586 enclosure_logical_id = 0; 587 slot = 0; 588 } else if (channel == PCIE_CHANNEL) { 589 pcie_device = device; 590 sas_address = pcie_device->wwid; 591 device_name = 0; 592 enclosure_logical_id = 0; 593 slot = 0; 594 } else { 595 sas_device = device; 596 sas_address = sas_device->sas_address; 597 device_name = sas_device->device_name; 598 enclosure_logical_id = sas_device->enclosure_logical_id; 599 slot = sas_device->slot; 600 } 601 602 if (!ioc->req_boot_device.device) { 603 if (_scsih_is_boot_device(sas_address, device_name, 604 enclosure_logical_id, slot, 605 (ioc->bios_pg2.ReqBootDeviceForm & 606 MPI2_BIOSPAGE2_FORM_MASK), 607 &ioc->bios_pg2.RequestedBootDevice)) { 608 dinitprintk(ioc, 609 ioc_info(ioc, "%s: req_boot_device(0x%016llx)\n", 610 __func__, (u64)sas_address)); 611 ioc->req_boot_device.device = device; 612 ioc->req_boot_device.channel = channel; 613 } 614 } 615 616 if (!ioc->req_alt_boot_device.device) { 617 if (_scsih_is_boot_device(sas_address, device_name, 618 enclosure_logical_id, slot, 619 (ioc->bios_pg2.ReqAltBootDeviceForm & 620 MPI2_BIOSPAGE2_FORM_MASK), 621 &ioc->bios_pg2.RequestedAltBootDevice)) { 622 dinitprintk(ioc, 623 ioc_info(ioc, "%s: req_alt_boot_device(0x%016llx)\n", 624 __func__, (u64)sas_address)); 625 ioc->req_alt_boot_device.device = device; 626 ioc->req_alt_boot_device.channel = channel; 627 } 628 } 629 630 if (!ioc->current_boot_device.device) { 631 if (_scsih_is_boot_device(sas_address, device_name, 632 enclosure_logical_id, slot, 633 (ioc->bios_pg2.CurrentBootDeviceForm & 634 MPI2_BIOSPAGE2_FORM_MASK), 635 &ioc->bios_pg2.CurrentBootDevice)) { 636 dinitprintk(ioc, 637 ioc_info(ioc, "%s: current_boot_device(0x%016llx)\n", 638 __func__, (u64)sas_address)); 639 ioc->current_boot_device.device = device; 640 ioc->current_boot_device.channel = channel; 641 } 642 } 643 } 644 645 static struct _sas_device * 646 __mpt3sas_get_sdev_from_target(struct MPT3SAS_ADAPTER *ioc, 647 struct MPT3SAS_TARGET *tgt_priv) 648 { 649 struct _sas_device *ret; 650 651 assert_spin_locked(&ioc->sas_device_lock); 652 653 ret = tgt_priv->sas_dev; 654 if (ret) 655 sas_device_get(ret); 656 657 return ret; 658 } 659 660 static struct _sas_device * 661 mpt3sas_get_sdev_from_target(struct MPT3SAS_ADAPTER *ioc, 662 struct MPT3SAS_TARGET *tgt_priv) 663 { 664 struct _sas_device *ret; 665 unsigned long flags; 666 667 spin_lock_irqsave(&ioc->sas_device_lock, flags); 668 ret = __mpt3sas_get_sdev_from_target(ioc, tgt_priv); 669 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 670 671 return ret; 672 } 673 674 static struct _pcie_device * 675 __mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER *ioc, 676 struct MPT3SAS_TARGET *tgt_priv) 677 { 678 struct _pcie_device *ret; 679 680 assert_spin_locked(&ioc->pcie_device_lock); 681 682 ret = tgt_priv->pcie_dev; 683 if (ret) 684 pcie_device_get(ret); 685 686 return ret; 687 } 688 689 /** 690 * mpt3sas_get_pdev_from_target - pcie device search 691 * @ioc: per adapter object 692 * @tgt_priv: starget private object 693 * 694 * Context: This function will acquire ioc->pcie_device_lock and will release 695 * before returning the pcie_device object. 696 * 697 * This searches for pcie_device from target, then return pcie_device object. 698 */ 699 static struct _pcie_device * 700 mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER *ioc, 701 struct MPT3SAS_TARGET *tgt_priv) 702 { 703 struct _pcie_device *ret; 704 unsigned long flags; 705 706 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 707 ret = __mpt3sas_get_pdev_from_target(ioc, tgt_priv); 708 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 709 710 return ret; 711 } 712 713 714 /** 715 * __mpt3sas_get_sdev_by_rphy - sas device search 716 * @ioc: per adapter object 717 * @rphy: sas_rphy pointer 718 * 719 * Context: This function will acquire ioc->sas_device_lock and will release 720 * before returning the sas_device object. 721 * 722 * This searches for sas_device from rphy object 723 * then return sas_device object. 724 */ 725 struct _sas_device * 726 __mpt3sas_get_sdev_by_rphy(struct MPT3SAS_ADAPTER *ioc, 727 struct sas_rphy *rphy) 728 { 729 struct _sas_device *sas_device; 730 731 assert_spin_locked(&ioc->sas_device_lock); 732 733 list_for_each_entry(sas_device, &ioc->sas_device_list, list) { 734 if (sas_device->rphy != rphy) 735 continue; 736 sas_device_get(sas_device); 737 return sas_device; 738 } 739 740 sas_device = NULL; 741 list_for_each_entry(sas_device, &ioc->sas_device_init_list, list) { 742 if (sas_device->rphy != rphy) 743 continue; 744 sas_device_get(sas_device); 745 return sas_device; 746 } 747 748 return NULL; 749 } 750 751 /** 752 * __mpt3sas_get_sdev_by_addr - get _sas_device object corresponding to provided 753 * sas address from sas_device_list list 754 * @ioc: per adapter object 755 * @sas_address: device sas address 756 * @port: port number 757 * 758 * Search for _sas_device object corresponding to provided sas address, 759 * if available return _sas_device object address otherwise return NULL. 760 */ 761 struct _sas_device * 762 __mpt3sas_get_sdev_by_addr(struct MPT3SAS_ADAPTER *ioc, 763 u64 sas_address, struct hba_port *port) 764 { 765 struct _sas_device *sas_device; 766 767 if (!port) 768 return NULL; 769 770 assert_spin_locked(&ioc->sas_device_lock); 771 772 list_for_each_entry(sas_device, &ioc->sas_device_list, list) { 773 if (sas_device->sas_address != sas_address) 774 continue; 775 if (sas_device->port != port) 776 continue; 777 sas_device_get(sas_device); 778 return sas_device; 779 } 780 781 list_for_each_entry(sas_device, &ioc->sas_device_init_list, list) { 782 if (sas_device->sas_address != sas_address) 783 continue; 784 if (sas_device->port != port) 785 continue; 786 sas_device_get(sas_device); 787 return sas_device; 788 } 789 790 return NULL; 791 } 792 793 /** 794 * mpt3sas_get_sdev_by_addr - sas device search 795 * @ioc: per adapter object 796 * @sas_address: sas address 797 * @port: hba port entry 798 * Context: Calling function should acquire ioc->sas_device_lock 799 * 800 * This searches for sas_device based on sas_address & port number, 801 * then return sas_device object. 802 */ 803 struct _sas_device * 804 mpt3sas_get_sdev_by_addr(struct MPT3SAS_ADAPTER *ioc, 805 u64 sas_address, struct hba_port *port) 806 { 807 struct _sas_device *sas_device; 808 unsigned long flags; 809 810 spin_lock_irqsave(&ioc->sas_device_lock, flags); 811 sas_device = __mpt3sas_get_sdev_by_addr(ioc, 812 sas_address, port); 813 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 814 815 return sas_device; 816 } 817 818 static struct _sas_device * 819 __mpt3sas_get_sdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) 820 { 821 struct _sas_device *sas_device; 822 823 assert_spin_locked(&ioc->sas_device_lock); 824 825 list_for_each_entry(sas_device, &ioc->sas_device_list, list) 826 if (sas_device->handle == handle) 827 goto found_device; 828 829 list_for_each_entry(sas_device, &ioc->sas_device_init_list, list) 830 if (sas_device->handle == handle) 831 goto found_device; 832 833 return NULL; 834 835 found_device: 836 sas_device_get(sas_device); 837 return sas_device; 838 } 839 840 /** 841 * mpt3sas_get_sdev_by_handle - sas device search 842 * @ioc: per adapter object 843 * @handle: sas device handle (assigned by firmware) 844 * Context: Calling function should acquire ioc->sas_device_lock 845 * 846 * This searches for sas_device based on sas_address, then return sas_device 847 * object. 848 */ 849 struct _sas_device * 850 mpt3sas_get_sdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) 851 { 852 struct _sas_device *sas_device; 853 unsigned long flags; 854 855 spin_lock_irqsave(&ioc->sas_device_lock, flags); 856 sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); 857 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 858 859 return sas_device; 860 } 861 862 /** 863 * _scsih_display_enclosure_chassis_info - display device location info 864 * @ioc: per adapter object 865 * @sas_device: per sas device object 866 * @sdev: scsi device struct 867 * @starget: scsi target struct 868 */ 869 static void 870 _scsih_display_enclosure_chassis_info(struct MPT3SAS_ADAPTER *ioc, 871 struct _sas_device *sas_device, struct scsi_device *sdev, 872 struct scsi_target *starget) 873 { 874 if (sdev) { 875 if (sas_device->enclosure_handle != 0) 876 sdev_printk(KERN_INFO, sdev, 877 "enclosure logical id (0x%016llx), slot(%d) \n", 878 (unsigned long long) 879 sas_device->enclosure_logical_id, 880 sas_device->slot); 881 if (sas_device->connector_name[0] != '\0') 882 sdev_printk(KERN_INFO, sdev, 883 "enclosure level(0x%04x), connector name( %s)\n", 884 sas_device->enclosure_level, 885 sas_device->connector_name); 886 if (sas_device->is_chassis_slot_valid) 887 sdev_printk(KERN_INFO, sdev, "chassis slot(0x%04x)\n", 888 sas_device->chassis_slot); 889 } else if (starget) { 890 if (sas_device->enclosure_handle != 0) 891 starget_printk(KERN_INFO, starget, 892 "enclosure logical id(0x%016llx), slot(%d) \n", 893 (unsigned long long) 894 sas_device->enclosure_logical_id, 895 sas_device->slot); 896 if (sas_device->connector_name[0] != '\0') 897 starget_printk(KERN_INFO, starget, 898 "enclosure level(0x%04x), connector name( %s)\n", 899 sas_device->enclosure_level, 900 sas_device->connector_name); 901 if (sas_device->is_chassis_slot_valid) 902 starget_printk(KERN_INFO, starget, 903 "chassis slot(0x%04x)\n", 904 sas_device->chassis_slot); 905 } else { 906 if (sas_device->enclosure_handle != 0) 907 ioc_info(ioc, "enclosure logical id(0x%016llx), slot(%d)\n", 908 (u64)sas_device->enclosure_logical_id, 909 sas_device->slot); 910 if (sas_device->connector_name[0] != '\0') 911 ioc_info(ioc, "enclosure level(0x%04x), connector name( %s)\n", 912 sas_device->enclosure_level, 913 sas_device->connector_name); 914 if (sas_device->is_chassis_slot_valid) 915 ioc_info(ioc, "chassis slot(0x%04x)\n", 916 sas_device->chassis_slot); 917 } 918 } 919 920 /** 921 * _scsih_sas_device_remove - remove sas_device from list. 922 * @ioc: per adapter object 923 * @sas_device: the sas_device object 924 * Context: This function will acquire ioc->sas_device_lock. 925 * 926 * If sas_device is on the list, remove it and decrement its reference count. 927 */ 928 static void 929 _scsih_sas_device_remove(struct MPT3SAS_ADAPTER *ioc, 930 struct _sas_device *sas_device) 931 { 932 unsigned long flags; 933 934 if (!sas_device) 935 return; 936 ioc_info(ioc, "removing handle(0x%04x), sas_addr(0x%016llx)\n", 937 sas_device->handle, (u64)sas_device->sas_address); 938 939 _scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL); 940 941 /* 942 * The lock serializes access to the list, but we still need to verify 943 * that nobody removed the entry while we were waiting on the lock. 944 */ 945 spin_lock_irqsave(&ioc->sas_device_lock, flags); 946 if (!list_empty(&sas_device->list)) { 947 list_del_init(&sas_device->list); 948 sas_device_put(sas_device); 949 } 950 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 951 } 952 953 /** 954 * _scsih_device_remove_by_handle - removing device object by handle 955 * @ioc: per adapter object 956 * @handle: device handle 957 */ 958 static void 959 _scsih_device_remove_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) 960 { 961 struct _sas_device *sas_device; 962 unsigned long flags; 963 964 if (ioc->shost_recovery) 965 return; 966 967 spin_lock_irqsave(&ioc->sas_device_lock, flags); 968 sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); 969 if (sas_device) { 970 list_del_init(&sas_device->list); 971 sas_device_put(sas_device); 972 } 973 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 974 if (sas_device) { 975 _scsih_remove_device(ioc, sas_device); 976 sas_device_put(sas_device); 977 } 978 } 979 980 /** 981 * mpt3sas_device_remove_by_sas_address - removing device object by 982 * sas address & port number 983 * @ioc: per adapter object 984 * @sas_address: device sas_address 985 * @port: hba port entry 986 * 987 * Return nothing. 988 */ 989 void 990 mpt3sas_device_remove_by_sas_address(struct MPT3SAS_ADAPTER *ioc, 991 u64 sas_address, struct hba_port *port) 992 { 993 struct _sas_device *sas_device; 994 unsigned long flags; 995 996 if (ioc->shost_recovery) 997 return; 998 999 spin_lock_irqsave(&ioc->sas_device_lock, flags); 1000 sas_device = __mpt3sas_get_sdev_by_addr(ioc, sas_address, port); 1001 if (sas_device) { 1002 list_del_init(&sas_device->list); 1003 sas_device_put(sas_device); 1004 } 1005 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 1006 if (sas_device) { 1007 _scsih_remove_device(ioc, sas_device); 1008 sas_device_put(sas_device); 1009 } 1010 } 1011 1012 /** 1013 * _scsih_sas_device_add - insert sas_device to the list. 1014 * @ioc: per adapter object 1015 * @sas_device: the sas_device object 1016 * Context: This function will acquire ioc->sas_device_lock. 1017 * 1018 * Adding new object to the ioc->sas_device_list. 1019 */ 1020 static void 1021 _scsih_sas_device_add(struct MPT3SAS_ADAPTER *ioc, 1022 struct _sas_device *sas_device) 1023 { 1024 unsigned long flags; 1025 1026 dewtprintk(ioc, 1027 ioc_info(ioc, "%s: handle(0x%04x), sas_addr(0x%016llx)\n", 1028 __func__, sas_device->handle, 1029 (u64)sas_device->sas_address)); 1030 1031 dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device, 1032 NULL, NULL)); 1033 1034 spin_lock_irqsave(&ioc->sas_device_lock, flags); 1035 sas_device_get(sas_device); 1036 list_add_tail(&sas_device->list, &ioc->sas_device_list); 1037 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 1038 1039 if (ioc->hide_drives) { 1040 clear_bit(sas_device->handle, ioc->pend_os_device_add); 1041 return; 1042 } 1043 1044 if (!mpt3sas_transport_port_add(ioc, sas_device->handle, 1045 sas_device->sas_address_parent, sas_device->port)) { 1046 _scsih_sas_device_remove(ioc, sas_device); 1047 } else if (!sas_device->starget) { 1048 /* 1049 * When asyn scanning is enabled, its not possible to remove 1050 * devices while scanning is turned on due to an oops in 1051 * scsi_sysfs_add_sdev()->add_device()->sysfs_addrm_start() 1052 */ 1053 if (!ioc->is_driver_loading) { 1054 mpt3sas_transport_port_remove(ioc, 1055 sas_device->sas_address, 1056 sas_device->sas_address_parent, 1057 sas_device->port); 1058 _scsih_sas_device_remove(ioc, sas_device); 1059 } 1060 } else 1061 clear_bit(sas_device->handle, ioc->pend_os_device_add); 1062 } 1063 1064 /** 1065 * _scsih_sas_device_init_add - insert sas_device to the list. 1066 * @ioc: per adapter object 1067 * @sas_device: the sas_device object 1068 * Context: This function will acquire ioc->sas_device_lock. 1069 * 1070 * Adding new object at driver load time to the ioc->sas_device_init_list. 1071 */ 1072 static void 1073 _scsih_sas_device_init_add(struct MPT3SAS_ADAPTER *ioc, 1074 struct _sas_device *sas_device) 1075 { 1076 unsigned long flags; 1077 1078 dewtprintk(ioc, 1079 ioc_info(ioc, "%s: handle(0x%04x), sas_addr(0x%016llx)\n", 1080 __func__, sas_device->handle, 1081 (u64)sas_device->sas_address)); 1082 1083 dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device, 1084 NULL, NULL)); 1085 1086 spin_lock_irqsave(&ioc->sas_device_lock, flags); 1087 sas_device_get(sas_device); 1088 list_add_tail(&sas_device->list, &ioc->sas_device_init_list); 1089 _scsih_determine_boot_device(ioc, sas_device, 0); 1090 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 1091 } 1092 1093 1094 static struct _pcie_device * 1095 __mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid) 1096 { 1097 struct _pcie_device *pcie_device; 1098 1099 assert_spin_locked(&ioc->pcie_device_lock); 1100 1101 list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) 1102 if (pcie_device->wwid == wwid) 1103 goto found_device; 1104 1105 list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list) 1106 if (pcie_device->wwid == wwid) 1107 goto found_device; 1108 1109 return NULL; 1110 1111 found_device: 1112 pcie_device_get(pcie_device); 1113 return pcie_device; 1114 } 1115 1116 1117 /** 1118 * mpt3sas_get_pdev_by_wwid - pcie device search 1119 * @ioc: per adapter object 1120 * @wwid: wwid 1121 * 1122 * Context: This function will acquire ioc->pcie_device_lock and will release 1123 * before returning the pcie_device object. 1124 * 1125 * This searches for pcie_device based on wwid, then return pcie_device object. 1126 */ 1127 static struct _pcie_device * 1128 mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid) 1129 { 1130 struct _pcie_device *pcie_device; 1131 unsigned long flags; 1132 1133 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 1134 pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid); 1135 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 1136 1137 return pcie_device; 1138 } 1139 1140 1141 static struct _pcie_device * 1142 __mpt3sas_get_pdev_by_idchannel(struct MPT3SAS_ADAPTER *ioc, int id, 1143 int channel) 1144 { 1145 struct _pcie_device *pcie_device; 1146 1147 assert_spin_locked(&ioc->pcie_device_lock); 1148 1149 list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) 1150 if (pcie_device->id == id && pcie_device->channel == channel) 1151 goto found_device; 1152 1153 list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list) 1154 if (pcie_device->id == id && pcie_device->channel == channel) 1155 goto found_device; 1156 1157 return NULL; 1158 1159 found_device: 1160 pcie_device_get(pcie_device); 1161 return pcie_device; 1162 } 1163 1164 static struct _pcie_device * 1165 __mpt3sas_get_pdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) 1166 { 1167 struct _pcie_device *pcie_device; 1168 1169 assert_spin_locked(&ioc->pcie_device_lock); 1170 1171 list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) 1172 if (pcie_device->handle == handle) 1173 goto found_device; 1174 1175 list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list) 1176 if (pcie_device->handle == handle) 1177 goto found_device; 1178 1179 return NULL; 1180 1181 found_device: 1182 pcie_device_get(pcie_device); 1183 return pcie_device; 1184 } 1185 1186 1187 /** 1188 * mpt3sas_get_pdev_by_handle - pcie device search 1189 * @ioc: per adapter object 1190 * @handle: Firmware device handle 1191 * 1192 * Context: This function will acquire ioc->pcie_device_lock and will release 1193 * before returning the pcie_device object. 1194 * 1195 * This searches for pcie_device based on handle, then return pcie_device 1196 * object. 1197 */ 1198 struct _pcie_device * 1199 mpt3sas_get_pdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) 1200 { 1201 struct _pcie_device *pcie_device; 1202 unsigned long flags; 1203 1204 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 1205 pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle); 1206 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 1207 1208 return pcie_device; 1209 } 1210 1211 /** 1212 * _scsih_set_nvme_max_shutdown_latency - Update max_shutdown_latency. 1213 * @ioc: per adapter object 1214 * Context: This function will acquire ioc->pcie_device_lock 1215 * 1216 * Update ioc->max_shutdown_latency to that NVMe drives RTD3 Entry Latency 1217 * which has reported maximum among all available NVMe drives. 1218 * Minimum max_shutdown_latency will be six seconds. 1219 */ 1220 static void 1221 _scsih_set_nvme_max_shutdown_latency(struct MPT3SAS_ADAPTER *ioc) 1222 { 1223 struct _pcie_device *pcie_device; 1224 unsigned long flags; 1225 u16 shutdown_latency = IO_UNIT_CONTROL_SHUTDOWN_TIMEOUT; 1226 1227 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 1228 list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) { 1229 if (pcie_device->shutdown_latency) { 1230 if (shutdown_latency < pcie_device->shutdown_latency) 1231 shutdown_latency = 1232 pcie_device->shutdown_latency; 1233 } 1234 } 1235 ioc->max_shutdown_latency = shutdown_latency; 1236 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 1237 } 1238 1239 /** 1240 * _scsih_pcie_device_remove - remove pcie_device from list. 1241 * @ioc: per adapter object 1242 * @pcie_device: the pcie_device object 1243 * Context: This function will acquire ioc->pcie_device_lock. 1244 * 1245 * If pcie_device is on the list, remove it and decrement its reference count. 1246 */ 1247 static void 1248 _scsih_pcie_device_remove(struct MPT3SAS_ADAPTER *ioc, 1249 struct _pcie_device *pcie_device) 1250 { 1251 unsigned long flags; 1252 int was_on_pcie_device_list = 0; 1253 u8 update_latency = 0; 1254 1255 if (!pcie_device) 1256 return; 1257 ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n", 1258 pcie_device->handle, (u64)pcie_device->wwid); 1259 if (pcie_device->enclosure_handle != 0) 1260 ioc_info(ioc, "removing enclosure logical id(0x%016llx), slot(%d)\n", 1261 (u64)pcie_device->enclosure_logical_id, 1262 pcie_device->slot); 1263 if (pcie_device->connector_name[0] != '\0') 1264 ioc_info(ioc, "removing enclosure level(0x%04x), connector name( %s)\n", 1265 pcie_device->enclosure_level, 1266 pcie_device->connector_name); 1267 1268 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 1269 if (!list_empty(&pcie_device->list)) { 1270 list_del_init(&pcie_device->list); 1271 was_on_pcie_device_list = 1; 1272 } 1273 if (pcie_device->shutdown_latency == ioc->max_shutdown_latency) 1274 update_latency = 1; 1275 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 1276 if (was_on_pcie_device_list) { 1277 kfree(pcie_device->serial_number); 1278 pcie_device_put(pcie_device); 1279 } 1280 1281 /* 1282 * This device's RTD3 Entry Latency matches IOC's 1283 * max_shutdown_latency. Recalculate IOC's max_shutdown_latency 1284 * from the available drives as current drive is getting removed. 1285 */ 1286 if (update_latency) 1287 _scsih_set_nvme_max_shutdown_latency(ioc); 1288 } 1289 1290 1291 /** 1292 * _scsih_pcie_device_remove_by_handle - removing pcie device object by handle 1293 * @ioc: per adapter object 1294 * @handle: device handle 1295 */ 1296 static void 1297 _scsih_pcie_device_remove_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) 1298 { 1299 struct _pcie_device *pcie_device; 1300 unsigned long flags; 1301 int was_on_pcie_device_list = 0; 1302 u8 update_latency = 0; 1303 1304 if (ioc->shost_recovery) 1305 return; 1306 1307 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 1308 pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle); 1309 if (pcie_device) { 1310 if (!list_empty(&pcie_device->list)) { 1311 list_del_init(&pcie_device->list); 1312 was_on_pcie_device_list = 1; 1313 pcie_device_put(pcie_device); 1314 } 1315 if (pcie_device->shutdown_latency == ioc->max_shutdown_latency) 1316 update_latency = 1; 1317 } 1318 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 1319 if (was_on_pcie_device_list) { 1320 _scsih_pcie_device_remove_from_sml(ioc, pcie_device); 1321 pcie_device_put(pcie_device); 1322 } 1323 1324 /* 1325 * This device's RTD3 Entry Latency matches IOC's 1326 * max_shutdown_latency. Recalculate IOC's max_shutdown_latency 1327 * from the available drives as current drive is getting removed. 1328 */ 1329 if (update_latency) 1330 _scsih_set_nvme_max_shutdown_latency(ioc); 1331 } 1332 1333 /** 1334 * _scsih_pcie_device_add - add pcie_device object 1335 * @ioc: per adapter object 1336 * @pcie_device: pcie_device object 1337 * 1338 * This is added to the pcie_device_list link list. 1339 */ 1340 static void 1341 _scsih_pcie_device_add(struct MPT3SAS_ADAPTER *ioc, 1342 struct _pcie_device *pcie_device) 1343 { 1344 unsigned long flags; 1345 1346 dewtprintk(ioc, 1347 ioc_info(ioc, "%s: handle (0x%04x), wwid(0x%016llx)\n", 1348 __func__, 1349 pcie_device->handle, (u64)pcie_device->wwid)); 1350 if (pcie_device->enclosure_handle != 0) 1351 dewtprintk(ioc, 1352 ioc_info(ioc, "%s: enclosure logical id(0x%016llx), slot( %d)\n", 1353 __func__, 1354 (u64)pcie_device->enclosure_logical_id, 1355 pcie_device->slot)); 1356 if (pcie_device->connector_name[0] != '\0') 1357 dewtprintk(ioc, 1358 ioc_info(ioc, "%s: enclosure level(0x%04x), connector name( %s)\n", 1359 __func__, pcie_device->enclosure_level, 1360 pcie_device->connector_name)); 1361 1362 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 1363 pcie_device_get(pcie_device); 1364 list_add_tail(&pcie_device->list, &ioc->pcie_device_list); 1365 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 1366 1367 if (pcie_device->access_status == 1368 MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED) { 1369 clear_bit(pcie_device->handle, ioc->pend_os_device_add); 1370 return; 1371 } 1372 if (scsi_add_device(ioc->shost, PCIE_CHANNEL, pcie_device->id, 0)) { 1373 _scsih_pcie_device_remove(ioc, pcie_device); 1374 } else if (!pcie_device->starget) { 1375 if (!ioc->is_driver_loading) { 1376 /*TODO-- Need to find out whether this condition will occur or not*/ 1377 clear_bit(pcie_device->handle, ioc->pend_os_device_add); 1378 } 1379 } else 1380 clear_bit(pcie_device->handle, ioc->pend_os_device_add); 1381 } 1382 1383 /* 1384 * _scsih_pcie_device_init_add - insert pcie_device to the init list. 1385 * @ioc: per adapter object 1386 * @pcie_device: the pcie_device object 1387 * Context: This function will acquire ioc->pcie_device_lock. 1388 * 1389 * Adding new object at driver load time to the ioc->pcie_device_init_list. 1390 */ 1391 static void 1392 _scsih_pcie_device_init_add(struct MPT3SAS_ADAPTER *ioc, 1393 struct _pcie_device *pcie_device) 1394 { 1395 unsigned long flags; 1396 1397 dewtprintk(ioc, 1398 ioc_info(ioc, "%s: handle (0x%04x), wwid(0x%016llx)\n", 1399 __func__, 1400 pcie_device->handle, (u64)pcie_device->wwid)); 1401 if (pcie_device->enclosure_handle != 0) 1402 dewtprintk(ioc, 1403 ioc_info(ioc, "%s: enclosure logical id(0x%016llx), slot( %d)\n", 1404 __func__, 1405 (u64)pcie_device->enclosure_logical_id, 1406 pcie_device->slot)); 1407 if (pcie_device->connector_name[0] != '\0') 1408 dewtprintk(ioc, 1409 ioc_info(ioc, "%s: enclosure level(0x%04x), connector name( %s)\n", 1410 __func__, pcie_device->enclosure_level, 1411 pcie_device->connector_name)); 1412 1413 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 1414 pcie_device_get(pcie_device); 1415 list_add_tail(&pcie_device->list, &ioc->pcie_device_init_list); 1416 if (pcie_device->access_status != 1417 MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED) 1418 _scsih_determine_boot_device(ioc, pcie_device, PCIE_CHANNEL); 1419 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 1420 } 1421 /** 1422 * _scsih_raid_device_find_by_id - raid device search 1423 * @ioc: per adapter object 1424 * @id: sas device target id 1425 * @channel: sas device channel 1426 * Context: Calling function should acquire ioc->raid_device_lock 1427 * 1428 * This searches for raid_device based on target id, then return raid_device 1429 * object. 1430 */ 1431 static struct _raid_device * 1432 _scsih_raid_device_find_by_id(struct MPT3SAS_ADAPTER *ioc, int id, int channel) 1433 { 1434 struct _raid_device *raid_device, *r; 1435 1436 r = NULL; 1437 list_for_each_entry(raid_device, &ioc->raid_device_list, list) { 1438 if (raid_device->id == id && raid_device->channel == channel) { 1439 r = raid_device; 1440 goto out; 1441 } 1442 } 1443 1444 out: 1445 return r; 1446 } 1447 1448 /** 1449 * mpt3sas_raid_device_find_by_handle - raid device search 1450 * @ioc: per adapter object 1451 * @handle: sas device handle (assigned by firmware) 1452 * Context: Calling function should acquire ioc->raid_device_lock 1453 * 1454 * This searches for raid_device based on handle, then return raid_device 1455 * object. 1456 */ 1457 struct _raid_device * 1458 mpt3sas_raid_device_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) 1459 { 1460 struct _raid_device *raid_device, *r; 1461 1462 r = NULL; 1463 list_for_each_entry(raid_device, &ioc->raid_device_list, list) { 1464 if (raid_device->handle != handle) 1465 continue; 1466 r = raid_device; 1467 goto out; 1468 } 1469 1470 out: 1471 return r; 1472 } 1473 1474 /** 1475 * _scsih_raid_device_find_by_wwid - raid device search 1476 * @ioc: per adapter object 1477 * @wwid: ? 1478 * Context: Calling function should acquire ioc->raid_device_lock 1479 * 1480 * This searches for raid_device based on wwid, then return raid_device 1481 * object. 1482 */ 1483 static struct _raid_device * 1484 _scsih_raid_device_find_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid) 1485 { 1486 struct _raid_device *raid_device, *r; 1487 1488 r = NULL; 1489 list_for_each_entry(raid_device, &ioc->raid_device_list, list) { 1490 if (raid_device->wwid != wwid) 1491 continue; 1492 r = raid_device; 1493 goto out; 1494 } 1495 1496 out: 1497 return r; 1498 } 1499 1500 /** 1501 * _scsih_raid_device_add - add raid_device object 1502 * @ioc: per adapter object 1503 * @raid_device: raid_device object 1504 * 1505 * This is added to the raid_device_list link list. 1506 */ 1507 static void 1508 _scsih_raid_device_add(struct MPT3SAS_ADAPTER *ioc, 1509 struct _raid_device *raid_device) 1510 { 1511 unsigned long flags; 1512 1513 dewtprintk(ioc, 1514 ioc_info(ioc, "%s: handle(0x%04x), wwid(0x%016llx)\n", 1515 __func__, 1516 raid_device->handle, (u64)raid_device->wwid)); 1517 1518 spin_lock_irqsave(&ioc->raid_device_lock, flags); 1519 list_add_tail(&raid_device->list, &ioc->raid_device_list); 1520 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 1521 } 1522 1523 /** 1524 * _scsih_raid_device_remove - delete raid_device object 1525 * @ioc: per adapter object 1526 * @raid_device: raid_device object 1527 * 1528 */ 1529 static void 1530 _scsih_raid_device_remove(struct MPT3SAS_ADAPTER *ioc, 1531 struct _raid_device *raid_device) 1532 { 1533 unsigned long flags; 1534 1535 spin_lock_irqsave(&ioc->raid_device_lock, flags); 1536 list_del(&raid_device->list); 1537 kfree(raid_device); 1538 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 1539 } 1540 1541 /** 1542 * mpt3sas_scsih_expander_find_by_handle - expander device search 1543 * @ioc: per adapter object 1544 * @handle: expander handle (assigned by firmware) 1545 * Context: Calling function should acquire ioc->sas_device_lock 1546 * 1547 * This searches for expander device based on handle, then returns the 1548 * sas_node object. 1549 */ 1550 struct _sas_node * 1551 mpt3sas_scsih_expander_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) 1552 { 1553 struct _sas_node *sas_expander, *r; 1554 1555 r = NULL; 1556 list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) { 1557 if (sas_expander->handle != handle) 1558 continue; 1559 r = sas_expander; 1560 goto out; 1561 } 1562 out: 1563 return r; 1564 } 1565 1566 /** 1567 * mpt3sas_scsih_enclosure_find_by_handle - exclosure device search 1568 * @ioc: per adapter object 1569 * @handle: enclosure handle (assigned by firmware) 1570 * Context: Calling function should acquire ioc->sas_device_lock 1571 * 1572 * This searches for enclosure device based on handle, then returns the 1573 * enclosure object. 1574 */ 1575 static struct _enclosure_node * 1576 mpt3sas_scsih_enclosure_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) 1577 { 1578 struct _enclosure_node *enclosure_dev, *r; 1579 1580 r = NULL; 1581 list_for_each_entry(enclosure_dev, &ioc->enclosure_list, list) { 1582 if (le16_to_cpu(enclosure_dev->pg0.EnclosureHandle) != handle) 1583 continue; 1584 r = enclosure_dev; 1585 goto out; 1586 } 1587 out: 1588 return r; 1589 } 1590 /** 1591 * mpt3sas_scsih_expander_find_by_sas_address - expander device search 1592 * @ioc: per adapter object 1593 * @sas_address: sas address 1594 * @port: hba port entry 1595 * Context: Calling function should acquire ioc->sas_node_lock. 1596 * 1597 * This searches for expander device based on sas_address & port number, 1598 * then returns the sas_node object. 1599 */ 1600 struct _sas_node * 1601 mpt3sas_scsih_expander_find_by_sas_address(struct MPT3SAS_ADAPTER *ioc, 1602 u64 sas_address, struct hba_port *port) 1603 { 1604 struct _sas_node *sas_expander, *r = NULL; 1605 1606 if (!port) 1607 return r; 1608 1609 list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) { 1610 if (sas_expander->sas_address != sas_address) 1611 continue; 1612 if (sas_expander->port != port) 1613 continue; 1614 r = sas_expander; 1615 goto out; 1616 } 1617 out: 1618 return r; 1619 } 1620 1621 /** 1622 * _scsih_expander_node_add - insert expander device to the list. 1623 * @ioc: per adapter object 1624 * @sas_expander: the sas_device object 1625 * Context: This function will acquire ioc->sas_node_lock. 1626 * 1627 * Adding new object to the ioc->sas_expander_list. 1628 */ 1629 static void 1630 _scsih_expander_node_add(struct MPT3SAS_ADAPTER *ioc, 1631 struct _sas_node *sas_expander) 1632 { 1633 unsigned long flags; 1634 1635 spin_lock_irqsave(&ioc->sas_node_lock, flags); 1636 list_add_tail(&sas_expander->list, &ioc->sas_expander_list); 1637 spin_unlock_irqrestore(&ioc->sas_node_lock, flags); 1638 } 1639 1640 /** 1641 * _scsih_is_end_device - determines if device is an end device 1642 * @device_info: bitfield providing information about the device. 1643 * Context: none 1644 * 1645 * Return: 1 if end device. 1646 */ 1647 static int 1648 _scsih_is_end_device(u32 device_info) 1649 { 1650 if (device_info & MPI2_SAS_DEVICE_INFO_END_DEVICE && 1651 ((device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET) | 1652 (device_info & MPI2_SAS_DEVICE_INFO_STP_TARGET) | 1653 (device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE))) 1654 return 1; 1655 else 1656 return 0; 1657 } 1658 1659 /** 1660 * _scsih_is_nvme_pciescsi_device - determines if 1661 * device is an pcie nvme/scsi device 1662 * @device_info: bitfield providing information about the device. 1663 * Context: none 1664 * 1665 * Returns 1 if device is pcie device type nvme/scsi. 1666 */ 1667 static int 1668 _scsih_is_nvme_pciescsi_device(u32 device_info) 1669 { 1670 if (((device_info & MPI26_PCIE_DEVINFO_MASK_DEVICE_TYPE) 1671 == MPI26_PCIE_DEVINFO_NVME) || 1672 ((device_info & MPI26_PCIE_DEVINFO_MASK_DEVICE_TYPE) 1673 == MPI26_PCIE_DEVINFO_SCSI)) 1674 return 1; 1675 else 1676 return 0; 1677 } 1678 1679 /** 1680 * _scsih_scsi_lookup_find_by_target - search for matching channel:id 1681 * @ioc: per adapter object 1682 * @id: target id 1683 * @channel: channel 1684 * Context: This function will acquire ioc->scsi_lookup_lock. 1685 * 1686 * This will search for a matching channel:id in the scsi_lookup array, 1687 * returning 1 if found. 1688 */ 1689 static u8 1690 _scsih_scsi_lookup_find_by_target(struct MPT3SAS_ADAPTER *ioc, int id, 1691 int channel) 1692 { 1693 int smid; 1694 struct scsi_cmnd *scmd; 1695 1696 for (smid = 1; 1697 smid <= ioc->shost->can_queue; smid++) { 1698 scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); 1699 if (!scmd) 1700 continue; 1701 if (scmd->device->id == id && 1702 scmd->device->channel == channel) 1703 return 1; 1704 } 1705 return 0; 1706 } 1707 1708 /** 1709 * _scsih_scsi_lookup_find_by_lun - search for matching channel:id:lun 1710 * @ioc: per adapter object 1711 * @id: target id 1712 * @lun: lun number 1713 * @channel: channel 1714 * Context: This function will acquire ioc->scsi_lookup_lock. 1715 * 1716 * This will search for a matching channel:id:lun in the scsi_lookup array, 1717 * returning 1 if found. 1718 */ 1719 static u8 1720 _scsih_scsi_lookup_find_by_lun(struct MPT3SAS_ADAPTER *ioc, int id, 1721 unsigned int lun, int channel) 1722 { 1723 int smid; 1724 struct scsi_cmnd *scmd; 1725 1726 for (smid = 1; smid <= ioc->shost->can_queue; smid++) { 1727 1728 scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); 1729 if (!scmd) 1730 continue; 1731 if (scmd->device->id == id && 1732 scmd->device->channel == channel && 1733 scmd->device->lun == lun) 1734 return 1; 1735 } 1736 return 0; 1737 } 1738 1739 /** 1740 * mpt3sas_scsih_scsi_lookup_get - returns scmd entry 1741 * @ioc: per adapter object 1742 * @smid: system request message index 1743 * 1744 * Return: the smid stored scmd pointer. 1745 * Then will dereference the stored scmd pointer. 1746 */ 1747 struct scsi_cmnd * 1748 mpt3sas_scsih_scsi_lookup_get(struct MPT3SAS_ADAPTER *ioc, u16 smid) 1749 { 1750 struct scsi_cmnd *scmd = NULL; 1751 struct scsiio_tracker *st; 1752 Mpi25SCSIIORequest_t *mpi_request; 1753 u16 tag = smid - 1; 1754 1755 if (smid > 0 && 1756 smid <= ioc->scsiio_depth - INTERNAL_SCSIIO_CMDS_COUNT) { 1757 u32 unique_tag = 1758 ioc->io_queue_num[tag] << BLK_MQ_UNIQUE_TAG_BITS | tag; 1759 1760 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); 1761 1762 /* 1763 * If SCSI IO request is outstanding at driver level then 1764 * DevHandle filed must be non-zero. If DevHandle is zero 1765 * then it means that this smid is free at driver level, 1766 * so return NULL. 1767 */ 1768 if (!mpi_request->DevHandle) 1769 return scmd; 1770 1771 scmd = scsi_host_find_tag(ioc->shost, unique_tag); 1772 if (scmd) { 1773 st = scsi_cmd_priv(scmd); 1774 if (st->cb_idx == 0xFF || st->smid == 0) 1775 scmd = NULL; 1776 } 1777 } 1778 return scmd; 1779 } 1780 1781 /** 1782 * scsih_change_queue_depth - setting device queue depth 1783 * @sdev: scsi device struct 1784 * @qdepth: requested queue depth 1785 * 1786 * Return: queue depth. 1787 */ 1788 static int 1789 scsih_change_queue_depth(struct scsi_device *sdev, int qdepth) 1790 { 1791 struct Scsi_Host *shost = sdev->host; 1792 int max_depth; 1793 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); 1794 struct MPT3SAS_DEVICE *sas_device_priv_data; 1795 struct MPT3SAS_TARGET *sas_target_priv_data; 1796 struct _sas_device *sas_device; 1797 unsigned long flags; 1798 1799 max_depth = shost->can_queue; 1800 1801 /* 1802 * limit max device queue for SATA to 32 if enable_sdev_max_qd 1803 * is disabled. 1804 */ 1805 if (ioc->enable_sdev_max_qd || ioc->is_gen35_ioc) 1806 goto not_sata; 1807 1808 sas_device_priv_data = sdev->hostdata; 1809 if (!sas_device_priv_data) 1810 goto not_sata; 1811 sas_target_priv_data = sas_device_priv_data->sas_target; 1812 if (!sas_target_priv_data) 1813 goto not_sata; 1814 if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)) 1815 goto not_sata; 1816 1817 spin_lock_irqsave(&ioc->sas_device_lock, flags); 1818 sas_device = __mpt3sas_get_sdev_from_target(ioc, sas_target_priv_data); 1819 if (sas_device) { 1820 if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE) 1821 max_depth = MPT3SAS_SATA_QUEUE_DEPTH; 1822 1823 sas_device_put(sas_device); 1824 } 1825 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 1826 1827 not_sata: 1828 1829 if (!sdev->tagged_supported) 1830 max_depth = 1; 1831 if (qdepth > max_depth) 1832 qdepth = max_depth; 1833 scsi_change_queue_depth(sdev, qdepth); 1834 sdev_printk(KERN_INFO, sdev, 1835 "qdepth(%d), tagged(%d), scsi_level(%d), cmd_que(%d)\n", 1836 sdev->queue_depth, sdev->tagged_supported, 1837 sdev->scsi_level, ((sdev->inquiry[7] & 2) >> 1)); 1838 return sdev->queue_depth; 1839 } 1840 1841 /** 1842 * mpt3sas_scsih_change_queue_depth - setting device queue depth 1843 * @sdev: scsi device struct 1844 * @qdepth: requested queue depth 1845 * 1846 * Returns nothing. 1847 */ 1848 void 1849 mpt3sas_scsih_change_queue_depth(struct scsi_device *sdev, int qdepth) 1850 { 1851 struct Scsi_Host *shost = sdev->host; 1852 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); 1853 1854 if (ioc->enable_sdev_max_qd) 1855 qdepth = shost->can_queue; 1856 1857 scsih_change_queue_depth(sdev, qdepth); 1858 } 1859 1860 /** 1861 * scsih_target_alloc - target add routine 1862 * @starget: scsi target struct 1863 * 1864 * Return: 0 if ok. Any other return is assumed to be an error and 1865 * the device is ignored. 1866 */ 1867 static int 1868 scsih_target_alloc(struct scsi_target *starget) 1869 { 1870 struct Scsi_Host *shost = dev_to_shost(&starget->dev); 1871 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); 1872 struct MPT3SAS_TARGET *sas_target_priv_data; 1873 struct _sas_device *sas_device; 1874 struct _raid_device *raid_device; 1875 struct _pcie_device *pcie_device; 1876 unsigned long flags; 1877 struct sas_rphy *rphy; 1878 1879 sas_target_priv_data = kzalloc(sizeof(*sas_target_priv_data), 1880 GFP_KERNEL); 1881 if (!sas_target_priv_data) 1882 return -ENOMEM; 1883 1884 starget->hostdata = sas_target_priv_data; 1885 sas_target_priv_data->starget = starget; 1886 sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE; 1887 1888 /* RAID volumes */ 1889 if (starget->channel == RAID_CHANNEL) { 1890 spin_lock_irqsave(&ioc->raid_device_lock, flags); 1891 raid_device = _scsih_raid_device_find_by_id(ioc, starget->id, 1892 starget->channel); 1893 if (raid_device) { 1894 sas_target_priv_data->handle = raid_device->handle; 1895 sas_target_priv_data->sas_address = raid_device->wwid; 1896 sas_target_priv_data->flags |= MPT_TARGET_FLAGS_VOLUME; 1897 if (ioc->is_warpdrive) 1898 sas_target_priv_data->raid_device = raid_device; 1899 raid_device->starget = starget; 1900 } 1901 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 1902 return 0; 1903 } 1904 1905 /* PCIe devices */ 1906 if (starget->channel == PCIE_CHANNEL) { 1907 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 1908 pcie_device = __mpt3sas_get_pdev_by_idchannel(ioc, starget->id, 1909 starget->channel); 1910 if (pcie_device) { 1911 sas_target_priv_data->handle = pcie_device->handle; 1912 sas_target_priv_data->sas_address = pcie_device->wwid; 1913 sas_target_priv_data->port = NULL; 1914 sas_target_priv_data->pcie_dev = pcie_device; 1915 pcie_device->starget = starget; 1916 pcie_device->id = starget->id; 1917 pcie_device->channel = starget->channel; 1918 sas_target_priv_data->flags |= 1919 MPT_TARGET_FLAGS_PCIE_DEVICE; 1920 if (pcie_device->fast_path) 1921 sas_target_priv_data->flags |= 1922 MPT_TARGET_FASTPATH_IO; 1923 } 1924 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 1925 return 0; 1926 } 1927 1928 /* sas/sata devices */ 1929 spin_lock_irqsave(&ioc->sas_device_lock, flags); 1930 rphy = dev_to_rphy(starget->dev.parent); 1931 sas_device = __mpt3sas_get_sdev_by_rphy(ioc, rphy); 1932 1933 if (sas_device) { 1934 sas_target_priv_data->handle = sas_device->handle; 1935 sas_target_priv_data->sas_address = sas_device->sas_address; 1936 sas_target_priv_data->port = sas_device->port; 1937 sas_target_priv_data->sas_dev = sas_device; 1938 sas_device->starget = starget; 1939 sas_device->id = starget->id; 1940 sas_device->channel = starget->channel; 1941 if (test_bit(sas_device->handle, ioc->pd_handles)) 1942 sas_target_priv_data->flags |= 1943 MPT_TARGET_FLAGS_RAID_COMPONENT; 1944 if (sas_device->fast_path) 1945 sas_target_priv_data->flags |= 1946 MPT_TARGET_FASTPATH_IO; 1947 } 1948 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 1949 1950 return 0; 1951 } 1952 1953 /** 1954 * scsih_target_destroy - target destroy routine 1955 * @starget: scsi target struct 1956 */ 1957 static void 1958 scsih_target_destroy(struct scsi_target *starget) 1959 { 1960 struct Scsi_Host *shost = dev_to_shost(&starget->dev); 1961 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); 1962 struct MPT3SAS_TARGET *sas_target_priv_data; 1963 struct _sas_device *sas_device; 1964 struct _raid_device *raid_device; 1965 struct _pcie_device *pcie_device; 1966 unsigned long flags; 1967 1968 sas_target_priv_data = starget->hostdata; 1969 if (!sas_target_priv_data) 1970 return; 1971 1972 if (starget->channel == RAID_CHANNEL) { 1973 spin_lock_irqsave(&ioc->raid_device_lock, flags); 1974 raid_device = _scsih_raid_device_find_by_id(ioc, starget->id, 1975 starget->channel); 1976 if (raid_device) { 1977 raid_device->starget = NULL; 1978 raid_device->sdev = NULL; 1979 } 1980 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 1981 goto out; 1982 } 1983 1984 if (starget->channel == PCIE_CHANNEL) { 1985 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 1986 pcie_device = __mpt3sas_get_pdev_from_target(ioc, 1987 sas_target_priv_data); 1988 if (pcie_device && (pcie_device->starget == starget) && 1989 (pcie_device->id == starget->id) && 1990 (pcie_device->channel == starget->channel)) 1991 pcie_device->starget = NULL; 1992 1993 if (pcie_device) { 1994 /* 1995 * Corresponding get() is in _scsih_target_alloc() 1996 */ 1997 sas_target_priv_data->pcie_dev = NULL; 1998 pcie_device_put(pcie_device); 1999 pcie_device_put(pcie_device); 2000 } 2001 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 2002 goto out; 2003 } 2004 2005 spin_lock_irqsave(&ioc->sas_device_lock, flags); 2006 sas_device = __mpt3sas_get_sdev_from_target(ioc, sas_target_priv_data); 2007 if (sas_device && (sas_device->starget == starget) && 2008 (sas_device->id == starget->id) && 2009 (sas_device->channel == starget->channel)) 2010 sas_device->starget = NULL; 2011 2012 if (sas_device) { 2013 /* 2014 * Corresponding get() is in _scsih_target_alloc() 2015 */ 2016 sas_target_priv_data->sas_dev = NULL; 2017 sas_device_put(sas_device); 2018 2019 sas_device_put(sas_device); 2020 } 2021 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 2022 2023 out: 2024 kfree(sas_target_priv_data); 2025 starget->hostdata = NULL; 2026 } 2027 2028 /** 2029 * scsih_slave_alloc - device add routine 2030 * @sdev: scsi device struct 2031 * 2032 * Return: 0 if ok. Any other return is assumed to be an error and 2033 * the device is ignored. 2034 */ 2035 static int 2036 scsih_slave_alloc(struct scsi_device *sdev) 2037 { 2038 struct Scsi_Host *shost; 2039 struct MPT3SAS_ADAPTER *ioc; 2040 struct MPT3SAS_TARGET *sas_target_priv_data; 2041 struct MPT3SAS_DEVICE *sas_device_priv_data; 2042 struct scsi_target *starget; 2043 struct _raid_device *raid_device; 2044 struct _sas_device *sas_device; 2045 struct _pcie_device *pcie_device; 2046 unsigned long flags; 2047 2048 sas_device_priv_data = kzalloc(sizeof(*sas_device_priv_data), 2049 GFP_KERNEL); 2050 if (!sas_device_priv_data) 2051 return -ENOMEM; 2052 2053 sas_device_priv_data->lun = sdev->lun; 2054 sas_device_priv_data->flags = MPT_DEVICE_FLAGS_INIT; 2055 2056 starget = scsi_target(sdev); 2057 sas_target_priv_data = starget->hostdata; 2058 sas_target_priv_data->num_luns++; 2059 sas_device_priv_data->sas_target = sas_target_priv_data; 2060 sdev->hostdata = sas_device_priv_data; 2061 if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT)) 2062 sdev->no_uld_attach = 1; 2063 2064 shost = dev_to_shost(&starget->dev); 2065 ioc = shost_priv(shost); 2066 if (starget->channel == RAID_CHANNEL) { 2067 spin_lock_irqsave(&ioc->raid_device_lock, flags); 2068 raid_device = _scsih_raid_device_find_by_id(ioc, 2069 starget->id, starget->channel); 2070 if (raid_device) 2071 raid_device->sdev = sdev; /* raid is single lun */ 2072 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 2073 } 2074 if (starget->channel == PCIE_CHANNEL) { 2075 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 2076 pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, 2077 sas_target_priv_data->sas_address); 2078 if (pcie_device && (pcie_device->starget == NULL)) { 2079 sdev_printk(KERN_INFO, sdev, 2080 "%s : pcie_device->starget set to starget @ %d\n", 2081 __func__, __LINE__); 2082 pcie_device->starget = starget; 2083 } 2084 2085 if (pcie_device) 2086 pcie_device_put(pcie_device); 2087 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 2088 2089 } else if (!(sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)) { 2090 spin_lock_irqsave(&ioc->sas_device_lock, flags); 2091 sas_device = __mpt3sas_get_sdev_by_addr(ioc, 2092 sas_target_priv_data->sas_address, 2093 sas_target_priv_data->port); 2094 if (sas_device && (sas_device->starget == NULL)) { 2095 sdev_printk(KERN_INFO, sdev, 2096 "%s : sas_device->starget set to starget @ %d\n", 2097 __func__, __LINE__); 2098 sas_device->starget = starget; 2099 } 2100 2101 if (sas_device) 2102 sas_device_put(sas_device); 2103 2104 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 2105 } 2106 2107 return 0; 2108 } 2109 2110 /** 2111 * scsih_slave_destroy - device destroy routine 2112 * @sdev: scsi device struct 2113 */ 2114 static void 2115 scsih_slave_destroy(struct scsi_device *sdev) 2116 { 2117 struct MPT3SAS_TARGET *sas_target_priv_data; 2118 struct scsi_target *starget; 2119 struct Scsi_Host *shost; 2120 struct MPT3SAS_ADAPTER *ioc; 2121 struct _sas_device *sas_device; 2122 struct _pcie_device *pcie_device; 2123 unsigned long flags; 2124 2125 if (!sdev->hostdata) 2126 return; 2127 2128 starget = scsi_target(sdev); 2129 sas_target_priv_data = starget->hostdata; 2130 sas_target_priv_data->num_luns--; 2131 2132 shost = dev_to_shost(&starget->dev); 2133 ioc = shost_priv(shost); 2134 2135 if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) { 2136 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 2137 pcie_device = __mpt3sas_get_pdev_from_target(ioc, 2138 sas_target_priv_data); 2139 if (pcie_device && !sas_target_priv_data->num_luns) 2140 pcie_device->starget = NULL; 2141 2142 if (pcie_device) 2143 pcie_device_put(pcie_device); 2144 2145 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 2146 2147 } else if (!(sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)) { 2148 spin_lock_irqsave(&ioc->sas_device_lock, flags); 2149 sas_device = __mpt3sas_get_sdev_from_target(ioc, 2150 sas_target_priv_data); 2151 if (sas_device && !sas_target_priv_data->num_luns) 2152 sas_device->starget = NULL; 2153 2154 if (sas_device) 2155 sas_device_put(sas_device); 2156 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 2157 } 2158 2159 kfree(sdev->hostdata); 2160 sdev->hostdata = NULL; 2161 } 2162 2163 /** 2164 * _scsih_display_sata_capabilities - sata capabilities 2165 * @ioc: per adapter object 2166 * @handle: device handle 2167 * @sdev: scsi device struct 2168 */ 2169 static void 2170 _scsih_display_sata_capabilities(struct MPT3SAS_ADAPTER *ioc, 2171 u16 handle, struct scsi_device *sdev) 2172 { 2173 Mpi2ConfigReply_t mpi_reply; 2174 Mpi2SasDevicePage0_t sas_device_pg0; 2175 u32 ioc_status; 2176 u16 flags; 2177 u32 device_info; 2178 2179 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, 2180 MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { 2181 ioc_err(ioc, "failure at %s:%d/%s()!\n", 2182 __FILE__, __LINE__, __func__); 2183 return; 2184 } 2185 2186 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 2187 MPI2_IOCSTATUS_MASK; 2188 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 2189 ioc_err(ioc, "failure at %s:%d/%s()!\n", 2190 __FILE__, __LINE__, __func__); 2191 return; 2192 } 2193 2194 flags = le16_to_cpu(sas_device_pg0.Flags); 2195 device_info = le32_to_cpu(sas_device_pg0.DeviceInfo); 2196 2197 sdev_printk(KERN_INFO, sdev, 2198 "atapi(%s), ncq(%s), asyn_notify(%s), smart(%s), fua(%s), " 2199 "sw_preserve(%s)\n", 2200 (device_info & MPI2_SAS_DEVICE_INFO_ATAPI_DEVICE) ? "y" : "n", 2201 (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_NCQ_SUPPORTED) ? "y" : "n", 2202 (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_ASYNCHRONOUS_NOTIFY) ? "y" : 2203 "n", 2204 (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_SMART_SUPPORTED) ? "y" : "n", 2205 (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_FUA_SUPPORTED) ? "y" : "n", 2206 (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_SW_PRESERVE) ? "y" : "n"); 2207 } 2208 2209 /* 2210 * raid transport support - 2211 * Enabled for SLES11 and newer, in older kernels the driver will panic when 2212 * unloading the driver followed by a load - I believe that the subroutine 2213 * raid_class_release() is not cleaning up properly. 2214 */ 2215 2216 /** 2217 * scsih_is_raid - return boolean indicating device is raid volume 2218 * @dev: the device struct object 2219 */ 2220 static int 2221 scsih_is_raid(struct device *dev) 2222 { 2223 struct scsi_device *sdev = to_scsi_device(dev); 2224 struct MPT3SAS_ADAPTER *ioc = shost_priv(sdev->host); 2225 2226 if (ioc->is_warpdrive) 2227 return 0; 2228 return (sdev->channel == RAID_CHANNEL) ? 1 : 0; 2229 } 2230 2231 static int 2232 scsih_is_nvme(struct device *dev) 2233 { 2234 struct scsi_device *sdev = to_scsi_device(dev); 2235 2236 return (sdev->channel == PCIE_CHANNEL) ? 1 : 0; 2237 } 2238 2239 /** 2240 * scsih_get_resync - get raid volume resync percent complete 2241 * @dev: the device struct object 2242 */ 2243 static void 2244 scsih_get_resync(struct device *dev) 2245 { 2246 struct scsi_device *sdev = to_scsi_device(dev); 2247 struct MPT3SAS_ADAPTER *ioc = shost_priv(sdev->host); 2248 static struct _raid_device *raid_device; 2249 unsigned long flags; 2250 Mpi2RaidVolPage0_t vol_pg0; 2251 Mpi2ConfigReply_t mpi_reply; 2252 u32 volume_status_flags; 2253 u8 percent_complete; 2254 u16 handle; 2255 2256 percent_complete = 0; 2257 handle = 0; 2258 if (ioc->is_warpdrive) 2259 goto out; 2260 2261 spin_lock_irqsave(&ioc->raid_device_lock, flags); 2262 raid_device = _scsih_raid_device_find_by_id(ioc, sdev->id, 2263 sdev->channel); 2264 if (raid_device) { 2265 handle = raid_device->handle; 2266 percent_complete = raid_device->percent_complete; 2267 } 2268 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 2269 2270 if (!handle) 2271 goto out; 2272 2273 if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, &vol_pg0, 2274 MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle, 2275 sizeof(Mpi2RaidVolPage0_t))) { 2276 ioc_err(ioc, "failure at %s:%d/%s()!\n", 2277 __FILE__, __LINE__, __func__); 2278 percent_complete = 0; 2279 goto out; 2280 } 2281 2282 volume_status_flags = le32_to_cpu(vol_pg0.VolumeStatusFlags); 2283 if (!(volume_status_flags & 2284 MPI2_RAIDVOL0_STATUS_FLAG_RESYNC_IN_PROGRESS)) 2285 percent_complete = 0; 2286 2287 out: 2288 2289 switch (ioc->hba_mpi_version_belonged) { 2290 case MPI2_VERSION: 2291 raid_set_resync(mpt2sas_raid_template, dev, percent_complete); 2292 break; 2293 case MPI25_VERSION: 2294 case MPI26_VERSION: 2295 raid_set_resync(mpt3sas_raid_template, dev, percent_complete); 2296 break; 2297 } 2298 } 2299 2300 /** 2301 * scsih_get_state - get raid volume level 2302 * @dev: the device struct object 2303 */ 2304 static void 2305 scsih_get_state(struct device *dev) 2306 { 2307 struct scsi_device *sdev = to_scsi_device(dev); 2308 struct MPT3SAS_ADAPTER *ioc = shost_priv(sdev->host); 2309 static struct _raid_device *raid_device; 2310 unsigned long flags; 2311 Mpi2RaidVolPage0_t vol_pg0; 2312 Mpi2ConfigReply_t mpi_reply; 2313 u32 volstate; 2314 enum raid_state state = RAID_STATE_UNKNOWN; 2315 u16 handle = 0; 2316 2317 spin_lock_irqsave(&ioc->raid_device_lock, flags); 2318 raid_device = _scsih_raid_device_find_by_id(ioc, sdev->id, 2319 sdev->channel); 2320 if (raid_device) 2321 handle = raid_device->handle; 2322 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 2323 2324 if (!raid_device) 2325 goto out; 2326 2327 if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, &vol_pg0, 2328 MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle, 2329 sizeof(Mpi2RaidVolPage0_t))) { 2330 ioc_err(ioc, "failure at %s:%d/%s()!\n", 2331 __FILE__, __LINE__, __func__); 2332 goto out; 2333 } 2334 2335 volstate = le32_to_cpu(vol_pg0.VolumeStatusFlags); 2336 if (volstate & MPI2_RAIDVOL0_STATUS_FLAG_RESYNC_IN_PROGRESS) { 2337 state = RAID_STATE_RESYNCING; 2338 goto out; 2339 } 2340 2341 switch (vol_pg0.VolumeState) { 2342 case MPI2_RAID_VOL_STATE_OPTIMAL: 2343 case MPI2_RAID_VOL_STATE_ONLINE: 2344 state = RAID_STATE_ACTIVE; 2345 break; 2346 case MPI2_RAID_VOL_STATE_DEGRADED: 2347 state = RAID_STATE_DEGRADED; 2348 break; 2349 case MPI2_RAID_VOL_STATE_FAILED: 2350 case MPI2_RAID_VOL_STATE_MISSING: 2351 state = RAID_STATE_OFFLINE; 2352 break; 2353 } 2354 out: 2355 switch (ioc->hba_mpi_version_belonged) { 2356 case MPI2_VERSION: 2357 raid_set_state(mpt2sas_raid_template, dev, state); 2358 break; 2359 case MPI25_VERSION: 2360 case MPI26_VERSION: 2361 raid_set_state(mpt3sas_raid_template, dev, state); 2362 break; 2363 } 2364 } 2365 2366 /** 2367 * _scsih_set_level - set raid level 2368 * @ioc: ? 2369 * @sdev: scsi device struct 2370 * @volume_type: volume type 2371 */ 2372 static void 2373 _scsih_set_level(struct MPT3SAS_ADAPTER *ioc, 2374 struct scsi_device *sdev, u8 volume_type) 2375 { 2376 enum raid_level level = RAID_LEVEL_UNKNOWN; 2377 2378 switch (volume_type) { 2379 case MPI2_RAID_VOL_TYPE_RAID0: 2380 level = RAID_LEVEL_0; 2381 break; 2382 case MPI2_RAID_VOL_TYPE_RAID10: 2383 level = RAID_LEVEL_10; 2384 break; 2385 case MPI2_RAID_VOL_TYPE_RAID1E: 2386 level = RAID_LEVEL_1E; 2387 break; 2388 case MPI2_RAID_VOL_TYPE_RAID1: 2389 level = RAID_LEVEL_1; 2390 break; 2391 } 2392 2393 switch (ioc->hba_mpi_version_belonged) { 2394 case MPI2_VERSION: 2395 raid_set_level(mpt2sas_raid_template, 2396 &sdev->sdev_gendev, level); 2397 break; 2398 case MPI25_VERSION: 2399 case MPI26_VERSION: 2400 raid_set_level(mpt3sas_raid_template, 2401 &sdev->sdev_gendev, level); 2402 break; 2403 } 2404 } 2405 2406 2407 /** 2408 * _scsih_get_volume_capabilities - volume capabilities 2409 * @ioc: per adapter object 2410 * @raid_device: the raid_device object 2411 * 2412 * Return: 0 for success, else 1 2413 */ 2414 static int 2415 _scsih_get_volume_capabilities(struct MPT3SAS_ADAPTER *ioc, 2416 struct _raid_device *raid_device) 2417 { 2418 Mpi2RaidVolPage0_t *vol_pg0; 2419 Mpi2RaidPhysDiskPage0_t pd_pg0; 2420 Mpi2SasDevicePage0_t sas_device_pg0; 2421 Mpi2ConfigReply_t mpi_reply; 2422 u16 sz; 2423 u8 num_pds; 2424 2425 if ((mpt3sas_config_get_number_pds(ioc, raid_device->handle, 2426 &num_pds)) || !num_pds) { 2427 dfailprintk(ioc, 2428 ioc_warn(ioc, "failure at %s:%d/%s()!\n", 2429 __FILE__, __LINE__, __func__)); 2430 return 1; 2431 } 2432 2433 raid_device->num_pds = num_pds; 2434 sz = struct_size(vol_pg0, PhysDisk, num_pds); 2435 vol_pg0 = kzalloc(sz, GFP_KERNEL); 2436 if (!vol_pg0) { 2437 dfailprintk(ioc, 2438 ioc_warn(ioc, "failure at %s:%d/%s()!\n", 2439 __FILE__, __LINE__, __func__)); 2440 return 1; 2441 } 2442 2443 if ((mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, vol_pg0, 2444 MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, raid_device->handle, sz))) { 2445 dfailprintk(ioc, 2446 ioc_warn(ioc, "failure at %s:%d/%s()!\n", 2447 __FILE__, __LINE__, __func__)); 2448 kfree(vol_pg0); 2449 return 1; 2450 } 2451 2452 raid_device->volume_type = vol_pg0->VolumeType; 2453 2454 /* figure out what the underlying devices are by 2455 * obtaining the device_info bits for the 1st device 2456 */ 2457 if (!(mpt3sas_config_get_phys_disk_pg0(ioc, &mpi_reply, 2458 &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_PHYSDISKNUM, 2459 vol_pg0->PhysDisk[0].PhysDiskNum))) { 2460 if (!(mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, 2461 &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, 2462 le16_to_cpu(pd_pg0.DevHandle)))) { 2463 raid_device->device_info = 2464 le32_to_cpu(sas_device_pg0.DeviceInfo); 2465 } 2466 } 2467 2468 kfree(vol_pg0); 2469 return 0; 2470 } 2471 2472 /** 2473 * _scsih_enable_tlr - setting TLR flags 2474 * @ioc: per adapter object 2475 * @sdev: scsi device struct 2476 * 2477 * Enabling Transaction Layer Retries for tape devices when 2478 * vpd page 0x90 is present 2479 * 2480 */ 2481 static void 2482 _scsih_enable_tlr(struct MPT3SAS_ADAPTER *ioc, struct scsi_device *sdev) 2483 { 2484 2485 /* only for TAPE */ 2486 if (sdev->type != TYPE_TAPE) 2487 return; 2488 2489 if (!(ioc->facts.IOCCapabilities & MPI2_IOCFACTS_CAPABILITY_TLR)) 2490 return; 2491 2492 sas_enable_tlr(sdev); 2493 sdev_printk(KERN_INFO, sdev, "TLR %s\n", 2494 sas_is_tlr_enabled(sdev) ? "Enabled" : "Disabled"); 2495 return; 2496 2497 } 2498 2499 /** 2500 * scsih_device_configure - device configure routine. 2501 * @sdev: scsi device struct 2502 * @lim: queue limits 2503 * 2504 * Return: 0 if ok. Any other return is assumed to be an error and 2505 * the device is ignored. 2506 */ 2507 static int 2508 scsih_device_configure(struct scsi_device *sdev, struct queue_limits *lim) 2509 { 2510 struct Scsi_Host *shost = sdev->host; 2511 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); 2512 struct MPT3SAS_DEVICE *sas_device_priv_data; 2513 struct MPT3SAS_TARGET *sas_target_priv_data; 2514 struct _sas_device *sas_device; 2515 struct _pcie_device *pcie_device; 2516 struct _raid_device *raid_device; 2517 unsigned long flags; 2518 int qdepth; 2519 u8 ssp_target = 0; 2520 char *ds = ""; 2521 char *r_level = ""; 2522 u16 handle, volume_handle = 0; 2523 u64 volume_wwid = 0; 2524 2525 qdepth = 1; 2526 sas_device_priv_data = sdev->hostdata; 2527 sas_device_priv_data->configured_lun = 1; 2528 sas_device_priv_data->flags &= ~MPT_DEVICE_FLAGS_INIT; 2529 sas_target_priv_data = sas_device_priv_data->sas_target; 2530 handle = sas_target_priv_data->handle; 2531 2532 /* raid volume handling */ 2533 if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME) { 2534 2535 spin_lock_irqsave(&ioc->raid_device_lock, flags); 2536 raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); 2537 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 2538 if (!raid_device) { 2539 dfailprintk(ioc, 2540 ioc_warn(ioc, "failure at %s:%d/%s()!\n", 2541 __FILE__, __LINE__, __func__)); 2542 return 1; 2543 } 2544 2545 if (_scsih_get_volume_capabilities(ioc, raid_device)) { 2546 dfailprintk(ioc, 2547 ioc_warn(ioc, "failure at %s:%d/%s()!\n", 2548 __FILE__, __LINE__, __func__)); 2549 return 1; 2550 } 2551 2552 /* 2553 * WARPDRIVE: Initialize the required data for Direct IO 2554 */ 2555 mpt3sas_init_warpdrive_properties(ioc, raid_device); 2556 2557 /* RAID Queue Depth Support 2558 * IS volume = underlying qdepth of drive type, either 2559 * MPT3SAS_SAS_QUEUE_DEPTH or MPT3SAS_SATA_QUEUE_DEPTH 2560 * IM/IME/R10 = 128 (MPT3SAS_RAID_QUEUE_DEPTH) 2561 */ 2562 if (raid_device->device_info & 2563 MPI2_SAS_DEVICE_INFO_SSP_TARGET) { 2564 qdepth = MPT3SAS_SAS_QUEUE_DEPTH; 2565 ds = "SSP"; 2566 } else { 2567 qdepth = MPT3SAS_SATA_QUEUE_DEPTH; 2568 if (raid_device->device_info & 2569 MPI2_SAS_DEVICE_INFO_SATA_DEVICE) 2570 ds = "SATA"; 2571 else 2572 ds = "STP"; 2573 } 2574 2575 switch (raid_device->volume_type) { 2576 case MPI2_RAID_VOL_TYPE_RAID0: 2577 r_level = "RAID0"; 2578 break; 2579 case MPI2_RAID_VOL_TYPE_RAID1E: 2580 qdepth = MPT3SAS_RAID_QUEUE_DEPTH; 2581 if (ioc->manu_pg10.OEMIdentifier && 2582 (le32_to_cpu(ioc->manu_pg10.GenericFlags0) & 2583 MFG10_GF0_R10_DISPLAY) && 2584 !(raid_device->num_pds % 2)) 2585 r_level = "RAID10"; 2586 else 2587 r_level = "RAID1E"; 2588 break; 2589 case MPI2_RAID_VOL_TYPE_RAID1: 2590 qdepth = MPT3SAS_RAID_QUEUE_DEPTH; 2591 r_level = "RAID1"; 2592 break; 2593 case MPI2_RAID_VOL_TYPE_RAID10: 2594 qdepth = MPT3SAS_RAID_QUEUE_DEPTH; 2595 r_level = "RAID10"; 2596 break; 2597 case MPI2_RAID_VOL_TYPE_UNKNOWN: 2598 default: 2599 qdepth = MPT3SAS_RAID_QUEUE_DEPTH; 2600 r_level = "RAIDX"; 2601 break; 2602 } 2603 2604 if (!ioc->hide_ir_msg) 2605 sdev_printk(KERN_INFO, sdev, 2606 "%s: handle(0x%04x), wwid(0x%016llx)," 2607 " pd_count(%d), type(%s)\n", 2608 r_level, raid_device->handle, 2609 (unsigned long long)raid_device->wwid, 2610 raid_device->num_pds, ds); 2611 2612 if (shost->max_sectors > MPT3SAS_RAID_MAX_SECTORS) { 2613 lim->max_hw_sectors = MPT3SAS_RAID_MAX_SECTORS; 2614 sdev_printk(KERN_INFO, sdev, 2615 "Set queue's max_sector to: %u\n", 2616 MPT3SAS_RAID_MAX_SECTORS); 2617 } 2618 2619 mpt3sas_scsih_change_queue_depth(sdev, qdepth); 2620 2621 /* raid transport support */ 2622 if (!ioc->is_warpdrive) 2623 _scsih_set_level(ioc, sdev, raid_device->volume_type); 2624 return 0; 2625 } 2626 2627 /* non-raid handling */ 2628 if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) { 2629 if (mpt3sas_config_get_volume_handle(ioc, handle, 2630 &volume_handle)) { 2631 dfailprintk(ioc, 2632 ioc_warn(ioc, "failure at %s:%d/%s()!\n", 2633 __FILE__, __LINE__, __func__)); 2634 return 1; 2635 } 2636 if (volume_handle && mpt3sas_config_get_volume_wwid(ioc, 2637 volume_handle, &volume_wwid)) { 2638 dfailprintk(ioc, 2639 ioc_warn(ioc, "failure at %s:%d/%s()!\n", 2640 __FILE__, __LINE__, __func__)); 2641 return 1; 2642 } 2643 } 2644 2645 /* PCIe handling */ 2646 if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) { 2647 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 2648 pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, 2649 sas_device_priv_data->sas_target->sas_address); 2650 if (!pcie_device) { 2651 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 2652 dfailprintk(ioc, 2653 ioc_warn(ioc, "failure at %s:%d/%s()!\n", 2654 __FILE__, __LINE__, __func__)); 2655 return 1; 2656 } 2657 2658 qdepth = ioc->max_nvme_qd; 2659 ds = "NVMe"; 2660 sdev_printk(KERN_INFO, sdev, 2661 "%s: handle(0x%04x), wwid(0x%016llx), port(%d)\n", 2662 ds, handle, (unsigned long long)pcie_device->wwid, 2663 pcie_device->port_num); 2664 if (pcie_device->enclosure_handle != 0) 2665 sdev_printk(KERN_INFO, sdev, 2666 "%s: enclosure logical id(0x%016llx), slot(%d)\n", 2667 ds, 2668 (unsigned long long)pcie_device->enclosure_logical_id, 2669 pcie_device->slot); 2670 if (pcie_device->connector_name[0] != '\0') 2671 sdev_printk(KERN_INFO, sdev, 2672 "%s: enclosure level(0x%04x)," 2673 "connector name( %s)\n", ds, 2674 pcie_device->enclosure_level, 2675 pcie_device->connector_name); 2676 2677 if (pcie_device->nvme_mdts) 2678 lim->max_hw_sectors = pcie_device->nvme_mdts / 512; 2679 2680 pcie_device_put(pcie_device); 2681 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 2682 mpt3sas_scsih_change_queue_depth(sdev, qdepth); 2683 lim->virt_boundary_mask = ioc->page_size - 1; 2684 return 0; 2685 } 2686 2687 spin_lock_irqsave(&ioc->sas_device_lock, flags); 2688 sas_device = __mpt3sas_get_sdev_by_addr(ioc, 2689 sas_device_priv_data->sas_target->sas_address, 2690 sas_device_priv_data->sas_target->port); 2691 if (!sas_device) { 2692 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 2693 dfailprintk(ioc, 2694 ioc_warn(ioc, "failure at %s:%d/%s()!\n", 2695 __FILE__, __LINE__, __func__)); 2696 return 1; 2697 } 2698 2699 sas_device->volume_handle = volume_handle; 2700 sas_device->volume_wwid = volume_wwid; 2701 if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET) { 2702 qdepth = (sas_device->port_type > 1) ? 2703 ioc->max_wideport_qd : ioc->max_narrowport_qd; 2704 ssp_target = 1; 2705 if (sas_device->device_info & 2706 MPI2_SAS_DEVICE_INFO_SEP) { 2707 sdev_printk(KERN_WARNING, sdev, 2708 "set ignore_delay_remove for handle(0x%04x)\n", 2709 sas_device_priv_data->sas_target->handle); 2710 sas_device_priv_data->ignore_delay_remove = 1; 2711 ds = "SES"; 2712 } else 2713 ds = "SSP"; 2714 } else { 2715 qdepth = ioc->max_sata_qd; 2716 if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_STP_TARGET) 2717 ds = "STP"; 2718 else if (sas_device->device_info & 2719 MPI2_SAS_DEVICE_INFO_SATA_DEVICE) 2720 ds = "SATA"; 2721 } 2722 2723 sdev_printk(KERN_INFO, sdev, "%s: handle(0x%04x), " \ 2724 "sas_addr(0x%016llx), phy(%d), device_name(0x%016llx)\n", 2725 ds, handle, (unsigned long long)sas_device->sas_address, 2726 sas_device->phy, (unsigned long long)sas_device->device_name); 2727 2728 _scsih_display_enclosure_chassis_info(NULL, sas_device, sdev, NULL); 2729 2730 sas_device_put(sas_device); 2731 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 2732 2733 if (!ssp_target) 2734 _scsih_display_sata_capabilities(ioc, handle, sdev); 2735 2736 2737 mpt3sas_scsih_change_queue_depth(sdev, qdepth); 2738 2739 if (ssp_target) { 2740 sas_read_port_mode_page(sdev); 2741 _scsih_enable_tlr(ioc, sdev); 2742 } 2743 2744 return 0; 2745 } 2746 2747 /** 2748 * scsih_bios_param - fetch head, sector, cylinder info for a disk 2749 * @sdev: scsi device struct 2750 * @bdev: pointer to block device context 2751 * @capacity: device size (in 512 byte sectors) 2752 * @params: three element array to place output: 2753 * params[0] number of heads (max 255) 2754 * params[1] number of sectors (max 63) 2755 * params[2] number of cylinders 2756 */ 2757 static int 2758 scsih_bios_param(struct scsi_device *sdev, struct block_device *bdev, 2759 sector_t capacity, int params[]) 2760 { 2761 int heads; 2762 int sectors; 2763 sector_t cylinders; 2764 ulong dummy; 2765 2766 heads = 64; 2767 sectors = 32; 2768 2769 dummy = heads * sectors; 2770 cylinders = capacity; 2771 sector_div(cylinders, dummy); 2772 2773 /* 2774 * Handle extended translation size for logical drives 2775 * > 1Gb 2776 */ 2777 if ((ulong)capacity >= 0x200000) { 2778 heads = 255; 2779 sectors = 63; 2780 dummy = heads * sectors; 2781 cylinders = capacity; 2782 sector_div(cylinders, dummy); 2783 } 2784 2785 /* return result */ 2786 params[0] = heads; 2787 params[1] = sectors; 2788 params[2] = cylinders; 2789 2790 return 0; 2791 } 2792 2793 /** 2794 * _scsih_response_code - translation of device response code 2795 * @ioc: per adapter object 2796 * @response_code: response code returned by the device 2797 */ 2798 static void 2799 _scsih_response_code(struct MPT3SAS_ADAPTER *ioc, u8 response_code) 2800 { 2801 char *desc; 2802 2803 switch (response_code) { 2804 case MPI2_SCSITASKMGMT_RSP_TM_COMPLETE: 2805 desc = "task management request completed"; 2806 break; 2807 case MPI2_SCSITASKMGMT_RSP_INVALID_FRAME: 2808 desc = "invalid frame"; 2809 break; 2810 case MPI2_SCSITASKMGMT_RSP_TM_NOT_SUPPORTED: 2811 desc = "task management request not supported"; 2812 break; 2813 case MPI2_SCSITASKMGMT_RSP_TM_FAILED: 2814 desc = "task management request failed"; 2815 break; 2816 case MPI2_SCSITASKMGMT_RSP_TM_SUCCEEDED: 2817 desc = "task management request succeeded"; 2818 break; 2819 case MPI2_SCSITASKMGMT_RSP_TM_INVALID_LUN: 2820 desc = "invalid lun"; 2821 break; 2822 case 0xA: 2823 desc = "overlapped tag attempted"; 2824 break; 2825 case MPI2_SCSITASKMGMT_RSP_IO_QUEUED_ON_IOC: 2826 desc = "task queued, however not sent to target"; 2827 break; 2828 default: 2829 desc = "unknown"; 2830 break; 2831 } 2832 ioc_warn(ioc, "response_code(0x%01x): %s\n", response_code, desc); 2833 } 2834 2835 /** 2836 * _scsih_tm_done - tm completion routine 2837 * @ioc: per adapter object 2838 * @smid: system request message index 2839 * @msix_index: MSIX table index supplied by the OS 2840 * @reply: reply message frame(lower 32bit addr) 2841 * Context: none. 2842 * 2843 * The callback handler when using scsih_issue_tm. 2844 * 2845 * Return: 1 meaning mf should be freed from _base_interrupt 2846 * 0 means the mf is freed from this function. 2847 */ 2848 static u8 2849 _scsih_tm_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply) 2850 { 2851 MPI2DefaultReply_t *mpi_reply; 2852 2853 if (ioc->tm_cmds.status == MPT3_CMD_NOT_USED) 2854 return 1; 2855 if (ioc->tm_cmds.smid != smid) 2856 return 1; 2857 ioc->tm_cmds.status |= MPT3_CMD_COMPLETE; 2858 mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply); 2859 if (mpi_reply) { 2860 memcpy(ioc->tm_cmds.reply, mpi_reply, mpi_reply->MsgLength*4); 2861 ioc->tm_cmds.status |= MPT3_CMD_REPLY_VALID; 2862 } 2863 ioc->tm_cmds.status &= ~MPT3_CMD_PENDING; 2864 complete(&ioc->tm_cmds.done); 2865 return 1; 2866 } 2867 2868 /** 2869 * mpt3sas_scsih_set_tm_flag - set per target tm_busy 2870 * @ioc: per adapter object 2871 * @handle: device handle 2872 * 2873 * During taskmangement request, we need to freeze the device queue. 2874 */ 2875 void 2876 mpt3sas_scsih_set_tm_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle) 2877 { 2878 struct MPT3SAS_DEVICE *sas_device_priv_data; 2879 struct scsi_device *sdev; 2880 u8 skip = 0; 2881 2882 shost_for_each_device(sdev, ioc->shost) { 2883 if (skip) 2884 continue; 2885 sas_device_priv_data = sdev->hostdata; 2886 if (!sas_device_priv_data) 2887 continue; 2888 if (sas_device_priv_data->sas_target->handle == handle) { 2889 sas_device_priv_data->sas_target->tm_busy = 1; 2890 skip = 1; 2891 ioc->ignore_loginfos = 1; 2892 } 2893 } 2894 } 2895 2896 /** 2897 * mpt3sas_scsih_clear_tm_flag - clear per target tm_busy 2898 * @ioc: per adapter object 2899 * @handle: device handle 2900 * 2901 * During taskmangement request, we need to freeze the device queue. 2902 */ 2903 void 2904 mpt3sas_scsih_clear_tm_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle) 2905 { 2906 struct MPT3SAS_DEVICE *sas_device_priv_data; 2907 struct scsi_device *sdev; 2908 u8 skip = 0; 2909 2910 shost_for_each_device(sdev, ioc->shost) { 2911 if (skip) 2912 continue; 2913 sas_device_priv_data = sdev->hostdata; 2914 if (!sas_device_priv_data) 2915 continue; 2916 if (sas_device_priv_data->sas_target->handle == handle) { 2917 sas_device_priv_data->sas_target->tm_busy = 0; 2918 skip = 1; 2919 ioc->ignore_loginfos = 0; 2920 } 2921 } 2922 } 2923 2924 /** 2925 * scsih_tm_cmd_map_status - map the target reset & LUN reset TM status 2926 * @ioc: per adapter object 2927 * @channel: the channel assigned by the OS 2928 * @id: the id assigned by the OS 2929 * @lun: lun number 2930 * @type: MPI2_SCSITASKMGMT_TASKTYPE__XXX (defined in mpi2_init.h) 2931 * @smid_task: smid assigned to the task 2932 * 2933 * Look whether TM has aborted the timed out SCSI command, if 2934 * TM has aborted the IO then return SUCCESS else return FAILED. 2935 */ 2936 static int 2937 scsih_tm_cmd_map_status(struct MPT3SAS_ADAPTER *ioc, uint channel, 2938 uint id, uint lun, u8 type, u16 smid_task) 2939 { 2940 2941 if (smid_task <= ioc->shost->can_queue) { 2942 switch (type) { 2943 case MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET: 2944 if (!(_scsih_scsi_lookup_find_by_target(ioc, 2945 id, channel))) 2946 return SUCCESS; 2947 break; 2948 case MPI2_SCSITASKMGMT_TASKTYPE_ABRT_TASK_SET: 2949 case MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET: 2950 if (!(_scsih_scsi_lookup_find_by_lun(ioc, id, 2951 lun, channel))) 2952 return SUCCESS; 2953 break; 2954 default: 2955 return SUCCESS; 2956 } 2957 } else if (smid_task == ioc->scsih_cmds.smid) { 2958 if ((ioc->scsih_cmds.status & MPT3_CMD_COMPLETE) || 2959 (ioc->scsih_cmds.status & MPT3_CMD_NOT_USED)) 2960 return SUCCESS; 2961 } else if (smid_task == ioc->ctl_cmds.smid) { 2962 if ((ioc->ctl_cmds.status & MPT3_CMD_COMPLETE) || 2963 (ioc->ctl_cmds.status & MPT3_CMD_NOT_USED)) 2964 return SUCCESS; 2965 } 2966 2967 return FAILED; 2968 } 2969 2970 /** 2971 * scsih_tm_post_processing - post processing of target & LUN reset 2972 * @ioc: per adapter object 2973 * @handle: device handle 2974 * @channel: the channel assigned by the OS 2975 * @id: the id assigned by the OS 2976 * @lun: lun number 2977 * @type: MPI2_SCSITASKMGMT_TASKTYPE__XXX (defined in mpi2_init.h) 2978 * @smid_task: smid assigned to the task 2979 * 2980 * Post processing of target & LUN reset. Due to interrupt latency 2981 * issue it possible that interrupt for aborted IO might not be 2982 * received yet. So before returning failure status, poll the 2983 * reply descriptor pools for the reply of timed out SCSI command. 2984 * Return FAILED status if reply for timed out is not received 2985 * otherwise return SUCCESS. 2986 */ 2987 static int 2988 scsih_tm_post_processing(struct MPT3SAS_ADAPTER *ioc, u16 handle, 2989 uint channel, uint id, uint lun, u8 type, u16 smid_task) 2990 { 2991 int rc; 2992 2993 rc = scsih_tm_cmd_map_status(ioc, channel, id, lun, type, smid_task); 2994 if (rc == SUCCESS) 2995 return rc; 2996 2997 ioc_info(ioc, 2998 "Poll ReplyDescriptor queues for completion of" 2999 " smid(%d), task_type(0x%02x), handle(0x%04x)\n", 3000 smid_task, type, handle); 3001 3002 /* 3003 * Due to interrupt latency issues, driver may receive interrupt for 3004 * TM first and then for aborted SCSI IO command. So, poll all the 3005 * ReplyDescriptor pools before returning the FAILED status to SML. 3006 */ 3007 mpt3sas_base_mask_interrupts(ioc); 3008 mpt3sas_base_sync_reply_irqs(ioc, 1); 3009 mpt3sas_base_unmask_interrupts(ioc); 3010 3011 return scsih_tm_cmd_map_status(ioc, channel, id, lun, type, smid_task); 3012 } 3013 3014 /** 3015 * mpt3sas_scsih_issue_tm - main routine for sending tm requests 3016 * @ioc: per adapter struct 3017 * @handle: device handle 3018 * @channel: the channel assigned by the OS 3019 * @id: the id assigned by the OS 3020 * @lun: lun number 3021 * @type: MPI2_SCSITASKMGMT_TASKTYPE__XXX (defined in mpi2_init.h) 3022 * @smid_task: smid assigned to the task 3023 * @msix_task: MSIX table index supplied by the OS 3024 * @timeout: timeout in seconds 3025 * @tr_method: Target Reset Method 3026 * Context: user 3027 * 3028 * A generic API for sending task management requests to firmware. 3029 * 3030 * The callback index is set inside `ioc->tm_cb_idx`. 3031 * The caller is responsible to check for outstanding commands. 3032 * 3033 * Return: SUCCESS or FAILED. 3034 */ 3035 int 3036 mpt3sas_scsih_issue_tm(struct MPT3SAS_ADAPTER *ioc, u16 handle, uint channel, 3037 uint id, u64 lun, u8 type, u16 smid_task, u16 msix_task, 3038 u8 timeout, u8 tr_method) 3039 { 3040 Mpi2SCSITaskManagementRequest_t *mpi_request; 3041 Mpi2SCSITaskManagementReply_t *mpi_reply; 3042 Mpi25SCSIIORequest_t *request; 3043 u16 smid = 0; 3044 u32 ioc_state; 3045 int rc; 3046 u8 issue_reset = 0; 3047 3048 lockdep_assert_held(&ioc->tm_cmds.mutex); 3049 3050 if (ioc->tm_cmds.status != MPT3_CMD_NOT_USED) { 3051 ioc_info(ioc, "%s: tm_cmd busy!!!\n", __func__); 3052 return FAILED; 3053 } 3054 3055 if (ioc->shost_recovery || ioc->remove_host || 3056 ioc->pci_error_recovery) { 3057 ioc_info(ioc, "%s: host reset in progress!\n", __func__); 3058 return FAILED; 3059 } 3060 3061 ioc_state = mpt3sas_base_get_iocstate(ioc, 0); 3062 if (ioc_state & MPI2_DOORBELL_USED) { 3063 dhsprintk(ioc, ioc_info(ioc, "unexpected doorbell active!\n")); 3064 rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER); 3065 return (!rc) ? SUCCESS : FAILED; 3066 } 3067 3068 if ((ioc_state & MPI2_IOC_STATE_MASK) == MPI2_IOC_STATE_FAULT) { 3069 mpt3sas_print_fault_code(ioc, ioc_state & 3070 MPI2_DOORBELL_DATA_MASK); 3071 rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER); 3072 return (!rc) ? SUCCESS : FAILED; 3073 } else if ((ioc_state & MPI2_IOC_STATE_MASK) == 3074 MPI2_IOC_STATE_COREDUMP) { 3075 mpt3sas_print_coredump_info(ioc, ioc_state & 3076 MPI2_DOORBELL_DATA_MASK); 3077 rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER); 3078 return (!rc) ? SUCCESS : FAILED; 3079 } 3080 3081 smid = mpt3sas_base_get_smid_hpr(ioc, ioc->tm_cb_idx); 3082 if (!smid) { 3083 ioc_err(ioc, "%s: failed obtaining a smid\n", __func__); 3084 return FAILED; 3085 } 3086 3087 dtmprintk(ioc, 3088 ioc_info(ioc, "sending tm: handle(0x%04x), task_type(0x%02x), smid(%d), timeout(%d), tr_method(0x%x)\n", 3089 handle, type, smid_task, timeout, tr_method)); 3090 ioc->tm_cmds.status = MPT3_CMD_PENDING; 3091 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); 3092 ioc->tm_cmds.smid = smid; 3093 memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t)); 3094 memset(ioc->tm_cmds.reply, 0, sizeof(Mpi2SCSITaskManagementReply_t)); 3095 mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT; 3096 mpi_request->DevHandle = cpu_to_le16(handle); 3097 mpi_request->TaskType = type; 3098 if (type == MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK || 3099 type == MPI2_SCSITASKMGMT_TASKTYPE_QUERY_TASK) 3100 mpi_request->MsgFlags = tr_method; 3101 mpi_request->TaskMID = cpu_to_le16(smid_task); 3102 int_to_scsilun(lun, (struct scsi_lun *)mpi_request->LUN); 3103 mpt3sas_scsih_set_tm_flag(ioc, handle); 3104 init_completion(&ioc->tm_cmds.done); 3105 ioc->put_smid_hi_priority(ioc, smid, msix_task); 3106 wait_for_completion_timeout(&ioc->tm_cmds.done, timeout*HZ); 3107 if (!(ioc->tm_cmds.status & MPT3_CMD_COMPLETE)) { 3108 mpt3sas_check_cmd_timeout(ioc, 3109 ioc->tm_cmds.status, mpi_request, 3110 sizeof(Mpi2SCSITaskManagementRequest_t)/4, issue_reset); 3111 if (issue_reset) { 3112 rc = mpt3sas_base_hard_reset_handler(ioc, 3113 FORCE_BIG_HAMMER); 3114 rc = (!rc) ? SUCCESS : FAILED; 3115 goto out; 3116 } 3117 } 3118 3119 /* sync IRQs in case those were busy during flush. */ 3120 mpt3sas_base_sync_reply_irqs(ioc, 0); 3121 3122 if (ioc->tm_cmds.status & MPT3_CMD_REPLY_VALID) { 3123 mpt3sas_trigger_master(ioc, MASTER_TRIGGER_TASK_MANAGMENT); 3124 mpi_reply = ioc->tm_cmds.reply; 3125 dtmprintk(ioc, 3126 ioc_info(ioc, "complete tm: ioc_status(0x%04x), loginfo(0x%08x), term_count(0x%08x)\n", 3127 le16_to_cpu(mpi_reply->IOCStatus), 3128 le32_to_cpu(mpi_reply->IOCLogInfo), 3129 le32_to_cpu(mpi_reply->TerminationCount))); 3130 if (ioc->logging_level & MPT_DEBUG_TM) { 3131 _scsih_response_code(ioc, mpi_reply->ResponseCode); 3132 if (mpi_reply->IOCStatus) 3133 _debug_dump_mf(mpi_request, 3134 sizeof(Mpi2SCSITaskManagementRequest_t)/4); 3135 } 3136 } 3137 3138 switch (type) { 3139 case MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK: 3140 rc = SUCCESS; 3141 /* 3142 * If DevHandle filed in smid_task's entry of request pool 3143 * doesn't match with device handle on which this task abort 3144 * TM is received then it means that TM has successfully 3145 * aborted the timed out command. Since smid_task's entry in 3146 * request pool will be memset to zero once the timed out 3147 * command is returned to the SML. If the command is not 3148 * aborted then smid_task’s entry won’t be cleared and it 3149 * will have same DevHandle value on which this task abort TM 3150 * is received and driver will return the TM status as FAILED. 3151 */ 3152 request = mpt3sas_base_get_msg_frame(ioc, smid_task); 3153 if (le16_to_cpu(request->DevHandle) != handle) 3154 break; 3155 3156 ioc_info(ioc, "Task abort tm failed: handle(0x%04x)," 3157 "timeout(%d) tr_method(0x%x) smid(%d) msix_index(%d)\n", 3158 handle, timeout, tr_method, smid_task, msix_task); 3159 rc = FAILED; 3160 break; 3161 3162 case MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET: 3163 case MPI2_SCSITASKMGMT_TASKTYPE_ABRT_TASK_SET: 3164 case MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET: 3165 rc = scsih_tm_post_processing(ioc, handle, channel, id, lun, 3166 type, smid_task); 3167 break; 3168 case MPI2_SCSITASKMGMT_TASKTYPE_QUERY_TASK: 3169 rc = SUCCESS; 3170 break; 3171 default: 3172 rc = FAILED; 3173 break; 3174 } 3175 3176 out: 3177 mpt3sas_scsih_clear_tm_flag(ioc, handle); 3178 ioc->tm_cmds.status = MPT3_CMD_NOT_USED; 3179 return rc; 3180 } 3181 3182 int mpt3sas_scsih_issue_locked_tm(struct MPT3SAS_ADAPTER *ioc, u16 handle, 3183 uint channel, uint id, u64 lun, u8 type, u16 smid_task, 3184 u16 msix_task, u8 timeout, u8 tr_method) 3185 { 3186 int ret; 3187 3188 mutex_lock(&ioc->tm_cmds.mutex); 3189 ret = mpt3sas_scsih_issue_tm(ioc, handle, channel, id, lun, type, 3190 smid_task, msix_task, timeout, tr_method); 3191 mutex_unlock(&ioc->tm_cmds.mutex); 3192 3193 return ret; 3194 } 3195 3196 /** 3197 * _scsih_tm_display_info - displays info about the device 3198 * @ioc: per adapter struct 3199 * @scmd: pointer to scsi command object 3200 * 3201 * Called by task management callback handlers. 3202 */ 3203 static void 3204 _scsih_tm_display_info(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd) 3205 { 3206 struct scsi_target *starget = scmd->device->sdev_target; 3207 struct MPT3SAS_TARGET *priv_target = starget->hostdata; 3208 struct _sas_device *sas_device = NULL; 3209 struct _pcie_device *pcie_device = NULL; 3210 unsigned long flags; 3211 char *device_str = NULL; 3212 3213 if (!priv_target) 3214 return; 3215 if (ioc->hide_ir_msg) 3216 device_str = "WarpDrive"; 3217 else 3218 device_str = "volume"; 3219 3220 scsi_print_command(scmd); 3221 if (priv_target->flags & MPT_TARGET_FLAGS_VOLUME) { 3222 starget_printk(KERN_INFO, starget, 3223 "%s handle(0x%04x), %s wwid(0x%016llx)\n", 3224 device_str, priv_target->handle, 3225 device_str, (unsigned long long)priv_target->sas_address); 3226 3227 } else if (priv_target->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) { 3228 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 3229 pcie_device = __mpt3sas_get_pdev_from_target(ioc, priv_target); 3230 if (pcie_device) { 3231 starget_printk(KERN_INFO, starget, 3232 "handle(0x%04x), wwid(0x%016llx), port(%d)\n", 3233 pcie_device->handle, 3234 (unsigned long long)pcie_device->wwid, 3235 pcie_device->port_num); 3236 if (pcie_device->enclosure_handle != 0) 3237 starget_printk(KERN_INFO, starget, 3238 "enclosure logical id(0x%016llx), slot(%d)\n", 3239 (unsigned long long) 3240 pcie_device->enclosure_logical_id, 3241 pcie_device->slot); 3242 if (pcie_device->connector_name[0] != '\0') 3243 starget_printk(KERN_INFO, starget, 3244 "enclosure level(0x%04x), connector name( %s)\n", 3245 pcie_device->enclosure_level, 3246 pcie_device->connector_name); 3247 pcie_device_put(pcie_device); 3248 } 3249 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 3250 3251 } else { 3252 spin_lock_irqsave(&ioc->sas_device_lock, flags); 3253 sas_device = __mpt3sas_get_sdev_from_target(ioc, priv_target); 3254 if (sas_device) { 3255 if (priv_target->flags & 3256 MPT_TARGET_FLAGS_RAID_COMPONENT) { 3257 starget_printk(KERN_INFO, starget, 3258 "volume handle(0x%04x), " 3259 "volume wwid(0x%016llx)\n", 3260 sas_device->volume_handle, 3261 (unsigned long long)sas_device->volume_wwid); 3262 } 3263 starget_printk(KERN_INFO, starget, 3264 "handle(0x%04x), sas_address(0x%016llx), phy(%d)\n", 3265 sas_device->handle, 3266 (unsigned long long)sas_device->sas_address, 3267 sas_device->phy); 3268 3269 _scsih_display_enclosure_chassis_info(NULL, sas_device, 3270 NULL, starget); 3271 3272 sas_device_put(sas_device); 3273 } 3274 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 3275 } 3276 } 3277 3278 /** 3279 * scsih_abort - eh threads main abort routine 3280 * @scmd: pointer to scsi command object 3281 * 3282 * Return: SUCCESS if command aborted else FAILED 3283 */ 3284 static int 3285 scsih_abort(struct scsi_cmnd *scmd) 3286 { 3287 struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host); 3288 struct MPT3SAS_DEVICE *sas_device_priv_data; 3289 struct scsiio_tracker *st = scsi_cmd_priv(scmd); 3290 u16 handle; 3291 int r; 3292 3293 u8 timeout = 30; 3294 struct _pcie_device *pcie_device = NULL; 3295 sdev_printk(KERN_INFO, scmd->device, "attempting task abort!" 3296 "scmd(0x%p), outstanding for %u ms & timeout %u ms\n", 3297 scmd, jiffies_to_msecs(jiffies - scmd->jiffies_at_alloc), 3298 (scsi_cmd_to_rq(scmd)->timeout / HZ) * 1000); 3299 _scsih_tm_display_info(ioc, scmd); 3300 3301 sas_device_priv_data = scmd->device->hostdata; 3302 if (!sas_device_priv_data || !sas_device_priv_data->sas_target || 3303 ioc->remove_host) { 3304 sdev_printk(KERN_INFO, scmd->device, 3305 "device been deleted! scmd(0x%p)\n", scmd); 3306 scmd->result = DID_NO_CONNECT << 16; 3307 scsi_done(scmd); 3308 r = SUCCESS; 3309 goto out; 3310 } 3311 3312 /* check for completed command */ 3313 if (st == NULL || st->cb_idx == 0xFF) { 3314 sdev_printk(KERN_INFO, scmd->device, "No reference found at " 3315 "driver, assuming scmd(0x%p) might have completed\n", scmd); 3316 scmd->result = DID_RESET << 16; 3317 r = SUCCESS; 3318 goto out; 3319 } 3320 3321 /* for hidden raid components and volumes this is not supported */ 3322 if (sas_device_priv_data->sas_target->flags & 3323 MPT_TARGET_FLAGS_RAID_COMPONENT || 3324 sas_device_priv_data->sas_target->flags & MPT_TARGET_FLAGS_VOLUME) { 3325 scmd->result = DID_RESET << 16; 3326 r = FAILED; 3327 goto out; 3328 } 3329 3330 mpt3sas_halt_firmware(ioc); 3331 3332 handle = sas_device_priv_data->sas_target->handle; 3333 pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle); 3334 if (pcie_device && (!ioc->tm_custom_handling) && 3335 (!(mpt3sas_scsih_is_pcie_scsi_device(pcie_device->device_info)))) 3336 timeout = ioc->nvme_abort_timeout; 3337 r = mpt3sas_scsih_issue_locked_tm(ioc, handle, scmd->device->channel, 3338 scmd->device->id, scmd->device->lun, 3339 MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK, 3340 st->smid, st->msix_io, timeout, 0); 3341 /* Command must be cleared after abort */ 3342 if (r == SUCCESS && st->cb_idx != 0xFF) 3343 r = FAILED; 3344 out: 3345 sdev_printk(KERN_INFO, scmd->device, "task abort: %s scmd(0x%p)\n", 3346 ((r == SUCCESS) ? "SUCCESS" : "FAILED"), scmd); 3347 if (pcie_device) 3348 pcie_device_put(pcie_device); 3349 return r; 3350 } 3351 3352 /** 3353 * scsih_dev_reset - eh threads main device reset routine 3354 * @scmd: pointer to scsi command object 3355 * 3356 * Return: SUCCESS if command aborted else FAILED 3357 */ 3358 static int 3359 scsih_dev_reset(struct scsi_cmnd *scmd) 3360 { 3361 struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host); 3362 struct MPT3SAS_DEVICE *sas_device_priv_data; 3363 struct _sas_device *sas_device = NULL; 3364 struct _pcie_device *pcie_device = NULL; 3365 u16 handle; 3366 u8 tr_method = 0; 3367 u8 tr_timeout = 30; 3368 int r; 3369 3370 struct scsi_target *starget = scmd->device->sdev_target; 3371 struct MPT3SAS_TARGET *target_priv_data = starget->hostdata; 3372 3373 sdev_printk(KERN_INFO, scmd->device, 3374 "attempting device reset! scmd(0x%p)\n", scmd); 3375 _scsih_tm_display_info(ioc, scmd); 3376 3377 sas_device_priv_data = scmd->device->hostdata; 3378 if (!sas_device_priv_data || !sas_device_priv_data->sas_target || 3379 ioc->remove_host) { 3380 sdev_printk(KERN_INFO, scmd->device, 3381 "device been deleted! scmd(0x%p)\n", scmd); 3382 scmd->result = DID_NO_CONNECT << 16; 3383 scsi_done(scmd); 3384 r = SUCCESS; 3385 goto out; 3386 } 3387 3388 /* for hidden raid components obtain the volume_handle */ 3389 handle = 0; 3390 if (sas_device_priv_data->sas_target->flags & 3391 MPT_TARGET_FLAGS_RAID_COMPONENT) { 3392 sas_device = mpt3sas_get_sdev_from_target(ioc, 3393 target_priv_data); 3394 if (sas_device) 3395 handle = sas_device->volume_handle; 3396 } else 3397 handle = sas_device_priv_data->sas_target->handle; 3398 3399 if (!handle) { 3400 scmd->result = DID_RESET << 16; 3401 r = FAILED; 3402 goto out; 3403 } 3404 3405 pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle); 3406 3407 if (pcie_device && (!ioc->tm_custom_handling) && 3408 (!(mpt3sas_scsih_is_pcie_scsi_device(pcie_device->device_info)))) { 3409 tr_timeout = pcie_device->reset_timeout; 3410 tr_method = MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE; 3411 } else 3412 tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET; 3413 3414 r = mpt3sas_scsih_issue_locked_tm(ioc, handle, scmd->device->channel, 3415 scmd->device->id, scmd->device->lun, 3416 MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET, 0, 0, 3417 tr_timeout, tr_method); 3418 /* Check for busy commands after reset */ 3419 if (r == SUCCESS && scsi_device_busy(scmd->device)) 3420 r = FAILED; 3421 out: 3422 sdev_printk(KERN_INFO, scmd->device, "device reset: %s scmd(0x%p)\n", 3423 ((r == SUCCESS) ? "SUCCESS" : "FAILED"), scmd); 3424 3425 if (sas_device) 3426 sas_device_put(sas_device); 3427 if (pcie_device) 3428 pcie_device_put(pcie_device); 3429 3430 return r; 3431 } 3432 3433 /** 3434 * scsih_target_reset - eh threads main target reset routine 3435 * @scmd: pointer to scsi command object 3436 * 3437 * Return: SUCCESS if command aborted else FAILED 3438 */ 3439 static int 3440 scsih_target_reset(struct scsi_cmnd *scmd) 3441 { 3442 struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host); 3443 struct MPT3SAS_DEVICE *sas_device_priv_data; 3444 struct _sas_device *sas_device = NULL; 3445 struct _pcie_device *pcie_device = NULL; 3446 u16 handle; 3447 u8 tr_method = 0; 3448 u8 tr_timeout = 30; 3449 int r; 3450 struct scsi_target *starget = scmd->device->sdev_target; 3451 struct MPT3SAS_TARGET *target_priv_data = starget->hostdata; 3452 3453 starget_printk(KERN_INFO, starget, 3454 "attempting target reset! scmd(0x%p)\n", scmd); 3455 _scsih_tm_display_info(ioc, scmd); 3456 3457 sas_device_priv_data = scmd->device->hostdata; 3458 if (!sas_device_priv_data || !sas_device_priv_data->sas_target || 3459 ioc->remove_host) { 3460 starget_printk(KERN_INFO, starget, 3461 "target been deleted! scmd(0x%p)\n", scmd); 3462 scmd->result = DID_NO_CONNECT << 16; 3463 scsi_done(scmd); 3464 r = SUCCESS; 3465 goto out; 3466 } 3467 3468 /* for hidden raid components obtain the volume_handle */ 3469 handle = 0; 3470 if (sas_device_priv_data->sas_target->flags & 3471 MPT_TARGET_FLAGS_RAID_COMPONENT) { 3472 sas_device = mpt3sas_get_sdev_from_target(ioc, 3473 target_priv_data); 3474 if (sas_device) 3475 handle = sas_device->volume_handle; 3476 } else 3477 handle = sas_device_priv_data->sas_target->handle; 3478 3479 if (!handle) { 3480 scmd->result = DID_RESET << 16; 3481 r = FAILED; 3482 goto out; 3483 } 3484 3485 pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle); 3486 3487 if (pcie_device && (!ioc->tm_custom_handling) && 3488 (!(mpt3sas_scsih_is_pcie_scsi_device(pcie_device->device_info)))) { 3489 tr_timeout = pcie_device->reset_timeout; 3490 tr_method = MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE; 3491 } else 3492 tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET; 3493 r = mpt3sas_scsih_issue_locked_tm(ioc, handle, scmd->device->channel, 3494 scmd->device->id, 0, 3495 MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET, 0, 0, 3496 tr_timeout, tr_method); 3497 /* Check for busy commands after reset */ 3498 if (r == SUCCESS && atomic_read(&starget->target_busy)) 3499 r = FAILED; 3500 out: 3501 starget_printk(KERN_INFO, starget, "target reset: %s scmd(0x%p)\n", 3502 ((r == SUCCESS) ? "SUCCESS" : "FAILED"), scmd); 3503 3504 if (sas_device) 3505 sas_device_put(sas_device); 3506 if (pcie_device) 3507 pcie_device_put(pcie_device); 3508 return r; 3509 } 3510 3511 3512 /** 3513 * scsih_host_reset - eh threads main host reset routine 3514 * @scmd: pointer to scsi command object 3515 * 3516 * Return: SUCCESS if command aborted else FAILED 3517 */ 3518 static int 3519 scsih_host_reset(struct scsi_cmnd *scmd) 3520 { 3521 struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host); 3522 int r, retval; 3523 3524 ioc_info(ioc, "attempting host reset! scmd(0x%p)\n", scmd); 3525 scsi_print_command(scmd); 3526 3527 if (ioc->is_driver_loading || ioc->remove_host) { 3528 ioc_info(ioc, "Blocking the host reset\n"); 3529 r = FAILED; 3530 goto out; 3531 } 3532 3533 retval = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER); 3534 r = (retval < 0) ? FAILED : SUCCESS; 3535 out: 3536 ioc_info(ioc, "host reset: %s scmd(0x%p)\n", 3537 r == SUCCESS ? "SUCCESS" : "FAILED", scmd); 3538 3539 return r; 3540 } 3541 3542 /** 3543 * _scsih_fw_event_add - insert and queue up fw_event 3544 * @ioc: per adapter object 3545 * @fw_event: object describing the event 3546 * Context: This function will acquire ioc->fw_event_lock. 3547 * 3548 * This adds the firmware event object into link list, then queues it up to 3549 * be processed from user context. 3550 */ 3551 static void 3552 _scsih_fw_event_add(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) 3553 { 3554 unsigned long flags; 3555 3556 if (ioc->firmware_event_thread == NULL) 3557 return; 3558 3559 spin_lock_irqsave(&ioc->fw_event_lock, flags); 3560 fw_event_work_get(fw_event); 3561 INIT_LIST_HEAD(&fw_event->list); 3562 list_add_tail(&fw_event->list, &ioc->fw_event_list); 3563 INIT_WORK(&fw_event->work, _firmware_event_work); 3564 fw_event_work_get(fw_event); 3565 queue_work(ioc->firmware_event_thread, &fw_event->work); 3566 spin_unlock_irqrestore(&ioc->fw_event_lock, flags); 3567 } 3568 3569 /** 3570 * _scsih_fw_event_del_from_list - delete fw_event from the list 3571 * @ioc: per adapter object 3572 * @fw_event: object describing the event 3573 * Context: This function will acquire ioc->fw_event_lock. 3574 * 3575 * If the fw_event is on the fw_event_list, remove it and do a put. 3576 */ 3577 static void 3578 _scsih_fw_event_del_from_list(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work 3579 *fw_event) 3580 { 3581 unsigned long flags; 3582 3583 spin_lock_irqsave(&ioc->fw_event_lock, flags); 3584 if (!list_empty(&fw_event->list)) { 3585 list_del_init(&fw_event->list); 3586 fw_event_work_put(fw_event); 3587 } 3588 spin_unlock_irqrestore(&ioc->fw_event_lock, flags); 3589 } 3590 3591 3592 /** 3593 * mpt3sas_send_trigger_data_event - send event for processing trigger data 3594 * @ioc: per adapter object 3595 * @event_data: trigger event data 3596 */ 3597 void 3598 mpt3sas_send_trigger_data_event(struct MPT3SAS_ADAPTER *ioc, 3599 struct SL_WH_TRIGGERS_EVENT_DATA_T *event_data) 3600 { 3601 struct fw_event_work *fw_event; 3602 u16 sz; 3603 3604 if (ioc->is_driver_loading) 3605 return; 3606 sz = sizeof(*event_data); 3607 fw_event = alloc_fw_event_work(sz); 3608 if (!fw_event) 3609 return; 3610 fw_event->event = MPT3SAS_PROCESS_TRIGGER_DIAG; 3611 fw_event->ioc = ioc; 3612 memcpy(fw_event->event_data, event_data, sizeof(*event_data)); 3613 _scsih_fw_event_add(ioc, fw_event); 3614 fw_event_work_put(fw_event); 3615 } 3616 3617 /** 3618 * _scsih_error_recovery_delete_devices - remove devices not responding 3619 * @ioc: per adapter object 3620 */ 3621 static void 3622 _scsih_error_recovery_delete_devices(struct MPT3SAS_ADAPTER *ioc) 3623 { 3624 struct fw_event_work *fw_event; 3625 3626 fw_event = alloc_fw_event_work(0); 3627 if (!fw_event) 3628 return; 3629 fw_event->event = MPT3SAS_REMOVE_UNRESPONDING_DEVICES; 3630 fw_event->ioc = ioc; 3631 _scsih_fw_event_add(ioc, fw_event); 3632 fw_event_work_put(fw_event); 3633 } 3634 3635 /** 3636 * mpt3sas_port_enable_complete - port enable completed (fake event) 3637 * @ioc: per adapter object 3638 */ 3639 void 3640 mpt3sas_port_enable_complete(struct MPT3SAS_ADAPTER *ioc) 3641 { 3642 struct fw_event_work *fw_event; 3643 3644 fw_event = alloc_fw_event_work(0); 3645 if (!fw_event) 3646 return; 3647 fw_event->event = MPT3SAS_PORT_ENABLE_COMPLETE; 3648 fw_event->ioc = ioc; 3649 _scsih_fw_event_add(ioc, fw_event); 3650 fw_event_work_put(fw_event); 3651 } 3652 3653 static struct fw_event_work *dequeue_next_fw_event(struct MPT3SAS_ADAPTER *ioc) 3654 { 3655 unsigned long flags; 3656 struct fw_event_work *fw_event = NULL; 3657 3658 spin_lock_irqsave(&ioc->fw_event_lock, flags); 3659 if (!list_empty(&ioc->fw_event_list)) { 3660 fw_event = list_first_entry(&ioc->fw_event_list, 3661 struct fw_event_work, list); 3662 list_del_init(&fw_event->list); 3663 fw_event_work_put(fw_event); 3664 } 3665 spin_unlock_irqrestore(&ioc->fw_event_lock, flags); 3666 3667 return fw_event; 3668 } 3669 3670 /** 3671 * _scsih_fw_event_cleanup_queue - cleanup event queue 3672 * @ioc: per adapter object 3673 * 3674 * Walk the firmware event queue, either killing timers, or waiting 3675 * for outstanding events to complete 3676 * 3677 * Context: task, can sleep 3678 */ 3679 static void 3680 _scsih_fw_event_cleanup_queue(struct MPT3SAS_ADAPTER *ioc) 3681 { 3682 struct fw_event_work *fw_event; 3683 3684 if ((list_empty(&ioc->fw_event_list) && !ioc->current_event) || 3685 !ioc->firmware_event_thread) 3686 return; 3687 /* 3688 * Set current running event as ignore, so that 3689 * current running event will exit quickly. 3690 * As diag reset has occurred it is of no use 3691 * to process remaining stale event data entries. 3692 */ 3693 if (ioc->shost_recovery && ioc->current_event) 3694 ioc->current_event->ignore = 1; 3695 3696 ioc->fw_events_cleanup = 1; 3697 while ((fw_event = dequeue_next_fw_event(ioc)) || 3698 (fw_event = ioc->current_event)) { 3699 3700 /* 3701 * Don't call cancel_work_sync() for current_event 3702 * other than MPT3SAS_REMOVE_UNRESPONDING_DEVICES; 3703 * otherwise we may observe deadlock if current 3704 * hard reset issued as part of processing the current_event. 3705 * 3706 * Orginal logic of cleaning the current_event is added 3707 * for handling the back to back host reset issued by the user. 3708 * i.e. during back to back host reset, driver use to process 3709 * the two instances of MPT3SAS_REMOVE_UNRESPONDING_DEVICES 3710 * event back to back and this made the drives to unregister 3711 * the devices from SML. 3712 */ 3713 3714 if (fw_event == ioc->current_event && 3715 ioc->current_event->event != 3716 MPT3SAS_REMOVE_UNRESPONDING_DEVICES) { 3717 ioc->current_event = NULL; 3718 continue; 3719 } 3720 3721 /* 3722 * Driver has to clear ioc->start_scan flag when 3723 * it is cleaning up MPT3SAS_PORT_ENABLE_COMPLETE, 3724 * otherwise scsi_scan_host() API waits for the 3725 * 5 minute timer to expire. If we exit from 3726 * scsi_scan_host() early then we can issue the 3727 * new port enable request as part of current diag reset. 3728 */ 3729 if (fw_event->event == MPT3SAS_PORT_ENABLE_COMPLETE) { 3730 ioc->port_enable_cmds.status |= MPT3_CMD_RESET; 3731 ioc->start_scan = 0; 3732 } 3733 3734 /* 3735 * Wait on the fw_event to complete. If this returns 1, then 3736 * the event was never executed, and we need a put for the 3737 * reference the work had on the fw_event. 3738 * 3739 * If it did execute, we wait for it to finish, and the put will 3740 * happen from _firmware_event_work() 3741 */ 3742 if (cancel_work_sync(&fw_event->work)) 3743 fw_event_work_put(fw_event); 3744 3745 } 3746 ioc->fw_events_cleanup = 0; 3747 } 3748 3749 /** 3750 * _scsih_internal_device_block - block the sdev device 3751 * @sdev: per device object 3752 * @sas_device_priv_data : per device driver private data 3753 * 3754 * make sure device is blocked without error, if not 3755 * print an error 3756 */ 3757 static void 3758 _scsih_internal_device_block(struct scsi_device *sdev, 3759 struct MPT3SAS_DEVICE *sas_device_priv_data) 3760 { 3761 int r = 0; 3762 3763 sdev_printk(KERN_INFO, sdev, "device_block, handle(0x%04x)\n", 3764 sas_device_priv_data->sas_target->handle); 3765 sas_device_priv_data->block = 1; 3766 3767 r = scsi_internal_device_block_nowait(sdev); 3768 if (r == -EINVAL) 3769 sdev_printk(KERN_WARNING, sdev, 3770 "device_block failed with return(%d) for handle(0x%04x)\n", 3771 r, sas_device_priv_data->sas_target->handle); 3772 } 3773 3774 /** 3775 * _scsih_internal_device_unblock - unblock the sdev device 3776 * @sdev: per device object 3777 * @sas_device_priv_data : per device driver private data 3778 * make sure device is unblocked without error, if not retry 3779 * by blocking and then unblocking 3780 */ 3781 3782 static void 3783 _scsih_internal_device_unblock(struct scsi_device *sdev, 3784 struct MPT3SAS_DEVICE *sas_device_priv_data) 3785 { 3786 int r = 0; 3787 3788 sdev_printk(KERN_WARNING, sdev, "device_unblock and setting to running, " 3789 "handle(0x%04x)\n", sas_device_priv_data->sas_target->handle); 3790 sas_device_priv_data->block = 0; 3791 r = scsi_internal_device_unblock_nowait(sdev, SDEV_RUNNING); 3792 if (r == -EINVAL) { 3793 /* The device has been set to SDEV_RUNNING by SD layer during 3794 * device addition but the request queue is still stopped by 3795 * our earlier block call. We need to perform a block again 3796 * to get the device to SDEV_BLOCK and then to SDEV_RUNNING */ 3797 3798 sdev_printk(KERN_WARNING, sdev, 3799 "device_unblock failed with return(%d) for handle(0x%04x) " 3800 "performing a block followed by an unblock\n", 3801 r, sas_device_priv_data->sas_target->handle); 3802 sas_device_priv_data->block = 1; 3803 r = scsi_internal_device_block_nowait(sdev); 3804 if (r) 3805 sdev_printk(KERN_WARNING, sdev, "retried device_block " 3806 "failed with return(%d) for handle(0x%04x)\n", 3807 r, sas_device_priv_data->sas_target->handle); 3808 3809 sas_device_priv_data->block = 0; 3810 r = scsi_internal_device_unblock_nowait(sdev, SDEV_RUNNING); 3811 if (r) 3812 sdev_printk(KERN_WARNING, sdev, "retried device_unblock" 3813 " failed with return(%d) for handle(0x%04x)\n", 3814 r, sas_device_priv_data->sas_target->handle); 3815 } 3816 } 3817 3818 /** 3819 * _scsih_ublock_io_all_device - unblock every device 3820 * @ioc: per adapter object 3821 * 3822 * change the device state from block to running 3823 */ 3824 static void 3825 _scsih_ublock_io_all_device(struct MPT3SAS_ADAPTER *ioc) 3826 { 3827 struct MPT3SAS_DEVICE *sas_device_priv_data; 3828 struct scsi_device *sdev; 3829 3830 shost_for_each_device(sdev, ioc->shost) { 3831 sas_device_priv_data = sdev->hostdata; 3832 if (!sas_device_priv_data) 3833 continue; 3834 if (!sas_device_priv_data->block) 3835 continue; 3836 3837 dewtprintk(ioc, sdev_printk(KERN_INFO, sdev, 3838 "device_running, handle(0x%04x)\n", 3839 sas_device_priv_data->sas_target->handle)); 3840 _scsih_internal_device_unblock(sdev, sas_device_priv_data); 3841 } 3842 } 3843 3844 3845 /** 3846 * _scsih_ublock_io_device - prepare device to be deleted 3847 * @ioc: per adapter object 3848 * @sas_address: sas address 3849 * @port: hba port entry 3850 * 3851 * unblock then put device in offline state 3852 */ 3853 static void 3854 _scsih_ublock_io_device(struct MPT3SAS_ADAPTER *ioc, 3855 u64 sas_address, struct hba_port *port) 3856 { 3857 struct MPT3SAS_DEVICE *sas_device_priv_data; 3858 struct scsi_device *sdev; 3859 3860 shost_for_each_device(sdev, ioc->shost) { 3861 sas_device_priv_data = sdev->hostdata; 3862 if (!sas_device_priv_data || !sas_device_priv_data->sas_target) 3863 continue; 3864 if (sas_device_priv_data->sas_target->sas_address 3865 != sas_address) 3866 continue; 3867 if (sas_device_priv_data->sas_target->port != port) 3868 continue; 3869 if (sas_device_priv_data->block) 3870 _scsih_internal_device_unblock(sdev, 3871 sas_device_priv_data); 3872 } 3873 } 3874 3875 /** 3876 * _scsih_block_io_all_device - set the device state to SDEV_BLOCK 3877 * @ioc: per adapter object 3878 * 3879 * During device pull we need to appropriately set the sdev state. 3880 */ 3881 static void 3882 _scsih_block_io_all_device(struct MPT3SAS_ADAPTER *ioc) 3883 { 3884 struct MPT3SAS_DEVICE *sas_device_priv_data; 3885 struct scsi_device *sdev; 3886 3887 shost_for_each_device(sdev, ioc->shost) { 3888 sas_device_priv_data = sdev->hostdata; 3889 if (!sas_device_priv_data) 3890 continue; 3891 if (sas_device_priv_data->block) 3892 continue; 3893 if (sas_device_priv_data->ignore_delay_remove) { 3894 sdev_printk(KERN_INFO, sdev, 3895 "%s skip device_block for SES handle(0x%04x)\n", 3896 __func__, sas_device_priv_data->sas_target->handle); 3897 continue; 3898 } 3899 _scsih_internal_device_block(sdev, sas_device_priv_data); 3900 } 3901 } 3902 3903 /** 3904 * _scsih_block_io_device - set the device state to SDEV_BLOCK 3905 * @ioc: per adapter object 3906 * @handle: device handle 3907 * 3908 * During device pull we need to appropriately set the sdev state. 3909 */ 3910 static void 3911 _scsih_block_io_device(struct MPT3SAS_ADAPTER *ioc, u16 handle) 3912 { 3913 struct MPT3SAS_DEVICE *sas_device_priv_data; 3914 struct scsi_device *sdev; 3915 struct _sas_device *sas_device; 3916 3917 sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); 3918 3919 shost_for_each_device(sdev, ioc->shost) { 3920 sas_device_priv_data = sdev->hostdata; 3921 if (!sas_device_priv_data) 3922 continue; 3923 if (sas_device_priv_data->sas_target->handle != handle) 3924 continue; 3925 if (sas_device_priv_data->block) 3926 continue; 3927 if (sas_device && sas_device->pend_sas_rphy_add) 3928 continue; 3929 if (sas_device_priv_data->ignore_delay_remove) { 3930 sdev_printk(KERN_INFO, sdev, 3931 "%s skip device_block for SES handle(0x%04x)\n", 3932 __func__, sas_device_priv_data->sas_target->handle); 3933 continue; 3934 } 3935 _scsih_internal_device_block(sdev, sas_device_priv_data); 3936 } 3937 3938 if (sas_device) 3939 sas_device_put(sas_device); 3940 } 3941 3942 /** 3943 * _scsih_block_io_to_children_attached_to_ex 3944 * @ioc: per adapter object 3945 * @sas_expander: the sas_device object 3946 * 3947 * This routine set sdev state to SDEV_BLOCK for all devices 3948 * attached to this expander. This function called when expander is 3949 * pulled. 3950 */ 3951 static void 3952 _scsih_block_io_to_children_attached_to_ex(struct MPT3SAS_ADAPTER *ioc, 3953 struct _sas_node *sas_expander) 3954 { 3955 struct _sas_port *mpt3sas_port; 3956 struct _sas_device *sas_device; 3957 struct _sas_node *expander_sibling; 3958 unsigned long flags; 3959 3960 if (!sas_expander) 3961 return; 3962 3963 list_for_each_entry(mpt3sas_port, 3964 &sas_expander->sas_port_list, port_list) { 3965 if (mpt3sas_port->remote_identify.device_type == 3966 SAS_END_DEVICE) { 3967 spin_lock_irqsave(&ioc->sas_device_lock, flags); 3968 sas_device = __mpt3sas_get_sdev_by_addr(ioc, 3969 mpt3sas_port->remote_identify.sas_address, 3970 mpt3sas_port->hba_port); 3971 if (sas_device) { 3972 set_bit(sas_device->handle, 3973 ioc->blocking_handles); 3974 sas_device_put(sas_device); 3975 } 3976 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 3977 } 3978 } 3979 3980 list_for_each_entry(mpt3sas_port, 3981 &sas_expander->sas_port_list, port_list) { 3982 3983 if (mpt3sas_port->remote_identify.device_type == 3984 SAS_EDGE_EXPANDER_DEVICE || 3985 mpt3sas_port->remote_identify.device_type == 3986 SAS_FANOUT_EXPANDER_DEVICE) { 3987 expander_sibling = 3988 mpt3sas_scsih_expander_find_by_sas_address( 3989 ioc, mpt3sas_port->remote_identify.sas_address, 3990 mpt3sas_port->hba_port); 3991 _scsih_block_io_to_children_attached_to_ex(ioc, 3992 expander_sibling); 3993 } 3994 } 3995 } 3996 3997 /** 3998 * _scsih_block_io_to_children_attached_directly 3999 * @ioc: per adapter object 4000 * @event_data: topology change event data 4001 * 4002 * This routine set sdev state to SDEV_BLOCK for all devices 4003 * direct attached during device pull. 4004 */ 4005 static void 4006 _scsih_block_io_to_children_attached_directly(struct MPT3SAS_ADAPTER *ioc, 4007 Mpi2EventDataSasTopologyChangeList_t *event_data) 4008 { 4009 int i; 4010 u16 handle; 4011 u16 reason_code; 4012 4013 for (i = 0; i < event_data->NumEntries; i++) { 4014 handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); 4015 if (!handle) 4016 continue; 4017 reason_code = event_data->PHY[i].PhyStatus & 4018 MPI2_EVENT_SAS_TOPO_RC_MASK; 4019 if (reason_code == MPI2_EVENT_SAS_TOPO_RC_DELAY_NOT_RESPONDING) 4020 _scsih_block_io_device(ioc, handle); 4021 } 4022 } 4023 4024 /** 4025 * _scsih_block_io_to_pcie_children_attached_directly 4026 * @ioc: per adapter object 4027 * @event_data: topology change event data 4028 * 4029 * This routine set sdev state to SDEV_BLOCK for all devices 4030 * direct attached during device pull/reconnect. 4031 */ 4032 static void 4033 _scsih_block_io_to_pcie_children_attached_directly(struct MPT3SAS_ADAPTER *ioc, 4034 Mpi26EventDataPCIeTopologyChangeList_t *event_data) 4035 { 4036 int i; 4037 u16 handle; 4038 u16 reason_code; 4039 4040 for (i = 0; i < event_data->NumEntries; i++) { 4041 handle = 4042 le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle); 4043 if (!handle) 4044 continue; 4045 reason_code = event_data->PortEntry[i].PortStatus; 4046 if (reason_code == 4047 MPI26_EVENT_PCIE_TOPO_PS_DELAY_NOT_RESPONDING) 4048 _scsih_block_io_device(ioc, handle); 4049 } 4050 } 4051 /** 4052 * _scsih_tm_tr_send - send task management request 4053 * @ioc: per adapter object 4054 * @handle: device handle 4055 * Context: interrupt time. 4056 * 4057 * This code is to initiate the device removal handshake protocol 4058 * with controller firmware. This function will issue target reset 4059 * using high priority request queue. It will send a sas iounit 4060 * control request (MPI2_SAS_OP_REMOVE_DEVICE) from this completion. 4061 * 4062 * This is designed to send muliple task management request at the same 4063 * time to the fifo. If the fifo is full, we will append the request, 4064 * and process it in a future completion. 4065 */ 4066 static void 4067 _scsih_tm_tr_send(struct MPT3SAS_ADAPTER *ioc, u16 handle) 4068 { 4069 Mpi2SCSITaskManagementRequest_t *mpi_request; 4070 u16 smid; 4071 struct _sas_device *sas_device = NULL; 4072 struct _pcie_device *pcie_device = NULL; 4073 struct MPT3SAS_TARGET *sas_target_priv_data = NULL; 4074 u64 sas_address = 0; 4075 unsigned long flags; 4076 struct _tr_list *delayed_tr; 4077 u32 ioc_state; 4078 u8 tr_method = 0; 4079 struct hba_port *port = NULL; 4080 4081 if (ioc->pci_error_recovery) { 4082 dewtprintk(ioc, 4083 ioc_info(ioc, "%s: host in pci error recovery: handle(0x%04x)\n", 4084 __func__, handle)); 4085 return; 4086 } 4087 ioc_state = mpt3sas_base_get_iocstate(ioc, 1); 4088 if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) { 4089 dewtprintk(ioc, 4090 ioc_info(ioc, "%s: host is not operational: handle(0x%04x)\n", 4091 __func__, handle)); 4092 return; 4093 } 4094 4095 /* if PD, then return */ 4096 if (test_bit(handle, ioc->pd_handles)) 4097 return; 4098 4099 clear_bit(handle, ioc->pend_os_device_add); 4100 4101 spin_lock_irqsave(&ioc->sas_device_lock, flags); 4102 sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); 4103 if (sas_device && sas_device->starget && 4104 sas_device->starget->hostdata) { 4105 sas_target_priv_data = sas_device->starget->hostdata; 4106 sas_target_priv_data->deleted = 1; 4107 sas_address = sas_device->sas_address; 4108 port = sas_device->port; 4109 } 4110 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 4111 if (!sas_device) { 4112 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 4113 pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle); 4114 if (pcie_device && pcie_device->starget && 4115 pcie_device->starget->hostdata) { 4116 sas_target_priv_data = pcie_device->starget->hostdata; 4117 sas_target_priv_data->deleted = 1; 4118 sas_address = pcie_device->wwid; 4119 } 4120 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 4121 if (pcie_device && (!ioc->tm_custom_handling) && 4122 (!(mpt3sas_scsih_is_pcie_scsi_device( 4123 pcie_device->device_info)))) 4124 tr_method = 4125 MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE; 4126 else 4127 tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET; 4128 } 4129 if (sas_target_priv_data) { 4130 dewtprintk(ioc, 4131 ioc_info(ioc, "setting delete flag: handle(0x%04x), sas_addr(0x%016llx)\n", 4132 handle, (u64)sas_address)); 4133 if (sas_device) { 4134 if (sas_device->enclosure_handle != 0) 4135 dewtprintk(ioc, 4136 ioc_info(ioc, "setting delete flag:enclosure logical id(0x%016llx), slot(%d)\n", 4137 (u64)sas_device->enclosure_logical_id, 4138 sas_device->slot)); 4139 if (sas_device->connector_name[0] != '\0') 4140 dewtprintk(ioc, 4141 ioc_info(ioc, "setting delete flag: enclosure level(0x%04x), connector name( %s)\n", 4142 sas_device->enclosure_level, 4143 sas_device->connector_name)); 4144 } else if (pcie_device) { 4145 if (pcie_device->enclosure_handle != 0) 4146 dewtprintk(ioc, 4147 ioc_info(ioc, "setting delete flag: logical id(0x%016llx), slot(%d)\n", 4148 (u64)pcie_device->enclosure_logical_id, 4149 pcie_device->slot)); 4150 if (pcie_device->connector_name[0] != '\0') 4151 dewtprintk(ioc, 4152 ioc_info(ioc, "setting delete flag:, enclosure level(0x%04x), connector name( %s)\n", 4153 pcie_device->enclosure_level, 4154 pcie_device->connector_name)); 4155 } 4156 _scsih_ublock_io_device(ioc, sas_address, port); 4157 sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE; 4158 } 4159 4160 smid = mpt3sas_base_get_smid_hpr(ioc, ioc->tm_tr_cb_idx); 4161 if (!smid) { 4162 delayed_tr = kzalloc(sizeof(*delayed_tr), GFP_ATOMIC); 4163 if (!delayed_tr) 4164 goto out; 4165 INIT_LIST_HEAD(&delayed_tr->list); 4166 delayed_tr->handle = handle; 4167 list_add_tail(&delayed_tr->list, &ioc->delayed_tr_list); 4168 dewtprintk(ioc, 4169 ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n", 4170 handle)); 4171 goto out; 4172 } 4173 4174 dewtprintk(ioc, 4175 ioc_info(ioc, "tr_send:handle(0x%04x), (open), smid(%d), cb(%d)\n", 4176 handle, smid, ioc->tm_tr_cb_idx)); 4177 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); 4178 memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t)); 4179 mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT; 4180 mpi_request->DevHandle = cpu_to_le16(handle); 4181 mpi_request->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET; 4182 mpi_request->MsgFlags = tr_method; 4183 set_bit(handle, ioc->device_remove_in_progress); 4184 ioc->put_smid_hi_priority(ioc, smid, 0); 4185 mpt3sas_trigger_master(ioc, MASTER_TRIGGER_DEVICE_REMOVAL); 4186 4187 out: 4188 if (sas_device) 4189 sas_device_put(sas_device); 4190 if (pcie_device) 4191 pcie_device_put(pcie_device); 4192 } 4193 4194 /** 4195 * _scsih_tm_tr_complete - 4196 * @ioc: per adapter object 4197 * @smid: system request message index 4198 * @msix_index: MSIX table index supplied by the OS 4199 * @reply: reply message frame(lower 32bit addr) 4200 * Context: interrupt time. 4201 * 4202 * This is the target reset completion routine. 4203 * This code is part of the code to initiate the device removal 4204 * handshake protocol with controller firmware. 4205 * It will send a sas iounit control request (MPI2_SAS_OP_REMOVE_DEVICE) 4206 * 4207 * Return: 1 meaning mf should be freed from _base_interrupt 4208 * 0 means the mf is freed from this function. 4209 */ 4210 static u8 4211 _scsih_tm_tr_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, 4212 u32 reply) 4213 { 4214 u16 handle; 4215 Mpi2SCSITaskManagementRequest_t *mpi_request_tm; 4216 Mpi2SCSITaskManagementReply_t *mpi_reply = 4217 mpt3sas_base_get_reply_virt_addr(ioc, reply); 4218 Mpi2SasIoUnitControlRequest_t *mpi_request; 4219 u16 smid_sas_ctrl; 4220 u32 ioc_state; 4221 struct _sc_list *delayed_sc; 4222 4223 if (ioc->pci_error_recovery) { 4224 dewtprintk(ioc, 4225 ioc_info(ioc, "%s: host in pci error recovery\n", 4226 __func__)); 4227 return 1; 4228 } 4229 ioc_state = mpt3sas_base_get_iocstate(ioc, 1); 4230 if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) { 4231 dewtprintk(ioc, 4232 ioc_info(ioc, "%s: host is not operational\n", 4233 __func__)); 4234 return 1; 4235 } 4236 if (unlikely(!mpi_reply)) { 4237 ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n", 4238 __FILE__, __LINE__, __func__); 4239 return 1; 4240 } 4241 mpi_request_tm = mpt3sas_base_get_msg_frame(ioc, smid); 4242 handle = le16_to_cpu(mpi_request_tm->DevHandle); 4243 if (handle != le16_to_cpu(mpi_reply->DevHandle)) { 4244 dewtprintk(ioc, 4245 ioc_err(ioc, "spurious interrupt: handle(0x%04x:0x%04x), smid(%d)!!!\n", 4246 handle, 4247 le16_to_cpu(mpi_reply->DevHandle), smid)); 4248 return 0; 4249 } 4250 4251 mpt3sas_trigger_master(ioc, MASTER_TRIGGER_TASK_MANAGMENT); 4252 dewtprintk(ioc, 4253 ioc_info(ioc, "tr_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x), completed(%d)\n", 4254 handle, smid, le16_to_cpu(mpi_reply->IOCStatus), 4255 le32_to_cpu(mpi_reply->IOCLogInfo), 4256 le32_to_cpu(mpi_reply->TerminationCount))); 4257 4258 smid_sas_ctrl = mpt3sas_base_get_smid(ioc, ioc->tm_sas_control_cb_idx); 4259 if (!smid_sas_ctrl) { 4260 delayed_sc = kzalloc(sizeof(*delayed_sc), GFP_ATOMIC); 4261 if (!delayed_sc) 4262 return _scsih_check_for_pending_tm(ioc, smid); 4263 INIT_LIST_HEAD(&delayed_sc->list); 4264 delayed_sc->handle = le16_to_cpu(mpi_request_tm->DevHandle); 4265 list_add_tail(&delayed_sc->list, &ioc->delayed_sc_list); 4266 dewtprintk(ioc, 4267 ioc_info(ioc, "DELAYED:sc:handle(0x%04x), (open)\n", 4268 handle)); 4269 return _scsih_check_for_pending_tm(ioc, smid); 4270 } 4271 4272 dewtprintk(ioc, 4273 ioc_info(ioc, "sc_send:handle(0x%04x), (open), smid(%d), cb(%d)\n", 4274 handle, smid_sas_ctrl, ioc->tm_sas_control_cb_idx)); 4275 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid_sas_ctrl); 4276 memset(mpi_request, 0, sizeof(Mpi2SasIoUnitControlRequest_t)); 4277 mpi_request->Function = MPI2_FUNCTION_SAS_IO_UNIT_CONTROL; 4278 mpi_request->Operation = MPI2_SAS_OP_REMOVE_DEVICE; 4279 mpi_request->DevHandle = mpi_request_tm->DevHandle; 4280 ioc->put_smid_default(ioc, smid_sas_ctrl); 4281 4282 return _scsih_check_for_pending_tm(ioc, smid); 4283 } 4284 4285 /** _scsih_allow_scmd_to_device - check whether scmd needs to 4286 * issue to IOC or not. 4287 * @ioc: per adapter object 4288 * @scmd: pointer to scsi command object 4289 * 4290 * Returns true if scmd can be issued to IOC otherwise returns false. 4291 */ 4292 inline bool _scsih_allow_scmd_to_device(struct MPT3SAS_ADAPTER *ioc, 4293 struct scsi_cmnd *scmd) 4294 { 4295 4296 if (ioc->pci_error_recovery) 4297 return false; 4298 4299 if (ioc->hba_mpi_version_belonged == MPI2_VERSION) { 4300 if (ioc->remove_host) 4301 return false; 4302 4303 return true; 4304 } 4305 4306 if (ioc->remove_host) { 4307 4308 switch (scmd->cmnd[0]) { 4309 case SYNCHRONIZE_CACHE: 4310 case START_STOP: 4311 return true; 4312 default: 4313 return false; 4314 } 4315 } 4316 4317 return true; 4318 } 4319 4320 /** 4321 * _scsih_sas_control_complete - completion routine 4322 * @ioc: per adapter object 4323 * @smid: system request message index 4324 * @msix_index: MSIX table index supplied by the OS 4325 * @reply: reply message frame(lower 32bit addr) 4326 * Context: interrupt time. 4327 * 4328 * This is the sas iounit control completion routine. 4329 * This code is part of the code to initiate the device removal 4330 * handshake protocol with controller firmware. 4331 * 4332 * Return: 1 meaning mf should be freed from _base_interrupt 4333 * 0 means the mf is freed from this function. 4334 */ 4335 static u8 4336 _scsih_sas_control_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid, 4337 u8 msix_index, u32 reply) 4338 { 4339 Mpi2SasIoUnitControlReply_t *mpi_reply = 4340 mpt3sas_base_get_reply_virt_addr(ioc, reply); 4341 4342 if (likely(mpi_reply)) { 4343 dewtprintk(ioc, 4344 ioc_info(ioc, "sc_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x)\n", 4345 le16_to_cpu(mpi_reply->DevHandle), smid, 4346 le16_to_cpu(mpi_reply->IOCStatus), 4347 le32_to_cpu(mpi_reply->IOCLogInfo))); 4348 if (le16_to_cpu(mpi_reply->IOCStatus) == 4349 MPI2_IOCSTATUS_SUCCESS) { 4350 clear_bit(le16_to_cpu(mpi_reply->DevHandle), 4351 ioc->device_remove_in_progress); 4352 } 4353 } else { 4354 ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n", 4355 __FILE__, __LINE__, __func__); 4356 } 4357 return mpt3sas_check_for_pending_internal_cmds(ioc, smid); 4358 } 4359 4360 /** 4361 * _scsih_tm_tr_volume_send - send target reset request for volumes 4362 * @ioc: per adapter object 4363 * @handle: device handle 4364 * Context: interrupt time. 4365 * 4366 * This is designed to send muliple task management request at the same 4367 * time to the fifo. If the fifo is full, we will append the request, 4368 * and process it in a future completion. 4369 */ 4370 static void 4371 _scsih_tm_tr_volume_send(struct MPT3SAS_ADAPTER *ioc, u16 handle) 4372 { 4373 Mpi2SCSITaskManagementRequest_t *mpi_request; 4374 u16 smid; 4375 struct _tr_list *delayed_tr; 4376 4377 if (ioc->pci_error_recovery) { 4378 dewtprintk(ioc, 4379 ioc_info(ioc, "%s: host reset in progress!\n", 4380 __func__)); 4381 return; 4382 } 4383 4384 smid = mpt3sas_base_get_smid_hpr(ioc, ioc->tm_tr_volume_cb_idx); 4385 if (!smid) { 4386 delayed_tr = kzalloc(sizeof(*delayed_tr), GFP_ATOMIC); 4387 if (!delayed_tr) 4388 return; 4389 INIT_LIST_HEAD(&delayed_tr->list); 4390 delayed_tr->handle = handle; 4391 list_add_tail(&delayed_tr->list, &ioc->delayed_tr_volume_list); 4392 dewtprintk(ioc, 4393 ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n", 4394 handle)); 4395 return; 4396 } 4397 4398 dewtprintk(ioc, 4399 ioc_info(ioc, "tr_send:handle(0x%04x), (open), smid(%d), cb(%d)\n", 4400 handle, smid, ioc->tm_tr_volume_cb_idx)); 4401 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); 4402 memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t)); 4403 mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT; 4404 mpi_request->DevHandle = cpu_to_le16(handle); 4405 mpi_request->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET; 4406 ioc->put_smid_hi_priority(ioc, smid, 0); 4407 } 4408 4409 /** 4410 * _scsih_tm_volume_tr_complete - target reset completion 4411 * @ioc: per adapter object 4412 * @smid: system request message index 4413 * @msix_index: MSIX table index supplied by the OS 4414 * @reply: reply message frame(lower 32bit addr) 4415 * Context: interrupt time. 4416 * 4417 * Return: 1 meaning mf should be freed from _base_interrupt 4418 * 0 means the mf is freed from this function. 4419 */ 4420 static u8 4421 _scsih_tm_volume_tr_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid, 4422 u8 msix_index, u32 reply) 4423 { 4424 u16 handle; 4425 Mpi2SCSITaskManagementRequest_t *mpi_request_tm; 4426 Mpi2SCSITaskManagementReply_t *mpi_reply = 4427 mpt3sas_base_get_reply_virt_addr(ioc, reply); 4428 4429 if (ioc->shost_recovery || ioc->pci_error_recovery) { 4430 dewtprintk(ioc, 4431 ioc_info(ioc, "%s: host reset in progress!\n", 4432 __func__)); 4433 return 1; 4434 } 4435 if (unlikely(!mpi_reply)) { 4436 ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n", 4437 __FILE__, __LINE__, __func__); 4438 return 1; 4439 } 4440 4441 mpi_request_tm = mpt3sas_base_get_msg_frame(ioc, smid); 4442 handle = le16_to_cpu(mpi_request_tm->DevHandle); 4443 if (handle != le16_to_cpu(mpi_reply->DevHandle)) { 4444 dewtprintk(ioc, 4445 ioc_err(ioc, "spurious interrupt: handle(0x%04x:0x%04x), smid(%d)!!!\n", 4446 handle, le16_to_cpu(mpi_reply->DevHandle), 4447 smid)); 4448 return 0; 4449 } 4450 4451 dewtprintk(ioc, 4452 ioc_info(ioc, "tr_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x), completed(%d)\n", 4453 handle, smid, le16_to_cpu(mpi_reply->IOCStatus), 4454 le32_to_cpu(mpi_reply->IOCLogInfo), 4455 le32_to_cpu(mpi_reply->TerminationCount))); 4456 4457 return _scsih_check_for_pending_tm(ioc, smid); 4458 } 4459 4460 /** 4461 * _scsih_issue_delayed_event_ack - issue delayed Event ACK messages 4462 * @ioc: per adapter object 4463 * @smid: system request message index 4464 * @event: Event ID 4465 * @event_context: used to track events uniquely 4466 * 4467 * Context - processed in interrupt context. 4468 */ 4469 static void 4470 _scsih_issue_delayed_event_ack(struct MPT3SAS_ADAPTER *ioc, u16 smid, U16 event, 4471 U32 event_context) 4472 { 4473 Mpi2EventAckRequest_t *ack_request; 4474 int i = smid - ioc->internal_smid; 4475 unsigned long flags; 4476 4477 /* Without releasing the smid just update the 4478 * call back index and reuse the same smid for 4479 * processing this delayed request 4480 */ 4481 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); 4482 ioc->internal_lookup[i].cb_idx = ioc->base_cb_idx; 4483 spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags); 4484 4485 dewtprintk(ioc, 4486 ioc_info(ioc, "EVENT ACK: event(0x%04x), smid(%d), cb(%d)\n", 4487 le16_to_cpu(event), smid, ioc->base_cb_idx)); 4488 ack_request = mpt3sas_base_get_msg_frame(ioc, smid); 4489 memset(ack_request, 0, sizeof(Mpi2EventAckRequest_t)); 4490 ack_request->Function = MPI2_FUNCTION_EVENT_ACK; 4491 ack_request->Event = event; 4492 ack_request->EventContext = event_context; 4493 ack_request->VF_ID = 0; /* TODO */ 4494 ack_request->VP_ID = 0; 4495 ioc->put_smid_default(ioc, smid); 4496 } 4497 4498 /** 4499 * _scsih_issue_delayed_sas_io_unit_ctrl - issue delayed 4500 * sas_io_unit_ctrl messages 4501 * @ioc: per adapter object 4502 * @smid: system request message index 4503 * @handle: device handle 4504 * 4505 * Context - processed in interrupt context. 4506 */ 4507 static void 4508 _scsih_issue_delayed_sas_io_unit_ctrl(struct MPT3SAS_ADAPTER *ioc, 4509 u16 smid, u16 handle) 4510 { 4511 Mpi2SasIoUnitControlRequest_t *mpi_request; 4512 u32 ioc_state; 4513 int i = smid - ioc->internal_smid; 4514 unsigned long flags; 4515 4516 if (ioc->remove_host) { 4517 dewtprintk(ioc, 4518 ioc_info(ioc, "%s: host has been removed\n", 4519 __func__)); 4520 return; 4521 } else if (ioc->pci_error_recovery) { 4522 dewtprintk(ioc, 4523 ioc_info(ioc, "%s: host in pci error recovery\n", 4524 __func__)); 4525 return; 4526 } 4527 ioc_state = mpt3sas_base_get_iocstate(ioc, 1); 4528 if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) { 4529 dewtprintk(ioc, 4530 ioc_info(ioc, "%s: host is not operational\n", 4531 __func__)); 4532 return; 4533 } 4534 4535 /* Without releasing the smid just update the 4536 * call back index and reuse the same smid for 4537 * processing this delayed request 4538 */ 4539 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); 4540 ioc->internal_lookup[i].cb_idx = ioc->tm_sas_control_cb_idx; 4541 spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags); 4542 4543 dewtprintk(ioc, 4544 ioc_info(ioc, "sc_send:handle(0x%04x), (open), smid(%d), cb(%d)\n", 4545 handle, smid, ioc->tm_sas_control_cb_idx)); 4546 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); 4547 memset(mpi_request, 0, sizeof(Mpi2SasIoUnitControlRequest_t)); 4548 mpi_request->Function = MPI2_FUNCTION_SAS_IO_UNIT_CONTROL; 4549 mpi_request->Operation = MPI2_SAS_OP_REMOVE_DEVICE; 4550 mpi_request->DevHandle = cpu_to_le16(handle); 4551 ioc->put_smid_default(ioc, smid); 4552 } 4553 4554 /** 4555 * mpt3sas_check_for_pending_internal_cmds - check for pending internal messages 4556 * @ioc: per adapter object 4557 * @smid: system request message index 4558 * 4559 * Context: Executed in interrupt context 4560 * 4561 * This will check delayed internal messages list, and process the 4562 * next request. 4563 * 4564 * Return: 1 meaning mf should be freed from _base_interrupt 4565 * 0 means the mf is freed from this function. 4566 */ 4567 u8 4568 mpt3sas_check_for_pending_internal_cmds(struct MPT3SAS_ADAPTER *ioc, u16 smid) 4569 { 4570 struct _sc_list *delayed_sc; 4571 struct _event_ack_list *delayed_event_ack; 4572 4573 if (!list_empty(&ioc->delayed_event_ack_list)) { 4574 delayed_event_ack = list_entry(ioc->delayed_event_ack_list.next, 4575 struct _event_ack_list, list); 4576 _scsih_issue_delayed_event_ack(ioc, smid, 4577 delayed_event_ack->Event, delayed_event_ack->EventContext); 4578 list_del(&delayed_event_ack->list); 4579 kfree(delayed_event_ack); 4580 return 0; 4581 } 4582 4583 if (!list_empty(&ioc->delayed_sc_list)) { 4584 delayed_sc = list_entry(ioc->delayed_sc_list.next, 4585 struct _sc_list, list); 4586 _scsih_issue_delayed_sas_io_unit_ctrl(ioc, smid, 4587 delayed_sc->handle); 4588 list_del(&delayed_sc->list); 4589 kfree(delayed_sc); 4590 return 0; 4591 } 4592 return 1; 4593 } 4594 4595 /** 4596 * _scsih_check_for_pending_tm - check for pending task management 4597 * @ioc: per adapter object 4598 * @smid: system request message index 4599 * 4600 * This will check delayed target reset list, and feed the 4601 * next reqeust. 4602 * 4603 * Return: 1 meaning mf should be freed from _base_interrupt 4604 * 0 means the mf is freed from this function. 4605 */ 4606 static u8 4607 _scsih_check_for_pending_tm(struct MPT3SAS_ADAPTER *ioc, u16 smid) 4608 { 4609 struct _tr_list *delayed_tr; 4610 4611 if (!list_empty(&ioc->delayed_tr_volume_list)) { 4612 delayed_tr = list_entry(ioc->delayed_tr_volume_list.next, 4613 struct _tr_list, list); 4614 mpt3sas_base_free_smid(ioc, smid); 4615 _scsih_tm_tr_volume_send(ioc, delayed_tr->handle); 4616 list_del(&delayed_tr->list); 4617 kfree(delayed_tr); 4618 return 0; 4619 } 4620 4621 if (!list_empty(&ioc->delayed_tr_list)) { 4622 delayed_tr = list_entry(ioc->delayed_tr_list.next, 4623 struct _tr_list, list); 4624 mpt3sas_base_free_smid(ioc, smid); 4625 _scsih_tm_tr_send(ioc, delayed_tr->handle); 4626 list_del(&delayed_tr->list); 4627 kfree(delayed_tr); 4628 return 0; 4629 } 4630 4631 return 1; 4632 } 4633 4634 /** 4635 * _scsih_check_topo_delete_events - sanity check on topo events 4636 * @ioc: per adapter object 4637 * @event_data: the event data payload 4638 * 4639 * This routine added to better handle cable breaker. 4640 * 4641 * This handles the case where driver receives multiple expander 4642 * add and delete events in a single shot. When there is a delete event 4643 * the routine will void any pending add events waiting in the event queue. 4644 */ 4645 static void 4646 _scsih_check_topo_delete_events(struct MPT3SAS_ADAPTER *ioc, 4647 Mpi2EventDataSasTopologyChangeList_t *event_data) 4648 { 4649 struct fw_event_work *fw_event; 4650 Mpi2EventDataSasTopologyChangeList_t *local_event_data; 4651 u16 expander_handle; 4652 struct _sas_node *sas_expander; 4653 unsigned long flags; 4654 int i, reason_code; 4655 u16 handle; 4656 4657 for (i = 0 ; i < event_data->NumEntries; i++) { 4658 handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); 4659 if (!handle) 4660 continue; 4661 reason_code = event_data->PHY[i].PhyStatus & 4662 MPI2_EVENT_SAS_TOPO_RC_MASK; 4663 if (reason_code == MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING) 4664 _scsih_tm_tr_send(ioc, handle); 4665 } 4666 4667 expander_handle = le16_to_cpu(event_data->ExpanderDevHandle); 4668 if (expander_handle < ioc->sas_hba.num_phys) { 4669 _scsih_block_io_to_children_attached_directly(ioc, event_data); 4670 return; 4671 } 4672 if (event_data->ExpStatus == 4673 MPI2_EVENT_SAS_TOPO_ES_DELAY_NOT_RESPONDING) { 4674 /* put expander attached devices into blocking state */ 4675 spin_lock_irqsave(&ioc->sas_node_lock, flags); 4676 sas_expander = mpt3sas_scsih_expander_find_by_handle(ioc, 4677 expander_handle); 4678 _scsih_block_io_to_children_attached_to_ex(ioc, sas_expander); 4679 spin_unlock_irqrestore(&ioc->sas_node_lock, flags); 4680 do { 4681 handle = find_first_bit(ioc->blocking_handles, 4682 ioc->facts.MaxDevHandle); 4683 if (handle < ioc->facts.MaxDevHandle) 4684 _scsih_block_io_device(ioc, handle); 4685 } while (test_and_clear_bit(handle, ioc->blocking_handles)); 4686 } else if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_RESPONDING) 4687 _scsih_block_io_to_children_attached_directly(ioc, event_data); 4688 4689 if (event_data->ExpStatus != MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING) 4690 return; 4691 4692 /* mark ignore flag for pending events */ 4693 spin_lock_irqsave(&ioc->fw_event_lock, flags); 4694 list_for_each_entry(fw_event, &ioc->fw_event_list, list) { 4695 if (fw_event->event != MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST || 4696 fw_event->ignore) 4697 continue; 4698 local_event_data = (Mpi2EventDataSasTopologyChangeList_t *) 4699 fw_event->event_data; 4700 if (local_event_data->ExpStatus == 4701 MPI2_EVENT_SAS_TOPO_ES_ADDED || 4702 local_event_data->ExpStatus == 4703 MPI2_EVENT_SAS_TOPO_ES_RESPONDING) { 4704 if (le16_to_cpu(local_event_data->ExpanderDevHandle) == 4705 expander_handle) { 4706 dewtprintk(ioc, 4707 ioc_info(ioc, "setting ignoring flag\n")); 4708 fw_event->ignore = 1; 4709 } 4710 } 4711 } 4712 spin_unlock_irqrestore(&ioc->fw_event_lock, flags); 4713 } 4714 4715 /** 4716 * _scsih_check_pcie_topo_remove_events - sanity check on topo 4717 * events 4718 * @ioc: per adapter object 4719 * @event_data: the event data payload 4720 * 4721 * This handles the case where driver receives multiple switch 4722 * or device add and delete events in a single shot. When there 4723 * is a delete event the routine will void any pending add 4724 * events waiting in the event queue. 4725 */ 4726 static void 4727 _scsih_check_pcie_topo_remove_events(struct MPT3SAS_ADAPTER *ioc, 4728 Mpi26EventDataPCIeTopologyChangeList_t *event_data) 4729 { 4730 struct fw_event_work *fw_event; 4731 Mpi26EventDataPCIeTopologyChangeList_t *local_event_data; 4732 unsigned long flags; 4733 int i, reason_code; 4734 u16 handle, switch_handle; 4735 4736 for (i = 0; i < event_data->NumEntries; i++) { 4737 handle = 4738 le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle); 4739 if (!handle) 4740 continue; 4741 reason_code = event_data->PortEntry[i].PortStatus; 4742 if (reason_code == MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING) 4743 _scsih_tm_tr_send(ioc, handle); 4744 } 4745 4746 switch_handle = le16_to_cpu(event_data->SwitchDevHandle); 4747 if (!switch_handle) { 4748 _scsih_block_io_to_pcie_children_attached_directly( 4749 ioc, event_data); 4750 return; 4751 } 4752 /* TODO We are not supporting cascaded PCIe Switch removal yet*/ 4753 if ((event_data->SwitchStatus 4754 == MPI26_EVENT_PCIE_TOPO_SS_DELAY_NOT_RESPONDING) || 4755 (event_data->SwitchStatus == 4756 MPI26_EVENT_PCIE_TOPO_SS_RESPONDING)) 4757 _scsih_block_io_to_pcie_children_attached_directly( 4758 ioc, event_data); 4759 4760 if (event_data->SwitchStatus != MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING) 4761 return; 4762 4763 /* mark ignore flag for pending events */ 4764 spin_lock_irqsave(&ioc->fw_event_lock, flags); 4765 list_for_each_entry(fw_event, &ioc->fw_event_list, list) { 4766 if (fw_event->event != MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST || 4767 fw_event->ignore) 4768 continue; 4769 local_event_data = 4770 (Mpi26EventDataPCIeTopologyChangeList_t *) 4771 fw_event->event_data; 4772 if (local_event_data->SwitchStatus == 4773 MPI2_EVENT_SAS_TOPO_ES_ADDED || 4774 local_event_data->SwitchStatus == 4775 MPI2_EVENT_SAS_TOPO_ES_RESPONDING) { 4776 if (le16_to_cpu(local_event_data->SwitchDevHandle) == 4777 switch_handle) { 4778 dewtprintk(ioc, 4779 ioc_info(ioc, "setting ignoring flag for switch event\n")); 4780 fw_event->ignore = 1; 4781 } 4782 } 4783 } 4784 spin_unlock_irqrestore(&ioc->fw_event_lock, flags); 4785 } 4786 4787 /** 4788 * _scsih_set_volume_delete_flag - setting volume delete flag 4789 * @ioc: per adapter object 4790 * @handle: device handle 4791 * 4792 * This returns nothing. 4793 */ 4794 static void 4795 _scsih_set_volume_delete_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle) 4796 { 4797 struct _raid_device *raid_device; 4798 struct MPT3SAS_TARGET *sas_target_priv_data; 4799 unsigned long flags; 4800 4801 spin_lock_irqsave(&ioc->raid_device_lock, flags); 4802 raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); 4803 if (raid_device && raid_device->starget && 4804 raid_device->starget->hostdata) { 4805 sas_target_priv_data = 4806 raid_device->starget->hostdata; 4807 sas_target_priv_data->deleted = 1; 4808 dewtprintk(ioc, 4809 ioc_info(ioc, "setting delete flag: handle(0x%04x), wwid(0x%016llx)\n", 4810 handle, (u64)raid_device->wwid)); 4811 } 4812 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 4813 } 4814 4815 /** 4816 * _scsih_set_volume_handle_for_tr - set handle for target reset to volume 4817 * @handle: input handle 4818 * @a: handle for volume a 4819 * @b: handle for volume b 4820 * 4821 * IR firmware only supports two raid volumes. The purpose of this 4822 * routine is to set the volume handle in either a or b. When the given 4823 * input handle is non-zero, or when a and b have not been set before. 4824 */ 4825 static void 4826 _scsih_set_volume_handle_for_tr(u16 handle, u16 *a, u16 *b) 4827 { 4828 if (!handle || handle == *a || handle == *b) 4829 return; 4830 if (!*a) 4831 *a = handle; 4832 else if (!*b) 4833 *b = handle; 4834 } 4835 4836 /** 4837 * _scsih_check_ir_config_unhide_events - check for UNHIDE events 4838 * @ioc: per adapter object 4839 * @event_data: the event data payload 4840 * Context: interrupt time. 4841 * 4842 * This routine will send target reset to volume, followed by target 4843 * resets to the PDs. This is called when a PD has been removed, or 4844 * volume has been deleted or removed. When the target reset is sent 4845 * to volume, the PD target resets need to be queued to start upon 4846 * completion of the volume target reset. 4847 */ 4848 static void 4849 _scsih_check_ir_config_unhide_events(struct MPT3SAS_ADAPTER *ioc, 4850 Mpi2EventDataIrConfigChangeList_t *event_data) 4851 { 4852 Mpi2EventIrConfigElement_t *element; 4853 int i; 4854 u16 handle, volume_handle, a, b; 4855 struct _tr_list *delayed_tr; 4856 4857 a = 0; 4858 b = 0; 4859 4860 if (ioc->is_warpdrive) 4861 return; 4862 4863 /* Volume Resets for Deleted or Removed */ 4864 element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; 4865 for (i = 0; i < event_data->NumElements; i++, element++) { 4866 if (le32_to_cpu(event_data->Flags) & 4867 MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG) 4868 continue; 4869 if (element->ReasonCode == 4870 MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED || 4871 element->ReasonCode == 4872 MPI2_EVENT_IR_CHANGE_RC_REMOVED) { 4873 volume_handle = le16_to_cpu(element->VolDevHandle); 4874 _scsih_set_volume_delete_flag(ioc, volume_handle); 4875 _scsih_set_volume_handle_for_tr(volume_handle, &a, &b); 4876 } 4877 } 4878 4879 /* Volume Resets for UNHIDE events */ 4880 element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; 4881 for (i = 0; i < event_data->NumElements; i++, element++) { 4882 if (le32_to_cpu(event_data->Flags) & 4883 MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG) 4884 continue; 4885 if (element->ReasonCode == MPI2_EVENT_IR_CHANGE_RC_UNHIDE) { 4886 volume_handle = le16_to_cpu(element->VolDevHandle); 4887 _scsih_set_volume_handle_for_tr(volume_handle, &a, &b); 4888 } 4889 } 4890 4891 if (a) 4892 _scsih_tm_tr_volume_send(ioc, a); 4893 if (b) 4894 _scsih_tm_tr_volume_send(ioc, b); 4895 4896 /* PD target resets */ 4897 element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; 4898 for (i = 0; i < event_data->NumElements; i++, element++) { 4899 if (element->ReasonCode != MPI2_EVENT_IR_CHANGE_RC_UNHIDE) 4900 continue; 4901 handle = le16_to_cpu(element->PhysDiskDevHandle); 4902 volume_handle = le16_to_cpu(element->VolDevHandle); 4903 clear_bit(handle, ioc->pd_handles); 4904 if (!volume_handle) 4905 _scsih_tm_tr_send(ioc, handle); 4906 else if (volume_handle == a || volume_handle == b) { 4907 delayed_tr = kzalloc(sizeof(*delayed_tr), GFP_ATOMIC); 4908 BUG_ON(!delayed_tr); 4909 INIT_LIST_HEAD(&delayed_tr->list); 4910 delayed_tr->handle = handle; 4911 list_add_tail(&delayed_tr->list, &ioc->delayed_tr_list); 4912 dewtprintk(ioc, 4913 ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n", 4914 handle)); 4915 } else 4916 _scsih_tm_tr_send(ioc, handle); 4917 } 4918 } 4919 4920 4921 /** 4922 * _scsih_check_volume_delete_events - set delete flag for volumes 4923 * @ioc: per adapter object 4924 * @event_data: the event data payload 4925 * Context: interrupt time. 4926 * 4927 * This will handle the case when the cable connected to entire volume is 4928 * pulled. We will take care of setting the deleted flag so normal IO will 4929 * not be sent. 4930 */ 4931 static void 4932 _scsih_check_volume_delete_events(struct MPT3SAS_ADAPTER *ioc, 4933 Mpi2EventDataIrVolume_t *event_data) 4934 { 4935 u32 state; 4936 4937 if (event_data->ReasonCode != MPI2_EVENT_IR_VOLUME_RC_STATE_CHANGED) 4938 return; 4939 state = le32_to_cpu(event_data->NewValue); 4940 if (state == MPI2_RAID_VOL_STATE_MISSING || state == 4941 MPI2_RAID_VOL_STATE_FAILED) 4942 _scsih_set_volume_delete_flag(ioc, 4943 le16_to_cpu(event_data->VolDevHandle)); 4944 } 4945 4946 /** 4947 * _scsih_temp_threshold_events - display temperature threshold exceeded events 4948 * @ioc: per adapter object 4949 * @event_data: the temp threshold event data 4950 * Context: interrupt time. 4951 */ 4952 static void 4953 _scsih_temp_threshold_events(struct MPT3SAS_ADAPTER *ioc, 4954 Mpi2EventDataTemperature_t *event_data) 4955 { 4956 u32 doorbell; 4957 if (ioc->temp_sensors_count >= event_data->SensorNum) { 4958 ioc_err(ioc, "Temperature Threshold flags %s%s%s%s exceeded for Sensor: %d !!!\n", 4959 le16_to_cpu(event_data->Status) & 0x1 ? "0 " : " ", 4960 le16_to_cpu(event_data->Status) & 0x2 ? "1 " : " ", 4961 le16_to_cpu(event_data->Status) & 0x4 ? "2 " : " ", 4962 le16_to_cpu(event_data->Status) & 0x8 ? "3 " : " ", 4963 event_data->SensorNum); 4964 ioc_err(ioc, "Current Temp In Celsius: %d\n", 4965 event_data->CurrentTemperature); 4966 if (ioc->hba_mpi_version_belonged != MPI2_VERSION) { 4967 doorbell = mpt3sas_base_get_iocstate(ioc, 0); 4968 if ((doorbell & MPI2_IOC_STATE_MASK) == 4969 MPI2_IOC_STATE_FAULT) { 4970 mpt3sas_print_fault_code(ioc, 4971 doorbell & MPI2_DOORBELL_DATA_MASK); 4972 } else if ((doorbell & MPI2_IOC_STATE_MASK) == 4973 MPI2_IOC_STATE_COREDUMP) { 4974 mpt3sas_print_coredump_info(ioc, 4975 doorbell & MPI2_DOORBELL_DATA_MASK); 4976 } 4977 } 4978 } 4979 } 4980 4981 static int _scsih_set_satl_pending(struct scsi_cmnd *scmd, bool pending) 4982 { 4983 struct MPT3SAS_DEVICE *priv = scmd->device->hostdata; 4984 4985 if (scmd->cmnd[0] != ATA_12 && scmd->cmnd[0] != ATA_16) 4986 return 0; 4987 4988 if (pending) 4989 return test_and_set_bit(0, &priv->ata_command_pending); 4990 4991 clear_bit(0, &priv->ata_command_pending); 4992 return 0; 4993 } 4994 4995 /** 4996 * _scsih_flush_running_cmds - completing outstanding commands. 4997 * @ioc: per adapter object 4998 * 4999 * The flushing out of all pending scmd commands following host reset, 5000 * where all IO is dropped to the floor. 5001 */ 5002 static void 5003 _scsih_flush_running_cmds(struct MPT3SAS_ADAPTER *ioc) 5004 { 5005 struct scsi_cmnd *scmd; 5006 struct scsiio_tracker *st; 5007 u16 smid; 5008 int count = 0; 5009 5010 for (smid = 1; smid <= ioc->scsiio_depth; smid++) { 5011 scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); 5012 if (!scmd) 5013 continue; 5014 count++; 5015 _scsih_set_satl_pending(scmd, false); 5016 st = scsi_cmd_priv(scmd); 5017 mpt3sas_base_clear_st(ioc, st); 5018 scsi_dma_unmap(scmd); 5019 if (ioc->pci_error_recovery || ioc->remove_host) 5020 scmd->result = DID_NO_CONNECT << 16; 5021 else 5022 scmd->result = DID_RESET << 16; 5023 scsi_done(scmd); 5024 } 5025 dtmprintk(ioc, ioc_info(ioc, "completing %d cmds\n", count)); 5026 } 5027 5028 /** 5029 * _scsih_setup_eedp - setup MPI request for EEDP transfer 5030 * @ioc: per adapter object 5031 * @scmd: pointer to scsi command object 5032 * @mpi_request: pointer to the SCSI_IO request message frame 5033 * 5034 * Supporting protection 1 and 3. 5035 */ 5036 static void 5037 _scsih_setup_eedp(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd, 5038 Mpi25SCSIIORequest_t *mpi_request) 5039 { 5040 u16 eedp_flags; 5041 Mpi25SCSIIORequest_t *mpi_request_3v = 5042 (Mpi25SCSIIORequest_t *)mpi_request; 5043 5044 switch (scsi_get_prot_op(scmd)) { 5045 case SCSI_PROT_READ_STRIP: 5046 eedp_flags = MPI2_SCSIIO_EEDPFLAGS_CHECK_REMOVE_OP; 5047 break; 5048 case SCSI_PROT_WRITE_INSERT: 5049 eedp_flags = MPI2_SCSIIO_EEDPFLAGS_INSERT_OP; 5050 break; 5051 default: 5052 return; 5053 } 5054 5055 if (scmd->prot_flags & SCSI_PROT_GUARD_CHECK) 5056 eedp_flags |= MPI2_SCSIIO_EEDPFLAGS_CHECK_GUARD; 5057 5058 if (scmd->prot_flags & SCSI_PROT_REF_CHECK) 5059 eedp_flags |= MPI2_SCSIIO_EEDPFLAGS_CHECK_REFTAG; 5060 5061 if (scmd->prot_flags & SCSI_PROT_REF_INCREMENT) { 5062 eedp_flags |= MPI2_SCSIIO_EEDPFLAGS_INC_PRI_REFTAG; 5063 5064 mpi_request->CDB.EEDP32.PrimaryReferenceTag = 5065 cpu_to_be32(scsi_prot_ref_tag(scmd)); 5066 } 5067 5068 mpi_request_3v->EEDPBlockSize = cpu_to_le16(scsi_prot_interval(scmd)); 5069 5070 if (ioc->is_gen35_ioc) 5071 eedp_flags |= MPI25_SCSIIO_EEDPFLAGS_APPTAG_DISABLE_MODE; 5072 mpi_request->EEDPFlags = cpu_to_le16(eedp_flags); 5073 } 5074 5075 /** 5076 * _scsih_eedp_error_handling - return sense code for EEDP errors 5077 * @scmd: pointer to scsi command object 5078 * @ioc_status: ioc status 5079 */ 5080 static void 5081 _scsih_eedp_error_handling(struct scsi_cmnd *scmd, u16 ioc_status) 5082 { 5083 u8 ascq; 5084 5085 switch (ioc_status) { 5086 case MPI2_IOCSTATUS_EEDP_GUARD_ERROR: 5087 ascq = 0x01; 5088 break; 5089 case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR: 5090 ascq = 0x02; 5091 break; 5092 case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR: 5093 ascq = 0x03; 5094 break; 5095 default: 5096 ascq = 0x00; 5097 break; 5098 } 5099 scsi_build_sense(scmd, 0, ILLEGAL_REQUEST, 0x10, ascq); 5100 set_host_byte(scmd, DID_ABORT); 5101 } 5102 5103 /** 5104 * scsih_qcmd - main scsi request entry point 5105 * @shost: SCSI host pointer 5106 * @scmd: pointer to scsi command object 5107 * 5108 * The callback index is set inside `ioc->scsi_io_cb_idx`. 5109 * 5110 * Return: 0 on success. If there's a failure, return either: 5111 * SCSI_MLQUEUE_DEVICE_BUSY if the device queue is full, or 5112 * SCSI_MLQUEUE_HOST_BUSY if the entire host queue is full 5113 */ 5114 static int 5115 scsih_qcmd(struct Scsi_Host *shost, struct scsi_cmnd *scmd) 5116 { 5117 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); 5118 struct MPT3SAS_DEVICE *sas_device_priv_data; 5119 struct MPT3SAS_TARGET *sas_target_priv_data; 5120 struct _raid_device *raid_device; 5121 struct request *rq = scsi_cmd_to_rq(scmd); 5122 int class; 5123 Mpi25SCSIIORequest_t *mpi_request; 5124 struct _pcie_device *pcie_device = NULL; 5125 u32 mpi_control; 5126 u16 smid; 5127 u16 handle; 5128 5129 if (ioc->logging_level & MPT_DEBUG_SCSI) 5130 scsi_print_command(scmd); 5131 5132 sas_device_priv_data = scmd->device->hostdata; 5133 if (!sas_device_priv_data || !sas_device_priv_data->sas_target) { 5134 scmd->result = DID_NO_CONNECT << 16; 5135 scsi_done(scmd); 5136 return 0; 5137 } 5138 5139 if (!(_scsih_allow_scmd_to_device(ioc, scmd))) { 5140 scmd->result = DID_NO_CONNECT << 16; 5141 scsi_done(scmd); 5142 return 0; 5143 } 5144 5145 sas_target_priv_data = sas_device_priv_data->sas_target; 5146 5147 /* invalid device handle */ 5148 handle = sas_target_priv_data->handle; 5149 5150 /* 5151 * Avoid error handling escallation when device is disconnected 5152 */ 5153 if (handle == MPT3SAS_INVALID_DEVICE_HANDLE || sas_device_priv_data->block) { 5154 if (scmd->device->host->shost_state == SHOST_RECOVERY && 5155 scmd->cmnd[0] == TEST_UNIT_READY) { 5156 scsi_build_sense(scmd, 0, UNIT_ATTENTION, 0x29, 0x07); 5157 scsi_done(scmd); 5158 return 0; 5159 } 5160 } 5161 5162 if (handle == MPT3SAS_INVALID_DEVICE_HANDLE) { 5163 scmd->result = DID_NO_CONNECT << 16; 5164 scsi_done(scmd); 5165 return 0; 5166 } 5167 5168 5169 if (ioc->shost_recovery || ioc->ioc_link_reset_in_progress) { 5170 /* host recovery or link resets sent via IOCTLs */ 5171 return SCSI_MLQUEUE_HOST_BUSY; 5172 } else if (sas_target_priv_data->deleted) { 5173 /* device has been deleted */ 5174 scmd->result = DID_NO_CONNECT << 16; 5175 scsi_done(scmd); 5176 return 0; 5177 } else if (sas_target_priv_data->tm_busy || 5178 sas_device_priv_data->block) { 5179 /* device busy with task management */ 5180 return SCSI_MLQUEUE_DEVICE_BUSY; 5181 } 5182 5183 /* 5184 * Bug work around for firmware SATL handling. The loop 5185 * is based on atomic operations and ensures consistency 5186 * since we're lockless at this point 5187 */ 5188 do { 5189 if (test_bit(0, &sas_device_priv_data->ata_command_pending)) 5190 return SCSI_MLQUEUE_DEVICE_BUSY; 5191 } while (_scsih_set_satl_pending(scmd, true)); 5192 5193 if (scmd->sc_data_direction == DMA_FROM_DEVICE) 5194 mpi_control = MPI2_SCSIIO_CONTROL_READ; 5195 else if (scmd->sc_data_direction == DMA_TO_DEVICE) 5196 mpi_control = MPI2_SCSIIO_CONTROL_WRITE; 5197 else 5198 mpi_control = MPI2_SCSIIO_CONTROL_NODATATRANSFER; 5199 5200 /* set tags */ 5201 mpi_control |= MPI2_SCSIIO_CONTROL_SIMPLEQ; 5202 /* NCQ Prio supported, make sure control indicated high priority */ 5203 if (sas_device_priv_data->ncq_prio_enable) { 5204 class = IOPRIO_PRIO_CLASS(req_get_ioprio(rq)); 5205 if (class == IOPRIO_CLASS_RT) 5206 mpi_control |= 1 << MPI2_SCSIIO_CONTROL_CMDPRI_SHIFT; 5207 } 5208 /* Make sure Device is not raid volume. 5209 * We do not expose raid functionality to upper layer for warpdrive. 5210 */ 5211 if (((!ioc->is_warpdrive && !scsih_is_raid(&scmd->device->sdev_gendev)) 5212 && !scsih_is_nvme(&scmd->device->sdev_gendev)) 5213 && sas_is_tlr_enabled(scmd->device) && scmd->cmd_len != 32) 5214 mpi_control |= MPI2_SCSIIO_CONTROL_TLR_ON; 5215 5216 smid = mpt3sas_base_get_smid_scsiio(ioc, ioc->scsi_io_cb_idx, scmd); 5217 if (!smid) { 5218 ioc_err(ioc, "%s: failed obtaining a smid\n", __func__); 5219 _scsih_set_satl_pending(scmd, false); 5220 goto out; 5221 } 5222 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); 5223 memset(mpi_request, 0, ioc->request_sz); 5224 _scsih_setup_eedp(ioc, scmd, mpi_request); 5225 5226 if (scmd->cmd_len == 32) 5227 mpi_control |= 4 << MPI2_SCSIIO_CONTROL_ADDCDBLEN_SHIFT; 5228 mpi_request->Function = MPI2_FUNCTION_SCSI_IO_REQUEST; 5229 if (sas_device_priv_data->sas_target->flags & 5230 MPT_TARGET_FLAGS_RAID_COMPONENT) 5231 mpi_request->Function = MPI2_FUNCTION_RAID_SCSI_IO_PASSTHROUGH; 5232 else 5233 mpi_request->Function = MPI2_FUNCTION_SCSI_IO_REQUEST; 5234 mpi_request->DevHandle = cpu_to_le16(handle); 5235 mpi_request->DataLength = cpu_to_le32(scsi_bufflen(scmd)); 5236 mpi_request->Control = cpu_to_le32(mpi_control); 5237 mpi_request->IoFlags = cpu_to_le16(scmd->cmd_len); 5238 mpi_request->MsgFlags = MPI2_SCSIIO_MSGFLAGS_SYSTEM_SENSE_ADDR; 5239 mpi_request->SenseBufferLength = SCSI_SENSE_BUFFERSIZE; 5240 mpi_request->SenseBufferLowAddress = 5241 mpt3sas_base_get_sense_buffer_dma(ioc, smid); 5242 mpi_request->SGLOffset0 = offsetof(Mpi25SCSIIORequest_t, SGL) / 4; 5243 int_to_scsilun(sas_device_priv_data->lun, (struct scsi_lun *) 5244 mpi_request->LUN); 5245 memcpy(mpi_request->CDB.CDB32, scmd->cmnd, scmd->cmd_len); 5246 5247 if (mpi_request->DataLength) { 5248 pcie_device = sas_target_priv_data->pcie_dev; 5249 if (ioc->build_sg_scmd(ioc, scmd, smid, pcie_device)) { 5250 mpt3sas_base_free_smid(ioc, smid); 5251 _scsih_set_satl_pending(scmd, false); 5252 goto out; 5253 } 5254 } else 5255 ioc->build_zero_len_sge(ioc, &mpi_request->SGL); 5256 5257 raid_device = sas_target_priv_data->raid_device; 5258 if (raid_device && raid_device->direct_io_enabled) 5259 mpt3sas_setup_direct_io(ioc, scmd, 5260 raid_device, mpi_request); 5261 5262 if (likely(mpi_request->Function == MPI2_FUNCTION_SCSI_IO_REQUEST)) { 5263 if (sas_target_priv_data->flags & MPT_TARGET_FASTPATH_IO) { 5264 mpi_request->IoFlags = cpu_to_le16(scmd->cmd_len | 5265 MPI25_SCSIIO_IOFLAGS_FAST_PATH); 5266 ioc->put_smid_fast_path(ioc, smid, handle); 5267 } else 5268 ioc->put_smid_scsi_io(ioc, smid, 5269 le16_to_cpu(mpi_request->DevHandle)); 5270 } else 5271 ioc->put_smid_default(ioc, smid); 5272 return 0; 5273 5274 out: 5275 return SCSI_MLQUEUE_HOST_BUSY; 5276 } 5277 5278 /** 5279 * _scsih_normalize_sense - normalize descriptor and fixed format sense data 5280 * @sense_buffer: sense data returned by target 5281 * @data: normalized skey/asc/ascq 5282 */ 5283 static void 5284 _scsih_normalize_sense(char *sense_buffer, struct sense_info *data) 5285 { 5286 if ((sense_buffer[0] & 0x7F) >= 0x72) { 5287 /* descriptor format */ 5288 data->skey = sense_buffer[1] & 0x0F; 5289 data->asc = sense_buffer[2]; 5290 data->ascq = sense_buffer[3]; 5291 } else { 5292 /* fixed format */ 5293 data->skey = sense_buffer[2] & 0x0F; 5294 data->asc = sense_buffer[12]; 5295 data->ascq = sense_buffer[13]; 5296 } 5297 } 5298 5299 /** 5300 * _scsih_scsi_ioc_info - translated non-successful SCSI_IO request 5301 * @ioc: per adapter object 5302 * @scmd: pointer to scsi command object 5303 * @mpi_reply: reply mf payload returned from firmware 5304 * @smid: ? 5305 * 5306 * scsi_status - SCSI Status code returned from target device 5307 * scsi_state - state info associated with SCSI_IO determined by ioc 5308 * ioc_status - ioc supplied status info 5309 */ 5310 static void 5311 _scsih_scsi_ioc_info(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd, 5312 Mpi2SCSIIOReply_t *mpi_reply, u16 smid) 5313 { 5314 u32 response_info; 5315 u8 *response_bytes; 5316 u16 ioc_status = le16_to_cpu(mpi_reply->IOCStatus) & 5317 MPI2_IOCSTATUS_MASK; 5318 u8 scsi_state = mpi_reply->SCSIState; 5319 u8 scsi_status = mpi_reply->SCSIStatus; 5320 char *desc_ioc_state = NULL; 5321 char *desc_scsi_status = NULL; 5322 char *desc_scsi_state = ioc->tmp_string; 5323 u32 log_info = le32_to_cpu(mpi_reply->IOCLogInfo); 5324 struct _sas_device *sas_device = NULL; 5325 struct _pcie_device *pcie_device = NULL; 5326 struct scsi_target *starget = scmd->device->sdev_target; 5327 struct MPT3SAS_TARGET *priv_target = starget->hostdata; 5328 char *device_str = NULL; 5329 5330 if (!priv_target) 5331 return; 5332 if (ioc->hide_ir_msg) 5333 device_str = "WarpDrive"; 5334 else 5335 device_str = "volume"; 5336 5337 if (log_info == 0x31170000) 5338 return; 5339 5340 switch (ioc_status) { 5341 case MPI2_IOCSTATUS_SUCCESS: 5342 desc_ioc_state = "success"; 5343 break; 5344 case MPI2_IOCSTATUS_INVALID_FUNCTION: 5345 desc_ioc_state = "invalid function"; 5346 break; 5347 case MPI2_IOCSTATUS_SCSI_RECOVERED_ERROR: 5348 desc_ioc_state = "scsi recovered error"; 5349 break; 5350 case MPI2_IOCSTATUS_SCSI_INVALID_DEVHANDLE: 5351 desc_ioc_state = "scsi invalid dev handle"; 5352 break; 5353 case MPI2_IOCSTATUS_SCSI_DEVICE_NOT_THERE: 5354 desc_ioc_state = "scsi device not there"; 5355 break; 5356 case MPI2_IOCSTATUS_SCSI_DATA_OVERRUN: 5357 desc_ioc_state = "scsi data overrun"; 5358 break; 5359 case MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN: 5360 desc_ioc_state = "scsi data underrun"; 5361 break; 5362 case MPI2_IOCSTATUS_SCSI_IO_DATA_ERROR: 5363 desc_ioc_state = "scsi io data error"; 5364 break; 5365 case MPI2_IOCSTATUS_SCSI_PROTOCOL_ERROR: 5366 desc_ioc_state = "scsi protocol error"; 5367 break; 5368 case MPI2_IOCSTATUS_SCSI_TASK_TERMINATED: 5369 desc_ioc_state = "scsi task terminated"; 5370 break; 5371 case MPI2_IOCSTATUS_SCSI_RESIDUAL_MISMATCH: 5372 desc_ioc_state = "scsi residual mismatch"; 5373 break; 5374 case MPI2_IOCSTATUS_SCSI_TASK_MGMT_FAILED: 5375 desc_ioc_state = "scsi task mgmt failed"; 5376 break; 5377 case MPI2_IOCSTATUS_SCSI_IOC_TERMINATED: 5378 desc_ioc_state = "scsi ioc terminated"; 5379 break; 5380 case MPI2_IOCSTATUS_SCSI_EXT_TERMINATED: 5381 desc_ioc_state = "scsi ext terminated"; 5382 break; 5383 case MPI2_IOCSTATUS_EEDP_GUARD_ERROR: 5384 desc_ioc_state = "eedp guard error"; 5385 break; 5386 case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR: 5387 desc_ioc_state = "eedp ref tag error"; 5388 break; 5389 case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR: 5390 desc_ioc_state = "eedp app tag error"; 5391 break; 5392 case MPI2_IOCSTATUS_INSUFFICIENT_POWER: 5393 desc_ioc_state = "insufficient power"; 5394 break; 5395 default: 5396 desc_ioc_state = "unknown"; 5397 break; 5398 } 5399 5400 switch (scsi_status) { 5401 case MPI2_SCSI_STATUS_GOOD: 5402 desc_scsi_status = "good"; 5403 break; 5404 case MPI2_SCSI_STATUS_CHECK_CONDITION: 5405 desc_scsi_status = "check condition"; 5406 break; 5407 case MPI2_SCSI_STATUS_CONDITION_MET: 5408 desc_scsi_status = "condition met"; 5409 break; 5410 case MPI2_SCSI_STATUS_BUSY: 5411 desc_scsi_status = "busy"; 5412 break; 5413 case MPI2_SCSI_STATUS_INTERMEDIATE: 5414 desc_scsi_status = "intermediate"; 5415 break; 5416 case MPI2_SCSI_STATUS_INTERMEDIATE_CONDMET: 5417 desc_scsi_status = "intermediate condmet"; 5418 break; 5419 case MPI2_SCSI_STATUS_RESERVATION_CONFLICT: 5420 desc_scsi_status = "reservation conflict"; 5421 break; 5422 case MPI2_SCSI_STATUS_COMMAND_TERMINATED: 5423 desc_scsi_status = "command terminated"; 5424 break; 5425 case MPI2_SCSI_STATUS_TASK_SET_FULL: 5426 desc_scsi_status = "task set full"; 5427 break; 5428 case MPI2_SCSI_STATUS_ACA_ACTIVE: 5429 desc_scsi_status = "aca active"; 5430 break; 5431 case MPI2_SCSI_STATUS_TASK_ABORTED: 5432 desc_scsi_status = "task aborted"; 5433 break; 5434 default: 5435 desc_scsi_status = "unknown"; 5436 break; 5437 } 5438 5439 desc_scsi_state[0] = '\0'; 5440 if (!scsi_state) 5441 desc_scsi_state = " "; 5442 if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID) 5443 strcat(desc_scsi_state, "response info "); 5444 if (scsi_state & MPI2_SCSI_STATE_TERMINATED) 5445 strcat(desc_scsi_state, "state terminated "); 5446 if (scsi_state & MPI2_SCSI_STATE_NO_SCSI_STATUS) 5447 strcat(desc_scsi_state, "no status "); 5448 if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_FAILED) 5449 strcat(desc_scsi_state, "autosense failed "); 5450 if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) 5451 strcat(desc_scsi_state, "autosense valid "); 5452 5453 scsi_print_command(scmd); 5454 5455 if (priv_target->flags & MPT_TARGET_FLAGS_VOLUME) { 5456 ioc_warn(ioc, "\t%s wwid(0x%016llx)\n", 5457 device_str, (u64)priv_target->sas_address); 5458 } else if (priv_target->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) { 5459 pcie_device = mpt3sas_get_pdev_from_target(ioc, priv_target); 5460 if (pcie_device) { 5461 ioc_info(ioc, "\twwid(0x%016llx), port(%d)\n", 5462 (u64)pcie_device->wwid, pcie_device->port_num); 5463 if (pcie_device->enclosure_handle != 0) 5464 ioc_info(ioc, "\tenclosure logical id(0x%016llx), slot(%d)\n", 5465 (u64)pcie_device->enclosure_logical_id, 5466 pcie_device->slot); 5467 if (pcie_device->connector_name[0]) 5468 ioc_info(ioc, "\tenclosure level(0x%04x), connector name( %s)\n", 5469 pcie_device->enclosure_level, 5470 pcie_device->connector_name); 5471 pcie_device_put(pcie_device); 5472 } 5473 } else { 5474 sas_device = mpt3sas_get_sdev_from_target(ioc, priv_target); 5475 if (sas_device) { 5476 ioc_warn(ioc, "\tsas_address(0x%016llx), phy(%d)\n", 5477 (u64)sas_device->sas_address, sas_device->phy); 5478 5479 _scsih_display_enclosure_chassis_info(ioc, sas_device, 5480 NULL, NULL); 5481 5482 sas_device_put(sas_device); 5483 } 5484 } 5485 5486 ioc_warn(ioc, "\thandle(0x%04x), ioc_status(%s)(0x%04x), smid(%d)\n", 5487 le16_to_cpu(mpi_reply->DevHandle), 5488 desc_ioc_state, ioc_status, smid); 5489 ioc_warn(ioc, "\trequest_len(%d), underflow(%d), resid(%d)\n", 5490 scsi_bufflen(scmd), scmd->underflow, scsi_get_resid(scmd)); 5491 ioc_warn(ioc, "\ttag(%d), transfer_count(%d), sc->result(0x%08x)\n", 5492 le16_to_cpu(mpi_reply->TaskTag), 5493 le32_to_cpu(mpi_reply->TransferCount), scmd->result); 5494 ioc_warn(ioc, "\tscsi_status(%s)(0x%02x), scsi_state(%s)(0x%02x)\n", 5495 desc_scsi_status, scsi_status, desc_scsi_state, scsi_state); 5496 5497 if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) { 5498 struct sense_info data; 5499 _scsih_normalize_sense(scmd->sense_buffer, &data); 5500 ioc_warn(ioc, "\t[sense_key,asc,ascq]: [0x%02x,0x%02x,0x%02x], count(%d)\n", 5501 data.skey, data.asc, data.ascq, 5502 le32_to_cpu(mpi_reply->SenseCount)); 5503 } 5504 if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID) { 5505 response_info = le32_to_cpu(mpi_reply->ResponseInfo); 5506 response_bytes = (u8 *)&response_info; 5507 _scsih_response_code(ioc, response_bytes[0]); 5508 } 5509 } 5510 5511 /** 5512 * _scsih_turn_on_pfa_led - illuminate PFA LED 5513 * @ioc: per adapter object 5514 * @handle: device handle 5515 * Context: process 5516 */ 5517 static void 5518 _scsih_turn_on_pfa_led(struct MPT3SAS_ADAPTER *ioc, u16 handle) 5519 { 5520 Mpi2SepReply_t mpi_reply; 5521 Mpi2SepRequest_t mpi_request; 5522 struct _sas_device *sas_device; 5523 5524 sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); 5525 if (!sas_device) 5526 return; 5527 5528 memset(&mpi_request, 0, sizeof(Mpi2SepRequest_t)); 5529 mpi_request.Function = MPI2_FUNCTION_SCSI_ENCLOSURE_PROCESSOR; 5530 mpi_request.Action = MPI2_SEP_REQ_ACTION_WRITE_STATUS; 5531 mpi_request.SlotStatus = 5532 cpu_to_le32(MPI2_SEP_REQ_SLOTSTATUS_PREDICTED_FAULT); 5533 mpi_request.DevHandle = cpu_to_le16(handle); 5534 mpi_request.Flags = MPI2_SEP_REQ_FLAGS_DEVHANDLE_ADDRESS; 5535 if ((mpt3sas_base_scsi_enclosure_processor(ioc, &mpi_reply, 5536 &mpi_request)) != 0) { 5537 ioc_err(ioc, "failure at %s:%d/%s()!\n", 5538 __FILE__, __LINE__, __func__); 5539 goto out; 5540 } 5541 sas_device->pfa_led_on = 1; 5542 5543 if (mpi_reply.IOCStatus || mpi_reply.IOCLogInfo) { 5544 dewtprintk(ioc, 5545 ioc_info(ioc, "enclosure_processor: ioc_status (0x%04x), loginfo(0x%08x)\n", 5546 le16_to_cpu(mpi_reply.IOCStatus), 5547 le32_to_cpu(mpi_reply.IOCLogInfo))); 5548 goto out; 5549 } 5550 out: 5551 sas_device_put(sas_device); 5552 } 5553 5554 /** 5555 * _scsih_turn_off_pfa_led - turn off Fault LED 5556 * @ioc: per adapter object 5557 * @sas_device: sas device whose PFA LED has to turned off 5558 * Context: process 5559 */ 5560 static void 5561 _scsih_turn_off_pfa_led(struct MPT3SAS_ADAPTER *ioc, 5562 struct _sas_device *sas_device) 5563 { 5564 Mpi2SepReply_t mpi_reply; 5565 Mpi2SepRequest_t mpi_request; 5566 5567 memset(&mpi_request, 0, sizeof(Mpi2SepRequest_t)); 5568 mpi_request.Function = MPI2_FUNCTION_SCSI_ENCLOSURE_PROCESSOR; 5569 mpi_request.Action = MPI2_SEP_REQ_ACTION_WRITE_STATUS; 5570 mpi_request.SlotStatus = 0; 5571 mpi_request.Slot = cpu_to_le16(sas_device->slot); 5572 mpi_request.DevHandle = 0; 5573 mpi_request.EnclosureHandle = cpu_to_le16(sas_device->enclosure_handle); 5574 mpi_request.Flags = MPI2_SEP_REQ_FLAGS_ENCLOSURE_SLOT_ADDRESS; 5575 if ((mpt3sas_base_scsi_enclosure_processor(ioc, &mpi_reply, 5576 &mpi_request)) != 0) { 5577 ioc_err(ioc, "failure at %s:%d/%s()!\n", 5578 __FILE__, __LINE__, __func__); 5579 return; 5580 } 5581 5582 if (mpi_reply.IOCStatus || mpi_reply.IOCLogInfo) { 5583 dewtprintk(ioc, 5584 ioc_info(ioc, "enclosure_processor: ioc_status (0x%04x), loginfo(0x%08x)\n", 5585 le16_to_cpu(mpi_reply.IOCStatus), 5586 le32_to_cpu(mpi_reply.IOCLogInfo))); 5587 return; 5588 } 5589 } 5590 5591 /** 5592 * _scsih_send_event_to_turn_on_pfa_led - fire delayed event 5593 * @ioc: per adapter object 5594 * @handle: device handle 5595 * Context: interrupt. 5596 */ 5597 static void 5598 _scsih_send_event_to_turn_on_pfa_led(struct MPT3SAS_ADAPTER *ioc, u16 handle) 5599 { 5600 struct fw_event_work *fw_event; 5601 5602 fw_event = alloc_fw_event_work(0); 5603 if (!fw_event) 5604 return; 5605 fw_event->event = MPT3SAS_TURN_ON_PFA_LED; 5606 fw_event->device_handle = handle; 5607 fw_event->ioc = ioc; 5608 _scsih_fw_event_add(ioc, fw_event); 5609 fw_event_work_put(fw_event); 5610 } 5611 5612 /** 5613 * _scsih_smart_predicted_fault - process smart errors 5614 * @ioc: per adapter object 5615 * @handle: device handle 5616 * Context: interrupt. 5617 */ 5618 static void 5619 _scsih_smart_predicted_fault(struct MPT3SAS_ADAPTER *ioc, u16 handle) 5620 { 5621 struct scsi_target *starget; 5622 struct MPT3SAS_TARGET *sas_target_priv_data; 5623 Mpi2EventNotificationReply_t *event_reply; 5624 Mpi2EventDataSasDeviceStatusChange_t *event_data; 5625 struct _sas_device *sas_device; 5626 ssize_t sz; 5627 unsigned long flags; 5628 5629 /* only handle non-raid devices */ 5630 spin_lock_irqsave(&ioc->sas_device_lock, flags); 5631 sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); 5632 if (!sas_device) 5633 goto out_unlock; 5634 5635 starget = sas_device->starget; 5636 sas_target_priv_data = starget->hostdata; 5637 5638 if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) || 5639 ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME))) 5640 goto out_unlock; 5641 5642 _scsih_display_enclosure_chassis_info(NULL, sas_device, NULL, starget); 5643 5644 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 5645 5646 if (ioc->pdev->subsystem_vendor == PCI_VENDOR_ID_IBM) 5647 _scsih_send_event_to_turn_on_pfa_led(ioc, handle); 5648 5649 /* insert into event log */ 5650 sz = offsetof(Mpi2EventNotificationReply_t, EventData) + 5651 sizeof(Mpi2EventDataSasDeviceStatusChange_t); 5652 event_reply = kzalloc(sz, GFP_ATOMIC); 5653 if (!event_reply) { 5654 ioc_err(ioc, "failure at %s:%d/%s()!\n", 5655 __FILE__, __LINE__, __func__); 5656 goto out; 5657 } 5658 5659 event_reply->Function = MPI2_FUNCTION_EVENT_NOTIFICATION; 5660 event_reply->Event = 5661 cpu_to_le16(MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE); 5662 event_reply->MsgLength = sz/4; 5663 event_reply->EventDataLength = 5664 cpu_to_le16(sizeof(Mpi2EventDataSasDeviceStatusChange_t)/4); 5665 event_data = (Mpi2EventDataSasDeviceStatusChange_t *) 5666 event_reply->EventData; 5667 event_data->ReasonCode = MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA; 5668 event_data->ASC = 0x5D; 5669 event_data->DevHandle = cpu_to_le16(handle); 5670 event_data->SASAddress = cpu_to_le64(sas_target_priv_data->sas_address); 5671 mpt3sas_ctl_add_to_event_log(ioc, event_reply); 5672 kfree(event_reply); 5673 out: 5674 if (sas_device) 5675 sas_device_put(sas_device); 5676 return; 5677 5678 out_unlock: 5679 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 5680 goto out; 5681 } 5682 5683 /** 5684 * _scsih_io_done - scsi request callback 5685 * @ioc: per adapter object 5686 * @smid: system request message index 5687 * @msix_index: MSIX table index supplied by the OS 5688 * @reply: reply message frame(lower 32bit addr) 5689 * 5690 * Callback handler when using _scsih_qcmd. 5691 * 5692 * Return: 1 meaning mf should be freed from _base_interrupt 5693 * 0 means the mf is freed from this function. 5694 */ 5695 static u8 5696 _scsih_io_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply) 5697 { 5698 Mpi25SCSIIORequest_t *mpi_request; 5699 Mpi2SCSIIOReply_t *mpi_reply; 5700 struct scsi_cmnd *scmd; 5701 struct scsiio_tracker *st; 5702 u16 ioc_status; 5703 u32 xfer_cnt; 5704 u8 scsi_state; 5705 u8 scsi_status; 5706 u32 log_info; 5707 struct MPT3SAS_DEVICE *sas_device_priv_data; 5708 u32 response_code = 0; 5709 5710 mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply); 5711 5712 scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); 5713 if (scmd == NULL) 5714 return 1; 5715 5716 _scsih_set_satl_pending(scmd, false); 5717 5718 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); 5719 5720 if (mpi_reply == NULL) { 5721 scmd->result = DID_OK << 16; 5722 goto out; 5723 } 5724 5725 sas_device_priv_data = scmd->device->hostdata; 5726 if (!sas_device_priv_data || !sas_device_priv_data->sas_target || 5727 sas_device_priv_data->sas_target->deleted) { 5728 scmd->result = DID_NO_CONNECT << 16; 5729 goto out; 5730 } 5731 ioc_status = le16_to_cpu(mpi_reply->IOCStatus); 5732 5733 /* 5734 * WARPDRIVE: If direct_io is set then it is directIO, 5735 * the failed direct I/O should be redirected to volume 5736 */ 5737 st = scsi_cmd_priv(scmd); 5738 if (st->direct_io && 5739 ((ioc_status & MPI2_IOCSTATUS_MASK) 5740 != MPI2_IOCSTATUS_SCSI_TASK_TERMINATED)) { 5741 st->direct_io = 0; 5742 st->scmd = scmd; 5743 memcpy(mpi_request->CDB.CDB32, scmd->cmnd, scmd->cmd_len); 5744 mpi_request->DevHandle = 5745 cpu_to_le16(sas_device_priv_data->sas_target->handle); 5746 ioc->put_smid_scsi_io(ioc, smid, 5747 sas_device_priv_data->sas_target->handle); 5748 return 0; 5749 } 5750 /* turning off TLR */ 5751 scsi_state = mpi_reply->SCSIState; 5752 if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID) 5753 response_code = 5754 le32_to_cpu(mpi_reply->ResponseInfo) & 0xFF; 5755 if (!sas_device_priv_data->tlr_snoop_check) { 5756 sas_device_priv_data->tlr_snoop_check++; 5757 if ((!ioc->is_warpdrive && 5758 !scsih_is_raid(&scmd->device->sdev_gendev) && 5759 !scsih_is_nvme(&scmd->device->sdev_gendev)) 5760 && sas_is_tlr_enabled(scmd->device) && 5761 response_code == MPI2_SCSITASKMGMT_RSP_INVALID_FRAME) { 5762 sas_disable_tlr(scmd->device); 5763 sdev_printk(KERN_INFO, scmd->device, "TLR disabled\n"); 5764 } 5765 } 5766 5767 xfer_cnt = le32_to_cpu(mpi_reply->TransferCount); 5768 scsi_set_resid(scmd, scsi_bufflen(scmd) - xfer_cnt); 5769 if (ioc_status & MPI2_IOCSTATUS_FLAG_LOG_INFO_AVAILABLE) 5770 log_info = le32_to_cpu(mpi_reply->IOCLogInfo); 5771 else 5772 log_info = 0; 5773 ioc_status &= MPI2_IOCSTATUS_MASK; 5774 scsi_status = mpi_reply->SCSIStatus; 5775 5776 if (ioc_status == MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN && xfer_cnt == 0 && 5777 (scsi_status == MPI2_SCSI_STATUS_BUSY || 5778 scsi_status == MPI2_SCSI_STATUS_RESERVATION_CONFLICT || 5779 scsi_status == MPI2_SCSI_STATUS_TASK_SET_FULL)) { 5780 ioc_status = MPI2_IOCSTATUS_SUCCESS; 5781 } 5782 5783 if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) { 5784 struct sense_info data; 5785 const void *sense_data = mpt3sas_base_get_sense_buffer(ioc, 5786 smid); 5787 u32 sz = min_t(u32, SCSI_SENSE_BUFFERSIZE, 5788 le32_to_cpu(mpi_reply->SenseCount)); 5789 memcpy(scmd->sense_buffer, sense_data, sz); 5790 _scsih_normalize_sense(scmd->sense_buffer, &data); 5791 /* failure prediction threshold exceeded */ 5792 if (data.asc == 0x5D) 5793 _scsih_smart_predicted_fault(ioc, 5794 le16_to_cpu(mpi_reply->DevHandle)); 5795 mpt3sas_trigger_scsi(ioc, data.skey, data.asc, data.ascq); 5796 5797 if ((ioc->logging_level & MPT_DEBUG_REPLY) && 5798 ((scmd->sense_buffer[2] == UNIT_ATTENTION) || 5799 (scmd->sense_buffer[2] == MEDIUM_ERROR) || 5800 (scmd->sense_buffer[2] == HARDWARE_ERROR))) 5801 _scsih_scsi_ioc_info(ioc, scmd, mpi_reply, smid); 5802 } 5803 switch (ioc_status) { 5804 case MPI2_IOCSTATUS_BUSY: 5805 case MPI2_IOCSTATUS_INSUFFICIENT_RESOURCES: 5806 scmd->result = SAM_STAT_BUSY; 5807 break; 5808 5809 case MPI2_IOCSTATUS_SCSI_DEVICE_NOT_THERE: 5810 scmd->result = DID_NO_CONNECT << 16; 5811 break; 5812 5813 case MPI2_IOCSTATUS_SCSI_IOC_TERMINATED: 5814 if (sas_device_priv_data->block) { 5815 scmd->result = DID_TRANSPORT_DISRUPTED << 16; 5816 goto out; 5817 } 5818 if (log_info == 0x31110630) { 5819 if (scmd->retries > 2) { 5820 scmd->result = DID_NO_CONNECT << 16; 5821 scsi_device_set_state(scmd->device, 5822 SDEV_OFFLINE); 5823 } else { 5824 scmd->result = DID_SOFT_ERROR << 16; 5825 scmd->device->expecting_cc_ua = 1; 5826 } 5827 break; 5828 } else if (log_info == VIRTUAL_IO_FAILED_RETRY) { 5829 scmd->result = DID_RESET << 16; 5830 break; 5831 } else if ((scmd->device->channel == RAID_CHANNEL) && 5832 (scsi_state == (MPI2_SCSI_STATE_TERMINATED | 5833 MPI2_SCSI_STATE_NO_SCSI_STATUS))) { 5834 scmd->result = DID_RESET << 16; 5835 break; 5836 } 5837 scmd->result = DID_SOFT_ERROR << 16; 5838 break; 5839 case MPI2_IOCSTATUS_SCSI_TASK_TERMINATED: 5840 case MPI2_IOCSTATUS_SCSI_EXT_TERMINATED: 5841 scmd->result = DID_RESET << 16; 5842 break; 5843 5844 case MPI2_IOCSTATUS_SCSI_RESIDUAL_MISMATCH: 5845 if ((xfer_cnt == 0) || (scmd->underflow > xfer_cnt)) 5846 scmd->result = DID_SOFT_ERROR << 16; 5847 else 5848 scmd->result = (DID_OK << 16) | scsi_status; 5849 break; 5850 5851 case MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN: 5852 scmd->result = (DID_OK << 16) | scsi_status; 5853 5854 if ((scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID)) 5855 break; 5856 5857 if (xfer_cnt < scmd->underflow) { 5858 if (scsi_status == SAM_STAT_BUSY) 5859 scmd->result = SAM_STAT_BUSY; 5860 else 5861 scmd->result = DID_SOFT_ERROR << 16; 5862 } else if (scsi_state & (MPI2_SCSI_STATE_AUTOSENSE_FAILED | 5863 MPI2_SCSI_STATE_NO_SCSI_STATUS)) 5864 scmd->result = DID_SOFT_ERROR << 16; 5865 else if (scsi_state & MPI2_SCSI_STATE_TERMINATED) 5866 scmd->result = DID_RESET << 16; 5867 else if (!xfer_cnt && scmd->cmnd[0] == REPORT_LUNS) { 5868 mpi_reply->SCSIState = MPI2_SCSI_STATE_AUTOSENSE_VALID; 5869 mpi_reply->SCSIStatus = SAM_STAT_CHECK_CONDITION; 5870 scsi_build_sense(scmd, 0, ILLEGAL_REQUEST, 5871 0x20, 0); 5872 } 5873 break; 5874 5875 case MPI2_IOCSTATUS_SCSI_DATA_OVERRUN: 5876 scsi_set_resid(scmd, 0); 5877 fallthrough; 5878 case MPI2_IOCSTATUS_SCSI_RECOVERED_ERROR: 5879 case MPI2_IOCSTATUS_SUCCESS: 5880 scmd->result = (DID_OK << 16) | scsi_status; 5881 if (response_code == 5882 MPI2_SCSITASKMGMT_RSP_INVALID_FRAME || 5883 (scsi_state & (MPI2_SCSI_STATE_AUTOSENSE_FAILED | 5884 MPI2_SCSI_STATE_NO_SCSI_STATUS))) 5885 scmd->result = DID_SOFT_ERROR << 16; 5886 else if (scsi_state & MPI2_SCSI_STATE_TERMINATED) 5887 scmd->result = DID_RESET << 16; 5888 break; 5889 5890 case MPI2_IOCSTATUS_EEDP_GUARD_ERROR: 5891 case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR: 5892 case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR: 5893 _scsih_eedp_error_handling(scmd, ioc_status); 5894 break; 5895 5896 case MPI2_IOCSTATUS_SCSI_PROTOCOL_ERROR: 5897 case MPI2_IOCSTATUS_INVALID_FUNCTION: 5898 case MPI2_IOCSTATUS_INVALID_SGL: 5899 case MPI2_IOCSTATUS_INTERNAL_ERROR: 5900 case MPI2_IOCSTATUS_INVALID_FIELD: 5901 case MPI2_IOCSTATUS_INVALID_STATE: 5902 case MPI2_IOCSTATUS_SCSI_IO_DATA_ERROR: 5903 case MPI2_IOCSTATUS_SCSI_TASK_MGMT_FAILED: 5904 case MPI2_IOCSTATUS_INSUFFICIENT_POWER: 5905 default: 5906 scmd->result = DID_SOFT_ERROR << 16; 5907 break; 5908 5909 } 5910 5911 if (scmd->result && (ioc->logging_level & MPT_DEBUG_REPLY)) 5912 _scsih_scsi_ioc_info(ioc , scmd, mpi_reply, smid); 5913 5914 out: 5915 5916 scsi_dma_unmap(scmd); 5917 mpt3sas_base_free_smid(ioc, smid); 5918 scsi_done(scmd); 5919 return 0; 5920 } 5921 5922 /** 5923 * _scsih_update_vphys_after_reset - update the Port's 5924 * vphys_list after reset 5925 * @ioc: per adapter object 5926 * 5927 * Returns nothing. 5928 */ 5929 static void 5930 _scsih_update_vphys_after_reset(struct MPT3SAS_ADAPTER *ioc) 5931 { 5932 u16 sz, ioc_status; 5933 int i; 5934 Mpi2ConfigReply_t mpi_reply; 5935 Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL; 5936 u16 attached_handle; 5937 u64 attached_sas_addr; 5938 u8 found = 0, port_id; 5939 Mpi2SasPhyPage0_t phy_pg0; 5940 struct hba_port *port, *port_next, *mport; 5941 struct virtual_phy *vphy, *vphy_next; 5942 struct _sas_device *sas_device; 5943 5944 /* 5945 * Mark all the vphys objects as dirty. 5946 */ 5947 list_for_each_entry_safe(port, port_next, 5948 &ioc->port_table_list, list) { 5949 if (!port->vphys_mask) 5950 continue; 5951 list_for_each_entry_safe(vphy, vphy_next, 5952 &port->vphys_list, list) { 5953 vphy->flags |= MPT_VPHY_FLAG_DIRTY_PHY; 5954 } 5955 } 5956 5957 /* 5958 * Read SASIOUnitPage0 to get each HBA Phy's data. 5959 */ 5960 sz = struct_size(sas_iounit_pg0, PhyData, ioc->sas_hba.num_phys); 5961 sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL); 5962 if (!sas_iounit_pg0) { 5963 ioc_err(ioc, "failure at %s:%d/%s()!\n", 5964 __FILE__, __LINE__, __func__); 5965 return; 5966 } 5967 if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply, 5968 sas_iounit_pg0, sz)) != 0) 5969 goto out; 5970 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; 5971 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) 5972 goto out; 5973 /* 5974 * Loop over each HBA Phy. 5975 */ 5976 for (i = 0; i < ioc->sas_hba.num_phys; i++) { 5977 /* 5978 * Check whether Phy's Negotiation Link Rate is > 1.5G or not. 5979 */ 5980 if ((sas_iounit_pg0->PhyData[i].NegotiatedLinkRate >> 4) < 5981 MPI2_SAS_NEG_LINK_RATE_1_5) 5982 continue; 5983 /* 5984 * Check whether Phy is connected to SEP device or not, 5985 * if it is SEP device then read the Phy's SASPHYPage0 data to 5986 * determine whether Phy is a virtual Phy or not. if it is 5987 * virtual phy then it is conformed that the attached remote 5988 * device is a HBA's vSES device. 5989 */ 5990 if (!(le32_to_cpu( 5991 sas_iounit_pg0->PhyData[i].ControllerPhyDeviceInfo) & 5992 MPI2_SAS_DEVICE_INFO_SEP)) 5993 continue; 5994 5995 if ((mpt3sas_config_get_phy_pg0(ioc, &mpi_reply, &phy_pg0, 5996 i))) { 5997 ioc_err(ioc, "failure at %s:%d/%s()!\n", 5998 __FILE__, __LINE__, __func__); 5999 continue; 6000 } 6001 6002 if (!(le32_to_cpu(phy_pg0.PhyInfo) & 6003 MPI2_SAS_PHYINFO_VIRTUAL_PHY)) 6004 continue; 6005 /* 6006 * Get the vSES device's SAS Address. 6007 */ 6008 attached_handle = le16_to_cpu( 6009 sas_iounit_pg0->PhyData[i].AttachedDevHandle); 6010 if (_scsih_get_sas_address(ioc, attached_handle, 6011 &attached_sas_addr) != 0) { 6012 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6013 __FILE__, __LINE__, __func__); 6014 continue; 6015 } 6016 6017 found = 0; 6018 port = port_next = NULL; 6019 /* 6020 * Loop over each virtual_phy object from 6021 * each port's vphys_list. 6022 */ 6023 list_for_each_entry_safe(port, 6024 port_next, &ioc->port_table_list, list) { 6025 if (!port->vphys_mask) 6026 continue; 6027 list_for_each_entry_safe(vphy, vphy_next, 6028 &port->vphys_list, list) { 6029 /* 6030 * Continue with next virtual_phy object 6031 * if the object is not marked as dirty. 6032 */ 6033 if (!(vphy->flags & MPT_VPHY_FLAG_DIRTY_PHY)) 6034 continue; 6035 6036 /* 6037 * Continue with next virtual_phy object 6038 * if the object's SAS Address is not equals 6039 * to current Phy's vSES device SAS Address. 6040 */ 6041 if (vphy->sas_address != attached_sas_addr) 6042 continue; 6043 /* 6044 * Enable current Phy number bit in object's 6045 * phy_mask field. 6046 */ 6047 if (!(vphy->phy_mask & (1 << i))) 6048 vphy->phy_mask = (1 << i); 6049 /* 6050 * Get hba_port object from hba_port table 6051 * corresponding to current phy's Port ID. 6052 * if there is no hba_port object corresponding 6053 * to Phy's Port ID then create a new hba_port 6054 * object & add to hba_port table. 6055 */ 6056 port_id = sas_iounit_pg0->PhyData[i].Port; 6057 mport = mpt3sas_get_port_by_id(ioc, port_id, 1); 6058 if (!mport) { 6059 mport = kzalloc( 6060 sizeof(struct hba_port), GFP_KERNEL); 6061 if (!mport) 6062 break; 6063 mport->port_id = port_id; 6064 ioc_info(ioc, 6065 "%s: hba_port entry: %p, port: %d is added to hba_port list\n", 6066 __func__, mport, mport->port_id); 6067 list_add_tail(&mport->list, 6068 &ioc->port_table_list); 6069 } 6070 /* 6071 * If mport & port pointers are not pointing to 6072 * same hba_port object then it means that vSES 6073 * device's Port ID got changed after reset and 6074 * hence move current virtual_phy object from 6075 * port's vphys_list to mport's vphys_list. 6076 */ 6077 if (port != mport) { 6078 if (!mport->vphys_mask) 6079 INIT_LIST_HEAD( 6080 &mport->vphys_list); 6081 mport->vphys_mask |= (1 << i); 6082 port->vphys_mask &= ~(1 << i); 6083 list_move(&vphy->list, 6084 &mport->vphys_list); 6085 sas_device = mpt3sas_get_sdev_by_addr( 6086 ioc, attached_sas_addr, port); 6087 if (sas_device) 6088 sas_device->port = mport; 6089 } 6090 /* 6091 * Earlier while updating the hba_port table, 6092 * it is determined that there is no other 6093 * direct attached device with mport's Port ID, 6094 * Hence mport was marked as dirty. Only vSES 6095 * device has this Port ID, so unmark the mport 6096 * as dirt. 6097 */ 6098 if (mport->flags & HBA_PORT_FLAG_DIRTY_PORT) { 6099 mport->sas_address = 0; 6100 mport->phy_mask = 0; 6101 mport->flags &= 6102 ~HBA_PORT_FLAG_DIRTY_PORT; 6103 } 6104 /* 6105 * Unmark current virtual_phy object as dirty. 6106 */ 6107 vphy->flags &= ~MPT_VPHY_FLAG_DIRTY_PHY; 6108 found = 1; 6109 break; 6110 } 6111 if (found) 6112 break; 6113 } 6114 } 6115 out: 6116 kfree(sas_iounit_pg0); 6117 } 6118 6119 /** 6120 * _scsih_get_port_table_after_reset - Construct temporary port table 6121 * @ioc: per adapter object 6122 * @port_table: address where port table needs to be constructed 6123 * 6124 * return number of HBA port entries available after reset. 6125 */ 6126 static int 6127 _scsih_get_port_table_after_reset(struct MPT3SAS_ADAPTER *ioc, 6128 struct hba_port *port_table) 6129 { 6130 u16 sz, ioc_status; 6131 int i, j; 6132 Mpi2ConfigReply_t mpi_reply; 6133 Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL; 6134 u16 attached_handle; 6135 u64 attached_sas_addr; 6136 u8 found = 0, port_count = 0, port_id; 6137 6138 sz = struct_size(sas_iounit_pg0, PhyData, ioc->sas_hba.num_phys); 6139 sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL); 6140 if (!sas_iounit_pg0) { 6141 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6142 __FILE__, __LINE__, __func__); 6143 return port_count; 6144 } 6145 6146 if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply, 6147 sas_iounit_pg0, sz)) != 0) 6148 goto out; 6149 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; 6150 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) 6151 goto out; 6152 for (i = 0; i < ioc->sas_hba.num_phys; i++) { 6153 found = 0; 6154 if ((sas_iounit_pg0->PhyData[i].NegotiatedLinkRate >> 4) < 6155 MPI2_SAS_NEG_LINK_RATE_1_5) 6156 continue; 6157 attached_handle = 6158 le16_to_cpu(sas_iounit_pg0->PhyData[i].AttachedDevHandle); 6159 if (_scsih_get_sas_address( 6160 ioc, attached_handle, &attached_sas_addr) != 0) { 6161 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6162 __FILE__, __LINE__, __func__); 6163 continue; 6164 } 6165 6166 for (j = 0; j < port_count; j++) { 6167 port_id = sas_iounit_pg0->PhyData[i].Port; 6168 if (port_table[j].port_id == port_id && 6169 port_table[j].sas_address == attached_sas_addr) { 6170 port_table[j].phy_mask |= (1 << i); 6171 found = 1; 6172 break; 6173 } 6174 } 6175 6176 if (found) 6177 continue; 6178 6179 port_id = sas_iounit_pg0->PhyData[i].Port; 6180 port_table[port_count].port_id = port_id; 6181 port_table[port_count].phy_mask = (1 << i); 6182 port_table[port_count].sas_address = attached_sas_addr; 6183 port_count++; 6184 } 6185 out: 6186 kfree(sas_iounit_pg0); 6187 return port_count; 6188 } 6189 6190 enum hba_port_matched_codes { 6191 NOT_MATCHED = 0, 6192 MATCHED_WITH_ADDR_AND_PHYMASK, 6193 MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT, 6194 MATCHED_WITH_ADDR_AND_SUBPHYMASK, 6195 MATCHED_WITH_ADDR, 6196 }; 6197 6198 /** 6199 * _scsih_look_and_get_matched_port_entry - Get matched hba port entry 6200 * from HBA port table 6201 * @ioc: per adapter object 6202 * @port_entry: hba port entry from temporary port table which needs to be 6203 * searched for matched entry in the HBA port table 6204 * @matched_port_entry: save matched hba port entry here 6205 * @count: count of matched entries 6206 * 6207 * return type of matched entry found. 6208 */ 6209 static enum hba_port_matched_codes 6210 _scsih_look_and_get_matched_port_entry(struct MPT3SAS_ADAPTER *ioc, 6211 struct hba_port *port_entry, 6212 struct hba_port **matched_port_entry, int *count) 6213 { 6214 struct hba_port *port_table_entry, *matched_port = NULL; 6215 enum hba_port_matched_codes matched_code = NOT_MATCHED; 6216 int lcount = 0; 6217 *matched_port_entry = NULL; 6218 6219 list_for_each_entry(port_table_entry, &ioc->port_table_list, list) { 6220 if (!(port_table_entry->flags & HBA_PORT_FLAG_DIRTY_PORT)) 6221 continue; 6222 6223 if ((port_table_entry->sas_address == port_entry->sas_address) 6224 && (port_table_entry->phy_mask == port_entry->phy_mask)) { 6225 matched_code = MATCHED_WITH_ADDR_AND_PHYMASK; 6226 matched_port = port_table_entry; 6227 break; 6228 } 6229 6230 if ((port_table_entry->sas_address == port_entry->sas_address) 6231 && (port_table_entry->phy_mask & port_entry->phy_mask) 6232 && (port_table_entry->port_id == port_entry->port_id)) { 6233 matched_code = MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT; 6234 matched_port = port_table_entry; 6235 continue; 6236 } 6237 6238 if ((port_table_entry->sas_address == port_entry->sas_address) 6239 && (port_table_entry->phy_mask & port_entry->phy_mask)) { 6240 if (matched_code == 6241 MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT) 6242 continue; 6243 matched_code = MATCHED_WITH_ADDR_AND_SUBPHYMASK; 6244 matched_port = port_table_entry; 6245 continue; 6246 } 6247 6248 if (port_table_entry->sas_address == port_entry->sas_address) { 6249 if (matched_code == 6250 MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT) 6251 continue; 6252 if (matched_code == MATCHED_WITH_ADDR_AND_SUBPHYMASK) 6253 continue; 6254 matched_code = MATCHED_WITH_ADDR; 6255 matched_port = port_table_entry; 6256 lcount++; 6257 } 6258 } 6259 6260 *matched_port_entry = matched_port; 6261 if (matched_code == MATCHED_WITH_ADDR) 6262 *count = lcount; 6263 return matched_code; 6264 } 6265 6266 /** 6267 * _scsih_del_phy_part_of_anther_port - remove phy if it 6268 * is a part of anther port 6269 *@ioc: per adapter object 6270 *@port_table: port table after reset 6271 *@index: hba port entry index 6272 *@port_count: number of ports available after host reset 6273 *@offset: HBA phy bit offset 6274 * 6275 */ 6276 static void 6277 _scsih_del_phy_part_of_anther_port(struct MPT3SAS_ADAPTER *ioc, 6278 struct hba_port *port_table, 6279 int index, u8 port_count, int offset) 6280 { 6281 struct _sas_node *sas_node = &ioc->sas_hba; 6282 u32 i, found = 0; 6283 6284 for (i = 0; i < port_count; i++) { 6285 if (i == index) 6286 continue; 6287 6288 if (port_table[i].phy_mask & (1 << offset)) { 6289 mpt3sas_transport_del_phy_from_an_existing_port( 6290 ioc, sas_node, &sas_node->phy[offset]); 6291 found = 1; 6292 break; 6293 } 6294 } 6295 if (!found) 6296 port_table[index].phy_mask |= (1 << offset); 6297 } 6298 6299 /** 6300 * _scsih_add_or_del_phys_from_existing_port - add/remove phy to/from 6301 * right port 6302 *@ioc: per adapter object 6303 *@hba_port_entry: hba port table entry 6304 *@port_table: temporary port table 6305 *@index: hba port entry index 6306 *@port_count: number of ports available after host reset 6307 * 6308 */ 6309 static void 6310 _scsih_add_or_del_phys_from_existing_port(struct MPT3SAS_ADAPTER *ioc, 6311 struct hba_port *hba_port_entry, struct hba_port *port_table, 6312 int index, int port_count) 6313 { 6314 u32 phy_mask, offset = 0; 6315 struct _sas_node *sas_node = &ioc->sas_hba; 6316 6317 phy_mask = hba_port_entry->phy_mask ^ port_table[index].phy_mask; 6318 6319 for (offset = 0; offset < ioc->sas_hba.num_phys; offset++) { 6320 if (phy_mask & (1 << offset)) { 6321 if (!(port_table[index].phy_mask & (1 << offset))) { 6322 _scsih_del_phy_part_of_anther_port( 6323 ioc, port_table, index, port_count, 6324 offset); 6325 continue; 6326 } 6327 if (sas_node->phy[offset].phy_belongs_to_port) 6328 mpt3sas_transport_del_phy_from_an_existing_port( 6329 ioc, sas_node, &sas_node->phy[offset]); 6330 mpt3sas_transport_add_phy_to_an_existing_port( 6331 ioc, sas_node, &sas_node->phy[offset], 6332 hba_port_entry->sas_address, 6333 hba_port_entry); 6334 } 6335 } 6336 } 6337 6338 /** 6339 * _scsih_del_dirty_vphy - delete virtual_phy objects marked as dirty. 6340 * @ioc: per adapter object 6341 * 6342 * Returns nothing. 6343 */ 6344 static void 6345 _scsih_del_dirty_vphy(struct MPT3SAS_ADAPTER *ioc) 6346 { 6347 struct hba_port *port, *port_next; 6348 struct virtual_phy *vphy, *vphy_next; 6349 6350 list_for_each_entry_safe(port, port_next, 6351 &ioc->port_table_list, list) { 6352 if (!port->vphys_mask) 6353 continue; 6354 list_for_each_entry_safe(vphy, vphy_next, 6355 &port->vphys_list, list) { 6356 if (vphy->flags & MPT_VPHY_FLAG_DIRTY_PHY) { 6357 drsprintk(ioc, ioc_info(ioc, 6358 "Deleting vphy %p entry from port id: %d\t, Phy_mask 0x%08x\n", 6359 vphy, port->port_id, 6360 vphy->phy_mask)); 6361 port->vphys_mask &= ~vphy->phy_mask; 6362 list_del(&vphy->list); 6363 kfree(vphy); 6364 } 6365 } 6366 if (!port->vphys_mask && !port->sas_address) 6367 port->flags |= HBA_PORT_FLAG_DIRTY_PORT; 6368 } 6369 } 6370 6371 /** 6372 * _scsih_del_dirty_port_entries - delete dirty port entries from port list 6373 * after host reset 6374 *@ioc: per adapter object 6375 * 6376 */ 6377 static void 6378 _scsih_del_dirty_port_entries(struct MPT3SAS_ADAPTER *ioc) 6379 { 6380 struct hba_port *port, *port_next; 6381 6382 list_for_each_entry_safe(port, port_next, 6383 &ioc->port_table_list, list) { 6384 if (!(port->flags & HBA_PORT_FLAG_DIRTY_PORT) || 6385 port->flags & HBA_PORT_FLAG_NEW_PORT) 6386 continue; 6387 6388 drsprintk(ioc, ioc_info(ioc, 6389 "Deleting port table entry %p having Port: %d\t Phy_mask 0x%08x\n", 6390 port, port->port_id, port->phy_mask)); 6391 list_del(&port->list); 6392 kfree(port); 6393 } 6394 } 6395 6396 /** 6397 * _scsih_sas_port_refresh - Update HBA port table after host reset 6398 * @ioc: per adapter object 6399 */ 6400 static void 6401 _scsih_sas_port_refresh(struct MPT3SAS_ADAPTER *ioc) 6402 { 6403 u32 port_count = 0; 6404 struct hba_port *port_table; 6405 struct hba_port *port_table_entry; 6406 struct hba_port *port_entry = NULL; 6407 int i, j, count = 0, lcount = 0; 6408 int ret; 6409 u64 sas_addr; 6410 u8 num_phys; 6411 6412 drsprintk(ioc, ioc_info(ioc, 6413 "updating ports for sas_host(0x%016llx)\n", 6414 (unsigned long long)ioc->sas_hba.sas_address)); 6415 6416 mpt3sas_config_get_number_hba_phys(ioc, &num_phys); 6417 if (!num_phys) { 6418 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6419 __FILE__, __LINE__, __func__); 6420 return; 6421 } 6422 6423 if (num_phys > ioc->sas_hba.nr_phys_allocated) { 6424 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6425 __FILE__, __LINE__, __func__); 6426 return; 6427 } 6428 ioc->sas_hba.num_phys = num_phys; 6429 6430 port_table = kcalloc(ioc->sas_hba.num_phys, 6431 sizeof(struct hba_port), GFP_KERNEL); 6432 if (!port_table) 6433 return; 6434 6435 port_count = _scsih_get_port_table_after_reset(ioc, port_table); 6436 if (!port_count) 6437 return; 6438 6439 drsprintk(ioc, ioc_info(ioc, "New Port table\n")); 6440 for (j = 0; j < port_count; j++) 6441 drsprintk(ioc, ioc_info(ioc, 6442 "Port: %d\t Phy_mask 0x%08x\t sas_addr(0x%016llx)\n", 6443 port_table[j].port_id, 6444 port_table[j].phy_mask, port_table[j].sas_address)); 6445 6446 list_for_each_entry(port_table_entry, &ioc->port_table_list, list) 6447 port_table_entry->flags |= HBA_PORT_FLAG_DIRTY_PORT; 6448 6449 drsprintk(ioc, ioc_info(ioc, "Old Port table\n")); 6450 port_table_entry = NULL; 6451 list_for_each_entry(port_table_entry, &ioc->port_table_list, list) { 6452 drsprintk(ioc, ioc_info(ioc, 6453 "Port: %d\t Phy_mask 0x%08x\t sas_addr(0x%016llx)\n", 6454 port_table_entry->port_id, 6455 port_table_entry->phy_mask, 6456 port_table_entry->sas_address)); 6457 } 6458 6459 for (j = 0; j < port_count; j++) { 6460 ret = _scsih_look_and_get_matched_port_entry(ioc, 6461 &port_table[j], &port_entry, &count); 6462 if (!port_entry) { 6463 drsprintk(ioc, ioc_info(ioc, 6464 "No Matched entry for sas_addr(0x%16llx), Port:%d\n", 6465 port_table[j].sas_address, 6466 port_table[j].port_id)); 6467 continue; 6468 } 6469 6470 switch (ret) { 6471 case MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT: 6472 case MATCHED_WITH_ADDR_AND_SUBPHYMASK: 6473 _scsih_add_or_del_phys_from_existing_port(ioc, 6474 port_entry, port_table, j, port_count); 6475 break; 6476 case MATCHED_WITH_ADDR: 6477 sas_addr = port_table[j].sas_address; 6478 for (i = 0; i < port_count; i++) { 6479 if (port_table[i].sas_address == sas_addr) 6480 lcount++; 6481 } 6482 6483 if (count > 1 || lcount > 1) 6484 port_entry = NULL; 6485 else 6486 _scsih_add_or_del_phys_from_existing_port(ioc, 6487 port_entry, port_table, j, port_count); 6488 } 6489 6490 if (!port_entry) 6491 continue; 6492 6493 if (port_entry->port_id != port_table[j].port_id) 6494 port_entry->port_id = port_table[j].port_id; 6495 port_entry->flags &= ~HBA_PORT_FLAG_DIRTY_PORT; 6496 port_entry->phy_mask = port_table[j].phy_mask; 6497 } 6498 6499 port_table_entry = NULL; 6500 } 6501 6502 /** 6503 * _scsih_alloc_vphy - allocate virtual_phy object 6504 * @ioc: per adapter object 6505 * @port_id: Port ID number 6506 * @phy_num: HBA Phy number 6507 * 6508 * Returns allocated virtual_phy object. 6509 */ 6510 static struct virtual_phy * 6511 _scsih_alloc_vphy(struct MPT3SAS_ADAPTER *ioc, u8 port_id, u8 phy_num) 6512 { 6513 struct virtual_phy *vphy; 6514 struct hba_port *port; 6515 6516 port = mpt3sas_get_port_by_id(ioc, port_id, 0); 6517 if (!port) 6518 return NULL; 6519 6520 vphy = mpt3sas_get_vphy_by_phy(ioc, port, phy_num); 6521 if (!vphy) { 6522 vphy = kzalloc(sizeof(struct virtual_phy), GFP_KERNEL); 6523 if (!vphy) 6524 return NULL; 6525 6526 if (!port->vphys_mask) 6527 INIT_LIST_HEAD(&port->vphys_list); 6528 6529 /* 6530 * Enable bit corresponding to HBA phy number on its 6531 * parent hba_port object's vphys_mask field. 6532 */ 6533 port->vphys_mask |= (1 << phy_num); 6534 vphy->phy_mask |= (1 << phy_num); 6535 6536 list_add_tail(&vphy->list, &port->vphys_list); 6537 6538 ioc_info(ioc, 6539 "vphy entry: %p, port id: %d, phy:%d is added to port's vphys_list\n", 6540 vphy, port->port_id, phy_num); 6541 } 6542 return vphy; 6543 } 6544 6545 /** 6546 * _scsih_sas_host_refresh - refreshing sas host object contents 6547 * @ioc: per adapter object 6548 * Context: user 6549 * 6550 * During port enable, fw will send topology events for every device. Its 6551 * possible that the handles may change from the previous setting, so this 6552 * code keeping handles updating if changed. 6553 */ 6554 static void 6555 _scsih_sas_host_refresh(struct MPT3SAS_ADAPTER *ioc) 6556 { 6557 u16 sz; 6558 u16 ioc_status; 6559 int i; 6560 Mpi2ConfigReply_t mpi_reply; 6561 Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL; 6562 u16 attached_handle; 6563 u8 link_rate, port_id; 6564 struct hba_port *port; 6565 Mpi2SasPhyPage0_t phy_pg0; 6566 6567 dtmprintk(ioc, 6568 ioc_info(ioc, "updating handles for sas_host(0x%016llx)\n", 6569 (u64)ioc->sas_hba.sas_address)); 6570 6571 sz = struct_size(sas_iounit_pg0, PhyData, ioc->sas_hba.num_phys); 6572 sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL); 6573 if (!sas_iounit_pg0) { 6574 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6575 __FILE__, __LINE__, __func__); 6576 return; 6577 } 6578 6579 if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply, 6580 sas_iounit_pg0, sz)) != 0) 6581 goto out; 6582 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; 6583 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) 6584 goto out; 6585 for (i = 0; i < ioc->sas_hba.num_phys ; i++) { 6586 link_rate = sas_iounit_pg0->PhyData[i].NegotiatedLinkRate >> 4; 6587 if (i == 0) 6588 ioc->sas_hba.handle = le16_to_cpu( 6589 sas_iounit_pg0->PhyData[0].ControllerDevHandle); 6590 port_id = sas_iounit_pg0->PhyData[i].Port; 6591 if (!(mpt3sas_get_port_by_id(ioc, port_id, 0))) { 6592 port = kzalloc(sizeof(struct hba_port), GFP_KERNEL); 6593 if (!port) 6594 goto out; 6595 6596 port->port_id = port_id; 6597 ioc_info(ioc, 6598 "hba_port entry: %p, port: %d is added to hba_port list\n", 6599 port, port->port_id); 6600 if (ioc->shost_recovery) 6601 port->flags = HBA_PORT_FLAG_NEW_PORT; 6602 list_add_tail(&port->list, &ioc->port_table_list); 6603 } 6604 /* 6605 * Check whether current Phy belongs to HBA vSES device or not. 6606 */ 6607 if (le32_to_cpu(sas_iounit_pg0->PhyData[i].ControllerPhyDeviceInfo) & 6608 MPI2_SAS_DEVICE_INFO_SEP && 6609 (link_rate >= MPI2_SAS_NEG_LINK_RATE_1_5)) { 6610 if ((mpt3sas_config_get_phy_pg0(ioc, &mpi_reply, 6611 &phy_pg0, i))) { 6612 ioc_err(ioc, 6613 "failure at %s:%d/%s()!\n", 6614 __FILE__, __LINE__, __func__); 6615 goto out; 6616 } 6617 if (!(le32_to_cpu(phy_pg0.PhyInfo) & 6618 MPI2_SAS_PHYINFO_VIRTUAL_PHY)) 6619 continue; 6620 /* 6621 * Allocate a virtual_phy object for vSES device, if 6622 * this vSES device is hot added. 6623 */ 6624 if (!_scsih_alloc_vphy(ioc, port_id, i)) 6625 goto out; 6626 ioc->sas_hba.phy[i].hba_vphy = 1; 6627 } 6628 6629 /* 6630 * Add new HBA phys to STL if these new phys got added as part 6631 * of HBA Firmware upgrade/downgrade operation. 6632 */ 6633 if (!ioc->sas_hba.phy[i].phy) { 6634 if ((mpt3sas_config_get_phy_pg0(ioc, &mpi_reply, 6635 &phy_pg0, i))) { 6636 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6637 __FILE__, __LINE__, __func__); 6638 continue; 6639 } 6640 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 6641 MPI2_IOCSTATUS_MASK; 6642 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 6643 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6644 __FILE__, __LINE__, __func__); 6645 continue; 6646 } 6647 ioc->sas_hba.phy[i].phy_id = i; 6648 mpt3sas_transport_add_host_phy(ioc, 6649 &ioc->sas_hba.phy[i], phy_pg0, 6650 ioc->sas_hba.parent_dev); 6651 continue; 6652 } 6653 ioc->sas_hba.phy[i].handle = ioc->sas_hba.handle; 6654 attached_handle = le16_to_cpu(sas_iounit_pg0->PhyData[i]. 6655 AttachedDevHandle); 6656 if (attached_handle && link_rate < MPI2_SAS_NEG_LINK_RATE_1_5) 6657 link_rate = MPI2_SAS_NEG_LINK_RATE_1_5; 6658 ioc->sas_hba.phy[i].port = 6659 mpt3sas_get_port_by_id(ioc, port_id, 0); 6660 mpt3sas_transport_update_links(ioc, ioc->sas_hba.sas_address, 6661 attached_handle, i, link_rate, 6662 ioc->sas_hba.phy[i].port); 6663 } 6664 /* 6665 * Clear the phy details if this phy got disabled as part of 6666 * HBA Firmware upgrade/downgrade operation. 6667 */ 6668 for (i = ioc->sas_hba.num_phys; 6669 i < ioc->sas_hba.nr_phys_allocated; i++) { 6670 if (ioc->sas_hba.phy[i].phy && 6671 ioc->sas_hba.phy[i].phy->negotiated_linkrate >= 6672 SAS_LINK_RATE_1_5_GBPS) 6673 mpt3sas_transport_update_links(ioc, 6674 ioc->sas_hba.sas_address, 0, i, 6675 MPI2_SAS_NEG_LINK_RATE_PHY_DISABLED, NULL); 6676 } 6677 out: 6678 kfree(sas_iounit_pg0); 6679 } 6680 6681 /** 6682 * _scsih_sas_host_add - create sas host object 6683 * @ioc: per adapter object 6684 * 6685 * Creating host side data object, stored in ioc->sas_hba 6686 */ 6687 static void 6688 _scsih_sas_host_add(struct MPT3SAS_ADAPTER *ioc) 6689 { 6690 int i; 6691 Mpi2ConfigReply_t mpi_reply; 6692 Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL; 6693 Mpi2SasIOUnitPage1_t *sas_iounit_pg1 = NULL; 6694 Mpi2SasPhyPage0_t phy_pg0; 6695 Mpi2SasDevicePage0_t sas_device_pg0; 6696 Mpi2SasEnclosurePage0_t enclosure_pg0; 6697 u16 ioc_status; 6698 u16 sz; 6699 u8 device_missing_delay; 6700 u8 num_phys, port_id; 6701 struct hba_port *port; 6702 6703 mpt3sas_config_get_number_hba_phys(ioc, &num_phys); 6704 if (!num_phys) { 6705 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6706 __FILE__, __LINE__, __func__); 6707 return; 6708 } 6709 6710 ioc->sas_hba.nr_phys_allocated = max_t(u8, 6711 MPT_MAX_HBA_NUM_PHYS, num_phys); 6712 ioc->sas_hba.phy = kcalloc(ioc->sas_hba.nr_phys_allocated, 6713 sizeof(struct _sas_phy), GFP_KERNEL); 6714 if (!ioc->sas_hba.phy) { 6715 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6716 __FILE__, __LINE__, __func__); 6717 goto out; 6718 } 6719 ioc->sas_hba.num_phys = num_phys; 6720 6721 /* sas_iounit page 0 */ 6722 sz = struct_size(sas_iounit_pg0, PhyData, ioc->sas_hba.num_phys); 6723 sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL); 6724 if (!sas_iounit_pg0) { 6725 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6726 __FILE__, __LINE__, __func__); 6727 return; 6728 } 6729 if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply, 6730 sas_iounit_pg0, sz))) { 6731 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6732 __FILE__, __LINE__, __func__); 6733 goto out; 6734 } 6735 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 6736 MPI2_IOCSTATUS_MASK; 6737 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 6738 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6739 __FILE__, __LINE__, __func__); 6740 goto out; 6741 } 6742 6743 /* sas_iounit page 1 */ 6744 sz = struct_size(sas_iounit_pg1, PhyData, ioc->sas_hba.num_phys); 6745 sas_iounit_pg1 = kzalloc(sz, GFP_KERNEL); 6746 if (!sas_iounit_pg1) { 6747 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6748 __FILE__, __LINE__, __func__); 6749 goto out; 6750 } 6751 if ((mpt3sas_config_get_sas_iounit_pg1(ioc, &mpi_reply, 6752 sas_iounit_pg1, sz))) { 6753 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6754 __FILE__, __LINE__, __func__); 6755 goto out; 6756 } 6757 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 6758 MPI2_IOCSTATUS_MASK; 6759 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 6760 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6761 __FILE__, __LINE__, __func__); 6762 goto out; 6763 } 6764 6765 ioc->io_missing_delay = 6766 sas_iounit_pg1->IODeviceMissingDelay; 6767 device_missing_delay = 6768 sas_iounit_pg1->ReportDeviceMissingDelay; 6769 if (device_missing_delay & MPI2_SASIOUNIT1_REPORT_MISSING_UNIT_16) 6770 ioc->device_missing_delay = (device_missing_delay & 6771 MPI2_SASIOUNIT1_REPORT_MISSING_TIMEOUT_MASK) * 16; 6772 else 6773 ioc->device_missing_delay = device_missing_delay & 6774 MPI2_SASIOUNIT1_REPORT_MISSING_TIMEOUT_MASK; 6775 6776 ioc->sas_hba.parent_dev = &ioc->shost->shost_gendev; 6777 for (i = 0; i < ioc->sas_hba.num_phys ; i++) { 6778 if ((mpt3sas_config_get_phy_pg0(ioc, &mpi_reply, &phy_pg0, 6779 i))) { 6780 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6781 __FILE__, __LINE__, __func__); 6782 goto out; 6783 } 6784 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 6785 MPI2_IOCSTATUS_MASK; 6786 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 6787 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6788 __FILE__, __LINE__, __func__); 6789 goto out; 6790 } 6791 6792 if (i == 0) 6793 ioc->sas_hba.handle = le16_to_cpu(sas_iounit_pg0-> 6794 PhyData[0].ControllerDevHandle); 6795 6796 port_id = sas_iounit_pg0->PhyData[i].Port; 6797 if (!(mpt3sas_get_port_by_id(ioc, port_id, 0))) { 6798 port = kzalloc(sizeof(struct hba_port), GFP_KERNEL); 6799 if (!port) 6800 goto out; 6801 6802 port->port_id = port_id; 6803 ioc_info(ioc, 6804 "hba_port entry: %p, port: %d is added to hba_port list\n", 6805 port, port->port_id); 6806 list_add_tail(&port->list, 6807 &ioc->port_table_list); 6808 } 6809 6810 /* 6811 * Check whether current Phy belongs to HBA vSES device or not. 6812 */ 6813 if ((le32_to_cpu(phy_pg0.PhyInfo) & 6814 MPI2_SAS_PHYINFO_VIRTUAL_PHY) && 6815 (phy_pg0.NegotiatedLinkRate >> 4) >= 6816 MPI2_SAS_NEG_LINK_RATE_1_5) { 6817 /* 6818 * Allocate a virtual_phy object for vSES device. 6819 */ 6820 if (!_scsih_alloc_vphy(ioc, port_id, i)) 6821 goto out; 6822 ioc->sas_hba.phy[i].hba_vphy = 1; 6823 } 6824 6825 ioc->sas_hba.phy[i].handle = ioc->sas_hba.handle; 6826 ioc->sas_hba.phy[i].phy_id = i; 6827 ioc->sas_hba.phy[i].port = 6828 mpt3sas_get_port_by_id(ioc, port_id, 0); 6829 mpt3sas_transport_add_host_phy(ioc, &ioc->sas_hba.phy[i], 6830 phy_pg0, ioc->sas_hba.parent_dev); 6831 } 6832 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, 6833 MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, ioc->sas_hba.handle))) { 6834 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6835 __FILE__, __LINE__, __func__); 6836 goto out; 6837 } 6838 ioc->sas_hba.enclosure_handle = 6839 le16_to_cpu(sas_device_pg0.EnclosureHandle); 6840 ioc->sas_hba.sas_address = le64_to_cpu(sas_device_pg0.SASAddress); 6841 ioc_info(ioc, "host_add: handle(0x%04x), sas_addr(0x%016llx), phys(%d)\n", 6842 ioc->sas_hba.handle, 6843 (u64)ioc->sas_hba.sas_address, 6844 ioc->sas_hba.num_phys); 6845 6846 if (ioc->sas_hba.enclosure_handle) { 6847 if (!(mpt3sas_config_get_enclosure_pg0(ioc, &mpi_reply, 6848 &enclosure_pg0, MPI2_SAS_ENCLOS_PGAD_FORM_HANDLE, 6849 ioc->sas_hba.enclosure_handle))) 6850 ioc->sas_hba.enclosure_logical_id = 6851 le64_to_cpu(enclosure_pg0.EnclosureLogicalID); 6852 } 6853 6854 out: 6855 kfree(sas_iounit_pg1); 6856 kfree(sas_iounit_pg0); 6857 } 6858 6859 /** 6860 * _scsih_expander_add - creating expander object 6861 * @ioc: per adapter object 6862 * @handle: expander handle 6863 * 6864 * Creating expander object, stored in ioc->sas_expander_list. 6865 * 6866 * Return: 0 for success, else error. 6867 */ 6868 static int 6869 _scsih_expander_add(struct MPT3SAS_ADAPTER *ioc, u16 handle) 6870 { 6871 struct _sas_node *sas_expander; 6872 struct _enclosure_node *enclosure_dev; 6873 Mpi2ConfigReply_t mpi_reply; 6874 Mpi2ExpanderPage0_t expander_pg0; 6875 Mpi2ExpanderPage1_t expander_pg1; 6876 u32 ioc_status; 6877 u16 parent_handle; 6878 u64 sas_address, sas_address_parent = 0; 6879 int i; 6880 unsigned long flags; 6881 struct _sas_port *mpt3sas_port = NULL; 6882 u8 port_id; 6883 6884 int rc = 0; 6885 6886 if (!handle) 6887 return -1; 6888 6889 if (ioc->shost_recovery || ioc->pci_error_recovery) 6890 return -1; 6891 6892 if ((mpt3sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0, 6893 MPI2_SAS_EXPAND_PGAD_FORM_HNDL, handle))) { 6894 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6895 __FILE__, __LINE__, __func__); 6896 return -1; 6897 } 6898 6899 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 6900 MPI2_IOCSTATUS_MASK; 6901 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 6902 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6903 __FILE__, __LINE__, __func__); 6904 return -1; 6905 } 6906 6907 /* handle out of order topology events */ 6908 parent_handle = le16_to_cpu(expander_pg0.ParentDevHandle); 6909 if (_scsih_get_sas_address(ioc, parent_handle, &sas_address_parent) 6910 != 0) { 6911 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6912 __FILE__, __LINE__, __func__); 6913 return -1; 6914 } 6915 6916 port_id = expander_pg0.PhysicalPort; 6917 if (sas_address_parent != ioc->sas_hba.sas_address) { 6918 spin_lock_irqsave(&ioc->sas_node_lock, flags); 6919 sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc, 6920 sas_address_parent, 6921 mpt3sas_get_port_by_id(ioc, port_id, 0)); 6922 spin_unlock_irqrestore(&ioc->sas_node_lock, flags); 6923 if (!sas_expander) { 6924 rc = _scsih_expander_add(ioc, parent_handle); 6925 if (rc != 0) 6926 return rc; 6927 } 6928 } 6929 6930 spin_lock_irqsave(&ioc->sas_node_lock, flags); 6931 sas_address = le64_to_cpu(expander_pg0.SASAddress); 6932 sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc, 6933 sas_address, mpt3sas_get_port_by_id(ioc, port_id, 0)); 6934 spin_unlock_irqrestore(&ioc->sas_node_lock, flags); 6935 6936 if (sas_expander) 6937 return 0; 6938 6939 sas_expander = kzalloc(sizeof(struct _sas_node), 6940 GFP_KERNEL); 6941 if (!sas_expander) { 6942 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6943 __FILE__, __LINE__, __func__); 6944 return -1; 6945 } 6946 6947 sas_expander->handle = handle; 6948 sas_expander->num_phys = expander_pg0.NumPhys; 6949 sas_expander->sas_address_parent = sas_address_parent; 6950 sas_expander->sas_address = sas_address; 6951 sas_expander->port = mpt3sas_get_port_by_id(ioc, port_id, 0); 6952 if (!sas_expander->port) { 6953 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6954 __FILE__, __LINE__, __func__); 6955 rc = -1; 6956 goto out_fail; 6957 } 6958 6959 ioc_info(ioc, "expander_add: handle(0x%04x), parent(0x%04x), sas_addr(0x%016llx), phys(%d)\n", 6960 handle, parent_handle, 6961 (u64)sas_expander->sas_address, sas_expander->num_phys); 6962 6963 if (!sas_expander->num_phys) { 6964 rc = -1; 6965 goto out_fail; 6966 } 6967 sas_expander->phy = kcalloc(sas_expander->num_phys, 6968 sizeof(struct _sas_phy), GFP_KERNEL); 6969 if (!sas_expander->phy) { 6970 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6971 __FILE__, __LINE__, __func__); 6972 rc = -1; 6973 goto out_fail; 6974 } 6975 6976 INIT_LIST_HEAD(&sas_expander->sas_port_list); 6977 mpt3sas_port = mpt3sas_transport_port_add(ioc, handle, 6978 sas_address_parent, sas_expander->port); 6979 if (!mpt3sas_port) { 6980 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6981 __FILE__, __LINE__, __func__); 6982 rc = -1; 6983 goto out_fail; 6984 } 6985 sas_expander->parent_dev = &mpt3sas_port->rphy->dev; 6986 sas_expander->rphy = mpt3sas_port->rphy; 6987 6988 for (i = 0 ; i < sas_expander->num_phys ; i++) { 6989 if ((mpt3sas_config_get_expander_pg1(ioc, &mpi_reply, 6990 &expander_pg1, i, handle))) { 6991 ioc_err(ioc, "failure at %s:%d/%s()!\n", 6992 __FILE__, __LINE__, __func__); 6993 rc = -1; 6994 goto out_fail; 6995 } 6996 sas_expander->phy[i].handle = handle; 6997 sas_expander->phy[i].phy_id = i; 6998 sas_expander->phy[i].port = 6999 mpt3sas_get_port_by_id(ioc, port_id, 0); 7000 7001 if ((mpt3sas_transport_add_expander_phy(ioc, 7002 &sas_expander->phy[i], expander_pg1, 7003 sas_expander->parent_dev))) { 7004 ioc_err(ioc, "failure at %s:%d/%s()!\n", 7005 __FILE__, __LINE__, __func__); 7006 rc = -1; 7007 goto out_fail; 7008 } 7009 } 7010 7011 if (sas_expander->enclosure_handle) { 7012 enclosure_dev = 7013 mpt3sas_scsih_enclosure_find_by_handle(ioc, 7014 sas_expander->enclosure_handle); 7015 if (enclosure_dev) 7016 sas_expander->enclosure_logical_id = 7017 le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); 7018 } 7019 7020 _scsih_expander_node_add(ioc, sas_expander); 7021 return 0; 7022 7023 out_fail: 7024 7025 if (mpt3sas_port) 7026 mpt3sas_transport_port_remove(ioc, sas_expander->sas_address, 7027 sas_address_parent, sas_expander->port); 7028 kfree(sas_expander); 7029 return rc; 7030 } 7031 7032 /** 7033 * mpt3sas_expander_remove - removing expander object 7034 * @ioc: per adapter object 7035 * @sas_address: expander sas_address 7036 * @port: hba port entry 7037 */ 7038 void 7039 mpt3sas_expander_remove(struct MPT3SAS_ADAPTER *ioc, u64 sas_address, 7040 struct hba_port *port) 7041 { 7042 struct _sas_node *sas_expander; 7043 unsigned long flags; 7044 7045 if (ioc->shost_recovery) 7046 return; 7047 7048 if (!port) 7049 return; 7050 7051 spin_lock_irqsave(&ioc->sas_node_lock, flags); 7052 sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc, 7053 sas_address, port); 7054 spin_unlock_irqrestore(&ioc->sas_node_lock, flags); 7055 if (sas_expander) 7056 _scsih_expander_node_remove(ioc, sas_expander); 7057 } 7058 7059 /** 7060 * _scsih_done - internal SCSI_IO callback handler. 7061 * @ioc: per adapter object 7062 * @smid: system request message index 7063 * @msix_index: MSIX table index supplied by the OS 7064 * @reply: reply message frame(lower 32bit addr) 7065 * 7066 * Callback handler when sending internal generated SCSI_IO. 7067 * The callback index passed is `ioc->scsih_cb_idx` 7068 * 7069 * Return: 1 meaning mf should be freed from _base_interrupt 7070 * 0 means the mf is freed from this function. 7071 */ 7072 static u8 7073 _scsih_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply) 7074 { 7075 MPI2DefaultReply_t *mpi_reply; 7076 7077 mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply); 7078 if (ioc->scsih_cmds.status == MPT3_CMD_NOT_USED) 7079 return 1; 7080 if (ioc->scsih_cmds.smid != smid) 7081 return 1; 7082 ioc->scsih_cmds.status |= MPT3_CMD_COMPLETE; 7083 if (mpi_reply) { 7084 memcpy(ioc->scsih_cmds.reply, mpi_reply, 7085 mpi_reply->MsgLength*4); 7086 ioc->scsih_cmds.status |= MPT3_CMD_REPLY_VALID; 7087 } 7088 ioc->scsih_cmds.status &= ~MPT3_CMD_PENDING; 7089 complete(&ioc->scsih_cmds.done); 7090 return 1; 7091 } 7092 7093 7094 7095 7096 #define MPT3_MAX_LUNS (255) 7097 7098 7099 /** 7100 * _scsih_check_access_status - check access flags 7101 * @ioc: per adapter object 7102 * @sas_address: sas address 7103 * @handle: sas device handle 7104 * @access_status: errors returned during discovery of the device 7105 * 7106 * Return: 0 for success, else failure 7107 */ 7108 static u8 7109 _scsih_check_access_status(struct MPT3SAS_ADAPTER *ioc, u64 sas_address, 7110 u16 handle, u8 access_status) 7111 { 7112 u8 rc = 1; 7113 char *desc = NULL; 7114 7115 switch (access_status) { 7116 case MPI2_SAS_DEVICE0_ASTATUS_NO_ERRORS: 7117 case MPI2_SAS_DEVICE0_ASTATUS_SATA_NEEDS_INITIALIZATION: 7118 rc = 0; 7119 break; 7120 case MPI2_SAS_DEVICE0_ASTATUS_SATA_CAPABILITY_FAILED: 7121 desc = "sata capability failed"; 7122 break; 7123 case MPI2_SAS_DEVICE0_ASTATUS_SATA_AFFILIATION_CONFLICT: 7124 desc = "sata affiliation conflict"; 7125 break; 7126 case MPI2_SAS_DEVICE0_ASTATUS_ROUTE_NOT_ADDRESSABLE: 7127 desc = "route not addressable"; 7128 break; 7129 case MPI2_SAS_DEVICE0_ASTATUS_SMP_ERROR_NOT_ADDRESSABLE: 7130 desc = "smp error not addressable"; 7131 break; 7132 case MPI2_SAS_DEVICE0_ASTATUS_DEVICE_BLOCKED: 7133 desc = "device blocked"; 7134 break; 7135 case MPI2_SAS_DEVICE0_ASTATUS_SATA_INIT_FAILED: 7136 case MPI2_SAS_DEVICE0_ASTATUS_SIF_UNKNOWN: 7137 case MPI2_SAS_DEVICE0_ASTATUS_SIF_AFFILIATION_CONFLICT: 7138 case MPI2_SAS_DEVICE0_ASTATUS_SIF_DIAG: 7139 case MPI2_SAS_DEVICE0_ASTATUS_SIF_IDENTIFICATION: 7140 case MPI2_SAS_DEVICE0_ASTATUS_SIF_CHECK_POWER: 7141 case MPI2_SAS_DEVICE0_ASTATUS_SIF_PIO_SN: 7142 case MPI2_SAS_DEVICE0_ASTATUS_SIF_MDMA_SN: 7143 case MPI2_SAS_DEVICE0_ASTATUS_SIF_UDMA_SN: 7144 case MPI2_SAS_DEVICE0_ASTATUS_SIF_ZONING_VIOLATION: 7145 case MPI2_SAS_DEVICE0_ASTATUS_SIF_NOT_ADDRESSABLE: 7146 case MPI2_SAS_DEVICE0_ASTATUS_SIF_MAX: 7147 desc = "sata initialization failed"; 7148 break; 7149 default: 7150 desc = "unknown"; 7151 break; 7152 } 7153 7154 if (!rc) 7155 return 0; 7156 7157 ioc_err(ioc, "discovery errors(%s): sas_address(0x%016llx), handle(0x%04x)\n", 7158 desc, (u64)sas_address, handle); 7159 return rc; 7160 } 7161 7162 /** 7163 * _scsih_check_device - checking device responsiveness 7164 * @ioc: per adapter object 7165 * @parent_sas_address: sas address of parent expander or sas host 7166 * @handle: attached device handle 7167 * @phy_number: phy number 7168 * @link_rate: new link rate 7169 */ 7170 static void 7171 _scsih_check_device(struct MPT3SAS_ADAPTER *ioc, 7172 u64 parent_sas_address, u16 handle, u8 phy_number, u8 link_rate) 7173 { 7174 Mpi2ConfigReply_t mpi_reply; 7175 Mpi2SasDevicePage0_t sas_device_pg0; 7176 struct _sas_device *sas_device = NULL; 7177 struct _enclosure_node *enclosure_dev = NULL; 7178 u32 ioc_status; 7179 unsigned long flags; 7180 u64 sas_address; 7181 struct scsi_target *starget; 7182 struct MPT3SAS_TARGET *sas_target_priv_data; 7183 u32 device_info; 7184 struct hba_port *port; 7185 7186 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, 7187 MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) 7188 return; 7189 7190 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; 7191 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) 7192 return; 7193 7194 /* wide port handling ~ we need only handle device once for the phy that 7195 * is matched in sas device page zero 7196 */ 7197 if (phy_number != sas_device_pg0.PhyNum) 7198 return; 7199 7200 /* check if this is end device */ 7201 device_info = le32_to_cpu(sas_device_pg0.DeviceInfo); 7202 if (!(_scsih_is_end_device(device_info))) 7203 return; 7204 7205 spin_lock_irqsave(&ioc->sas_device_lock, flags); 7206 sas_address = le64_to_cpu(sas_device_pg0.SASAddress); 7207 port = mpt3sas_get_port_by_id(ioc, sas_device_pg0.PhysicalPort, 0); 7208 if (!port) 7209 goto out_unlock; 7210 sas_device = __mpt3sas_get_sdev_by_addr(ioc, 7211 sas_address, port); 7212 7213 if (!sas_device) 7214 goto out_unlock; 7215 7216 if (unlikely(sas_device->handle != handle)) { 7217 starget = sas_device->starget; 7218 sas_target_priv_data = starget->hostdata; 7219 starget_printk(KERN_INFO, starget, 7220 "handle changed from(0x%04x) to (0x%04x)!!!\n", 7221 sas_device->handle, handle); 7222 sas_target_priv_data->handle = handle; 7223 sas_device->handle = handle; 7224 if (le16_to_cpu(sas_device_pg0.Flags) & 7225 MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) { 7226 sas_device->enclosure_level = 7227 sas_device_pg0.EnclosureLevel; 7228 memcpy(sas_device->connector_name, 7229 sas_device_pg0.ConnectorName, 4); 7230 sas_device->connector_name[4] = '\0'; 7231 } else { 7232 sas_device->enclosure_level = 0; 7233 sas_device->connector_name[0] = '\0'; 7234 } 7235 7236 sas_device->enclosure_handle = 7237 le16_to_cpu(sas_device_pg0.EnclosureHandle); 7238 sas_device->is_chassis_slot_valid = 0; 7239 enclosure_dev = mpt3sas_scsih_enclosure_find_by_handle(ioc, 7240 sas_device->enclosure_handle); 7241 if (enclosure_dev) { 7242 sas_device->enclosure_logical_id = 7243 le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); 7244 if (le16_to_cpu(enclosure_dev->pg0.Flags) & 7245 MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) { 7246 sas_device->is_chassis_slot_valid = 1; 7247 sas_device->chassis_slot = 7248 enclosure_dev->pg0.ChassisSlot; 7249 } 7250 } 7251 } 7252 7253 /* check if device is present */ 7254 if (!(le16_to_cpu(sas_device_pg0.Flags) & 7255 MPI2_SAS_DEVICE0_FLAGS_DEVICE_PRESENT)) { 7256 ioc_err(ioc, "device is not present handle(0x%04x), flags!!!\n", 7257 handle); 7258 goto out_unlock; 7259 } 7260 7261 /* check if there were any issues with discovery */ 7262 if (_scsih_check_access_status(ioc, sas_address, handle, 7263 sas_device_pg0.AccessStatus)) 7264 goto out_unlock; 7265 7266 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 7267 _scsih_ublock_io_device(ioc, sas_address, port); 7268 7269 if (sas_device) 7270 sas_device_put(sas_device); 7271 return; 7272 7273 out_unlock: 7274 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 7275 if (sas_device) 7276 sas_device_put(sas_device); 7277 } 7278 7279 /** 7280 * _scsih_add_device - creating sas device object 7281 * @ioc: per adapter object 7282 * @handle: sas device handle 7283 * @phy_num: phy number end device attached to 7284 * @is_pd: is this hidden raid component 7285 * 7286 * Creating end device object, stored in ioc->sas_device_list. 7287 * 7288 * Return: 0 for success, non-zero for failure. 7289 */ 7290 static int 7291 _scsih_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle, u8 phy_num, 7292 u8 is_pd) 7293 { 7294 Mpi2ConfigReply_t mpi_reply; 7295 Mpi2SasDevicePage0_t sas_device_pg0; 7296 struct _sas_device *sas_device; 7297 struct _enclosure_node *enclosure_dev = NULL; 7298 u32 ioc_status; 7299 u64 sas_address; 7300 u32 device_info; 7301 u8 port_id; 7302 7303 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, 7304 MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { 7305 ioc_err(ioc, "failure at %s:%d/%s()!\n", 7306 __FILE__, __LINE__, __func__); 7307 return -1; 7308 } 7309 7310 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 7311 MPI2_IOCSTATUS_MASK; 7312 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 7313 ioc_err(ioc, "failure at %s:%d/%s()!\n", 7314 __FILE__, __LINE__, __func__); 7315 return -1; 7316 } 7317 7318 /* check if this is end device */ 7319 device_info = le32_to_cpu(sas_device_pg0.DeviceInfo); 7320 if (!(_scsih_is_end_device(device_info))) 7321 return -1; 7322 set_bit(handle, ioc->pend_os_device_add); 7323 sas_address = le64_to_cpu(sas_device_pg0.SASAddress); 7324 7325 /* check if device is present */ 7326 if (!(le16_to_cpu(sas_device_pg0.Flags) & 7327 MPI2_SAS_DEVICE0_FLAGS_DEVICE_PRESENT)) { 7328 ioc_err(ioc, "device is not present handle(0x04%x)!!!\n", 7329 handle); 7330 return -1; 7331 } 7332 7333 /* check if there were any issues with discovery */ 7334 if (_scsih_check_access_status(ioc, sas_address, handle, 7335 sas_device_pg0.AccessStatus)) 7336 return -1; 7337 7338 port_id = sas_device_pg0.PhysicalPort; 7339 sas_device = mpt3sas_get_sdev_by_addr(ioc, 7340 sas_address, mpt3sas_get_port_by_id(ioc, port_id, 0)); 7341 if (sas_device) { 7342 clear_bit(handle, ioc->pend_os_device_add); 7343 sas_device_put(sas_device); 7344 return -1; 7345 } 7346 7347 if (sas_device_pg0.EnclosureHandle) { 7348 enclosure_dev = 7349 mpt3sas_scsih_enclosure_find_by_handle(ioc, 7350 le16_to_cpu(sas_device_pg0.EnclosureHandle)); 7351 if (enclosure_dev == NULL) 7352 ioc_info(ioc, "Enclosure handle(0x%04x) doesn't match with enclosure device!\n", 7353 sas_device_pg0.EnclosureHandle); 7354 } 7355 7356 sas_device = kzalloc(sizeof(struct _sas_device), 7357 GFP_KERNEL); 7358 if (!sas_device) { 7359 ioc_err(ioc, "failure at %s:%d/%s()!\n", 7360 __FILE__, __LINE__, __func__); 7361 return 0; 7362 } 7363 7364 kref_init(&sas_device->refcount); 7365 sas_device->handle = handle; 7366 if (_scsih_get_sas_address(ioc, 7367 le16_to_cpu(sas_device_pg0.ParentDevHandle), 7368 &sas_device->sas_address_parent) != 0) 7369 ioc_err(ioc, "failure at %s:%d/%s()!\n", 7370 __FILE__, __LINE__, __func__); 7371 sas_device->enclosure_handle = 7372 le16_to_cpu(sas_device_pg0.EnclosureHandle); 7373 if (sas_device->enclosure_handle != 0) 7374 sas_device->slot = 7375 le16_to_cpu(sas_device_pg0.Slot); 7376 sas_device->device_info = device_info; 7377 sas_device->sas_address = sas_address; 7378 sas_device->phy = sas_device_pg0.PhyNum; 7379 sas_device->fast_path = (le16_to_cpu(sas_device_pg0.Flags) & 7380 MPI25_SAS_DEVICE0_FLAGS_FAST_PATH_CAPABLE) ? 1 : 0; 7381 sas_device->port = mpt3sas_get_port_by_id(ioc, port_id, 0); 7382 if (!sas_device->port) { 7383 ioc_err(ioc, "failure at %s:%d/%s()!\n", 7384 __FILE__, __LINE__, __func__); 7385 goto out; 7386 } 7387 7388 if (le16_to_cpu(sas_device_pg0.Flags) 7389 & MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) { 7390 sas_device->enclosure_level = 7391 sas_device_pg0.EnclosureLevel; 7392 memcpy(sas_device->connector_name, 7393 sas_device_pg0.ConnectorName, 4); 7394 sas_device->connector_name[4] = '\0'; 7395 } else { 7396 sas_device->enclosure_level = 0; 7397 sas_device->connector_name[0] = '\0'; 7398 } 7399 /* get enclosure_logical_id & chassis_slot*/ 7400 sas_device->is_chassis_slot_valid = 0; 7401 if (enclosure_dev) { 7402 sas_device->enclosure_logical_id = 7403 le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); 7404 if (le16_to_cpu(enclosure_dev->pg0.Flags) & 7405 MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) { 7406 sas_device->is_chassis_slot_valid = 1; 7407 sas_device->chassis_slot = 7408 enclosure_dev->pg0.ChassisSlot; 7409 } 7410 } 7411 7412 /* get device name */ 7413 sas_device->device_name = le64_to_cpu(sas_device_pg0.DeviceName); 7414 sas_device->port_type = sas_device_pg0.MaxPortConnections; 7415 ioc_info(ioc, 7416 "handle(0x%0x) sas_address(0x%016llx) port_type(0x%0x)\n", 7417 handle, sas_device->sas_address, sas_device->port_type); 7418 7419 if (ioc->wait_for_discovery_to_complete) 7420 _scsih_sas_device_init_add(ioc, sas_device); 7421 else 7422 _scsih_sas_device_add(ioc, sas_device); 7423 7424 out: 7425 sas_device_put(sas_device); 7426 return 0; 7427 } 7428 7429 /** 7430 * _scsih_remove_device - removing sas device object 7431 * @ioc: per adapter object 7432 * @sas_device: the sas_device object 7433 */ 7434 static void 7435 _scsih_remove_device(struct MPT3SAS_ADAPTER *ioc, 7436 struct _sas_device *sas_device) 7437 { 7438 struct MPT3SAS_TARGET *sas_target_priv_data; 7439 7440 if ((ioc->pdev->subsystem_vendor == PCI_VENDOR_ID_IBM) && 7441 (sas_device->pfa_led_on)) { 7442 _scsih_turn_off_pfa_led(ioc, sas_device); 7443 sas_device->pfa_led_on = 0; 7444 } 7445 7446 dewtprintk(ioc, 7447 ioc_info(ioc, "%s: enter: handle(0x%04x), sas_addr(0x%016llx)\n", 7448 __func__, 7449 sas_device->handle, (u64)sas_device->sas_address)); 7450 7451 dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device, 7452 NULL, NULL)); 7453 7454 if (sas_device->starget && sas_device->starget->hostdata) { 7455 sas_target_priv_data = sas_device->starget->hostdata; 7456 sas_target_priv_data->deleted = 1; 7457 _scsih_ublock_io_device(ioc, sas_device->sas_address, 7458 sas_device->port); 7459 sas_target_priv_data->handle = 7460 MPT3SAS_INVALID_DEVICE_HANDLE; 7461 } 7462 7463 if (!ioc->hide_drives) 7464 mpt3sas_transport_port_remove(ioc, 7465 sas_device->sas_address, 7466 sas_device->sas_address_parent, 7467 sas_device->port); 7468 7469 ioc_info(ioc, "removing handle(0x%04x), sas_addr(0x%016llx)\n", 7470 sas_device->handle, (u64)sas_device->sas_address); 7471 7472 _scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL); 7473 7474 dewtprintk(ioc, 7475 ioc_info(ioc, "%s: exit: handle(0x%04x), sas_addr(0x%016llx)\n", 7476 __func__, 7477 sas_device->handle, (u64)sas_device->sas_address)); 7478 dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device, 7479 NULL, NULL)); 7480 } 7481 7482 /** 7483 * _scsih_sas_topology_change_event_debug - debug for topology event 7484 * @ioc: per adapter object 7485 * @event_data: event data payload 7486 * Context: user. 7487 */ 7488 static void 7489 _scsih_sas_topology_change_event_debug(struct MPT3SAS_ADAPTER *ioc, 7490 Mpi2EventDataSasTopologyChangeList_t *event_data) 7491 { 7492 int i; 7493 u16 handle; 7494 u16 reason_code; 7495 u8 phy_number; 7496 char *status_str = NULL; 7497 u8 link_rate, prev_link_rate; 7498 7499 switch (event_data->ExpStatus) { 7500 case MPI2_EVENT_SAS_TOPO_ES_ADDED: 7501 status_str = "add"; 7502 break; 7503 case MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING: 7504 status_str = "remove"; 7505 break; 7506 case MPI2_EVENT_SAS_TOPO_ES_RESPONDING: 7507 case 0: 7508 status_str = "responding"; 7509 break; 7510 case MPI2_EVENT_SAS_TOPO_ES_DELAY_NOT_RESPONDING: 7511 status_str = "remove delay"; 7512 break; 7513 default: 7514 status_str = "unknown status"; 7515 break; 7516 } 7517 ioc_info(ioc, "sas topology change: (%s)\n", status_str); 7518 pr_info("\thandle(0x%04x), enclosure_handle(0x%04x) " \ 7519 "start_phy(%02d), count(%d)\n", 7520 le16_to_cpu(event_data->ExpanderDevHandle), 7521 le16_to_cpu(event_data->EnclosureHandle), 7522 event_data->StartPhyNum, event_data->NumEntries); 7523 for (i = 0; i < event_data->NumEntries; i++) { 7524 handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); 7525 if (!handle) 7526 continue; 7527 phy_number = event_data->StartPhyNum + i; 7528 reason_code = event_data->PHY[i].PhyStatus & 7529 MPI2_EVENT_SAS_TOPO_RC_MASK; 7530 switch (reason_code) { 7531 case MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED: 7532 status_str = "target add"; 7533 break; 7534 case MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING: 7535 status_str = "target remove"; 7536 break; 7537 case MPI2_EVENT_SAS_TOPO_RC_DELAY_NOT_RESPONDING: 7538 status_str = "delay target remove"; 7539 break; 7540 case MPI2_EVENT_SAS_TOPO_RC_PHY_CHANGED: 7541 status_str = "link rate change"; 7542 break; 7543 case MPI2_EVENT_SAS_TOPO_RC_NO_CHANGE: 7544 status_str = "target responding"; 7545 break; 7546 default: 7547 status_str = "unknown"; 7548 break; 7549 } 7550 link_rate = event_data->PHY[i].LinkRate >> 4; 7551 prev_link_rate = event_data->PHY[i].LinkRate & 0xF; 7552 pr_info("\tphy(%02d), attached_handle(0x%04x): %s:" \ 7553 " link rate: new(0x%02x), old(0x%02x)\n", phy_number, 7554 handle, status_str, link_rate, prev_link_rate); 7555 7556 } 7557 } 7558 7559 /** 7560 * _scsih_sas_topology_change_event - handle topology changes 7561 * @ioc: per adapter object 7562 * @fw_event: The fw_event_work object 7563 * Context: user. 7564 * 7565 */ 7566 static int 7567 _scsih_sas_topology_change_event(struct MPT3SAS_ADAPTER *ioc, 7568 struct fw_event_work *fw_event) 7569 { 7570 int i; 7571 u16 parent_handle, handle; 7572 u16 reason_code; 7573 u8 phy_number, max_phys; 7574 struct _sas_node *sas_expander; 7575 u64 sas_address; 7576 unsigned long flags; 7577 u8 link_rate, prev_link_rate; 7578 struct hba_port *port; 7579 Mpi2EventDataSasTopologyChangeList_t *event_data = 7580 (Mpi2EventDataSasTopologyChangeList_t *) 7581 fw_event->event_data; 7582 7583 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) 7584 _scsih_sas_topology_change_event_debug(ioc, event_data); 7585 7586 if (ioc->shost_recovery || ioc->remove_host || ioc->pci_error_recovery) 7587 return 0; 7588 7589 if (!ioc->sas_hba.num_phys) 7590 _scsih_sas_host_add(ioc); 7591 else 7592 _scsih_sas_host_refresh(ioc); 7593 7594 if (fw_event->ignore) { 7595 dewtprintk(ioc, ioc_info(ioc, "ignoring expander event\n")); 7596 return 0; 7597 } 7598 7599 parent_handle = le16_to_cpu(event_data->ExpanderDevHandle); 7600 port = mpt3sas_get_port_by_id(ioc, event_data->PhysicalPort, 0); 7601 7602 /* handle expander add */ 7603 if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_ADDED) 7604 if (_scsih_expander_add(ioc, parent_handle) != 0) 7605 return 0; 7606 7607 spin_lock_irqsave(&ioc->sas_node_lock, flags); 7608 sas_expander = mpt3sas_scsih_expander_find_by_handle(ioc, 7609 parent_handle); 7610 if (sas_expander) { 7611 sas_address = sas_expander->sas_address; 7612 max_phys = sas_expander->num_phys; 7613 port = sas_expander->port; 7614 } else if (parent_handle < ioc->sas_hba.num_phys) { 7615 sas_address = ioc->sas_hba.sas_address; 7616 max_phys = ioc->sas_hba.num_phys; 7617 } else { 7618 spin_unlock_irqrestore(&ioc->sas_node_lock, flags); 7619 return 0; 7620 } 7621 spin_unlock_irqrestore(&ioc->sas_node_lock, flags); 7622 7623 /* handle siblings events */ 7624 for (i = 0; i < event_data->NumEntries; i++) { 7625 if (fw_event->ignore) { 7626 dewtprintk(ioc, 7627 ioc_info(ioc, "ignoring expander event\n")); 7628 return 0; 7629 } 7630 if (ioc->remove_host || ioc->pci_error_recovery) 7631 return 0; 7632 phy_number = event_data->StartPhyNum + i; 7633 if (phy_number >= max_phys) 7634 continue; 7635 reason_code = event_data->PHY[i].PhyStatus & 7636 MPI2_EVENT_SAS_TOPO_RC_MASK; 7637 if ((event_data->PHY[i].PhyStatus & 7638 MPI2_EVENT_SAS_TOPO_PHYSTATUS_VACANT) && (reason_code != 7639 MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING)) 7640 continue; 7641 handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); 7642 if (!handle) 7643 continue; 7644 link_rate = event_data->PHY[i].LinkRate >> 4; 7645 prev_link_rate = event_data->PHY[i].LinkRate & 0xF; 7646 switch (reason_code) { 7647 case MPI2_EVENT_SAS_TOPO_RC_PHY_CHANGED: 7648 7649 if (ioc->shost_recovery) 7650 break; 7651 7652 if (link_rate == prev_link_rate) 7653 break; 7654 7655 mpt3sas_transport_update_links(ioc, sas_address, 7656 handle, phy_number, link_rate, port); 7657 7658 if (link_rate < MPI2_SAS_NEG_LINK_RATE_1_5) 7659 break; 7660 7661 _scsih_check_device(ioc, sas_address, handle, 7662 phy_number, link_rate); 7663 7664 if (!test_bit(handle, ioc->pend_os_device_add)) 7665 break; 7666 7667 fallthrough; 7668 7669 case MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED: 7670 7671 if (ioc->shost_recovery) 7672 break; 7673 7674 mpt3sas_transport_update_links(ioc, sas_address, 7675 handle, phy_number, link_rate, port); 7676 7677 _scsih_add_device(ioc, handle, phy_number, 0); 7678 7679 break; 7680 case MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING: 7681 7682 _scsih_device_remove_by_handle(ioc, handle); 7683 break; 7684 } 7685 } 7686 7687 /* handle expander removal */ 7688 if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING && 7689 sas_expander) 7690 mpt3sas_expander_remove(ioc, sas_address, port); 7691 7692 return 0; 7693 } 7694 7695 /** 7696 * _scsih_sas_device_status_change_event_debug - debug for device event 7697 * @ioc: ? 7698 * @event_data: event data payload 7699 * Context: user. 7700 */ 7701 static void 7702 _scsih_sas_device_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc, 7703 Mpi2EventDataSasDeviceStatusChange_t *event_data) 7704 { 7705 char *reason_str = NULL; 7706 7707 switch (event_data->ReasonCode) { 7708 case MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA: 7709 reason_str = "smart data"; 7710 break; 7711 case MPI2_EVENT_SAS_DEV_STAT_RC_UNSUPPORTED: 7712 reason_str = "unsupported device discovered"; 7713 break; 7714 case MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET: 7715 reason_str = "internal device reset"; 7716 break; 7717 case MPI2_EVENT_SAS_DEV_STAT_RC_TASK_ABORT_INTERNAL: 7718 reason_str = "internal task abort"; 7719 break; 7720 case MPI2_EVENT_SAS_DEV_STAT_RC_ABORT_TASK_SET_INTERNAL: 7721 reason_str = "internal task abort set"; 7722 break; 7723 case MPI2_EVENT_SAS_DEV_STAT_RC_CLEAR_TASK_SET_INTERNAL: 7724 reason_str = "internal clear task set"; 7725 break; 7726 case MPI2_EVENT_SAS_DEV_STAT_RC_QUERY_TASK_INTERNAL: 7727 reason_str = "internal query task"; 7728 break; 7729 case MPI2_EVENT_SAS_DEV_STAT_RC_SATA_INIT_FAILURE: 7730 reason_str = "sata init failure"; 7731 break; 7732 case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_INTERNAL_DEV_RESET: 7733 reason_str = "internal device reset complete"; 7734 break; 7735 case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_TASK_ABORT_INTERNAL: 7736 reason_str = "internal task abort complete"; 7737 break; 7738 case MPI2_EVENT_SAS_DEV_STAT_RC_ASYNC_NOTIFICATION: 7739 reason_str = "internal async notification"; 7740 break; 7741 case MPI2_EVENT_SAS_DEV_STAT_RC_EXPANDER_REDUCED_FUNCTIONALITY: 7742 reason_str = "expander reduced functionality"; 7743 break; 7744 case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_EXPANDER_REDUCED_FUNCTIONALITY: 7745 reason_str = "expander reduced functionality complete"; 7746 break; 7747 default: 7748 reason_str = "unknown reason"; 7749 break; 7750 } 7751 ioc_info(ioc, "device status change: (%s)\thandle(0x%04x), sas address(0x%016llx), tag(%d)", 7752 reason_str, le16_to_cpu(event_data->DevHandle), 7753 (u64)le64_to_cpu(event_data->SASAddress), 7754 le16_to_cpu(event_data->TaskTag)); 7755 if (event_data->ReasonCode == MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA) 7756 pr_cont(", ASC(0x%x), ASCQ(0x%x)\n", 7757 event_data->ASC, event_data->ASCQ); 7758 pr_cont("\n"); 7759 } 7760 7761 /** 7762 * _scsih_sas_device_status_change_event - handle device status change 7763 * @ioc: per adapter object 7764 * @event_data: The fw event 7765 * Context: user. 7766 */ 7767 static void 7768 _scsih_sas_device_status_change_event(struct MPT3SAS_ADAPTER *ioc, 7769 Mpi2EventDataSasDeviceStatusChange_t *event_data) 7770 { 7771 struct MPT3SAS_TARGET *target_priv_data; 7772 struct _sas_device *sas_device; 7773 u64 sas_address; 7774 unsigned long flags; 7775 7776 /* In MPI Revision K (0xC), the internal device reset complete was 7777 * implemented, so avoid setting tm_busy flag for older firmware. 7778 */ 7779 if ((ioc->facts.HeaderVersion >> 8) < 0xC) 7780 return; 7781 7782 if (event_data->ReasonCode != 7783 MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET && 7784 event_data->ReasonCode != 7785 MPI2_EVENT_SAS_DEV_STAT_RC_CMP_INTERNAL_DEV_RESET) 7786 return; 7787 7788 spin_lock_irqsave(&ioc->sas_device_lock, flags); 7789 sas_address = le64_to_cpu(event_data->SASAddress); 7790 sas_device = __mpt3sas_get_sdev_by_addr(ioc, 7791 sas_address, 7792 mpt3sas_get_port_by_id(ioc, event_data->PhysicalPort, 0)); 7793 7794 if (!sas_device || !sas_device->starget) 7795 goto out; 7796 7797 target_priv_data = sas_device->starget->hostdata; 7798 if (!target_priv_data) 7799 goto out; 7800 7801 if (event_data->ReasonCode == 7802 MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET) 7803 target_priv_data->tm_busy = 1; 7804 else 7805 target_priv_data->tm_busy = 0; 7806 7807 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) 7808 ioc_info(ioc, 7809 "%s tm_busy flag for handle(0x%04x)\n", 7810 (target_priv_data->tm_busy == 1) ? "Enable" : "Disable", 7811 target_priv_data->handle); 7812 7813 out: 7814 if (sas_device) 7815 sas_device_put(sas_device); 7816 7817 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 7818 } 7819 7820 7821 /** 7822 * _scsih_check_pcie_access_status - check access flags 7823 * @ioc: per adapter object 7824 * @wwid: wwid 7825 * @handle: sas device handle 7826 * @access_status: errors returned during discovery of the device 7827 * 7828 * Return: 0 for success, else failure 7829 */ 7830 static u8 7831 _scsih_check_pcie_access_status(struct MPT3SAS_ADAPTER *ioc, u64 wwid, 7832 u16 handle, u8 access_status) 7833 { 7834 u8 rc = 1; 7835 char *desc = NULL; 7836 7837 switch (access_status) { 7838 case MPI26_PCIEDEV0_ASTATUS_NO_ERRORS: 7839 case MPI26_PCIEDEV0_ASTATUS_NEEDS_INITIALIZATION: 7840 rc = 0; 7841 break; 7842 case MPI26_PCIEDEV0_ASTATUS_CAPABILITY_FAILED: 7843 desc = "PCIe device capability failed"; 7844 break; 7845 case MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED: 7846 desc = "PCIe device blocked"; 7847 ioc_info(ioc, 7848 "Device with Access Status (%s): wwid(0x%016llx), " 7849 "handle(0x%04x)\n ll only be added to the internal list", 7850 desc, (u64)wwid, handle); 7851 rc = 0; 7852 break; 7853 case MPI26_PCIEDEV0_ASTATUS_MEMORY_SPACE_ACCESS_FAILED: 7854 desc = "PCIe device mem space access failed"; 7855 break; 7856 case MPI26_PCIEDEV0_ASTATUS_UNSUPPORTED_DEVICE: 7857 desc = "PCIe device unsupported"; 7858 break; 7859 case MPI26_PCIEDEV0_ASTATUS_MSIX_REQUIRED: 7860 desc = "PCIe device MSIx Required"; 7861 break; 7862 case MPI26_PCIEDEV0_ASTATUS_INIT_FAIL_MAX: 7863 desc = "PCIe device init fail max"; 7864 break; 7865 case MPI26_PCIEDEV0_ASTATUS_UNKNOWN: 7866 desc = "PCIe device status unknown"; 7867 break; 7868 case MPI26_PCIEDEV0_ASTATUS_NVME_READY_TIMEOUT: 7869 desc = "nvme ready timeout"; 7870 break; 7871 case MPI26_PCIEDEV0_ASTATUS_NVME_DEVCFG_UNSUPPORTED: 7872 desc = "nvme device configuration unsupported"; 7873 break; 7874 case MPI26_PCIEDEV0_ASTATUS_NVME_IDENTIFY_FAILED: 7875 desc = "nvme identify failed"; 7876 break; 7877 case MPI26_PCIEDEV0_ASTATUS_NVME_QCONFIG_FAILED: 7878 desc = "nvme qconfig failed"; 7879 break; 7880 case MPI26_PCIEDEV0_ASTATUS_NVME_QCREATION_FAILED: 7881 desc = "nvme qcreation failed"; 7882 break; 7883 case MPI26_PCIEDEV0_ASTATUS_NVME_EVENTCFG_FAILED: 7884 desc = "nvme eventcfg failed"; 7885 break; 7886 case MPI26_PCIEDEV0_ASTATUS_NVME_GET_FEATURE_STAT_FAILED: 7887 desc = "nvme get feature stat failed"; 7888 break; 7889 case MPI26_PCIEDEV0_ASTATUS_NVME_IDLE_TIMEOUT: 7890 desc = "nvme idle timeout"; 7891 break; 7892 case MPI26_PCIEDEV0_ASTATUS_NVME_FAILURE_STATUS: 7893 desc = "nvme failure status"; 7894 break; 7895 default: 7896 ioc_err(ioc, "NVMe discovery error(0x%02x): wwid(0x%016llx), handle(0x%04x)\n", 7897 access_status, (u64)wwid, handle); 7898 return rc; 7899 } 7900 7901 if (!rc) 7902 return rc; 7903 7904 ioc_info(ioc, "NVMe discovery error(%s): wwid(0x%016llx), handle(0x%04x)\n", 7905 desc, (u64)wwid, handle); 7906 return rc; 7907 } 7908 7909 /** 7910 * _scsih_pcie_device_remove_from_sml - removing pcie device 7911 * from SML and free up associated memory 7912 * @ioc: per adapter object 7913 * @pcie_device: the pcie_device object 7914 */ 7915 static void 7916 _scsih_pcie_device_remove_from_sml(struct MPT3SAS_ADAPTER *ioc, 7917 struct _pcie_device *pcie_device) 7918 { 7919 struct MPT3SAS_TARGET *sas_target_priv_data; 7920 7921 dewtprintk(ioc, 7922 ioc_info(ioc, "%s: enter: handle(0x%04x), wwid(0x%016llx)\n", 7923 __func__, 7924 pcie_device->handle, (u64)pcie_device->wwid)); 7925 if (pcie_device->enclosure_handle != 0) 7926 dewtprintk(ioc, 7927 ioc_info(ioc, "%s: enter: enclosure logical id(0x%016llx), slot(%d)\n", 7928 __func__, 7929 (u64)pcie_device->enclosure_logical_id, 7930 pcie_device->slot)); 7931 if (pcie_device->connector_name[0] != '\0') 7932 dewtprintk(ioc, 7933 ioc_info(ioc, "%s: enter: enclosure level(0x%04x), connector name(%s)\n", 7934 __func__, 7935 pcie_device->enclosure_level, 7936 pcie_device->connector_name)); 7937 7938 if (pcie_device->starget && pcie_device->starget->hostdata) { 7939 sas_target_priv_data = pcie_device->starget->hostdata; 7940 sas_target_priv_data->deleted = 1; 7941 _scsih_ublock_io_device(ioc, pcie_device->wwid, NULL); 7942 sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE; 7943 } 7944 7945 ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n", 7946 pcie_device->handle, (u64)pcie_device->wwid); 7947 if (pcie_device->enclosure_handle != 0) 7948 ioc_info(ioc, "removing : enclosure logical id(0x%016llx), slot(%d)\n", 7949 (u64)pcie_device->enclosure_logical_id, 7950 pcie_device->slot); 7951 if (pcie_device->connector_name[0] != '\0') 7952 ioc_info(ioc, "removing: enclosure level(0x%04x), connector name( %s)\n", 7953 pcie_device->enclosure_level, 7954 pcie_device->connector_name); 7955 7956 if (pcie_device->starget && (pcie_device->access_status != 7957 MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED)) 7958 scsi_remove_target(&pcie_device->starget->dev); 7959 dewtprintk(ioc, 7960 ioc_info(ioc, "%s: exit: handle(0x%04x), wwid(0x%016llx)\n", 7961 __func__, 7962 pcie_device->handle, (u64)pcie_device->wwid)); 7963 if (pcie_device->enclosure_handle != 0) 7964 dewtprintk(ioc, 7965 ioc_info(ioc, "%s: exit: enclosure logical id(0x%016llx), slot(%d)\n", 7966 __func__, 7967 (u64)pcie_device->enclosure_logical_id, 7968 pcie_device->slot)); 7969 if (pcie_device->connector_name[0] != '\0') 7970 dewtprintk(ioc, 7971 ioc_info(ioc, "%s: exit: enclosure level(0x%04x), connector name( %s)\n", 7972 __func__, 7973 pcie_device->enclosure_level, 7974 pcie_device->connector_name)); 7975 7976 kfree(pcie_device->serial_number); 7977 } 7978 7979 7980 /** 7981 * _scsih_pcie_check_device - checking device responsiveness 7982 * @ioc: per adapter object 7983 * @handle: attached device handle 7984 */ 7985 static void 7986 _scsih_pcie_check_device(struct MPT3SAS_ADAPTER *ioc, u16 handle) 7987 { 7988 Mpi2ConfigReply_t mpi_reply; 7989 Mpi26PCIeDevicePage0_t pcie_device_pg0; 7990 u32 ioc_status; 7991 struct _pcie_device *pcie_device; 7992 u64 wwid; 7993 unsigned long flags; 7994 struct scsi_target *starget; 7995 struct MPT3SAS_TARGET *sas_target_priv_data; 7996 u32 device_info; 7997 7998 if ((mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply, 7999 &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_HANDLE, handle))) 8000 return; 8001 8002 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; 8003 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) 8004 return; 8005 8006 /* check if this is end device */ 8007 device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo); 8008 if (!(_scsih_is_nvme_pciescsi_device(device_info))) 8009 return; 8010 8011 wwid = le64_to_cpu(pcie_device_pg0.WWID); 8012 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 8013 pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid); 8014 8015 if (!pcie_device) { 8016 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 8017 return; 8018 } 8019 8020 if (unlikely(pcie_device->handle != handle)) { 8021 starget = pcie_device->starget; 8022 sas_target_priv_data = starget->hostdata; 8023 pcie_device->access_status = pcie_device_pg0.AccessStatus; 8024 starget_printk(KERN_INFO, starget, 8025 "handle changed from(0x%04x) to (0x%04x)!!!\n", 8026 pcie_device->handle, handle); 8027 sas_target_priv_data->handle = handle; 8028 pcie_device->handle = handle; 8029 8030 if (le32_to_cpu(pcie_device_pg0.Flags) & 8031 MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) { 8032 pcie_device->enclosure_level = 8033 pcie_device_pg0.EnclosureLevel; 8034 memcpy(&pcie_device->connector_name[0], 8035 &pcie_device_pg0.ConnectorName[0], 4); 8036 } else { 8037 pcie_device->enclosure_level = 0; 8038 pcie_device->connector_name[0] = '\0'; 8039 } 8040 } 8041 8042 /* check if device is present */ 8043 if (!(le32_to_cpu(pcie_device_pg0.Flags) & 8044 MPI26_PCIEDEV0_FLAGS_DEVICE_PRESENT)) { 8045 ioc_info(ioc, "device is not present handle(0x%04x), flags!!!\n", 8046 handle); 8047 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 8048 pcie_device_put(pcie_device); 8049 return; 8050 } 8051 8052 /* check if there were any issues with discovery */ 8053 if (_scsih_check_pcie_access_status(ioc, wwid, handle, 8054 pcie_device_pg0.AccessStatus)) { 8055 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 8056 pcie_device_put(pcie_device); 8057 return; 8058 } 8059 8060 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 8061 pcie_device_put(pcie_device); 8062 8063 _scsih_ublock_io_device(ioc, wwid, NULL); 8064 8065 return; 8066 } 8067 8068 /** 8069 * _scsih_pcie_add_device - creating pcie device object 8070 * @ioc: per adapter object 8071 * @handle: pcie device handle 8072 * 8073 * Creating end device object, stored in ioc->pcie_device_list. 8074 * 8075 * Return: 1 means queue the event later, 0 means complete the event 8076 */ 8077 static int 8078 _scsih_pcie_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle) 8079 { 8080 Mpi26PCIeDevicePage0_t pcie_device_pg0; 8081 Mpi26PCIeDevicePage2_t pcie_device_pg2; 8082 Mpi2ConfigReply_t mpi_reply; 8083 struct _pcie_device *pcie_device; 8084 struct _enclosure_node *enclosure_dev; 8085 u32 ioc_status; 8086 u64 wwid; 8087 8088 if ((mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply, 8089 &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_HANDLE, handle))) { 8090 ioc_err(ioc, "failure at %s:%d/%s()!\n", 8091 __FILE__, __LINE__, __func__); 8092 return 0; 8093 } 8094 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 8095 MPI2_IOCSTATUS_MASK; 8096 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 8097 ioc_err(ioc, "failure at %s:%d/%s()!\n", 8098 __FILE__, __LINE__, __func__); 8099 return 0; 8100 } 8101 8102 set_bit(handle, ioc->pend_os_device_add); 8103 wwid = le64_to_cpu(pcie_device_pg0.WWID); 8104 8105 /* check if device is present */ 8106 if (!(le32_to_cpu(pcie_device_pg0.Flags) & 8107 MPI26_PCIEDEV0_FLAGS_DEVICE_PRESENT)) { 8108 ioc_err(ioc, "device is not present handle(0x04%x)!!!\n", 8109 handle); 8110 return 0; 8111 } 8112 8113 /* check if there were any issues with discovery */ 8114 if (_scsih_check_pcie_access_status(ioc, wwid, handle, 8115 pcie_device_pg0.AccessStatus)) 8116 return 0; 8117 8118 if (!(_scsih_is_nvme_pciescsi_device(le32_to_cpu 8119 (pcie_device_pg0.DeviceInfo)))) 8120 return 0; 8121 8122 pcie_device = mpt3sas_get_pdev_by_wwid(ioc, wwid); 8123 if (pcie_device) { 8124 clear_bit(handle, ioc->pend_os_device_add); 8125 pcie_device_put(pcie_device); 8126 return 0; 8127 } 8128 8129 /* PCIe Device Page 2 contains read-only information about a 8130 * specific NVMe device; therefore, this page is only 8131 * valid for NVMe devices and skip for pcie devices of type scsi. 8132 */ 8133 if (!(mpt3sas_scsih_is_pcie_scsi_device( 8134 le32_to_cpu(pcie_device_pg0.DeviceInfo)))) { 8135 if (mpt3sas_config_get_pcie_device_pg2(ioc, &mpi_reply, 8136 &pcie_device_pg2, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, 8137 handle)) { 8138 ioc_err(ioc, 8139 "failure at %s:%d/%s()!\n", __FILE__, 8140 __LINE__, __func__); 8141 return 0; 8142 } 8143 8144 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 8145 MPI2_IOCSTATUS_MASK; 8146 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 8147 ioc_err(ioc, 8148 "failure at %s:%d/%s()!\n", __FILE__, 8149 __LINE__, __func__); 8150 return 0; 8151 } 8152 } 8153 8154 pcie_device = kzalloc(sizeof(struct _pcie_device), GFP_KERNEL); 8155 if (!pcie_device) { 8156 ioc_err(ioc, "failure at %s:%d/%s()!\n", 8157 __FILE__, __LINE__, __func__); 8158 return 0; 8159 } 8160 8161 kref_init(&pcie_device->refcount); 8162 pcie_device->id = ioc->pcie_target_id++; 8163 pcie_device->channel = PCIE_CHANNEL; 8164 pcie_device->handle = handle; 8165 pcie_device->access_status = pcie_device_pg0.AccessStatus; 8166 pcie_device->device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo); 8167 pcie_device->wwid = wwid; 8168 pcie_device->port_num = pcie_device_pg0.PortNum; 8169 pcie_device->fast_path = (le32_to_cpu(pcie_device_pg0.Flags) & 8170 MPI26_PCIEDEV0_FLAGS_FAST_PATH_CAPABLE) ? 1 : 0; 8171 8172 pcie_device->enclosure_handle = 8173 le16_to_cpu(pcie_device_pg0.EnclosureHandle); 8174 if (pcie_device->enclosure_handle != 0) 8175 pcie_device->slot = le16_to_cpu(pcie_device_pg0.Slot); 8176 8177 if (le32_to_cpu(pcie_device_pg0.Flags) & 8178 MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) { 8179 pcie_device->enclosure_level = pcie_device_pg0.EnclosureLevel; 8180 memcpy(&pcie_device->connector_name[0], 8181 &pcie_device_pg0.ConnectorName[0], 4); 8182 } else { 8183 pcie_device->enclosure_level = 0; 8184 pcie_device->connector_name[0] = '\0'; 8185 } 8186 8187 /* get enclosure_logical_id */ 8188 if (pcie_device->enclosure_handle) { 8189 enclosure_dev = 8190 mpt3sas_scsih_enclosure_find_by_handle(ioc, 8191 pcie_device->enclosure_handle); 8192 if (enclosure_dev) 8193 pcie_device->enclosure_logical_id = 8194 le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); 8195 } 8196 /* TODO -- Add device name once FW supports it */ 8197 if (!(mpt3sas_scsih_is_pcie_scsi_device( 8198 le32_to_cpu(pcie_device_pg0.DeviceInfo)))) { 8199 pcie_device->nvme_mdts = 8200 le32_to_cpu(pcie_device_pg2.MaximumDataTransferSize); 8201 pcie_device->shutdown_latency = 8202 le16_to_cpu(pcie_device_pg2.ShutdownLatency); 8203 /* 8204 * Set IOC's max_shutdown_latency to drive's RTD3 Entry Latency 8205 * if drive's RTD3 Entry Latency is greater then IOC's 8206 * max_shutdown_latency. 8207 */ 8208 if (pcie_device->shutdown_latency > ioc->max_shutdown_latency) 8209 ioc->max_shutdown_latency = 8210 pcie_device->shutdown_latency; 8211 if (pcie_device_pg2.ControllerResetTO) 8212 pcie_device->reset_timeout = 8213 pcie_device_pg2.ControllerResetTO; 8214 else 8215 pcie_device->reset_timeout = 30; 8216 } else 8217 pcie_device->reset_timeout = 30; 8218 8219 if (ioc->wait_for_discovery_to_complete) 8220 _scsih_pcie_device_init_add(ioc, pcie_device); 8221 else 8222 _scsih_pcie_device_add(ioc, pcie_device); 8223 8224 pcie_device_put(pcie_device); 8225 return 0; 8226 } 8227 8228 /** 8229 * _scsih_pcie_topology_change_event_debug - debug for topology 8230 * event 8231 * @ioc: per adapter object 8232 * @event_data: event data payload 8233 * Context: user. 8234 */ 8235 static void 8236 _scsih_pcie_topology_change_event_debug(struct MPT3SAS_ADAPTER *ioc, 8237 Mpi26EventDataPCIeTopologyChangeList_t *event_data) 8238 { 8239 int i; 8240 u16 handle; 8241 u16 reason_code; 8242 u8 port_number; 8243 char *status_str = NULL; 8244 u8 link_rate, prev_link_rate; 8245 8246 switch (event_data->SwitchStatus) { 8247 case MPI26_EVENT_PCIE_TOPO_SS_ADDED: 8248 status_str = "add"; 8249 break; 8250 case MPI26_EVENT_PCIE_TOPO_SS_NOT_RESPONDING: 8251 status_str = "remove"; 8252 break; 8253 case MPI26_EVENT_PCIE_TOPO_SS_RESPONDING: 8254 case 0: 8255 status_str = "responding"; 8256 break; 8257 case MPI26_EVENT_PCIE_TOPO_SS_DELAY_NOT_RESPONDING: 8258 status_str = "remove delay"; 8259 break; 8260 default: 8261 status_str = "unknown status"; 8262 break; 8263 } 8264 ioc_info(ioc, "pcie topology change: (%s)\n", status_str); 8265 pr_info("\tswitch_handle(0x%04x), enclosure_handle(0x%04x)" 8266 "start_port(%02d), count(%d)\n", 8267 le16_to_cpu(event_data->SwitchDevHandle), 8268 le16_to_cpu(event_data->EnclosureHandle), 8269 event_data->StartPortNum, event_data->NumEntries); 8270 for (i = 0; i < event_data->NumEntries; i++) { 8271 handle = 8272 le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle); 8273 if (!handle) 8274 continue; 8275 port_number = event_data->StartPortNum + i; 8276 reason_code = event_data->PortEntry[i].PortStatus; 8277 switch (reason_code) { 8278 case MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED: 8279 status_str = "target add"; 8280 break; 8281 case MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING: 8282 status_str = "target remove"; 8283 break; 8284 case MPI26_EVENT_PCIE_TOPO_PS_DELAY_NOT_RESPONDING: 8285 status_str = "delay target remove"; 8286 break; 8287 case MPI26_EVENT_PCIE_TOPO_PS_PORT_CHANGED: 8288 status_str = "link rate change"; 8289 break; 8290 case MPI26_EVENT_PCIE_TOPO_PS_NO_CHANGE: 8291 status_str = "target responding"; 8292 break; 8293 default: 8294 status_str = "unknown"; 8295 break; 8296 } 8297 link_rate = event_data->PortEntry[i].CurrentPortInfo & 8298 MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK; 8299 prev_link_rate = event_data->PortEntry[i].PreviousPortInfo & 8300 MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK; 8301 pr_info("\tport(%02d), attached_handle(0x%04x): %s:" 8302 " link rate: new(0x%02x), old(0x%02x)\n", port_number, 8303 handle, status_str, link_rate, prev_link_rate); 8304 } 8305 } 8306 8307 /** 8308 * _scsih_pcie_topology_change_event - handle PCIe topology 8309 * changes 8310 * @ioc: per adapter object 8311 * @fw_event: The fw_event_work object 8312 * Context: user. 8313 * 8314 */ 8315 static void 8316 _scsih_pcie_topology_change_event(struct MPT3SAS_ADAPTER *ioc, 8317 struct fw_event_work *fw_event) 8318 { 8319 int i; 8320 u16 handle; 8321 u16 reason_code; 8322 u8 link_rate, prev_link_rate; 8323 unsigned long flags; 8324 int rc; 8325 Mpi26EventDataPCIeTopologyChangeList_t *event_data = 8326 (Mpi26EventDataPCIeTopologyChangeList_t *) fw_event->event_data; 8327 struct _pcie_device *pcie_device; 8328 8329 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) 8330 _scsih_pcie_topology_change_event_debug(ioc, event_data); 8331 8332 if (ioc->shost_recovery || ioc->remove_host || 8333 ioc->pci_error_recovery) 8334 return; 8335 8336 if (fw_event->ignore) { 8337 dewtprintk(ioc, ioc_info(ioc, "ignoring switch event\n")); 8338 return; 8339 } 8340 8341 /* handle siblings events */ 8342 for (i = 0; i < event_data->NumEntries; i++) { 8343 if (fw_event->ignore) { 8344 dewtprintk(ioc, 8345 ioc_info(ioc, "ignoring switch event\n")); 8346 return; 8347 } 8348 if (ioc->remove_host || ioc->pci_error_recovery) 8349 return; 8350 reason_code = event_data->PortEntry[i].PortStatus; 8351 handle = 8352 le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle); 8353 if (!handle) 8354 continue; 8355 8356 link_rate = event_data->PortEntry[i].CurrentPortInfo 8357 & MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK; 8358 prev_link_rate = event_data->PortEntry[i].PreviousPortInfo 8359 & MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK; 8360 8361 switch (reason_code) { 8362 case MPI26_EVENT_PCIE_TOPO_PS_PORT_CHANGED: 8363 if (ioc->shost_recovery) 8364 break; 8365 if (link_rate == prev_link_rate) 8366 break; 8367 if (link_rate < MPI26_EVENT_PCIE_TOPO_PI_RATE_2_5) 8368 break; 8369 8370 _scsih_pcie_check_device(ioc, handle); 8371 8372 /* This code after this point handles the test case 8373 * where a device has been added, however its returning 8374 * BUSY for sometime. Then before the Device Missing 8375 * Delay expires and the device becomes READY, the 8376 * device is removed and added back. 8377 */ 8378 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 8379 pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle); 8380 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 8381 8382 if (pcie_device) { 8383 pcie_device_put(pcie_device); 8384 break; 8385 } 8386 8387 if (!test_bit(handle, ioc->pend_os_device_add)) 8388 break; 8389 8390 dewtprintk(ioc, 8391 ioc_info(ioc, "handle(0x%04x) device not found: convert event to a device add\n", 8392 handle)); 8393 event_data->PortEntry[i].PortStatus &= 0xF0; 8394 event_data->PortEntry[i].PortStatus |= 8395 MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED; 8396 fallthrough; 8397 case MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED: 8398 if (ioc->shost_recovery) 8399 break; 8400 if (link_rate < MPI26_EVENT_PCIE_TOPO_PI_RATE_2_5) 8401 break; 8402 8403 rc = _scsih_pcie_add_device(ioc, handle); 8404 if (!rc) { 8405 /* mark entry vacant */ 8406 /* TODO This needs to be reviewed and fixed, 8407 * we dont have an entry 8408 * to make an event void like vacant 8409 */ 8410 event_data->PortEntry[i].PortStatus |= 8411 MPI26_EVENT_PCIE_TOPO_PS_NO_CHANGE; 8412 } 8413 break; 8414 case MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING: 8415 _scsih_pcie_device_remove_by_handle(ioc, handle); 8416 break; 8417 } 8418 } 8419 } 8420 8421 /** 8422 * _scsih_pcie_device_status_change_event_debug - debug for device event 8423 * @ioc: ? 8424 * @event_data: event data payload 8425 * Context: user. 8426 */ 8427 static void 8428 _scsih_pcie_device_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc, 8429 Mpi26EventDataPCIeDeviceStatusChange_t *event_data) 8430 { 8431 char *reason_str = NULL; 8432 8433 switch (event_data->ReasonCode) { 8434 case MPI26_EVENT_PCIDEV_STAT_RC_SMART_DATA: 8435 reason_str = "smart data"; 8436 break; 8437 case MPI26_EVENT_PCIDEV_STAT_RC_UNSUPPORTED: 8438 reason_str = "unsupported device discovered"; 8439 break; 8440 case MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET: 8441 reason_str = "internal device reset"; 8442 break; 8443 case MPI26_EVENT_PCIDEV_STAT_RC_TASK_ABORT_INTERNAL: 8444 reason_str = "internal task abort"; 8445 break; 8446 case MPI26_EVENT_PCIDEV_STAT_RC_ABORT_TASK_SET_INTERNAL: 8447 reason_str = "internal task abort set"; 8448 break; 8449 case MPI26_EVENT_PCIDEV_STAT_RC_CLEAR_TASK_SET_INTERNAL: 8450 reason_str = "internal clear task set"; 8451 break; 8452 case MPI26_EVENT_PCIDEV_STAT_RC_QUERY_TASK_INTERNAL: 8453 reason_str = "internal query task"; 8454 break; 8455 case MPI26_EVENT_PCIDEV_STAT_RC_DEV_INIT_FAILURE: 8456 reason_str = "device init failure"; 8457 break; 8458 case MPI26_EVENT_PCIDEV_STAT_RC_CMP_INTERNAL_DEV_RESET: 8459 reason_str = "internal device reset complete"; 8460 break; 8461 case MPI26_EVENT_PCIDEV_STAT_RC_CMP_TASK_ABORT_INTERNAL: 8462 reason_str = "internal task abort complete"; 8463 break; 8464 case MPI26_EVENT_PCIDEV_STAT_RC_ASYNC_NOTIFICATION: 8465 reason_str = "internal async notification"; 8466 break; 8467 case MPI26_EVENT_PCIDEV_STAT_RC_PCIE_HOT_RESET_FAILED: 8468 reason_str = "pcie hot reset failed"; 8469 break; 8470 default: 8471 reason_str = "unknown reason"; 8472 break; 8473 } 8474 8475 ioc_info(ioc, "PCIE device status change: (%s)\n" 8476 "\thandle(0x%04x), WWID(0x%016llx), tag(%d)", 8477 reason_str, le16_to_cpu(event_data->DevHandle), 8478 (u64)le64_to_cpu(event_data->WWID), 8479 le16_to_cpu(event_data->TaskTag)); 8480 if (event_data->ReasonCode == MPI26_EVENT_PCIDEV_STAT_RC_SMART_DATA) 8481 pr_cont(", ASC(0x%x), ASCQ(0x%x)\n", 8482 event_data->ASC, event_data->ASCQ); 8483 pr_cont("\n"); 8484 } 8485 8486 /** 8487 * _scsih_pcie_device_status_change_event - handle device status 8488 * change 8489 * @ioc: per adapter object 8490 * @fw_event: The fw_event_work object 8491 * Context: user. 8492 */ 8493 static void 8494 _scsih_pcie_device_status_change_event(struct MPT3SAS_ADAPTER *ioc, 8495 struct fw_event_work *fw_event) 8496 { 8497 struct MPT3SAS_TARGET *target_priv_data; 8498 struct _pcie_device *pcie_device; 8499 u64 wwid; 8500 unsigned long flags; 8501 Mpi26EventDataPCIeDeviceStatusChange_t *event_data = 8502 (Mpi26EventDataPCIeDeviceStatusChange_t *)fw_event->event_data; 8503 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) 8504 _scsih_pcie_device_status_change_event_debug(ioc, 8505 event_data); 8506 8507 if (event_data->ReasonCode != 8508 MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET && 8509 event_data->ReasonCode != 8510 MPI26_EVENT_PCIDEV_STAT_RC_CMP_INTERNAL_DEV_RESET) 8511 return; 8512 8513 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 8514 wwid = le64_to_cpu(event_data->WWID); 8515 pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid); 8516 8517 if (!pcie_device || !pcie_device->starget) 8518 goto out; 8519 8520 target_priv_data = pcie_device->starget->hostdata; 8521 if (!target_priv_data) 8522 goto out; 8523 8524 if (event_data->ReasonCode == 8525 MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET) 8526 target_priv_data->tm_busy = 1; 8527 else 8528 target_priv_data->tm_busy = 0; 8529 out: 8530 if (pcie_device) 8531 pcie_device_put(pcie_device); 8532 8533 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 8534 } 8535 8536 /** 8537 * _scsih_sas_enclosure_dev_status_change_event_debug - debug for enclosure 8538 * event 8539 * @ioc: per adapter object 8540 * @event_data: event data payload 8541 * Context: user. 8542 */ 8543 static void 8544 _scsih_sas_enclosure_dev_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc, 8545 Mpi2EventDataSasEnclDevStatusChange_t *event_data) 8546 { 8547 char *reason_str = NULL; 8548 8549 switch (event_data->ReasonCode) { 8550 case MPI2_EVENT_SAS_ENCL_RC_ADDED: 8551 reason_str = "enclosure add"; 8552 break; 8553 case MPI2_EVENT_SAS_ENCL_RC_NOT_RESPONDING: 8554 reason_str = "enclosure remove"; 8555 break; 8556 default: 8557 reason_str = "unknown reason"; 8558 break; 8559 } 8560 8561 ioc_info(ioc, "enclosure status change: (%s)\n" 8562 "\thandle(0x%04x), enclosure logical id(0x%016llx) number slots(%d)\n", 8563 reason_str, 8564 le16_to_cpu(event_data->EnclosureHandle), 8565 (u64)le64_to_cpu(event_data->EnclosureLogicalID), 8566 le16_to_cpu(event_data->StartSlot)); 8567 } 8568 8569 /** 8570 * _scsih_sas_enclosure_dev_status_change_event - handle enclosure events 8571 * @ioc: per adapter object 8572 * @fw_event: The fw_event_work object 8573 * Context: user. 8574 */ 8575 static void 8576 _scsih_sas_enclosure_dev_status_change_event(struct MPT3SAS_ADAPTER *ioc, 8577 struct fw_event_work *fw_event) 8578 { 8579 Mpi2ConfigReply_t mpi_reply; 8580 struct _enclosure_node *enclosure_dev = NULL; 8581 Mpi2EventDataSasEnclDevStatusChange_t *event_data = 8582 (Mpi2EventDataSasEnclDevStatusChange_t *)fw_event->event_data; 8583 int rc; 8584 u16 enclosure_handle = le16_to_cpu(event_data->EnclosureHandle); 8585 8586 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) 8587 _scsih_sas_enclosure_dev_status_change_event_debug(ioc, 8588 (Mpi2EventDataSasEnclDevStatusChange_t *) 8589 fw_event->event_data); 8590 if (ioc->shost_recovery) 8591 return; 8592 8593 if (enclosure_handle) 8594 enclosure_dev = 8595 mpt3sas_scsih_enclosure_find_by_handle(ioc, 8596 enclosure_handle); 8597 switch (event_data->ReasonCode) { 8598 case MPI2_EVENT_SAS_ENCL_RC_ADDED: 8599 if (!enclosure_dev) { 8600 enclosure_dev = 8601 kzalloc(sizeof(struct _enclosure_node), 8602 GFP_KERNEL); 8603 if (!enclosure_dev) { 8604 ioc_info(ioc, "failure at %s:%d/%s()!\n", 8605 __FILE__, __LINE__, __func__); 8606 return; 8607 } 8608 rc = mpt3sas_config_get_enclosure_pg0(ioc, &mpi_reply, 8609 &enclosure_dev->pg0, 8610 MPI2_SAS_ENCLOS_PGAD_FORM_HANDLE, 8611 enclosure_handle); 8612 8613 if (rc || (le16_to_cpu(mpi_reply.IOCStatus) & 8614 MPI2_IOCSTATUS_MASK)) { 8615 kfree(enclosure_dev); 8616 return; 8617 } 8618 8619 list_add_tail(&enclosure_dev->list, 8620 &ioc->enclosure_list); 8621 } 8622 break; 8623 case MPI2_EVENT_SAS_ENCL_RC_NOT_RESPONDING: 8624 if (enclosure_dev) { 8625 list_del(&enclosure_dev->list); 8626 kfree(enclosure_dev); 8627 } 8628 break; 8629 default: 8630 break; 8631 } 8632 } 8633 8634 /** 8635 * _scsih_sas_broadcast_primitive_event - handle broadcast events 8636 * @ioc: per adapter object 8637 * @fw_event: The fw_event_work object 8638 * Context: user. 8639 */ 8640 static void 8641 _scsih_sas_broadcast_primitive_event(struct MPT3SAS_ADAPTER *ioc, 8642 struct fw_event_work *fw_event) 8643 { 8644 struct scsi_cmnd *scmd; 8645 struct scsi_device *sdev; 8646 struct scsiio_tracker *st; 8647 u16 smid, handle; 8648 u32 lun; 8649 struct MPT3SAS_DEVICE *sas_device_priv_data; 8650 u32 termination_count; 8651 u32 query_count; 8652 Mpi2SCSITaskManagementReply_t *mpi_reply; 8653 Mpi2EventDataSasBroadcastPrimitive_t *event_data = 8654 (Mpi2EventDataSasBroadcastPrimitive_t *) 8655 fw_event->event_data; 8656 u16 ioc_status; 8657 unsigned long flags; 8658 int r; 8659 u8 max_retries = 0; 8660 u8 task_abort_retries; 8661 8662 mutex_lock(&ioc->tm_cmds.mutex); 8663 ioc_info(ioc, "%s: enter: phy number(%d), width(%d)\n", 8664 __func__, event_data->PhyNum, event_data->PortWidth); 8665 8666 _scsih_block_io_all_device(ioc); 8667 8668 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); 8669 mpi_reply = ioc->tm_cmds.reply; 8670 broadcast_aen_retry: 8671 8672 /* sanity checks for retrying this loop */ 8673 if (max_retries++ == 5) { 8674 dewtprintk(ioc, ioc_info(ioc, "%s: giving up\n", __func__)); 8675 goto out; 8676 } else if (max_retries > 1) 8677 dewtprintk(ioc, 8678 ioc_info(ioc, "%s: %d retry\n", 8679 __func__, max_retries - 1)); 8680 8681 termination_count = 0; 8682 query_count = 0; 8683 for (smid = 1; smid <= ioc->scsiio_depth; smid++) { 8684 if (ioc->shost_recovery) 8685 goto out; 8686 scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); 8687 if (!scmd) 8688 continue; 8689 st = scsi_cmd_priv(scmd); 8690 sdev = scmd->device; 8691 sas_device_priv_data = sdev->hostdata; 8692 if (!sas_device_priv_data || !sas_device_priv_data->sas_target) 8693 continue; 8694 /* skip hidden raid components */ 8695 if (sas_device_priv_data->sas_target->flags & 8696 MPT_TARGET_FLAGS_RAID_COMPONENT) 8697 continue; 8698 /* skip volumes */ 8699 if (sas_device_priv_data->sas_target->flags & 8700 MPT_TARGET_FLAGS_VOLUME) 8701 continue; 8702 /* skip PCIe devices */ 8703 if (sas_device_priv_data->sas_target->flags & 8704 MPT_TARGET_FLAGS_PCIE_DEVICE) 8705 continue; 8706 8707 handle = sas_device_priv_data->sas_target->handle; 8708 lun = sas_device_priv_data->lun; 8709 query_count++; 8710 8711 if (ioc->shost_recovery) 8712 goto out; 8713 8714 spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags); 8715 r = mpt3sas_scsih_issue_tm(ioc, handle, 0, 0, lun, 8716 MPI2_SCSITASKMGMT_TASKTYPE_QUERY_TASK, st->smid, 8717 st->msix_io, 30, 0); 8718 if (r == FAILED) { 8719 sdev_printk(KERN_WARNING, sdev, 8720 "mpt3sas_scsih_issue_tm: FAILED when sending " 8721 "QUERY_TASK: scmd(%p)\n", scmd); 8722 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); 8723 goto broadcast_aen_retry; 8724 } 8725 ioc_status = le16_to_cpu(mpi_reply->IOCStatus) 8726 & MPI2_IOCSTATUS_MASK; 8727 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 8728 sdev_printk(KERN_WARNING, sdev, 8729 "query task: FAILED with IOCSTATUS(0x%04x), scmd(%p)\n", 8730 ioc_status, scmd); 8731 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); 8732 goto broadcast_aen_retry; 8733 } 8734 8735 /* see if IO is still owned by IOC and target */ 8736 if (mpi_reply->ResponseCode == 8737 MPI2_SCSITASKMGMT_RSP_TM_SUCCEEDED || 8738 mpi_reply->ResponseCode == 8739 MPI2_SCSITASKMGMT_RSP_IO_QUEUED_ON_IOC) { 8740 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); 8741 continue; 8742 } 8743 task_abort_retries = 0; 8744 tm_retry: 8745 if (task_abort_retries++ == 60) { 8746 dewtprintk(ioc, 8747 ioc_info(ioc, "%s: ABORT_TASK: giving up\n", 8748 __func__)); 8749 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); 8750 goto broadcast_aen_retry; 8751 } 8752 8753 if (ioc->shost_recovery) 8754 goto out_no_lock; 8755 8756 r = mpt3sas_scsih_issue_tm(ioc, handle, sdev->channel, sdev->id, 8757 sdev->lun, MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK, 8758 st->smid, st->msix_io, 30, 0); 8759 if (r == FAILED || st->cb_idx != 0xFF) { 8760 sdev_printk(KERN_WARNING, sdev, 8761 "mpt3sas_scsih_issue_tm: ABORT_TASK: FAILED : " 8762 "scmd(%p)\n", scmd); 8763 goto tm_retry; 8764 } 8765 8766 if (task_abort_retries > 1) 8767 sdev_printk(KERN_WARNING, sdev, 8768 "mpt3sas_scsih_issue_tm: ABORT_TASK: RETRIES (%d):" 8769 " scmd(%p)\n", 8770 task_abort_retries - 1, scmd); 8771 8772 termination_count += le32_to_cpu(mpi_reply->TerminationCount); 8773 spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); 8774 } 8775 8776 if (ioc->broadcast_aen_pending) { 8777 dewtprintk(ioc, 8778 ioc_info(ioc, 8779 "%s: loop back due to pending AEN\n", 8780 __func__)); 8781 ioc->broadcast_aen_pending = 0; 8782 goto broadcast_aen_retry; 8783 } 8784 8785 out: 8786 spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags); 8787 out_no_lock: 8788 8789 dewtprintk(ioc, 8790 ioc_info(ioc, "%s - exit, query_count = %d termination_count = %d\n", 8791 __func__, query_count, termination_count)); 8792 8793 ioc->broadcast_aen_busy = 0; 8794 if (!ioc->shost_recovery) 8795 _scsih_ublock_io_all_device(ioc); 8796 mutex_unlock(&ioc->tm_cmds.mutex); 8797 } 8798 8799 /** 8800 * _scsih_sas_discovery_event - handle discovery events 8801 * @ioc: per adapter object 8802 * @fw_event: The fw_event_work object 8803 * Context: user. 8804 */ 8805 static void 8806 _scsih_sas_discovery_event(struct MPT3SAS_ADAPTER *ioc, 8807 struct fw_event_work *fw_event) 8808 { 8809 Mpi2EventDataSasDiscovery_t *event_data = 8810 (Mpi2EventDataSasDiscovery_t *) fw_event->event_data; 8811 8812 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) { 8813 ioc_info(ioc, "discovery event: (%s)", 8814 event_data->ReasonCode == MPI2_EVENT_SAS_DISC_RC_STARTED ? 8815 "start" : "stop"); 8816 if (event_data->DiscoveryStatus) 8817 pr_cont("discovery_status(0x%08x)", 8818 le32_to_cpu(event_data->DiscoveryStatus)); 8819 pr_cont("\n"); 8820 } 8821 8822 if (event_data->ReasonCode == MPI2_EVENT_SAS_DISC_RC_STARTED && 8823 !ioc->sas_hba.num_phys) { 8824 if (disable_discovery > 0 && ioc->shost_recovery) { 8825 /* Wait for the reset to complete */ 8826 while (ioc->shost_recovery) 8827 ssleep(1); 8828 } 8829 _scsih_sas_host_add(ioc); 8830 } 8831 } 8832 8833 /** 8834 * _scsih_sas_device_discovery_error_event - display SAS device discovery error 8835 * events 8836 * @ioc: per adapter object 8837 * @fw_event: The fw_event_work object 8838 * Context: user. 8839 */ 8840 static void 8841 _scsih_sas_device_discovery_error_event(struct MPT3SAS_ADAPTER *ioc, 8842 struct fw_event_work *fw_event) 8843 { 8844 Mpi25EventDataSasDeviceDiscoveryError_t *event_data = 8845 (Mpi25EventDataSasDeviceDiscoveryError_t *)fw_event->event_data; 8846 8847 switch (event_data->ReasonCode) { 8848 case MPI25_EVENT_SAS_DISC_ERR_SMP_FAILED: 8849 ioc_warn(ioc, "SMP command sent to the expander (handle:0x%04x, sas_address:0x%016llx, physical_port:0x%02x) has failed\n", 8850 le16_to_cpu(event_data->DevHandle), 8851 (u64)le64_to_cpu(event_data->SASAddress), 8852 event_data->PhysicalPort); 8853 break; 8854 case MPI25_EVENT_SAS_DISC_ERR_SMP_TIMEOUT: 8855 ioc_warn(ioc, "SMP command sent to the expander (handle:0x%04x, sas_address:0x%016llx, physical_port:0x%02x) has timed out\n", 8856 le16_to_cpu(event_data->DevHandle), 8857 (u64)le64_to_cpu(event_data->SASAddress), 8858 event_data->PhysicalPort); 8859 break; 8860 default: 8861 break; 8862 } 8863 } 8864 8865 /** 8866 * _scsih_pcie_enumeration_event - handle enumeration events 8867 * @ioc: per adapter object 8868 * @fw_event: The fw_event_work object 8869 * Context: user. 8870 */ 8871 static void 8872 _scsih_pcie_enumeration_event(struct MPT3SAS_ADAPTER *ioc, 8873 struct fw_event_work *fw_event) 8874 { 8875 Mpi26EventDataPCIeEnumeration_t *event_data = 8876 (Mpi26EventDataPCIeEnumeration_t *)fw_event->event_data; 8877 8878 if (!(ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)) 8879 return; 8880 8881 ioc_info(ioc, "pcie enumeration event: (%s) Flag 0x%02x", 8882 (event_data->ReasonCode == MPI26_EVENT_PCIE_ENUM_RC_STARTED) ? 8883 "started" : "completed", 8884 event_data->Flags); 8885 if (event_data->EnumerationStatus) 8886 pr_cont("enumeration_status(0x%08x)", 8887 le32_to_cpu(event_data->EnumerationStatus)); 8888 pr_cont("\n"); 8889 } 8890 8891 /** 8892 * _scsih_ir_fastpath - turn on fastpath for IR physdisk 8893 * @ioc: per adapter object 8894 * @handle: device handle for physical disk 8895 * @phys_disk_num: physical disk number 8896 * 8897 * Return: 0 for success, else failure. 8898 */ 8899 static int 8900 _scsih_ir_fastpath(struct MPT3SAS_ADAPTER *ioc, u16 handle, u8 phys_disk_num) 8901 { 8902 Mpi2RaidActionRequest_t *mpi_request; 8903 Mpi2RaidActionReply_t *mpi_reply; 8904 u16 smid; 8905 u8 issue_reset = 0; 8906 int rc = 0; 8907 u16 ioc_status; 8908 u32 log_info; 8909 8910 if (ioc->hba_mpi_version_belonged == MPI2_VERSION) 8911 return rc; 8912 8913 mutex_lock(&ioc->scsih_cmds.mutex); 8914 8915 if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) { 8916 ioc_err(ioc, "%s: scsih_cmd in use\n", __func__); 8917 rc = -EAGAIN; 8918 goto out; 8919 } 8920 ioc->scsih_cmds.status = MPT3_CMD_PENDING; 8921 8922 smid = mpt3sas_base_get_smid(ioc, ioc->scsih_cb_idx); 8923 if (!smid) { 8924 ioc_err(ioc, "%s: failed obtaining a smid\n", __func__); 8925 ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; 8926 rc = -EAGAIN; 8927 goto out; 8928 } 8929 8930 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); 8931 ioc->scsih_cmds.smid = smid; 8932 memset(mpi_request, 0, sizeof(Mpi2RaidActionRequest_t)); 8933 8934 mpi_request->Function = MPI2_FUNCTION_RAID_ACTION; 8935 mpi_request->Action = MPI2_RAID_ACTION_PHYSDISK_HIDDEN; 8936 mpi_request->PhysDiskNum = phys_disk_num; 8937 8938 dewtprintk(ioc, 8939 ioc_info(ioc, "IR RAID_ACTION: turning fast path on for handle(0x%04x), phys_disk_num (0x%02x)\n", 8940 handle, phys_disk_num)); 8941 8942 init_completion(&ioc->scsih_cmds.done); 8943 ioc->put_smid_default(ioc, smid); 8944 wait_for_completion_timeout(&ioc->scsih_cmds.done, 10*HZ); 8945 8946 if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) { 8947 mpt3sas_check_cmd_timeout(ioc, 8948 ioc->scsih_cmds.status, mpi_request, 8949 sizeof(Mpi2RaidActionRequest_t)/4, issue_reset); 8950 rc = -EFAULT; 8951 goto out; 8952 } 8953 8954 if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) { 8955 8956 mpi_reply = ioc->scsih_cmds.reply; 8957 ioc_status = le16_to_cpu(mpi_reply->IOCStatus); 8958 if (ioc_status & MPI2_IOCSTATUS_FLAG_LOG_INFO_AVAILABLE) 8959 log_info = le32_to_cpu(mpi_reply->IOCLogInfo); 8960 else 8961 log_info = 0; 8962 ioc_status &= MPI2_IOCSTATUS_MASK; 8963 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 8964 dewtprintk(ioc, 8965 ioc_info(ioc, "IR RAID_ACTION: failed: ioc_status(0x%04x), loginfo(0x%08x)!!!\n", 8966 ioc_status, log_info)); 8967 rc = -EFAULT; 8968 } else 8969 dewtprintk(ioc, 8970 ioc_info(ioc, "IR RAID_ACTION: completed successfully\n")); 8971 } 8972 8973 out: 8974 ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; 8975 mutex_unlock(&ioc->scsih_cmds.mutex); 8976 8977 if (issue_reset) 8978 mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER); 8979 return rc; 8980 } 8981 8982 /** 8983 * _scsih_reprobe_lun - reprobing lun 8984 * @sdev: scsi device struct 8985 * @no_uld_attach: sdev->no_uld_attach flag setting 8986 * 8987 **/ 8988 static void 8989 _scsih_reprobe_lun(struct scsi_device *sdev, void *no_uld_attach) 8990 { 8991 sdev->no_uld_attach = no_uld_attach ? 1 : 0; 8992 sdev_printk(KERN_INFO, sdev, "%s raid component\n", 8993 sdev->no_uld_attach ? "hiding" : "exposing"); 8994 WARN_ON(scsi_device_reprobe(sdev)); 8995 } 8996 8997 /** 8998 * _scsih_sas_volume_add - add new volume 8999 * @ioc: per adapter object 9000 * @element: IR config element data 9001 * Context: user. 9002 */ 9003 static void 9004 _scsih_sas_volume_add(struct MPT3SAS_ADAPTER *ioc, 9005 Mpi2EventIrConfigElement_t *element) 9006 { 9007 struct _raid_device *raid_device; 9008 unsigned long flags; 9009 u64 wwid; 9010 u16 handle = le16_to_cpu(element->VolDevHandle); 9011 int rc; 9012 9013 mpt3sas_config_get_volume_wwid(ioc, handle, &wwid); 9014 if (!wwid) { 9015 ioc_err(ioc, "failure at %s:%d/%s()!\n", 9016 __FILE__, __LINE__, __func__); 9017 return; 9018 } 9019 9020 spin_lock_irqsave(&ioc->raid_device_lock, flags); 9021 raid_device = _scsih_raid_device_find_by_wwid(ioc, wwid); 9022 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 9023 9024 if (raid_device) 9025 return; 9026 9027 raid_device = kzalloc(sizeof(struct _raid_device), GFP_KERNEL); 9028 if (!raid_device) { 9029 ioc_err(ioc, "failure at %s:%d/%s()!\n", 9030 __FILE__, __LINE__, __func__); 9031 return; 9032 } 9033 9034 raid_device->id = ioc->sas_id++; 9035 raid_device->channel = RAID_CHANNEL; 9036 raid_device->handle = handle; 9037 raid_device->wwid = wwid; 9038 _scsih_raid_device_add(ioc, raid_device); 9039 if (!ioc->wait_for_discovery_to_complete) { 9040 rc = scsi_add_device(ioc->shost, RAID_CHANNEL, 9041 raid_device->id, 0); 9042 if (rc) 9043 _scsih_raid_device_remove(ioc, raid_device); 9044 } else { 9045 spin_lock_irqsave(&ioc->raid_device_lock, flags); 9046 _scsih_determine_boot_device(ioc, raid_device, 1); 9047 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 9048 } 9049 } 9050 9051 /** 9052 * _scsih_sas_volume_delete - delete volume 9053 * @ioc: per adapter object 9054 * @handle: volume device handle 9055 * Context: user. 9056 */ 9057 static void 9058 _scsih_sas_volume_delete(struct MPT3SAS_ADAPTER *ioc, u16 handle) 9059 { 9060 struct _raid_device *raid_device; 9061 unsigned long flags; 9062 struct MPT3SAS_TARGET *sas_target_priv_data; 9063 struct scsi_target *starget = NULL; 9064 9065 spin_lock_irqsave(&ioc->raid_device_lock, flags); 9066 raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); 9067 if (raid_device) { 9068 if (raid_device->starget) { 9069 starget = raid_device->starget; 9070 sas_target_priv_data = starget->hostdata; 9071 sas_target_priv_data->deleted = 1; 9072 } 9073 ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n", 9074 raid_device->handle, (u64)raid_device->wwid); 9075 list_del(&raid_device->list); 9076 kfree(raid_device); 9077 } 9078 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 9079 if (starget) 9080 scsi_remove_target(&starget->dev); 9081 } 9082 9083 /** 9084 * _scsih_sas_pd_expose - expose pd component to /dev/sdX 9085 * @ioc: per adapter object 9086 * @element: IR config element data 9087 * Context: user. 9088 */ 9089 static void 9090 _scsih_sas_pd_expose(struct MPT3SAS_ADAPTER *ioc, 9091 Mpi2EventIrConfigElement_t *element) 9092 { 9093 struct _sas_device *sas_device; 9094 struct scsi_target *starget = NULL; 9095 struct MPT3SAS_TARGET *sas_target_priv_data; 9096 unsigned long flags; 9097 u16 handle = le16_to_cpu(element->PhysDiskDevHandle); 9098 9099 spin_lock_irqsave(&ioc->sas_device_lock, flags); 9100 sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); 9101 if (sas_device) { 9102 sas_device->volume_handle = 0; 9103 sas_device->volume_wwid = 0; 9104 clear_bit(handle, ioc->pd_handles); 9105 if (sas_device->starget && sas_device->starget->hostdata) { 9106 starget = sas_device->starget; 9107 sas_target_priv_data = starget->hostdata; 9108 sas_target_priv_data->flags &= 9109 ~MPT_TARGET_FLAGS_RAID_COMPONENT; 9110 } 9111 } 9112 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 9113 if (!sas_device) 9114 return; 9115 9116 /* exposing raid component */ 9117 if (starget) 9118 starget_for_each_device(starget, NULL, _scsih_reprobe_lun); 9119 9120 sas_device_put(sas_device); 9121 } 9122 9123 /** 9124 * _scsih_sas_pd_hide - hide pd component from /dev/sdX 9125 * @ioc: per adapter object 9126 * @element: IR config element data 9127 * Context: user. 9128 */ 9129 static void 9130 _scsih_sas_pd_hide(struct MPT3SAS_ADAPTER *ioc, 9131 Mpi2EventIrConfigElement_t *element) 9132 { 9133 struct _sas_device *sas_device; 9134 struct scsi_target *starget = NULL; 9135 struct MPT3SAS_TARGET *sas_target_priv_data; 9136 unsigned long flags; 9137 u16 handle = le16_to_cpu(element->PhysDiskDevHandle); 9138 u16 volume_handle = 0; 9139 u64 volume_wwid = 0; 9140 9141 mpt3sas_config_get_volume_handle(ioc, handle, &volume_handle); 9142 if (volume_handle) 9143 mpt3sas_config_get_volume_wwid(ioc, volume_handle, 9144 &volume_wwid); 9145 9146 spin_lock_irqsave(&ioc->sas_device_lock, flags); 9147 sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); 9148 if (sas_device) { 9149 set_bit(handle, ioc->pd_handles); 9150 if (sas_device->starget && sas_device->starget->hostdata) { 9151 starget = sas_device->starget; 9152 sas_target_priv_data = starget->hostdata; 9153 sas_target_priv_data->flags |= 9154 MPT_TARGET_FLAGS_RAID_COMPONENT; 9155 sas_device->volume_handle = volume_handle; 9156 sas_device->volume_wwid = volume_wwid; 9157 } 9158 } 9159 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 9160 if (!sas_device) 9161 return; 9162 9163 /* hiding raid component */ 9164 _scsih_ir_fastpath(ioc, handle, element->PhysDiskNum); 9165 9166 if (starget) 9167 starget_for_each_device(starget, (void *)1, _scsih_reprobe_lun); 9168 9169 sas_device_put(sas_device); 9170 } 9171 9172 /** 9173 * _scsih_sas_pd_delete - delete pd component 9174 * @ioc: per adapter object 9175 * @element: IR config element data 9176 * Context: user. 9177 */ 9178 static void 9179 _scsih_sas_pd_delete(struct MPT3SAS_ADAPTER *ioc, 9180 Mpi2EventIrConfigElement_t *element) 9181 { 9182 u16 handle = le16_to_cpu(element->PhysDiskDevHandle); 9183 9184 _scsih_device_remove_by_handle(ioc, handle); 9185 } 9186 9187 /** 9188 * _scsih_sas_pd_add - remove pd component 9189 * @ioc: per adapter object 9190 * @element: IR config element data 9191 * Context: user. 9192 */ 9193 static void 9194 _scsih_sas_pd_add(struct MPT3SAS_ADAPTER *ioc, 9195 Mpi2EventIrConfigElement_t *element) 9196 { 9197 struct _sas_device *sas_device; 9198 u16 handle = le16_to_cpu(element->PhysDiskDevHandle); 9199 Mpi2ConfigReply_t mpi_reply; 9200 Mpi2SasDevicePage0_t sas_device_pg0; 9201 u32 ioc_status; 9202 u64 sas_address; 9203 u16 parent_handle; 9204 9205 set_bit(handle, ioc->pd_handles); 9206 9207 sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); 9208 if (sas_device) { 9209 _scsih_ir_fastpath(ioc, handle, element->PhysDiskNum); 9210 sas_device_put(sas_device); 9211 return; 9212 } 9213 9214 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0, 9215 MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { 9216 ioc_err(ioc, "failure at %s:%d/%s()!\n", 9217 __FILE__, __LINE__, __func__); 9218 return; 9219 } 9220 9221 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 9222 MPI2_IOCSTATUS_MASK; 9223 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 9224 ioc_err(ioc, "failure at %s:%d/%s()!\n", 9225 __FILE__, __LINE__, __func__); 9226 return; 9227 } 9228 9229 parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle); 9230 if (!_scsih_get_sas_address(ioc, parent_handle, &sas_address)) 9231 mpt3sas_transport_update_links(ioc, sas_address, handle, 9232 sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5, 9233 mpt3sas_get_port_by_id(ioc, 9234 sas_device_pg0.PhysicalPort, 0)); 9235 9236 _scsih_ir_fastpath(ioc, handle, element->PhysDiskNum); 9237 _scsih_add_device(ioc, handle, 0, 1); 9238 } 9239 9240 /** 9241 * _scsih_sas_ir_config_change_event_debug - debug for IR Config Change events 9242 * @ioc: per adapter object 9243 * @event_data: event data payload 9244 * Context: user. 9245 */ 9246 static void 9247 _scsih_sas_ir_config_change_event_debug(struct MPT3SAS_ADAPTER *ioc, 9248 Mpi2EventDataIrConfigChangeList_t *event_data) 9249 { 9250 Mpi2EventIrConfigElement_t *element; 9251 u8 element_type; 9252 int i; 9253 char *reason_str = NULL, *element_str = NULL; 9254 9255 element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; 9256 9257 ioc_info(ioc, "raid config change: (%s), elements(%d)\n", 9258 le32_to_cpu(event_data->Flags) & MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG ? 9259 "foreign" : "native", 9260 event_data->NumElements); 9261 for (i = 0; i < event_data->NumElements; i++, element++) { 9262 switch (element->ReasonCode) { 9263 case MPI2_EVENT_IR_CHANGE_RC_ADDED: 9264 reason_str = "add"; 9265 break; 9266 case MPI2_EVENT_IR_CHANGE_RC_REMOVED: 9267 reason_str = "remove"; 9268 break; 9269 case MPI2_EVENT_IR_CHANGE_RC_NO_CHANGE: 9270 reason_str = "no change"; 9271 break; 9272 case MPI2_EVENT_IR_CHANGE_RC_HIDE: 9273 reason_str = "hide"; 9274 break; 9275 case MPI2_EVENT_IR_CHANGE_RC_UNHIDE: 9276 reason_str = "unhide"; 9277 break; 9278 case MPI2_EVENT_IR_CHANGE_RC_VOLUME_CREATED: 9279 reason_str = "volume_created"; 9280 break; 9281 case MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED: 9282 reason_str = "volume_deleted"; 9283 break; 9284 case MPI2_EVENT_IR_CHANGE_RC_PD_CREATED: 9285 reason_str = "pd_created"; 9286 break; 9287 case MPI2_EVENT_IR_CHANGE_RC_PD_DELETED: 9288 reason_str = "pd_deleted"; 9289 break; 9290 default: 9291 reason_str = "unknown reason"; 9292 break; 9293 } 9294 element_type = le16_to_cpu(element->ElementFlags) & 9295 MPI2_EVENT_IR_CHANGE_EFLAGS_ELEMENT_TYPE_MASK; 9296 switch (element_type) { 9297 case MPI2_EVENT_IR_CHANGE_EFLAGS_VOLUME_ELEMENT: 9298 element_str = "volume"; 9299 break; 9300 case MPI2_EVENT_IR_CHANGE_EFLAGS_VOLPHYSDISK_ELEMENT: 9301 element_str = "phys disk"; 9302 break; 9303 case MPI2_EVENT_IR_CHANGE_EFLAGS_HOTSPARE_ELEMENT: 9304 element_str = "hot spare"; 9305 break; 9306 default: 9307 element_str = "unknown element"; 9308 break; 9309 } 9310 pr_info("\t(%s:%s), vol handle(0x%04x), " \ 9311 "pd handle(0x%04x), pd num(0x%02x)\n", element_str, 9312 reason_str, le16_to_cpu(element->VolDevHandle), 9313 le16_to_cpu(element->PhysDiskDevHandle), 9314 element->PhysDiskNum); 9315 } 9316 } 9317 9318 /** 9319 * _scsih_sas_ir_config_change_event - handle ir configuration change events 9320 * @ioc: per adapter object 9321 * @fw_event: The fw_event_work object 9322 * Context: user. 9323 */ 9324 static void 9325 _scsih_sas_ir_config_change_event(struct MPT3SAS_ADAPTER *ioc, 9326 struct fw_event_work *fw_event) 9327 { 9328 Mpi2EventIrConfigElement_t *element; 9329 int i; 9330 u8 foreign_config; 9331 Mpi2EventDataIrConfigChangeList_t *event_data = 9332 (Mpi2EventDataIrConfigChangeList_t *) 9333 fw_event->event_data; 9334 9335 if ((ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) && 9336 (!ioc->hide_ir_msg)) 9337 _scsih_sas_ir_config_change_event_debug(ioc, event_data); 9338 9339 foreign_config = (le32_to_cpu(event_data->Flags) & 9340 MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG) ? 1 : 0; 9341 9342 element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; 9343 if (ioc->shost_recovery && 9344 ioc->hba_mpi_version_belonged != MPI2_VERSION) { 9345 for (i = 0; i < event_data->NumElements; i++, element++) { 9346 if (element->ReasonCode == MPI2_EVENT_IR_CHANGE_RC_HIDE) 9347 _scsih_ir_fastpath(ioc, 9348 le16_to_cpu(element->PhysDiskDevHandle), 9349 element->PhysDiskNum); 9350 } 9351 return; 9352 } 9353 9354 for (i = 0; i < event_data->NumElements; i++, element++) { 9355 9356 switch (element->ReasonCode) { 9357 case MPI2_EVENT_IR_CHANGE_RC_VOLUME_CREATED: 9358 case MPI2_EVENT_IR_CHANGE_RC_ADDED: 9359 if (!foreign_config) 9360 _scsih_sas_volume_add(ioc, element); 9361 break; 9362 case MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED: 9363 case MPI2_EVENT_IR_CHANGE_RC_REMOVED: 9364 if (!foreign_config) 9365 _scsih_sas_volume_delete(ioc, 9366 le16_to_cpu(element->VolDevHandle)); 9367 break; 9368 case MPI2_EVENT_IR_CHANGE_RC_PD_CREATED: 9369 if (!ioc->is_warpdrive) 9370 _scsih_sas_pd_hide(ioc, element); 9371 break; 9372 case MPI2_EVENT_IR_CHANGE_RC_PD_DELETED: 9373 if (!ioc->is_warpdrive) 9374 _scsih_sas_pd_expose(ioc, element); 9375 break; 9376 case MPI2_EVENT_IR_CHANGE_RC_HIDE: 9377 if (!ioc->is_warpdrive) 9378 _scsih_sas_pd_add(ioc, element); 9379 break; 9380 case MPI2_EVENT_IR_CHANGE_RC_UNHIDE: 9381 if (!ioc->is_warpdrive) 9382 _scsih_sas_pd_delete(ioc, element); 9383 break; 9384 } 9385 } 9386 } 9387 9388 /** 9389 * _scsih_sas_ir_volume_event - IR volume event 9390 * @ioc: per adapter object 9391 * @fw_event: The fw_event_work object 9392 * Context: user. 9393 */ 9394 static void 9395 _scsih_sas_ir_volume_event(struct MPT3SAS_ADAPTER *ioc, 9396 struct fw_event_work *fw_event) 9397 { 9398 u64 wwid; 9399 unsigned long flags; 9400 struct _raid_device *raid_device; 9401 u16 handle; 9402 u32 state; 9403 int rc; 9404 Mpi2EventDataIrVolume_t *event_data = 9405 (Mpi2EventDataIrVolume_t *) fw_event->event_data; 9406 9407 if (ioc->shost_recovery) 9408 return; 9409 9410 if (event_data->ReasonCode != MPI2_EVENT_IR_VOLUME_RC_STATE_CHANGED) 9411 return; 9412 9413 handle = le16_to_cpu(event_data->VolDevHandle); 9414 state = le32_to_cpu(event_data->NewValue); 9415 if (!ioc->hide_ir_msg) 9416 dewtprintk(ioc, 9417 ioc_info(ioc, "%s: handle(0x%04x), old(0x%08x), new(0x%08x)\n", 9418 __func__, handle, 9419 le32_to_cpu(event_data->PreviousValue), 9420 state)); 9421 switch (state) { 9422 case MPI2_RAID_VOL_STATE_MISSING: 9423 case MPI2_RAID_VOL_STATE_FAILED: 9424 _scsih_sas_volume_delete(ioc, handle); 9425 break; 9426 9427 case MPI2_RAID_VOL_STATE_ONLINE: 9428 case MPI2_RAID_VOL_STATE_DEGRADED: 9429 case MPI2_RAID_VOL_STATE_OPTIMAL: 9430 9431 spin_lock_irqsave(&ioc->raid_device_lock, flags); 9432 raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); 9433 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 9434 9435 if (raid_device) 9436 break; 9437 9438 mpt3sas_config_get_volume_wwid(ioc, handle, &wwid); 9439 if (!wwid) { 9440 ioc_err(ioc, "failure at %s:%d/%s()!\n", 9441 __FILE__, __LINE__, __func__); 9442 break; 9443 } 9444 9445 raid_device = kzalloc(sizeof(struct _raid_device), GFP_KERNEL); 9446 if (!raid_device) { 9447 ioc_err(ioc, "failure at %s:%d/%s()!\n", 9448 __FILE__, __LINE__, __func__); 9449 break; 9450 } 9451 9452 raid_device->id = ioc->sas_id++; 9453 raid_device->channel = RAID_CHANNEL; 9454 raid_device->handle = handle; 9455 raid_device->wwid = wwid; 9456 _scsih_raid_device_add(ioc, raid_device); 9457 rc = scsi_add_device(ioc->shost, RAID_CHANNEL, 9458 raid_device->id, 0); 9459 if (rc) 9460 _scsih_raid_device_remove(ioc, raid_device); 9461 break; 9462 9463 case MPI2_RAID_VOL_STATE_INITIALIZING: 9464 default: 9465 break; 9466 } 9467 } 9468 9469 /** 9470 * _scsih_sas_ir_physical_disk_event - PD event 9471 * @ioc: per adapter object 9472 * @fw_event: The fw_event_work object 9473 * Context: user. 9474 */ 9475 static void 9476 _scsih_sas_ir_physical_disk_event(struct MPT3SAS_ADAPTER *ioc, 9477 struct fw_event_work *fw_event) 9478 { 9479 u16 handle, parent_handle; 9480 u32 state; 9481 struct _sas_device *sas_device; 9482 Mpi2ConfigReply_t mpi_reply; 9483 Mpi2SasDevicePage0_t sas_device_pg0; 9484 u32 ioc_status; 9485 Mpi2EventDataIrPhysicalDisk_t *event_data = 9486 (Mpi2EventDataIrPhysicalDisk_t *) fw_event->event_data; 9487 u64 sas_address; 9488 9489 if (ioc->shost_recovery) 9490 return; 9491 9492 if (event_data->ReasonCode != MPI2_EVENT_IR_PHYSDISK_RC_STATE_CHANGED) 9493 return; 9494 9495 handle = le16_to_cpu(event_data->PhysDiskDevHandle); 9496 state = le32_to_cpu(event_data->NewValue); 9497 9498 if (!ioc->hide_ir_msg) 9499 dewtprintk(ioc, 9500 ioc_info(ioc, "%s: handle(0x%04x), old(0x%08x), new(0x%08x)\n", 9501 __func__, handle, 9502 le32_to_cpu(event_data->PreviousValue), 9503 state)); 9504 9505 switch (state) { 9506 case MPI2_RAID_PD_STATE_ONLINE: 9507 case MPI2_RAID_PD_STATE_DEGRADED: 9508 case MPI2_RAID_PD_STATE_REBUILDING: 9509 case MPI2_RAID_PD_STATE_OPTIMAL: 9510 case MPI2_RAID_PD_STATE_HOT_SPARE: 9511 9512 if (!ioc->is_warpdrive) 9513 set_bit(handle, ioc->pd_handles); 9514 9515 sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); 9516 if (sas_device) { 9517 sas_device_put(sas_device); 9518 return; 9519 } 9520 9521 if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, 9522 &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, 9523 handle))) { 9524 ioc_err(ioc, "failure at %s:%d/%s()!\n", 9525 __FILE__, __LINE__, __func__); 9526 return; 9527 } 9528 9529 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 9530 MPI2_IOCSTATUS_MASK; 9531 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 9532 ioc_err(ioc, "failure at %s:%d/%s()!\n", 9533 __FILE__, __LINE__, __func__); 9534 return; 9535 } 9536 9537 parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle); 9538 if (!_scsih_get_sas_address(ioc, parent_handle, &sas_address)) 9539 mpt3sas_transport_update_links(ioc, sas_address, handle, 9540 sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5, 9541 mpt3sas_get_port_by_id(ioc, 9542 sas_device_pg0.PhysicalPort, 0)); 9543 9544 _scsih_add_device(ioc, handle, 0, 1); 9545 9546 break; 9547 9548 case MPI2_RAID_PD_STATE_OFFLINE: 9549 case MPI2_RAID_PD_STATE_NOT_CONFIGURED: 9550 case MPI2_RAID_PD_STATE_NOT_COMPATIBLE: 9551 default: 9552 break; 9553 } 9554 } 9555 9556 /** 9557 * _scsih_sas_ir_operation_status_event_debug - debug for IR op event 9558 * @ioc: per adapter object 9559 * @event_data: event data payload 9560 * Context: user. 9561 */ 9562 static void 9563 _scsih_sas_ir_operation_status_event_debug(struct MPT3SAS_ADAPTER *ioc, 9564 Mpi2EventDataIrOperationStatus_t *event_data) 9565 { 9566 char *reason_str = NULL; 9567 9568 switch (event_data->RAIDOperation) { 9569 case MPI2_EVENT_IR_RAIDOP_RESYNC: 9570 reason_str = "resync"; 9571 break; 9572 case MPI2_EVENT_IR_RAIDOP_ONLINE_CAP_EXPANSION: 9573 reason_str = "online capacity expansion"; 9574 break; 9575 case MPI2_EVENT_IR_RAIDOP_CONSISTENCY_CHECK: 9576 reason_str = "consistency check"; 9577 break; 9578 case MPI2_EVENT_IR_RAIDOP_BACKGROUND_INIT: 9579 reason_str = "background init"; 9580 break; 9581 case MPI2_EVENT_IR_RAIDOP_MAKE_DATA_CONSISTENT: 9582 reason_str = "make data consistent"; 9583 break; 9584 } 9585 9586 if (!reason_str) 9587 return; 9588 9589 ioc_info(ioc, "raid operational status: (%s)\thandle(0x%04x), percent complete(%d)\n", 9590 reason_str, 9591 le16_to_cpu(event_data->VolDevHandle), 9592 event_data->PercentComplete); 9593 } 9594 9595 /** 9596 * _scsih_sas_ir_operation_status_event - handle RAID operation events 9597 * @ioc: per adapter object 9598 * @fw_event: The fw_event_work object 9599 * Context: user. 9600 */ 9601 static void 9602 _scsih_sas_ir_operation_status_event(struct MPT3SAS_ADAPTER *ioc, 9603 struct fw_event_work *fw_event) 9604 { 9605 Mpi2EventDataIrOperationStatus_t *event_data = 9606 (Mpi2EventDataIrOperationStatus_t *) 9607 fw_event->event_data; 9608 static struct _raid_device *raid_device; 9609 unsigned long flags; 9610 u16 handle; 9611 9612 if ((ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) && 9613 (!ioc->hide_ir_msg)) 9614 _scsih_sas_ir_operation_status_event_debug(ioc, 9615 event_data); 9616 9617 /* code added for raid transport support */ 9618 if (event_data->RAIDOperation == MPI2_EVENT_IR_RAIDOP_RESYNC) { 9619 9620 spin_lock_irqsave(&ioc->raid_device_lock, flags); 9621 handle = le16_to_cpu(event_data->VolDevHandle); 9622 raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); 9623 if (raid_device) 9624 raid_device->percent_complete = 9625 event_data->PercentComplete; 9626 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 9627 } 9628 } 9629 9630 /** 9631 * _scsih_prep_device_scan - initialize parameters prior to device scan 9632 * @ioc: per adapter object 9633 * 9634 * Set the deleted flag prior to device scan. If the device is found during 9635 * the scan, then we clear the deleted flag. 9636 */ 9637 static void 9638 _scsih_prep_device_scan(struct MPT3SAS_ADAPTER *ioc) 9639 { 9640 struct MPT3SAS_DEVICE *sas_device_priv_data; 9641 struct scsi_device *sdev; 9642 9643 shost_for_each_device(sdev, ioc->shost) { 9644 sas_device_priv_data = sdev->hostdata; 9645 if (sas_device_priv_data && sas_device_priv_data->sas_target) 9646 sas_device_priv_data->sas_target->deleted = 1; 9647 } 9648 } 9649 9650 /** 9651 * _scsih_update_device_qdepth - Update QD during Reset. 9652 * @ioc: per adapter object 9653 * 9654 */ 9655 static void 9656 _scsih_update_device_qdepth(struct MPT3SAS_ADAPTER *ioc) 9657 { 9658 struct MPT3SAS_DEVICE *sas_device_priv_data; 9659 struct MPT3SAS_TARGET *sas_target_priv_data; 9660 struct _sas_device *sas_device; 9661 struct scsi_device *sdev; 9662 u16 qdepth; 9663 9664 ioc_info(ioc, "Update devices with firmware reported queue depth\n"); 9665 shost_for_each_device(sdev, ioc->shost) { 9666 sas_device_priv_data = sdev->hostdata; 9667 if (sas_device_priv_data && sas_device_priv_data->sas_target) { 9668 sas_target_priv_data = sas_device_priv_data->sas_target; 9669 sas_device = sas_device_priv_data->sas_target->sas_dev; 9670 if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) 9671 qdepth = ioc->max_nvme_qd; 9672 else if (sas_device && 9673 sas_device->device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET) 9674 qdepth = (sas_device->port_type > 1) ? 9675 ioc->max_wideport_qd : ioc->max_narrowport_qd; 9676 else if (sas_device && 9677 sas_device->device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE) 9678 qdepth = ioc->max_sata_qd; 9679 else 9680 continue; 9681 mpt3sas_scsih_change_queue_depth(sdev, qdepth); 9682 } 9683 } 9684 } 9685 9686 /** 9687 * _scsih_mark_responding_sas_device - mark a sas_devices as responding 9688 * @ioc: per adapter object 9689 * @sas_device_pg0: SAS Device page 0 9690 * 9691 * After host reset, find out whether devices are still responding. 9692 * Used in _scsih_remove_unresponsive_sas_devices. 9693 */ 9694 static void 9695 _scsih_mark_responding_sas_device(struct MPT3SAS_ADAPTER *ioc, 9696 Mpi2SasDevicePage0_t *sas_device_pg0) 9697 { 9698 struct MPT3SAS_TARGET *sas_target_priv_data = NULL; 9699 struct scsi_target *starget; 9700 struct _sas_device *sas_device = NULL; 9701 struct _enclosure_node *enclosure_dev = NULL; 9702 unsigned long flags; 9703 struct hba_port *port = mpt3sas_get_port_by_id( 9704 ioc, sas_device_pg0->PhysicalPort, 0); 9705 9706 if (sas_device_pg0->EnclosureHandle) { 9707 enclosure_dev = 9708 mpt3sas_scsih_enclosure_find_by_handle(ioc, 9709 le16_to_cpu(sas_device_pg0->EnclosureHandle)); 9710 if (enclosure_dev == NULL) 9711 ioc_info(ioc, "Enclosure handle(0x%04x) doesn't match with enclosure device!\n", 9712 sas_device_pg0->EnclosureHandle); 9713 } 9714 spin_lock_irqsave(&ioc->sas_device_lock, flags); 9715 list_for_each_entry(sas_device, &ioc->sas_device_list, list) { 9716 if (sas_device->sas_address != le64_to_cpu( 9717 sas_device_pg0->SASAddress)) 9718 continue; 9719 if (sas_device->slot != le16_to_cpu(sas_device_pg0->Slot)) 9720 continue; 9721 if (sas_device->port != port) 9722 continue; 9723 sas_device->responding = 1; 9724 starget = sas_device->starget; 9725 if (starget && starget->hostdata) { 9726 sas_target_priv_data = starget->hostdata; 9727 sas_target_priv_data->tm_busy = 0; 9728 sas_target_priv_data->deleted = 0; 9729 } else 9730 sas_target_priv_data = NULL; 9731 if (starget) { 9732 starget_printk(KERN_INFO, starget, 9733 "handle(0x%04x), sas_addr(0x%016llx)\n", 9734 le16_to_cpu(sas_device_pg0->DevHandle), 9735 (unsigned long long) 9736 sas_device->sas_address); 9737 9738 if (sas_device->enclosure_handle != 0) 9739 starget_printk(KERN_INFO, starget, 9740 "enclosure logical id(0x%016llx), slot(%d)\n", 9741 (unsigned long long) 9742 sas_device->enclosure_logical_id, 9743 sas_device->slot); 9744 } 9745 if (le16_to_cpu(sas_device_pg0->Flags) & 9746 MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) { 9747 sas_device->enclosure_level = 9748 sas_device_pg0->EnclosureLevel; 9749 memcpy(&sas_device->connector_name[0], 9750 &sas_device_pg0->ConnectorName[0], 4); 9751 } else { 9752 sas_device->enclosure_level = 0; 9753 sas_device->connector_name[0] = '\0'; 9754 } 9755 9756 sas_device->enclosure_handle = 9757 le16_to_cpu(sas_device_pg0->EnclosureHandle); 9758 sas_device->is_chassis_slot_valid = 0; 9759 if (enclosure_dev) { 9760 sas_device->enclosure_logical_id = le64_to_cpu( 9761 enclosure_dev->pg0.EnclosureLogicalID); 9762 if (le16_to_cpu(enclosure_dev->pg0.Flags) & 9763 MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) { 9764 sas_device->is_chassis_slot_valid = 1; 9765 sas_device->chassis_slot = 9766 enclosure_dev->pg0.ChassisSlot; 9767 } 9768 } 9769 9770 if (sas_device->handle == le16_to_cpu( 9771 sas_device_pg0->DevHandle)) 9772 goto out; 9773 pr_info("\thandle changed from(0x%04x)!!!\n", 9774 sas_device->handle); 9775 sas_device->handle = le16_to_cpu( 9776 sas_device_pg0->DevHandle); 9777 if (sas_target_priv_data) 9778 sas_target_priv_data->handle = 9779 le16_to_cpu(sas_device_pg0->DevHandle); 9780 goto out; 9781 } 9782 out: 9783 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 9784 } 9785 9786 /** 9787 * _scsih_create_enclosure_list_after_reset - Free Existing list, 9788 * And create enclosure list by scanning all Enclosure Page(0)s 9789 * @ioc: per adapter object 9790 */ 9791 static void 9792 _scsih_create_enclosure_list_after_reset(struct MPT3SAS_ADAPTER *ioc) 9793 { 9794 struct _enclosure_node *enclosure_dev; 9795 Mpi2ConfigReply_t mpi_reply; 9796 u16 enclosure_handle; 9797 int rc; 9798 9799 /* Free existing enclosure list */ 9800 mpt3sas_free_enclosure_list(ioc); 9801 9802 /* Re constructing enclosure list after reset*/ 9803 enclosure_handle = 0xFFFF; 9804 do { 9805 enclosure_dev = 9806 kzalloc(sizeof(struct _enclosure_node), GFP_KERNEL); 9807 if (!enclosure_dev) { 9808 ioc_err(ioc, "failure at %s:%d/%s()!\n", 9809 __FILE__, __LINE__, __func__); 9810 return; 9811 } 9812 rc = mpt3sas_config_get_enclosure_pg0(ioc, &mpi_reply, 9813 &enclosure_dev->pg0, 9814 MPI2_SAS_ENCLOS_PGAD_FORM_GET_NEXT_HANDLE, 9815 enclosure_handle); 9816 9817 if (rc || (le16_to_cpu(mpi_reply.IOCStatus) & 9818 MPI2_IOCSTATUS_MASK)) { 9819 kfree(enclosure_dev); 9820 return; 9821 } 9822 list_add_tail(&enclosure_dev->list, 9823 &ioc->enclosure_list); 9824 enclosure_handle = 9825 le16_to_cpu(enclosure_dev->pg0.EnclosureHandle); 9826 } while (1); 9827 } 9828 9829 /** 9830 * _scsih_search_responding_sas_devices - 9831 * @ioc: per adapter object 9832 * 9833 * After host reset, find out whether devices are still responding. 9834 * If not remove. 9835 */ 9836 static void 9837 _scsih_search_responding_sas_devices(struct MPT3SAS_ADAPTER *ioc) 9838 { 9839 Mpi2SasDevicePage0_t sas_device_pg0; 9840 Mpi2ConfigReply_t mpi_reply; 9841 u16 ioc_status; 9842 u16 handle; 9843 u32 device_info; 9844 9845 ioc_info(ioc, "search for end-devices: start\n"); 9846 9847 if (list_empty(&ioc->sas_device_list)) 9848 goto out; 9849 9850 handle = 0xFFFF; 9851 while (!(mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, 9852 &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_GET_NEXT_HANDLE, 9853 handle))) { 9854 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 9855 MPI2_IOCSTATUS_MASK; 9856 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) 9857 break; 9858 handle = le16_to_cpu(sas_device_pg0.DevHandle); 9859 device_info = le32_to_cpu(sas_device_pg0.DeviceInfo); 9860 if (!(_scsih_is_end_device(device_info))) 9861 continue; 9862 _scsih_mark_responding_sas_device(ioc, &sas_device_pg0); 9863 } 9864 9865 out: 9866 ioc_info(ioc, "search for end-devices: complete\n"); 9867 } 9868 9869 /** 9870 * _scsih_mark_responding_pcie_device - mark a pcie_device as responding 9871 * @ioc: per adapter object 9872 * @pcie_device_pg0: PCIe Device page 0 9873 * 9874 * After host reset, find out whether devices are still responding. 9875 * Used in _scsih_remove_unresponding_devices. 9876 */ 9877 static void 9878 _scsih_mark_responding_pcie_device(struct MPT3SAS_ADAPTER *ioc, 9879 Mpi26PCIeDevicePage0_t *pcie_device_pg0) 9880 { 9881 struct MPT3SAS_TARGET *sas_target_priv_data = NULL; 9882 struct scsi_target *starget; 9883 struct _pcie_device *pcie_device; 9884 unsigned long flags; 9885 9886 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 9887 list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) { 9888 if ((pcie_device->wwid == le64_to_cpu(pcie_device_pg0->WWID)) 9889 && (pcie_device->slot == le16_to_cpu( 9890 pcie_device_pg0->Slot))) { 9891 pcie_device->access_status = 9892 pcie_device_pg0->AccessStatus; 9893 pcie_device->responding = 1; 9894 starget = pcie_device->starget; 9895 if (starget && starget->hostdata) { 9896 sas_target_priv_data = starget->hostdata; 9897 sas_target_priv_data->tm_busy = 0; 9898 sas_target_priv_data->deleted = 0; 9899 } else 9900 sas_target_priv_data = NULL; 9901 if (starget) { 9902 starget_printk(KERN_INFO, starget, 9903 "handle(0x%04x), wwid(0x%016llx) ", 9904 pcie_device->handle, 9905 (unsigned long long)pcie_device->wwid); 9906 if (pcie_device->enclosure_handle != 0) 9907 starget_printk(KERN_INFO, starget, 9908 "enclosure logical id(0x%016llx), " 9909 "slot(%d)\n", 9910 (unsigned long long) 9911 pcie_device->enclosure_logical_id, 9912 pcie_device->slot); 9913 } 9914 9915 if (((le32_to_cpu(pcie_device_pg0->Flags)) & 9916 MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) && 9917 (ioc->hba_mpi_version_belonged != MPI2_VERSION)) { 9918 pcie_device->enclosure_level = 9919 pcie_device_pg0->EnclosureLevel; 9920 memcpy(&pcie_device->connector_name[0], 9921 &pcie_device_pg0->ConnectorName[0], 4); 9922 } else { 9923 pcie_device->enclosure_level = 0; 9924 pcie_device->connector_name[0] = '\0'; 9925 } 9926 9927 if (pcie_device->handle == le16_to_cpu( 9928 pcie_device_pg0->DevHandle)) 9929 goto out; 9930 pr_info("\thandle changed from(0x%04x)!!!\n", 9931 pcie_device->handle); 9932 pcie_device->handle = le16_to_cpu( 9933 pcie_device_pg0->DevHandle); 9934 if (sas_target_priv_data) 9935 sas_target_priv_data->handle = 9936 le16_to_cpu(pcie_device_pg0->DevHandle); 9937 goto out; 9938 } 9939 } 9940 9941 out: 9942 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 9943 } 9944 9945 /** 9946 * _scsih_search_responding_pcie_devices - 9947 * @ioc: per adapter object 9948 * 9949 * After host reset, find out whether devices are still responding. 9950 * If not remove. 9951 */ 9952 static void 9953 _scsih_search_responding_pcie_devices(struct MPT3SAS_ADAPTER *ioc) 9954 { 9955 Mpi26PCIeDevicePage0_t pcie_device_pg0; 9956 Mpi2ConfigReply_t mpi_reply; 9957 u16 ioc_status; 9958 u16 handle; 9959 u32 device_info; 9960 9961 ioc_info(ioc, "search for end-devices: start\n"); 9962 9963 if (list_empty(&ioc->pcie_device_list)) 9964 goto out; 9965 9966 handle = 0xFFFF; 9967 while (!(mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply, 9968 &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_GET_NEXT_HANDLE, 9969 handle))) { 9970 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 9971 MPI2_IOCSTATUS_MASK; 9972 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 9973 ioc_info(ioc, "\tbreak from %s: ioc_status(0x%04x), loginfo(0x%08x)\n", 9974 __func__, ioc_status, 9975 le32_to_cpu(mpi_reply.IOCLogInfo)); 9976 break; 9977 } 9978 handle = le16_to_cpu(pcie_device_pg0.DevHandle); 9979 device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo); 9980 if (!(_scsih_is_nvme_pciescsi_device(device_info))) 9981 continue; 9982 _scsih_mark_responding_pcie_device(ioc, &pcie_device_pg0); 9983 } 9984 out: 9985 ioc_info(ioc, "search for PCIe end-devices: complete\n"); 9986 } 9987 9988 /** 9989 * _scsih_mark_responding_raid_device - mark a raid_device as responding 9990 * @ioc: per adapter object 9991 * @wwid: world wide identifier for raid volume 9992 * @handle: device handle 9993 * 9994 * After host reset, find out whether devices are still responding. 9995 * Used in _scsih_remove_unresponsive_raid_devices. 9996 */ 9997 static void 9998 _scsih_mark_responding_raid_device(struct MPT3SAS_ADAPTER *ioc, u64 wwid, 9999 u16 handle) 10000 { 10001 struct MPT3SAS_TARGET *sas_target_priv_data = NULL; 10002 struct scsi_target *starget; 10003 struct _raid_device *raid_device; 10004 unsigned long flags; 10005 10006 spin_lock_irqsave(&ioc->raid_device_lock, flags); 10007 list_for_each_entry(raid_device, &ioc->raid_device_list, list) { 10008 if (raid_device->wwid == wwid && raid_device->starget) { 10009 starget = raid_device->starget; 10010 if (starget && starget->hostdata) { 10011 sas_target_priv_data = starget->hostdata; 10012 sas_target_priv_data->deleted = 0; 10013 } else 10014 sas_target_priv_data = NULL; 10015 raid_device->responding = 1; 10016 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 10017 starget_printk(KERN_INFO, raid_device->starget, 10018 "handle(0x%04x), wwid(0x%016llx)\n", handle, 10019 (unsigned long long)raid_device->wwid); 10020 10021 /* 10022 * WARPDRIVE: The handles of the PDs might have changed 10023 * across the host reset so re-initialize the 10024 * required data for Direct IO 10025 */ 10026 mpt3sas_init_warpdrive_properties(ioc, raid_device); 10027 spin_lock_irqsave(&ioc->raid_device_lock, flags); 10028 if (raid_device->handle == handle) { 10029 spin_unlock_irqrestore(&ioc->raid_device_lock, 10030 flags); 10031 return; 10032 } 10033 pr_info("\thandle changed from(0x%04x)!!!\n", 10034 raid_device->handle); 10035 raid_device->handle = handle; 10036 if (sas_target_priv_data) 10037 sas_target_priv_data->handle = handle; 10038 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 10039 return; 10040 } 10041 } 10042 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 10043 } 10044 10045 /** 10046 * _scsih_search_responding_raid_devices - 10047 * @ioc: per adapter object 10048 * 10049 * After host reset, find out whether devices are still responding. 10050 * If not remove. 10051 */ 10052 static void 10053 _scsih_search_responding_raid_devices(struct MPT3SAS_ADAPTER *ioc) 10054 { 10055 Mpi2RaidVolPage1_t volume_pg1; 10056 Mpi2RaidVolPage0_t volume_pg0; 10057 Mpi2RaidPhysDiskPage0_t pd_pg0; 10058 Mpi2ConfigReply_t mpi_reply; 10059 u16 ioc_status; 10060 u16 handle; 10061 u8 phys_disk_num; 10062 10063 if (!ioc->ir_firmware) 10064 return; 10065 10066 ioc_info(ioc, "search for raid volumes: start\n"); 10067 10068 if (list_empty(&ioc->raid_device_list)) 10069 goto out; 10070 10071 handle = 0xFFFF; 10072 while (!(mpt3sas_config_get_raid_volume_pg1(ioc, &mpi_reply, 10073 &volume_pg1, MPI2_RAID_VOLUME_PGAD_FORM_GET_NEXT_HANDLE, handle))) { 10074 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 10075 MPI2_IOCSTATUS_MASK; 10076 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) 10077 break; 10078 handle = le16_to_cpu(volume_pg1.DevHandle); 10079 10080 if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, 10081 &volume_pg0, MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle, 10082 sizeof(Mpi2RaidVolPage0_t))) 10083 continue; 10084 10085 if (volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_OPTIMAL || 10086 volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_ONLINE || 10087 volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_DEGRADED) 10088 _scsih_mark_responding_raid_device(ioc, 10089 le64_to_cpu(volume_pg1.WWID), handle); 10090 } 10091 10092 /* refresh the pd_handles */ 10093 if (!ioc->is_warpdrive) { 10094 phys_disk_num = 0xFF; 10095 memset(ioc->pd_handles, 0, ioc->pd_handles_sz); 10096 while (!(mpt3sas_config_get_phys_disk_pg0(ioc, &mpi_reply, 10097 &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_GET_NEXT_PHYSDISKNUM, 10098 phys_disk_num))) { 10099 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 10100 MPI2_IOCSTATUS_MASK; 10101 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) 10102 break; 10103 phys_disk_num = pd_pg0.PhysDiskNum; 10104 handle = le16_to_cpu(pd_pg0.DevHandle); 10105 set_bit(handle, ioc->pd_handles); 10106 } 10107 } 10108 out: 10109 ioc_info(ioc, "search for responding raid volumes: complete\n"); 10110 } 10111 10112 /** 10113 * _scsih_mark_responding_expander - mark a expander as responding 10114 * @ioc: per adapter object 10115 * @expander_pg0:SAS Expander Config Page0 10116 * 10117 * After host reset, find out whether devices are still responding. 10118 * Used in _scsih_remove_unresponsive_expanders. 10119 */ 10120 static void 10121 _scsih_mark_responding_expander(struct MPT3SAS_ADAPTER *ioc, 10122 Mpi2ExpanderPage0_t *expander_pg0) 10123 { 10124 struct _sas_node *sas_expander = NULL; 10125 unsigned long flags; 10126 int i; 10127 struct _enclosure_node *enclosure_dev = NULL; 10128 u16 handle = le16_to_cpu(expander_pg0->DevHandle); 10129 u16 enclosure_handle = le16_to_cpu(expander_pg0->EnclosureHandle); 10130 u64 sas_address = le64_to_cpu(expander_pg0->SASAddress); 10131 struct hba_port *port = mpt3sas_get_port_by_id( 10132 ioc, expander_pg0->PhysicalPort, 0); 10133 10134 if (enclosure_handle) 10135 enclosure_dev = 10136 mpt3sas_scsih_enclosure_find_by_handle(ioc, 10137 enclosure_handle); 10138 10139 spin_lock_irqsave(&ioc->sas_node_lock, flags); 10140 list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) { 10141 if (sas_expander->sas_address != sas_address) 10142 continue; 10143 if (sas_expander->port != port) 10144 continue; 10145 sas_expander->responding = 1; 10146 10147 if (enclosure_dev) { 10148 sas_expander->enclosure_logical_id = 10149 le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); 10150 sas_expander->enclosure_handle = 10151 le16_to_cpu(expander_pg0->EnclosureHandle); 10152 } 10153 10154 if (sas_expander->handle == handle) 10155 goto out; 10156 pr_info("\texpander(0x%016llx): handle changed" \ 10157 " from(0x%04x) to (0x%04x)!!!\n", 10158 (unsigned long long)sas_expander->sas_address, 10159 sas_expander->handle, handle); 10160 sas_expander->handle = handle; 10161 for (i = 0 ; i < sas_expander->num_phys ; i++) 10162 sas_expander->phy[i].handle = handle; 10163 goto out; 10164 } 10165 out: 10166 spin_unlock_irqrestore(&ioc->sas_node_lock, flags); 10167 } 10168 10169 /** 10170 * _scsih_search_responding_expanders - 10171 * @ioc: per adapter object 10172 * 10173 * After host reset, find out whether devices are still responding. 10174 * If not remove. 10175 */ 10176 static void 10177 _scsih_search_responding_expanders(struct MPT3SAS_ADAPTER *ioc) 10178 { 10179 Mpi2ExpanderPage0_t expander_pg0; 10180 Mpi2ConfigReply_t mpi_reply; 10181 u16 ioc_status; 10182 u64 sas_address; 10183 u16 handle; 10184 u8 port; 10185 10186 ioc_info(ioc, "search for expanders: start\n"); 10187 10188 if (list_empty(&ioc->sas_expander_list)) 10189 goto out; 10190 10191 handle = 0xFFFF; 10192 while (!(mpt3sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0, 10193 MPI2_SAS_EXPAND_PGAD_FORM_GET_NEXT_HNDL, handle))) { 10194 10195 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 10196 MPI2_IOCSTATUS_MASK; 10197 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) 10198 break; 10199 10200 handle = le16_to_cpu(expander_pg0.DevHandle); 10201 sas_address = le64_to_cpu(expander_pg0.SASAddress); 10202 port = expander_pg0.PhysicalPort; 10203 pr_info( 10204 "\texpander present: handle(0x%04x), sas_addr(0x%016llx), port:%d\n", 10205 handle, (unsigned long long)sas_address, 10206 (ioc->multipath_on_hba ? 10207 port : MULTIPATH_DISABLED_PORT_ID)); 10208 _scsih_mark_responding_expander(ioc, &expander_pg0); 10209 } 10210 10211 out: 10212 ioc_info(ioc, "search for expanders: complete\n"); 10213 } 10214 10215 /** 10216 * _scsih_remove_unresponding_devices - removing unresponding devices 10217 * @ioc: per adapter object 10218 */ 10219 static void 10220 _scsih_remove_unresponding_devices(struct MPT3SAS_ADAPTER *ioc) 10221 { 10222 struct _sas_device *sas_device, *sas_device_next; 10223 struct _sas_node *sas_expander, *sas_expander_next; 10224 struct _raid_device *raid_device, *raid_device_next; 10225 struct _pcie_device *pcie_device, *pcie_device_next; 10226 struct list_head tmp_list; 10227 unsigned long flags; 10228 LIST_HEAD(head); 10229 10230 ioc_info(ioc, "removing unresponding devices: start\n"); 10231 10232 /* removing unresponding end devices */ 10233 ioc_info(ioc, "removing unresponding devices: end-devices\n"); 10234 /* 10235 * Iterate, pulling off devices marked as non-responding. We become the 10236 * owner for the reference the list had on any object we prune. 10237 */ 10238 spin_lock_irqsave(&ioc->sas_device_lock, flags); 10239 10240 /* 10241 * Clean up the sas_device_init_list list as 10242 * driver goes for fresh scan as part of diag reset. 10243 */ 10244 list_for_each_entry_safe(sas_device, sas_device_next, 10245 &ioc->sas_device_init_list, list) { 10246 list_del_init(&sas_device->list); 10247 sas_device_put(sas_device); 10248 } 10249 10250 list_for_each_entry_safe(sas_device, sas_device_next, 10251 &ioc->sas_device_list, list) { 10252 if (!sas_device->responding) 10253 list_move_tail(&sas_device->list, &head); 10254 else 10255 sas_device->responding = 0; 10256 } 10257 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 10258 10259 /* 10260 * Now, uninitialize and remove the unresponding devices we pruned. 10261 */ 10262 list_for_each_entry_safe(sas_device, sas_device_next, &head, list) { 10263 _scsih_remove_device(ioc, sas_device); 10264 list_del_init(&sas_device->list); 10265 sas_device_put(sas_device); 10266 } 10267 10268 ioc_info(ioc, "Removing unresponding devices: pcie end-devices\n"); 10269 INIT_LIST_HEAD(&head); 10270 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 10271 /* 10272 * Clean up the pcie_device_init_list list as 10273 * driver goes for fresh scan as part of diag reset. 10274 */ 10275 list_for_each_entry_safe(pcie_device, pcie_device_next, 10276 &ioc->pcie_device_init_list, list) { 10277 list_del_init(&pcie_device->list); 10278 pcie_device_put(pcie_device); 10279 } 10280 10281 list_for_each_entry_safe(pcie_device, pcie_device_next, 10282 &ioc->pcie_device_list, list) { 10283 if (!pcie_device->responding) 10284 list_move_tail(&pcie_device->list, &head); 10285 else 10286 pcie_device->responding = 0; 10287 } 10288 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 10289 10290 list_for_each_entry_safe(pcie_device, pcie_device_next, &head, list) { 10291 _scsih_pcie_device_remove_from_sml(ioc, pcie_device); 10292 list_del_init(&pcie_device->list); 10293 pcie_device_put(pcie_device); 10294 } 10295 10296 /* removing unresponding volumes */ 10297 if (ioc->ir_firmware) { 10298 ioc_info(ioc, "removing unresponding devices: volumes\n"); 10299 list_for_each_entry_safe(raid_device, raid_device_next, 10300 &ioc->raid_device_list, list) { 10301 if (!raid_device->responding) 10302 _scsih_sas_volume_delete(ioc, 10303 raid_device->handle); 10304 else 10305 raid_device->responding = 0; 10306 } 10307 } 10308 10309 /* removing unresponding expanders */ 10310 ioc_info(ioc, "removing unresponding devices: expanders\n"); 10311 spin_lock_irqsave(&ioc->sas_node_lock, flags); 10312 INIT_LIST_HEAD(&tmp_list); 10313 list_for_each_entry_safe(sas_expander, sas_expander_next, 10314 &ioc->sas_expander_list, list) { 10315 if (!sas_expander->responding) 10316 list_move_tail(&sas_expander->list, &tmp_list); 10317 else 10318 sas_expander->responding = 0; 10319 } 10320 spin_unlock_irqrestore(&ioc->sas_node_lock, flags); 10321 list_for_each_entry_safe(sas_expander, sas_expander_next, &tmp_list, 10322 list) { 10323 _scsih_expander_node_remove(ioc, sas_expander); 10324 } 10325 10326 ioc_info(ioc, "removing unresponding devices: complete\n"); 10327 10328 /* unblock devices */ 10329 _scsih_ublock_io_all_device(ioc); 10330 } 10331 10332 static void 10333 _scsih_refresh_expander_links(struct MPT3SAS_ADAPTER *ioc, 10334 struct _sas_node *sas_expander, u16 handle) 10335 { 10336 Mpi2ExpanderPage1_t expander_pg1; 10337 Mpi2ConfigReply_t mpi_reply; 10338 int i; 10339 10340 for (i = 0 ; i < sas_expander->num_phys ; i++) { 10341 if ((mpt3sas_config_get_expander_pg1(ioc, &mpi_reply, 10342 &expander_pg1, i, handle))) { 10343 ioc_err(ioc, "failure at %s:%d/%s()!\n", 10344 __FILE__, __LINE__, __func__); 10345 return; 10346 } 10347 10348 mpt3sas_transport_update_links(ioc, sas_expander->sas_address, 10349 le16_to_cpu(expander_pg1.AttachedDevHandle), i, 10350 expander_pg1.NegotiatedLinkRate >> 4, 10351 sas_expander->port); 10352 } 10353 } 10354 10355 /** 10356 * _scsih_scan_for_devices_after_reset - scan for devices after host reset 10357 * @ioc: per adapter object 10358 */ 10359 static void 10360 _scsih_scan_for_devices_after_reset(struct MPT3SAS_ADAPTER *ioc) 10361 { 10362 Mpi2ExpanderPage0_t expander_pg0; 10363 Mpi2SasDevicePage0_t sas_device_pg0; 10364 Mpi26PCIeDevicePage0_t pcie_device_pg0; 10365 Mpi2RaidVolPage1_t volume_pg1; 10366 Mpi2RaidVolPage0_t volume_pg0; 10367 Mpi2RaidPhysDiskPage0_t pd_pg0; 10368 Mpi2EventIrConfigElement_t element; 10369 Mpi2ConfigReply_t mpi_reply; 10370 u8 phys_disk_num, port_id; 10371 u16 ioc_status; 10372 u16 handle, parent_handle; 10373 u64 sas_address; 10374 struct _sas_device *sas_device; 10375 struct _pcie_device *pcie_device; 10376 struct _sas_node *expander_device; 10377 static struct _raid_device *raid_device; 10378 u8 retry_count; 10379 unsigned long flags; 10380 10381 ioc_info(ioc, "scan devices: start\n"); 10382 10383 _scsih_sas_host_refresh(ioc); 10384 10385 ioc_info(ioc, "\tscan devices: expanders start\n"); 10386 10387 /* expanders */ 10388 handle = 0xFFFF; 10389 while (!(mpt3sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0, 10390 MPI2_SAS_EXPAND_PGAD_FORM_GET_NEXT_HNDL, handle))) { 10391 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 10392 MPI2_IOCSTATUS_MASK; 10393 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 10394 ioc_info(ioc, "\tbreak from expander scan: ioc_status(0x%04x), loginfo(0x%08x)\n", 10395 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); 10396 break; 10397 } 10398 handle = le16_to_cpu(expander_pg0.DevHandle); 10399 spin_lock_irqsave(&ioc->sas_node_lock, flags); 10400 port_id = expander_pg0.PhysicalPort; 10401 expander_device = mpt3sas_scsih_expander_find_by_sas_address( 10402 ioc, le64_to_cpu(expander_pg0.SASAddress), 10403 mpt3sas_get_port_by_id(ioc, port_id, 0)); 10404 spin_unlock_irqrestore(&ioc->sas_node_lock, flags); 10405 if (expander_device) 10406 _scsih_refresh_expander_links(ioc, expander_device, 10407 handle); 10408 else { 10409 ioc_info(ioc, "\tBEFORE adding expander: handle (0x%04x), sas_addr(0x%016llx)\n", 10410 handle, 10411 (u64)le64_to_cpu(expander_pg0.SASAddress)); 10412 _scsih_expander_add(ioc, handle); 10413 ioc_info(ioc, "\tAFTER adding expander: handle (0x%04x), sas_addr(0x%016llx)\n", 10414 handle, 10415 (u64)le64_to_cpu(expander_pg0.SASAddress)); 10416 } 10417 } 10418 10419 ioc_info(ioc, "\tscan devices: expanders complete\n"); 10420 10421 if (!ioc->ir_firmware) 10422 goto skip_to_sas; 10423 10424 ioc_info(ioc, "\tscan devices: phys disk start\n"); 10425 10426 /* phys disk */ 10427 phys_disk_num = 0xFF; 10428 while (!(mpt3sas_config_get_phys_disk_pg0(ioc, &mpi_reply, 10429 &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_GET_NEXT_PHYSDISKNUM, 10430 phys_disk_num))) { 10431 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 10432 MPI2_IOCSTATUS_MASK; 10433 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 10434 ioc_info(ioc, "\tbreak from phys disk scan: ioc_status(0x%04x), loginfo(0x%08x)\n", 10435 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); 10436 break; 10437 } 10438 phys_disk_num = pd_pg0.PhysDiskNum; 10439 handle = le16_to_cpu(pd_pg0.DevHandle); 10440 sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); 10441 if (sas_device) { 10442 sas_device_put(sas_device); 10443 continue; 10444 } 10445 if (mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, 10446 &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, 10447 handle) != 0) 10448 continue; 10449 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 10450 MPI2_IOCSTATUS_MASK; 10451 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 10452 ioc_info(ioc, "\tbreak from phys disk scan ioc_status(0x%04x), loginfo(0x%08x)\n", 10453 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); 10454 break; 10455 } 10456 parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle); 10457 if (!_scsih_get_sas_address(ioc, parent_handle, 10458 &sas_address)) { 10459 ioc_info(ioc, "\tBEFORE adding phys disk: handle (0x%04x), sas_addr(0x%016llx)\n", 10460 handle, 10461 (u64)le64_to_cpu(sas_device_pg0.SASAddress)); 10462 port_id = sas_device_pg0.PhysicalPort; 10463 mpt3sas_transport_update_links(ioc, sas_address, 10464 handle, sas_device_pg0.PhyNum, 10465 MPI2_SAS_NEG_LINK_RATE_1_5, 10466 mpt3sas_get_port_by_id(ioc, port_id, 0)); 10467 set_bit(handle, ioc->pd_handles); 10468 retry_count = 0; 10469 /* This will retry adding the end device. 10470 * _scsih_add_device() will decide on retries and 10471 * return "1" when it should be retried 10472 */ 10473 while (_scsih_add_device(ioc, handle, retry_count++, 10474 1)) { 10475 ssleep(1); 10476 } 10477 ioc_info(ioc, "\tAFTER adding phys disk: handle (0x%04x), sas_addr(0x%016llx)\n", 10478 handle, 10479 (u64)le64_to_cpu(sas_device_pg0.SASAddress)); 10480 } 10481 } 10482 10483 ioc_info(ioc, "\tscan devices: phys disk complete\n"); 10484 10485 ioc_info(ioc, "\tscan devices: volumes start\n"); 10486 10487 /* volumes */ 10488 handle = 0xFFFF; 10489 while (!(mpt3sas_config_get_raid_volume_pg1(ioc, &mpi_reply, 10490 &volume_pg1, MPI2_RAID_VOLUME_PGAD_FORM_GET_NEXT_HANDLE, handle))) { 10491 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 10492 MPI2_IOCSTATUS_MASK; 10493 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 10494 ioc_info(ioc, "\tbreak from volume scan: ioc_status(0x%04x), loginfo(0x%08x)\n", 10495 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); 10496 break; 10497 } 10498 handle = le16_to_cpu(volume_pg1.DevHandle); 10499 spin_lock_irqsave(&ioc->raid_device_lock, flags); 10500 raid_device = _scsih_raid_device_find_by_wwid(ioc, 10501 le64_to_cpu(volume_pg1.WWID)); 10502 spin_unlock_irqrestore(&ioc->raid_device_lock, flags); 10503 if (raid_device) 10504 continue; 10505 if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, 10506 &volume_pg0, MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle, 10507 sizeof(Mpi2RaidVolPage0_t))) 10508 continue; 10509 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 10510 MPI2_IOCSTATUS_MASK; 10511 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 10512 ioc_info(ioc, "\tbreak from volume scan: ioc_status(0x%04x), loginfo(0x%08x)\n", 10513 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); 10514 break; 10515 } 10516 if (volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_OPTIMAL || 10517 volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_ONLINE || 10518 volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_DEGRADED) { 10519 memset(&element, 0, sizeof(Mpi2EventIrConfigElement_t)); 10520 element.ReasonCode = MPI2_EVENT_IR_CHANGE_RC_ADDED; 10521 element.VolDevHandle = volume_pg1.DevHandle; 10522 ioc_info(ioc, "\tBEFORE adding volume: handle (0x%04x)\n", 10523 volume_pg1.DevHandle); 10524 _scsih_sas_volume_add(ioc, &element); 10525 ioc_info(ioc, "\tAFTER adding volume: handle (0x%04x)\n", 10526 volume_pg1.DevHandle); 10527 } 10528 } 10529 10530 ioc_info(ioc, "\tscan devices: volumes complete\n"); 10531 10532 skip_to_sas: 10533 10534 ioc_info(ioc, "\tscan devices: end devices start\n"); 10535 10536 /* sas devices */ 10537 handle = 0xFFFF; 10538 while (!(mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, 10539 &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_GET_NEXT_HANDLE, 10540 handle))) { 10541 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & 10542 MPI2_IOCSTATUS_MASK; 10543 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 10544 ioc_info(ioc, "\tbreak from end device scan: ioc_status(0x%04x), loginfo(0x%08x)\n", 10545 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); 10546 break; 10547 } 10548 handle = le16_to_cpu(sas_device_pg0.DevHandle); 10549 if (!(_scsih_is_end_device( 10550 le32_to_cpu(sas_device_pg0.DeviceInfo)))) 10551 continue; 10552 port_id = sas_device_pg0.PhysicalPort; 10553 sas_device = mpt3sas_get_sdev_by_addr(ioc, 10554 le64_to_cpu(sas_device_pg0.SASAddress), 10555 mpt3sas_get_port_by_id(ioc, port_id, 0)); 10556 if (sas_device) { 10557 sas_device_put(sas_device); 10558 continue; 10559 } 10560 parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle); 10561 if (!_scsih_get_sas_address(ioc, parent_handle, &sas_address)) { 10562 ioc_info(ioc, "\tBEFORE adding end device: handle (0x%04x), sas_addr(0x%016llx)\n", 10563 handle, 10564 (u64)le64_to_cpu(sas_device_pg0.SASAddress)); 10565 mpt3sas_transport_update_links(ioc, sas_address, handle, 10566 sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5, 10567 mpt3sas_get_port_by_id(ioc, port_id, 0)); 10568 retry_count = 0; 10569 /* This will retry adding the end device. 10570 * _scsih_add_device() will decide on retries and 10571 * return "1" when it should be retried 10572 */ 10573 while (_scsih_add_device(ioc, handle, retry_count++, 10574 0)) { 10575 ssleep(1); 10576 } 10577 ioc_info(ioc, "\tAFTER adding end device: handle (0x%04x), sas_addr(0x%016llx)\n", 10578 handle, 10579 (u64)le64_to_cpu(sas_device_pg0.SASAddress)); 10580 } 10581 } 10582 ioc_info(ioc, "\tscan devices: end devices complete\n"); 10583 ioc_info(ioc, "\tscan devices: pcie end devices start\n"); 10584 10585 /* pcie devices */ 10586 handle = 0xFFFF; 10587 while (!(mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply, 10588 &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_GET_NEXT_HANDLE, 10589 handle))) { 10590 ioc_status = le16_to_cpu(mpi_reply.IOCStatus) 10591 & MPI2_IOCSTATUS_MASK; 10592 if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { 10593 ioc_info(ioc, "\tbreak from pcie end device scan: ioc_status(0x%04x), loginfo(0x%08x)\n", 10594 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); 10595 break; 10596 } 10597 handle = le16_to_cpu(pcie_device_pg0.DevHandle); 10598 if (!(_scsih_is_nvme_pciescsi_device( 10599 le32_to_cpu(pcie_device_pg0.DeviceInfo)))) 10600 continue; 10601 pcie_device = mpt3sas_get_pdev_by_wwid(ioc, 10602 le64_to_cpu(pcie_device_pg0.WWID)); 10603 if (pcie_device) { 10604 pcie_device_put(pcie_device); 10605 continue; 10606 } 10607 retry_count = 0; 10608 parent_handle = le16_to_cpu(pcie_device_pg0.ParentDevHandle); 10609 _scsih_pcie_add_device(ioc, handle); 10610 10611 ioc_info(ioc, "\tAFTER adding pcie end device: handle (0x%04x), wwid(0x%016llx)\n", 10612 handle, (u64)le64_to_cpu(pcie_device_pg0.WWID)); 10613 } 10614 10615 ioc_info(ioc, "\tpcie devices: pcie end devices complete\n"); 10616 ioc_info(ioc, "scan devices: complete\n"); 10617 } 10618 10619 /** 10620 * mpt3sas_scsih_pre_reset_handler - reset callback handler (for scsih) 10621 * @ioc: per adapter object 10622 * 10623 * The handler for doing any required cleanup or initialization. 10624 */ 10625 void mpt3sas_scsih_pre_reset_handler(struct MPT3SAS_ADAPTER *ioc) 10626 { 10627 dtmprintk(ioc, ioc_info(ioc, "%s: MPT3_IOC_PRE_RESET\n", __func__)); 10628 } 10629 10630 /** 10631 * mpt3sas_scsih_clear_outstanding_scsi_tm_commands - clears outstanding 10632 * scsi & tm cmds. 10633 * @ioc: per adapter object 10634 * 10635 * The handler for doing any required cleanup or initialization. 10636 */ 10637 void 10638 mpt3sas_scsih_clear_outstanding_scsi_tm_commands(struct MPT3SAS_ADAPTER *ioc) 10639 { 10640 dtmprintk(ioc, 10641 ioc_info(ioc, "%s: clear outstanding scsi & tm cmds\n", __func__)); 10642 if (ioc->scsih_cmds.status & MPT3_CMD_PENDING) { 10643 ioc->scsih_cmds.status |= MPT3_CMD_RESET; 10644 mpt3sas_base_free_smid(ioc, ioc->scsih_cmds.smid); 10645 complete(&ioc->scsih_cmds.done); 10646 } 10647 if (ioc->tm_cmds.status & MPT3_CMD_PENDING) { 10648 ioc->tm_cmds.status |= MPT3_CMD_RESET; 10649 mpt3sas_base_free_smid(ioc, ioc->tm_cmds.smid); 10650 complete(&ioc->tm_cmds.done); 10651 } 10652 10653 memset(ioc->pend_os_device_add, 0, ioc->pend_os_device_add_sz); 10654 memset(ioc->device_remove_in_progress, 0, 10655 ioc->device_remove_in_progress_sz); 10656 _scsih_fw_event_cleanup_queue(ioc); 10657 _scsih_flush_running_cmds(ioc); 10658 } 10659 10660 /** 10661 * mpt3sas_scsih_reset_done_handler - reset callback handler (for scsih) 10662 * @ioc: per adapter object 10663 * 10664 * The handler for doing any required cleanup or initialization. 10665 */ 10666 void 10667 mpt3sas_scsih_reset_done_handler(struct MPT3SAS_ADAPTER *ioc) 10668 { 10669 dtmprintk(ioc, ioc_info(ioc, "%s: MPT3_IOC_DONE_RESET\n", __func__)); 10670 if (!(disable_discovery > 0 && !ioc->sas_hba.num_phys)) { 10671 if (ioc->multipath_on_hba) { 10672 _scsih_sas_port_refresh(ioc); 10673 _scsih_update_vphys_after_reset(ioc); 10674 } 10675 _scsih_prep_device_scan(ioc); 10676 _scsih_create_enclosure_list_after_reset(ioc); 10677 _scsih_search_responding_sas_devices(ioc); 10678 _scsih_search_responding_pcie_devices(ioc); 10679 _scsih_search_responding_raid_devices(ioc); 10680 _scsih_search_responding_expanders(ioc); 10681 _scsih_error_recovery_delete_devices(ioc); 10682 } 10683 } 10684 10685 /** 10686 * _mpt3sas_fw_work - delayed task for processing firmware events 10687 * @ioc: per adapter object 10688 * @fw_event: The fw_event_work object 10689 * Context: user. 10690 */ 10691 static void 10692 _mpt3sas_fw_work(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) 10693 { 10694 ioc->current_event = fw_event; 10695 _scsih_fw_event_del_from_list(ioc, fw_event); 10696 10697 /* the queue is being flushed so ignore this event */ 10698 if (ioc->remove_host || ioc->pci_error_recovery) { 10699 fw_event_work_put(fw_event); 10700 ioc->current_event = NULL; 10701 return; 10702 } 10703 10704 switch (fw_event->event) { 10705 case MPT3SAS_PROCESS_TRIGGER_DIAG: 10706 mpt3sas_process_trigger_data(ioc, 10707 (struct SL_WH_TRIGGERS_EVENT_DATA_T *) 10708 fw_event->event_data); 10709 break; 10710 case MPT3SAS_REMOVE_UNRESPONDING_DEVICES: 10711 while (scsi_host_in_recovery(ioc->shost) || 10712 ioc->shost_recovery) { 10713 /* 10714 * If we're unloading or cancelling the work, bail. 10715 * Otherwise, this can become an infinite loop. 10716 */ 10717 if (ioc->remove_host || ioc->fw_events_cleanup) 10718 goto out; 10719 ssleep(1); 10720 } 10721 _scsih_remove_unresponding_devices(ioc); 10722 _scsih_del_dirty_vphy(ioc); 10723 _scsih_del_dirty_port_entries(ioc); 10724 if (ioc->is_gen35_ioc) 10725 _scsih_update_device_qdepth(ioc); 10726 _scsih_scan_for_devices_after_reset(ioc); 10727 /* 10728 * If diag reset has occurred during the driver load 10729 * then driver has to complete the driver load operation 10730 * by executing the following items: 10731 *- Register the devices from sas_device_init_list to SML 10732 *- clear is_driver_loading flag, 10733 *- start the watchdog thread. 10734 * In happy driver load path, above things are taken care of when 10735 * driver executes scsih_scan_finished(). 10736 */ 10737 if (ioc->is_driver_loading) 10738 _scsih_complete_devices_scanning(ioc); 10739 _scsih_set_nvme_max_shutdown_latency(ioc); 10740 break; 10741 case MPT3SAS_PORT_ENABLE_COMPLETE: 10742 ioc->start_scan = 0; 10743 if (missing_delay[0] != -1 && missing_delay[1] != -1) 10744 mpt3sas_base_update_missing_delay(ioc, missing_delay[0], 10745 missing_delay[1]); 10746 dewtprintk(ioc, 10747 ioc_info(ioc, "port enable: complete from worker thread\n")); 10748 break; 10749 case MPT3SAS_TURN_ON_PFA_LED: 10750 _scsih_turn_on_pfa_led(ioc, fw_event->device_handle); 10751 break; 10752 case MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST: 10753 _scsih_sas_topology_change_event(ioc, fw_event); 10754 break; 10755 case MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE: 10756 if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) 10757 _scsih_sas_device_status_change_event_debug(ioc, 10758 (Mpi2EventDataSasDeviceStatusChange_t *) 10759 fw_event->event_data); 10760 break; 10761 case MPI2_EVENT_SAS_DISCOVERY: 10762 _scsih_sas_discovery_event(ioc, fw_event); 10763 break; 10764 case MPI2_EVENT_SAS_DEVICE_DISCOVERY_ERROR: 10765 _scsih_sas_device_discovery_error_event(ioc, fw_event); 10766 break; 10767 case MPI2_EVENT_SAS_BROADCAST_PRIMITIVE: 10768 _scsih_sas_broadcast_primitive_event(ioc, fw_event); 10769 break; 10770 case MPI2_EVENT_SAS_ENCL_DEVICE_STATUS_CHANGE: 10771 _scsih_sas_enclosure_dev_status_change_event(ioc, 10772 fw_event); 10773 break; 10774 case MPI2_EVENT_IR_CONFIGURATION_CHANGE_LIST: 10775 _scsih_sas_ir_config_change_event(ioc, fw_event); 10776 break; 10777 case MPI2_EVENT_IR_VOLUME: 10778 _scsih_sas_ir_volume_event(ioc, fw_event); 10779 break; 10780 case MPI2_EVENT_IR_PHYSICAL_DISK: 10781 _scsih_sas_ir_physical_disk_event(ioc, fw_event); 10782 break; 10783 case MPI2_EVENT_IR_OPERATION_STATUS: 10784 _scsih_sas_ir_operation_status_event(ioc, fw_event); 10785 break; 10786 case MPI2_EVENT_PCIE_DEVICE_STATUS_CHANGE: 10787 _scsih_pcie_device_status_change_event(ioc, fw_event); 10788 break; 10789 case MPI2_EVENT_PCIE_ENUMERATION: 10790 _scsih_pcie_enumeration_event(ioc, fw_event); 10791 break; 10792 case MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST: 10793 _scsih_pcie_topology_change_event(ioc, fw_event); 10794 ioc->current_event = NULL; 10795 return; 10796 } 10797 out: 10798 fw_event_work_put(fw_event); 10799 ioc->current_event = NULL; 10800 } 10801 10802 /** 10803 * _firmware_event_work 10804 * @work: The fw_event_work object 10805 * Context: user. 10806 * 10807 * wrappers for the work thread handling firmware events 10808 */ 10809 10810 static void 10811 _firmware_event_work(struct work_struct *work) 10812 { 10813 struct fw_event_work *fw_event = container_of(work, 10814 struct fw_event_work, work); 10815 10816 _mpt3sas_fw_work(fw_event->ioc, fw_event); 10817 } 10818 10819 /** 10820 * mpt3sas_scsih_event_callback - firmware event handler (called at ISR time) 10821 * @ioc: per adapter object 10822 * @msix_index: MSIX table index supplied by the OS 10823 * @reply: reply message frame(lower 32bit addr) 10824 * Context: interrupt. 10825 * 10826 * This function merely adds a new work task into ioc->firmware_event_thread. 10827 * The tasks are worked from _firmware_event_work in user context. 10828 * 10829 * Return: 1 meaning mf should be freed from _base_interrupt 10830 * 0 means the mf is freed from this function. 10831 */ 10832 u8 10833 mpt3sas_scsih_event_callback(struct MPT3SAS_ADAPTER *ioc, u8 msix_index, 10834 u32 reply) 10835 { 10836 struct fw_event_work *fw_event; 10837 Mpi2EventNotificationReply_t *mpi_reply; 10838 u16 event; 10839 u16 sz; 10840 Mpi26EventDataActiveCableExcept_t *ActiveCableEventData; 10841 10842 /* events turned off due to host reset */ 10843 if (ioc->pci_error_recovery) 10844 return 1; 10845 10846 mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply); 10847 10848 if (unlikely(!mpi_reply)) { 10849 ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n", 10850 __FILE__, __LINE__, __func__); 10851 return 1; 10852 } 10853 10854 event = le16_to_cpu(mpi_reply->Event); 10855 10856 if (event != MPI2_EVENT_LOG_ENTRY_ADDED) 10857 mpt3sas_trigger_event(ioc, event, 0); 10858 10859 switch (event) { 10860 /* handle these */ 10861 case MPI2_EVENT_SAS_BROADCAST_PRIMITIVE: 10862 { 10863 Mpi2EventDataSasBroadcastPrimitive_t *baen_data = 10864 (Mpi2EventDataSasBroadcastPrimitive_t *) 10865 mpi_reply->EventData; 10866 10867 if (baen_data->Primitive != 10868 MPI2_EVENT_PRIMITIVE_ASYNCHRONOUS_EVENT) 10869 return 1; 10870 10871 if (ioc->broadcast_aen_busy) { 10872 ioc->broadcast_aen_pending++; 10873 return 1; 10874 } else 10875 ioc->broadcast_aen_busy = 1; 10876 break; 10877 } 10878 10879 case MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST: 10880 _scsih_check_topo_delete_events(ioc, 10881 (Mpi2EventDataSasTopologyChangeList_t *) 10882 mpi_reply->EventData); 10883 /* 10884 * No need to add the topology change list 10885 * event to fw event work queue when 10886 * diag reset is going on. Since during diag 10887 * reset driver scan the devices by reading 10888 * sas device page0's not by processing the 10889 * events. 10890 */ 10891 if (ioc->shost_recovery) 10892 return 1; 10893 break; 10894 case MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST: 10895 _scsih_check_pcie_topo_remove_events(ioc, 10896 (Mpi26EventDataPCIeTopologyChangeList_t *) 10897 mpi_reply->EventData); 10898 if (ioc->shost_recovery) 10899 return 1; 10900 break; 10901 case MPI2_EVENT_IR_CONFIGURATION_CHANGE_LIST: 10902 _scsih_check_ir_config_unhide_events(ioc, 10903 (Mpi2EventDataIrConfigChangeList_t *) 10904 mpi_reply->EventData); 10905 break; 10906 case MPI2_EVENT_IR_VOLUME: 10907 _scsih_check_volume_delete_events(ioc, 10908 (Mpi2EventDataIrVolume_t *) 10909 mpi_reply->EventData); 10910 break; 10911 case MPI2_EVENT_LOG_ENTRY_ADDED: 10912 { 10913 Mpi2EventDataLogEntryAdded_t *log_entry; 10914 u32 log_code; 10915 10916 if (!ioc->is_warpdrive) 10917 break; 10918 10919 log_entry = (Mpi2EventDataLogEntryAdded_t *) 10920 mpi_reply->EventData; 10921 log_code = le32_to_cpu(*(__le32 *)log_entry->LogData); 10922 10923 if (le16_to_cpu(log_entry->LogEntryQualifier) 10924 != MPT2_WARPDRIVE_LOGENTRY) 10925 break; 10926 10927 switch (log_code) { 10928 case MPT2_WARPDRIVE_LC_SSDT: 10929 ioc_warn(ioc, "WarpDrive Warning: IO Throttling has occurred in the WarpDrive subsystem. Check WarpDrive documentation for additional details.\n"); 10930 break; 10931 case MPT2_WARPDRIVE_LC_SSDLW: 10932 ioc_warn(ioc, "WarpDrive Warning: Program/Erase Cycles for the WarpDrive subsystem in degraded range. Check WarpDrive documentation for additional details.\n"); 10933 break; 10934 case MPT2_WARPDRIVE_LC_SSDLF: 10935 ioc_err(ioc, "WarpDrive Fatal Error: There are no Program/Erase Cycles for the WarpDrive subsystem. The storage device will be in read-only mode. Check WarpDrive documentation for additional details.\n"); 10936 break; 10937 case MPT2_WARPDRIVE_LC_BRMF: 10938 ioc_err(ioc, "WarpDrive Fatal Error: The Backup Rail Monitor has failed on the WarpDrive subsystem. Check WarpDrive documentation for additional details.\n"); 10939 break; 10940 } 10941 10942 break; 10943 } 10944 case MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE: 10945 _scsih_sas_device_status_change_event(ioc, 10946 (Mpi2EventDataSasDeviceStatusChange_t *) 10947 mpi_reply->EventData); 10948 break; 10949 case MPI2_EVENT_IR_OPERATION_STATUS: 10950 case MPI2_EVENT_SAS_DISCOVERY: 10951 case MPI2_EVENT_SAS_DEVICE_DISCOVERY_ERROR: 10952 case MPI2_EVENT_SAS_ENCL_DEVICE_STATUS_CHANGE: 10953 case MPI2_EVENT_IR_PHYSICAL_DISK: 10954 case MPI2_EVENT_PCIE_ENUMERATION: 10955 case MPI2_EVENT_PCIE_DEVICE_STATUS_CHANGE: 10956 break; 10957 10958 case MPI2_EVENT_TEMP_THRESHOLD: 10959 _scsih_temp_threshold_events(ioc, 10960 (Mpi2EventDataTemperature_t *) 10961 mpi_reply->EventData); 10962 break; 10963 case MPI2_EVENT_ACTIVE_CABLE_EXCEPTION: 10964 ActiveCableEventData = 10965 (Mpi26EventDataActiveCableExcept_t *) mpi_reply->EventData; 10966 switch (ActiveCableEventData->ReasonCode) { 10967 case MPI26_EVENT_ACTIVE_CABLE_INSUFFICIENT_POWER: 10968 ioc_notice(ioc, "Currently an active cable with ReceptacleID %d\n", 10969 ActiveCableEventData->ReceptacleID); 10970 pr_notice("cannot be powered and devices connected\n"); 10971 pr_notice("to this active cable will not be seen\n"); 10972 pr_notice("This active cable requires %d mW of power\n", 10973 le32_to_cpu( 10974 ActiveCableEventData->ActiveCablePowerRequirement)); 10975 break; 10976 10977 case MPI26_EVENT_ACTIVE_CABLE_DEGRADED: 10978 ioc_notice(ioc, "Currently a cable with ReceptacleID %d\n", 10979 ActiveCableEventData->ReceptacleID); 10980 pr_notice( 10981 "is not running at optimal speed(12 Gb/s rate)\n"); 10982 break; 10983 } 10984 10985 break; 10986 10987 default: /* ignore the rest */ 10988 return 1; 10989 } 10990 10991 sz = le16_to_cpu(mpi_reply->EventDataLength) * 4; 10992 fw_event = alloc_fw_event_work(sz); 10993 if (!fw_event) { 10994 ioc_err(ioc, "failure at %s:%d/%s()!\n", 10995 __FILE__, __LINE__, __func__); 10996 return 1; 10997 } 10998 10999 memcpy(fw_event->event_data, mpi_reply->EventData, sz); 11000 fw_event->ioc = ioc; 11001 fw_event->VF_ID = mpi_reply->VF_ID; 11002 fw_event->VP_ID = mpi_reply->VP_ID; 11003 fw_event->event = event; 11004 _scsih_fw_event_add(ioc, fw_event); 11005 fw_event_work_put(fw_event); 11006 return 1; 11007 } 11008 11009 /** 11010 * _scsih_expander_node_remove - removing expander device from list. 11011 * @ioc: per adapter object 11012 * @sas_expander: the sas_device object 11013 * 11014 * Removing object and freeing associated memory from the 11015 * ioc->sas_expander_list. 11016 */ 11017 static void 11018 _scsih_expander_node_remove(struct MPT3SAS_ADAPTER *ioc, 11019 struct _sas_node *sas_expander) 11020 { 11021 struct _sas_port *mpt3sas_port, *next; 11022 unsigned long flags; 11023 int port_id; 11024 11025 /* remove sibling ports attached to this expander */ 11026 list_for_each_entry_safe(mpt3sas_port, next, 11027 &sas_expander->sas_port_list, port_list) { 11028 if (ioc->shost_recovery) 11029 return; 11030 if (mpt3sas_port->remote_identify.device_type == 11031 SAS_END_DEVICE) 11032 mpt3sas_device_remove_by_sas_address(ioc, 11033 mpt3sas_port->remote_identify.sas_address, 11034 mpt3sas_port->hba_port); 11035 else if (mpt3sas_port->remote_identify.device_type == 11036 SAS_EDGE_EXPANDER_DEVICE || 11037 mpt3sas_port->remote_identify.device_type == 11038 SAS_FANOUT_EXPANDER_DEVICE) 11039 mpt3sas_expander_remove(ioc, 11040 mpt3sas_port->remote_identify.sas_address, 11041 mpt3sas_port->hba_port); 11042 } 11043 11044 port_id = sas_expander->port->port_id; 11045 11046 mpt3sas_transport_port_remove(ioc, sas_expander->sas_address, 11047 sas_expander->sas_address_parent, sas_expander->port); 11048 11049 ioc_info(ioc, 11050 "expander_remove: handle(0x%04x), sas_addr(0x%016llx), port:%d\n", 11051 sas_expander->handle, (unsigned long long) 11052 sas_expander->sas_address, 11053 port_id); 11054 11055 spin_lock_irqsave(&ioc->sas_node_lock, flags); 11056 list_del(&sas_expander->list); 11057 spin_unlock_irqrestore(&ioc->sas_node_lock, flags); 11058 11059 kfree(sas_expander->phy); 11060 kfree(sas_expander); 11061 } 11062 11063 /** 11064 * _scsih_nvme_shutdown - NVMe shutdown notification 11065 * @ioc: per adapter object 11066 * 11067 * Sending IoUnitControl request with shutdown operation code to alert IOC that 11068 * the host system is shutting down so that IOC can issue NVMe shutdown to 11069 * NVMe drives attached to it. 11070 */ 11071 static void 11072 _scsih_nvme_shutdown(struct MPT3SAS_ADAPTER *ioc) 11073 { 11074 Mpi26IoUnitControlRequest_t *mpi_request; 11075 Mpi26IoUnitControlReply_t *mpi_reply; 11076 u16 smid; 11077 11078 /* are there any NVMe devices ? */ 11079 if (list_empty(&ioc->pcie_device_list)) 11080 return; 11081 11082 mutex_lock(&ioc->scsih_cmds.mutex); 11083 11084 if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) { 11085 ioc_err(ioc, "%s: scsih_cmd in use\n", __func__); 11086 goto out; 11087 } 11088 11089 ioc->scsih_cmds.status = MPT3_CMD_PENDING; 11090 11091 smid = mpt3sas_base_get_smid(ioc, ioc->scsih_cb_idx); 11092 if (!smid) { 11093 ioc_err(ioc, 11094 "%s: failed obtaining a smid\n", __func__); 11095 ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; 11096 goto out; 11097 } 11098 11099 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); 11100 ioc->scsih_cmds.smid = smid; 11101 memset(mpi_request, 0, sizeof(Mpi26IoUnitControlRequest_t)); 11102 mpi_request->Function = MPI2_FUNCTION_IO_UNIT_CONTROL; 11103 mpi_request->Operation = MPI26_CTRL_OP_SHUTDOWN; 11104 11105 init_completion(&ioc->scsih_cmds.done); 11106 ioc->put_smid_default(ioc, smid); 11107 /* Wait for max_shutdown_latency seconds */ 11108 ioc_info(ioc, 11109 "Io Unit Control shutdown (sending), Shutdown latency %d sec\n", 11110 ioc->max_shutdown_latency); 11111 wait_for_completion_timeout(&ioc->scsih_cmds.done, 11112 ioc->max_shutdown_latency*HZ); 11113 11114 if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) { 11115 ioc_err(ioc, "%s: timeout\n", __func__); 11116 goto out; 11117 } 11118 11119 if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) { 11120 mpi_reply = ioc->scsih_cmds.reply; 11121 ioc_info(ioc, "Io Unit Control shutdown (complete):" 11122 "ioc_status(0x%04x), loginfo(0x%08x)\n", 11123 le16_to_cpu(mpi_reply->IOCStatus), 11124 le32_to_cpu(mpi_reply->IOCLogInfo)); 11125 } 11126 out: 11127 ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; 11128 mutex_unlock(&ioc->scsih_cmds.mutex); 11129 } 11130 11131 11132 /** 11133 * _scsih_ir_shutdown - IR shutdown notification 11134 * @ioc: per adapter object 11135 * 11136 * Sending RAID Action to alert the Integrated RAID subsystem of the IOC that 11137 * the host system is shutting down. 11138 */ 11139 static void 11140 _scsih_ir_shutdown(struct MPT3SAS_ADAPTER *ioc) 11141 { 11142 Mpi2RaidActionRequest_t *mpi_request; 11143 Mpi2RaidActionReply_t *mpi_reply; 11144 u16 smid; 11145 11146 /* is IR firmware build loaded ? */ 11147 if (!ioc->ir_firmware) 11148 return; 11149 11150 /* are there any volumes ? */ 11151 if (list_empty(&ioc->raid_device_list)) 11152 return; 11153 11154 mutex_lock(&ioc->scsih_cmds.mutex); 11155 11156 if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) { 11157 ioc_err(ioc, "%s: scsih_cmd in use\n", __func__); 11158 goto out; 11159 } 11160 ioc->scsih_cmds.status = MPT3_CMD_PENDING; 11161 11162 smid = mpt3sas_base_get_smid(ioc, ioc->scsih_cb_idx); 11163 if (!smid) { 11164 ioc_err(ioc, "%s: failed obtaining a smid\n", __func__); 11165 ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; 11166 goto out; 11167 } 11168 11169 mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); 11170 ioc->scsih_cmds.smid = smid; 11171 memset(mpi_request, 0, sizeof(Mpi2RaidActionRequest_t)); 11172 11173 mpi_request->Function = MPI2_FUNCTION_RAID_ACTION; 11174 mpi_request->Action = MPI2_RAID_ACTION_SYSTEM_SHUTDOWN_INITIATED; 11175 11176 if (!ioc->hide_ir_msg) 11177 ioc_info(ioc, "IR shutdown (sending)\n"); 11178 init_completion(&ioc->scsih_cmds.done); 11179 ioc->put_smid_default(ioc, smid); 11180 wait_for_completion_timeout(&ioc->scsih_cmds.done, 10*HZ); 11181 11182 if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) { 11183 ioc_err(ioc, "%s: timeout\n", __func__); 11184 goto out; 11185 } 11186 11187 if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) { 11188 mpi_reply = ioc->scsih_cmds.reply; 11189 if (!ioc->hide_ir_msg) 11190 ioc_info(ioc, "IR shutdown (complete): ioc_status(0x%04x), loginfo(0x%08x)\n", 11191 le16_to_cpu(mpi_reply->IOCStatus), 11192 le32_to_cpu(mpi_reply->IOCLogInfo)); 11193 } 11194 11195 out: 11196 ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; 11197 mutex_unlock(&ioc->scsih_cmds.mutex); 11198 } 11199 11200 /** 11201 * _scsih_get_shost_and_ioc - get shost and ioc 11202 * and verify whether they are NULL or not 11203 * @pdev: PCI device struct 11204 * @shost: address of scsi host pointer 11205 * @ioc: address of HBA adapter pointer 11206 * 11207 * Return zero if *shost and *ioc are not NULL otherwise return error number. 11208 */ 11209 static int 11210 _scsih_get_shost_and_ioc(struct pci_dev *pdev, 11211 struct Scsi_Host **shost, struct MPT3SAS_ADAPTER **ioc) 11212 { 11213 *shost = pci_get_drvdata(pdev); 11214 if (*shost == NULL) { 11215 dev_err(&pdev->dev, "pdev's driver data is null\n"); 11216 return -ENXIO; 11217 } 11218 11219 *ioc = shost_priv(*shost); 11220 if (*ioc == NULL) { 11221 dev_err(&pdev->dev, "shost's private data is null\n"); 11222 return -ENXIO; 11223 } 11224 11225 return 0; 11226 } 11227 11228 /** 11229 * scsih_remove - detach and remove add host 11230 * @pdev: PCI device struct 11231 * 11232 * Routine called when unloading the driver. 11233 */ 11234 static void scsih_remove(struct pci_dev *pdev) 11235 { 11236 struct Scsi_Host *shost; 11237 struct MPT3SAS_ADAPTER *ioc; 11238 struct _sas_port *mpt3sas_port, *next_port; 11239 struct _raid_device *raid_device, *next; 11240 struct MPT3SAS_TARGET *sas_target_priv_data; 11241 struct _pcie_device *pcie_device, *pcienext; 11242 struct workqueue_struct *wq; 11243 unsigned long flags; 11244 Mpi2ConfigReply_t mpi_reply; 11245 struct hba_port *port, *port_next; 11246 11247 if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc)) 11248 return; 11249 11250 ioc->remove_host = 1; 11251 11252 if (!pci_device_is_present(pdev)) { 11253 mpt3sas_base_pause_mq_polling(ioc); 11254 _scsih_flush_running_cmds(ioc); 11255 } 11256 11257 _scsih_fw_event_cleanup_queue(ioc); 11258 11259 spin_lock_irqsave(&ioc->fw_event_lock, flags); 11260 wq = ioc->firmware_event_thread; 11261 ioc->firmware_event_thread = NULL; 11262 spin_unlock_irqrestore(&ioc->fw_event_lock, flags); 11263 if (wq) 11264 destroy_workqueue(wq); 11265 /* 11266 * Copy back the unmodified ioc page1. so that on next driver load, 11267 * current modified changes on ioc page1 won't take effect. 11268 */ 11269 if (ioc->is_aero_ioc) 11270 mpt3sas_config_set_ioc_pg1(ioc, &mpi_reply, 11271 &ioc->ioc_pg1_copy); 11272 /* release all the volumes */ 11273 _scsih_ir_shutdown(ioc); 11274 mpt3sas_destroy_debugfs(ioc); 11275 sas_remove_host(shost); 11276 list_for_each_entry_safe(raid_device, next, &ioc->raid_device_list, 11277 list) { 11278 if (raid_device->starget) { 11279 sas_target_priv_data = 11280 raid_device->starget->hostdata; 11281 sas_target_priv_data->deleted = 1; 11282 scsi_remove_target(&raid_device->starget->dev); 11283 } 11284 ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n", 11285 raid_device->handle, (u64)raid_device->wwid); 11286 _scsih_raid_device_remove(ioc, raid_device); 11287 } 11288 list_for_each_entry_safe(pcie_device, pcienext, &ioc->pcie_device_list, 11289 list) { 11290 _scsih_pcie_device_remove_from_sml(ioc, pcie_device); 11291 list_del_init(&pcie_device->list); 11292 pcie_device_put(pcie_device); 11293 } 11294 11295 /* free ports attached to the sas_host */ 11296 list_for_each_entry_safe(mpt3sas_port, next_port, 11297 &ioc->sas_hba.sas_port_list, port_list) { 11298 if (mpt3sas_port->remote_identify.device_type == 11299 SAS_END_DEVICE) 11300 mpt3sas_device_remove_by_sas_address(ioc, 11301 mpt3sas_port->remote_identify.sas_address, 11302 mpt3sas_port->hba_port); 11303 else if (mpt3sas_port->remote_identify.device_type == 11304 SAS_EDGE_EXPANDER_DEVICE || 11305 mpt3sas_port->remote_identify.device_type == 11306 SAS_FANOUT_EXPANDER_DEVICE) 11307 mpt3sas_expander_remove(ioc, 11308 mpt3sas_port->remote_identify.sas_address, 11309 mpt3sas_port->hba_port); 11310 } 11311 11312 list_for_each_entry_safe(port, port_next, 11313 &ioc->port_table_list, list) { 11314 list_del(&port->list); 11315 kfree(port); 11316 } 11317 11318 /* free phys attached to the sas_host */ 11319 if (ioc->sas_hba.num_phys) { 11320 kfree(ioc->sas_hba.phy); 11321 ioc->sas_hba.phy = NULL; 11322 ioc->sas_hba.num_phys = 0; 11323 } 11324 11325 mpt3sas_base_detach(ioc); 11326 mpt3sas_ctl_release(ioc); 11327 spin_lock(&gioc_lock); 11328 list_del(&ioc->list); 11329 spin_unlock(&gioc_lock); 11330 scsi_host_put(shost); 11331 } 11332 11333 /** 11334 * scsih_shutdown - routine call during system shutdown 11335 * @pdev: PCI device struct 11336 */ 11337 static void 11338 scsih_shutdown(struct pci_dev *pdev) 11339 { 11340 struct Scsi_Host *shost; 11341 struct MPT3SAS_ADAPTER *ioc; 11342 struct workqueue_struct *wq; 11343 unsigned long flags; 11344 Mpi2ConfigReply_t mpi_reply; 11345 11346 if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc)) 11347 return; 11348 11349 ioc->remove_host = 1; 11350 11351 if (!pci_device_is_present(pdev)) { 11352 mpt3sas_base_pause_mq_polling(ioc); 11353 _scsih_flush_running_cmds(ioc); 11354 } 11355 11356 _scsih_fw_event_cleanup_queue(ioc); 11357 11358 spin_lock_irqsave(&ioc->fw_event_lock, flags); 11359 wq = ioc->firmware_event_thread; 11360 ioc->firmware_event_thread = NULL; 11361 spin_unlock_irqrestore(&ioc->fw_event_lock, flags); 11362 if (wq) 11363 destroy_workqueue(wq); 11364 /* 11365 * Copy back the unmodified ioc page1 so that on next driver load, 11366 * current modified changes on ioc page1 won't take effect. 11367 */ 11368 if (ioc->is_aero_ioc) 11369 mpt3sas_config_set_ioc_pg1(ioc, &mpi_reply, 11370 &ioc->ioc_pg1_copy); 11371 11372 _scsih_ir_shutdown(ioc); 11373 _scsih_nvme_shutdown(ioc); 11374 mpt3sas_base_mask_interrupts(ioc); 11375 mpt3sas_base_stop_watchdog(ioc); 11376 ioc->shost_recovery = 1; 11377 mpt3sas_base_make_ioc_ready(ioc, SOFT_RESET); 11378 ioc->shost_recovery = 0; 11379 mpt3sas_base_free_irq(ioc); 11380 mpt3sas_base_disable_msix(ioc); 11381 } 11382 11383 11384 /** 11385 * _scsih_probe_boot_devices - reports 1st device 11386 * @ioc: per adapter object 11387 * 11388 * If specified in bios page 2, this routine reports the 1st 11389 * device scsi-ml or sas transport for persistent boot device 11390 * purposes. Please refer to function _scsih_determine_boot_device() 11391 */ 11392 static void 11393 _scsih_probe_boot_devices(struct MPT3SAS_ADAPTER *ioc) 11394 { 11395 u32 channel; 11396 void *device; 11397 struct _sas_device *sas_device; 11398 struct _raid_device *raid_device; 11399 struct _pcie_device *pcie_device; 11400 u16 handle; 11401 u64 sas_address_parent; 11402 u64 sas_address; 11403 unsigned long flags; 11404 int rc; 11405 int tid; 11406 struct hba_port *port; 11407 11408 /* no Bios, return immediately */ 11409 if (!ioc->bios_pg3.BiosVersion) 11410 return; 11411 11412 device = NULL; 11413 if (ioc->req_boot_device.device) { 11414 device = ioc->req_boot_device.device; 11415 channel = ioc->req_boot_device.channel; 11416 } else if (ioc->req_alt_boot_device.device) { 11417 device = ioc->req_alt_boot_device.device; 11418 channel = ioc->req_alt_boot_device.channel; 11419 } else if (ioc->current_boot_device.device) { 11420 device = ioc->current_boot_device.device; 11421 channel = ioc->current_boot_device.channel; 11422 } 11423 11424 if (!device) 11425 return; 11426 11427 if (channel == RAID_CHANNEL) { 11428 raid_device = device; 11429 /* 11430 * If this boot vd is already registered with SML then 11431 * no need to register it again as part of device scanning 11432 * after diag reset during driver load operation. 11433 */ 11434 if (raid_device->starget) 11435 return; 11436 rc = scsi_add_device(ioc->shost, RAID_CHANNEL, 11437 raid_device->id, 0); 11438 if (rc) 11439 _scsih_raid_device_remove(ioc, raid_device); 11440 } else if (channel == PCIE_CHANNEL) { 11441 pcie_device = device; 11442 /* 11443 * If this boot NVMe device is already registered with SML then 11444 * no need to register it again as part of device scanning 11445 * after diag reset during driver load operation. 11446 */ 11447 if (pcie_device->starget) 11448 return; 11449 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 11450 tid = pcie_device->id; 11451 list_move_tail(&pcie_device->list, &ioc->pcie_device_list); 11452 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 11453 rc = scsi_add_device(ioc->shost, PCIE_CHANNEL, tid, 0); 11454 if (rc) 11455 _scsih_pcie_device_remove(ioc, pcie_device); 11456 } else { 11457 sas_device = device; 11458 /* 11459 * If this boot sas/sata device is already registered with SML 11460 * then no need to register it again as part of device scanning 11461 * after diag reset during driver load operation. 11462 */ 11463 if (sas_device->starget) 11464 return; 11465 spin_lock_irqsave(&ioc->sas_device_lock, flags); 11466 handle = sas_device->handle; 11467 sas_address_parent = sas_device->sas_address_parent; 11468 sas_address = sas_device->sas_address; 11469 port = sas_device->port; 11470 list_move_tail(&sas_device->list, &ioc->sas_device_list); 11471 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 11472 11473 if (ioc->hide_drives) 11474 return; 11475 11476 if (!port) 11477 return; 11478 11479 if (!mpt3sas_transport_port_add(ioc, handle, 11480 sas_address_parent, port)) { 11481 _scsih_sas_device_remove(ioc, sas_device); 11482 } else if (!sas_device->starget) { 11483 if (!ioc->is_driver_loading) { 11484 mpt3sas_transport_port_remove(ioc, 11485 sas_address, 11486 sas_address_parent, port); 11487 _scsih_sas_device_remove(ioc, sas_device); 11488 } 11489 } 11490 } 11491 } 11492 11493 /** 11494 * _scsih_probe_raid - reporting raid volumes to scsi-ml 11495 * @ioc: per adapter object 11496 * 11497 * Called during initial loading of the driver. 11498 */ 11499 static void 11500 _scsih_probe_raid(struct MPT3SAS_ADAPTER *ioc) 11501 { 11502 struct _raid_device *raid_device, *raid_next; 11503 int rc; 11504 11505 list_for_each_entry_safe(raid_device, raid_next, 11506 &ioc->raid_device_list, list) { 11507 if (raid_device->starget) 11508 continue; 11509 rc = scsi_add_device(ioc->shost, RAID_CHANNEL, 11510 raid_device->id, 0); 11511 if (rc) 11512 _scsih_raid_device_remove(ioc, raid_device); 11513 } 11514 } 11515 11516 static struct _sas_device *get_next_sas_device(struct MPT3SAS_ADAPTER *ioc) 11517 { 11518 struct _sas_device *sas_device = NULL; 11519 unsigned long flags; 11520 11521 spin_lock_irqsave(&ioc->sas_device_lock, flags); 11522 if (!list_empty(&ioc->sas_device_init_list)) { 11523 sas_device = list_first_entry(&ioc->sas_device_init_list, 11524 struct _sas_device, list); 11525 sas_device_get(sas_device); 11526 } 11527 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 11528 11529 return sas_device; 11530 } 11531 11532 static void sas_device_make_active(struct MPT3SAS_ADAPTER *ioc, 11533 struct _sas_device *sas_device) 11534 { 11535 unsigned long flags; 11536 11537 spin_lock_irqsave(&ioc->sas_device_lock, flags); 11538 11539 /* 11540 * Since we dropped the lock during the call to port_add(), we need to 11541 * be careful here that somebody else didn't move or delete this item 11542 * while we were busy with other things. 11543 * 11544 * If it was on the list, we need a put() for the reference the list 11545 * had. Either way, we need a get() for the destination list. 11546 */ 11547 if (!list_empty(&sas_device->list)) { 11548 list_del_init(&sas_device->list); 11549 sas_device_put(sas_device); 11550 } 11551 11552 sas_device_get(sas_device); 11553 list_add_tail(&sas_device->list, &ioc->sas_device_list); 11554 11555 spin_unlock_irqrestore(&ioc->sas_device_lock, flags); 11556 } 11557 11558 /** 11559 * _scsih_probe_sas - reporting sas devices to sas transport 11560 * @ioc: per adapter object 11561 * 11562 * Called during initial loading of the driver. 11563 */ 11564 static void 11565 _scsih_probe_sas(struct MPT3SAS_ADAPTER *ioc) 11566 { 11567 struct _sas_device *sas_device; 11568 11569 if (ioc->hide_drives) 11570 return; 11571 11572 while ((sas_device = get_next_sas_device(ioc))) { 11573 if (!mpt3sas_transport_port_add(ioc, sas_device->handle, 11574 sas_device->sas_address_parent, sas_device->port)) { 11575 _scsih_sas_device_remove(ioc, sas_device); 11576 sas_device_put(sas_device); 11577 continue; 11578 } else if (!sas_device->starget) { 11579 /* 11580 * When asyn scanning is enabled, its not possible to 11581 * remove devices while scanning is turned on due to an 11582 * oops in scsi_sysfs_add_sdev()->add_device()-> 11583 * sysfs_addrm_start() 11584 */ 11585 if (!ioc->is_driver_loading) { 11586 mpt3sas_transport_port_remove(ioc, 11587 sas_device->sas_address, 11588 sas_device->sas_address_parent, 11589 sas_device->port); 11590 _scsih_sas_device_remove(ioc, sas_device); 11591 sas_device_put(sas_device); 11592 continue; 11593 } 11594 } 11595 sas_device_make_active(ioc, sas_device); 11596 sas_device_put(sas_device); 11597 } 11598 } 11599 11600 /** 11601 * get_next_pcie_device - Get the next pcie device 11602 * @ioc: per adapter object 11603 * 11604 * Get the next pcie device from pcie_device_init_list list. 11605 * 11606 * Return: pcie device structure if pcie_device_init_list list is not empty 11607 * otherwise returns NULL 11608 */ 11609 static struct _pcie_device *get_next_pcie_device(struct MPT3SAS_ADAPTER *ioc) 11610 { 11611 struct _pcie_device *pcie_device = NULL; 11612 unsigned long flags; 11613 11614 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 11615 if (!list_empty(&ioc->pcie_device_init_list)) { 11616 pcie_device = list_first_entry(&ioc->pcie_device_init_list, 11617 struct _pcie_device, list); 11618 pcie_device_get(pcie_device); 11619 } 11620 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 11621 11622 return pcie_device; 11623 } 11624 11625 /** 11626 * pcie_device_make_active - Add pcie device to pcie_device_list list 11627 * @ioc: per adapter object 11628 * @pcie_device: pcie device object 11629 * 11630 * Add the pcie device which has registered with SCSI Transport Later to 11631 * pcie_device_list list 11632 */ 11633 static void pcie_device_make_active(struct MPT3SAS_ADAPTER *ioc, 11634 struct _pcie_device *pcie_device) 11635 { 11636 unsigned long flags; 11637 11638 spin_lock_irqsave(&ioc->pcie_device_lock, flags); 11639 11640 if (!list_empty(&pcie_device->list)) { 11641 list_del_init(&pcie_device->list); 11642 pcie_device_put(pcie_device); 11643 } 11644 pcie_device_get(pcie_device); 11645 list_add_tail(&pcie_device->list, &ioc->pcie_device_list); 11646 11647 spin_unlock_irqrestore(&ioc->pcie_device_lock, flags); 11648 } 11649 11650 /** 11651 * _scsih_probe_pcie - reporting PCIe devices to scsi-ml 11652 * @ioc: per adapter object 11653 * 11654 * Called during initial loading of the driver. 11655 */ 11656 static void 11657 _scsih_probe_pcie(struct MPT3SAS_ADAPTER *ioc) 11658 { 11659 struct _pcie_device *pcie_device; 11660 int rc; 11661 11662 /* PCIe Device List */ 11663 while ((pcie_device = get_next_pcie_device(ioc))) { 11664 if (pcie_device->starget) { 11665 pcie_device_put(pcie_device); 11666 continue; 11667 } 11668 if (pcie_device->access_status == 11669 MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED) { 11670 pcie_device_make_active(ioc, pcie_device); 11671 pcie_device_put(pcie_device); 11672 continue; 11673 } 11674 rc = scsi_add_device(ioc->shost, PCIE_CHANNEL, 11675 pcie_device->id, 0); 11676 if (rc) { 11677 _scsih_pcie_device_remove(ioc, pcie_device); 11678 pcie_device_put(pcie_device); 11679 continue; 11680 } else if (!pcie_device->starget) { 11681 /* 11682 * When async scanning is enabled, its not possible to 11683 * remove devices while scanning is turned on due to an 11684 * oops in scsi_sysfs_add_sdev()->add_device()-> 11685 * sysfs_addrm_start() 11686 */ 11687 if (!ioc->is_driver_loading) { 11688 /* TODO-- Need to find out whether this condition will 11689 * occur or not 11690 */ 11691 _scsih_pcie_device_remove(ioc, pcie_device); 11692 pcie_device_put(pcie_device); 11693 continue; 11694 } 11695 } 11696 pcie_device_make_active(ioc, pcie_device); 11697 pcie_device_put(pcie_device); 11698 } 11699 } 11700 11701 /** 11702 * _scsih_probe_devices - probing for devices 11703 * @ioc: per adapter object 11704 * 11705 * Called during initial loading of the driver. 11706 */ 11707 static void 11708 _scsih_probe_devices(struct MPT3SAS_ADAPTER *ioc) 11709 { 11710 u16 volume_mapping_flags; 11711 11712 if (!(ioc->facts.ProtocolFlags & MPI2_IOCFACTS_PROTOCOL_SCSI_INITIATOR)) 11713 return; /* return when IOC doesn't support initiator mode */ 11714 11715 _scsih_probe_boot_devices(ioc); 11716 11717 if (ioc->ir_firmware) { 11718 volume_mapping_flags = 11719 le16_to_cpu(ioc->ioc_pg8.IRVolumeMappingFlags) & 11720 MPI2_IOCPAGE8_IRFLAGS_MASK_VOLUME_MAPPING_MODE; 11721 if (volume_mapping_flags == 11722 MPI2_IOCPAGE8_IRFLAGS_LOW_VOLUME_MAPPING) { 11723 _scsih_probe_raid(ioc); 11724 _scsih_probe_sas(ioc); 11725 } else { 11726 _scsih_probe_sas(ioc); 11727 _scsih_probe_raid(ioc); 11728 } 11729 } else { 11730 _scsih_probe_sas(ioc); 11731 _scsih_probe_pcie(ioc); 11732 } 11733 } 11734 11735 /** 11736 * scsih_scan_start - scsi lld callback for .scan_start 11737 * @shost: SCSI host pointer 11738 * 11739 * The shost has the ability to discover targets on its own instead 11740 * of scanning the entire bus. In our implemention, we will kick off 11741 * firmware discovery. 11742 */ 11743 static void 11744 scsih_scan_start(struct Scsi_Host *shost) 11745 { 11746 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); 11747 int rc; 11748 if (diag_buffer_enable != -1 && diag_buffer_enable != 0) 11749 mpt3sas_enable_diag_buffer(ioc, diag_buffer_enable); 11750 else if (ioc->manu_pg11.HostTraceBufferMaxSizeKB != 0) 11751 mpt3sas_enable_diag_buffer(ioc, 1); 11752 11753 if (disable_discovery > 0) 11754 return; 11755 11756 ioc->start_scan = 1; 11757 rc = mpt3sas_port_enable(ioc); 11758 11759 if (rc != 0) 11760 ioc_info(ioc, "port enable: FAILED\n"); 11761 } 11762 11763 /** 11764 * _scsih_complete_devices_scanning - add the devices to sml and 11765 * complete ioc initialization. 11766 * @ioc: per adapter object 11767 * 11768 * Return nothing. 11769 */ 11770 static void _scsih_complete_devices_scanning(struct MPT3SAS_ADAPTER *ioc) 11771 { 11772 11773 if (ioc->wait_for_discovery_to_complete) { 11774 ioc->wait_for_discovery_to_complete = 0; 11775 _scsih_probe_devices(ioc); 11776 } 11777 11778 mpt3sas_base_start_watchdog(ioc); 11779 ioc->is_driver_loading = 0; 11780 } 11781 11782 /** 11783 * scsih_scan_finished - scsi lld callback for .scan_finished 11784 * @shost: SCSI host pointer 11785 * @time: elapsed time of the scan in jiffies 11786 * 11787 * This function will be called periodicallyn until it returns 1 with the 11788 * scsi_host and the elapsed time of the scan in jiffies. In our implemention, 11789 * we wait for firmware discovery to complete, then return 1. 11790 */ 11791 static int 11792 scsih_scan_finished(struct Scsi_Host *shost, unsigned long time) 11793 { 11794 struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); 11795 u32 ioc_state; 11796 int issue_hard_reset = 0; 11797 11798 if (disable_discovery > 0) { 11799 ioc->is_driver_loading = 0; 11800 ioc->wait_for_discovery_to_complete = 0; 11801 return 1; 11802 } 11803 11804 if (time >= (300 * HZ)) { 11805 ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED; 11806 ioc_info(ioc, "port enable: FAILED with timeout (timeout=300s)\n"); 11807 ioc->is_driver_loading = 0; 11808 return 1; 11809 } 11810 11811 if (ioc->start_scan) { 11812 ioc_state = mpt3sas_base_get_iocstate(ioc, 0); 11813 if ((ioc_state & MPI2_IOC_STATE_MASK) == MPI2_IOC_STATE_FAULT) { 11814 mpt3sas_print_fault_code(ioc, ioc_state & 11815 MPI2_DOORBELL_DATA_MASK); 11816 issue_hard_reset = 1; 11817 goto out; 11818 } else if ((ioc_state & MPI2_IOC_STATE_MASK) == 11819 MPI2_IOC_STATE_COREDUMP) { 11820 mpt3sas_base_coredump_info(ioc, ioc_state & 11821 MPI2_DOORBELL_DATA_MASK); 11822 mpt3sas_base_wait_for_coredump_completion(ioc, __func__); 11823 issue_hard_reset = 1; 11824 goto out; 11825 } 11826 return 0; 11827 } 11828 11829 if (ioc->port_enable_cmds.status & MPT3_CMD_RESET) { 11830 ioc_info(ioc, 11831 "port enable: aborted due to diag reset\n"); 11832 ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED; 11833 goto out; 11834 } 11835 if (ioc->start_scan_failed) { 11836 ioc_info(ioc, "port enable: FAILED with (ioc_status=0x%08x)\n", 11837 ioc->start_scan_failed); 11838 ioc->is_driver_loading = 0; 11839 ioc->wait_for_discovery_to_complete = 0; 11840 ioc->remove_host = 1; 11841 return 1; 11842 } 11843 11844 ioc_info(ioc, "port enable: SUCCESS\n"); 11845 ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED; 11846 _scsih_complete_devices_scanning(ioc); 11847 11848 out: 11849 if (issue_hard_reset) { 11850 ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED; 11851 if (mpt3sas_base_hard_reset_handler(ioc, SOFT_RESET)) 11852 ioc->is_driver_loading = 0; 11853 } 11854 return 1; 11855 } 11856 11857 /** 11858 * scsih_map_queues - map reply queues with request queues 11859 * @shost: SCSI host pointer 11860 */ 11861 static void scsih_map_queues(struct Scsi_Host *shost) 11862 { 11863 struct MPT3SAS_ADAPTER *ioc = 11864 (struct MPT3SAS_ADAPTER *)shost->hostdata; 11865 struct blk_mq_queue_map *map; 11866 int i, qoff, offset; 11867 int nr_msix_vectors = ioc->iopoll_q_start_index; 11868 int iopoll_q_count = ioc->reply_queue_count - nr_msix_vectors; 11869 11870 if (shost->nr_hw_queues == 1) 11871 return; 11872 11873 for (i = 0, qoff = 0; i < shost->nr_maps; i++) { 11874 map = &shost->tag_set.map[i]; 11875 map->nr_queues = 0; 11876 offset = 0; 11877 if (i == HCTX_TYPE_DEFAULT) { 11878 map->nr_queues = 11879 nr_msix_vectors - ioc->high_iops_queues; 11880 offset = ioc->high_iops_queues; 11881 } else if (i == HCTX_TYPE_POLL) 11882 map->nr_queues = iopoll_q_count; 11883 11884 if (!map->nr_queues) 11885 BUG_ON(i == HCTX_TYPE_DEFAULT); 11886 11887 /* 11888 * The poll queue(s) doesn't have an IRQ (and hence IRQ 11889 * affinity), so use the regular blk-mq cpu mapping 11890 */ 11891 map->queue_offset = qoff; 11892 if (i != HCTX_TYPE_POLL) 11893 blk_mq_pci_map_queues(map, ioc->pdev, offset); 11894 else 11895 blk_mq_map_queues(map); 11896 11897 qoff += map->nr_queues; 11898 } 11899 } 11900 11901 /* shost template for SAS 2.0 HBA devices */ 11902 static const struct scsi_host_template mpt2sas_driver_template = { 11903 .module = THIS_MODULE, 11904 .name = "Fusion MPT SAS Host", 11905 .proc_name = MPT2SAS_DRIVER_NAME, 11906 .queuecommand = scsih_qcmd, 11907 .target_alloc = scsih_target_alloc, 11908 .slave_alloc = scsih_slave_alloc, 11909 .device_configure = scsih_device_configure, 11910 .target_destroy = scsih_target_destroy, 11911 .slave_destroy = scsih_slave_destroy, 11912 .scan_finished = scsih_scan_finished, 11913 .scan_start = scsih_scan_start, 11914 .change_queue_depth = scsih_change_queue_depth, 11915 .eh_abort_handler = scsih_abort, 11916 .eh_device_reset_handler = scsih_dev_reset, 11917 .eh_target_reset_handler = scsih_target_reset, 11918 .eh_host_reset_handler = scsih_host_reset, 11919 .bios_param = scsih_bios_param, 11920 .can_queue = 1, 11921 .this_id = -1, 11922 .sg_tablesize = MPT2SAS_SG_DEPTH, 11923 .max_sectors = 32767, 11924 .cmd_per_lun = 7, 11925 .shost_groups = mpt3sas_host_groups, 11926 .sdev_groups = mpt3sas_dev_groups, 11927 .track_queue_depth = 1, 11928 .cmd_size = sizeof(struct scsiio_tracker), 11929 }; 11930 11931 /* raid transport support for SAS 2.0 HBA devices */ 11932 static struct raid_function_template mpt2sas_raid_functions = { 11933 .cookie = &mpt2sas_driver_template, 11934 .is_raid = scsih_is_raid, 11935 .get_resync = scsih_get_resync, 11936 .get_state = scsih_get_state, 11937 }; 11938 11939 /* shost template for SAS 3.0 HBA devices */ 11940 static const struct scsi_host_template mpt3sas_driver_template = { 11941 .module = THIS_MODULE, 11942 .name = "Fusion MPT SAS Host", 11943 .proc_name = MPT3SAS_DRIVER_NAME, 11944 .queuecommand = scsih_qcmd, 11945 .target_alloc = scsih_target_alloc, 11946 .slave_alloc = scsih_slave_alloc, 11947 .device_configure = scsih_device_configure, 11948 .target_destroy = scsih_target_destroy, 11949 .slave_destroy = scsih_slave_destroy, 11950 .scan_finished = scsih_scan_finished, 11951 .scan_start = scsih_scan_start, 11952 .change_queue_depth = scsih_change_queue_depth, 11953 .eh_abort_handler = scsih_abort, 11954 .eh_device_reset_handler = scsih_dev_reset, 11955 .eh_target_reset_handler = scsih_target_reset, 11956 .eh_host_reset_handler = scsih_host_reset, 11957 .bios_param = scsih_bios_param, 11958 .can_queue = 1, 11959 .this_id = -1, 11960 .sg_tablesize = MPT3SAS_SG_DEPTH, 11961 .max_sectors = 32767, 11962 .max_segment_size = 0xffffffff, 11963 .cmd_per_lun = 128, 11964 .shost_groups = mpt3sas_host_groups, 11965 .sdev_groups = mpt3sas_dev_groups, 11966 .track_queue_depth = 1, 11967 .cmd_size = sizeof(struct scsiio_tracker), 11968 .map_queues = scsih_map_queues, 11969 .mq_poll = mpt3sas_blk_mq_poll, 11970 }; 11971 11972 /* raid transport support for SAS 3.0 HBA devices */ 11973 static struct raid_function_template mpt3sas_raid_functions = { 11974 .cookie = &mpt3sas_driver_template, 11975 .is_raid = scsih_is_raid, 11976 .get_resync = scsih_get_resync, 11977 .get_state = scsih_get_state, 11978 }; 11979 11980 /** 11981 * _scsih_determine_hba_mpi_version - determine in which MPI version class 11982 * this device belongs to. 11983 * @pdev: PCI device struct 11984 * 11985 * return MPI2_VERSION for SAS 2.0 HBA devices, 11986 * MPI25_VERSION for SAS 3.0 HBA devices, and 11987 * MPI26 VERSION for Cutlass & Invader SAS 3.0 HBA devices 11988 */ 11989 static u16 11990 _scsih_determine_hba_mpi_version(struct pci_dev *pdev) 11991 { 11992 11993 switch (pdev->device) { 11994 case MPI2_MFGPAGE_DEVID_SSS6200: 11995 case MPI2_MFGPAGE_DEVID_SAS2004: 11996 case MPI2_MFGPAGE_DEVID_SAS2008: 11997 case MPI2_MFGPAGE_DEVID_SAS2108_1: 11998 case MPI2_MFGPAGE_DEVID_SAS2108_2: 11999 case MPI2_MFGPAGE_DEVID_SAS2108_3: 12000 case MPI2_MFGPAGE_DEVID_SAS2116_1: 12001 case MPI2_MFGPAGE_DEVID_SAS2116_2: 12002 case MPI2_MFGPAGE_DEVID_SAS2208_1: 12003 case MPI2_MFGPAGE_DEVID_SAS2208_2: 12004 case MPI2_MFGPAGE_DEVID_SAS2208_3: 12005 case MPI2_MFGPAGE_DEVID_SAS2208_4: 12006 case MPI2_MFGPAGE_DEVID_SAS2208_5: 12007 case MPI2_MFGPAGE_DEVID_SAS2208_6: 12008 case MPI2_MFGPAGE_DEVID_SAS2308_1: 12009 case MPI2_MFGPAGE_DEVID_SAS2308_2: 12010 case MPI2_MFGPAGE_DEVID_SAS2308_3: 12011 case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP: 12012 case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1: 12013 return MPI2_VERSION; 12014 case MPI25_MFGPAGE_DEVID_SAS3004: 12015 case MPI25_MFGPAGE_DEVID_SAS3008: 12016 case MPI25_MFGPAGE_DEVID_SAS3108_1: 12017 case MPI25_MFGPAGE_DEVID_SAS3108_2: 12018 case MPI25_MFGPAGE_DEVID_SAS3108_5: 12019 case MPI25_MFGPAGE_DEVID_SAS3108_6: 12020 return MPI25_VERSION; 12021 case MPI26_MFGPAGE_DEVID_SAS3216: 12022 case MPI26_MFGPAGE_DEVID_SAS3224: 12023 case MPI26_MFGPAGE_DEVID_SAS3316_1: 12024 case MPI26_MFGPAGE_DEVID_SAS3316_2: 12025 case MPI26_MFGPAGE_DEVID_SAS3316_3: 12026 case MPI26_MFGPAGE_DEVID_SAS3316_4: 12027 case MPI26_MFGPAGE_DEVID_SAS3324_1: 12028 case MPI26_MFGPAGE_DEVID_SAS3324_2: 12029 case MPI26_MFGPAGE_DEVID_SAS3324_3: 12030 case MPI26_MFGPAGE_DEVID_SAS3324_4: 12031 case MPI26_MFGPAGE_DEVID_SAS3508: 12032 case MPI26_MFGPAGE_DEVID_SAS3508_1: 12033 case MPI26_MFGPAGE_DEVID_SAS3408: 12034 case MPI26_MFGPAGE_DEVID_SAS3516: 12035 case MPI26_MFGPAGE_DEVID_SAS3516_1: 12036 case MPI26_MFGPAGE_DEVID_SAS3416: 12037 case MPI26_MFGPAGE_DEVID_SAS3616: 12038 case MPI26_ATLAS_PCIe_SWITCH_DEVID: 12039 case MPI26_MFGPAGE_DEVID_CFG_SEC_3916: 12040 case MPI26_MFGPAGE_DEVID_HARD_SEC_3916: 12041 case MPI26_MFGPAGE_DEVID_CFG_SEC_3816: 12042 case MPI26_MFGPAGE_DEVID_HARD_SEC_3816: 12043 case MPI26_MFGPAGE_DEVID_INVALID0_3916: 12044 case MPI26_MFGPAGE_DEVID_INVALID1_3916: 12045 case MPI26_MFGPAGE_DEVID_INVALID0_3816: 12046 case MPI26_MFGPAGE_DEVID_INVALID1_3816: 12047 return MPI26_VERSION; 12048 } 12049 return 0; 12050 } 12051 12052 /** 12053 * _scsih_probe - attach and add scsi host 12054 * @pdev: PCI device struct 12055 * @id: pci device id 12056 * 12057 * Return: 0 success, anything else error. 12058 */ 12059 static int 12060 _scsih_probe(struct pci_dev *pdev, const struct pci_device_id *id) 12061 { 12062 struct MPT3SAS_ADAPTER *ioc; 12063 struct Scsi_Host *shost = NULL; 12064 int rv; 12065 u16 hba_mpi_version; 12066 int iopoll_q_count = 0; 12067 12068 /* Determine in which MPI version class this pci device belongs */ 12069 hba_mpi_version = _scsih_determine_hba_mpi_version(pdev); 12070 if (hba_mpi_version == 0) 12071 return -ENODEV; 12072 12073 /* Enumerate only SAS 2.0 HBA's if hbas_to_enumerate is one, 12074 * for other generation HBA's return with -ENODEV 12075 */ 12076 if ((hbas_to_enumerate == 1) && (hba_mpi_version != MPI2_VERSION)) 12077 return -ENODEV; 12078 12079 /* Enumerate only SAS 3.0 HBA's if hbas_to_enumerate is two, 12080 * for other generation HBA's return with -ENODEV 12081 */ 12082 if ((hbas_to_enumerate == 2) && (!(hba_mpi_version == MPI25_VERSION 12083 || hba_mpi_version == MPI26_VERSION))) 12084 return -ENODEV; 12085 12086 switch (hba_mpi_version) { 12087 case MPI2_VERSION: 12088 pci_disable_link_state(pdev, PCIE_LINK_STATE_L0S | 12089 PCIE_LINK_STATE_L1 | PCIE_LINK_STATE_CLKPM); 12090 /* Use mpt2sas driver host template for SAS 2.0 HBA's */ 12091 shost = scsi_host_alloc(&mpt2sas_driver_template, 12092 sizeof(struct MPT3SAS_ADAPTER)); 12093 if (!shost) 12094 return -ENODEV; 12095 ioc = shost_priv(shost); 12096 memset(ioc, 0, sizeof(struct MPT3SAS_ADAPTER)); 12097 ioc->hba_mpi_version_belonged = hba_mpi_version; 12098 ioc->id = mpt2_ids++; 12099 sprintf(ioc->driver_name, "%s", MPT2SAS_DRIVER_NAME); 12100 switch (pdev->device) { 12101 case MPI2_MFGPAGE_DEVID_SSS6200: 12102 ioc->is_warpdrive = 1; 12103 ioc->hide_ir_msg = 1; 12104 break; 12105 case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP: 12106 case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1: 12107 ioc->is_mcpu_endpoint = 1; 12108 break; 12109 default: 12110 ioc->mfg_pg10_hide_flag = MFG_PAGE10_EXPOSE_ALL_DISKS; 12111 break; 12112 } 12113 12114 if (multipath_on_hba == -1 || multipath_on_hba == 0) 12115 ioc->multipath_on_hba = 0; 12116 else 12117 ioc->multipath_on_hba = 1; 12118 12119 break; 12120 case MPI25_VERSION: 12121 case MPI26_VERSION: 12122 /* Use mpt3sas driver host template for SAS 3.0 HBA's */ 12123 shost = scsi_host_alloc(&mpt3sas_driver_template, 12124 sizeof(struct MPT3SAS_ADAPTER)); 12125 if (!shost) 12126 return -ENODEV; 12127 ioc = shost_priv(shost); 12128 memset(ioc, 0, sizeof(struct MPT3SAS_ADAPTER)); 12129 ioc->hba_mpi_version_belonged = hba_mpi_version; 12130 ioc->id = mpt3_ids++; 12131 sprintf(ioc->driver_name, "%s", MPT3SAS_DRIVER_NAME); 12132 switch (pdev->device) { 12133 case MPI26_MFGPAGE_DEVID_SAS3508: 12134 case MPI26_MFGPAGE_DEVID_SAS3508_1: 12135 case MPI26_MFGPAGE_DEVID_SAS3408: 12136 case MPI26_MFGPAGE_DEVID_SAS3516: 12137 case MPI26_MFGPAGE_DEVID_SAS3516_1: 12138 case MPI26_MFGPAGE_DEVID_SAS3416: 12139 case MPI26_MFGPAGE_DEVID_SAS3616: 12140 case MPI26_ATLAS_PCIe_SWITCH_DEVID: 12141 ioc->is_gen35_ioc = 1; 12142 break; 12143 case MPI26_MFGPAGE_DEVID_INVALID0_3816: 12144 case MPI26_MFGPAGE_DEVID_INVALID0_3916: 12145 dev_err(&pdev->dev, 12146 "HBA with DeviceId 0x%04x, sub VendorId 0x%04x, sub DeviceId 0x%04x is Invalid", 12147 pdev->device, pdev->subsystem_vendor, 12148 pdev->subsystem_device); 12149 return 1; 12150 case MPI26_MFGPAGE_DEVID_INVALID1_3816: 12151 case MPI26_MFGPAGE_DEVID_INVALID1_3916: 12152 dev_err(&pdev->dev, 12153 "HBA with DeviceId 0x%04x, sub VendorId 0x%04x, sub DeviceId 0x%04x is Tampered", 12154 pdev->device, pdev->subsystem_vendor, 12155 pdev->subsystem_device); 12156 return 1; 12157 case MPI26_MFGPAGE_DEVID_CFG_SEC_3816: 12158 case MPI26_MFGPAGE_DEVID_CFG_SEC_3916: 12159 dev_info(&pdev->dev, 12160 "HBA is in Configurable Secure mode\n"); 12161 fallthrough; 12162 case MPI26_MFGPAGE_DEVID_HARD_SEC_3816: 12163 case MPI26_MFGPAGE_DEVID_HARD_SEC_3916: 12164 ioc->is_aero_ioc = ioc->is_gen35_ioc = 1; 12165 break; 12166 default: 12167 ioc->is_gen35_ioc = ioc->is_aero_ioc = 0; 12168 } 12169 if ((ioc->hba_mpi_version_belonged == MPI25_VERSION && 12170 pdev->revision >= SAS3_PCI_DEVICE_C0_REVISION) || 12171 (ioc->hba_mpi_version_belonged == MPI26_VERSION)) { 12172 ioc->combined_reply_queue = 1; 12173 if (ioc->is_gen35_ioc) 12174 ioc->combined_reply_index_count = 12175 MPT3_SUP_REPLY_POST_HOST_INDEX_REG_COUNT_G35; 12176 else 12177 ioc->combined_reply_index_count = 12178 MPT3_SUP_REPLY_POST_HOST_INDEX_REG_COUNT_G3; 12179 } 12180 12181 switch (ioc->is_gen35_ioc) { 12182 case 0: 12183 if (multipath_on_hba == -1 || multipath_on_hba == 0) 12184 ioc->multipath_on_hba = 0; 12185 else 12186 ioc->multipath_on_hba = 1; 12187 break; 12188 case 1: 12189 if (multipath_on_hba == -1 || multipath_on_hba > 0) 12190 ioc->multipath_on_hba = 1; 12191 else 12192 ioc->multipath_on_hba = 0; 12193 break; 12194 default: 12195 break; 12196 } 12197 12198 break; 12199 default: 12200 return -ENODEV; 12201 } 12202 12203 INIT_LIST_HEAD(&ioc->list); 12204 spin_lock(&gioc_lock); 12205 list_add_tail(&ioc->list, &mpt3sas_ioc_list); 12206 spin_unlock(&gioc_lock); 12207 ioc->shost = shost; 12208 ioc->pdev = pdev; 12209 ioc->scsi_io_cb_idx = scsi_io_cb_idx; 12210 ioc->tm_cb_idx = tm_cb_idx; 12211 ioc->ctl_cb_idx = ctl_cb_idx; 12212 ioc->base_cb_idx = base_cb_idx; 12213 ioc->port_enable_cb_idx = port_enable_cb_idx; 12214 ioc->transport_cb_idx = transport_cb_idx; 12215 ioc->scsih_cb_idx = scsih_cb_idx; 12216 ioc->config_cb_idx = config_cb_idx; 12217 ioc->tm_tr_cb_idx = tm_tr_cb_idx; 12218 ioc->tm_tr_volume_cb_idx = tm_tr_volume_cb_idx; 12219 ioc->tm_sas_control_cb_idx = tm_sas_control_cb_idx; 12220 ioc->logging_level = logging_level; 12221 ioc->schedule_dead_ioc_flush_running_cmds = &_scsih_flush_running_cmds; 12222 /* Host waits for minimum of six seconds */ 12223 ioc->max_shutdown_latency = IO_UNIT_CONTROL_SHUTDOWN_TIMEOUT; 12224 /* 12225 * Enable MEMORY MOVE support flag. 12226 */ 12227 ioc->drv_support_bitmap |= MPT_DRV_SUPPORT_BITMAP_MEMMOVE; 12228 /* Enable ADDITIONAL QUERY support flag. */ 12229 ioc->drv_support_bitmap |= MPT_DRV_SUPPORT_BITMAP_ADDNLQUERY; 12230 12231 ioc->enable_sdev_max_qd = enable_sdev_max_qd; 12232 12233 /* misc semaphores and spin locks */ 12234 mutex_init(&ioc->reset_in_progress_mutex); 12235 mutex_init(&ioc->hostdiag_unlock_mutex); 12236 /* initializing pci_access_mutex lock */ 12237 mutex_init(&ioc->pci_access_mutex); 12238 spin_lock_init(&ioc->ioc_reset_in_progress_lock); 12239 spin_lock_init(&ioc->scsi_lookup_lock); 12240 spin_lock_init(&ioc->sas_device_lock); 12241 spin_lock_init(&ioc->sas_node_lock); 12242 spin_lock_init(&ioc->fw_event_lock); 12243 spin_lock_init(&ioc->raid_device_lock); 12244 spin_lock_init(&ioc->pcie_device_lock); 12245 spin_lock_init(&ioc->diag_trigger_lock); 12246 12247 INIT_LIST_HEAD(&ioc->sas_device_list); 12248 INIT_LIST_HEAD(&ioc->sas_device_init_list); 12249 INIT_LIST_HEAD(&ioc->sas_expander_list); 12250 INIT_LIST_HEAD(&ioc->enclosure_list); 12251 INIT_LIST_HEAD(&ioc->pcie_device_list); 12252 INIT_LIST_HEAD(&ioc->pcie_device_init_list); 12253 INIT_LIST_HEAD(&ioc->fw_event_list); 12254 INIT_LIST_HEAD(&ioc->raid_device_list); 12255 INIT_LIST_HEAD(&ioc->sas_hba.sas_port_list); 12256 INIT_LIST_HEAD(&ioc->delayed_tr_list); 12257 INIT_LIST_HEAD(&ioc->delayed_sc_list); 12258 INIT_LIST_HEAD(&ioc->delayed_event_ack_list); 12259 INIT_LIST_HEAD(&ioc->delayed_tr_volume_list); 12260 INIT_LIST_HEAD(&ioc->reply_queue_list); 12261 INIT_LIST_HEAD(&ioc->port_table_list); 12262 12263 sprintf(ioc->name, "%s_cm%d", ioc->driver_name, ioc->id); 12264 12265 /* init shost parameters */ 12266 shost->max_cmd_len = 32; 12267 shost->max_lun = max_lun; 12268 shost->transportt = mpt3sas_transport_template; 12269 shost->unique_id = ioc->id; 12270 12271 if (ioc->is_mcpu_endpoint) { 12272 /* mCPU MPI support 64K max IO */ 12273 shost->max_sectors = 128; 12274 ioc_info(ioc, "The max_sectors value is set to %d\n", 12275 shost->max_sectors); 12276 } else { 12277 if (max_sectors != 0xFFFF) { 12278 if (max_sectors < 64) { 12279 shost->max_sectors = 64; 12280 ioc_warn(ioc, "Invalid value %d passed for max_sectors, range is 64 to 32767. Assigning value of 64.\n", 12281 max_sectors); 12282 } else if (max_sectors > 32767) { 12283 shost->max_sectors = 32767; 12284 ioc_warn(ioc, "Invalid value %d passed for max_sectors, range is 64 to 32767.Assigning default value of 32767.\n", 12285 max_sectors); 12286 } else { 12287 shost->max_sectors = max_sectors & 0xFFFE; 12288 ioc_info(ioc, "The max_sectors value is set to %d\n", 12289 shost->max_sectors); 12290 } 12291 } 12292 } 12293 /* register EEDP capabilities with SCSI layer */ 12294 if (prot_mask >= 0) 12295 scsi_host_set_prot(shost, (prot_mask & 0x07)); 12296 else 12297 scsi_host_set_prot(shost, SHOST_DIF_TYPE1_PROTECTION 12298 | SHOST_DIF_TYPE2_PROTECTION 12299 | SHOST_DIF_TYPE3_PROTECTION); 12300 12301 scsi_host_set_guard(shost, SHOST_DIX_GUARD_CRC); 12302 12303 /* event thread */ 12304 ioc->firmware_event_thread = alloc_ordered_workqueue( 12305 "fw_event_%s%d", 0, ioc->driver_name, ioc->id); 12306 if (!ioc->firmware_event_thread) { 12307 ioc_err(ioc, "failure at %s:%d/%s()!\n", 12308 __FILE__, __LINE__, __func__); 12309 rv = -ENODEV; 12310 goto out_thread_fail; 12311 } 12312 12313 shost->host_tagset = 0; 12314 12315 if (ioc->is_gen35_ioc && host_tagset_enable) 12316 shost->host_tagset = 1; 12317 12318 ioc->is_driver_loading = 1; 12319 if ((mpt3sas_base_attach(ioc))) { 12320 ioc_err(ioc, "failure at %s:%d/%s()!\n", 12321 __FILE__, __LINE__, __func__); 12322 rv = -ENODEV; 12323 goto out_attach_fail; 12324 } 12325 12326 if (ioc->is_warpdrive) { 12327 if (ioc->mfg_pg10_hide_flag == MFG_PAGE10_EXPOSE_ALL_DISKS) 12328 ioc->hide_drives = 0; 12329 else if (ioc->mfg_pg10_hide_flag == MFG_PAGE10_HIDE_ALL_DISKS) 12330 ioc->hide_drives = 1; 12331 else { 12332 if (mpt3sas_get_num_volumes(ioc)) 12333 ioc->hide_drives = 1; 12334 else 12335 ioc->hide_drives = 0; 12336 } 12337 } else 12338 ioc->hide_drives = 0; 12339 12340 shost->nr_hw_queues = 1; 12341 12342 if (shost->host_tagset) { 12343 shost->nr_hw_queues = 12344 ioc->reply_queue_count - ioc->high_iops_queues; 12345 12346 iopoll_q_count = 12347 ioc->reply_queue_count - ioc->iopoll_q_start_index; 12348 12349 shost->nr_maps = iopoll_q_count ? 3 : 1; 12350 12351 dev_info(&ioc->pdev->dev, 12352 "Max SCSIIO MPT commands: %d shared with nr_hw_queues = %d\n", 12353 shost->can_queue, shost->nr_hw_queues); 12354 } 12355 12356 rv = scsi_add_host(shost, &pdev->dev); 12357 if (rv) { 12358 ioc_err(ioc, "failure at %s:%d/%s()!\n", 12359 __FILE__, __LINE__, __func__); 12360 goto out_add_shost_fail; 12361 } 12362 12363 scsi_scan_host(shost); 12364 mpt3sas_setup_debugfs(ioc); 12365 return 0; 12366 out_add_shost_fail: 12367 mpt3sas_base_detach(ioc); 12368 out_attach_fail: 12369 destroy_workqueue(ioc->firmware_event_thread); 12370 out_thread_fail: 12371 spin_lock(&gioc_lock); 12372 list_del(&ioc->list); 12373 spin_unlock(&gioc_lock); 12374 scsi_host_put(shost); 12375 return rv; 12376 } 12377 12378 /** 12379 * scsih_suspend - power management suspend main entry point 12380 * @dev: Device struct 12381 * 12382 * Return: 0 success, anything else error. 12383 */ 12384 static int __maybe_unused 12385 scsih_suspend(struct device *dev) 12386 { 12387 struct pci_dev *pdev = to_pci_dev(dev); 12388 struct Scsi_Host *shost; 12389 struct MPT3SAS_ADAPTER *ioc; 12390 int rc; 12391 12392 rc = _scsih_get_shost_and_ioc(pdev, &shost, &ioc); 12393 if (rc) 12394 return rc; 12395 12396 mpt3sas_base_stop_watchdog(ioc); 12397 scsi_block_requests(shost); 12398 _scsih_nvme_shutdown(ioc); 12399 ioc_info(ioc, "pdev=0x%p, slot=%s, entering operating state\n", 12400 pdev, pci_name(pdev)); 12401 12402 mpt3sas_base_free_resources(ioc); 12403 return 0; 12404 } 12405 12406 /** 12407 * scsih_resume - power management resume main entry point 12408 * @dev: Device struct 12409 * 12410 * Return: 0 success, anything else error. 12411 */ 12412 static int __maybe_unused 12413 scsih_resume(struct device *dev) 12414 { 12415 struct pci_dev *pdev = to_pci_dev(dev); 12416 struct Scsi_Host *shost; 12417 struct MPT3SAS_ADAPTER *ioc; 12418 pci_power_t device_state = pdev->current_state; 12419 int r; 12420 12421 r = _scsih_get_shost_and_ioc(pdev, &shost, &ioc); 12422 if (r) 12423 return r; 12424 12425 ioc_info(ioc, "pdev=0x%p, slot=%s, previous operating state [D%d]\n", 12426 pdev, pci_name(pdev), device_state); 12427 12428 ioc->pdev = pdev; 12429 r = mpt3sas_base_map_resources(ioc); 12430 if (r) 12431 return r; 12432 ioc_info(ioc, "Issuing Hard Reset as part of OS Resume\n"); 12433 mpt3sas_base_hard_reset_handler(ioc, SOFT_RESET); 12434 scsi_unblock_requests(shost); 12435 mpt3sas_base_start_watchdog(ioc); 12436 return 0; 12437 } 12438 12439 /** 12440 * scsih_pci_error_detected - Called when a PCI error is detected. 12441 * @pdev: PCI device struct 12442 * @state: PCI channel state 12443 * 12444 * Description: Called when a PCI error is detected. 12445 * 12446 * Return: PCI_ERS_RESULT_NEED_RESET or PCI_ERS_RESULT_DISCONNECT. 12447 */ 12448 static pci_ers_result_t 12449 scsih_pci_error_detected(struct pci_dev *pdev, pci_channel_state_t state) 12450 { 12451 struct Scsi_Host *shost; 12452 struct MPT3SAS_ADAPTER *ioc; 12453 12454 if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc)) 12455 return PCI_ERS_RESULT_DISCONNECT; 12456 12457 ioc_info(ioc, "PCI error: detected callback, state(%d)!!\n", state); 12458 12459 switch (state) { 12460 case pci_channel_io_normal: 12461 return PCI_ERS_RESULT_CAN_RECOVER; 12462 case pci_channel_io_frozen: 12463 /* Fatal error, prepare for slot reset */ 12464 ioc->pci_error_recovery = 1; 12465 scsi_block_requests(ioc->shost); 12466 mpt3sas_base_stop_watchdog(ioc); 12467 mpt3sas_base_free_resources(ioc); 12468 return PCI_ERS_RESULT_NEED_RESET; 12469 case pci_channel_io_perm_failure: 12470 /* Permanent error, prepare for device removal */ 12471 ioc->pci_error_recovery = 1; 12472 mpt3sas_base_stop_watchdog(ioc); 12473 mpt3sas_base_pause_mq_polling(ioc); 12474 _scsih_flush_running_cmds(ioc); 12475 return PCI_ERS_RESULT_DISCONNECT; 12476 } 12477 return PCI_ERS_RESULT_NEED_RESET; 12478 } 12479 12480 /** 12481 * scsih_pci_slot_reset - Called when PCI slot has been reset. 12482 * @pdev: PCI device struct 12483 * 12484 * Description: This routine is called by the pci error recovery 12485 * code after the PCI slot has been reset, just before we 12486 * should resume normal operations. 12487 */ 12488 static pci_ers_result_t 12489 scsih_pci_slot_reset(struct pci_dev *pdev) 12490 { 12491 struct Scsi_Host *shost; 12492 struct MPT3SAS_ADAPTER *ioc; 12493 int rc; 12494 12495 if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc)) 12496 return PCI_ERS_RESULT_DISCONNECT; 12497 12498 ioc_info(ioc, "PCI error: slot reset callback!!\n"); 12499 12500 ioc->pci_error_recovery = 0; 12501 ioc->pdev = pdev; 12502 pci_restore_state(pdev); 12503 rc = mpt3sas_base_map_resources(ioc); 12504 if (rc) 12505 return PCI_ERS_RESULT_DISCONNECT; 12506 12507 ioc_info(ioc, "Issuing Hard Reset as part of PCI Slot Reset\n"); 12508 rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER); 12509 12510 ioc_warn(ioc, "hard reset: %s\n", 12511 (rc == 0) ? "success" : "failed"); 12512 12513 if (!rc) 12514 return PCI_ERS_RESULT_RECOVERED; 12515 else 12516 return PCI_ERS_RESULT_DISCONNECT; 12517 } 12518 12519 /** 12520 * scsih_pci_resume() - resume normal ops after PCI reset 12521 * @pdev: pointer to PCI device 12522 * 12523 * Called when the error recovery driver tells us that its 12524 * OK to resume normal operation. Use completion to allow 12525 * halted scsi ops to resume. 12526 */ 12527 static void 12528 scsih_pci_resume(struct pci_dev *pdev) 12529 { 12530 struct Scsi_Host *shost; 12531 struct MPT3SAS_ADAPTER *ioc; 12532 12533 if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc)) 12534 return; 12535 12536 ioc_info(ioc, "PCI error: resume callback!!\n"); 12537 12538 mpt3sas_base_start_watchdog(ioc); 12539 scsi_unblock_requests(ioc->shost); 12540 } 12541 12542 /** 12543 * scsih_pci_mmio_enabled - Enable MMIO and dump debug registers 12544 * @pdev: pointer to PCI device 12545 */ 12546 static pci_ers_result_t 12547 scsih_pci_mmio_enabled(struct pci_dev *pdev) 12548 { 12549 struct Scsi_Host *shost; 12550 struct MPT3SAS_ADAPTER *ioc; 12551 12552 if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc)) 12553 return PCI_ERS_RESULT_DISCONNECT; 12554 12555 ioc_info(ioc, "PCI error: mmio enabled callback!!\n"); 12556 12557 /* TODO - dump whatever for debugging purposes */ 12558 12559 /* This called only if scsih_pci_error_detected returns 12560 * PCI_ERS_RESULT_CAN_RECOVER. Read/write to the device still 12561 * works, no need to reset slot. 12562 */ 12563 return PCI_ERS_RESULT_RECOVERED; 12564 } 12565 12566 /* 12567 * The pci device ids are defined in mpi/mpi2_cnfg.h. 12568 */ 12569 static const struct pci_device_id mpt3sas_pci_table[] = { 12570 /* Spitfire ~ 2004 */ 12571 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2004, 12572 PCI_ANY_ID, PCI_ANY_ID }, 12573 /* Falcon ~ 2008 */ 12574 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2008, 12575 PCI_ANY_ID, PCI_ANY_ID }, 12576 /* Liberator ~ 2108 */ 12577 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_1, 12578 PCI_ANY_ID, PCI_ANY_ID }, 12579 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_2, 12580 PCI_ANY_ID, PCI_ANY_ID }, 12581 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_3, 12582 PCI_ANY_ID, PCI_ANY_ID }, 12583 /* Meteor ~ 2116 */ 12584 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2116_1, 12585 PCI_ANY_ID, PCI_ANY_ID }, 12586 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2116_2, 12587 PCI_ANY_ID, PCI_ANY_ID }, 12588 /* Thunderbolt ~ 2208 */ 12589 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_1, 12590 PCI_ANY_ID, PCI_ANY_ID }, 12591 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_2, 12592 PCI_ANY_ID, PCI_ANY_ID }, 12593 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_3, 12594 PCI_ANY_ID, PCI_ANY_ID }, 12595 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_4, 12596 PCI_ANY_ID, PCI_ANY_ID }, 12597 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_5, 12598 PCI_ANY_ID, PCI_ANY_ID }, 12599 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_6, 12600 PCI_ANY_ID, PCI_ANY_ID }, 12601 /* Mustang ~ 2308 */ 12602 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_1, 12603 PCI_ANY_ID, PCI_ANY_ID }, 12604 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_2, 12605 PCI_ANY_ID, PCI_ANY_ID }, 12606 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_3, 12607 PCI_ANY_ID, PCI_ANY_ID }, 12608 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP, 12609 PCI_ANY_ID, PCI_ANY_ID }, 12610 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1, 12611 PCI_ANY_ID, PCI_ANY_ID }, 12612 /* SSS6200 */ 12613 { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SSS6200, 12614 PCI_ANY_ID, PCI_ANY_ID }, 12615 /* Fury ~ 3004 and 3008 */ 12616 { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3004, 12617 PCI_ANY_ID, PCI_ANY_ID }, 12618 { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3008, 12619 PCI_ANY_ID, PCI_ANY_ID }, 12620 /* Invader ~ 3108 */ 12621 { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_1, 12622 PCI_ANY_ID, PCI_ANY_ID }, 12623 { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_2, 12624 PCI_ANY_ID, PCI_ANY_ID }, 12625 { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_5, 12626 PCI_ANY_ID, PCI_ANY_ID }, 12627 { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_6, 12628 PCI_ANY_ID, PCI_ANY_ID }, 12629 /* Cutlass ~ 3216 and 3224 */ 12630 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3216, 12631 PCI_ANY_ID, PCI_ANY_ID }, 12632 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3224, 12633 PCI_ANY_ID, PCI_ANY_ID }, 12634 /* Intruder ~ 3316 and 3324 */ 12635 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_1, 12636 PCI_ANY_ID, PCI_ANY_ID }, 12637 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_2, 12638 PCI_ANY_ID, PCI_ANY_ID }, 12639 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_3, 12640 PCI_ANY_ID, PCI_ANY_ID }, 12641 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_4, 12642 PCI_ANY_ID, PCI_ANY_ID }, 12643 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_1, 12644 PCI_ANY_ID, PCI_ANY_ID }, 12645 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_2, 12646 PCI_ANY_ID, PCI_ANY_ID }, 12647 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_3, 12648 PCI_ANY_ID, PCI_ANY_ID }, 12649 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_4, 12650 PCI_ANY_ID, PCI_ANY_ID }, 12651 /* Ventura, Crusader, Harpoon & Tomcat ~ 3516, 3416, 3508 & 3408*/ 12652 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3508, 12653 PCI_ANY_ID, PCI_ANY_ID }, 12654 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3508_1, 12655 PCI_ANY_ID, PCI_ANY_ID }, 12656 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3408, 12657 PCI_ANY_ID, PCI_ANY_ID }, 12658 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3516, 12659 PCI_ANY_ID, PCI_ANY_ID }, 12660 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3516_1, 12661 PCI_ANY_ID, PCI_ANY_ID }, 12662 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3416, 12663 PCI_ANY_ID, PCI_ANY_ID }, 12664 /* Mercator ~ 3616*/ 12665 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3616, 12666 PCI_ANY_ID, PCI_ANY_ID }, 12667 12668 /* Aero SI 0x00E1 Configurable Secure 12669 * 0x00E2 Hard Secure 12670 */ 12671 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_CFG_SEC_3916, 12672 PCI_ANY_ID, PCI_ANY_ID }, 12673 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_HARD_SEC_3916, 12674 PCI_ANY_ID, PCI_ANY_ID }, 12675 12676 /* 12677 * Aero SI –> 0x00E0 Invalid, 0x00E3 Tampered 12678 */ 12679 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID0_3916, 12680 PCI_ANY_ID, PCI_ANY_ID }, 12681 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID1_3916, 12682 PCI_ANY_ID, PCI_ANY_ID }, 12683 12684 /* Atlas PCIe Switch Management Port */ 12685 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_ATLAS_PCIe_SWITCH_DEVID, 12686 PCI_ANY_ID, PCI_ANY_ID }, 12687 12688 /* Sea SI 0x00E5 Configurable Secure 12689 * 0x00E6 Hard Secure 12690 */ 12691 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_CFG_SEC_3816, 12692 PCI_ANY_ID, PCI_ANY_ID }, 12693 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_HARD_SEC_3816, 12694 PCI_ANY_ID, PCI_ANY_ID }, 12695 12696 /* 12697 * ATTO Branded ExpressSAS H12xx GT 12698 */ 12699 { MPI2_MFGPAGE_VENDORID_ATTO, MPI26_MFGPAGE_DEVID_HARD_SEC_3816, 12700 PCI_ANY_ID, PCI_ANY_ID }, 12701 12702 /* 12703 * Sea SI –> 0x00E4 Invalid, 0x00E7 Tampered 12704 */ 12705 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID0_3816, 12706 PCI_ANY_ID, PCI_ANY_ID }, 12707 { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID1_3816, 12708 PCI_ANY_ID, PCI_ANY_ID }, 12709 12710 {0} /* Terminating entry */ 12711 }; 12712 MODULE_DEVICE_TABLE(pci, mpt3sas_pci_table); 12713 12714 static struct pci_error_handlers _mpt3sas_err_handler = { 12715 .error_detected = scsih_pci_error_detected, 12716 .mmio_enabled = scsih_pci_mmio_enabled, 12717 .slot_reset = scsih_pci_slot_reset, 12718 .resume = scsih_pci_resume, 12719 }; 12720 12721 static SIMPLE_DEV_PM_OPS(scsih_pm_ops, scsih_suspend, scsih_resume); 12722 12723 static struct pci_driver mpt3sas_driver = { 12724 .name = MPT3SAS_DRIVER_NAME, 12725 .id_table = mpt3sas_pci_table, 12726 .probe = _scsih_probe, 12727 .remove = scsih_remove, 12728 .shutdown = scsih_shutdown, 12729 .err_handler = &_mpt3sas_err_handler, 12730 .driver.pm = &scsih_pm_ops, 12731 }; 12732 12733 /** 12734 * scsih_init - main entry point for this driver. 12735 * 12736 * Return: 0 success, anything else error. 12737 */ 12738 static int 12739 scsih_init(void) 12740 { 12741 mpt2_ids = 0; 12742 mpt3_ids = 0; 12743 12744 mpt3sas_base_initialize_callback_handler(); 12745 12746 /* queuecommand callback hander */ 12747 scsi_io_cb_idx = mpt3sas_base_register_callback_handler(_scsih_io_done); 12748 12749 /* task management callback handler */ 12750 tm_cb_idx = mpt3sas_base_register_callback_handler(_scsih_tm_done); 12751 12752 /* base internal commands callback handler */ 12753 base_cb_idx = mpt3sas_base_register_callback_handler(mpt3sas_base_done); 12754 port_enable_cb_idx = mpt3sas_base_register_callback_handler( 12755 mpt3sas_port_enable_done); 12756 12757 /* transport internal commands callback handler */ 12758 transport_cb_idx = mpt3sas_base_register_callback_handler( 12759 mpt3sas_transport_done); 12760 12761 /* scsih internal commands callback handler */ 12762 scsih_cb_idx = mpt3sas_base_register_callback_handler(_scsih_done); 12763 12764 /* configuration page API internal commands callback handler */ 12765 config_cb_idx = mpt3sas_base_register_callback_handler( 12766 mpt3sas_config_done); 12767 12768 /* ctl module callback handler */ 12769 ctl_cb_idx = mpt3sas_base_register_callback_handler(mpt3sas_ctl_done); 12770 12771 tm_tr_cb_idx = mpt3sas_base_register_callback_handler( 12772 _scsih_tm_tr_complete); 12773 12774 tm_tr_volume_cb_idx = mpt3sas_base_register_callback_handler( 12775 _scsih_tm_volume_tr_complete); 12776 12777 tm_sas_control_cb_idx = mpt3sas_base_register_callback_handler( 12778 _scsih_sas_control_complete); 12779 12780 mpt3sas_init_debugfs(); 12781 return 0; 12782 } 12783 12784 /** 12785 * scsih_exit - exit point for this driver (when it is a module). 12786 * 12787 * Return: 0 success, anything else error. 12788 */ 12789 static void 12790 scsih_exit(void) 12791 { 12792 12793 mpt3sas_base_release_callback_handler(scsi_io_cb_idx); 12794 mpt3sas_base_release_callback_handler(tm_cb_idx); 12795 mpt3sas_base_release_callback_handler(base_cb_idx); 12796 mpt3sas_base_release_callback_handler(port_enable_cb_idx); 12797 mpt3sas_base_release_callback_handler(transport_cb_idx); 12798 mpt3sas_base_release_callback_handler(scsih_cb_idx); 12799 mpt3sas_base_release_callback_handler(config_cb_idx); 12800 mpt3sas_base_release_callback_handler(ctl_cb_idx); 12801 12802 mpt3sas_base_release_callback_handler(tm_tr_cb_idx); 12803 mpt3sas_base_release_callback_handler(tm_tr_volume_cb_idx); 12804 mpt3sas_base_release_callback_handler(tm_sas_control_cb_idx); 12805 12806 /* raid transport support */ 12807 if (hbas_to_enumerate != 1) 12808 raid_class_release(mpt3sas_raid_template); 12809 if (hbas_to_enumerate != 2) 12810 raid_class_release(mpt2sas_raid_template); 12811 sas_release_transport(mpt3sas_transport_template); 12812 mpt3sas_exit_debugfs(); 12813 } 12814 12815 /** 12816 * _mpt3sas_init - main entry point for this driver. 12817 * 12818 * Return: 0 success, anything else error. 12819 */ 12820 static int __init 12821 _mpt3sas_init(void) 12822 { 12823 int error; 12824 12825 pr_info("%s version %s loaded\n", MPT3SAS_DRIVER_NAME, 12826 MPT3SAS_DRIVER_VERSION); 12827 12828 mpt3sas_transport_template = 12829 sas_attach_transport(&mpt3sas_transport_functions); 12830 if (!mpt3sas_transport_template) 12831 return -ENODEV; 12832 12833 /* No need attach mpt3sas raid functions template 12834 * if hbas_to_enumarate value is one. 12835 */ 12836 if (hbas_to_enumerate != 1) { 12837 mpt3sas_raid_template = 12838 raid_class_attach(&mpt3sas_raid_functions); 12839 if (!mpt3sas_raid_template) { 12840 sas_release_transport(mpt3sas_transport_template); 12841 return -ENODEV; 12842 } 12843 } 12844 12845 /* No need to attach mpt2sas raid functions template 12846 * if hbas_to_enumarate value is two 12847 */ 12848 if (hbas_to_enumerate != 2) { 12849 mpt2sas_raid_template = 12850 raid_class_attach(&mpt2sas_raid_functions); 12851 if (!mpt2sas_raid_template) { 12852 sas_release_transport(mpt3sas_transport_template); 12853 return -ENODEV; 12854 } 12855 } 12856 12857 error = scsih_init(); 12858 if (error) { 12859 scsih_exit(); 12860 return error; 12861 } 12862 12863 mpt3sas_ctl_init(hbas_to_enumerate); 12864 12865 error = pci_register_driver(&mpt3sas_driver); 12866 if (error) { 12867 mpt3sas_ctl_exit(hbas_to_enumerate); 12868 scsih_exit(); 12869 } 12870 12871 return error; 12872 } 12873 12874 /** 12875 * _mpt3sas_exit - exit point for this driver (when it is a module). 12876 * 12877 */ 12878 static void __exit 12879 _mpt3sas_exit(void) 12880 { 12881 pr_info("mpt3sas version %s unloading\n", 12882 MPT3SAS_DRIVER_VERSION); 12883 12884 pci_unregister_driver(&mpt3sas_driver); 12885 12886 mpt3sas_ctl_exit(hbas_to_enumerate); 12887 12888 scsih_exit(); 12889 } 12890 12891 module_init(_mpt3sas_init); 12892 module_exit(_mpt3sas_exit); 12893