1 /* 2 * zfcp device driver 3 * 4 * Implementation of FSF commands. 5 * 6 * Copyright IBM Corp. 2002, 2010 7 */ 8 9 #define KMSG_COMPONENT "zfcp" 10 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt 11 12 #include <linux/blktrace_api.h> 13 #include <linux/slab.h> 14 #include <scsi/fc/fc_els.h> 15 #include "zfcp_ext.h" 16 #include "zfcp_fc.h" 17 #include "zfcp_dbf.h" 18 #include "zfcp_qdio.h" 19 #include "zfcp_reqlist.h" 20 21 struct kmem_cache *zfcp_fsf_qtcb_cache; 22 23 static void zfcp_fsf_request_timeout_handler(unsigned long data) 24 { 25 struct zfcp_adapter *adapter = (struct zfcp_adapter *) data; 26 zfcp_qdio_siosl(adapter); 27 zfcp_erp_adapter_reopen(adapter, ZFCP_STATUS_COMMON_ERP_FAILED, 28 "fsrth_1"); 29 } 30 31 static void zfcp_fsf_start_timer(struct zfcp_fsf_req *fsf_req, 32 unsigned long timeout) 33 { 34 fsf_req->timer.function = zfcp_fsf_request_timeout_handler; 35 fsf_req->timer.data = (unsigned long) fsf_req->adapter; 36 fsf_req->timer.expires = jiffies + timeout; 37 add_timer(&fsf_req->timer); 38 } 39 40 static void zfcp_fsf_start_erp_timer(struct zfcp_fsf_req *fsf_req) 41 { 42 BUG_ON(!fsf_req->erp_action); 43 fsf_req->timer.function = zfcp_erp_timeout_handler; 44 fsf_req->timer.data = (unsigned long) fsf_req->erp_action; 45 fsf_req->timer.expires = jiffies + 30 * HZ; 46 add_timer(&fsf_req->timer); 47 } 48 49 /* association between FSF command and FSF QTCB type */ 50 static u32 fsf_qtcb_type[] = { 51 [FSF_QTCB_FCP_CMND] = FSF_IO_COMMAND, 52 [FSF_QTCB_ABORT_FCP_CMND] = FSF_SUPPORT_COMMAND, 53 [FSF_QTCB_OPEN_PORT_WITH_DID] = FSF_SUPPORT_COMMAND, 54 [FSF_QTCB_OPEN_LUN] = FSF_SUPPORT_COMMAND, 55 [FSF_QTCB_CLOSE_LUN] = FSF_SUPPORT_COMMAND, 56 [FSF_QTCB_CLOSE_PORT] = FSF_SUPPORT_COMMAND, 57 [FSF_QTCB_CLOSE_PHYSICAL_PORT] = FSF_SUPPORT_COMMAND, 58 [FSF_QTCB_SEND_ELS] = FSF_SUPPORT_COMMAND, 59 [FSF_QTCB_SEND_GENERIC] = FSF_SUPPORT_COMMAND, 60 [FSF_QTCB_EXCHANGE_CONFIG_DATA] = FSF_CONFIG_COMMAND, 61 [FSF_QTCB_EXCHANGE_PORT_DATA] = FSF_PORT_COMMAND, 62 [FSF_QTCB_DOWNLOAD_CONTROL_FILE] = FSF_SUPPORT_COMMAND, 63 [FSF_QTCB_UPLOAD_CONTROL_FILE] = FSF_SUPPORT_COMMAND 64 }; 65 66 static void zfcp_fsf_class_not_supp(struct zfcp_fsf_req *req) 67 { 68 dev_err(&req->adapter->ccw_device->dev, "FCP device not " 69 "operational because of an unsupported FC class\n"); 70 zfcp_erp_adapter_shutdown(req->adapter, 0, "fscns_1"); 71 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 72 } 73 74 /** 75 * zfcp_fsf_req_free - free memory used by fsf request 76 * @fsf_req: pointer to struct zfcp_fsf_req 77 */ 78 void zfcp_fsf_req_free(struct zfcp_fsf_req *req) 79 { 80 if (likely(req->pool)) { 81 if (likely(req->qtcb)) 82 mempool_free(req->qtcb, req->adapter->pool.qtcb_pool); 83 mempool_free(req, req->pool); 84 return; 85 } 86 87 if (likely(req->qtcb)) 88 kmem_cache_free(zfcp_fsf_qtcb_cache, req->qtcb); 89 kfree(req); 90 } 91 92 static void zfcp_fsf_status_read_port_closed(struct zfcp_fsf_req *req) 93 { 94 unsigned long flags; 95 struct fsf_status_read_buffer *sr_buf = req->data; 96 struct zfcp_adapter *adapter = req->adapter; 97 struct zfcp_port *port; 98 int d_id = ntoh24(sr_buf->d_id); 99 100 read_lock_irqsave(&adapter->port_list_lock, flags); 101 list_for_each_entry(port, &adapter->port_list, list) 102 if (port->d_id == d_id) { 103 zfcp_erp_port_reopen(port, 0, "fssrpc1"); 104 break; 105 } 106 read_unlock_irqrestore(&adapter->port_list_lock, flags); 107 } 108 109 static void zfcp_fsf_link_down_info_eval(struct zfcp_fsf_req *req, 110 struct fsf_link_down_info *link_down) 111 { 112 struct zfcp_adapter *adapter = req->adapter; 113 114 if (atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED) 115 return; 116 117 atomic_set_mask(ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED, &adapter->status); 118 119 zfcp_scsi_schedule_rports_block(adapter); 120 121 if (!link_down) 122 goto out; 123 124 switch (link_down->error_code) { 125 case FSF_PSQ_LINK_NO_LIGHT: 126 dev_warn(&req->adapter->ccw_device->dev, 127 "There is no light signal from the local " 128 "fibre channel cable\n"); 129 break; 130 case FSF_PSQ_LINK_WRAP_PLUG: 131 dev_warn(&req->adapter->ccw_device->dev, 132 "There is a wrap plug instead of a fibre " 133 "channel cable\n"); 134 break; 135 case FSF_PSQ_LINK_NO_FCP: 136 dev_warn(&req->adapter->ccw_device->dev, 137 "The adjacent fibre channel node does not " 138 "support FCP\n"); 139 break; 140 case FSF_PSQ_LINK_FIRMWARE_UPDATE: 141 dev_warn(&req->adapter->ccw_device->dev, 142 "The FCP device is suspended because of a " 143 "firmware update\n"); 144 break; 145 case FSF_PSQ_LINK_INVALID_WWPN: 146 dev_warn(&req->adapter->ccw_device->dev, 147 "The FCP device detected a WWPN that is " 148 "duplicate or not valid\n"); 149 break; 150 case FSF_PSQ_LINK_NO_NPIV_SUPPORT: 151 dev_warn(&req->adapter->ccw_device->dev, 152 "The fibre channel fabric does not support NPIV\n"); 153 break; 154 case FSF_PSQ_LINK_NO_FCP_RESOURCES: 155 dev_warn(&req->adapter->ccw_device->dev, 156 "The FCP adapter cannot support more NPIV ports\n"); 157 break; 158 case FSF_PSQ_LINK_NO_FABRIC_RESOURCES: 159 dev_warn(&req->adapter->ccw_device->dev, 160 "The adjacent switch cannot support " 161 "more NPIV ports\n"); 162 break; 163 case FSF_PSQ_LINK_FABRIC_LOGIN_UNABLE: 164 dev_warn(&req->adapter->ccw_device->dev, 165 "The FCP adapter could not log in to the " 166 "fibre channel fabric\n"); 167 break; 168 case FSF_PSQ_LINK_WWPN_ASSIGNMENT_CORRUPTED: 169 dev_warn(&req->adapter->ccw_device->dev, 170 "The WWPN assignment file on the FCP adapter " 171 "has been damaged\n"); 172 break; 173 case FSF_PSQ_LINK_MODE_TABLE_CURRUPTED: 174 dev_warn(&req->adapter->ccw_device->dev, 175 "The mode table on the FCP adapter " 176 "has been damaged\n"); 177 break; 178 case FSF_PSQ_LINK_NO_WWPN_ASSIGNMENT: 179 dev_warn(&req->adapter->ccw_device->dev, 180 "All NPIV ports on the FCP adapter have " 181 "been assigned\n"); 182 break; 183 default: 184 dev_warn(&req->adapter->ccw_device->dev, 185 "The link between the FCP adapter and " 186 "the FC fabric is down\n"); 187 } 188 out: 189 zfcp_erp_set_adapter_status(adapter, ZFCP_STATUS_COMMON_ERP_FAILED); 190 } 191 192 static void zfcp_fsf_status_read_link_down(struct zfcp_fsf_req *req) 193 { 194 struct fsf_status_read_buffer *sr_buf = req->data; 195 struct fsf_link_down_info *ldi = 196 (struct fsf_link_down_info *) &sr_buf->payload; 197 198 switch (sr_buf->status_subtype) { 199 case FSF_STATUS_READ_SUB_NO_PHYSICAL_LINK: 200 zfcp_fsf_link_down_info_eval(req, ldi); 201 break; 202 case FSF_STATUS_READ_SUB_FDISC_FAILED: 203 zfcp_fsf_link_down_info_eval(req, ldi); 204 break; 205 case FSF_STATUS_READ_SUB_FIRMWARE_UPDATE: 206 zfcp_fsf_link_down_info_eval(req, NULL); 207 }; 208 } 209 210 static void zfcp_fsf_status_read_handler(struct zfcp_fsf_req *req) 211 { 212 struct zfcp_adapter *adapter = req->adapter; 213 struct fsf_status_read_buffer *sr_buf = req->data; 214 215 if (req->status & ZFCP_STATUS_FSFREQ_DISMISSED) { 216 zfcp_dbf_hba_fsf_uss("fssrh_1", req); 217 mempool_free(virt_to_page(sr_buf), adapter->pool.sr_data); 218 zfcp_fsf_req_free(req); 219 return; 220 } 221 222 zfcp_dbf_hba_fsf_uss("fssrh_4", req); 223 224 switch (sr_buf->status_type) { 225 case FSF_STATUS_READ_PORT_CLOSED: 226 zfcp_fsf_status_read_port_closed(req); 227 break; 228 case FSF_STATUS_READ_INCOMING_ELS: 229 zfcp_fc_incoming_els(req); 230 break; 231 case FSF_STATUS_READ_SENSE_DATA_AVAIL: 232 break; 233 case FSF_STATUS_READ_BIT_ERROR_THRESHOLD: 234 dev_warn(&adapter->ccw_device->dev, 235 "The error threshold for checksum statistics " 236 "has been exceeded\n"); 237 zfcp_dbf_hba_bit_err("fssrh_3", req); 238 break; 239 case FSF_STATUS_READ_LINK_DOWN: 240 zfcp_fsf_status_read_link_down(req); 241 zfcp_fc_enqueue_event(adapter, FCH_EVT_LINKDOWN, 0); 242 break; 243 case FSF_STATUS_READ_LINK_UP: 244 dev_info(&adapter->ccw_device->dev, 245 "The local link has been restored\n"); 246 /* All ports should be marked as ready to run again */ 247 zfcp_erp_set_adapter_status(adapter, 248 ZFCP_STATUS_COMMON_RUNNING); 249 zfcp_erp_adapter_reopen(adapter, 250 ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED | 251 ZFCP_STATUS_COMMON_ERP_FAILED, 252 "fssrh_2"); 253 zfcp_fc_enqueue_event(adapter, FCH_EVT_LINKUP, 0); 254 255 break; 256 case FSF_STATUS_READ_NOTIFICATION_LOST: 257 if (sr_buf->status_subtype & FSF_STATUS_READ_SUB_ACT_UPDATED) 258 zfcp_cfdc_adapter_access_changed(adapter); 259 if (sr_buf->status_subtype & FSF_STATUS_READ_SUB_INCOMING_ELS) 260 zfcp_fc_conditional_port_scan(adapter); 261 break; 262 case FSF_STATUS_READ_CFDC_UPDATED: 263 zfcp_cfdc_adapter_access_changed(adapter); 264 break; 265 case FSF_STATUS_READ_FEATURE_UPDATE_ALERT: 266 adapter->adapter_features = sr_buf->payload.word[0]; 267 break; 268 } 269 270 mempool_free(virt_to_page(sr_buf), adapter->pool.sr_data); 271 zfcp_fsf_req_free(req); 272 273 atomic_inc(&adapter->stat_miss); 274 queue_work(adapter->work_queue, &adapter->stat_work); 275 } 276 277 static void zfcp_fsf_fsfstatus_qual_eval(struct zfcp_fsf_req *req) 278 { 279 switch (req->qtcb->header.fsf_status_qual.word[0]) { 280 case FSF_SQ_FCP_RSP_AVAILABLE: 281 case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: 282 case FSF_SQ_NO_RETRY_POSSIBLE: 283 case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: 284 return; 285 case FSF_SQ_COMMAND_ABORTED: 286 break; 287 case FSF_SQ_NO_RECOM: 288 dev_err(&req->adapter->ccw_device->dev, 289 "The FCP adapter reported a problem " 290 "that cannot be recovered\n"); 291 zfcp_qdio_siosl(req->adapter); 292 zfcp_erp_adapter_shutdown(req->adapter, 0, "fsfsqe1"); 293 break; 294 } 295 /* all non-return stats set FSFREQ_ERROR*/ 296 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 297 } 298 299 static void zfcp_fsf_fsfstatus_eval(struct zfcp_fsf_req *req) 300 { 301 if (unlikely(req->status & ZFCP_STATUS_FSFREQ_ERROR)) 302 return; 303 304 switch (req->qtcb->header.fsf_status) { 305 case FSF_UNKNOWN_COMMAND: 306 dev_err(&req->adapter->ccw_device->dev, 307 "The FCP adapter does not recognize the command 0x%x\n", 308 req->qtcb->header.fsf_command); 309 zfcp_erp_adapter_shutdown(req->adapter, 0, "fsfse_1"); 310 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 311 break; 312 case FSF_ADAPTER_STATUS_AVAILABLE: 313 zfcp_fsf_fsfstatus_qual_eval(req); 314 break; 315 } 316 } 317 318 static void zfcp_fsf_protstatus_eval(struct zfcp_fsf_req *req) 319 { 320 struct zfcp_adapter *adapter = req->adapter; 321 struct fsf_qtcb *qtcb = req->qtcb; 322 union fsf_prot_status_qual *psq = &qtcb->prefix.prot_status_qual; 323 324 zfcp_dbf_hba_fsf_response(req); 325 326 if (req->status & ZFCP_STATUS_FSFREQ_DISMISSED) { 327 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 328 return; 329 } 330 331 switch (qtcb->prefix.prot_status) { 332 case FSF_PROT_GOOD: 333 case FSF_PROT_FSF_STATUS_PRESENTED: 334 return; 335 case FSF_PROT_QTCB_VERSION_ERROR: 336 dev_err(&adapter->ccw_device->dev, 337 "QTCB version 0x%x not supported by FCP adapter " 338 "(0x%x to 0x%x)\n", FSF_QTCB_CURRENT_VERSION, 339 psq->word[0], psq->word[1]); 340 zfcp_erp_adapter_shutdown(adapter, 0, "fspse_1"); 341 break; 342 case FSF_PROT_ERROR_STATE: 343 case FSF_PROT_SEQ_NUMB_ERROR: 344 zfcp_erp_adapter_reopen(adapter, 0, "fspse_2"); 345 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 346 break; 347 case FSF_PROT_UNSUPP_QTCB_TYPE: 348 dev_err(&adapter->ccw_device->dev, 349 "The QTCB type is not supported by the FCP adapter\n"); 350 zfcp_erp_adapter_shutdown(adapter, 0, "fspse_3"); 351 break; 352 case FSF_PROT_HOST_CONNECTION_INITIALIZING: 353 atomic_set_mask(ZFCP_STATUS_ADAPTER_HOST_CON_INIT, 354 &adapter->status); 355 break; 356 case FSF_PROT_DUPLICATE_REQUEST_ID: 357 dev_err(&adapter->ccw_device->dev, 358 "0x%Lx is an ambiguous request identifier\n", 359 (unsigned long long)qtcb->bottom.support.req_handle); 360 zfcp_erp_adapter_shutdown(adapter, 0, "fspse_4"); 361 break; 362 case FSF_PROT_LINK_DOWN: 363 zfcp_fsf_link_down_info_eval(req, &psq->link_down_info); 364 /* go through reopen to flush pending requests */ 365 zfcp_erp_adapter_reopen(adapter, 0, "fspse_6"); 366 break; 367 case FSF_PROT_REEST_QUEUE: 368 /* All ports should be marked as ready to run again */ 369 zfcp_erp_set_adapter_status(adapter, 370 ZFCP_STATUS_COMMON_RUNNING); 371 zfcp_erp_adapter_reopen(adapter, 372 ZFCP_STATUS_ADAPTER_LINK_UNPLUGGED | 373 ZFCP_STATUS_COMMON_ERP_FAILED, 374 "fspse_8"); 375 break; 376 default: 377 dev_err(&adapter->ccw_device->dev, 378 "0x%x is not a valid transfer protocol status\n", 379 qtcb->prefix.prot_status); 380 zfcp_qdio_siosl(adapter); 381 zfcp_erp_adapter_shutdown(adapter, 0, "fspse_9"); 382 } 383 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 384 } 385 386 /** 387 * zfcp_fsf_req_complete - process completion of a FSF request 388 * @fsf_req: The FSF request that has been completed. 389 * 390 * When a request has been completed either from the FCP adapter, 391 * or it has been dismissed due to a queue shutdown, this function 392 * is called to process the completion status and trigger further 393 * events related to the FSF request. 394 */ 395 static void zfcp_fsf_req_complete(struct zfcp_fsf_req *req) 396 { 397 if (unlikely(req->fsf_command == FSF_QTCB_UNSOLICITED_STATUS)) { 398 zfcp_fsf_status_read_handler(req); 399 return; 400 } 401 402 del_timer(&req->timer); 403 zfcp_fsf_protstatus_eval(req); 404 zfcp_fsf_fsfstatus_eval(req); 405 req->handler(req); 406 407 if (req->erp_action) 408 zfcp_erp_notify(req->erp_action, 0); 409 410 if (likely(req->status & ZFCP_STATUS_FSFREQ_CLEANUP)) 411 zfcp_fsf_req_free(req); 412 else 413 complete(&req->completion); 414 } 415 416 /** 417 * zfcp_fsf_req_dismiss_all - dismiss all fsf requests 418 * @adapter: pointer to struct zfcp_adapter 419 * 420 * Never ever call this without shutting down the adapter first. 421 * Otherwise the adapter would continue using and corrupting s390 storage. 422 * Included BUG_ON() call to ensure this is done. 423 * ERP is supposed to be the only user of this function. 424 */ 425 void zfcp_fsf_req_dismiss_all(struct zfcp_adapter *adapter) 426 { 427 struct zfcp_fsf_req *req, *tmp; 428 LIST_HEAD(remove_queue); 429 430 BUG_ON(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP); 431 zfcp_reqlist_move(adapter->req_list, &remove_queue); 432 433 list_for_each_entry_safe(req, tmp, &remove_queue, list) { 434 list_del(&req->list); 435 req->status |= ZFCP_STATUS_FSFREQ_DISMISSED; 436 zfcp_fsf_req_complete(req); 437 } 438 } 439 440 #define ZFCP_FSF_PORTSPEED_1GBIT (1 << 0) 441 #define ZFCP_FSF_PORTSPEED_2GBIT (1 << 1) 442 #define ZFCP_FSF_PORTSPEED_4GBIT (1 << 2) 443 #define ZFCP_FSF_PORTSPEED_10GBIT (1 << 3) 444 #define ZFCP_FSF_PORTSPEED_8GBIT (1 << 4) 445 #define ZFCP_FSF_PORTSPEED_16GBIT (1 << 5) 446 #define ZFCP_FSF_PORTSPEED_NOT_NEGOTIATED (1 << 15) 447 448 static u32 zfcp_fsf_convert_portspeed(u32 fsf_speed) 449 { 450 u32 fdmi_speed = 0; 451 if (fsf_speed & ZFCP_FSF_PORTSPEED_1GBIT) 452 fdmi_speed |= FC_PORTSPEED_1GBIT; 453 if (fsf_speed & ZFCP_FSF_PORTSPEED_2GBIT) 454 fdmi_speed |= FC_PORTSPEED_2GBIT; 455 if (fsf_speed & ZFCP_FSF_PORTSPEED_4GBIT) 456 fdmi_speed |= FC_PORTSPEED_4GBIT; 457 if (fsf_speed & ZFCP_FSF_PORTSPEED_10GBIT) 458 fdmi_speed |= FC_PORTSPEED_10GBIT; 459 if (fsf_speed & ZFCP_FSF_PORTSPEED_8GBIT) 460 fdmi_speed |= FC_PORTSPEED_8GBIT; 461 if (fsf_speed & ZFCP_FSF_PORTSPEED_16GBIT) 462 fdmi_speed |= FC_PORTSPEED_16GBIT; 463 if (fsf_speed & ZFCP_FSF_PORTSPEED_NOT_NEGOTIATED) 464 fdmi_speed |= FC_PORTSPEED_NOT_NEGOTIATED; 465 return fdmi_speed; 466 } 467 468 static int zfcp_fsf_exchange_config_evaluate(struct zfcp_fsf_req *req) 469 { 470 struct fsf_qtcb_bottom_config *bottom = &req->qtcb->bottom.config; 471 struct zfcp_adapter *adapter = req->adapter; 472 struct Scsi_Host *shost = adapter->scsi_host; 473 struct fc_els_flogi *nsp, *plogi; 474 475 /* adjust pointers for missing command code */ 476 nsp = (struct fc_els_flogi *) ((u8 *)&bottom->nport_serv_param 477 - sizeof(u32)); 478 plogi = (struct fc_els_flogi *) ((u8 *)&bottom->plogi_payload 479 - sizeof(u32)); 480 481 if (req->data) 482 memcpy(req->data, bottom, sizeof(*bottom)); 483 484 fc_host_port_name(shost) = nsp->fl_wwpn; 485 fc_host_node_name(shost) = nsp->fl_wwnn; 486 fc_host_port_id(shost) = ntoh24(bottom->s_id); 487 fc_host_speed(shost) = 488 zfcp_fsf_convert_portspeed(bottom->fc_link_speed); 489 fc_host_supported_classes(shost) = FC_COS_CLASS2 | FC_COS_CLASS3; 490 491 adapter->hydra_version = bottom->adapter_type; 492 adapter->timer_ticks = bottom->timer_interval & ZFCP_FSF_TIMER_INT_MASK; 493 adapter->stat_read_buf_num = max(bottom->status_read_buf_num, 494 (u16)FSF_STATUS_READS_RECOM); 495 496 if (fc_host_permanent_port_name(shost) == -1) 497 fc_host_permanent_port_name(shost) = fc_host_port_name(shost); 498 499 switch (bottom->fc_topology) { 500 case FSF_TOPO_P2P: 501 adapter->peer_d_id = ntoh24(bottom->peer_d_id); 502 adapter->peer_wwpn = plogi->fl_wwpn; 503 adapter->peer_wwnn = plogi->fl_wwnn; 504 fc_host_port_type(shost) = FC_PORTTYPE_PTP; 505 break; 506 case FSF_TOPO_FABRIC: 507 fc_host_port_type(shost) = FC_PORTTYPE_NPORT; 508 break; 509 case FSF_TOPO_AL: 510 fc_host_port_type(shost) = FC_PORTTYPE_NLPORT; 511 /* fall through */ 512 default: 513 dev_err(&adapter->ccw_device->dev, 514 "Unknown or unsupported arbitrated loop " 515 "fibre channel topology detected\n"); 516 zfcp_erp_adapter_shutdown(adapter, 0, "fsece_1"); 517 return -EIO; 518 } 519 520 zfcp_scsi_set_prot(adapter); 521 522 return 0; 523 } 524 525 static void zfcp_fsf_exchange_config_data_handler(struct zfcp_fsf_req *req) 526 { 527 struct zfcp_adapter *adapter = req->adapter; 528 struct fsf_qtcb *qtcb = req->qtcb; 529 struct fsf_qtcb_bottom_config *bottom = &qtcb->bottom.config; 530 struct Scsi_Host *shost = adapter->scsi_host; 531 532 if (req->status & ZFCP_STATUS_FSFREQ_ERROR) 533 return; 534 535 adapter->fsf_lic_version = bottom->lic_version; 536 adapter->adapter_features = bottom->adapter_features; 537 adapter->connection_features = bottom->connection_features; 538 adapter->peer_wwpn = 0; 539 adapter->peer_wwnn = 0; 540 adapter->peer_d_id = 0; 541 542 switch (qtcb->header.fsf_status) { 543 case FSF_GOOD: 544 if (zfcp_fsf_exchange_config_evaluate(req)) 545 return; 546 547 if (bottom->max_qtcb_size < sizeof(struct fsf_qtcb)) { 548 dev_err(&adapter->ccw_device->dev, 549 "FCP adapter maximum QTCB size (%d bytes) " 550 "is too small\n", 551 bottom->max_qtcb_size); 552 zfcp_erp_adapter_shutdown(adapter, 0, "fsecdh1"); 553 return; 554 } 555 atomic_set_mask(ZFCP_STATUS_ADAPTER_XCONFIG_OK, 556 &adapter->status); 557 break; 558 case FSF_EXCHANGE_CONFIG_DATA_INCOMPLETE: 559 fc_host_node_name(shost) = 0; 560 fc_host_port_name(shost) = 0; 561 fc_host_port_id(shost) = 0; 562 fc_host_speed(shost) = FC_PORTSPEED_UNKNOWN; 563 fc_host_port_type(shost) = FC_PORTTYPE_UNKNOWN; 564 adapter->hydra_version = 0; 565 566 /* avoids adapter shutdown to be able to recognize 567 * events such as LINK UP */ 568 atomic_set_mask(ZFCP_STATUS_ADAPTER_XCONFIG_OK, 569 &adapter->status); 570 zfcp_fsf_link_down_info_eval(req, 571 &qtcb->header.fsf_status_qual.link_down_info); 572 break; 573 default: 574 zfcp_erp_adapter_shutdown(adapter, 0, "fsecdh3"); 575 return; 576 } 577 578 if (adapter->adapter_features & FSF_FEATURE_HBAAPI_MANAGEMENT) { 579 adapter->hardware_version = bottom->hardware_version; 580 memcpy(fc_host_serial_number(shost), bottom->serial_number, 581 min(FC_SERIAL_NUMBER_SIZE, 17)); 582 EBCASC(fc_host_serial_number(shost), 583 min(FC_SERIAL_NUMBER_SIZE, 17)); 584 } 585 586 if (FSF_QTCB_CURRENT_VERSION < bottom->low_qtcb_version) { 587 dev_err(&adapter->ccw_device->dev, 588 "The FCP adapter only supports newer " 589 "control block versions\n"); 590 zfcp_erp_adapter_shutdown(adapter, 0, "fsecdh4"); 591 return; 592 } 593 if (FSF_QTCB_CURRENT_VERSION > bottom->high_qtcb_version) { 594 dev_err(&adapter->ccw_device->dev, 595 "The FCP adapter only supports older " 596 "control block versions\n"); 597 zfcp_erp_adapter_shutdown(adapter, 0, "fsecdh5"); 598 } 599 } 600 601 static void zfcp_fsf_exchange_port_evaluate(struct zfcp_fsf_req *req) 602 { 603 struct zfcp_adapter *adapter = req->adapter; 604 struct fsf_qtcb_bottom_port *bottom = &req->qtcb->bottom.port; 605 struct Scsi_Host *shost = adapter->scsi_host; 606 607 if (req->data) 608 memcpy(req->data, bottom, sizeof(*bottom)); 609 610 if (adapter->connection_features & FSF_FEATURE_NPIV_MODE) { 611 fc_host_permanent_port_name(shost) = bottom->wwpn; 612 fc_host_port_type(shost) = FC_PORTTYPE_NPIV; 613 } else 614 fc_host_permanent_port_name(shost) = fc_host_port_name(shost); 615 fc_host_maxframe_size(shost) = bottom->maximum_frame_size; 616 fc_host_supported_speeds(shost) = 617 zfcp_fsf_convert_portspeed(bottom->supported_speed); 618 memcpy(fc_host_supported_fc4s(shost), bottom->supported_fc4_types, 619 FC_FC4_LIST_SIZE); 620 memcpy(fc_host_active_fc4s(shost), bottom->active_fc4_types, 621 FC_FC4_LIST_SIZE); 622 } 623 624 static void zfcp_fsf_exchange_port_data_handler(struct zfcp_fsf_req *req) 625 { 626 struct fsf_qtcb *qtcb = req->qtcb; 627 628 if (req->status & ZFCP_STATUS_FSFREQ_ERROR) 629 return; 630 631 switch (qtcb->header.fsf_status) { 632 case FSF_GOOD: 633 zfcp_fsf_exchange_port_evaluate(req); 634 break; 635 case FSF_EXCHANGE_CONFIG_DATA_INCOMPLETE: 636 zfcp_fsf_exchange_port_evaluate(req); 637 zfcp_fsf_link_down_info_eval(req, 638 &qtcb->header.fsf_status_qual.link_down_info); 639 break; 640 } 641 } 642 643 static struct zfcp_fsf_req *zfcp_fsf_alloc(mempool_t *pool) 644 { 645 struct zfcp_fsf_req *req; 646 647 if (likely(pool)) 648 req = mempool_alloc(pool, GFP_ATOMIC); 649 else 650 req = kmalloc(sizeof(*req), GFP_ATOMIC); 651 652 if (unlikely(!req)) 653 return NULL; 654 655 memset(req, 0, sizeof(*req)); 656 req->pool = pool; 657 return req; 658 } 659 660 static struct fsf_qtcb *zfcp_qtcb_alloc(mempool_t *pool) 661 { 662 struct fsf_qtcb *qtcb; 663 664 if (likely(pool)) 665 qtcb = mempool_alloc(pool, GFP_ATOMIC); 666 else 667 qtcb = kmem_cache_alloc(zfcp_fsf_qtcb_cache, GFP_ATOMIC); 668 669 if (unlikely(!qtcb)) 670 return NULL; 671 672 memset(qtcb, 0, sizeof(*qtcb)); 673 return qtcb; 674 } 675 676 static struct zfcp_fsf_req *zfcp_fsf_req_create(struct zfcp_qdio *qdio, 677 u32 fsf_cmd, u8 sbtype, 678 mempool_t *pool) 679 { 680 struct zfcp_adapter *adapter = qdio->adapter; 681 struct zfcp_fsf_req *req = zfcp_fsf_alloc(pool); 682 683 if (unlikely(!req)) 684 return ERR_PTR(-ENOMEM); 685 686 if (adapter->req_no == 0) 687 adapter->req_no++; 688 689 INIT_LIST_HEAD(&req->list); 690 init_timer(&req->timer); 691 init_completion(&req->completion); 692 693 req->adapter = adapter; 694 req->fsf_command = fsf_cmd; 695 req->req_id = adapter->req_no; 696 697 if (likely(fsf_cmd != FSF_QTCB_UNSOLICITED_STATUS)) { 698 if (likely(pool)) 699 req->qtcb = zfcp_qtcb_alloc(adapter->pool.qtcb_pool); 700 else 701 req->qtcb = zfcp_qtcb_alloc(NULL); 702 703 if (unlikely(!req->qtcb)) { 704 zfcp_fsf_req_free(req); 705 return ERR_PTR(-ENOMEM); 706 } 707 708 req->seq_no = adapter->fsf_req_seq_no; 709 req->qtcb->prefix.req_seq_no = adapter->fsf_req_seq_no; 710 req->qtcb->prefix.req_id = req->req_id; 711 req->qtcb->prefix.ulp_info = 26; 712 req->qtcb->prefix.qtcb_type = fsf_qtcb_type[req->fsf_command]; 713 req->qtcb->prefix.qtcb_version = FSF_QTCB_CURRENT_VERSION; 714 req->qtcb->header.req_handle = req->req_id; 715 req->qtcb->header.fsf_command = req->fsf_command; 716 } 717 718 zfcp_qdio_req_init(adapter->qdio, &req->qdio_req, req->req_id, sbtype, 719 req->qtcb, sizeof(struct fsf_qtcb)); 720 721 return req; 722 } 723 724 static int zfcp_fsf_req_send(struct zfcp_fsf_req *req) 725 { 726 struct zfcp_adapter *adapter = req->adapter; 727 struct zfcp_qdio *qdio = adapter->qdio; 728 int with_qtcb = (req->qtcb != NULL); 729 int req_id = req->req_id; 730 731 zfcp_reqlist_add(adapter->req_list, req); 732 733 req->qdio_req.qdio_outb_usage = atomic_read(&qdio->req_q_free); 734 req->issued = get_tod_clock(); 735 if (zfcp_qdio_send(qdio, &req->qdio_req)) { 736 del_timer(&req->timer); 737 /* lookup request again, list might have changed */ 738 zfcp_reqlist_find_rm(adapter->req_list, req_id); 739 zfcp_erp_adapter_reopen(adapter, 0, "fsrs__1"); 740 return -EIO; 741 } 742 743 /* Don't increase for unsolicited status */ 744 if (with_qtcb) 745 adapter->fsf_req_seq_no++; 746 adapter->req_no++; 747 748 return 0; 749 } 750 751 /** 752 * zfcp_fsf_status_read - send status read request 753 * @adapter: pointer to struct zfcp_adapter 754 * @req_flags: request flags 755 * Returns: 0 on success, ERROR otherwise 756 */ 757 int zfcp_fsf_status_read(struct zfcp_qdio *qdio) 758 { 759 struct zfcp_adapter *adapter = qdio->adapter; 760 struct zfcp_fsf_req *req; 761 struct fsf_status_read_buffer *sr_buf; 762 struct page *page; 763 int retval = -EIO; 764 765 spin_lock_irq(&qdio->req_q_lock); 766 if (zfcp_qdio_sbal_get(qdio)) 767 goto out; 768 769 req = zfcp_fsf_req_create(qdio, FSF_QTCB_UNSOLICITED_STATUS, 0, 770 adapter->pool.status_read_req); 771 if (IS_ERR(req)) { 772 retval = PTR_ERR(req); 773 goto out; 774 } 775 776 page = mempool_alloc(adapter->pool.sr_data, GFP_ATOMIC); 777 if (!page) { 778 retval = -ENOMEM; 779 goto failed_buf; 780 } 781 sr_buf = page_address(page); 782 memset(sr_buf, 0, sizeof(*sr_buf)); 783 req->data = sr_buf; 784 785 zfcp_qdio_fill_next(qdio, &req->qdio_req, sr_buf, sizeof(*sr_buf)); 786 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 787 788 retval = zfcp_fsf_req_send(req); 789 if (retval) 790 goto failed_req_send; 791 792 goto out; 793 794 failed_req_send: 795 req->data = NULL; 796 mempool_free(virt_to_page(sr_buf), adapter->pool.sr_data); 797 failed_buf: 798 zfcp_dbf_hba_fsf_uss("fssr__1", req); 799 zfcp_fsf_req_free(req); 800 out: 801 spin_unlock_irq(&qdio->req_q_lock); 802 return retval; 803 } 804 805 static void zfcp_fsf_abort_fcp_command_handler(struct zfcp_fsf_req *req) 806 { 807 struct scsi_device *sdev = req->data; 808 struct zfcp_scsi_dev *zfcp_sdev; 809 union fsf_status_qual *fsq = &req->qtcb->header.fsf_status_qual; 810 811 if (req->status & ZFCP_STATUS_FSFREQ_ERROR) 812 return; 813 814 zfcp_sdev = sdev_to_zfcp(sdev); 815 816 switch (req->qtcb->header.fsf_status) { 817 case FSF_PORT_HANDLE_NOT_VALID: 818 if (fsq->word[0] == fsq->word[1]) { 819 zfcp_erp_adapter_reopen(zfcp_sdev->port->adapter, 0, 820 "fsafch1"); 821 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 822 } 823 break; 824 case FSF_LUN_HANDLE_NOT_VALID: 825 if (fsq->word[0] == fsq->word[1]) { 826 zfcp_erp_port_reopen(zfcp_sdev->port, 0, "fsafch2"); 827 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 828 } 829 break; 830 case FSF_FCP_COMMAND_DOES_NOT_EXIST: 831 req->status |= ZFCP_STATUS_FSFREQ_ABORTNOTNEEDED; 832 break; 833 case FSF_PORT_BOXED: 834 zfcp_erp_set_port_status(zfcp_sdev->port, 835 ZFCP_STATUS_COMMON_ACCESS_BOXED); 836 zfcp_erp_port_reopen(zfcp_sdev->port, 837 ZFCP_STATUS_COMMON_ERP_FAILED, "fsafch3"); 838 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 839 break; 840 case FSF_LUN_BOXED: 841 zfcp_erp_set_lun_status(sdev, ZFCP_STATUS_COMMON_ACCESS_BOXED); 842 zfcp_erp_lun_reopen(sdev, ZFCP_STATUS_COMMON_ERP_FAILED, 843 "fsafch4"); 844 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 845 break; 846 case FSF_ADAPTER_STATUS_AVAILABLE: 847 switch (fsq->word[0]) { 848 case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: 849 zfcp_fc_test_link(zfcp_sdev->port); 850 /* fall through */ 851 case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: 852 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 853 break; 854 } 855 break; 856 case FSF_GOOD: 857 req->status |= ZFCP_STATUS_FSFREQ_ABORTSUCCEEDED; 858 break; 859 } 860 } 861 862 /** 863 * zfcp_fsf_abort_fcp_cmnd - abort running SCSI command 864 * @scmnd: The SCSI command to abort 865 * Returns: pointer to struct zfcp_fsf_req 866 */ 867 868 struct zfcp_fsf_req *zfcp_fsf_abort_fcp_cmnd(struct scsi_cmnd *scmnd) 869 { 870 struct zfcp_fsf_req *req = NULL; 871 struct scsi_device *sdev = scmnd->device; 872 struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(sdev); 873 struct zfcp_qdio *qdio = zfcp_sdev->port->adapter->qdio; 874 unsigned long old_req_id = (unsigned long) scmnd->host_scribble; 875 876 spin_lock_irq(&qdio->req_q_lock); 877 if (zfcp_qdio_sbal_get(qdio)) 878 goto out; 879 req = zfcp_fsf_req_create(qdio, FSF_QTCB_ABORT_FCP_CMND, 880 SBAL_SFLAGS0_TYPE_READ, 881 qdio->adapter->pool.scsi_abort); 882 if (IS_ERR(req)) { 883 req = NULL; 884 goto out; 885 } 886 887 if (unlikely(!(atomic_read(&zfcp_sdev->status) & 888 ZFCP_STATUS_COMMON_UNBLOCKED))) 889 goto out_error_free; 890 891 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 892 893 req->data = sdev; 894 req->handler = zfcp_fsf_abort_fcp_command_handler; 895 req->qtcb->header.lun_handle = zfcp_sdev->lun_handle; 896 req->qtcb->header.port_handle = zfcp_sdev->port->handle; 897 req->qtcb->bottom.support.req_handle = (u64) old_req_id; 898 899 zfcp_fsf_start_timer(req, ZFCP_SCSI_ER_TIMEOUT); 900 if (!zfcp_fsf_req_send(req)) 901 goto out; 902 903 out_error_free: 904 zfcp_fsf_req_free(req); 905 req = NULL; 906 out: 907 spin_unlock_irq(&qdio->req_q_lock); 908 return req; 909 } 910 911 static void zfcp_fsf_send_ct_handler(struct zfcp_fsf_req *req) 912 { 913 struct zfcp_adapter *adapter = req->adapter; 914 struct zfcp_fsf_ct_els *ct = req->data; 915 struct fsf_qtcb_header *header = &req->qtcb->header; 916 917 ct->status = -EINVAL; 918 919 if (req->status & ZFCP_STATUS_FSFREQ_ERROR) 920 goto skip_fsfstatus; 921 922 switch (header->fsf_status) { 923 case FSF_GOOD: 924 zfcp_dbf_san_res("fsscth2", req); 925 ct->status = 0; 926 break; 927 case FSF_SERVICE_CLASS_NOT_SUPPORTED: 928 zfcp_fsf_class_not_supp(req); 929 break; 930 case FSF_ADAPTER_STATUS_AVAILABLE: 931 switch (header->fsf_status_qual.word[0]){ 932 case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: 933 case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: 934 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 935 break; 936 } 937 break; 938 case FSF_ACCESS_DENIED: 939 break; 940 case FSF_PORT_BOXED: 941 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 942 break; 943 case FSF_PORT_HANDLE_NOT_VALID: 944 zfcp_erp_adapter_reopen(adapter, 0, "fsscth1"); 945 /* fall through */ 946 case FSF_GENERIC_COMMAND_REJECTED: 947 case FSF_PAYLOAD_SIZE_MISMATCH: 948 case FSF_REQUEST_SIZE_TOO_LARGE: 949 case FSF_RESPONSE_SIZE_TOO_LARGE: 950 case FSF_SBAL_MISMATCH: 951 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 952 break; 953 } 954 955 skip_fsfstatus: 956 if (ct->handler) 957 ct->handler(ct->handler_data); 958 } 959 960 static void zfcp_fsf_setup_ct_els_unchained(struct zfcp_qdio *qdio, 961 struct zfcp_qdio_req *q_req, 962 struct scatterlist *sg_req, 963 struct scatterlist *sg_resp) 964 { 965 zfcp_qdio_fill_next(qdio, q_req, sg_virt(sg_req), sg_req->length); 966 zfcp_qdio_fill_next(qdio, q_req, sg_virt(sg_resp), sg_resp->length); 967 zfcp_qdio_set_sbale_last(qdio, q_req); 968 } 969 970 static int zfcp_fsf_setup_ct_els_sbals(struct zfcp_fsf_req *req, 971 struct scatterlist *sg_req, 972 struct scatterlist *sg_resp) 973 { 974 struct zfcp_adapter *adapter = req->adapter; 975 struct zfcp_qdio *qdio = adapter->qdio; 976 struct fsf_qtcb *qtcb = req->qtcb; 977 u32 feat = adapter->adapter_features; 978 979 if (zfcp_adapter_multi_buffer_active(adapter)) { 980 if (zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, sg_req)) 981 return -EIO; 982 if (zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, sg_resp)) 983 return -EIO; 984 985 zfcp_qdio_set_data_div(qdio, &req->qdio_req, 986 zfcp_qdio_sbale_count(sg_req)); 987 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 988 zfcp_qdio_set_scount(qdio, &req->qdio_req); 989 return 0; 990 } 991 992 /* use single, unchained SBAL if it can hold the request */ 993 if (zfcp_qdio_sg_one_sbale(sg_req) && zfcp_qdio_sg_one_sbale(sg_resp)) { 994 zfcp_fsf_setup_ct_els_unchained(qdio, &req->qdio_req, 995 sg_req, sg_resp); 996 return 0; 997 } 998 999 if (!(feat & FSF_FEATURE_ELS_CT_CHAINED_SBALS)) 1000 return -EOPNOTSUPP; 1001 1002 if (zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, sg_req)) 1003 return -EIO; 1004 1005 qtcb->bottom.support.req_buf_length = zfcp_qdio_real_bytes(sg_req); 1006 1007 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 1008 zfcp_qdio_skip_to_last_sbale(qdio, &req->qdio_req); 1009 1010 if (zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, sg_resp)) 1011 return -EIO; 1012 1013 qtcb->bottom.support.resp_buf_length = zfcp_qdio_real_bytes(sg_resp); 1014 1015 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 1016 1017 return 0; 1018 } 1019 1020 static int zfcp_fsf_setup_ct_els(struct zfcp_fsf_req *req, 1021 struct scatterlist *sg_req, 1022 struct scatterlist *sg_resp, 1023 unsigned int timeout) 1024 { 1025 int ret; 1026 1027 ret = zfcp_fsf_setup_ct_els_sbals(req, sg_req, sg_resp); 1028 if (ret) 1029 return ret; 1030 1031 /* common settings for ct/gs and els requests */ 1032 if (timeout > 255) 1033 timeout = 255; /* max value accepted by hardware */ 1034 req->qtcb->bottom.support.service_class = FSF_CLASS_3; 1035 req->qtcb->bottom.support.timeout = timeout; 1036 zfcp_fsf_start_timer(req, (timeout + 10) * HZ); 1037 1038 return 0; 1039 } 1040 1041 /** 1042 * zfcp_fsf_send_ct - initiate a Generic Service request (FC-GS) 1043 * @ct: pointer to struct zfcp_send_ct with data for request 1044 * @pool: if non-null this mempool is used to allocate struct zfcp_fsf_req 1045 */ 1046 int zfcp_fsf_send_ct(struct zfcp_fc_wka_port *wka_port, 1047 struct zfcp_fsf_ct_els *ct, mempool_t *pool, 1048 unsigned int timeout) 1049 { 1050 struct zfcp_qdio *qdio = wka_port->adapter->qdio; 1051 struct zfcp_fsf_req *req; 1052 int ret = -EIO; 1053 1054 spin_lock_irq(&qdio->req_q_lock); 1055 if (zfcp_qdio_sbal_get(qdio)) 1056 goto out; 1057 1058 req = zfcp_fsf_req_create(qdio, FSF_QTCB_SEND_GENERIC, 1059 SBAL_SFLAGS0_TYPE_WRITE_READ, pool); 1060 1061 if (IS_ERR(req)) { 1062 ret = PTR_ERR(req); 1063 goto out; 1064 } 1065 1066 req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; 1067 ret = zfcp_fsf_setup_ct_els(req, ct->req, ct->resp, timeout); 1068 if (ret) 1069 goto failed_send; 1070 1071 req->handler = zfcp_fsf_send_ct_handler; 1072 req->qtcb->header.port_handle = wka_port->handle; 1073 req->data = ct; 1074 1075 zfcp_dbf_san_req("fssct_1", req, wka_port->d_id); 1076 1077 ret = zfcp_fsf_req_send(req); 1078 if (ret) 1079 goto failed_send; 1080 1081 goto out; 1082 1083 failed_send: 1084 zfcp_fsf_req_free(req); 1085 out: 1086 spin_unlock_irq(&qdio->req_q_lock); 1087 return ret; 1088 } 1089 1090 static void zfcp_fsf_send_els_handler(struct zfcp_fsf_req *req) 1091 { 1092 struct zfcp_fsf_ct_els *send_els = req->data; 1093 struct zfcp_port *port = send_els->port; 1094 struct fsf_qtcb_header *header = &req->qtcb->header; 1095 1096 send_els->status = -EINVAL; 1097 1098 if (req->status & ZFCP_STATUS_FSFREQ_ERROR) 1099 goto skip_fsfstatus; 1100 1101 switch (header->fsf_status) { 1102 case FSF_GOOD: 1103 zfcp_dbf_san_res("fsselh1", req); 1104 send_els->status = 0; 1105 break; 1106 case FSF_SERVICE_CLASS_NOT_SUPPORTED: 1107 zfcp_fsf_class_not_supp(req); 1108 break; 1109 case FSF_ADAPTER_STATUS_AVAILABLE: 1110 switch (header->fsf_status_qual.word[0]){ 1111 case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: 1112 case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: 1113 case FSF_SQ_RETRY_IF_POSSIBLE: 1114 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1115 break; 1116 } 1117 break; 1118 case FSF_ELS_COMMAND_REJECTED: 1119 case FSF_PAYLOAD_SIZE_MISMATCH: 1120 case FSF_REQUEST_SIZE_TOO_LARGE: 1121 case FSF_RESPONSE_SIZE_TOO_LARGE: 1122 break; 1123 case FSF_ACCESS_DENIED: 1124 if (port) { 1125 zfcp_cfdc_port_denied(port, &header->fsf_status_qual); 1126 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1127 } 1128 break; 1129 case FSF_SBAL_MISMATCH: 1130 /* should never occur, avoided in zfcp_fsf_send_els */ 1131 /* fall through */ 1132 default: 1133 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1134 break; 1135 } 1136 skip_fsfstatus: 1137 if (send_els->handler) 1138 send_els->handler(send_els->handler_data); 1139 } 1140 1141 /** 1142 * zfcp_fsf_send_els - initiate an ELS command (FC-FS) 1143 * @els: pointer to struct zfcp_send_els with data for the command 1144 */ 1145 int zfcp_fsf_send_els(struct zfcp_adapter *adapter, u32 d_id, 1146 struct zfcp_fsf_ct_els *els, unsigned int timeout) 1147 { 1148 struct zfcp_fsf_req *req; 1149 struct zfcp_qdio *qdio = adapter->qdio; 1150 int ret = -EIO; 1151 1152 spin_lock_irq(&qdio->req_q_lock); 1153 if (zfcp_qdio_sbal_get(qdio)) 1154 goto out; 1155 1156 req = zfcp_fsf_req_create(qdio, FSF_QTCB_SEND_ELS, 1157 SBAL_SFLAGS0_TYPE_WRITE_READ, NULL); 1158 1159 if (IS_ERR(req)) { 1160 ret = PTR_ERR(req); 1161 goto out; 1162 } 1163 1164 req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; 1165 1166 if (!zfcp_adapter_multi_buffer_active(adapter)) 1167 zfcp_qdio_sbal_limit(qdio, &req->qdio_req, 2); 1168 1169 ret = zfcp_fsf_setup_ct_els(req, els->req, els->resp, timeout); 1170 1171 if (ret) 1172 goto failed_send; 1173 1174 hton24(req->qtcb->bottom.support.d_id, d_id); 1175 req->handler = zfcp_fsf_send_els_handler; 1176 req->data = els; 1177 1178 zfcp_dbf_san_req("fssels1", req, d_id); 1179 1180 ret = zfcp_fsf_req_send(req); 1181 if (ret) 1182 goto failed_send; 1183 1184 goto out; 1185 1186 failed_send: 1187 zfcp_fsf_req_free(req); 1188 out: 1189 spin_unlock_irq(&qdio->req_q_lock); 1190 return ret; 1191 } 1192 1193 int zfcp_fsf_exchange_config_data(struct zfcp_erp_action *erp_action) 1194 { 1195 struct zfcp_fsf_req *req; 1196 struct zfcp_qdio *qdio = erp_action->adapter->qdio; 1197 int retval = -EIO; 1198 1199 spin_lock_irq(&qdio->req_q_lock); 1200 if (zfcp_qdio_sbal_get(qdio)) 1201 goto out; 1202 1203 req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_CONFIG_DATA, 1204 SBAL_SFLAGS0_TYPE_READ, 1205 qdio->adapter->pool.erp_req); 1206 1207 if (IS_ERR(req)) { 1208 retval = PTR_ERR(req); 1209 goto out; 1210 } 1211 1212 req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; 1213 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 1214 1215 req->qtcb->bottom.config.feature_selection = 1216 FSF_FEATURE_CFDC | 1217 FSF_FEATURE_LUN_SHARING | 1218 FSF_FEATURE_NOTIFICATION_LOST | 1219 FSF_FEATURE_UPDATE_ALERT; 1220 req->erp_action = erp_action; 1221 req->handler = zfcp_fsf_exchange_config_data_handler; 1222 erp_action->fsf_req_id = req->req_id; 1223 1224 zfcp_fsf_start_erp_timer(req); 1225 retval = zfcp_fsf_req_send(req); 1226 if (retval) { 1227 zfcp_fsf_req_free(req); 1228 erp_action->fsf_req_id = 0; 1229 } 1230 out: 1231 spin_unlock_irq(&qdio->req_q_lock); 1232 return retval; 1233 } 1234 1235 int zfcp_fsf_exchange_config_data_sync(struct zfcp_qdio *qdio, 1236 struct fsf_qtcb_bottom_config *data) 1237 { 1238 struct zfcp_fsf_req *req = NULL; 1239 int retval = -EIO; 1240 1241 spin_lock_irq(&qdio->req_q_lock); 1242 if (zfcp_qdio_sbal_get(qdio)) 1243 goto out_unlock; 1244 1245 req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_CONFIG_DATA, 1246 SBAL_SFLAGS0_TYPE_READ, NULL); 1247 1248 if (IS_ERR(req)) { 1249 retval = PTR_ERR(req); 1250 goto out_unlock; 1251 } 1252 1253 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 1254 req->handler = zfcp_fsf_exchange_config_data_handler; 1255 1256 req->qtcb->bottom.config.feature_selection = 1257 FSF_FEATURE_CFDC | 1258 FSF_FEATURE_LUN_SHARING | 1259 FSF_FEATURE_NOTIFICATION_LOST | 1260 FSF_FEATURE_UPDATE_ALERT; 1261 1262 if (data) 1263 req->data = data; 1264 1265 zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT); 1266 retval = zfcp_fsf_req_send(req); 1267 spin_unlock_irq(&qdio->req_q_lock); 1268 if (!retval) 1269 wait_for_completion(&req->completion); 1270 1271 zfcp_fsf_req_free(req); 1272 return retval; 1273 1274 out_unlock: 1275 spin_unlock_irq(&qdio->req_q_lock); 1276 return retval; 1277 } 1278 1279 /** 1280 * zfcp_fsf_exchange_port_data - request information about local port 1281 * @erp_action: ERP action for the adapter for which port data is requested 1282 * Returns: 0 on success, error otherwise 1283 */ 1284 int zfcp_fsf_exchange_port_data(struct zfcp_erp_action *erp_action) 1285 { 1286 struct zfcp_qdio *qdio = erp_action->adapter->qdio; 1287 struct zfcp_fsf_req *req; 1288 int retval = -EIO; 1289 1290 if (!(qdio->adapter->adapter_features & FSF_FEATURE_HBAAPI_MANAGEMENT)) 1291 return -EOPNOTSUPP; 1292 1293 spin_lock_irq(&qdio->req_q_lock); 1294 if (zfcp_qdio_sbal_get(qdio)) 1295 goto out; 1296 1297 req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_PORT_DATA, 1298 SBAL_SFLAGS0_TYPE_READ, 1299 qdio->adapter->pool.erp_req); 1300 1301 if (IS_ERR(req)) { 1302 retval = PTR_ERR(req); 1303 goto out; 1304 } 1305 1306 req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; 1307 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 1308 1309 req->handler = zfcp_fsf_exchange_port_data_handler; 1310 req->erp_action = erp_action; 1311 erp_action->fsf_req_id = req->req_id; 1312 1313 zfcp_fsf_start_erp_timer(req); 1314 retval = zfcp_fsf_req_send(req); 1315 if (retval) { 1316 zfcp_fsf_req_free(req); 1317 erp_action->fsf_req_id = 0; 1318 } 1319 out: 1320 spin_unlock_irq(&qdio->req_q_lock); 1321 return retval; 1322 } 1323 1324 /** 1325 * zfcp_fsf_exchange_port_data_sync - request information about local port 1326 * @qdio: pointer to struct zfcp_qdio 1327 * @data: pointer to struct fsf_qtcb_bottom_port 1328 * Returns: 0 on success, error otherwise 1329 */ 1330 int zfcp_fsf_exchange_port_data_sync(struct zfcp_qdio *qdio, 1331 struct fsf_qtcb_bottom_port *data) 1332 { 1333 struct zfcp_fsf_req *req = NULL; 1334 int retval = -EIO; 1335 1336 if (!(qdio->adapter->adapter_features & FSF_FEATURE_HBAAPI_MANAGEMENT)) 1337 return -EOPNOTSUPP; 1338 1339 spin_lock_irq(&qdio->req_q_lock); 1340 if (zfcp_qdio_sbal_get(qdio)) 1341 goto out_unlock; 1342 1343 req = zfcp_fsf_req_create(qdio, FSF_QTCB_EXCHANGE_PORT_DATA, 1344 SBAL_SFLAGS0_TYPE_READ, NULL); 1345 1346 if (IS_ERR(req)) { 1347 retval = PTR_ERR(req); 1348 goto out_unlock; 1349 } 1350 1351 if (data) 1352 req->data = data; 1353 1354 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 1355 1356 req->handler = zfcp_fsf_exchange_port_data_handler; 1357 zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT); 1358 retval = zfcp_fsf_req_send(req); 1359 spin_unlock_irq(&qdio->req_q_lock); 1360 1361 if (!retval) 1362 wait_for_completion(&req->completion); 1363 1364 zfcp_fsf_req_free(req); 1365 1366 return retval; 1367 1368 out_unlock: 1369 spin_unlock_irq(&qdio->req_q_lock); 1370 return retval; 1371 } 1372 1373 static void zfcp_fsf_open_port_handler(struct zfcp_fsf_req *req) 1374 { 1375 struct zfcp_port *port = req->data; 1376 struct fsf_qtcb_header *header = &req->qtcb->header; 1377 struct fc_els_flogi *plogi; 1378 1379 if (req->status & ZFCP_STATUS_FSFREQ_ERROR) 1380 goto out; 1381 1382 switch (header->fsf_status) { 1383 case FSF_PORT_ALREADY_OPEN: 1384 break; 1385 case FSF_ACCESS_DENIED: 1386 zfcp_cfdc_port_denied(port, &header->fsf_status_qual); 1387 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1388 break; 1389 case FSF_MAXIMUM_NUMBER_OF_PORTS_EXCEEDED: 1390 dev_warn(&req->adapter->ccw_device->dev, 1391 "Not enough FCP adapter resources to open " 1392 "remote port 0x%016Lx\n", 1393 (unsigned long long)port->wwpn); 1394 zfcp_erp_set_port_status(port, 1395 ZFCP_STATUS_COMMON_ERP_FAILED); 1396 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1397 break; 1398 case FSF_ADAPTER_STATUS_AVAILABLE: 1399 switch (header->fsf_status_qual.word[0]) { 1400 case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: 1401 case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: 1402 case FSF_SQ_NO_RETRY_POSSIBLE: 1403 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1404 break; 1405 } 1406 break; 1407 case FSF_GOOD: 1408 port->handle = header->port_handle; 1409 atomic_set_mask(ZFCP_STATUS_COMMON_OPEN | 1410 ZFCP_STATUS_PORT_PHYS_OPEN, &port->status); 1411 atomic_clear_mask(ZFCP_STATUS_COMMON_ACCESS_DENIED | 1412 ZFCP_STATUS_COMMON_ACCESS_BOXED, 1413 &port->status); 1414 /* check whether D_ID has changed during open */ 1415 /* 1416 * FIXME: This check is not airtight, as the FCP channel does 1417 * not monitor closures of target port connections caused on 1418 * the remote side. Thus, they might miss out on invalidating 1419 * locally cached WWPNs (and other N_Port parameters) of gone 1420 * target ports. So, our heroic attempt to make things safe 1421 * could be undermined by 'open port' response data tagged with 1422 * obsolete WWPNs. Another reason to monitor potential 1423 * connection closures ourself at least (by interpreting 1424 * incoming ELS' and unsolicited status). It just crosses my 1425 * mind that one should be able to cross-check by means of 1426 * another GID_PN straight after a port has been opened. 1427 * Alternately, an ADISC/PDISC ELS should suffice, as well. 1428 */ 1429 plogi = (struct fc_els_flogi *) req->qtcb->bottom.support.els; 1430 if (req->qtcb->bottom.support.els1_length >= 1431 FSF_PLOGI_MIN_LEN) 1432 zfcp_fc_plogi_evaluate(port, plogi); 1433 break; 1434 case FSF_UNKNOWN_OP_SUBTYPE: 1435 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1436 break; 1437 } 1438 1439 out: 1440 put_device(&port->dev); 1441 } 1442 1443 /** 1444 * zfcp_fsf_open_port - create and send open port request 1445 * @erp_action: pointer to struct zfcp_erp_action 1446 * Returns: 0 on success, error otherwise 1447 */ 1448 int zfcp_fsf_open_port(struct zfcp_erp_action *erp_action) 1449 { 1450 struct zfcp_qdio *qdio = erp_action->adapter->qdio; 1451 struct zfcp_port *port = erp_action->port; 1452 struct zfcp_fsf_req *req; 1453 int retval = -EIO; 1454 1455 spin_lock_irq(&qdio->req_q_lock); 1456 if (zfcp_qdio_sbal_get(qdio)) 1457 goto out; 1458 1459 req = zfcp_fsf_req_create(qdio, FSF_QTCB_OPEN_PORT_WITH_DID, 1460 SBAL_SFLAGS0_TYPE_READ, 1461 qdio->adapter->pool.erp_req); 1462 1463 if (IS_ERR(req)) { 1464 retval = PTR_ERR(req); 1465 goto out; 1466 } 1467 1468 req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; 1469 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 1470 1471 req->handler = zfcp_fsf_open_port_handler; 1472 hton24(req->qtcb->bottom.support.d_id, port->d_id); 1473 req->data = port; 1474 req->erp_action = erp_action; 1475 erp_action->fsf_req_id = req->req_id; 1476 get_device(&port->dev); 1477 1478 zfcp_fsf_start_erp_timer(req); 1479 retval = zfcp_fsf_req_send(req); 1480 if (retval) { 1481 zfcp_fsf_req_free(req); 1482 erp_action->fsf_req_id = 0; 1483 put_device(&port->dev); 1484 } 1485 out: 1486 spin_unlock_irq(&qdio->req_q_lock); 1487 return retval; 1488 } 1489 1490 static void zfcp_fsf_close_port_handler(struct zfcp_fsf_req *req) 1491 { 1492 struct zfcp_port *port = req->data; 1493 1494 if (req->status & ZFCP_STATUS_FSFREQ_ERROR) 1495 return; 1496 1497 switch (req->qtcb->header.fsf_status) { 1498 case FSF_PORT_HANDLE_NOT_VALID: 1499 zfcp_erp_adapter_reopen(port->adapter, 0, "fscph_1"); 1500 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1501 break; 1502 case FSF_ADAPTER_STATUS_AVAILABLE: 1503 break; 1504 case FSF_GOOD: 1505 zfcp_erp_clear_port_status(port, ZFCP_STATUS_COMMON_OPEN); 1506 break; 1507 } 1508 } 1509 1510 /** 1511 * zfcp_fsf_close_port - create and send close port request 1512 * @erp_action: pointer to struct zfcp_erp_action 1513 * Returns: 0 on success, error otherwise 1514 */ 1515 int zfcp_fsf_close_port(struct zfcp_erp_action *erp_action) 1516 { 1517 struct zfcp_qdio *qdio = erp_action->adapter->qdio; 1518 struct zfcp_fsf_req *req; 1519 int retval = -EIO; 1520 1521 spin_lock_irq(&qdio->req_q_lock); 1522 if (zfcp_qdio_sbal_get(qdio)) 1523 goto out; 1524 1525 req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_PORT, 1526 SBAL_SFLAGS0_TYPE_READ, 1527 qdio->adapter->pool.erp_req); 1528 1529 if (IS_ERR(req)) { 1530 retval = PTR_ERR(req); 1531 goto out; 1532 } 1533 1534 req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; 1535 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 1536 1537 req->handler = zfcp_fsf_close_port_handler; 1538 req->data = erp_action->port; 1539 req->erp_action = erp_action; 1540 req->qtcb->header.port_handle = erp_action->port->handle; 1541 erp_action->fsf_req_id = req->req_id; 1542 1543 zfcp_fsf_start_erp_timer(req); 1544 retval = zfcp_fsf_req_send(req); 1545 if (retval) { 1546 zfcp_fsf_req_free(req); 1547 erp_action->fsf_req_id = 0; 1548 } 1549 out: 1550 spin_unlock_irq(&qdio->req_q_lock); 1551 return retval; 1552 } 1553 1554 static void zfcp_fsf_open_wka_port_handler(struct zfcp_fsf_req *req) 1555 { 1556 struct zfcp_fc_wka_port *wka_port = req->data; 1557 struct fsf_qtcb_header *header = &req->qtcb->header; 1558 1559 if (req->status & ZFCP_STATUS_FSFREQ_ERROR) { 1560 wka_port->status = ZFCP_FC_WKA_PORT_OFFLINE; 1561 goto out; 1562 } 1563 1564 switch (header->fsf_status) { 1565 case FSF_MAXIMUM_NUMBER_OF_PORTS_EXCEEDED: 1566 dev_warn(&req->adapter->ccw_device->dev, 1567 "Opening WKA port 0x%x failed\n", wka_port->d_id); 1568 /* fall through */ 1569 case FSF_ADAPTER_STATUS_AVAILABLE: 1570 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1571 /* fall through */ 1572 case FSF_ACCESS_DENIED: 1573 wka_port->status = ZFCP_FC_WKA_PORT_OFFLINE; 1574 break; 1575 case FSF_GOOD: 1576 wka_port->handle = header->port_handle; 1577 /* fall through */ 1578 case FSF_PORT_ALREADY_OPEN: 1579 wka_port->status = ZFCP_FC_WKA_PORT_ONLINE; 1580 } 1581 out: 1582 wake_up(&wka_port->completion_wq); 1583 } 1584 1585 /** 1586 * zfcp_fsf_open_wka_port - create and send open wka-port request 1587 * @wka_port: pointer to struct zfcp_fc_wka_port 1588 * Returns: 0 on success, error otherwise 1589 */ 1590 int zfcp_fsf_open_wka_port(struct zfcp_fc_wka_port *wka_port) 1591 { 1592 struct zfcp_qdio *qdio = wka_port->adapter->qdio; 1593 struct zfcp_fsf_req *req; 1594 int retval = -EIO; 1595 1596 spin_lock_irq(&qdio->req_q_lock); 1597 if (zfcp_qdio_sbal_get(qdio)) 1598 goto out; 1599 1600 req = zfcp_fsf_req_create(qdio, FSF_QTCB_OPEN_PORT_WITH_DID, 1601 SBAL_SFLAGS0_TYPE_READ, 1602 qdio->adapter->pool.erp_req); 1603 1604 if (IS_ERR(req)) { 1605 retval = PTR_ERR(req); 1606 goto out; 1607 } 1608 1609 req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; 1610 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 1611 1612 req->handler = zfcp_fsf_open_wka_port_handler; 1613 hton24(req->qtcb->bottom.support.d_id, wka_port->d_id); 1614 req->data = wka_port; 1615 1616 zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT); 1617 retval = zfcp_fsf_req_send(req); 1618 if (retval) 1619 zfcp_fsf_req_free(req); 1620 out: 1621 spin_unlock_irq(&qdio->req_q_lock); 1622 return retval; 1623 } 1624 1625 static void zfcp_fsf_close_wka_port_handler(struct zfcp_fsf_req *req) 1626 { 1627 struct zfcp_fc_wka_port *wka_port = req->data; 1628 1629 if (req->qtcb->header.fsf_status == FSF_PORT_HANDLE_NOT_VALID) { 1630 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1631 zfcp_erp_adapter_reopen(wka_port->adapter, 0, "fscwph1"); 1632 } 1633 1634 wka_port->status = ZFCP_FC_WKA_PORT_OFFLINE; 1635 wake_up(&wka_port->completion_wq); 1636 } 1637 1638 /** 1639 * zfcp_fsf_close_wka_port - create and send close wka port request 1640 * @wka_port: WKA port to open 1641 * Returns: 0 on success, error otherwise 1642 */ 1643 int zfcp_fsf_close_wka_port(struct zfcp_fc_wka_port *wka_port) 1644 { 1645 struct zfcp_qdio *qdio = wka_port->adapter->qdio; 1646 struct zfcp_fsf_req *req; 1647 int retval = -EIO; 1648 1649 spin_lock_irq(&qdio->req_q_lock); 1650 if (zfcp_qdio_sbal_get(qdio)) 1651 goto out; 1652 1653 req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_PORT, 1654 SBAL_SFLAGS0_TYPE_READ, 1655 qdio->adapter->pool.erp_req); 1656 1657 if (IS_ERR(req)) { 1658 retval = PTR_ERR(req); 1659 goto out; 1660 } 1661 1662 req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; 1663 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 1664 1665 req->handler = zfcp_fsf_close_wka_port_handler; 1666 req->data = wka_port; 1667 req->qtcb->header.port_handle = wka_port->handle; 1668 1669 zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT); 1670 retval = zfcp_fsf_req_send(req); 1671 if (retval) 1672 zfcp_fsf_req_free(req); 1673 out: 1674 spin_unlock_irq(&qdio->req_q_lock); 1675 return retval; 1676 } 1677 1678 static void zfcp_fsf_close_physical_port_handler(struct zfcp_fsf_req *req) 1679 { 1680 struct zfcp_port *port = req->data; 1681 struct fsf_qtcb_header *header = &req->qtcb->header; 1682 struct scsi_device *sdev; 1683 1684 if (req->status & ZFCP_STATUS_FSFREQ_ERROR) 1685 return; 1686 1687 switch (header->fsf_status) { 1688 case FSF_PORT_HANDLE_NOT_VALID: 1689 zfcp_erp_adapter_reopen(port->adapter, 0, "fscpph1"); 1690 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1691 break; 1692 case FSF_ACCESS_DENIED: 1693 zfcp_cfdc_port_denied(port, &header->fsf_status_qual); 1694 break; 1695 case FSF_PORT_BOXED: 1696 /* can't use generic zfcp_erp_modify_port_status because 1697 * ZFCP_STATUS_COMMON_OPEN must not be reset for the port */ 1698 atomic_clear_mask(ZFCP_STATUS_PORT_PHYS_OPEN, &port->status); 1699 shost_for_each_device(sdev, port->adapter->scsi_host) 1700 if (sdev_to_zfcp(sdev)->port == port) 1701 atomic_clear_mask(ZFCP_STATUS_COMMON_OPEN, 1702 &sdev_to_zfcp(sdev)->status); 1703 zfcp_erp_set_port_status(port, ZFCP_STATUS_COMMON_ACCESS_BOXED); 1704 zfcp_erp_port_reopen(port, ZFCP_STATUS_COMMON_ERP_FAILED, 1705 "fscpph2"); 1706 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1707 break; 1708 case FSF_ADAPTER_STATUS_AVAILABLE: 1709 switch (header->fsf_status_qual.word[0]) { 1710 case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: 1711 /* fall through */ 1712 case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: 1713 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1714 break; 1715 } 1716 break; 1717 case FSF_GOOD: 1718 /* can't use generic zfcp_erp_modify_port_status because 1719 * ZFCP_STATUS_COMMON_OPEN must not be reset for the port 1720 */ 1721 atomic_clear_mask(ZFCP_STATUS_PORT_PHYS_OPEN, &port->status); 1722 shost_for_each_device(sdev, port->adapter->scsi_host) 1723 if (sdev_to_zfcp(sdev)->port == port) 1724 atomic_clear_mask(ZFCP_STATUS_COMMON_OPEN, 1725 &sdev_to_zfcp(sdev)->status); 1726 break; 1727 } 1728 } 1729 1730 /** 1731 * zfcp_fsf_close_physical_port - close physical port 1732 * @erp_action: pointer to struct zfcp_erp_action 1733 * Returns: 0 on success 1734 */ 1735 int zfcp_fsf_close_physical_port(struct zfcp_erp_action *erp_action) 1736 { 1737 struct zfcp_qdio *qdio = erp_action->adapter->qdio; 1738 struct zfcp_fsf_req *req; 1739 int retval = -EIO; 1740 1741 spin_lock_irq(&qdio->req_q_lock); 1742 if (zfcp_qdio_sbal_get(qdio)) 1743 goto out; 1744 1745 req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_PHYSICAL_PORT, 1746 SBAL_SFLAGS0_TYPE_READ, 1747 qdio->adapter->pool.erp_req); 1748 1749 if (IS_ERR(req)) { 1750 retval = PTR_ERR(req); 1751 goto out; 1752 } 1753 1754 req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; 1755 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 1756 1757 req->data = erp_action->port; 1758 req->qtcb->header.port_handle = erp_action->port->handle; 1759 req->erp_action = erp_action; 1760 req->handler = zfcp_fsf_close_physical_port_handler; 1761 erp_action->fsf_req_id = req->req_id; 1762 1763 zfcp_fsf_start_erp_timer(req); 1764 retval = zfcp_fsf_req_send(req); 1765 if (retval) { 1766 zfcp_fsf_req_free(req); 1767 erp_action->fsf_req_id = 0; 1768 } 1769 out: 1770 spin_unlock_irq(&qdio->req_q_lock); 1771 return retval; 1772 } 1773 1774 static void zfcp_fsf_open_lun_handler(struct zfcp_fsf_req *req) 1775 { 1776 struct zfcp_adapter *adapter = req->adapter; 1777 struct scsi_device *sdev = req->data; 1778 struct zfcp_scsi_dev *zfcp_sdev; 1779 struct fsf_qtcb_header *header = &req->qtcb->header; 1780 struct fsf_qtcb_bottom_support *bottom = &req->qtcb->bottom.support; 1781 1782 if (req->status & ZFCP_STATUS_FSFREQ_ERROR) 1783 return; 1784 1785 zfcp_sdev = sdev_to_zfcp(sdev); 1786 1787 atomic_clear_mask(ZFCP_STATUS_COMMON_ACCESS_DENIED | 1788 ZFCP_STATUS_COMMON_ACCESS_BOXED | 1789 ZFCP_STATUS_LUN_SHARED | 1790 ZFCP_STATUS_LUN_READONLY, 1791 &zfcp_sdev->status); 1792 1793 switch (header->fsf_status) { 1794 1795 case FSF_PORT_HANDLE_NOT_VALID: 1796 zfcp_erp_adapter_reopen(adapter, 0, "fsouh_1"); 1797 /* fall through */ 1798 case FSF_LUN_ALREADY_OPEN: 1799 break; 1800 case FSF_ACCESS_DENIED: 1801 zfcp_cfdc_lun_denied(sdev, &header->fsf_status_qual); 1802 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1803 break; 1804 case FSF_PORT_BOXED: 1805 zfcp_erp_set_port_status(zfcp_sdev->port, 1806 ZFCP_STATUS_COMMON_ACCESS_BOXED); 1807 zfcp_erp_port_reopen(zfcp_sdev->port, 1808 ZFCP_STATUS_COMMON_ERP_FAILED, "fsouh_2"); 1809 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1810 break; 1811 case FSF_LUN_SHARING_VIOLATION: 1812 zfcp_cfdc_lun_shrng_vltn(sdev, &header->fsf_status_qual); 1813 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1814 break; 1815 case FSF_MAXIMUM_NUMBER_OF_LUNS_EXCEEDED: 1816 dev_warn(&adapter->ccw_device->dev, 1817 "No handle is available for LUN " 1818 "0x%016Lx on port 0x%016Lx\n", 1819 (unsigned long long)zfcp_scsi_dev_lun(sdev), 1820 (unsigned long long)zfcp_sdev->port->wwpn); 1821 zfcp_erp_set_lun_status(sdev, ZFCP_STATUS_COMMON_ERP_FAILED); 1822 /* fall through */ 1823 case FSF_INVALID_COMMAND_OPTION: 1824 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1825 break; 1826 case FSF_ADAPTER_STATUS_AVAILABLE: 1827 switch (header->fsf_status_qual.word[0]) { 1828 case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: 1829 zfcp_fc_test_link(zfcp_sdev->port); 1830 /* fall through */ 1831 case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: 1832 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1833 break; 1834 } 1835 break; 1836 1837 case FSF_GOOD: 1838 zfcp_sdev->lun_handle = header->lun_handle; 1839 atomic_set_mask(ZFCP_STATUS_COMMON_OPEN, &zfcp_sdev->status); 1840 zfcp_cfdc_open_lun_eval(sdev, bottom); 1841 break; 1842 } 1843 } 1844 1845 /** 1846 * zfcp_fsf_open_lun - open LUN 1847 * @erp_action: pointer to struct zfcp_erp_action 1848 * Returns: 0 on success, error otherwise 1849 */ 1850 int zfcp_fsf_open_lun(struct zfcp_erp_action *erp_action) 1851 { 1852 struct zfcp_adapter *adapter = erp_action->adapter; 1853 struct zfcp_qdio *qdio = adapter->qdio; 1854 struct zfcp_fsf_req *req; 1855 int retval = -EIO; 1856 1857 spin_lock_irq(&qdio->req_q_lock); 1858 if (zfcp_qdio_sbal_get(qdio)) 1859 goto out; 1860 1861 req = zfcp_fsf_req_create(qdio, FSF_QTCB_OPEN_LUN, 1862 SBAL_SFLAGS0_TYPE_READ, 1863 adapter->pool.erp_req); 1864 1865 if (IS_ERR(req)) { 1866 retval = PTR_ERR(req); 1867 goto out; 1868 } 1869 1870 req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; 1871 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 1872 1873 req->qtcb->header.port_handle = erp_action->port->handle; 1874 req->qtcb->bottom.support.fcp_lun = zfcp_scsi_dev_lun(erp_action->sdev); 1875 req->handler = zfcp_fsf_open_lun_handler; 1876 req->data = erp_action->sdev; 1877 req->erp_action = erp_action; 1878 erp_action->fsf_req_id = req->req_id; 1879 1880 if (!(adapter->connection_features & FSF_FEATURE_NPIV_MODE)) 1881 req->qtcb->bottom.support.option = FSF_OPEN_LUN_SUPPRESS_BOXING; 1882 1883 zfcp_fsf_start_erp_timer(req); 1884 retval = zfcp_fsf_req_send(req); 1885 if (retval) { 1886 zfcp_fsf_req_free(req); 1887 erp_action->fsf_req_id = 0; 1888 } 1889 out: 1890 spin_unlock_irq(&qdio->req_q_lock); 1891 return retval; 1892 } 1893 1894 static void zfcp_fsf_close_lun_handler(struct zfcp_fsf_req *req) 1895 { 1896 struct scsi_device *sdev = req->data; 1897 struct zfcp_scsi_dev *zfcp_sdev; 1898 1899 if (req->status & ZFCP_STATUS_FSFREQ_ERROR) 1900 return; 1901 1902 zfcp_sdev = sdev_to_zfcp(sdev); 1903 1904 switch (req->qtcb->header.fsf_status) { 1905 case FSF_PORT_HANDLE_NOT_VALID: 1906 zfcp_erp_adapter_reopen(zfcp_sdev->port->adapter, 0, "fscuh_1"); 1907 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1908 break; 1909 case FSF_LUN_HANDLE_NOT_VALID: 1910 zfcp_erp_port_reopen(zfcp_sdev->port, 0, "fscuh_2"); 1911 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1912 break; 1913 case FSF_PORT_BOXED: 1914 zfcp_erp_set_port_status(zfcp_sdev->port, 1915 ZFCP_STATUS_COMMON_ACCESS_BOXED); 1916 zfcp_erp_port_reopen(zfcp_sdev->port, 1917 ZFCP_STATUS_COMMON_ERP_FAILED, "fscuh_3"); 1918 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1919 break; 1920 case FSF_ADAPTER_STATUS_AVAILABLE: 1921 switch (req->qtcb->header.fsf_status_qual.word[0]) { 1922 case FSF_SQ_INVOKE_LINK_TEST_PROCEDURE: 1923 zfcp_fc_test_link(zfcp_sdev->port); 1924 /* fall through */ 1925 case FSF_SQ_ULP_DEPENDENT_ERP_REQUIRED: 1926 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 1927 break; 1928 } 1929 break; 1930 case FSF_GOOD: 1931 atomic_clear_mask(ZFCP_STATUS_COMMON_OPEN, &zfcp_sdev->status); 1932 break; 1933 } 1934 } 1935 1936 /** 1937 * zfcp_fsf_close_LUN - close LUN 1938 * @erp_action: pointer to erp_action triggering the "close LUN" 1939 * Returns: 0 on success, error otherwise 1940 */ 1941 int zfcp_fsf_close_lun(struct zfcp_erp_action *erp_action) 1942 { 1943 struct zfcp_qdio *qdio = erp_action->adapter->qdio; 1944 struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(erp_action->sdev); 1945 struct zfcp_fsf_req *req; 1946 int retval = -EIO; 1947 1948 spin_lock_irq(&qdio->req_q_lock); 1949 if (zfcp_qdio_sbal_get(qdio)) 1950 goto out; 1951 1952 req = zfcp_fsf_req_create(qdio, FSF_QTCB_CLOSE_LUN, 1953 SBAL_SFLAGS0_TYPE_READ, 1954 qdio->adapter->pool.erp_req); 1955 1956 if (IS_ERR(req)) { 1957 retval = PTR_ERR(req); 1958 goto out; 1959 } 1960 1961 req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; 1962 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 1963 1964 req->qtcb->header.port_handle = erp_action->port->handle; 1965 req->qtcb->header.lun_handle = zfcp_sdev->lun_handle; 1966 req->handler = zfcp_fsf_close_lun_handler; 1967 req->data = erp_action->sdev; 1968 req->erp_action = erp_action; 1969 erp_action->fsf_req_id = req->req_id; 1970 1971 zfcp_fsf_start_erp_timer(req); 1972 retval = zfcp_fsf_req_send(req); 1973 if (retval) { 1974 zfcp_fsf_req_free(req); 1975 erp_action->fsf_req_id = 0; 1976 } 1977 out: 1978 spin_unlock_irq(&qdio->req_q_lock); 1979 return retval; 1980 } 1981 1982 static void zfcp_fsf_update_lat(struct fsf_latency_record *lat_rec, u32 lat) 1983 { 1984 lat_rec->sum += lat; 1985 lat_rec->min = min(lat_rec->min, lat); 1986 lat_rec->max = max(lat_rec->max, lat); 1987 } 1988 1989 static void zfcp_fsf_req_trace(struct zfcp_fsf_req *req, struct scsi_cmnd *scsi) 1990 { 1991 struct fsf_qual_latency_info *lat_in; 1992 struct latency_cont *lat = NULL; 1993 struct zfcp_scsi_dev *zfcp_sdev; 1994 struct zfcp_blk_drv_data blktrc; 1995 int ticks = req->adapter->timer_ticks; 1996 1997 lat_in = &req->qtcb->prefix.prot_status_qual.latency_info; 1998 1999 blktrc.flags = 0; 2000 blktrc.magic = ZFCP_BLK_DRV_DATA_MAGIC; 2001 if (req->status & ZFCP_STATUS_FSFREQ_ERROR) 2002 blktrc.flags |= ZFCP_BLK_REQ_ERROR; 2003 blktrc.inb_usage = 0; 2004 blktrc.outb_usage = req->qdio_req.qdio_outb_usage; 2005 2006 if (req->adapter->adapter_features & FSF_FEATURE_MEASUREMENT_DATA && 2007 !(req->status & ZFCP_STATUS_FSFREQ_ERROR)) { 2008 zfcp_sdev = sdev_to_zfcp(scsi->device); 2009 blktrc.flags |= ZFCP_BLK_LAT_VALID; 2010 blktrc.channel_lat = lat_in->channel_lat * ticks; 2011 blktrc.fabric_lat = lat_in->fabric_lat * ticks; 2012 2013 switch (req->qtcb->bottom.io.data_direction) { 2014 case FSF_DATADIR_DIF_READ_STRIP: 2015 case FSF_DATADIR_DIF_READ_CONVERT: 2016 case FSF_DATADIR_READ: 2017 lat = &zfcp_sdev->latencies.read; 2018 break; 2019 case FSF_DATADIR_DIF_WRITE_INSERT: 2020 case FSF_DATADIR_DIF_WRITE_CONVERT: 2021 case FSF_DATADIR_WRITE: 2022 lat = &zfcp_sdev->latencies.write; 2023 break; 2024 case FSF_DATADIR_CMND: 2025 lat = &zfcp_sdev->latencies.cmd; 2026 break; 2027 } 2028 2029 if (lat) { 2030 spin_lock(&zfcp_sdev->latencies.lock); 2031 zfcp_fsf_update_lat(&lat->channel, lat_in->channel_lat); 2032 zfcp_fsf_update_lat(&lat->fabric, lat_in->fabric_lat); 2033 lat->counter++; 2034 spin_unlock(&zfcp_sdev->latencies.lock); 2035 } 2036 } 2037 2038 blk_add_driver_data(scsi->request->q, scsi->request, &blktrc, 2039 sizeof(blktrc)); 2040 } 2041 2042 static void zfcp_fsf_fcp_handler_common(struct zfcp_fsf_req *req) 2043 { 2044 struct scsi_cmnd *scmnd = req->data; 2045 struct scsi_device *sdev = scmnd->device; 2046 struct zfcp_scsi_dev *zfcp_sdev; 2047 struct fsf_qtcb_header *header = &req->qtcb->header; 2048 2049 if (unlikely(req->status & ZFCP_STATUS_FSFREQ_ERROR)) 2050 return; 2051 2052 zfcp_sdev = sdev_to_zfcp(sdev); 2053 2054 switch (header->fsf_status) { 2055 case FSF_HANDLE_MISMATCH: 2056 case FSF_PORT_HANDLE_NOT_VALID: 2057 zfcp_erp_adapter_reopen(zfcp_sdev->port->adapter, 0, "fssfch1"); 2058 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 2059 break; 2060 case FSF_FCPLUN_NOT_VALID: 2061 case FSF_LUN_HANDLE_NOT_VALID: 2062 zfcp_erp_port_reopen(zfcp_sdev->port, 0, "fssfch2"); 2063 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 2064 break; 2065 case FSF_SERVICE_CLASS_NOT_SUPPORTED: 2066 zfcp_fsf_class_not_supp(req); 2067 break; 2068 case FSF_ACCESS_DENIED: 2069 zfcp_cfdc_lun_denied(sdev, &header->fsf_status_qual); 2070 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 2071 break; 2072 case FSF_DIRECTION_INDICATOR_NOT_VALID: 2073 dev_err(&req->adapter->ccw_device->dev, 2074 "Incorrect direction %d, LUN 0x%016Lx on port " 2075 "0x%016Lx closed\n", 2076 req->qtcb->bottom.io.data_direction, 2077 (unsigned long long)zfcp_scsi_dev_lun(sdev), 2078 (unsigned long long)zfcp_sdev->port->wwpn); 2079 zfcp_erp_adapter_shutdown(zfcp_sdev->port->adapter, 0, 2080 "fssfch3"); 2081 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 2082 break; 2083 case FSF_CMND_LENGTH_NOT_VALID: 2084 dev_err(&req->adapter->ccw_device->dev, 2085 "Incorrect CDB length %d, LUN 0x%016Lx on " 2086 "port 0x%016Lx closed\n", 2087 req->qtcb->bottom.io.fcp_cmnd_length, 2088 (unsigned long long)zfcp_scsi_dev_lun(sdev), 2089 (unsigned long long)zfcp_sdev->port->wwpn); 2090 zfcp_erp_adapter_shutdown(zfcp_sdev->port->adapter, 0, 2091 "fssfch4"); 2092 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 2093 break; 2094 case FSF_PORT_BOXED: 2095 zfcp_erp_set_port_status(zfcp_sdev->port, 2096 ZFCP_STATUS_COMMON_ACCESS_BOXED); 2097 zfcp_erp_port_reopen(zfcp_sdev->port, 2098 ZFCP_STATUS_COMMON_ERP_FAILED, "fssfch5"); 2099 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 2100 break; 2101 case FSF_LUN_BOXED: 2102 zfcp_erp_set_lun_status(sdev, ZFCP_STATUS_COMMON_ACCESS_BOXED); 2103 zfcp_erp_lun_reopen(sdev, ZFCP_STATUS_COMMON_ERP_FAILED, 2104 "fssfch6"); 2105 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 2106 break; 2107 case FSF_ADAPTER_STATUS_AVAILABLE: 2108 if (header->fsf_status_qual.word[0] == 2109 FSF_SQ_INVOKE_LINK_TEST_PROCEDURE) 2110 zfcp_fc_test_link(zfcp_sdev->port); 2111 req->status |= ZFCP_STATUS_FSFREQ_ERROR; 2112 break; 2113 } 2114 } 2115 2116 static void zfcp_fsf_fcp_cmnd_handler(struct zfcp_fsf_req *req) 2117 { 2118 struct scsi_cmnd *scpnt; 2119 struct fcp_resp_with_ext *fcp_rsp; 2120 unsigned long flags; 2121 2122 read_lock_irqsave(&req->adapter->abort_lock, flags); 2123 2124 scpnt = req->data; 2125 if (unlikely(!scpnt)) { 2126 read_unlock_irqrestore(&req->adapter->abort_lock, flags); 2127 return; 2128 } 2129 2130 zfcp_fsf_fcp_handler_common(req); 2131 2132 if (unlikely(req->status & ZFCP_STATUS_FSFREQ_ERROR)) { 2133 set_host_byte(scpnt, DID_TRANSPORT_DISRUPTED); 2134 goto skip_fsfstatus; 2135 } 2136 2137 switch (req->qtcb->header.fsf_status) { 2138 case FSF_INCONSISTENT_PROT_DATA: 2139 case FSF_INVALID_PROT_PARM: 2140 set_host_byte(scpnt, DID_ERROR); 2141 goto skip_fsfstatus; 2142 case FSF_BLOCK_GUARD_CHECK_FAILURE: 2143 zfcp_scsi_dif_sense_error(scpnt, 0x1); 2144 goto skip_fsfstatus; 2145 case FSF_APP_TAG_CHECK_FAILURE: 2146 zfcp_scsi_dif_sense_error(scpnt, 0x2); 2147 goto skip_fsfstatus; 2148 case FSF_REF_TAG_CHECK_FAILURE: 2149 zfcp_scsi_dif_sense_error(scpnt, 0x3); 2150 goto skip_fsfstatus; 2151 } 2152 fcp_rsp = (struct fcp_resp_with_ext *) &req->qtcb->bottom.io.fcp_rsp; 2153 zfcp_fc_eval_fcp_rsp(fcp_rsp, scpnt); 2154 2155 skip_fsfstatus: 2156 zfcp_fsf_req_trace(req, scpnt); 2157 zfcp_dbf_scsi_result(scpnt, req); 2158 2159 scpnt->host_scribble = NULL; 2160 (scpnt->scsi_done) (scpnt); 2161 /* 2162 * We must hold this lock until scsi_done has been called. 2163 * Otherwise we may call scsi_done after abort regarding this 2164 * command has completed. 2165 * Note: scsi_done must not block! 2166 */ 2167 read_unlock_irqrestore(&req->adapter->abort_lock, flags); 2168 } 2169 2170 static int zfcp_fsf_set_data_dir(struct scsi_cmnd *scsi_cmnd, u32 *data_dir) 2171 { 2172 switch (scsi_get_prot_op(scsi_cmnd)) { 2173 case SCSI_PROT_NORMAL: 2174 switch (scsi_cmnd->sc_data_direction) { 2175 case DMA_NONE: 2176 *data_dir = FSF_DATADIR_CMND; 2177 break; 2178 case DMA_FROM_DEVICE: 2179 *data_dir = FSF_DATADIR_READ; 2180 break; 2181 case DMA_TO_DEVICE: 2182 *data_dir = FSF_DATADIR_WRITE; 2183 break; 2184 case DMA_BIDIRECTIONAL: 2185 return -EINVAL; 2186 } 2187 break; 2188 2189 case SCSI_PROT_READ_STRIP: 2190 *data_dir = FSF_DATADIR_DIF_READ_STRIP; 2191 break; 2192 case SCSI_PROT_WRITE_INSERT: 2193 *data_dir = FSF_DATADIR_DIF_WRITE_INSERT; 2194 break; 2195 case SCSI_PROT_READ_PASS: 2196 *data_dir = FSF_DATADIR_DIF_READ_CONVERT; 2197 break; 2198 case SCSI_PROT_WRITE_PASS: 2199 *data_dir = FSF_DATADIR_DIF_WRITE_CONVERT; 2200 break; 2201 default: 2202 return -EINVAL; 2203 } 2204 2205 return 0; 2206 } 2207 2208 /** 2209 * zfcp_fsf_fcp_cmnd - initiate an FCP command (for a SCSI command) 2210 * @scsi_cmnd: scsi command to be sent 2211 */ 2212 int zfcp_fsf_fcp_cmnd(struct scsi_cmnd *scsi_cmnd) 2213 { 2214 struct zfcp_fsf_req *req; 2215 struct fcp_cmnd *fcp_cmnd; 2216 u8 sbtype = SBAL_SFLAGS0_TYPE_READ; 2217 int retval = -EIO; 2218 struct scsi_device *sdev = scsi_cmnd->device; 2219 struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(sdev); 2220 struct zfcp_adapter *adapter = zfcp_sdev->port->adapter; 2221 struct zfcp_qdio *qdio = adapter->qdio; 2222 struct fsf_qtcb_bottom_io *io; 2223 unsigned long flags; 2224 2225 if (unlikely(!(atomic_read(&zfcp_sdev->status) & 2226 ZFCP_STATUS_COMMON_UNBLOCKED))) 2227 return -EBUSY; 2228 2229 spin_lock_irqsave(&qdio->req_q_lock, flags); 2230 if (atomic_read(&qdio->req_q_free) <= 0) { 2231 atomic_inc(&qdio->req_q_full); 2232 goto out; 2233 } 2234 2235 if (scsi_cmnd->sc_data_direction == DMA_TO_DEVICE) 2236 sbtype = SBAL_SFLAGS0_TYPE_WRITE; 2237 2238 req = zfcp_fsf_req_create(qdio, FSF_QTCB_FCP_CMND, 2239 sbtype, adapter->pool.scsi_req); 2240 2241 if (IS_ERR(req)) { 2242 retval = PTR_ERR(req); 2243 goto out; 2244 } 2245 2246 scsi_cmnd->host_scribble = (unsigned char *) req->req_id; 2247 2248 io = &req->qtcb->bottom.io; 2249 req->status |= ZFCP_STATUS_FSFREQ_CLEANUP; 2250 req->data = scsi_cmnd; 2251 req->handler = zfcp_fsf_fcp_cmnd_handler; 2252 req->qtcb->header.lun_handle = zfcp_sdev->lun_handle; 2253 req->qtcb->header.port_handle = zfcp_sdev->port->handle; 2254 io->service_class = FSF_CLASS_3; 2255 io->fcp_cmnd_length = FCP_CMND_LEN; 2256 2257 if (scsi_get_prot_op(scsi_cmnd) != SCSI_PROT_NORMAL) { 2258 io->data_block_length = scsi_cmnd->device->sector_size; 2259 io->ref_tag_value = scsi_get_lba(scsi_cmnd) & 0xFFFFFFFF; 2260 } 2261 2262 if (zfcp_fsf_set_data_dir(scsi_cmnd, &io->data_direction)) 2263 goto failed_scsi_cmnd; 2264 2265 fcp_cmnd = (struct fcp_cmnd *) &req->qtcb->bottom.io.fcp_cmnd; 2266 zfcp_fc_scsi_to_fcp(fcp_cmnd, scsi_cmnd, 0); 2267 2268 if (scsi_prot_sg_count(scsi_cmnd)) { 2269 zfcp_qdio_set_data_div(qdio, &req->qdio_req, 2270 scsi_prot_sg_count(scsi_cmnd)); 2271 retval = zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, 2272 scsi_prot_sglist(scsi_cmnd)); 2273 if (retval) 2274 goto failed_scsi_cmnd; 2275 io->prot_data_length = zfcp_qdio_real_bytes( 2276 scsi_prot_sglist(scsi_cmnd)); 2277 } 2278 2279 retval = zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, 2280 scsi_sglist(scsi_cmnd)); 2281 if (unlikely(retval)) 2282 goto failed_scsi_cmnd; 2283 2284 zfcp_qdio_set_sbale_last(adapter->qdio, &req->qdio_req); 2285 if (zfcp_adapter_multi_buffer_active(adapter)) 2286 zfcp_qdio_set_scount(qdio, &req->qdio_req); 2287 2288 retval = zfcp_fsf_req_send(req); 2289 if (unlikely(retval)) 2290 goto failed_scsi_cmnd; 2291 2292 goto out; 2293 2294 failed_scsi_cmnd: 2295 zfcp_fsf_req_free(req); 2296 scsi_cmnd->host_scribble = NULL; 2297 out: 2298 spin_unlock_irqrestore(&qdio->req_q_lock, flags); 2299 return retval; 2300 } 2301 2302 static void zfcp_fsf_fcp_task_mgmt_handler(struct zfcp_fsf_req *req) 2303 { 2304 struct fcp_resp_with_ext *fcp_rsp; 2305 struct fcp_resp_rsp_info *rsp_info; 2306 2307 zfcp_fsf_fcp_handler_common(req); 2308 2309 fcp_rsp = (struct fcp_resp_with_ext *) &req->qtcb->bottom.io.fcp_rsp; 2310 rsp_info = (struct fcp_resp_rsp_info *) &fcp_rsp[1]; 2311 2312 if ((rsp_info->rsp_code != FCP_TMF_CMPL) || 2313 (req->status & ZFCP_STATUS_FSFREQ_ERROR)) 2314 req->status |= ZFCP_STATUS_FSFREQ_TMFUNCFAILED; 2315 } 2316 2317 /** 2318 * zfcp_fsf_fcp_task_mgmt - send SCSI task management command 2319 * @scmnd: SCSI command to send the task management command for 2320 * @tm_flags: unsigned byte for task management flags 2321 * Returns: on success pointer to struct fsf_req, NULL otherwise 2322 */ 2323 struct zfcp_fsf_req *zfcp_fsf_fcp_task_mgmt(struct scsi_cmnd *scmnd, 2324 u8 tm_flags) 2325 { 2326 struct zfcp_fsf_req *req = NULL; 2327 struct fcp_cmnd *fcp_cmnd; 2328 struct zfcp_scsi_dev *zfcp_sdev = sdev_to_zfcp(scmnd->device); 2329 struct zfcp_qdio *qdio = zfcp_sdev->port->adapter->qdio; 2330 2331 if (unlikely(!(atomic_read(&zfcp_sdev->status) & 2332 ZFCP_STATUS_COMMON_UNBLOCKED))) 2333 return NULL; 2334 2335 spin_lock_irq(&qdio->req_q_lock); 2336 if (zfcp_qdio_sbal_get(qdio)) 2337 goto out; 2338 2339 req = zfcp_fsf_req_create(qdio, FSF_QTCB_FCP_CMND, 2340 SBAL_SFLAGS0_TYPE_WRITE, 2341 qdio->adapter->pool.scsi_req); 2342 2343 if (IS_ERR(req)) { 2344 req = NULL; 2345 goto out; 2346 } 2347 2348 req->data = scmnd; 2349 req->handler = zfcp_fsf_fcp_task_mgmt_handler; 2350 req->qtcb->header.lun_handle = zfcp_sdev->lun_handle; 2351 req->qtcb->header.port_handle = zfcp_sdev->port->handle; 2352 req->qtcb->bottom.io.data_direction = FSF_DATADIR_CMND; 2353 req->qtcb->bottom.io.service_class = FSF_CLASS_3; 2354 req->qtcb->bottom.io.fcp_cmnd_length = FCP_CMND_LEN; 2355 2356 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 2357 2358 fcp_cmnd = (struct fcp_cmnd *) &req->qtcb->bottom.io.fcp_cmnd; 2359 zfcp_fc_scsi_to_fcp(fcp_cmnd, scmnd, tm_flags); 2360 2361 zfcp_fsf_start_timer(req, ZFCP_SCSI_ER_TIMEOUT); 2362 if (!zfcp_fsf_req_send(req)) 2363 goto out; 2364 2365 zfcp_fsf_req_free(req); 2366 req = NULL; 2367 out: 2368 spin_unlock_irq(&qdio->req_q_lock); 2369 return req; 2370 } 2371 2372 static void zfcp_fsf_control_file_handler(struct zfcp_fsf_req *req) 2373 { 2374 } 2375 2376 /** 2377 * zfcp_fsf_control_file - control file upload/download 2378 * @adapter: pointer to struct zfcp_adapter 2379 * @fsf_cfdc: pointer to struct zfcp_fsf_cfdc 2380 * Returns: on success pointer to struct zfcp_fsf_req, NULL otherwise 2381 */ 2382 struct zfcp_fsf_req *zfcp_fsf_control_file(struct zfcp_adapter *adapter, 2383 struct zfcp_fsf_cfdc *fsf_cfdc) 2384 { 2385 struct zfcp_qdio *qdio = adapter->qdio; 2386 struct zfcp_fsf_req *req = NULL; 2387 struct fsf_qtcb_bottom_support *bottom; 2388 int retval = -EIO; 2389 u8 direction; 2390 2391 if (!(adapter->adapter_features & FSF_FEATURE_CFDC)) 2392 return ERR_PTR(-EOPNOTSUPP); 2393 2394 switch (fsf_cfdc->command) { 2395 case FSF_QTCB_DOWNLOAD_CONTROL_FILE: 2396 direction = SBAL_SFLAGS0_TYPE_WRITE; 2397 break; 2398 case FSF_QTCB_UPLOAD_CONTROL_FILE: 2399 direction = SBAL_SFLAGS0_TYPE_READ; 2400 break; 2401 default: 2402 return ERR_PTR(-EINVAL); 2403 } 2404 2405 spin_lock_irq(&qdio->req_q_lock); 2406 if (zfcp_qdio_sbal_get(qdio)) 2407 goto out; 2408 2409 req = zfcp_fsf_req_create(qdio, fsf_cfdc->command, direction, NULL); 2410 if (IS_ERR(req)) { 2411 retval = -EPERM; 2412 goto out; 2413 } 2414 2415 req->handler = zfcp_fsf_control_file_handler; 2416 2417 bottom = &req->qtcb->bottom.support; 2418 bottom->operation_subtype = FSF_CFDC_OPERATION_SUBTYPE; 2419 bottom->option = fsf_cfdc->option; 2420 2421 retval = zfcp_qdio_sbals_from_sg(qdio, &req->qdio_req, fsf_cfdc->sg); 2422 2423 if (retval || 2424 (zfcp_qdio_real_bytes(fsf_cfdc->sg) != ZFCP_CFDC_MAX_SIZE)) { 2425 zfcp_fsf_req_free(req); 2426 retval = -EIO; 2427 goto out; 2428 } 2429 zfcp_qdio_set_sbale_last(qdio, &req->qdio_req); 2430 if (zfcp_adapter_multi_buffer_active(adapter)) 2431 zfcp_qdio_set_scount(qdio, &req->qdio_req); 2432 2433 zfcp_fsf_start_timer(req, ZFCP_FSF_REQUEST_TIMEOUT); 2434 retval = zfcp_fsf_req_send(req); 2435 out: 2436 spin_unlock_irq(&qdio->req_q_lock); 2437 2438 if (!retval) { 2439 wait_for_completion(&req->completion); 2440 return req; 2441 } 2442 return ERR_PTR(retval); 2443 } 2444 2445 /** 2446 * zfcp_fsf_reqid_check - validate req_id contained in SBAL returned by QDIO 2447 * @adapter: pointer to struct zfcp_adapter 2448 * @sbal_idx: response queue index of SBAL to be processed 2449 */ 2450 void zfcp_fsf_reqid_check(struct zfcp_qdio *qdio, int sbal_idx) 2451 { 2452 struct zfcp_adapter *adapter = qdio->adapter; 2453 struct qdio_buffer *sbal = qdio->res_q[sbal_idx]; 2454 struct qdio_buffer_element *sbale; 2455 struct zfcp_fsf_req *fsf_req; 2456 unsigned long req_id; 2457 int idx; 2458 2459 for (idx = 0; idx < QDIO_MAX_ELEMENTS_PER_BUFFER; idx++) { 2460 2461 sbale = &sbal->element[idx]; 2462 req_id = (unsigned long) sbale->addr; 2463 fsf_req = zfcp_reqlist_find_rm(adapter->req_list, req_id); 2464 2465 if (!fsf_req) { 2466 /* 2467 * Unknown request means that we have potentially memory 2468 * corruption and must stop the machine immediately. 2469 */ 2470 zfcp_qdio_siosl(adapter); 2471 panic("error: unknown req_id (%lx) on adapter %s.\n", 2472 req_id, dev_name(&adapter->ccw_device->dev)); 2473 } 2474 2475 fsf_req->qdio_req.sbal_response = sbal_idx; 2476 zfcp_fsf_req_complete(fsf_req); 2477 2478 if (likely(sbale->eflags & SBAL_EFLAGS_LAST_ENTRY)) 2479 break; 2480 } 2481 } 2482 2483 struct zfcp_fsf_req *zfcp_fsf_get_req(struct zfcp_qdio *qdio, 2484 struct qdio_buffer *sbal) 2485 { 2486 struct qdio_buffer_element *sbale = &sbal->element[0]; 2487 u64 req_id = (unsigned long) sbale->addr; 2488 2489 return zfcp_reqlist_find(qdio->adapter->req_list, req_id); 2490 } 2491