1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3 * QLogic Fibre Channel HBA Driver
4 * Copyright (c) 2003-2014 QLogic Corporation
5 */
6 #include "qla_def.h"
7 #include "qla_target.h"
8
9 #include <linux/kthread.h>
10 #include <linux/vmalloc.h>
11 #include <linux/slab.h>
12 #include <linux/delay.h>
13
14 static int qla24xx_vport_disable(struct fc_vport *, bool);
15
16 /* SYSFS attributes --------------------------------------------------------- */
17
18 static ssize_t
qla2x00_sysfs_read_fw_dump(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)19 qla2x00_sysfs_read_fw_dump(struct file *filp, struct kobject *kobj,
20 struct bin_attribute *bin_attr,
21 char *buf, loff_t off, size_t count)
22 {
23 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
24 struct device, kobj)));
25 struct qla_hw_data *ha = vha->hw;
26 int rval = 0;
27
28 if (!(ha->fw_dump_reading || ha->mctp_dump_reading ||
29 ha->mpi_fw_dump_reading))
30 return 0;
31
32 mutex_lock(&ha->optrom_mutex);
33 if (IS_P3P_TYPE(ha)) {
34 if (off < ha->md_template_size) {
35 rval = memory_read_from_buffer(buf, count,
36 &off, ha->md_tmplt_hdr, ha->md_template_size);
37 } else {
38 off -= ha->md_template_size;
39 rval = memory_read_from_buffer(buf, count,
40 &off, ha->md_dump, ha->md_dump_size);
41 }
42 } else if (ha->mctp_dumped && ha->mctp_dump_reading) {
43 rval = memory_read_from_buffer(buf, count, &off, ha->mctp_dump,
44 MCTP_DUMP_SIZE);
45 } else if (ha->mpi_fw_dumped && ha->mpi_fw_dump_reading) {
46 rval = memory_read_from_buffer(buf, count, &off,
47 ha->mpi_fw_dump,
48 ha->mpi_fw_dump_len);
49 } else if (ha->fw_dump_reading) {
50 rval = memory_read_from_buffer(buf, count, &off, ha->fw_dump,
51 ha->fw_dump_len);
52 } else {
53 rval = 0;
54 }
55 mutex_unlock(&ha->optrom_mutex);
56 return rval;
57 }
58
59 static ssize_t
qla2x00_sysfs_write_fw_dump(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)60 qla2x00_sysfs_write_fw_dump(struct file *filp, struct kobject *kobj,
61 struct bin_attribute *bin_attr,
62 char *buf, loff_t off, size_t count)
63 {
64 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
65 struct device, kobj)));
66 struct qla_hw_data *ha = vha->hw;
67 int reading;
68
69 if (off != 0)
70 return (0);
71
72 reading = simple_strtol(buf, NULL, 10);
73 switch (reading) {
74 case 0:
75 if (!ha->fw_dump_reading)
76 break;
77
78 ql_log(ql_log_info, vha, 0x705d,
79 "Firmware dump cleared on (%ld).\n", vha->host_no);
80
81 if (IS_P3P_TYPE(ha)) {
82 qla82xx_md_free(vha);
83 qla82xx_md_prep(vha);
84 }
85 ha->fw_dump_reading = 0;
86 ha->fw_dumped = false;
87 break;
88 case 1:
89 if (ha->fw_dumped && !ha->fw_dump_reading) {
90 ha->fw_dump_reading = 1;
91
92 ql_log(ql_log_info, vha, 0x705e,
93 "Raw firmware dump ready for read on (%ld).\n",
94 vha->host_no);
95 }
96 break;
97 case 2:
98 qla2x00_alloc_fw_dump(vha);
99 break;
100 case 3:
101 if (IS_QLA82XX(ha)) {
102 qla82xx_idc_lock(ha);
103 qla82xx_set_reset_owner(vha);
104 qla82xx_idc_unlock(ha);
105 } else if (IS_QLA8044(ha)) {
106 qla8044_idc_lock(ha);
107 qla82xx_set_reset_owner(vha);
108 qla8044_idc_unlock(ha);
109 } else {
110 qla2x00_system_error(vha);
111 }
112 break;
113 case 4:
114 if (IS_P3P_TYPE(ha)) {
115 if (ha->md_tmplt_hdr)
116 ql_dbg(ql_dbg_user, vha, 0x705b,
117 "MiniDump supported with this firmware.\n");
118 else
119 ql_dbg(ql_dbg_user, vha, 0x709d,
120 "MiniDump not supported with this firmware.\n");
121 }
122 break;
123 case 5:
124 if (IS_P3P_TYPE(ha))
125 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
126 break;
127 case 6:
128 if (!ha->mctp_dump_reading)
129 break;
130 ql_log(ql_log_info, vha, 0x70c1,
131 "MCTP dump cleared on (%ld).\n", vha->host_no);
132 ha->mctp_dump_reading = 0;
133 ha->mctp_dumped = 0;
134 break;
135 case 7:
136 if (ha->mctp_dumped && !ha->mctp_dump_reading) {
137 ha->mctp_dump_reading = 1;
138 ql_log(ql_log_info, vha, 0x70c2,
139 "Raw mctp dump ready for read on (%ld).\n",
140 vha->host_no);
141 }
142 break;
143 case 8:
144 if (!ha->mpi_fw_dump_reading)
145 break;
146 ql_log(ql_log_info, vha, 0x70e7,
147 "MPI firmware dump cleared on (%ld).\n", vha->host_no);
148 ha->mpi_fw_dump_reading = 0;
149 ha->mpi_fw_dumped = 0;
150 break;
151 case 9:
152 if (ha->mpi_fw_dumped && !ha->mpi_fw_dump_reading) {
153 ha->mpi_fw_dump_reading = 1;
154 ql_log(ql_log_info, vha, 0x70e8,
155 "Raw MPI firmware dump ready for read on (%ld).\n",
156 vha->host_no);
157 }
158 break;
159 case 10:
160 if (IS_QLA27XX(ha) || IS_QLA28XX(ha)) {
161 ql_log(ql_log_info, vha, 0x70e9,
162 "Issuing MPI firmware dump on host#%ld.\n",
163 vha->host_no);
164 ha->isp_ops->mpi_fw_dump(vha, 0);
165 }
166 break;
167 }
168 return count;
169 }
170
171 static struct bin_attribute sysfs_fw_dump_attr = {
172 .attr = {
173 .name = "fw_dump",
174 .mode = S_IRUSR | S_IWUSR,
175 },
176 .size = 0,
177 .read = qla2x00_sysfs_read_fw_dump,
178 .write = qla2x00_sysfs_write_fw_dump,
179 };
180
181 static ssize_t
qla2x00_sysfs_read_nvram(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)182 qla2x00_sysfs_read_nvram(struct file *filp, struct kobject *kobj,
183 struct bin_attribute *bin_attr,
184 char *buf, loff_t off, size_t count)
185 {
186 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
187 struct device, kobj)));
188 struct qla_hw_data *ha = vha->hw;
189 uint32_t faddr;
190 struct active_regions active_regions = { };
191
192 if (!capable(CAP_SYS_ADMIN))
193 return 0;
194
195 mutex_lock(&ha->optrom_mutex);
196 if (qla2x00_chip_is_down(vha)) {
197 mutex_unlock(&ha->optrom_mutex);
198 return -EAGAIN;
199 }
200
201 if (!IS_NOCACHE_VPD_TYPE(ha)) {
202 mutex_unlock(&ha->optrom_mutex);
203 goto skip;
204 }
205
206 faddr = ha->flt_region_nvram;
207 if (IS_QLA28XX(ha)) {
208 qla28xx_get_aux_images(vha, &active_regions);
209 if (active_regions.aux.vpd_nvram == QLA27XX_SECONDARY_IMAGE)
210 faddr = ha->flt_region_nvram_sec;
211 }
212 ha->isp_ops->read_optrom(vha, ha->nvram, faddr << 2, ha->nvram_size);
213
214 mutex_unlock(&ha->optrom_mutex);
215
216 skip:
217 return memory_read_from_buffer(buf, count, &off, ha->nvram,
218 ha->nvram_size);
219 }
220
221 static ssize_t
qla2x00_sysfs_write_nvram(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)222 qla2x00_sysfs_write_nvram(struct file *filp, struct kobject *kobj,
223 struct bin_attribute *bin_attr,
224 char *buf, loff_t off, size_t count)
225 {
226 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
227 struct device, kobj)));
228 struct qla_hw_data *ha = vha->hw;
229 uint16_t cnt;
230
231 if (!capable(CAP_SYS_ADMIN) || off != 0 || count != ha->nvram_size ||
232 !ha->isp_ops->write_nvram)
233 return -EINVAL;
234
235 /* Checksum NVRAM. */
236 if (IS_FWI2_CAPABLE(ha)) {
237 __le32 *iter = (__force __le32 *)buf;
238 uint32_t chksum;
239
240 chksum = 0;
241 for (cnt = 0; cnt < ((count >> 2) - 1); cnt++, iter++)
242 chksum += le32_to_cpu(*iter);
243 chksum = ~chksum + 1;
244 *iter = cpu_to_le32(chksum);
245 } else {
246 uint8_t *iter;
247 uint8_t chksum;
248
249 iter = (uint8_t *)buf;
250 chksum = 0;
251 for (cnt = 0; cnt < count - 1; cnt++)
252 chksum += *iter++;
253 chksum = ~chksum + 1;
254 *iter = chksum;
255 }
256
257 if (qla2x00_wait_for_hba_online(vha) != QLA_SUCCESS) {
258 ql_log(ql_log_warn, vha, 0x705f,
259 "HBA not online, failing NVRAM update.\n");
260 return -EAGAIN;
261 }
262
263 mutex_lock(&ha->optrom_mutex);
264 if (qla2x00_chip_is_down(vha)) {
265 mutex_unlock(&ha->optrom_mutex);
266 return -EAGAIN;
267 }
268
269 /* Write NVRAM. */
270 ha->isp_ops->write_nvram(vha, buf, ha->nvram_base, count);
271 ha->isp_ops->read_nvram(vha, ha->nvram, ha->nvram_base,
272 count);
273 mutex_unlock(&ha->optrom_mutex);
274
275 ql_dbg(ql_dbg_user, vha, 0x7060,
276 "Setting ISP_ABORT_NEEDED\n");
277 /* NVRAM settings take effect immediately. */
278 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
279 qla2xxx_wake_dpc(vha);
280 qla2x00_wait_for_chip_reset(vha);
281
282 return count;
283 }
284
285 static struct bin_attribute sysfs_nvram_attr = {
286 .attr = {
287 .name = "nvram",
288 .mode = S_IRUSR | S_IWUSR,
289 },
290 .size = 512,
291 .read = qla2x00_sysfs_read_nvram,
292 .write = qla2x00_sysfs_write_nvram,
293 };
294
295 static ssize_t
qla2x00_sysfs_read_optrom(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)296 qla2x00_sysfs_read_optrom(struct file *filp, struct kobject *kobj,
297 struct bin_attribute *bin_attr,
298 char *buf, loff_t off, size_t count)
299 {
300 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
301 struct device, kobj)));
302 struct qla_hw_data *ha = vha->hw;
303 ssize_t rval = 0;
304
305 mutex_lock(&ha->optrom_mutex);
306
307 if (ha->optrom_state != QLA_SREADING)
308 goto out;
309
310 rval = memory_read_from_buffer(buf, count, &off, ha->optrom_buffer,
311 ha->optrom_region_size);
312
313 out:
314 mutex_unlock(&ha->optrom_mutex);
315
316 return rval;
317 }
318
319 static ssize_t
qla2x00_sysfs_write_optrom(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)320 qla2x00_sysfs_write_optrom(struct file *filp, struct kobject *kobj,
321 struct bin_attribute *bin_attr,
322 char *buf, loff_t off, size_t count)
323 {
324 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
325 struct device, kobj)));
326 struct qla_hw_data *ha = vha->hw;
327
328 mutex_lock(&ha->optrom_mutex);
329
330 if (ha->optrom_state != QLA_SWRITING) {
331 mutex_unlock(&ha->optrom_mutex);
332 return -EINVAL;
333 }
334 if (off > ha->optrom_region_size) {
335 mutex_unlock(&ha->optrom_mutex);
336 return -ERANGE;
337 }
338 if (off + count > ha->optrom_region_size)
339 count = ha->optrom_region_size - off;
340
341 memcpy(&ha->optrom_buffer[off], buf, count);
342 mutex_unlock(&ha->optrom_mutex);
343
344 return count;
345 }
346
347 static struct bin_attribute sysfs_optrom_attr = {
348 .attr = {
349 .name = "optrom",
350 .mode = S_IRUSR | S_IWUSR,
351 },
352 .size = 0,
353 .read = qla2x00_sysfs_read_optrom,
354 .write = qla2x00_sysfs_write_optrom,
355 };
356
357 static ssize_t
qla2x00_sysfs_write_optrom_ctl(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)358 qla2x00_sysfs_write_optrom_ctl(struct file *filp, struct kobject *kobj,
359 struct bin_attribute *bin_attr,
360 char *buf, loff_t off, size_t count)
361 {
362 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
363 struct device, kobj)));
364 struct qla_hw_data *ha = vha->hw;
365 uint32_t start = 0;
366 uint32_t size = ha->optrom_size;
367 int val, valid;
368 ssize_t rval = count;
369
370 if (off)
371 return -EINVAL;
372
373 if (unlikely(pci_channel_offline(ha->pdev)))
374 return -EAGAIN;
375
376 if (sscanf(buf, "%d:%x:%x", &val, &start, &size) < 1)
377 return -EINVAL;
378 if (start > ha->optrom_size)
379 return -EINVAL;
380 if (size > ha->optrom_size - start)
381 size = ha->optrom_size - start;
382
383 mutex_lock(&ha->optrom_mutex);
384 if (qla2x00_chip_is_down(vha)) {
385 mutex_unlock(&ha->optrom_mutex);
386 return -EAGAIN;
387 }
388 switch (val) {
389 case 0:
390 if (ha->optrom_state != QLA_SREADING &&
391 ha->optrom_state != QLA_SWRITING) {
392 rval = -EINVAL;
393 goto out;
394 }
395 ha->optrom_state = QLA_SWAITING;
396
397 ql_dbg(ql_dbg_user, vha, 0x7061,
398 "Freeing flash region allocation -- 0x%x bytes.\n",
399 ha->optrom_region_size);
400
401 vfree(ha->optrom_buffer);
402 ha->optrom_buffer = NULL;
403 break;
404 case 1:
405 if (ha->optrom_state != QLA_SWAITING) {
406 rval = -EINVAL;
407 goto out;
408 }
409
410 ha->optrom_region_start = start;
411 ha->optrom_region_size = size;
412
413 ha->optrom_state = QLA_SREADING;
414 ha->optrom_buffer = vzalloc(ha->optrom_region_size);
415 if (ha->optrom_buffer == NULL) {
416 ql_log(ql_log_warn, vha, 0x7062,
417 "Unable to allocate memory for optrom retrieval "
418 "(%x).\n", ha->optrom_region_size);
419
420 ha->optrom_state = QLA_SWAITING;
421 rval = -ENOMEM;
422 goto out;
423 }
424
425 if (qla2x00_wait_for_hba_online(vha) != QLA_SUCCESS) {
426 ql_log(ql_log_warn, vha, 0x7063,
427 "HBA not online, failing NVRAM update.\n");
428 rval = -EAGAIN;
429 goto out;
430 }
431
432 ql_dbg(ql_dbg_user, vha, 0x7064,
433 "Reading flash region -- 0x%x/0x%x.\n",
434 ha->optrom_region_start, ha->optrom_region_size);
435
436 ha->isp_ops->read_optrom(vha, ha->optrom_buffer,
437 ha->optrom_region_start, ha->optrom_region_size);
438 break;
439 case 2:
440 if (ha->optrom_state != QLA_SWAITING) {
441 rval = -EINVAL;
442 goto out;
443 }
444
445 /*
446 * We need to be more restrictive on which FLASH regions are
447 * allowed to be updated via user-space. Regions accessible
448 * via this method include:
449 *
450 * ISP21xx/ISP22xx/ISP23xx type boards:
451 *
452 * 0x000000 -> 0x020000 -- Boot code.
453 *
454 * ISP2322/ISP24xx type boards:
455 *
456 * 0x000000 -> 0x07ffff -- Boot code.
457 * 0x080000 -> 0x0fffff -- Firmware.
458 *
459 * ISP25xx type boards:
460 *
461 * 0x000000 -> 0x07ffff -- Boot code.
462 * 0x080000 -> 0x0fffff -- Firmware.
463 * 0x120000 -> 0x12ffff -- VPD and HBA parameters.
464 *
465 * > ISP25xx type boards:
466 *
467 * None -- should go through BSG.
468 */
469 valid = 0;
470 if (ha->optrom_size == OPTROM_SIZE_2300 && start == 0)
471 valid = 1;
472 else if (IS_QLA24XX_TYPE(ha) || IS_QLA25XX(ha))
473 valid = 1;
474 if (!valid) {
475 ql_log(ql_log_warn, vha, 0x7065,
476 "Invalid start region 0x%x/0x%x.\n", start, size);
477 rval = -EINVAL;
478 goto out;
479 }
480
481 ha->optrom_region_start = start;
482 ha->optrom_region_size = size;
483
484 ha->optrom_state = QLA_SWRITING;
485 ha->optrom_buffer = vzalloc(ha->optrom_region_size);
486 if (ha->optrom_buffer == NULL) {
487 ql_log(ql_log_warn, vha, 0x7066,
488 "Unable to allocate memory for optrom update "
489 "(%x)\n", ha->optrom_region_size);
490
491 ha->optrom_state = QLA_SWAITING;
492 rval = -ENOMEM;
493 goto out;
494 }
495
496 ql_dbg(ql_dbg_user, vha, 0x7067,
497 "Staging flash region write -- 0x%x/0x%x.\n",
498 ha->optrom_region_start, ha->optrom_region_size);
499
500 break;
501 case 3:
502 if (ha->optrom_state != QLA_SWRITING) {
503 rval = -EINVAL;
504 goto out;
505 }
506
507 if (qla2x00_wait_for_hba_online(vha) != QLA_SUCCESS) {
508 ql_log(ql_log_warn, vha, 0x7068,
509 "HBA not online, failing flash update.\n");
510 rval = -EAGAIN;
511 goto out;
512 }
513
514 ql_dbg(ql_dbg_user, vha, 0x7069,
515 "Writing flash region -- 0x%x/0x%x.\n",
516 ha->optrom_region_start, ha->optrom_region_size);
517
518 rval = ha->isp_ops->write_optrom(vha, ha->optrom_buffer,
519 ha->optrom_region_start, ha->optrom_region_size);
520 if (rval)
521 rval = -EIO;
522 break;
523 default:
524 rval = -EINVAL;
525 }
526
527 out:
528 mutex_unlock(&ha->optrom_mutex);
529 return rval;
530 }
531
532 static struct bin_attribute sysfs_optrom_ctl_attr = {
533 .attr = {
534 .name = "optrom_ctl",
535 .mode = S_IWUSR,
536 },
537 .size = 0,
538 .write = qla2x00_sysfs_write_optrom_ctl,
539 };
540
541 static ssize_t
qla2x00_sysfs_read_vpd(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)542 qla2x00_sysfs_read_vpd(struct file *filp, struct kobject *kobj,
543 struct bin_attribute *bin_attr,
544 char *buf, loff_t off, size_t count)
545 {
546 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
547 struct device, kobj)));
548 struct qla_hw_data *ha = vha->hw;
549 uint32_t faddr;
550 struct active_regions active_regions = { };
551
552 if (unlikely(pci_channel_offline(ha->pdev)))
553 return -EAGAIN;
554
555 if (!capable(CAP_SYS_ADMIN))
556 return -EINVAL;
557
558 if (!IS_NOCACHE_VPD_TYPE(ha))
559 goto skip;
560
561 faddr = ha->flt_region_vpd << 2;
562
563 if (IS_QLA28XX(ha)) {
564 qla28xx_get_aux_images(vha, &active_regions);
565 if (active_regions.aux.vpd_nvram == QLA27XX_SECONDARY_IMAGE)
566 faddr = ha->flt_region_vpd_sec << 2;
567
568 ql_dbg(ql_dbg_init, vha, 0x7070,
569 "Loading %s nvram image.\n",
570 active_regions.aux.vpd_nvram == QLA27XX_PRIMARY_IMAGE ?
571 "primary" : "secondary");
572 }
573
574 mutex_lock(&ha->optrom_mutex);
575 if (qla2x00_chip_is_down(vha)) {
576 mutex_unlock(&ha->optrom_mutex);
577 return -EAGAIN;
578 }
579
580 ha->isp_ops->read_optrom(vha, ha->vpd, faddr, ha->vpd_size);
581 mutex_unlock(&ha->optrom_mutex);
582
583 ha->isp_ops->read_optrom(vha, ha->vpd, faddr, ha->vpd_size);
584 skip:
585 return memory_read_from_buffer(buf, count, &off, ha->vpd, ha->vpd_size);
586 }
587
588 static ssize_t
qla2x00_sysfs_write_vpd(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)589 qla2x00_sysfs_write_vpd(struct file *filp, struct kobject *kobj,
590 struct bin_attribute *bin_attr,
591 char *buf, loff_t off, size_t count)
592 {
593 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
594 struct device, kobj)));
595 struct qla_hw_data *ha = vha->hw;
596 uint8_t *tmp_data;
597
598 if (unlikely(pci_channel_offline(ha->pdev)))
599 return 0;
600
601 if (qla2x00_chip_is_down(vha))
602 return 0;
603
604 if (!capable(CAP_SYS_ADMIN) || off != 0 || count != ha->vpd_size ||
605 !ha->isp_ops->write_nvram)
606 return 0;
607
608 if (qla2x00_wait_for_hba_online(vha) != QLA_SUCCESS) {
609 ql_log(ql_log_warn, vha, 0x706a,
610 "HBA not online, failing VPD update.\n");
611 return -EAGAIN;
612 }
613
614 mutex_lock(&ha->optrom_mutex);
615 if (qla2x00_chip_is_down(vha)) {
616 mutex_unlock(&ha->optrom_mutex);
617 return -EAGAIN;
618 }
619
620 /* Write NVRAM. */
621 ha->isp_ops->write_nvram(vha, buf, ha->vpd_base, count);
622 ha->isp_ops->read_nvram(vha, ha->vpd, ha->vpd_base, count);
623
624 /* Update flash version information for 4Gb & above. */
625 if (!IS_FWI2_CAPABLE(ha)) {
626 mutex_unlock(&ha->optrom_mutex);
627 return -EINVAL;
628 }
629
630 tmp_data = vmalloc(256);
631 if (!tmp_data) {
632 mutex_unlock(&ha->optrom_mutex);
633 ql_log(ql_log_warn, vha, 0x706b,
634 "Unable to allocate memory for VPD information update.\n");
635 return -ENOMEM;
636 }
637 ha->isp_ops->get_flash_version(vha, tmp_data);
638 vfree(tmp_data);
639
640 mutex_unlock(&ha->optrom_mutex);
641
642 return count;
643 }
644
645 static struct bin_attribute sysfs_vpd_attr = {
646 .attr = {
647 .name = "vpd",
648 .mode = S_IRUSR | S_IWUSR,
649 },
650 .size = 0,
651 .read = qla2x00_sysfs_read_vpd,
652 .write = qla2x00_sysfs_write_vpd,
653 };
654
655 static ssize_t
qla2x00_sysfs_read_sfp(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)656 qla2x00_sysfs_read_sfp(struct file *filp, struct kobject *kobj,
657 struct bin_attribute *bin_attr,
658 char *buf, loff_t off, size_t count)
659 {
660 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
661 struct device, kobj)));
662 int rval;
663
664 if (!capable(CAP_SYS_ADMIN) || off != 0 || count < SFP_DEV_SIZE)
665 return 0;
666
667 mutex_lock(&vha->hw->optrom_mutex);
668 if (qla2x00_chip_is_down(vha)) {
669 mutex_unlock(&vha->hw->optrom_mutex);
670 return 0;
671 }
672
673 rval = qla2x00_read_sfp_dev(vha, buf, count);
674 mutex_unlock(&vha->hw->optrom_mutex);
675
676 if (rval)
677 return -EIO;
678
679 return count;
680 }
681
682 static struct bin_attribute sysfs_sfp_attr = {
683 .attr = {
684 .name = "sfp",
685 .mode = S_IRUSR | S_IWUSR,
686 },
687 .size = SFP_DEV_SIZE,
688 .read = qla2x00_sysfs_read_sfp,
689 };
690
691 static ssize_t
qla2x00_sysfs_write_reset(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)692 qla2x00_sysfs_write_reset(struct file *filp, struct kobject *kobj,
693 struct bin_attribute *bin_attr,
694 char *buf, loff_t off, size_t count)
695 {
696 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
697 struct device, kobj)));
698 struct qla_hw_data *ha = vha->hw;
699 struct scsi_qla_host *base_vha = pci_get_drvdata(ha->pdev);
700 int type;
701 uint32_t idc_control;
702 uint8_t *tmp_data = NULL;
703
704 if (off != 0)
705 return -EINVAL;
706
707 type = simple_strtol(buf, NULL, 10);
708 switch (type) {
709 case 0x2025c:
710 ql_log(ql_log_info, vha, 0x706e,
711 "Issuing ISP reset.\n");
712
713 if (vha->hw->flags.port_isolated) {
714 ql_log(ql_log_info, vha, 0x706e,
715 "Port is isolated, returning.\n");
716 return -EINVAL;
717 }
718
719 scsi_block_requests(vha->host);
720 if (IS_QLA82XX(ha)) {
721 ha->flags.isp82xx_no_md_cap = 1;
722 qla82xx_idc_lock(ha);
723 qla82xx_set_reset_owner(vha);
724 qla82xx_idc_unlock(ha);
725 } else if (IS_QLA8044(ha)) {
726 qla8044_idc_lock(ha);
727 idc_control = qla8044_rd_reg(ha,
728 QLA8044_IDC_DRV_CTRL);
729 qla8044_wr_reg(ha, QLA8044_IDC_DRV_CTRL,
730 (idc_control | GRACEFUL_RESET_BIT1));
731 qla82xx_set_reset_owner(vha);
732 qla8044_idc_unlock(ha);
733 } else {
734 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
735 qla2xxx_wake_dpc(vha);
736 }
737 qla2x00_wait_for_chip_reset(vha);
738 scsi_unblock_requests(vha->host);
739 break;
740 case 0x2025d:
741 if (!IS_QLA81XX(ha) && !IS_QLA83XX(ha) &&
742 !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
743 return -EPERM;
744
745 ql_log(ql_log_info, vha, 0x706f,
746 "Issuing MPI reset.\n");
747
748 if (IS_QLA83XX(ha)) {
749 uint32_t idc_control;
750
751 qla83xx_idc_lock(vha, 0);
752 __qla83xx_get_idc_control(vha, &idc_control);
753 idc_control |= QLA83XX_IDC_GRACEFUL_RESET;
754 __qla83xx_set_idc_control(vha, idc_control);
755 qla83xx_wr_reg(vha, QLA83XX_IDC_DEV_STATE,
756 QLA8XXX_DEV_NEED_RESET);
757 qla83xx_idc_audit(vha, IDC_AUDIT_TIMESTAMP);
758 qla83xx_idc_unlock(vha, 0);
759 break;
760 } else {
761 /* Make sure FC side is not in reset */
762 WARN_ON_ONCE(qla2x00_wait_for_hba_online(vha) !=
763 QLA_SUCCESS);
764
765 /* Issue MPI reset */
766 scsi_block_requests(vha->host);
767 if (qla81xx_restart_mpi_firmware(vha) != QLA_SUCCESS)
768 ql_log(ql_log_warn, vha, 0x7070,
769 "MPI reset failed.\n");
770 scsi_unblock_requests(vha->host);
771 break;
772 }
773 break;
774 case 0x2025e:
775 if (!IS_P3P_TYPE(ha) || vha != base_vha) {
776 ql_log(ql_log_info, vha, 0x7071,
777 "FCoE ctx reset not supported.\n");
778 return -EPERM;
779 }
780
781 ql_log(ql_log_info, vha, 0x7072,
782 "Issuing FCoE ctx reset.\n");
783 set_bit(FCOE_CTX_RESET_NEEDED, &vha->dpc_flags);
784 qla2xxx_wake_dpc(vha);
785 qla2x00_wait_for_fcoe_ctx_reset(vha);
786 break;
787 case 0x2025f:
788 if (!IS_QLA8031(ha))
789 return -EPERM;
790 ql_log(ql_log_info, vha, 0x70bc,
791 "Disabling Reset by IDC control\n");
792 qla83xx_idc_lock(vha, 0);
793 __qla83xx_get_idc_control(vha, &idc_control);
794 idc_control |= QLA83XX_IDC_RESET_DISABLED;
795 __qla83xx_set_idc_control(vha, idc_control);
796 qla83xx_idc_unlock(vha, 0);
797 break;
798 case 0x20260:
799 if (!IS_QLA8031(ha))
800 return -EPERM;
801 ql_log(ql_log_info, vha, 0x70bd,
802 "Enabling Reset by IDC control\n");
803 qla83xx_idc_lock(vha, 0);
804 __qla83xx_get_idc_control(vha, &idc_control);
805 idc_control &= ~QLA83XX_IDC_RESET_DISABLED;
806 __qla83xx_set_idc_control(vha, idc_control);
807 qla83xx_idc_unlock(vha, 0);
808 break;
809 case 0x20261:
810 ql_dbg(ql_dbg_user, vha, 0x70e0,
811 "Updating cache versions without reset ");
812
813 tmp_data = vmalloc(256);
814 if (!tmp_data) {
815 ql_log(ql_log_warn, vha, 0x70e1,
816 "Unable to allocate memory for VPD information update.\n");
817 return -ENOMEM;
818 }
819 ha->isp_ops->get_flash_version(vha, tmp_data);
820 vfree(tmp_data);
821 break;
822 }
823 return count;
824 }
825
826 static struct bin_attribute sysfs_reset_attr = {
827 .attr = {
828 .name = "reset",
829 .mode = S_IWUSR,
830 },
831 .size = 0,
832 .write = qla2x00_sysfs_write_reset,
833 };
834
835 static ssize_t
qla2x00_issue_logo(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)836 qla2x00_issue_logo(struct file *filp, struct kobject *kobj,
837 struct bin_attribute *bin_attr,
838 char *buf, loff_t off, size_t count)
839 {
840 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
841 struct device, kobj)));
842 int type;
843 port_id_t did;
844
845 if (!capable(CAP_SYS_ADMIN))
846 return 0;
847
848 if (unlikely(pci_channel_offline(vha->hw->pdev)))
849 return 0;
850
851 if (qla2x00_chip_is_down(vha))
852 return 0;
853
854 type = simple_strtol(buf, NULL, 10);
855
856 did.b.domain = (type & 0x00ff0000) >> 16;
857 did.b.area = (type & 0x0000ff00) >> 8;
858 did.b.al_pa = (type & 0x000000ff);
859
860 ql_log(ql_log_info, vha, 0xd04d, "portid=%02x%02x%02x done\n",
861 did.b.domain, did.b.area, did.b.al_pa);
862
863 ql_log(ql_log_info, vha, 0x70e4, "%s: %d\n", __func__, type);
864
865 qla24xx_els_dcmd_iocb(vha, ELS_DCMD_LOGO, did);
866 return count;
867 }
868
869 static struct bin_attribute sysfs_issue_logo_attr = {
870 .attr = {
871 .name = "issue_logo",
872 .mode = S_IWUSR,
873 },
874 .size = 0,
875 .write = qla2x00_issue_logo,
876 };
877
878 static ssize_t
qla2x00_sysfs_read_xgmac_stats(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)879 qla2x00_sysfs_read_xgmac_stats(struct file *filp, struct kobject *kobj,
880 struct bin_attribute *bin_attr,
881 char *buf, loff_t off, size_t count)
882 {
883 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
884 struct device, kobj)));
885 struct qla_hw_data *ha = vha->hw;
886 int rval;
887 uint16_t actual_size;
888
889 if (!capable(CAP_SYS_ADMIN) || off != 0 || count > XGMAC_DATA_SIZE)
890 return 0;
891
892 if (unlikely(pci_channel_offline(ha->pdev)))
893 return 0;
894 mutex_lock(&vha->hw->optrom_mutex);
895 if (qla2x00_chip_is_down(vha)) {
896 mutex_unlock(&vha->hw->optrom_mutex);
897 return 0;
898 }
899
900 if (ha->xgmac_data)
901 goto do_read;
902
903 ha->xgmac_data = dma_alloc_coherent(&ha->pdev->dev, XGMAC_DATA_SIZE,
904 &ha->xgmac_data_dma, GFP_KERNEL);
905 if (!ha->xgmac_data) {
906 mutex_unlock(&vha->hw->optrom_mutex);
907 ql_log(ql_log_warn, vha, 0x7076,
908 "Unable to allocate memory for XGMAC read-data.\n");
909 return 0;
910 }
911
912 do_read:
913 actual_size = 0;
914 memset(ha->xgmac_data, 0, XGMAC_DATA_SIZE);
915
916 rval = qla2x00_get_xgmac_stats(vha, ha->xgmac_data_dma,
917 XGMAC_DATA_SIZE, &actual_size);
918
919 mutex_unlock(&vha->hw->optrom_mutex);
920 if (rval != QLA_SUCCESS) {
921 ql_log(ql_log_warn, vha, 0x7077,
922 "Unable to read XGMAC data (%x).\n", rval);
923 count = 0;
924 }
925
926 count = actual_size > count ? count : actual_size;
927 memcpy(buf, ha->xgmac_data, count);
928
929 return count;
930 }
931
932 static struct bin_attribute sysfs_xgmac_stats_attr = {
933 .attr = {
934 .name = "xgmac_stats",
935 .mode = S_IRUSR,
936 },
937 .size = 0,
938 .read = qla2x00_sysfs_read_xgmac_stats,
939 };
940
941 static ssize_t
qla2x00_sysfs_read_dcbx_tlv(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t off,size_t count)942 qla2x00_sysfs_read_dcbx_tlv(struct file *filp, struct kobject *kobj,
943 struct bin_attribute *bin_attr,
944 char *buf, loff_t off, size_t count)
945 {
946 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
947 struct device, kobj)));
948 struct qla_hw_data *ha = vha->hw;
949 int rval;
950
951 if (!capable(CAP_SYS_ADMIN) || off != 0 || count > DCBX_TLV_DATA_SIZE)
952 return 0;
953
954 mutex_lock(&vha->hw->optrom_mutex);
955 if (ha->dcbx_tlv)
956 goto do_read;
957 if (qla2x00_chip_is_down(vha)) {
958 mutex_unlock(&vha->hw->optrom_mutex);
959 return 0;
960 }
961
962 ha->dcbx_tlv = dma_alloc_coherent(&ha->pdev->dev, DCBX_TLV_DATA_SIZE,
963 &ha->dcbx_tlv_dma, GFP_KERNEL);
964 if (!ha->dcbx_tlv) {
965 mutex_unlock(&vha->hw->optrom_mutex);
966 ql_log(ql_log_warn, vha, 0x7078,
967 "Unable to allocate memory for DCBX TLV read-data.\n");
968 return -ENOMEM;
969 }
970
971 do_read:
972 memset(ha->dcbx_tlv, 0, DCBX_TLV_DATA_SIZE);
973
974 rval = qla2x00_get_dcbx_params(vha, ha->dcbx_tlv_dma,
975 DCBX_TLV_DATA_SIZE);
976
977 mutex_unlock(&vha->hw->optrom_mutex);
978
979 if (rval != QLA_SUCCESS) {
980 ql_log(ql_log_warn, vha, 0x7079,
981 "Unable to read DCBX TLV (%x).\n", rval);
982 return -EIO;
983 }
984
985 memcpy(buf, ha->dcbx_tlv, count);
986
987 return count;
988 }
989
990 static struct bin_attribute sysfs_dcbx_tlv_attr = {
991 .attr = {
992 .name = "dcbx_tlv",
993 .mode = S_IRUSR,
994 },
995 .size = 0,
996 .read = qla2x00_sysfs_read_dcbx_tlv,
997 };
998
999 static struct sysfs_entry {
1000 char *name;
1001 struct bin_attribute *attr;
1002 int type;
1003 } bin_file_entries[] = {
1004 { "fw_dump", &sysfs_fw_dump_attr, },
1005 { "nvram", &sysfs_nvram_attr, },
1006 { "optrom", &sysfs_optrom_attr, },
1007 { "optrom_ctl", &sysfs_optrom_ctl_attr, },
1008 { "vpd", &sysfs_vpd_attr, 1 },
1009 { "sfp", &sysfs_sfp_attr, 1 },
1010 { "reset", &sysfs_reset_attr, },
1011 { "issue_logo", &sysfs_issue_logo_attr, },
1012 { "xgmac_stats", &sysfs_xgmac_stats_attr, 3 },
1013 { "dcbx_tlv", &sysfs_dcbx_tlv_attr, 3 },
1014 { NULL },
1015 };
1016
1017 void
qla2x00_alloc_sysfs_attr(scsi_qla_host_t * vha)1018 qla2x00_alloc_sysfs_attr(scsi_qla_host_t *vha)
1019 {
1020 struct Scsi_Host *host = vha->host;
1021 struct sysfs_entry *iter;
1022 int ret;
1023
1024 for (iter = bin_file_entries; iter->name; iter++) {
1025 if (iter->type && !IS_FWI2_CAPABLE(vha->hw))
1026 continue;
1027 if (iter->type == 2 && !IS_QLA25XX(vha->hw))
1028 continue;
1029 if (iter->type == 3 && !(IS_CNA_CAPABLE(vha->hw)))
1030 continue;
1031
1032 ret = sysfs_create_bin_file(&host->shost_gendev.kobj,
1033 iter->attr);
1034 if (ret)
1035 ql_log(ql_log_warn, vha, 0x00f3,
1036 "Unable to create sysfs %s binary attribute (%d).\n",
1037 iter->name, ret);
1038 else
1039 ql_dbg(ql_dbg_init, vha, 0x00f4,
1040 "Successfully created sysfs %s binary attribute.\n",
1041 iter->name);
1042 }
1043 }
1044
1045 void
qla2x00_free_sysfs_attr(scsi_qla_host_t * vha,bool stop_beacon)1046 qla2x00_free_sysfs_attr(scsi_qla_host_t *vha, bool stop_beacon)
1047 {
1048 struct Scsi_Host *host = vha->host;
1049 struct sysfs_entry *iter;
1050 struct qla_hw_data *ha = vha->hw;
1051
1052 for (iter = bin_file_entries; iter->name; iter++) {
1053 if (iter->type && !IS_FWI2_CAPABLE(ha))
1054 continue;
1055 if (iter->type == 2 && !IS_QLA25XX(ha))
1056 continue;
1057 if (iter->type == 3 && !(IS_CNA_CAPABLE(ha)))
1058 continue;
1059
1060 sysfs_remove_bin_file(&host->shost_gendev.kobj,
1061 iter->attr);
1062 }
1063
1064 if (stop_beacon && ha->beacon_blink_led == 1)
1065 ha->isp_ops->beacon_off(vha);
1066 }
1067
1068 /* Scsi_Host attributes. */
1069
1070 static ssize_t
qla2x00_fw_version_show(struct device * dev,struct device_attribute * attr,char * buf)1071 qla2x00_fw_version_show(struct device *dev,
1072 struct device_attribute *attr, char *buf)
1073 {
1074 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1075 struct qla_hw_data *ha = vha->hw;
1076 char fw_str[128];
1077
1078 return scnprintf(buf, PAGE_SIZE, "%s\n",
1079 ha->isp_ops->fw_version_str(vha, fw_str, sizeof(fw_str)));
1080 }
1081
1082 static ssize_t
qla2x00_serial_num_show(struct device * dev,struct device_attribute * attr,char * buf)1083 qla2x00_serial_num_show(struct device *dev, struct device_attribute *attr,
1084 char *buf)
1085 {
1086 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1087 struct qla_hw_data *ha = vha->hw;
1088 uint32_t sn;
1089
1090 if (IS_QLAFX00(vha->hw)) {
1091 return scnprintf(buf, PAGE_SIZE, "%s\n",
1092 vha->hw->mr.serial_num);
1093 } else if (IS_FWI2_CAPABLE(ha)) {
1094 qla2xxx_get_vpd_field(vha, "SN", buf, PAGE_SIZE - 1);
1095 return strlen(strcat(buf, "\n"));
1096 }
1097
1098 sn = ((ha->serial0 & 0x1f) << 16) | (ha->serial2 << 8) | ha->serial1;
1099 return scnprintf(buf, PAGE_SIZE, "%c%05d\n", 'A' + sn / 100000,
1100 sn % 100000);
1101 }
1102
1103 static ssize_t
qla2x00_isp_name_show(struct device * dev,struct device_attribute * attr,char * buf)1104 qla2x00_isp_name_show(struct device *dev, struct device_attribute *attr,
1105 char *buf)
1106 {
1107 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1108
1109 return scnprintf(buf, PAGE_SIZE, "ISP%04X\n", vha->hw->pdev->device);
1110 }
1111
1112 static ssize_t
qla2x00_isp_id_show(struct device * dev,struct device_attribute * attr,char * buf)1113 qla2x00_isp_id_show(struct device *dev, struct device_attribute *attr,
1114 char *buf)
1115 {
1116 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1117 struct qla_hw_data *ha = vha->hw;
1118
1119 if (IS_QLAFX00(vha->hw))
1120 return scnprintf(buf, PAGE_SIZE, "%s\n",
1121 vha->hw->mr.hw_version);
1122
1123 return scnprintf(buf, PAGE_SIZE, "%04x %04x %04x %04x\n",
1124 ha->product_id[0], ha->product_id[1], ha->product_id[2],
1125 ha->product_id[3]);
1126 }
1127
1128 static ssize_t
qla2x00_model_name_show(struct device * dev,struct device_attribute * attr,char * buf)1129 qla2x00_model_name_show(struct device *dev, struct device_attribute *attr,
1130 char *buf)
1131 {
1132 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1133
1134 return scnprintf(buf, PAGE_SIZE, "%s\n", vha->hw->model_number);
1135 }
1136
1137 static ssize_t
qla2x00_model_desc_show(struct device * dev,struct device_attribute * attr,char * buf)1138 qla2x00_model_desc_show(struct device *dev, struct device_attribute *attr,
1139 char *buf)
1140 {
1141 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1142
1143 return scnprintf(buf, PAGE_SIZE, "%s\n", vha->hw->model_desc);
1144 }
1145
1146 static ssize_t
qla2x00_pci_info_show(struct device * dev,struct device_attribute * attr,char * buf)1147 qla2x00_pci_info_show(struct device *dev, struct device_attribute *attr,
1148 char *buf)
1149 {
1150 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1151 char pci_info[30];
1152
1153 return scnprintf(buf, PAGE_SIZE, "%s\n",
1154 vha->hw->isp_ops->pci_info_str(vha, pci_info,
1155 sizeof(pci_info)));
1156 }
1157
1158 static ssize_t
qla2x00_link_state_show(struct device * dev,struct device_attribute * attr,char * buf)1159 qla2x00_link_state_show(struct device *dev, struct device_attribute *attr,
1160 char *buf)
1161 {
1162 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1163 struct qla_hw_data *ha = vha->hw;
1164 int len = 0;
1165
1166 if (atomic_read(&vha->loop_state) == LOOP_DOWN ||
1167 atomic_read(&vha->loop_state) == LOOP_DEAD ||
1168 vha->device_flags & DFLG_NO_CABLE)
1169 len = scnprintf(buf, PAGE_SIZE, "Link Down\n");
1170 else if (atomic_read(&vha->loop_state) != LOOP_READY ||
1171 qla2x00_chip_is_down(vha))
1172 len = scnprintf(buf, PAGE_SIZE, "Unknown Link State\n");
1173 else {
1174 len = scnprintf(buf, PAGE_SIZE, "Link Up - ");
1175
1176 switch (ha->current_topology) {
1177 case ISP_CFG_NL:
1178 len += scnprintf(buf + len, PAGE_SIZE-len, "Loop\n");
1179 break;
1180 case ISP_CFG_FL:
1181 len += scnprintf(buf + len, PAGE_SIZE-len, "FL_Port\n");
1182 break;
1183 case ISP_CFG_N:
1184 len += scnprintf(buf + len, PAGE_SIZE-len,
1185 "N_Port to N_Port\n");
1186 break;
1187 case ISP_CFG_F:
1188 len += scnprintf(buf + len, PAGE_SIZE-len, "F_Port\n");
1189 break;
1190 default:
1191 len += scnprintf(buf + len, PAGE_SIZE-len, "Loop\n");
1192 break;
1193 }
1194 }
1195 return len;
1196 }
1197
1198 static ssize_t
qla2x00_zio_show(struct device * dev,struct device_attribute * attr,char * buf)1199 qla2x00_zio_show(struct device *dev, struct device_attribute *attr,
1200 char *buf)
1201 {
1202 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1203 int len = 0;
1204
1205 switch (vha->hw->zio_mode) {
1206 case QLA_ZIO_MODE_6:
1207 len += scnprintf(buf + len, PAGE_SIZE-len, "Mode 6\n");
1208 break;
1209 case QLA_ZIO_DISABLED:
1210 len += scnprintf(buf + len, PAGE_SIZE-len, "Disabled\n");
1211 break;
1212 }
1213 return len;
1214 }
1215
1216 static ssize_t
qla2x00_zio_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1217 qla2x00_zio_store(struct device *dev, struct device_attribute *attr,
1218 const char *buf, size_t count)
1219 {
1220 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1221 struct qla_hw_data *ha = vha->hw;
1222 int val = 0;
1223 uint16_t zio_mode;
1224
1225 if (!IS_ZIO_SUPPORTED(ha))
1226 return -ENOTSUPP;
1227
1228 if (sscanf(buf, "%d", &val) != 1)
1229 return -EINVAL;
1230
1231 if (val)
1232 zio_mode = QLA_ZIO_MODE_6;
1233 else
1234 zio_mode = QLA_ZIO_DISABLED;
1235
1236 /* Update per-hba values and queue a reset. */
1237 if (zio_mode != QLA_ZIO_DISABLED || ha->zio_mode != QLA_ZIO_DISABLED) {
1238 ha->zio_mode = zio_mode;
1239 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
1240 }
1241 return strlen(buf);
1242 }
1243
1244 static ssize_t
qla2x00_zio_timer_show(struct device * dev,struct device_attribute * attr,char * buf)1245 qla2x00_zio_timer_show(struct device *dev, struct device_attribute *attr,
1246 char *buf)
1247 {
1248 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1249
1250 return scnprintf(buf, PAGE_SIZE, "%d us\n", vha->hw->zio_timer * 100);
1251 }
1252
1253 static ssize_t
qla2x00_zio_timer_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1254 qla2x00_zio_timer_store(struct device *dev, struct device_attribute *attr,
1255 const char *buf, size_t count)
1256 {
1257 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1258 int val = 0;
1259 uint16_t zio_timer;
1260
1261 if (sscanf(buf, "%d", &val) != 1)
1262 return -EINVAL;
1263 if (val > 25500 || val < 100)
1264 return -ERANGE;
1265
1266 zio_timer = (uint16_t)(val / 100);
1267 vha->hw->zio_timer = zio_timer;
1268
1269 return strlen(buf);
1270 }
1271
1272 static ssize_t
qla_zio_threshold_show(struct device * dev,struct device_attribute * attr,char * buf)1273 qla_zio_threshold_show(struct device *dev, struct device_attribute *attr,
1274 char *buf)
1275 {
1276 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1277
1278 return scnprintf(buf, PAGE_SIZE, "%d exchanges\n",
1279 vha->hw->last_zio_threshold);
1280 }
1281
1282 static ssize_t
qla_zio_threshold_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1283 qla_zio_threshold_store(struct device *dev, struct device_attribute *attr,
1284 const char *buf, size_t count)
1285 {
1286 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1287 int val = 0;
1288
1289 if (vha->hw->zio_mode != QLA_ZIO_MODE_6)
1290 return -EINVAL;
1291 if (sscanf(buf, "%d", &val) != 1)
1292 return -EINVAL;
1293 if (val < 0 || val > 256)
1294 return -ERANGE;
1295
1296 atomic_set(&vha->hw->zio_threshold, val);
1297 return strlen(buf);
1298 }
1299
1300 static ssize_t
qla2x00_beacon_show(struct device * dev,struct device_attribute * attr,char * buf)1301 qla2x00_beacon_show(struct device *dev, struct device_attribute *attr,
1302 char *buf)
1303 {
1304 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1305 int len = 0;
1306
1307 if (vha->hw->beacon_blink_led)
1308 len += scnprintf(buf + len, PAGE_SIZE-len, "Enabled\n");
1309 else
1310 len += scnprintf(buf + len, PAGE_SIZE-len, "Disabled\n");
1311 return len;
1312 }
1313
1314 static ssize_t
qla2x00_beacon_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1315 qla2x00_beacon_store(struct device *dev, struct device_attribute *attr,
1316 const char *buf, size_t count)
1317 {
1318 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1319 struct qla_hw_data *ha = vha->hw;
1320 int val = 0;
1321 int rval;
1322
1323 if (IS_QLA2100(ha) || IS_QLA2200(ha))
1324 return -EPERM;
1325
1326 if (sscanf(buf, "%d", &val) != 1)
1327 return -EINVAL;
1328
1329 mutex_lock(&vha->hw->optrom_mutex);
1330 if (qla2x00_chip_is_down(vha)) {
1331 mutex_unlock(&vha->hw->optrom_mutex);
1332 ql_log(ql_log_warn, vha, 0x707a,
1333 "Abort ISP active -- ignoring beacon request.\n");
1334 return -EBUSY;
1335 }
1336
1337 if (val)
1338 rval = ha->isp_ops->beacon_on(vha);
1339 else
1340 rval = ha->isp_ops->beacon_off(vha);
1341
1342 if (rval != QLA_SUCCESS)
1343 count = 0;
1344
1345 mutex_unlock(&vha->hw->optrom_mutex);
1346
1347 return count;
1348 }
1349
1350 static ssize_t
qla2x00_beacon_config_show(struct device * dev,struct device_attribute * attr,char * buf)1351 qla2x00_beacon_config_show(struct device *dev, struct device_attribute *attr,
1352 char *buf)
1353 {
1354 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1355 struct qla_hw_data *ha = vha->hw;
1356 uint16_t led[3] = { 0 };
1357
1358 if (!IS_QLA2031(ha) && !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1359 return -EPERM;
1360
1361 if (ql26xx_led_config(vha, 0, led))
1362 return scnprintf(buf, PAGE_SIZE, "\n");
1363
1364 return scnprintf(buf, PAGE_SIZE, "%#04hx %#04hx %#04hx\n",
1365 led[0], led[1], led[2]);
1366 }
1367
1368 static ssize_t
qla2x00_beacon_config_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1369 qla2x00_beacon_config_store(struct device *dev, struct device_attribute *attr,
1370 const char *buf, size_t count)
1371 {
1372 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1373 struct qla_hw_data *ha = vha->hw;
1374 uint16_t options = BIT_0;
1375 uint16_t led[3] = { 0 };
1376 uint16_t word[4];
1377 int n;
1378
1379 if (!IS_QLA2031(ha) && !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1380 return -EPERM;
1381
1382 n = sscanf(buf, "%hx %hx %hx %hx", word+0, word+1, word+2, word+3);
1383 if (n == 4) {
1384 if (word[0] == 3) {
1385 options |= BIT_3|BIT_2|BIT_1;
1386 led[0] = word[1];
1387 led[1] = word[2];
1388 led[2] = word[3];
1389 goto write;
1390 }
1391 return -EINVAL;
1392 }
1393
1394 if (n == 2) {
1395 /* check led index */
1396 if (word[0] == 0) {
1397 options |= BIT_2;
1398 led[0] = word[1];
1399 goto write;
1400 }
1401 if (word[0] == 1) {
1402 options |= BIT_3;
1403 led[1] = word[1];
1404 goto write;
1405 }
1406 if (word[0] == 2) {
1407 options |= BIT_1;
1408 led[2] = word[1];
1409 goto write;
1410 }
1411 return -EINVAL;
1412 }
1413
1414 return -EINVAL;
1415
1416 write:
1417 if (ql26xx_led_config(vha, options, led))
1418 return -EFAULT;
1419
1420 return count;
1421 }
1422
1423 static ssize_t
qla2x00_optrom_bios_version_show(struct device * dev,struct device_attribute * attr,char * buf)1424 qla2x00_optrom_bios_version_show(struct device *dev,
1425 struct device_attribute *attr, char *buf)
1426 {
1427 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1428 struct qla_hw_data *ha = vha->hw;
1429
1430 return scnprintf(buf, PAGE_SIZE, "%d.%02d\n", ha->bios_revision[1],
1431 ha->bios_revision[0]);
1432 }
1433
1434 static ssize_t
qla2x00_optrom_efi_version_show(struct device * dev,struct device_attribute * attr,char * buf)1435 qla2x00_optrom_efi_version_show(struct device *dev,
1436 struct device_attribute *attr, char *buf)
1437 {
1438 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1439 struct qla_hw_data *ha = vha->hw;
1440
1441 return scnprintf(buf, PAGE_SIZE, "%d.%02d\n", ha->efi_revision[1],
1442 ha->efi_revision[0]);
1443 }
1444
1445 static ssize_t
qla2x00_optrom_fcode_version_show(struct device * dev,struct device_attribute * attr,char * buf)1446 qla2x00_optrom_fcode_version_show(struct device *dev,
1447 struct device_attribute *attr, char *buf)
1448 {
1449 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1450 struct qla_hw_data *ha = vha->hw;
1451
1452 return scnprintf(buf, PAGE_SIZE, "%d.%02d\n", ha->fcode_revision[1],
1453 ha->fcode_revision[0]);
1454 }
1455
1456 static ssize_t
qla2x00_optrom_fw_version_show(struct device * dev,struct device_attribute * attr,char * buf)1457 qla2x00_optrom_fw_version_show(struct device *dev,
1458 struct device_attribute *attr, char *buf)
1459 {
1460 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1461 struct qla_hw_data *ha = vha->hw;
1462
1463 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d %d\n",
1464 ha->fw_revision[0], ha->fw_revision[1], ha->fw_revision[2],
1465 ha->fw_revision[3]);
1466 }
1467
1468 static ssize_t
qla2x00_optrom_gold_fw_version_show(struct device * dev,struct device_attribute * attr,char * buf)1469 qla2x00_optrom_gold_fw_version_show(struct device *dev,
1470 struct device_attribute *attr, char *buf)
1471 {
1472 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1473 struct qla_hw_data *ha = vha->hw;
1474
1475 if (!IS_QLA81XX(ha) && !IS_QLA83XX(ha) &&
1476 !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1477 return scnprintf(buf, PAGE_SIZE, "\n");
1478
1479 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d (%d)\n",
1480 ha->gold_fw_version[0], ha->gold_fw_version[1],
1481 ha->gold_fw_version[2], ha->gold_fw_version[3]);
1482 }
1483
1484 static ssize_t
qla2x00_total_isp_aborts_show(struct device * dev,struct device_attribute * attr,char * buf)1485 qla2x00_total_isp_aborts_show(struct device *dev,
1486 struct device_attribute *attr, char *buf)
1487 {
1488 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1489
1490 return scnprintf(buf, PAGE_SIZE, "%d\n",
1491 vha->qla_stats.total_isp_aborts);
1492 }
1493
1494 static ssize_t
qla24xx_84xx_fw_version_show(struct device * dev,struct device_attribute * attr,char * buf)1495 qla24xx_84xx_fw_version_show(struct device *dev,
1496 struct device_attribute *attr, char *buf)
1497 {
1498 int rval = QLA_SUCCESS;
1499 uint16_t status[2] = { 0 };
1500 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1501 struct qla_hw_data *ha = vha->hw;
1502
1503 if (!IS_QLA84XX(ha))
1504 return scnprintf(buf, PAGE_SIZE, "\n");
1505
1506 if (!ha->cs84xx->op_fw_version) {
1507 rval = qla84xx_verify_chip(vha, status);
1508
1509 if (!rval && !status[0])
1510 return scnprintf(buf, PAGE_SIZE, "%u\n",
1511 (uint32_t)ha->cs84xx->op_fw_version);
1512 }
1513
1514 return scnprintf(buf, PAGE_SIZE, "\n");
1515 }
1516
1517 static ssize_t
qla2x00_serdes_version_show(struct device * dev,struct device_attribute * attr,char * buf)1518 qla2x00_serdes_version_show(struct device *dev, struct device_attribute *attr,
1519 char *buf)
1520 {
1521 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1522 struct qla_hw_data *ha = vha->hw;
1523
1524 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1525 return scnprintf(buf, PAGE_SIZE, "\n");
1526
1527 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d\n",
1528 ha->serdes_version[0], ha->serdes_version[1],
1529 ha->serdes_version[2]);
1530 }
1531
1532 static ssize_t
qla2x00_mpi_version_show(struct device * dev,struct device_attribute * attr,char * buf)1533 qla2x00_mpi_version_show(struct device *dev, struct device_attribute *attr,
1534 char *buf)
1535 {
1536 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1537 struct qla_hw_data *ha = vha->hw;
1538
1539 if (!IS_QLA81XX(ha) && !IS_QLA8031(ha) && !IS_QLA8044(ha) &&
1540 !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1541 return scnprintf(buf, PAGE_SIZE, "\n");
1542
1543 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d (%x)\n",
1544 ha->mpi_version[0], ha->mpi_version[1], ha->mpi_version[2],
1545 ha->mpi_capabilities);
1546 }
1547
1548 static ssize_t
qla2x00_phy_version_show(struct device * dev,struct device_attribute * attr,char * buf)1549 qla2x00_phy_version_show(struct device *dev, struct device_attribute *attr,
1550 char *buf)
1551 {
1552 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1553 struct qla_hw_data *ha = vha->hw;
1554
1555 if (!IS_QLA81XX(ha) && !IS_QLA8031(ha))
1556 return scnprintf(buf, PAGE_SIZE, "\n");
1557
1558 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d\n",
1559 ha->phy_version[0], ha->phy_version[1], ha->phy_version[2]);
1560 }
1561
1562 static ssize_t
qla2x00_flash_block_size_show(struct device * dev,struct device_attribute * attr,char * buf)1563 qla2x00_flash_block_size_show(struct device *dev,
1564 struct device_attribute *attr, char *buf)
1565 {
1566 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1567 struct qla_hw_data *ha = vha->hw;
1568
1569 return scnprintf(buf, PAGE_SIZE, "0x%x\n", ha->fdt_block_size);
1570 }
1571
1572 static ssize_t
qla2x00_vlan_id_show(struct device * dev,struct device_attribute * attr,char * buf)1573 qla2x00_vlan_id_show(struct device *dev, struct device_attribute *attr,
1574 char *buf)
1575 {
1576 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1577
1578 if (!IS_CNA_CAPABLE(vha->hw))
1579 return scnprintf(buf, PAGE_SIZE, "\n");
1580
1581 return scnprintf(buf, PAGE_SIZE, "%d\n", vha->fcoe_vlan_id);
1582 }
1583
1584 static ssize_t
qla2x00_vn_port_mac_address_show(struct device * dev,struct device_attribute * attr,char * buf)1585 qla2x00_vn_port_mac_address_show(struct device *dev,
1586 struct device_attribute *attr, char *buf)
1587 {
1588 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1589
1590 if (!IS_CNA_CAPABLE(vha->hw))
1591 return scnprintf(buf, PAGE_SIZE, "\n");
1592
1593 return scnprintf(buf, PAGE_SIZE, "%pMR\n", vha->fcoe_vn_port_mac);
1594 }
1595
1596 static ssize_t
qla2x00_fabric_param_show(struct device * dev,struct device_attribute * attr,char * buf)1597 qla2x00_fabric_param_show(struct device *dev, struct device_attribute *attr,
1598 char *buf)
1599 {
1600 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1601
1602 return scnprintf(buf, PAGE_SIZE, "%d\n", vha->hw->switch_cap);
1603 }
1604
1605 static ssize_t
qla2x00_thermal_temp_show(struct device * dev,struct device_attribute * attr,char * buf)1606 qla2x00_thermal_temp_show(struct device *dev,
1607 struct device_attribute *attr, char *buf)
1608 {
1609 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1610 uint16_t temp = 0;
1611 int rc;
1612
1613 mutex_lock(&vha->hw->optrom_mutex);
1614 if (qla2x00_chip_is_down(vha)) {
1615 mutex_unlock(&vha->hw->optrom_mutex);
1616 ql_log(ql_log_warn, vha, 0x70dc, "ISP reset active.\n");
1617 goto done;
1618 }
1619
1620 if (vha->hw->flags.eeh_busy) {
1621 mutex_unlock(&vha->hw->optrom_mutex);
1622 ql_log(ql_log_warn, vha, 0x70dd, "PCI EEH busy.\n");
1623 goto done;
1624 }
1625
1626 rc = qla2x00_get_thermal_temp(vha, &temp);
1627 mutex_unlock(&vha->hw->optrom_mutex);
1628 if (rc == QLA_SUCCESS)
1629 return scnprintf(buf, PAGE_SIZE, "%d\n", temp);
1630
1631 done:
1632 return scnprintf(buf, PAGE_SIZE, "\n");
1633 }
1634
1635 static ssize_t
qla2x00_fw_state_show(struct device * dev,struct device_attribute * attr,char * buf)1636 qla2x00_fw_state_show(struct device *dev, struct device_attribute *attr,
1637 char *buf)
1638 {
1639 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1640 int rval = QLA_FUNCTION_FAILED;
1641 uint16_t state[6];
1642 uint32_t pstate;
1643
1644 if (IS_QLAFX00(vha->hw)) {
1645 pstate = qlafx00_fw_state_show(dev, attr, buf);
1646 return scnprintf(buf, PAGE_SIZE, "0x%x\n", pstate);
1647 }
1648
1649 mutex_lock(&vha->hw->optrom_mutex);
1650 if (qla2x00_chip_is_down(vha)) {
1651 mutex_unlock(&vha->hw->optrom_mutex);
1652 ql_log(ql_log_warn, vha, 0x707c,
1653 "ISP reset active.\n");
1654 goto out;
1655 } else if (vha->hw->flags.eeh_busy) {
1656 mutex_unlock(&vha->hw->optrom_mutex);
1657 goto out;
1658 }
1659
1660 rval = qla2x00_get_firmware_state(vha, state);
1661 mutex_unlock(&vha->hw->optrom_mutex);
1662 out:
1663 if (rval != QLA_SUCCESS) {
1664 memset(state, -1, sizeof(state));
1665 rval = qla2x00_get_firmware_state(vha, state);
1666 }
1667
1668 return scnprintf(buf, PAGE_SIZE, "0x%x 0x%x 0x%x 0x%x 0x%x 0x%x\n",
1669 state[0], state[1], state[2], state[3], state[4], state[5]);
1670 }
1671
1672 static ssize_t
qla2x00_diag_requests_show(struct device * dev,struct device_attribute * attr,char * buf)1673 qla2x00_diag_requests_show(struct device *dev,
1674 struct device_attribute *attr, char *buf)
1675 {
1676 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1677
1678 if (!IS_BIDI_CAPABLE(vha->hw))
1679 return scnprintf(buf, PAGE_SIZE, "\n");
1680
1681 return scnprintf(buf, PAGE_SIZE, "%llu\n", vha->bidi_stats.io_count);
1682 }
1683
1684 static ssize_t
qla2x00_diag_megabytes_show(struct device * dev,struct device_attribute * attr,char * buf)1685 qla2x00_diag_megabytes_show(struct device *dev,
1686 struct device_attribute *attr, char *buf)
1687 {
1688 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1689
1690 if (!IS_BIDI_CAPABLE(vha->hw))
1691 return scnprintf(buf, PAGE_SIZE, "\n");
1692
1693 return scnprintf(buf, PAGE_SIZE, "%llu\n",
1694 vha->bidi_stats.transfer_bytes >> 20);
1695 }
1696
1697 static ssize_t
qla2x00_fw_dump_size_show(struct device * dev,struct device_attribute * attr,char * buf)1698 qla2x00_fw_dump_size_show(struct device *dev, struct device_attribute *attr,
1699 char *buf)
1700 {
1701 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1702 struct qla_hw_data *ha = vha->hw;
1703 uint32_t size;
1704
1705 if (!ha->fw_dumped)
1706 size = 0;
1707 else if (IS_P3P_TYPE(ha))
1708 size = ha->md_template_size + ha->md_dump_size;
1709 else
1710 size = ha->fw_dump_len;
1711
1712 return scnprintf(buf, PAGE_SIZE, "%d\n", size);
1713 }
1714
1715 static ssize_t
qla2x00_allow_cna_fw_dump_show(struct device * dev,struct device_attribute * attr,char * buf)1716 qla2x00_allow_cna_fw_dump_show(struct device *dev,
1717 struct device_attribute *attr, char *buf)
1718 {
1719 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1720
1721 if (!IS_P3P_TYPE(vha->hw))
1722 return scnprintf(buf, PAGE_SIZE, "\n");
1723 else
1724 return scnprintf(buf, PAGE_SIZE, "%s\n",
1725 vha->hw->allow_cna_fw_dump ? "true" : "false");
1726 }
1727
1728 static ssize_t
qla2x00_allow_cna_fw_dump_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1729 qla2x00_allow_cna_fw_dump_store(struct device *dev,
1730 struct device_attribute *attr, const char *buf, size_t count)
1731 {
1732 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1733 int val = 0;
1734
1735 if (!IS_P3P_TYPE(vha->hw))
1736 return -EINVAL;
1737
1738 if (sscanf(buf, "%d", &val) != 1)
1739 return -EINVAL;
1740
1741 vha->hw->allow_cna_fw_dump = val != 0;
1742
1743 return strlen(buf);
1744 }
1745
1746 static ssize_t
qla2x00_pep_version_show(struct device * dev,struct device_attribute * attr,char * buf)1747 qla2x00_pep_version_show(struct device *dev, struct device_attribute *attr,
1748 char *buf)
1749 {
1750 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1751 struct qla_hw_data *ha = vha->hw;
1752
1753 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1754 return scnprintf(buf, PAGE_SIZE, "\n");
1755
1756 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d\n",
1757 ha->pep_version[0], ha->pep_version[1], ha->pep_version[2]);
1758 }
1759
1760 static ssize_t
qla2x00_min_supported_speed_show(struct device * dev,struct device_attribute * attr,char * buf)1761 qla2x00_min_supported_speed_show(struct device *dev,
1762 struct device_attribute *attr, char *buf)
1763 {
1764 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1765 struct qla_hw_data *ha = vha->hw;
1766
1767 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1768 return scnprintf(buf, PAGE_SIZE, "\n");
1769
1770 return scnprintf(buf, PAGE_SIZE, "%s\n",
1771 ha->min_supported_speed == 6 ? "64Gps" :
1772 ha->min_supported_speed == 5 ? "32Gps" :
1773 ha->min_supported_speed == 4 ? "16Gps" :
1774 ha->min_supported_speed == 3 ? "8Gps" :
1775 ha->min_supported_speed == 2 ? "4Gps" :
1776 ha->min_supported_speed != 0 ? "unknown" : "");
1777 }
1778
1779 static ssize_t
qla2x00_max_supported_speed_show(struct device * dev,struct device_attribute * attr,char * buf)1780 qla2x00_max_supported_speed_show(struct device *dev,
1781 struct device_attribute *attr, char *buf)
1782 {
1783 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1784 struct qla_hw_data *ha = vha->hw;
1785
1786 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1787 return scnprintf(buf, PAGE_SIZE, "\n");
1788
1789 return scnprintf(buf, PAGE_SIZE, "%s\n",
1790 ha->max_supported_speed == 2 ? "64Gps" :
1791 ha->max_supported_speed == 1 ? "32Gps" :
1792 ha->max_supported_speed == 0 ? "16Gps" : "unknown");
1793 }
1794
1795 static ssize_t
qla2x00_port_speed_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1796 qla2x00_port_speed_store(struct device *dev, struct device_attribute *attr,
1797 const char *buf, size_t count)
1798 {
1799 struct scsi_qla_host *vha = shost_priv(dev_to_shost(dev));
1800 ulong type, speed;
1801 int oldspeed, rval;
1802 int mode = QLA_SET_DATA_RATE_LR;
1803 struct qla_hw_data *ha = vha->hw;
1804
1805 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha)) {
1806 ql_log(ql_log_warn, vha, 0x70d8,
1807 "Speed setting not supported \n");
1808 return -EINVAL;
1809 }
1810
1811 rval = kstrtol(buf, 10, &type);
1812 if (rval)
1813 return rval;
1814 speed = type;
1815 if (type == 40 || type == 80 || type == 160 ||
1816 type == 320) {
1817 ql_dbg(ql_dbg_user, vha, 0x70d9,
1818 "Setting will be affected after a loss of sync\n");
1819 type = type/10;
1820 mode = QLA_SET_DATA_RATE_NOLR;
1821 }
1822
1823 oldspeed = ha->set_data_rate;
1824
1825 switch (type) {
1826 case 0:
1827 ha->set_data_rate = PORT_SPEED_AUTO;
1828 break;
1829 case 4:
1830 ha->set_data_rate = PORT_SPEED_4GB;
1831 break;
1832 case 8:
1833 ha->set_data_rate = PORT_SPEED_8GB;
1834 break;
1835 case 16:
1836 ha->set_data_rate = PORT_SPEED_16GB;
1837 break;
1838 case 32:
1839 ha->set_data_rate = PORT_SPEED_32GB;
1840 break;
1841 default:
1842 ql_log(ql_log_warn, vha, 0x1199,
1843 "Unrecognized speed setting:%lx. Setting Autoneg\n",
1844 speed);
1845 ha->set_data_rate = PORT_SPEED_AUTO;
1846 }
1847
1848 if (qla2x00_chip_is_down(vha) || (oldspeed == ha->set_data_rate))
1849 return -EINVAL;
1850
1851 ql_log(ql_log_info, vha, 0x70da,
1852 "Setting speed to %lx Gbps \n", type);
1853
1854 rval = qla2x00_set_data_rate(vha, mode);
1855 if (rval != QLA_SUCCESS)
1856 return -EIO;
1857
1858 return strlen(buf);
1859 }
1860
1861 static const struct {
1862 u16 rate;
1863 char *str;
1864 } port_speed_str[] = {
1865 { PORT_SPEED_4GB, "4" },
1866 { PORT_SPEED_8GB, "8" },
1867 { PORT_SPEED_16GB, "16" },
1868 { PORT_SPEED_32GB, "32" },
1869 { PORT_SPEED_64GB, "64" },
1870 { PORT_SPEED_10GB, "10" },
1871 };
1872
1873 static ssize_t
qla2x00_port_speed_show(struct device * dev,struct device_attribute * attr,char * buf)1874 qla2x00_port_speed_show(struct device *dev, struct device_attribute *attr,
1875 char *buf)
1876 {
1877 struct scsi_qla_host *vha = shost_priv(dev_to_shost(dev));
1878 struct qla_hw_data *ha = vha->hw;
1879 ssize_t rval;
1880 u16 i;
1881 char *speed = "Unknown";
1882
1883 rval = qla2x00_get_data_rate(vha);
1884 if (rval != QLA_SUCCESS) {
1885 ql_log(ql_log_warn, vha, 0x70db,
1886 "Unable to get port speed rval:%zd\n", rval);
1887 return -EINVAL;
1888 }
1889
1890 for (i = 0; i < ARRAY_SIZE(port_speed_str); i++) {
1891 if (port_speed_str[i].rate != ha->link_data_rate)
1892 continue;
1893 speed = port_speed_str[i].str;
1894 break;
1895 }
1896
1897 return scnprintf(buf, PAGE_SIZE, "%s\n", speed);
1898 }
1899
1900 static ssize_t
qla2x00_mpi_pause_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1901 qla2x00_mpi_pause_store(struct device *dev,
1902 struct device_attribute *attr, const char *buf, size_t count)
1903 {
1904 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1905 int rval = 0;
1906
1907 if (sscanf(buf, "%d", &rval) != 1)
1908 return -EINVAL;
1909
1910 ql_log(ql_log_warn, vha, 0x7089, "Pausing MPI...\n");
1911
1912 rval = qla83xx_wr_reg(vha, 0x002012d4, 0x30000001);
1913
1914 if (rval != QLA_SUCCESS) {
1915 ql_log(ql_log_warn, vha, 0x708a, "Unable to pause MPI.\n");
1916 count = 0;
1917 }
1918
1919 return count;
1920 }
1921
1922 static DEVICE_ATTR(mpi_pause, S_IWUSR, NULL, qla2x00_mpi_pause_store);
1923
1924 /* ----- */
1925
1926 static ssize_t
qlini_mode_show(struct device * dev,struct device_attribute * attr,char * buf)1927 qlini_mode_show(struct device *dev, struct device_attribute *attr, char *buf)
1928 {
1929 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1930 int len = 0;
1931
1932 len += scnprintf(buf + len, PAGE_SIZE-len,
1933 "Supported options: enabled | disabled | dual | exclusive\n");
1934
1935 /* --- */
1936 len += scnprintf(buf + len, PAGE_SIZE-len, "Current selection: ");
1937
1938 switch (vha->qlini_mode) {
1939 case QLA2XXX_INI_MODE_EXCLUSIVE:
1940 len += scnprintf(buf + len, PAGE_SIZE-len,
1941 QLA2XXX_INI_MODE_STR_EXCLUSIVE);
1942 break;
1943 case QLA2XXX_INI_MODE_DISABLED:
1944 len += scnprintf(buf + len, PAGE_SIZE-len,
1945 QLA2XXX_INI_MODE_STR_DISABLED);
1946 break;
1947 case QLA2XXX_INI_MODE_ENABLED:
1948 len += scnprintf(buf + len, PAGE_SIZE-len,
1949 QLA2XXX_INI_MODE_STR_ENABLED);
1950 break;
1951 case QLA2XXX_INI_MODE_DUAL:
1952 len += scnprintf(buf + len, PAGE_SIZE-len,
1953 QLA2XXX_INI_MODE_STR_DUAL);
1954 break;
1955 }
1956 len += scnprintf(buf + len, PAGE_SIZE-len, "\n");
1957
1958 return len;
1959 }
1960
1961 static char *mode_to_str[] = {
1962 "exclusive",
1963 "disabled",
1964 "enabled",
1965 "dual",
1966 };
1967
1968 #define NEED_EXCH_OFFLOAD(_exchg) ((_exchg) > FW_DEF_EXCHANGES_CNT)
qla_set_ini_mode(scsi_qla_host_t * vha,int op)1969 static void qla_set_ini_mode(scsi_qla_host_t *vha, int op)
1970 {
1971 enum {
1972 NO_ACTION,
1973 MODE_CHANGE_ACCEPT,
1974 MODE_CHANGE_NO_ACTION,
1975 TARGET_STILL_ACTIVE,
1976 };
1977 int action = NO_ACTION;
1978 int set_mode = 0;
1979 u8 eo_toggle = 0; /* exchange offload flipped */
1980
1981 switch (vha->qlini_mode) {
1982 case QLA2XXX_INI_MODE_DISABLED:
1983 switch (op) {
1984 case QLA2XXX_INI_MODE_DISABLED:
1985 if (qla_tgt_mode_enabled(vha)) {
1986 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld) !=
1987 vha->hw->flags.exchoffld_enabled)
1988 eo_toggle = 1;
1989 if (((vha->ql2xexchoffld !=
1990 vha->u_ql2xexchoffld) &&
1991 NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld)) ||
1992 eo_toggle) {
1993 /*
1994 * The number of exchange to be offload
1995 * was tweaked or offload option was
1996 * flipped
1997 */
1998 action = MODE_CHANGE_ACCEPT;
1999 } else {
2000 action = MODE_CHANGE_NO_ACTION;
2001 }
2002 } else {
2003 action = MODE_CHANGE_NO_ACTION;
2004 }
2005 break;
2006 case QLA2XXX_INI_MODE_EXCLUSIVE:
2007 if (qla_tgt_mode_enabled(vha)) {
2008 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld) !=
2009 vha->hw->flags.exchoffld_enabled)
2010 eo_toggle = 1;
2011 if (((vha->ql2xexchoffld !=
2012 vha->u_ql2xexchoffld) &&
2013 NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld)) ||
2014 eo_toggle) {
2015 /*
2016 * The number of exchange to be offload
2017 * was tweaked or offload option was
2018 * flipped
2019 */
2020 action = MODE_CHANGE_ACCEPT;
2021 } else {
2022 action = MODE_CHANGE_NO_ACTION;
2023 }
2024 } else {
2025 action = MODE_CHANGE_ACCEPT;
2026 }
2027 break;
2028 case QLA2XXX_INI_MODE_DUAL:
2029 action = MODE_CHANGE_ACCEPT;
2030 /* active_mode is target only, reset it to dual */
2031 if (qla_tgt_mode_enabled(vha)) {
2032 set_mode = 1;
2033 action = MODE_CHANGE_ACCEPT;
2034 } else {
2035 action = MODE_CHANGE_NO_ACTION;
2036 }
2037 break;
2038
2039 case QLA2XXX_INI_MODE_ENABLED:
2040 if (qla_tgt_mode_enabled(vha))
2041 action = TARGET_STILL_ACTIVE;
2042 else {
2043 action = MODE_CHANGE_ACCEPT;
2044 set_mode = 1;
2045 }
2046 break;
2047 }
2048 break;
2049
2050 case QLA2XXX_INI_MODE_EXCLUSIVE:
2051 switch (op) {
2052 case QLA2XXX_INI_MODE_EXCLUSIVE:
2053 if (qla_tgt_mode_enabled(vha)) {
2054 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld) !=
2055 vha->hw->flags.exchoffld_enabled)
2056 eo_toggle = 1;
2057 if (((vha->ql2xexchoffld !=
2058 vha->u_ql2xexchoffld) &&
2059 NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld)) ||
2060 eo_toggle)
2061 /*
2062 * The number of exchange to be offload
2063 * was tweaked or offload option was
2064 * flipped
2065 */
2066 action = MODE_CHANGE_ACCEPT;
2067 else
2068 action = NO_ACTION;
2069 } else
2070 action = NO_ACTION;
2071
2072 break;
2073
2074 case QLA2XXX_INI_MODE_DISABLED:
2075 if (qla_tgt_mode_enabled(vha)) {
2076 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld) !=
2077 vha->hw->flags.exchoffld_enabled)
2078 eo_toggle = 1;
2079 if (((vha->ql2xexchoffld !=
2080 vha->u_ql2xexchoffld) &&
2081 NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld)) ||
2082 eo_toggle)
2083 action = MODE_CHANGE_ACCEPT;
2084 else
2085 action = MODE_CHANGE_NO_ACTION;
2086 } else
2087 action = MODE_CHANGE_NO_ACTION;
2088 break;
2089
2090 case QLA2XXX_INI_MODE_DUAL: /* exclusive -> dual */
2091 if (qla_tgt_mode_enabled(vha)) {
2092 action = MODE_CHANGE_ACCEPT;
2093 set_mode = 1;
2094 } else
2095 action = MODE_CHANGE_ACCEPT;
2096 break;
2097
2098 case QLA2XXX_INI_MODE_ENABLED:
2099 if (qla_tgt_mode_enabled(vha))
2100 action = TARGET_STILL_ACTIVE;
2101 else {
2102 if (vha->hw->flags.fw_started)
2103 action = MODE_CHANGE_NO_ACTION;
2104 else
2105 action = MODE_CHANGE_ACCEPT;
2106 }
2107 break;
2108 }
2109 break;
2110
2111 case QLA2XXX_INI_MODE_ENABLED:
2112 switch (op) {
2113 case QLA2XXX_INI_MODE_ENABLED:
2114 if (NEED_EXCH_OFFLOAD(vha->u_ql2xiniexchg) !=
2115 vha->hw->flags.exchoffld_enabled)
2116 eo_toggle = 1;
2117 if (((vha->ql2xiniexchg != vha->u_ql2xiniexchg) &&
2118 NEED_EXCH_OFFLOAD(vha->u_ql2xiniexchg)) ||
2119 eo_toggle)
2120 action = MODE_CHANGE_ACCEPT;
2121 else
2122 action = NO_ACTION;
2123 break;
2124 case QLA2XXX_INI_MODE_DUAL:
2125 case QLA2XXX_INI_MODE_DISABLED:
2126 action = MODE_CHANGE_ACCEPT;
2127 break;
2128 default:
2129 action = MODE_CHANGE_NO_ACTION;
2130 break;
2131 }
2132 break;
2133
2134 case QLA2XXX_INI_MODE_DUAL:
2135 switch (op) {
2136 case QLA2XXX_INI_MODE_DUAL:
2137 if (qla_tgt_mode_enabled(vha) ||
2138 qla_dual_mode_enabled(vha)) {
2139 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld +
2140 vha->u_ql2xiniexchg) !=
2141 vha->hw->flags.exchoffld_enabled)
2142 eo_toggle = 1;
2143
2144 if ((((vha->ql2xexchoffld +
2145 vha->ql2xiniexchg) !=
2146 (vha->u_ql2xiniexchg +
2147 vha->u_ql2xexchoffld)) &&
2148 NEED_EXCH_OFFLOAD(vha->u_ql2xiniexchg +
2149 vha->u_ql2xexchoffld)) || eo_toggle)
2150 action = MODE_CHANGE_ACCEPT;
2151 else
2152 action = NO_ACTION;
2153 } else {
2154 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld +
2155 vha->u_ql2xiniexchg) !=
2156 vha->hw->flags.exchoffld_enabled)
2157 eo_toggle = 1;
2158
2159 if ((((vha->ql2xexchoffld + vha->ql2xiniexchg)
2160 != (vha->u_ql2xiniexchg +
2161 vha->u_ql2xexchoffld)) &&
2162 NEED_EXCH_OFFLOAD(vha->u_ql2xiniexchg +
2163 vha->u_ql2xexchoffld)) || eo_toggle)
2164 action = MODE_CHANGE_NO_ACTION;
2165 else
2166 action = NO_ACTION;
2167 }
2168 break;
2169
2170 case QLA2XXX_INI_MODE_DISABLED:
2171 if (qla_tgt_mode_enabled(vha) ||
2172 qla_dual_mode_enabled(vha)) {
2173 /* turning off initiator mode */
2174 set_mode = 1;
2175 action = MODE_CHANGE_ACCEPT;
2176 } else {
2177 action = MODE_CHANGE_NO_ACTION;
2178 }
2179 break;
2180
2181 case QLA2XXX_INI_MODE_EXCLUSIVE:
2182 if (qla_tgt_mode_enabled(vha) ||
2183 qla_dual_mode_enabled(vha)) {
2184 set_mode = 1;
2185 action = MODE_CHANGE_ACCEPT;
2186 } else {
2187 action = MODE_CHANGE_ACCEPT;
2188 }
2189 break;
2190
2191 case QLA2XXX_INI_MODE_ENABLED:
2192 if (qla_tgt_mode_enabled(vha) ||
2193 qla_dual_mode_enabled(vha)) {
2194 action = TARGET_STILL_ACTIVE;
2195 } else {
2196 action = MODE_CHANGE_ACCEPT;
2197 }
2198 }
2199 break;
2200 }
2201
2202 switch (action) {
2203 case MODE_CHANGE_ACCEPT:
2204 ql_log(ql_log_warn, vha, 0xffff,
2205 "Mode change accepted. From %s to %s, Tgt exchg %d|%d. ini exchg %d|%d\n",
2206 mode_to_str[vha->qlini_mode], mode_to_str[op],
2207 vha->ql2xexchoffld, vha->u_ql2xexchoffld,
2208 vha->ql2xiniexchg, vha->u_ql2xiniexchg);
2209
2210 vha->qlini_mode = op;
2211 vha->ql2xexchoffld = vha->u_ql2xexchoffld;
2212 vha->ql2xiniexchg = vha->u_ql2xiniexchg;
2213 if (set_mode)
2214 qlt_set_mode(vha);
2215 vha->flags.online = 1;
2216 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
2217 break;
2218
2219 case MODE_CHANGE_NO_ACTION:
2220 ql_log(ql_log_warn, vha, 0xffff,
2221 "Mode is set. No action taken. From %s to %s, Tgt exchg %d|%d. ini exchg %d|%d\n",
2222 mode_to_str[vha->qlini_mode], mode_to_str[op],
2223 vha->ql2xexchoffld, vha->u_ql2xexchoffld,
2224 vha->ql2xiniexchg, vha->u_ql2xiniexchg);
2225 vha->qlini_mode = op;
2226 vha->ql2xexchoffld = vha->u_ql2xexchoffld;
2227 vha->ql2xiniexchg = vha->u_ql2xiniexchg;
2228 break;
2229
2230 case TARGET_STILL_ACTIVE:
2231 ql_log(ql_log_warn, vha, 0xffff,
2232 "Target Mode is active. Unable to change Mode.\n");
2233 break;
2234
2235 case NO_ACTION:
2236 default:
2237 ql_log(ql_log_warn, vha, 0xffff,
2238 "Mode unchange. No action taken. %d|%d pct %d|%d.\n",
2239 vha->qlini_mode, op,
2240 vha->ql2xexchoffld, vha->u_ql2xexchoffld);
2241 break;
2242 }
2243 }
2244
2245 static ssize_t
qlini_mode_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)2246 qlini_mode_store(struct device *dev, struct device_attribute *attr,
2247 const char *buf, size_t count)
2248 {
2249 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2250 int ini;
2251
2252 if (!buf)
2253 return -EINVAL;
2254
2255 if (strncasecmp(QLA2XXX_INI_MODE_STR_EXCLUSIVE, buf,
2256 strlen(QLA2XXX_INI_MODE_STR_EXCLUSIVE)) == 0)
2257 ini = QLA2XXX_INI_MODE_EXCLUSIVE;
2258 else if (strncasecmp(QLA2XXX_INI_MODE_STR_DISABLED, buf,
2259 strlen(QLA2XXX_INI_MODE_STR_DISABLED)) == 0)
2260 ini = QLA2XXX_INI_MODE_DISABLED;
2261 else if (strncasecmp(QLA2XXX_INI_MODE_STR_ENABLED, buf,
2262 strlen(QLA2XXX_INI_MODE_STR_ENABLED)) == 0)
2263 ini = QLA2XXX_INI_MODE_ENABLED;
2264 else if (strncasecmp(QLA2XXX_INI_MODE_STR_DUAL, buf,
2265 strlen(QLA2XXX_INI_MODE_STR_DUAL)) == 0)
2266 ini = QLA2XXX_INI_MODE_DUAL;
2267 else
2268 return -EINVAL;
2269
2270 qla_set_ini_mode(vha, ini);
2271 return strlen(buf);
2272 }
2273
2274 static ssize_t
ql2xexchoffld_show(struct device * dev,struct device_attribute * attr,char * buf)2275 ql2xexchoffld_show(struct device *dev, struct device_attribute *attr,
2276 char *buf)
2277 {
2278 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2279 int len = 0;
2280
2281 len += scnprintf(buf + len, PAGE_SIZE-len,
2282 "target exchange: new %d : current: %d\n\n",
2283 vha->u_ql2xexchoffld, vha->ql2xexchoffld);
2284
2285 len += scnprintf(buf + len, PAGE_SIZE-len,
2286 "Please (re)set operating mode via \"/sys/class/scsi_host/host%ld/qlini_mode\" to load new setting.\n",
2287 vha->host_no);
2288
2289 return len;
2290 }
2291
2292 static ssize_t
ql2xexchoffld_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)2293 ql2xexchoffld_store(struct device *dev, struct device_attribute *attr,
2294 const char *buf, size_t count)
2295 {
2296 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2297 int val = 0;
2298
2299 if (sscanf(buf, "%d", &val) != 1)
2300 return -EINVAL;
2301
2302 if (val > FW_MAX_EXCHANGES_CNT)
2303 val = FW_MAX_EXCHANGES_CNT;
2304 else if (val < 0)
2305 val = 0;
2306
2307 vha->u_ql2xexchoffld = val;
2308 return strlen(buf);
2309 }
2310
2311 static ssize_t
ql2xiniexchg_show(struct device * dev,struct device_attribute * attr,char * buf)2312 ql2xiniexchg_show(struct device *dev, struct device_attribute *attr,
2313 char *buf)
2314 {
2315 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2316 int len = 0;
2317
2318 len += scnprintf(buf + len, PAGE_SIZE-len,
2319 "target exchange: new %d : current: %d\n\n",
2320 vha->u_ql2xiniexchg, vha->ql2xiniexchg);
2321
2322 len += scnprintf(buf + len, PAGE_SIZE-len,
2323 "Please (re)set operating mode via \"/sys/class/scsi_host/host%ld/qlini_mode\" to load new setting.\n",
2324 vha->host_no);
2325
2326 return len;
2327 }
2328
2329 static ssize_t
ql2xiniexchg_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)2330 ql2xiniexchg_store(struct device *dev, struct device_attribute *attr,
2331 const char *buf, size_t count)
2332 {
2333 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2334 int val = 0;
2335
2336 if (sscanf(buf, "%d", &val) != 1)
2337 return -EINVAL;
2338
2339 if (val > FW_MAX_EXCHANGES_CNT)
2340 val = FW_MAX_EXCHANGES_CNT;
2341 else if (val < 0)
2342 val = 0;
2343
2344 vha->u_ql2xiniexchg = val;
2345 return strlen(buf);
2346 }
2347
2348 static ssize_t
qla2x00_dif_bundle_statistics_show(struct device * dev,struct device_attribute * attr,char * buf)2349 qla2x00_dif_bundle_statistics_show(struct device *dev,
2350 struct device_attribute *attr, char *buf)
2351 {
2352 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2353 struct qla_hw_data *ha = vha->hw;
2354
2355 return scnprintf(buf, PAGE_SIZE,
2356 "cross=%llu read=%llu write=%llu kalloc=%llu dma_alloc=%llu unusable=%u\n",
2357 ha->dif_bundle_crossed_pages, ha->dif_bundle_reads,
2358 ha->dif_bundle_writes, ha->dif_bundle_kallocs,
2359 ha->dif_bundle_dma_allocs, ha->pool.unusable.count);
2360 }
2361
2362 static ssize_t
qla2x00_fw_attr_show(struct device * dev,struct device_attribute * attr,char * buf)2363 qla2x00_fw_attr_show(struct device *dev,
2364 struct device_attribute *attr, char *buf)
2365 {
2366 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2367 struct qla_hw_data *ha = vha->hw;
2368
2369 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
2370 return scnprintf(buf, PAGE_SIZE, "\n");
2371
2372 return scnprintf(buf, PAGE_SIZE, "%llx\n",
2373 (uint64_t)ha->fw_attributes_ext[1] << 48 |
2374 (uint64_t)ha->fw_attributes_ext[0] << 32 |
2375 (uint64_t)ha->fw_attributes_h << 16 |
2376 (uint64_t)ha->fw_attributes);
2377 }
2378
2379 static ssize_t
qla2x00_port_no_show(struct device * dev,struct device_attribute * attr,char * buf)2380 qla2x00_port_no_show(struct device *dev, struct device_attribute *attr,
2381 char *buf)
2382 {
2383 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2384
2385 return scnprintf(buf, PAGE_SIZE, "%u\n", vha->hw->port_no);
2386 }
2387
2388 static ssize_t
qla2x00_dport_diagnostics_show(struct device * dev,struct device_attribute * attr,char * buf)2389 qla2x00_dport_diagnostics_show(struct device *dev,
2390 struct device_attribute *attr, char *buf)
2391 {
2392 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2393
2394 if (!IS_QLA83XX(vha->hw) && !IS_QLA27XX(vha->hw) &&
2395 !IS_QLA28XX(vha->hw))
2396 return scnprintf(buf, PAGE_SIZE, "\n");
2397
2398 if (!*vha->dport_data)
2399 return scnprintf(buf, PAGE_SIZE, "\n");
2400
2401 return scnprintf(buf, PAGE_SIZE, "%04x %04x %04x %04x\n",
2402 vha->dport_data[0], vha->dport_data[1],
2403 vha->dport_data[2], vha->dport_data[3]);
2404 }
2405 static DEVICE_ATTR(dport_diagnostics, 0444,
2406 qla2x00_dport_diagnostics_show, NULL);
2407
2408 static DEVICE_STRING_ATTR_RO(driver_version, S_IRUGO, qla2x00_version_str);
2409 static DEVICE_ATTR(fw_version, S_IRUGO, qla2x00_fw_version_show, NULL);
2410 static DEVICE_ATTR(serial_num, S_IRUGO, qla2x00_serial_num_show, NULL);
2411 static DEVICE_ATTR(isp_name, S_IRUGO, qla2x00_isp_name_show, NULL);
2412 static DEVICE_ATTR(isp_id, S_IRUGO, qla2x00_isp_id_show, NULL);
2413 static DEVICE_ATTR(model_name, S_IRUGO, qla2x00_model_name_show, NULL);
2414 static DEVICE_ATTR(model_desc, S_IRUGO, qla2x00_model_desc_show, NULL);
2415 static DEVICE_ATTR(pci_info, S_IRUGO, qla2x00_pci_info_show, NULL);
2416 static DEVICE_ATTR(link_state, S_IRUGO, qla2x00_link_state_show, NULL);
2417 static DEVICE_ATTR(zio, S_IRUGO | S_IWUSR, qla2x00_zio_show, qla2x00_zio_store);
2418 static DEVICE_ATTR(zio_timer, S_IRUGO | S_IWUSR, qla2x00_zio_timer_show,
2419 qla2x00_zio_timer_store);
2420 static DEVICE_ATTR(beacon, S_IRUGO | S_IWUSR, qla2x00_beacon_show,
2421 qla2x00_beacon_store);
2422 static DEVICE_ATTR(beacon_config, 0644, qla2x00_beacon_config_show,
2423 qla2x00_beacon_config_store);
2424 static DEVICE_ATTR(optrom_bios_version, S_IRUGO,
2425 qla2x00_optrom_bios_version_show, NULL);
2426 static DEVICE_ATTR(optrom_efi_version, S_IRUGO,
2427 qla2x00_optrom_efi_version_show, NULL);
2428 static DEVICE_ATTR(optrom_fcode_version, S_IRUGO,
2429 qla2x00_optrom_fcode_version_show, NULL);
2430 static DEVICE_ATTR(optrom_fw_version, S_IRUGO, qla2x00_optrom_fw_version_show,
2431 NULL);
2432 static DEVICE_ATTR(optrom_gold_fw_version, S_IRUGO,
2433 qla2x00_optrom_gold_fw_version_show, NULL);
2434 static DEVICE_ATTR(84xx_fw_version, S_IRUGO, qla24xx_84xx_fw_version_show,
2435 NULL);
2436 static DEVICE_ATTR(total_isp_aborts, S_IRUGO, qla2x00_total_isp_aborts_show,
2437 NULL);
2438 static DEVICE_ATTR(serdes_version, 0444, qla2x00_serdes_version_show, NULL);
2439 static DEVICE_ATTR(mpi_version, S_IRUGO, qla2x00_mpi_version_show, NULL);
2440 static DEVICE_ATTR(phy_version, S_IRUGO, qla2x00_phy_version_show, NULL);
2441 static DEVICE_ATTR(flash_block_size, S_IRUGO, qla2x00_flash_block_size_show,
2442 NULL);
2443 static DEVICE_ATTR(vlan_id, S_IRUGO, qla2x00_vlan_id_show, NULL);
2444 static DEVICE_ATTR(vn_port_mac_address, S_IRUGO,
2445 qla2x00_vn_port_mac_address_show, NULL);
2446 static DEVICE_ATTR(fabric_param, S_IRUGO, qla2x00_fabric_param_show, NULL);
2447 static DEVICE_ATTR(fw_state, S_IRUGO, qla2x00_fw_state_show, NULL);
2448 static DEVICE_ATTR(thermal_temp, S_IRUGO, qla2x00_thermal_temp_show, NULL);
2449 static DEVICE_ATTR(diag_requests, S_IRUGO, qla2x00_diag_requests_show, NULL);
2450 static DEVICE_ATTR(diag_megabytes, S_IRUGO, qla2x00_diag_megabytes_show, NULL);
2451 static DEVICE_ATTR(fw_dump_size, S_IRUGO, qla2x00_fw_dump_size_show, NULL);
2452 static DEVICE_ATTR(allow_cna_fw_dump, S_IRUGO | S_IWUSR,
2453 qla2x00_allow_cna_fw_dump_show,
2454 qla2x00_allow_cna_fw_dump_store);
2455 static DEVICE_ATTR(pep_version, S_IRUGO, qla2x00_pep_version_show, NULL);
2456 static DEVICE_ATTR(min_supported_speed, 0444,
2457 qla2x00_min_supported_speed_show, NULL);
2458 static DEVICE_ATTR(max_supported_speed, 0444,
2459 qla2x00_max_supported_speed_show, NULL);
2460 static DEVICE_ATTR(zio_threshold, 0644,
2461 qla_zio_threshold_show,
2462 qla_zio_threshold_store);
2463 static DEVICE_ATTR_RW(qlini_mode);
2464 static DEVICE_ATTR_RW(ql2xexchoffld);
2465 static DEVICE_ATTR_RW(ql2xiniexchg);
2466 static DEVICE_ATTR(dif_bundle_statistics, 0444,
2467 qla2x00_dif_bundle_statistics_show, NULL);
2468 static DEVICE_ATTR(port_speed, 0644, qla2x00_port_speed_show,
2469 qla2x00_port_speed_store);
2470 static DEVICE_ATTR(port_no, 0444, qla2x00_port_no_show, NULL);
2471 static DEVICE_ATTR(fw_attr, 0444, qla2x00_fw_attr_show, NULL);
2472
2473 static struct attribute *qla2x00_host_attrs[] = {
2474 &dev_attr_driver_version.attr.attr,
2475 &dev_attr_fw_version.attr,
2476 &dev_attr_serial_num.attr,
2477 &dev_attr_isp_name.attr,
2478 &dev_attr_isp_id.attr,
2479 &dev_attr_model_name.attr,
2480 &dev_attr_model_desc.attr,
2481 &dev_attr_pci_info.attr,
2482 &dev_attr_link_state.attr,
2483 &dev_attr_zio.attr,
2484 &dev_attr_zio_timer.attr,
2485 &dev_attr_beacon.attr,
2486 &dev_attr_beacon_config.attr,
2487 &dev_attr_optrom_bios_version.attr,
2488 &dev_attr_optrom_efi_version.attr,
2489 &dev_attr_optrom_fcode_version.attr,
2490 &dev_attr_optrom_fw_version.attr,
2491 &dev_attr_84xx_fw_version.attr,
2492 &dev_attr_total_isp_aborts.attr,
2493 &dev_attr_serdes_version.attr,
2494 &dev_attr_mpi_version.attr,
2495 &dev_attr_phy_version.attr,
2496 &dev_attr_flash_block_size.attr,
2497 &dev_attr_vlan_id.attr,
2498 &dev_attr_vn_port_mac_address.attr,
2499 &dev_attr_fabric_param.attr,
2500 &dev_attr_fw_state.attr,
2501 &dev_attr_optrom_gold_fw_version.attr,
2502 &dev_attr_thermal_temp.attr,
2503 &dev_attr_diag_requests.attr,
2504 &dev_attr_diag_megabytes.attr,
2505 &dev_attr_fw_dump_size.attr,
2506 &dev_attr_allow_cna_fw_dump.attr,
2507 &dev_attr_pep_version.attr,
2508 &dev_attr_min_supported_speed.attr,
2509 &dev_attr_max_supported_speed.attr,
2510 &dev_attr_zio_threshold.attr,
2511 &dev_attr_dif_bundle_statistics.attr,
2512 &dev_attr_port_speed.attr,
2513 &dev_attr_port_no.attr,
2514 &dev_attr_fw_attr.attr,
2515 &dev_attr_dport_diagnostics.attr,
2516 &dev_attr_mpi_pause.attr,
2517 &dev_attr_qlini_mode.attr,
2518 &dev_attr_ql2xiniexchg.attr,
2519 &dev_attr_ql2xexchoffld.attr,
2520 NULL,
2521 };
2522
qla_host_attr_is_visible(struct kobject * kobj,struct attribute * attr,int i)2523 static umode_t qla_host_attr_is_visible(struct kobject *kobj,
2524 struct attribute *attr, int i)
2525 {
2526 if (ql2x_ini_mode != QLA2XXX_INI_MODE_DUAL &&
2527 (attr == &dev_attr_qlini_mode.attr ||
2528 attr == &dev_attr_ql2xiniexchg.attr ||
2529 attr == &dev_attr_ql2xexchoffld.attr))
2530 return 0;
2531 return attr->mode;
2532 }
2533
2534 static const struct attribute_group qla2x00_host_attr_group = {
2535 .is_visible = qla_host_attr_is_visible,
2536 .attrs = qla2x00_host_attrs
2537 };
2538
2539 const struct attribute_group *qla2x00_host_groups[] = {
2540 &qla2x00_host_attr_group,
2541 NULL
2542 };
2543
2544 /* Host attributes. */
2545
2546 static void
qla2x00_get_host_port_id(struct Scsi_Host * shost)2547 qla2x00_get_host_port_id(struct Scsi_Host *shost)
2548 {
2549 scsi_qla_host_t *vha = shost_priv(shost);
2550
2551 fc_host_port_id(shost) = vha->d_id.b.domain << 16 |
2552 vha->d_id.b.area << 8 | vha->d_id.b.al_pa;
2553 }
2554
2555 static void
qla2x00_get_host_speed(struct Scsi_Host * shost)2556 qla2x00_get_host_speed(struct Scsi_Host *shost)
2557 {
2558 scsi_qla_host_t *vha = shost_priv(shost);
2559 u32 speed;
2560
2561 if (IS_QLAFX00(vha->hw)) {
2562 qlafx00_get_host_speed(shost);
2563 return;
2564 }
2565
2566 switch (vha->hw->link_data_rate) {
2567 case PORT_SPEED_1GB:
2568 speed = FC_PORTSPEED_1GBIT;
2569 break;
2570 case PORT_SPEED_2GB:
2571 speed = FC_PORTSPEED_2GBIT;
2572 break;
2573 case PORT_SPEED_4GB:
2574 speed = FC_PORTSPEED_4GBIT;
2575 break;
2576 case PORT_SPEED_8GB:
2577 speed = FC_PORTSPEED_8GBIT;
2578 break;
2579 case PORT_SPEED_10GB:
2580 speed = FC_PORTSPEED_10GBIT;
2581 break;
2582 case PORT_SPEED_16GB:
2583 speed = FC_PORTSPEED_16GBIT;
2584 break;
2585 case PORT_SPEED_32GB:
2586 speed = FC_PORTSPEED_32GBIT;
2587 break;
2588 case PORT_SPEED_64GB:
2589 speed = FC_PORTSPEED_64GBIT;
2590 break;
2591 default:
2592 speed = FC_PORTSPEED_UNKNOWN;
2593 break;
2594 }
2595
2596 fc_host_speed(shost) = speed;
2597 }
2598
2599 static void
qla2x00_get_host_port_type(struct Scsi_Host * shost)2600 qla2x00_get_host_port_type(struct Scsi_Host *shost)
2601 {
2602 scsi_qla_host_t *vha = shost_priv(shost);
2603 uint32_t port_type;
2604
2605 if (vha->vp_idx) {
2606 fc_host_port_type(shost) = FC_PORTTYPE_NPIV;
2607 return;
2608 }
2609 switch (vha->hw->current_topology) {
2610 case ISP_CFG_NL:
2611 port_type = FC_PORTTYPE_LPORT;
2612 break;
2613 case ISP_CFG_FL:
2614 port_type = FC_PORTTYPE_NLPORT;
2615 break;
2616 case ISP_CFG_N:
2617 port_type = FC_PORTTYPE_PTP;
2618 break;
2619 case ISP_CFG_F:
2620 port_type = FC_PORTTYPE_NPORT;
2621 break;
2622 default:
2623 port_type = FC_PORTTYPE_UNKNOWN;
2624 break;
2625 }
2626
2627 fc_host_port_type(shost) = port_type;
2628 }
2629
2630 static void
qla2x00_get_starget_node_name(struct scsi_target * starget)2631 qla2x00_get_starget_node_name(struct scsi_target *starget)
2632 {
2633 struct Scsi_Host *host = dev_to_shost(starget->dev.parent);
2634 scsi_qla_host_t *vha = shost_priv(host);
2635 fc_port_t *fcport;
2636 u64 node_name = 0;
2637
2638 list_for_each_entry(fcport, &vha->vp_fcports, list) {
2639 if (fcport->rport &&
2640 starget->id == fcport->rport->scsi_target_id) {
2641 node_name = wwn_to_u64(fcport->node_name);
2642 break;
2643 }
2644 }
2645
2646 fc_starget_node_name(starget) = node_name;
2647 }
2648
2649 static void
qla2x00_get_starget_port_name(struct scsi_target * starget)2650 qla2x00_get_starget_port_name(struct scsi_target *starget)
2651 {
2652 struct Scsi_Host *host = dev_to_shost(starget->dev.parent);
2653 scsi_qla_host_t *vha = shost_priv(host);
2654 fc_port_t *fcport;
2655 u64 port_name = 0;
2656
2657 list_for_each_entry(fcport, &vha->vp_fcports, list) {
2658 if (fcport->rport &&
2659 starget->id == fcport->rport->scsi_target_id) {
2660 port_name = wwn_to_u64(fcport->port_name);
2661 break;
2662 }
2663 }
2664
2665 fc_starget_port_name(starget) = port_name;
2666 }
2667
2668 static void
qla2x00_get_starget_port_id(struct scsi_target * starget)2669 qla2x00_get_starget_port_id(struct scsi_target *starget)
2670 {
2671 struct Scsi_Host *host = dev_to_shost(starget->dev.parent);
2672 scsi_qla_host_t *vha = shost_priv(host);
2673 fc_port_t *fcport;
2674 uint32_t port_id = ~0U;
2675
2676 list_for_each_entry(fcport, &vha->vp_fcports, list) {
2677 if (fcport->rport &&
2678 starget->id == fcport->rport->scsi_target_id) {
2679 port_id = fcport->d_id.b.domain << 16 |
2680 fcport->d_id.b.area << 8 | fcport->d_id.b.al_pa;
2681 break;
2682 }
2683 }
2684
2685 fc_starget_port_id(starget) = port_id;
2686 }
2687
2688 static inline void
qla2x00_set_rport_loss_tmo(struct fc_rport * rport,uint32_t timeout)2689 qla2x00_set_rport_loss_tmo(struct fc_rport *rport, uint32_t timeout)
2690 {
2691 fc_port_t *fcport = *(fc_port_t **)rport->dd_data;
2692
2693 rport->dev_loss_tmo = timeout ? timeout : 1;
2694
2695 if (IS_ENABLED(CONFIG_NVME_FC) && fcport && fcport->nvme_remote_port)
2696 nvme_fc_set_remoteport_devloss(fcport->nvme_remote_port,
2697 rport->dev_loss_tmo);
2698 }
2699
2700 static void
qla2x00_dev_loss_tmo_callbk(struct fc_rport * rport)2701 qla2x00_dev_loss_tmo_callbk(struct fc_rport *rport)
2702 {
2703 struct Scsi_Host *host = rport_to_shost(rport);
2704 fc_port_t *fcport = *(fc_port_t **)rport->dd_data;
2705 unsigned long flags;
2706
2707 if (!fcport)
2708 return;
2709
2710 ql_dbg(ql_dbg_async, fcport->vha, 0x5101,
2711 DBG_FCPORT_PRFMT(fcport, "dev_loss_tmo expiry, rport_state=%d",
2712 rport->port_state));
2713
2714 /*
2715 * Now that the rport has been deleted, set the fcport state to
2716 * FCS_DEVICE_DEAD, if the fcport is still lost.
2717 */
2718 if (fcport->scan_state != QLA_FCPORT_FOUND)
2719 qla2x00_set_fcport_state(fcport, FCS_DEVICE_DEAD);
2720
2721 /*
2722 * Transport has effectively 'deleted' the rport, clear
2723 * all local references.
2724 */
2725 spin_lock_irqsave(host->host_lock, flags);
2726 /* Confirm port has not reappeared before clearing pointers. */
2727 if (rport->port_state != FC_PORTSTATE_ONLINE) {
2728 fcport->rport = NULL;
2729 *((fc_port_t **)rport->dd_data) = NULL;
2730 }
2731 spin_unlock_irqrestore(host->host_lock, flags);
2732
2733 if (test_bit(ABORT_ISP_ACTIVE, &fcport->vha->dpc_flags))
2734 return;
2735
2736 if (unlikely(pci_channel_offline(fcport->vha->hw->pdev))) {
2737 /* Will wait for wind down of adapter */
2738 ql_dbg(ql_dbg_aer, fcport->vha, 0x900c,
2739 "%s pci offline detected (id %06x)\n", __func__,
2740 fcport->d_id.b24);
2741 qla_pci_set_eeh_busy(fcport->vha);
2742 qla2x00_eh_wait_for_pending_commands(fcport->vha, fcport->d_id.b24,
2743 0, WAIT_TARGET);
2744 return;
2745 }
2746 }
2747
2748 static void
qla2x00_terminate_rport_io(struct fc_rport * rport)2749 qla2x00_terminate_rport_io(struct fc_rport *rport)
2750 {
2751 fc_port_t *fcport = *(fc_port_t **)rport->dd_data;
2752 scsi_qla_host_t *vha;
2753
2754 if (!fcport)
2755 return;
2756
2757 if (test_bit(UNLOADING, &fcport->vha->dpc_flags))
2758 return;
2759
2760 if (test_bit(ABORT_ISP_ACTIVE, &fcport->vha->dpc_flags))
2761 return;
2762 vha = fcport->vha;
2763
2764 if (unlikely(pci_channel_offline(fcport->vha->hw->pdev))) {
2765 /* Will wait for wind down of adapter */
2766 ql_dbg(ql_dbg_aer, fcport->vha, 0x900b,
2767 "%s pci offline detected (id %06x)\n", __func__,
2768 fcport->d_id.b24);
2769 qla_pci_set_eeh_busy(vha);
2770 qla2x00_eh_wait_for_pending_commands(fcport->vha, fcport->d_id.b24,
2771 0, WAIT_TARGET);
2772 return;
2773 }
2774 /*
2775 * At this point all fcport's software-states are cleared. Perform any
2776 * final cleanup of firmware resources (PCBs and XCBs).
2777 *
2778 * Attempt to cleanup only lost devices.
2779 */
2780 if (fcport->loop_id != FC_NO_LOOP_ID) {
2781 if (IS_FWI2_CAPABLE(fcport->vha->hw) &&
2782 fcport->scan_state != QLA_FCPORT_FOUND) {
2783 if (fcport->loop_id != FC_NO_LOOP_ID)
2784 fcport->logout_on_delete = 1;
2785
2786 if (!EDIF_NEGOTIATION_PENDING(fcport)) {
2787 ql_dbg(ql_dbg_disc, fcport->vha, 0x911e,
2788 "%s %d schedule session deletion\n", __func__,
2789 __LINE__);
2790 qlt_schedule_sess_for_deletion(fcport);
2791 }
2792 } else if (!IS_FWI2_CAPABLE(fcport->vha->hw)) {
2793 qla2x00_port_logout(fcport->vha, fcport);
2794 }
2795 }
2796
2797 /* check for any straggling io left behind */
2798 if (qla2x00_eh_wait_for_pending_commands(fcport->vha, fcport->d_id.b24, 0, WAIT_TARGET)) {
2799 ql_log(ql_log_warn, vha, 0x300b,
2800 "IO not return. Resetting. \n");
2801 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
2802 qla2xxx_wake_dpc(vha);
2803 qla2x00_wait_for_chip_reset(vha);
2804 }
2805 }
2806
2807 static int
qla2x00_issue_lip(struct Scsi_Host * shost)2808 qla2x00_issue_lip(struct Scsi_Host *shost)
2809 {
2810 scsi_qla_host_t *vha = shost_priv(shost);
2811
2812 if (IS_QLAFX00(vha->hw))
2813 return 0;
2814
2815 if (vha->hw->flags.port_isolated)
2816 return 0;
2817
2818 qla2x00_loop_reset(vha);
2819 return 0;
2820 }
2821
2822 static struct fc_host_statistics *
qla2x00_get_fc_host_stats(struct Scsi_Host * shost)2823 qla2x00_get_fc_host_stats(struct Scsi_Host *shost)
2824 {
2825 scsi_qla_host_t *vha = shost_priv(shost);
2826 struct qla_hw_data *ha = vha->hw;
2827 struct scsi_qla_host *base_vha = pci_get_drvdata(ha->pdev);
2828 int rval;
2829 struct link_statistics *stats;
2830 dma_addr_t stats_dma;
2831 struct fc_host_statistics *p = &vha->fc_host_stat;
2832 struct qla_qpair *qpair;
2833 int i;
2834 u64 ib = 0, ob = 0, ir = 0, or = 0;
2835
2836 memset(p, -1, sizeof(*p));
2837
2838 if (IS_QLAFX00(vha->hw))
2839 goto done;
2840
2841 if (test_bit(UNLOADING, &vha->dpc_flags))
2842 goto done;
2843
2844 if (unlikely(pci_channel_offline(ha->pdev)))
2845 goto done;
2846
2847 if (qla2x00_chip_is_down(vha))
2848 goto done;
2849
2850 stats = dma_alloc_coherent(&ha->pdev->dev, sizeof(*stats), &stats_dma,
2851 GFP_KERNEL);
2852 if (!stats) {
2853 ql_log(ql_log_warn, vha, 0x707d,
2854 "Failed to allocate memory for stats.\n");
2855 goto done;
2856 }
2857
2858 rval = QLA_FUNCTION_FAILED;
2859 if (IS_FWI2_CAPABLE(ha)) {
2860 rval = qla24xx_get_isp_stats(base_vha, stats, stats_dma, 0);
2861 } else if (atomic_read(&base_vha->loop_state) == LOOP_READY &&
2862 !ha->dpc_active) {
2863 /* Must be in a 'READY' state for statistics retrieval. */
2864 rval = qla2x00_get_link_status(base_vha, base_vha->loop_id,
2865 stats, stats_dma);
2866 }
2867
2868 if (rval != QLA_SUCCESS)
2869 goto done_free;
2870
2871 /* --- */
2872 for (i = 0; i < vha->hw->max_qpairs; i++) {
2873 qpair = vha->hw->queue_pair_map[i];
2874 if (!qpair)
2875 continue;
2876 ir += qpair->counters.input_requests;
2877 or += qpair->counters.output_requests;
2878 ib += qpair->counters.input_bytes;
2879 ob += qpair->counters.output_bytes;
2880 }
2881 ir += ha->base_qpair->counters.input_requests;
2882 or += ha->base_qpair->counters.output_requests;
2883 ib += ha->base_qpair->counters.input_bytes;
2884 ob += ha->base_qpair->counters.output_bytes;
2885
2886 ir += vha->qla_stats.input_requests;
2887 or += vha->qla_stats.output_requests;
2888 ib += vha->qla_stats.input_bytes;
2889 ob += vha->qla_stats.output_bytes;
2890 /* --- */
2891
2892 p->link_failure_count = le32_to_cpu(stats->link_fail_cnt);
2893 p->loss_of_sync_count = le32_to_cpu(stats->loss_sync_cnt);
2894 p->loss_of_signal_count = le32_to_cpu(stats->loss_sig_cnt);
2895 p->prim_seq_protocol_err_count = le32_to_cpu(stats->prim_seq_err_cnt);
2896 p->invalid_tx_word_count = le32_to_cpu(stats->inval_xmit_word_cnt);
2897 p->invalid_crc_count = le32_to_cpu(stats->inval_crc_cnt);
2898 if (IS_FWI2_CAPABLE(ha)) {
2899 p->lip_count = le32_to_cpu(stats->lip_cnt);
2900 p->tx_frames = le32_to_cpu(stats->tx_frames);
2901 p->rx_frames = le32_to_cpu(stats->rx_frames);
2902 p->dumped_frames = le32_to_cpu(stats->discarded_frames);
2903 p->nos_count = le32_to_cpu(stats->nos_rcvd);
2904 p->error_frames =
2905 le32_to_cpu(stats->dropped_frames) +
2906 le32_to_cpu(stats->discarded_frames);
2907 if (IS_QLA83XX(ha) || IS_QLA27XX(ha) || IS_QLA28XX(ha)) {
2908 p->rx_words = le64_to_cpu(stats->fpm_recv_word_cnt);
2909 p->tx_words = le64_to_cpu(stats->fpm_xmit_word_cnt);
2910 } else {
2911 p->rx_words = ib >> 2;
2912 p->tx_words = ob >> 2;
2913 }
2914 }
2915
2916 p->fcp_control_requests = vha->qla_stats.control_requests;
2917 p->fcp_input_requests = ir;
2918 p->fcp_output_requests = or;
2919 p->fcp_input_megabytes = ib >> 20;
2920 p->fcp_output_megabytes = ob >> 20;
2921 p->seconds_since_last_reset =
2922 get_jiffies_64() - vha->qla_stats.jiffies_at_last_reset;
2923 do_div(p->seconds_since_last_reset, HZ);
2924
2925 done_free:
2926 dma_free_coherent(&ha->pdev->dev, sizeof(struct link_statistics),
2927 stats, stats_dma);
2928 done:
2929 return p;
2930 }
2931
2932 static void
qla2x00_reset_host_stats(struct Scsi_Host * shost)2933 qla2x00_reset_host_stats(struct Scsi_Host *shost)
2934 {
2935 scsi_qla_host_t *vha = shost_priv(shost);
2936 struct qla_hw_data *ha = vha->hw;
2937 struct scsi_qla_host *base_vha = pci_get_drvdata(ha->pdev);
2938 struct link_statistics *stats;
2939 dma_addr_t stats_dma;
2940 int i;
2941 struct qla_qpair *qpair;
2942
2943 memset(&vha->qla_stats, 0, sizeof(vha->qla_stats));
2944 memset(&vha->fc_host_stat, 0, sizeof(vha->fc_host_stat));
2945 for (i = 0; i < vha->hw->max_qpairs; i++) {
2946 qpair = vha->hw->queue_pair_map[i];
2947 if (!qpair)
2948 continue;
2949 memset(&qpair->counters, 0, sizeof(qpair->counters));
2950 }
2951 memset(&ha->base_qpair->counters, 0, sizeof(qpair->counters));
2952
2953 vha->qla_stats.jiffies_at_last_reset = get_jiffies_64();
2954
2955 if (IS_FWI2_CAPABLE(ha)) {
2956 int rval;
2957
2958 stats = dma_alloc_coherent(&ha->pdev->dev,
2959 sizeof(*stats), &stats_dma, GFP_KERNEL);
2960 if (!stats) {
2961 ql_log(ql_log_warn, vha, 0x70d7,
2962 "Failed to allocate memory for stats.\n");
2963 return;
2964 }
2965
2966 /* reset firmware statistics */
2967 rval = qla24xx_get_isp_stats(base_vha, stats, stats_dma, BIT_0);
2968 if (rval != QLA_SUCCESS)
2969 ql_log(ql_log_warn, vha, 0x70de,
2970 "Resetting ISP statistics failed: rval = %d\n",
2971 rval);
2972
2973 dma_free_coherent(&ha->pdev->dev, sizeof(*stats),
2974 stats, stats_dma);
2975 }
2976 }
2977
2978 static void
qla2x00_get_host_symbolic_name(struct Scsi_Host * shost)2979 qla2x00_get_host_symbolic_name(struct Scsi_Host *shost)
2980 {
2981 scsi_qla_host_t *vha = shost_priv(shost);
2982
2983 qla2x00_get_sym_node_name(vha, fc_host_symbolic_name(shost),
2984 sizeof(fc_host_symbolic_name(shost)));
2985 }
2986
2987 static void
qla2x00_set_host_system_hostname(struct Scsi_Host * shost)2988 qla2x00_set_host_system_hostname(struct Scsi_Host *shost)
2989 {
2990 scsi_qla_host_t *vha = shost_priv(shost);
2991
2992 set_bit(REGISTER_FDMI_NEEDED, &vha->dpc_flags);
2993 }
2994
2995 static void
qla2x00_get_host_fabric_name(struct Scsi_Host * shost)2996 qla2x00_get_host_fabric_name(struct Scsi_Host *shost)
2997 {
2998 scsi_qla_host_t *vha = shost_priv(shost);
2999 static const uint8_t node_name[WWN_SIZE] = {
3000 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF
3001 };
3002 u64 fabric_name = wwn_to_u64(node_name);
3003
3004 if (vha->device_flags & SWITCH_FOUND)
3005 fabric_name = wwn_to_u64(vha->fabric_node_name);
3006
3007 fc_host_fabric_name(shost) = fabric_name;
3008 }
3009
3010 static void
qla2x00_get_host_port_state(struct Scsi_Host * shost)3011 qla2x00_get_host_port_state(struct Scsi_Host *shost)
3012 {
3013 scsi_qla_host_t *vha = shost_priv(shost);
3014 struct scsi_qla_host *base_vha = pci_get_drvdata(vha->hw->pdev);
3015
3016 if (!base_vha->flags.online) {
3017 fc_host_port_state(shost) = FC_PORTSTATE_OFFLINE;
3018 return;
3019 }
3020
3021 switch (atomic_read(&base_vha->loop_state)) {
3022 case LOOP_UPDATE:
3023 fc_host_port_state(shost) = FC_PORTSTATE_DIAGNOSTICS;
3024 break;
3025 case LOOP_DOWN:
3026 if (test_bit(LOOP_RESYNC_NEEDED, &base_vha->dpc_flags))
3027 fc_host_port_state(shost) = FC_PORTSTATE_DIAGNOSTICS;
3028 else
3029 fc_host_port_state(shost) = FC_PORTSTATE_LINKDOWN;
3030 break;
3031 case LOOP_DEAD:
3032 fc_host_port_state(shost) = FC_PORTSTATE_LINKDOWN;
3033 break;
3034 case LOOP_READY:
3035 fc_host_port_state(shost) = FC_PORTSTATE_ONLINE;
3036 break;
3037 default:
3038 fc_host_port_state(shost) = FC_PORTSTATE_UNKNOWN;
3039 break;
3040 }
3041 }
3042
3043 static int
qla24xx_vport_create(struct fc_vport * fc_vport,bool disable)3044 qla24xx_vport_create(struct fc_vport *fc_vport, bool disable)
3045 {
3046 int ret = 0;
3047 uint8_t qos = 0;
3048 scsi_qla_host_t *base_vha = shost_priv(fc_vport->shost);
3049 scsi_qla_host_t *vha = NULL;
3050 struct qla_hw_data *ha = base_vha->hw;
3051 int cnt;
3052 struct req_que *req = ha->req_q_map[0];
3053 struct qla_qpair *qpair;
3054
3055 ret = qla24xx_vport_create_req_sanity_check(fc_vport);
3056 if (ret) {
3057 ql_log(ql_log_warn, vha, 0x707e,
3058 "Vport sanity check failed, status %x\n", ret);
3059 return (ret);
3060 }
3061
3062 vha = qla24xx_create_vhost(fc_vport);
3063 if (vha == NULL) {
3064 ql_log(ql_log_warn, vha, 0x707f, "Vport create host failed.\n");
3065 return FC_VPORT_FAILED;
3066 }
3067 if (disable) {
3068 atomic_set(&vha->vp_state, VP_OFFLINE);
3069 fc_vport_set_state(fc_vport, FC_VPORT_DISABLED);
3070 } else
3071 atomic_set(&vha->vp_state, VP_FAILED);
3072
3073 /* ready to create vport */
3074 ql_log(ql_log_info, vha, 0x7080,
3075 "VP entry id %d assigned.\n", vha->vp_idx);
3076
3077 /* initialized vport states */
3078 atomic_set(&vha->loop_state, LOOP_DOWN);
3079 vha->vp_err_state = VP_ERR_PORTDWN;
3080 vha->vp_prev_err_state = VP_ERR_UNKWN;
3081 /* Check if physical ha port is Up */
3082 if (atomic_read(&base_vha->loop_state) == LOOP_DOWN ||
3083 atomic_read(&base_vha->loop_state) == LOOP_DEAD) {
3084 /* Don't retry or attempt login of this virtual port */
3085 ql_dbg(ql_dbg_user, vha, 0x7081,
3086 "Vport loop state is not UP.\n");
3087 atomic_set(&vha->loop_state, LOOP_DEAD);
3088 if (!disable)
3089 fc_vport_set_state(fc_vport, FC_VPORT_LINKDOWN);
3090 }
3091
3092 if (IS_T10_PI_CAPABLE(ha) && ql2xenabledif) {
3093 if (ha->fw_attributes & BIT_4) {
3094 int prot = 0, guard;
3095
3096 vha->flags.difdix_supported = 1;
3097 ql_dbg(ql_dbg_user, vha, 0x7082,
3098 "Registered for DIF/DIX type 1 and 3 protection.\n");
3099 scsi_host_set_prot(vha->host,
3100 prot | SHOST_DIF_TYPE1_PROTECTION
3101 | SHOST_DIF_TYPE2_PROTECTION
3102 | SHOST_DIF_TYPE3_PROTECTION
3103 | SHOST_DIX_TYPE1_PROTECTION
3104 | SHOST_DIX_TYPE2_PROTECTION
3105 | SHOST_DIX_TYPE3_PROTECTION);
3106
3107 guard = SHOST_DIX_GUARD_CRC;
3108
3109 if (IS_PI_IPGUARD_CAPABLE(ha) &&
3110 (ql2xenabledif > 1 || IS_PI_DIFB_DIX0_CAPABLE(ha)))
3111 guard |= SHOST_DIX_GUARD_IP;
3112
3113 scsi_host_set_guard(vha->host, guard);
3114 } else
3115 vha->flags.difdix_supported = 0;
3116 }
3117
3118 if (scsi_add_host_with_dma(vha->host, &fc_vport->dev,
3119 &ha->pdev->dev)) {
3120 ql_dbg(ql_dbg_user, vha, 0x7083,
3121 "scsi_add_host failure for VP[%d].\n", vha->vp_idx);
3122 goto vport_create_failed_2;
3123 }
3124
3125 /* initialize attributes */
3126 fc_host_dev_loss_tmo(vha->host) = ha->port_down_retry_count;
3127 fc_host_node_name(vha->host) = wwn_to_u64(vha->node_name);
3128 fc_host_port_name(vha->host) = wwn_to_u64(vha->port_name);
3129 fc_host_supported_classes(vha->host) =
3130 fc_host_supported_classes(base_vha->host);
3131 fc_host_supported_speeds(vha->host) =
3132 fc_host_supported_speeds(base_vha->host);
3133
3134 qlt_vport_create(vha, ha);
3135 qla24xx_vport_disable(fc_vport, disable);
3136
3137 if (!ql2xmqsupport || !ha->npiv_info)
3138 goto vport_queue;
3139
3140 /* Create a request queue in QoS mode for the vport */
3141 for (cnt = 0; cnt < ha->nvram_npiv_size; cnt++) {
3142 if (memcmp(ha->npiv_info[cnt].port_name, vha->port_name, 8) == 0
3143 && memcmp(ha->npiv_info[cnt].node_name, vha->node_name,
3144 8) == 0) {
3145 qos = ha->npiv_info[cnt].q_qos;
3146 break;
3147 }
3148 }
3149
3150 if (qos) {
3151 qpair = qla2xxx_create_qpair(vha, qos, vha->vp_idx, true);
3152 if (!qpair)
3153 ql_log(ql_log_warn, vha, 0x7084,
3154 "Can't create qpair for VP[%d]\n",
3155 vha->vp_idx);
3156 else {
3157 ql_dbg(ql_dbg_multiq, vha, 0xc001,
3158 "Queue pair: %d Qos: %d) created for VP[%d]\n",
3159 qpair->id, qos, vha->vp_idx);
3160 ql_dbg(ql_dbg_user, vha, 0x7085,
3161 "Queue Pair: %d Qos: %d) created for VP[%d]\n",
3162 qpair->id, qos, vha->vp_idx);
3163 req = qpair->req;
3164 vha->qpair = qpair;
3165 }
3166 }
3167
3168 vport_queue:
3169 vha->req = req;
3170 return 0;
3171
3172 vport_create_failed_2:
3173 qla24xx_disable_vp(vha);
3174 qla24xx_deallocate_vp_id(vha);
3175 scsi_host_put(vha->host);
3176 return FC_VPORT_FAILED;
3177 }
3178
3179 static int
qla24xx_vport_delete(struct fc_vport * fc_vport)3180 qla24xx_vport_delete(struct fc_vport *fc_vport)
3181 {
3182 scsi_qla_host_t *vha = fc_vport->dd_data;
3183 struct qla_hw_data *ha = vha->hw;
3184 uint16_t id = vha->vp_idx;
3185
3186 set_bit(VPORT_DELETE, &vha->dpc_flags);
3187
3188 while (test_bit(LOOP_RESYNC_ACTIVE, &vha->dpc_flags))
3189 msleep(1000);
3190
3191
3192 qla24xx_disable_vp(vha);
3193 qla2x00_wait_for_sess_deletion(vha);
3194
3195 qla_nvme_delete(vha);
3196 qla_enode_stop(vha);
3197 qla_edb_stop(vha);
3198
3199 vha->flags.delete_progress = 1;
3200
3201 qlt_remove_target(ha, vha);
3202
3203 fc_remove_host(vha->host);
3204
3205 scsi_remove_host(vha->host);
3206
3207 /* Allow timer to run to drain queued items, when removing vp */
3208 qla24xx_deallocate_vp_id(vha);
3209
3210 if (vha->timer_active) {
3211 qla2x00_vp_stop_timer(vha);
3212 ql_dbg(ql_dbg_user, vha, 0x7086,
3213 "Timer for the VP[%d] has stopped\n", vha->vp_idx);
3214 }
3215
3216 qla2x00_free_fcports(vha);
3217
3218 mutex_lock(&ha->vport_lock);
3219 ha->cur_vport_count--;
3220 clear_bit(vha->vp_idx, ha->vp_idx_map);
3221 mutex_unlock(&ha->vport_lock);
3222
3223 dma_free_coherent(&ha->pdev->dev, vha->gnl.size, vha->gnl.l,
3224 vha->gnl.ldma);
3225
3226 vha->gnl.l = NULL;
3227
3228 vfree(vha->scan.l);
3229
3230 if (vha->qpair && vha->qpair->vp_idx == vha->vp_idx) {
3231 if (qla2xxx_delete_qpair(vha, vha->qpair) != QLA_SUCCESS)
3232 ql_log(ql_log_warn, vha, 0x7087,
3233 "Queue Pair delete failed.\n");
3234 }
3235
3236 ql_log(ql_log_info, vha, 0x7088, "VP[%d] deleted.\n", id);
3237 scsi_host_put(vha->host);
3238 return 0;
3239 }
3240
3241 static int
qla24xx_vport_disable(struct fc_vport * fc_vport,bool disable)3242 qla24xx_vport_disable(struct fc_vport *fc_vport, bool disable)
3243 {
3244 scsi_qla_host_t *vha = fc_vport->dd_data;
3245
3246 if (disable)
3247 qla24xx_disable_vp(vha);
3248 else
3249 qla24xx_enable_vp(vha);
3250
3251 return 0;
3252 }
3253
3254 struct fc_function_template qla2xxx_transport_functions = {
3255
3256 .show_host_node_name = 1,
3257 .show_host_port_name = 1,
3258 .show_host_supported_classes = 1,
3259 .show_host_supported_speeds = 1,
3260
3261 .get_host_port_id = qla2x00_get_host_port_id,
3262 .show_host_port_id = 1,
3263 .get_host_speed = qla2x00_get_host_speed,
3264 .show_host_speed = 1,
3265 .get_host_port_type = qla2x00_get_host_port_type,
3266 .show_host_port_type = 1,
3267 .get_host_symbolic_name = qla2x00_get_host_symbolic_name,
3268 .show_host_symbolic_name = 1,
3269 .set_host_system_hostname = qla2x00_set_host_system_hostname,
3270 .show_host_system_hostname = 1,
3271 .get_host_fabric_name = qla2x00_get_host_fabric_name,
3272 .show_host_fabric_name = 1,
3273 .get_host_port_state = qla2x00_get_host_port_state,
3274 .show_host_port_state = 1,
3275
3276 .dd_fcrport_size = sizeof(struct fc_port *),
3277 .show_rport_supported_classes = 1,
3278
3279 .get_starget_node_name = qla2x00_get_starget_node_name,
3280 .show_starget_node_name = 1,
3281 .get_starget_port_name = qla2x00_get_starget_port_name,
3282 .show_starget_port_name = 1,
3283 .get_starget_port_id = qla2x00_get_starget_port_id,
3284 .show_starget_port_id = 1,
3285
3286 .set_rport_dev_loss_tmo = qla2x00_set_rport_loss_tmo,
3287 .show_rport_dev_loss_tmo = 1,
3288
3289 .issue_fc_host_lip = qla2x00_issue_lip,
3290 .dev_loss_tmo_callbk = qla2x00_dev_loss_tmo_callbk,
3291 .terminate_rport_io = qla2x00_terminate_rport_io,
3292 .get_fc_host_stats = qla2x00_get_fc_host_stats,
3293 .reset_fc_host_stats = qla2x00_reset_host_stats,
3294
3295 .vport_create = qla24xx_vport_create,
3296 .vport_disable = qla24xx_vport_disable,
3297 .vport_delete = qla24xx_vport_delete,
3298 .bsg_request = qla24xx_bsg_request,
3299 .bsg_timeout = qla24xx_bsg_timeout,
3300 };
3301
3302 struct fc_function_template qla2xxx_transport_vport_functions = {
3303
3304 .show_host_node_name = 1,
3305 .show_host_port_name = 1,
3306 .show_host_supported_classes = 1,
3307 .show_host_supported_speeds = 1,
3308
3309 .get_host_port_id = qla2x00_get_host_port_id,
3310 .show_host_port_id = 1,
3311 .get_host_speed = qla2x00_get_host_speed,
3312 .show_host_speed = 1,
3313 .get_host_port_type = qla2x00_get_host_port_type,
3314 .show_host_port_type = 1,
3315 .get_host_symbolic_name = qla2x00_get_host_symbolic_name,
3316 .show_host_symbolic_name = 1,
3317 .set_host_system_hostname = qla2x00_set_host_system_hostname,
3318 .show_host_system_hostname = 1,
3319 .get_host_fabric_name = qla2x00_get_host_fabric_name,
3320 .show_host_fabric_name = 1,
3321 .get_host_port_state = qla2x00_get_host_port_state,
3322 .show_host_port_state = 1,
3323
3324 .dd_fcrport_size = sizeof(struct fc_port *),
3325 .show_rport_supported_classes = 1,
3326
3327 .get_starget_node_name = qla2x00_get_starget_node_name,
3328 .show_starget_node_name = 1,
3329 .get_starget_port_name = qla2x00_get_starget_port_name,
3330 .show_starget_port_name = 1,
3331 .get_starget_port_id = qla2x00_get_starget_port_id,
3332 .show_starget_port_id = 1,
3333
3334 .set_rport_dev_loss_tmo = qla2x00_set_rport_loss_tmo,
3335 .show_rport_dev_loss_tmo = 1,
3336
3337 .issue_fc_host_lip = qla2x00_issue_lip,
3338 .dev_loss_tmo_callbk = qla2x00_dev_loss_tmo_callbk,
3339 .terminate_rport_io = qla2x00_terminate_rport_io,
3340 .get_fc_host_stats = qla2x00_get_fc_host_stats,
3341 .reset_fc_host_stats = qla2x00_reset_host_stats,
3342
3343 .bsg_request = qla24xx_bsg_request,
3344 .bsg_timeout = qla24xx_bsg_timeout,
3345 };
3346
3347 static uint
qla2x00_get_host_supported_speeds(scsi_qla_host_t * vha,uint speeds)3348 qla2x00_get_host_supported_speeds(scsi_qla_host_t *vha, uint speeds)
3349 {
3350 uint supported_speeds = FC_PORTSPEED_UNKNOWN;
3351
3352 if (speeds & FDMI_PORT_SPEED_64GB)
3353 supported_speeds |= FC_PORTSPEED_64GBIT;
3354 if (speeds & FDMI_PORT_SPEED_32GB)
3355 supported_speeds |= FC_PORTSPEED_32GBIT;
3356 if (speeds & FDMI_PORT_SPEED_16GB)
3357 supported_speeds |= FC_PORTSPEED_16GBIT;
3358 if (speeds & FDMI_PORT_SPEED_8GB)
3359 supported_speeds |= FC_PORTSPEED_8GBIT;
3360 if (speeds & FDMI_PORT_SPEED_4GB)
3361 supported_speeds |= FC_PORTSPEED_4GBIT;
3362 if (speeds & FDMI_PORT_SPEED_2GB)
3363 supported_speeds |= FC_PORTSPEED_2GBIT;
3364 if (speeds & FDMI_PORT_SPEED_1GB)
3365 supported_speeds |= FC_PORTSPEED_1GBIT;
3366
3367 return supported_speeds;
3368 }
3369
3370 void
qla2x00_init_host_attr(scsi_qla_host_t * vha)3371 qla2x00_init_host_attr(scsi_qla_host_t *vha)
3372 {
3373 struct qla_hw_data *ha = vha->hw;
3374 u32 speeds = 0, fdmi_speed = 0;
3375
3376 fc_host_dev_loss_tmo(vha->host) = ha->port_down_retry_count;
3377 fc_host_node_name(vha->host) = wwn_to_u64(vha->node_name);
3378 fc_host_port_name(vha->host) = wwn_to_u64(vha->port_name);
3379 fc_host_supported_classes(vha->host) = ha->base_qpair->enable_class_2 ?
3380 (FC_COS_CLASS2|FC_COS_CLASS3) : FC_COS_CLASS3;
3381 fc_host_max_npiv_vports(vha->host) = ha->max_npiv_vports;
3382 fc_host_npiv_vports_inuse(vha->host) = ha->cur_vport_count;
3383
3384 fdmi_speed = qla25xx_fdmi_port_speed_capability(ha);
3385 speeds = qla2x00_get_host_supported_speeds(vha, fdmi_speed);
3386
3387 fc_host_supported_speeds(vha->host) = speeds;
3388 }
3389