xref: /linux/drivers/scsi/qla4xxx/ql4_bsg.c (revision b77e0ce62d63a761ffb7f7245a215a49f5921c2f)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * QLogic iSCSI HBA Driver
4  * Copyright (c) 2011-2013 QLogic Corporation
5  */
6 
7 #include "ql4_def.h"
8 #include "ql4_glbl.h"
9 #include "ql4_bsg.h"
10 
11 static int
12 qla4xxx_read_flash(struct bsg_job *bsg_job)
13 {
14 	struct Scsi_Host *host = iscsi_job_to_shost(bsg_job);
15 	struct scsi_qla_host *ha = to_qla_host(host);
16 	struct iscsi_bsg_reply *bsg_reply = bsg_job->reply;
17 	struct iscsi_bsg_request *bsg_req = bsg_job->request;
18 	uint32_t offset = 0;
19 	uint32_t length = 0;
20 	dma_addr_t flash_dma;
21 	uint8_t *flash = NULL;
22 	int rval = -EINVAL;
23 
24 	bsg_reply->reply_payload_rcv_len = 0;
25 
26 	if (unlikely(pci_channel_offline(ha->pdev)))
27 		goto leave;
28 
29 	if (ql4xxx_reset_active(ha)) {
30 		ql4_printk(KERN_ERR, ha, "%s: reset active\n", __func__);
31 		rval = -EBUSY;
32 		goto leave;
33 	}
34 
35 	if (ha->flash_state != QLFLASH_WAITING) {
36 		ql4_printk(KERN_ERR, ha, "%s: another flash operation "
37 			   "active\n", __func__);
38 		rval = -EBUSY;
39 		goto leave;
40 	}
41 
42 	ha->flash_state = QLFLASH_READING;
43 	offset = bsg_req->rqst_data.h_vendor.vendor_cmd[1];
44 	length = bsg_job->reply_payload.payload_len;
45 
46 	flash = dma_alloc_coherent(&ha->pdev->dev, length, &flash_dma,
47 				   GFP_KERNEL);
48 	if (!flash) {
49 		ql4_printk(KERN_ERR, ha, "%s: dma alloc failed for flash "
50 			   "data\n", __func__);
51 		rval = -ENOMEM;
52 		goto leave;
53 	}
54 
55 	rval = qla4xxx_get_flash(ha, flash_dma, offset, length);
56 	if (rval) {
57 		ql4_printk(KERN_ERR, ha, "%s: get flash failed\n", __func__);
58 		bsg_reply->result = DID_ERROR << 16;
59 		rval = -EIO;
60 	} else {
61 		bsg_reply->reply_payload_rcv_len =
62 			sg_copy_from_buffer(bsg_job->reply_payload.sg_list,
63 					    bsg_job->reply_payload.sg_cnt,
64 					    flash, length);
65 		bsg_reply->result = DID_OK << 16;
66 	}
67 
68 	bsg_job_done(bsg_job, bsg_reply->result,
69 		     bsg_reply->reply_payload_rcv_len);
70 	dma_free_coherent(&ha->pdev->dev, length, flash, flash_dma);
71 leave:
72 	ha->flash_state = QLFLASH_WAITING;
73 	return rval;
74 }
75 
76 static int
77 qla4xxx_update_flash(struct bsg_job *bsg_job)
78 {
79 	struct Scsi_Host *host = iscsi_job_to_shost(bsg_job);
80 	struct scsi_qla_host *ha = to_qla_host(host);
81 	struct iscsi_bsg_reply *bsg_reply = bsg_job->reply;
82 	struct iscsi_bsg_request *bsg_req = bsg_job->request;
83 	uint32_t length = 0;
84 	uint32_t offset = 0;
85 	uint32_t options = 0;
86 	dma_addr_t flash_dma;
87 	uint8_t *flash = NULL;
88 	int rval = -EINVAL;
89 
90 	bsg_reply->reply_payload_rcv_len = 0;
91 
92 	if (unlikely(pci_channel_offline(ha->pdev)))
93 		goto leave;
94 
95 	if (ql4xxx_reset_active(ha)) {
96 		ql4_printk(KERN_ERR, ha, "%s: reset active\n", __func__);
97 		rval = -EBUSY;
98 		goto leave;
99 	}
100 
101 	if (ha->flash_state != QLFLASH_WAITING) {
102 		ql4_printk(KERN_ERR, ha, "%s: another flash operation "
103 			   "active\n", __func__);
104 		rval = -EBUSY;
105 		goto leave;
106 	}
107 
108 	ha->flash_state = QLFLASH_WRITING;
109 	length = bsg_job->request_payload.payload_len;
110 	offset = bsg_req->rqst_data.h_vendor.vendor_cmd[1];
111 	options = bsg_req->rqst_data.h_vendor.vendor_cmd[2];
112 
113 	flash = dma_alloc_coherent(&ha->pdev->dev, length, &flash_dma,
114 				   GFP_KERNEL);
115 	if (!flash) {
116 		ql4_printk(KERN_ERR, ha, "%s: dma alloc failed for flash "
117 			   "data\n", __func__);
118 		rval = -ENOMEM;
119 		goto leave;
120 	}
121 
122 	sg_copy_to_buffer(bsg_job->request_payload.sg_list,
123 			  bsg_job->request_payload.sg_cnt, flash, length);
124 
125 	rval = qla4xxx_set_flash(ha, flash_dma, offset, length, options);
126 	if (rval) {
127 		ql4_printk(KERN_ERR, ha, "%s: set flash failed\n", __func__);
128 		bsg_reply->result = DID_ERROR << 16;
129 		rval = -EIO;
130 	} else
131 		bsg_reply->result = DID_OK << 16;
132 
133 	bsg_job_done(bsg_job, bsg_reply->result,
134 		     bsg_reply->reply_payload_rcv_len);
135 	dma_free_coherent(&ha->pdev->dev, length, flash, flash_dma);
136 leave:
137 	ha->flash_state = QLFLASH_WAITING;
138 	return rval;
139 }
140 
141 static int
142 qla4xxx_get_acb_state(struct bsg_job *bsg_job)
143 {
144 	struct Scsi_Host *host = iscsi_job_to_shost(bsg_job);
145 	struct scsi_qla_host *ha = to_qla_host(host);
146 	struct iscsi_bsg_request *bsg_req = bsg_job->request;
147 	struct iscsi_bsg_reply *bsg_reply = bsg_job->reply;
148 	uint32_t status[MBOX_REG_COUNT];
149 	uint32_t acb_idx;
150 	uint32_t ip_idx;
151 	int rval = -EINVAL;
152 
153 	bsg_reply->reply_payload_rcv_len = 0;
154 
155 	if (unlikely(pci_channel_offline(ha->pdev)))
156 		goto leave;
157 
158 	/* Only 4022 and above adapters are supported */
159 	if (is_qla4010(ha))
160 		goto leave;
161 
162 	if (ql4xxx_reset_active(ha)) {
163 		ql4_printk(KERN_ERR, ha, "%s: reset active\n", __func__);
164 		rval = -EBUSY;
165 		goto leave;
166 	}
167 
168 	if (bsg_job->reply_payload.payload_len < sizeof(status)) {
169 		ql4_printk(KERN_ERR, ha, "%s: invalid payload len %d\n",
170 			   __func__, bsg_job->reply_payload.payload_len);
171 		rval = -EINVAL;
172 		goto leave;
173 	}
174 
175 	acb_idx = bsg_req->rqst_data.h_vendor.vendor_cmd[1];
176 	ip_idx = bsg_req->rqst_data.h_vendor.vendor_cmd[2];
177 
178 	rval = qla4xxx_get_ip_state(ha, acb_idx, ip_idx, status);
179 	if (rval) {
180 		ql4_printk(KERN_ERR, ha, "%s: get ip state failed\n",
181 			   __func__);
182 		bsg_reply->result = DID_ERROR << 16;
183 		rval = -EIO;
184 	} else {
185 		bsg_reply->reply_payload_rcv_len =
186 			sg_copy_from_buffer(bsg_job->reply_payload.sg_list,
187 					    bsg_job->reply_payload.sg_cnt,
188 					    status, sizeof(status));
189 		bsg_reply->result = DID_OK << 16;
190 	}
191 
192 	bsg_job_done(bsg_job, bsg_reply->result,
193 		     bsg_reply->reply_payload_rcv_len);
194 leave:
195 	return rval;
196 }
197 
198 static int
199 qla4xxx_read_nvram(struct bsg_job *bsg_job)
200 {
201 	struct Scsi_Host *host = iscsi_job_to_shost(bsg_job);
202 	struct scsi_qla_host *ha = to_qla_host(host);
203 	struct iscsi_bsg_request *bsg_req = bsg_job->request;
204 	struct iscsi_bsg_reply *bsg_reply = bsg_job->reply;
205 	uint32_t offset = 0;
206 	uint32_t len = 0;
207 	uint32_t total_len = 0;
208 	dma_addr_t nvram_dma;
209 	uint8_t *nvram = NULL;
210 	int rval = -EINVAL;
211 
212 	bsg_reply->reply_payload_rcv_len = 0;
213 
214 	if (unlikely(pci_channel_offline(ha->pdev)))
215 		goto leave;
216 
217 	/* Only 40xx adapters are supported */
218 	if (!(is_qla4010(ha) || is_qla4022(ha) || is_qla4032(ha)))
219 		goto leave;
220 
221 	if (ql4xxx_reset_active(ha)) {
222 		ql4_printk(KERN_ERR, ha, "%s: reset active\n", __func__);
223 		rval = -EBUSY;
224 		goto leave;
225 	}
226 
227 	offset = bsg_req->rqst_data.h_vendor.vendor_cmd[1];
228 	len = bsg_job->reply_payload.payload_len;
229 	total_len = offset + len;
230 
231 	/* total len should not be greater than max NVRAM size */
232 	if ((is_qla4010(ha) && total_len > QL4010_NVRAM_SIZE) ||
233 	    ((is_qla4022(ha) || is_qla4032(ha)) &&
234 	     total_len > QL40X2_NVRAM_SIZE)) {
235 		ql4_printk(KERN_ERR, ha, "%s: offset+len greater than max"
236 			   " nvram size, offset=%d len=%d\n",
237 			   __func__, offset, len);
238 		goto leave;
239 	}
240 
241 	nvram = dma_alloc_coherent(&ha->pdev->dev, len, &nvram_dma,
242 				   GFP_KERNEL);
243 	if (!nvram) {
244 		ql4_printk(KERN_ERR, ha, "%s: dma alloc failed for nvram "
245 			   "data\n", __func__);
246 		rval = -ENOMEM;
247 		goto leave;
248 	}
249 
250 	rval = qla4xxx_get_nvram(ha, nvram_dma, offset, len);
251 	if (rval) {
252 		ql4_printk(KERN_ERR, ha, "%s: get nvram failed\n", __func__);
253 		bsg_reply->result = DID_ERROR << 16;
254 		rval = -EIO;
255 	} else {
256 		bsg_reply->reply_payload_rcv_len =
257 			sg_copy_from_buffer(bsg_job->reply_payload.sg_list,
258 					    bsg_job->reply_payload.sg_cnt,
259 					    nvram, len);
260 		bsg_reply->result = DID_OK << 16;
261 	}
262 
263 	bsg_job_done(bsg_job, bsg_reply->result,
264 		     bsg_reply->reply_payload_rcv_len);
265 	dma_free_coherent(&ha->pdev->dev, len, nvram, nvram_dma);
266 leave:
267 	return rval;
268 }
269 
270 static int
271 qla4xxx_update_nvram(struct bsg_job *bsg_job)
272 {
273 	struct Scsi_Host *host = iscsi_job_to_shost(bsg_job);
274 	struct scsi_qla_host *ha = to_qla_host(host);
275 	struct iscsi_bsg_request *bsg_req = bsg_job->request;
276 	struct iscsi_bsg_reply *bsg_reply = bsg_job->reply;
277 	uint32_t offset = 0;
278 	uint32_t len = 0;
279 	uint32_t total_len = 0;
280 	dma_addr_t nvram_dma;
281 	uint8_t *nvram = NULL;
282 	int rval = -EINVAL;
283 
284 	bsg_reply->reply_payload_rcv_len = 0;
285 
286 	if (unlikely(pci_channel_offline(ha->pdev)))
287 		goto leave;
288 
289 	if (!(is_qla4010(ha) || is_qla4022(ha) || is_qla4032(ha)))
290 		goto leave;
291 
292 	if (ql4xxx_reset_active(ha)) {
293 		ql4_printk(KERN_ERR, ha, "%s: reset active\n", __func__);
294 		rval = -EBUSY;
295 		goto leave;
296 	}
297 
298 	offset = bsg_req->rqst_data.h_vendor.vendor_cmd[1];
299 	len = bsg_job->request_payload.payload_len;
300 	total_len = offset + len;
301 
302 	/* total len should not be greater than max NVRAM size */
303 	if ((is_qla4010(ha) && total_len > QL4010_NVRAM_SIZE) ||
304 	    ((is_qla4022(ha) || is_qla4032(ha)) &&
305 	     total_len > QL40X2_NVRAM_SIZE)) {
306 		ql4_printk(KERN_ERR, ha, "%s: offset+len greater than max"
307 			   " nvram size, offset=%d len=%d\n",
308 			   __func__, offset, len);
309 		goto leave;
310 	}
311 
312 	nvram = dma_alloc_coherent(&ha->pdev->dev, len, &nvram_dma,
313 				   GFP_KERNEL);
314 	if (!nvram) {
315 		ql4_printk(KERN_ERR, ha, "%s: dma alloc failed for flash "
316 			   "data\n", __func__);
317 		rval = -ENOMEM;
318 		goto leave;
319 	}
320 
321 	sg_copy_to_buffer(bsg_job->request_payload.sg_list,
322 			  bsg_job->request_payload.sg_cnt, nvram, len);
323 
324 	rval = qla4xxx_set_nvram(ha, nvram_dma, offset, len);
325 	if (rval) {
326 		ql4_printk(KERN_ERR, ha, "%s: set nvram failed\n", __func__);
327 		bsg_reply->result = DID_ERROR << 16;
328 		rval = -EIO;
329 	} else
330 		bsg_reply->result = DID_OK << 16;
331 
332 	bsg_job_done(bsg_job, bsg_reply->result,
333 		     bsg_reply->reply_payload_rcv_len);
334 	dma_free_coherent(&ha->pdev->dev, len, nvram, nvram_dma);
335 leave:
336 	return rval;
337 }
338 
339 static int
340 qla4xxx_restore_defaults(struct bsg_job *bsg_job)
341 {
342 	struct Scsi_Host *host = iscsi_job_to_shost(bsg_job);
343 	struct scsi_qla_host *ha = to_qla_host(host);
344 	struct iscsi_bsg_request *bsg_req = bsg_job->request;
345 	struct iscsi_bsg_reply *bsg_reply = bsg_job->reply;
346 	uint32_t region = 0;
347 	uint32_t field0 = 0;
348 	uint32_t field1 = 0;
349 	int rval = -EINVAL;
350 
351 	bsg_reply->reply_payload_rcv_len = 0;
352 
353 	if (unlikely(pci_channel_offline(ha->pdev)))
354 		goto leave;
355 
356 	if (is_qla4010(ha))
357 		goto leave;
358 
359 	if (ql4xxx_reset_active(ha)) {
360 		ql4_printk(KERN_ERR, ha, "%s: reset active\n", __func__);
361 		rval = -EBUSY;
362 		goto leave;
363 	}
364 
365 	region = bsg_req->rqst_data.h_vendor.vendor_cmd[1];
366 	field0 = bsg_req->rqst_data.h_vendor.vendor_cmd[2];
367 	field1 = bsg_req->rqst_data.h_vendor.vendor_cmd[3];
368 
369 	rval = qla4xxx_restore_factory_defaults(ha, region, field0, field1);
370 	if (rval) {
371 		ql4_printk(KERN_ERR, ha, "%s: set nvram failed\n", __func__);
372 		bsg_reply->result = DID_ERROR << 16;
373 		rval = -EIO;
374 	} else
375 		bsg_reply->result = DID_OK << 16;
376 
377 	bsg_job_done(bsg_job, bsg_reply->result,
378 		     bsg_reply->reply_payload_rcv_len);
379 leave:
380 	return rval;
381 }
382 
383 static int
384 qla4xxx_bsg_get_acb(struct bsg_job *bsg_job)
385 {
386 	struct Scsi_Host *host = iscsi_job_to_shost(bsg_job);
387 	struct scsi_qla_host *ha = to_qla_host(host);
388 	struct iscsi_bsg_request *bsg_req = bsg_job->request;
389 	struct iscsi_bsg_reply *bsg_reply = bsg_job->reply;
390 	uint32_t acb_type = 0;
391 	uint32_t len = 0;
392 	dma_addr_t acb_dma;
393 	uint8_t *acb = NULL;
394 	int rval = -EINVAL;
395 
396 	bsg_reply->reply_payload_rcv_len = 0;
397 
398 	if (unlikely(pci_channel_offline(ha->pdev)))
399 		goto leave;
400 
401 	/* Only 4022 and above adapters are supported */
402 	if (is_qla4010(ha))
403 		goto leave;
404 
405 	if (ql4xxx_reset_active(ha)) {
406 		ql4_printk(KERN_ERR, ha, "%s: reset active\n", __func__);
407 		rval = -EBUSY;
408 		goto leave;
409 	}
410 
411 	acb_type = bsg_req->rqst_data.h_vendor.vendor_cmd[1];
412 	len = bsg_job->reply_payload.payload_len;
413 	if (len < sizeof(struct addr_ctrl_blk)) {
414 		ql4_printk(KERN_ERR, ha, "%s: invalid acb len %d\n",
415 			   __func__, len);
416 		rval = -EINVAL;
417 		goto leave;
418 	}
419 
420 	acb = dma_alloc_coherent(&ha->pdev->dev, len, &acb_dma, GFP_KERNEL);
421 	if (!acb) {
422 		ql4_printk(KERN_ERR, ha, "%s: dma alloc failed for acb "
423 			   "data\n", __func__);
424 		rval = -ENOMEM;
425 		goto leave;
426 	}
427 
428 	rval = qla4xxx_get_acb(ha, acb_dma, acb_type, len);
429 	if (rval) {
430 		ql4_printk(KERN_ERR, ha, "%s: get acb failed\n", __func__);
431 		bsg_reply->result = DID_ERROR << 16;
432 		rval = -EIO;
433 	} else {
434 		bsg_reply->reply_payload_rcv_len =
435 			sg_copy_from_buffer(bsg_job->reply_payload.sg_list,
436 					    bsg_job->reply_payload.sg_cnt,
437 					    acb, len);
438 		bsg_reply->result = DID_OK << 16;
439 	}
440 
441 	bsg_job_done(bsg_job, bsg_reply->result,
442 		     bsg_reply->reply_payload_rcv_len);
443 	dma_free_coherent(&ha->pdev->dev, len, acb, acb_dma);
444 leave:
445 	return rval;
446 }
447 
448 static void ql4xxx_execute_diag_cmd(struct bsg_job *bsg_job)
449 {
450 	struct Scsi_Host *host = iscsi_job_to_shost(bsg_job);
451 	struct scsi_qla_host *ha = to_qla_host(host);
452 	struct iscsi_bsg_request *bsg_req = bsg_job->request;
453 	struct iscsi_bsg_reply *bsg_reply = bsg_job->reply;
454 	uint8_t *rsp_ptr = NULL;
455 	uint32_t mbox_cmd[MBOX_REG_COUNT];
456 	uint32_t mbox_sts[MBOX_REG_COUNT];
457 	int status = QLA_ERROR;
458 
459 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: in\n", __func__));
460 
461 	if (test_bit(DPC_RESET_HA, &ha->dpc_flags)) {
462 		ql4_printk(KERN_INFO, ha, "%s: Adapter reset in progress. Invalid Request\n",
463 			   __func__);
464 		bsg_reply->result = DID_ERROR << 16;
465 		goto exit_diag_mem_test;
466 	}
467 
468 	bsg_reply->reply_payload_rcv_len = 0;
469 	memcpy(mbox_cmd, &bsg_req->rqst_data.h_vendor.vendor_cmd[1],
470 	       sizeof(uint32_t) * MBOX_REG_COUNT);
471 
472 	DEBUG2(ql4_printk(KERN_INFO, ha,
473 			  "%s: mbox_cmd: %08X %08X %08X %08X %08X %08X %08X %08X\n",
474 			  __func__, mbox_cmd[0], mbox_cmd[1], mbox_cmd[2],
475 			  mbox_cmd[3], mbox_cmd[4], mbox_cmd[5], mbox_cmd[6],
476 			  mbox_cmd[7]));
477 
478 	status = qla4xxx_mailbox_command(ha, MBOX_REG_COUNT, 8, &mbox_cmd[0],
479 					 &mbox_sts[0]);
480 
481 	DEBUG2(ql4_printk(KERN_INFO, ha,
482 			  "%s: mbox_sts: %08X %08X %08X %08X %08X %08X %08X %08X\n",
483 			  __func__, mbox_sts[0], mbox_sts[1], mbox_sts[2],
484 			  mbox_sts[3], mbox_sts[4], mbox_sts[5], mbox_sts[6],
485 			  mbox_sts[7]));
486 
487 	if (status == QLA_SUCCESS)
488 		bsg_reply->result = DID_OK << 16;
489 	else
490 		bsg_reply->result = DID_ERROR << 16;
491 
492 	/* Send mbox_sts to application */
493 	bsg_job->reply_len = sizeof(struct iscsi_bsg_reply) + sizeof(mbox_sts);
494 	rsp_ptr = ((uint8_t *)bsg_reply) + sizeof(struct iscsi_bsg_reply);
495 	memcpy(rsp_ptr, mbox_sts, sizeof(mbox_sts));
496 
497 exit_diag_mem_test:
498 	DEBUG2(ql4_printk(KERN_INFO, ha,
499 			  "%s: bsg_reply->result = x%x, status = %s\n",
500 			  __func__, bsg_reply->result, STATUS(status)));
501 
502 	bsg_job_done(bsg_job, bsg_reply->result,
503 		     bsg_reply->reply_payload_rcv_len);
504 }
505 
506 static int qla4_83xx_wait_for_loopback_config_comp(struct scsi_qla_host *ha,
507 						   int wait_for_link)
508 {
509 	int status = QLA_SUCCESS;
510 
511 	if (!wait_for_completion_timeout(&ha->idc_comp, (IDC_COMP_TOV * HZ))) {
512 		ql4_printk(KERN_INFO, ha, "%s: IDC Complete notification not received, Waiting for another %d timeout",
513 			   __func__, ha->idc_extend_tmo);
514 		if (ha->idc_extend_tmo) {
515 			if (!wait_for_completion_timeout(&ha->idc_comp,
516 						(ha->idc_extend_tmo * HZ))) {
517 				ha->notify_idc_comp = 0;
518 				ha->notify_link_up_comp = 0;
519 				ql4_printk(KERN_WARNING, ha, "%s: Aborting: IDC Complete notification not received",
520 					   __func__);
521 				status = QLA_ERROR;
522 				goto exit_wait;
523 			} else {
524 				DEBUG2(ql4_printk(KERN_INFO, ha,
525 						  "%s: IDC Complete notification received\n",
526 						  __func__));
527 			}
528 		}
529 	} else {
530 		DEBUG2(ql4_printk(KERN_INFO, ha,
531 				  "%s: IDC Complete notification received\n",
532 				  __func__));
533 	}
534 	ha->notify_idc_comp = 0;
535 
536 	if (wait_for_link) {
537 		if (!wait_for_completion_timeout(&ha->link_up_comp,
538 						 (IDC_COMP_TOV * HZ))) {
539 			ha->notify_link_up_comp = 0;
540 			ql4_printk(KERN_WARNING, ha, "%s: Aborting: LINK UP notification not received",
541 				   __func__);
542 			status = QLA_ERROR;
543 			goto exit_wait;
544 		} else {
545 			DEBUG2(ql4_printk(KERN_INFO, ha,
546 					  "%s: LINK UP notification received\n",
547 					  __func__));
548 		}
549 		ha->notify_link_up_comp = 0;
550 	}
551 
552 exit_wait:
553 	return status;
554 }
555 
556 static int qla4_83xx_pre_loopback_config(struct scsi_qla_host *ha,
557 					 uint32_t *mbox_cmd)
558 {
559 	uint32_t config = 0;
560 	int status = QLA_SUCCESS;
561 
562 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: in\n", __func__));
563 
564 	status = qla4_83xx_get_port_config(ha, &config);
565 	if (status != QLA_SUCCESS)
566 		goto exit_pre_loopback_config;
567 
568 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: Default port config=%08X\n",
569 			  __func__, config));
570 
571 	if ((config & ENABLE_INTERNAL_LOOPBACK) ||
572 	    (config & ENABLE_EXTERNAL_LOOPBACK)) {
573 		ql4_printk(KERN_INFO, ha, "%s: Loopback diagnostics already in progress. Invalid request\n",
574 			   __func__);
575 		goto exit_pre_loopback_config;
576 	}
577 
578 	if (mbox_cmd[1] == QL_DIAG_CMD_TEST_INT_LOOPBACK)
579 		config |= ENABLE_INTERNAL_LOOPBACK;
580 
581 	if (mbox_cmd[1] == QL_DIAG_CMD_TEST_EXT_LOOPBACK)
582 		config |= ENABLE_EXTERNAL_LOOPBACK;
583 
584 	config &= ~ENABLE_DCBX;
585 
586 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: New port config=%08X\n",
587 			  __func__, config));
588 
589 	ha->notify_idc_comp = 1;
590 	ha->notify_link_up_comp = 1;
591 
592 	/* get the link state */
593 	qla4xxx_get_firmware_state(ha);
594 
595 	status = qla4_83xx_set_port_config(ha, &config);
596 	if (status != QLA_SUCCESS) {
597 		ha->notify_idc_comp = 0;
598 		ha->notify_link_up_comp = 0;
599 		goto exit_pre_loopback_config;
600 	}
601 exit_pre_loopback_config:
602 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: status = %s\n", __func__,
603 			  STATUS(status)));
604 	return status;
605 }
606 
607 static int qla4_83xx_post_loopback_config(struct scsi_qla_host *ha,
608 					  uint32_t *mbox_cmd)
609 {
610 	int status = QLA_SUCCESS;
611 	uint32_t config = 0;
612 
613 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: in\n", __func__));
614 
615 	status = qla4_83xx_get_port_config(ha, &config);
616 	if (status != QLA_SUCCESS)
617 		goto exit_post_loopback_config;
618 
619 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: port config=%08X\n", __func__,
620 			  config));
621 
622 	if (mbox_cmd[1] == QL_DIAG_CMD_TEST_INT_LOOPBACK)
623 		config &= ~ENABLE_INTERNAL_LOOPBACK;
624 	else if (mbox_cmd[1] == QL_DIAG_CMD_TEST_EXT_LOOPBACK)
625 		config &= ~ENABLE_EXTERNAL_LOOPBACK;
626 
627 	config |= ENABLE_DCBX;
628 
629 	DEBUG2(ql4_printk(KERN_INFO, ha,
630 			  "%s: Restore default port config=%08X\n", __func__,
631 			  config));
632 
633 	ha->notify_idc_comp = 1;
634 	if (ha->addl_fw_state & FW_ADDSTATE_LINK_UP)
635 		ha->notify_link_up_comp = 1;
636 
637 	status = qla4_83xx_set_port_config(ha, &config);
638 	if (status != QLA_SUCCESS) {
639 		ql4_printk(KERN_INFO, ha, "%s: Scheduling adapter reset\n",
640 			   __func__);
641 		set_bit(DPC_RESET_HA, &ha->dpc_flags);
642 		clear_bit(AF_LOOPBACK, &ha->flags);
643 		goto exit_post_loopback_config;
644 	}
645 
646 exit_post_loopback_config:
647 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: status = %s\n", __func__,
648 			  STATUS(status)));
649 	return status;
650 }
651 
652 static void qla4xxx_execute_diag_loopback_cmd(struct bsg_job *bsg_job)
653 {
654 	struct Scsi_Host *host = iscsi_job_to_shost(bsg_job);
655 	struct scsi_qla_host *ha = to_qla_host(host);
656 	struct iscsi_bsg_request *bsg_req = bsg_job->request;
657 	struct iscsi_bsg_reply *bsg_reply = bsg_job->reply;
658 	uint8_t *rsp_ptr = NULL;
659 	uint32_t mbox_cmd[MBOX_REG_COUNT];
660 	uint32_t mbox_sts[MBOX_REG_COUNT];
661 	int wait_for_link = 1;
662 	int status = QLA_ERROR;
663 
664 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: in\n", __func__));
665 
666 	bsg_reply->reply_payload_rcv_len = 0;
667 
668 	if (test_bit(AF_LOOPBACK, &ha->flags)) {
669 		ql4_printk(KERN_INFO, ha, "%s: Loopback Diagnostics already in progress. Invalid Request\n",
670 			   __func__);
671 		bsg_reply->result = DID_ERROR << 16;
672 		goto exit_loopback_cmd;
673 	}
674 
675 	if (test_bit(DPC_RESET_HA, &ha->dpc_flags)) {
676 		ql4_printk(KERN_INFO, ha, "%s: Adapter reset in progress. Invalid Request\n",
677 			   __func__);
678 		bsg_reply->result = DID_ERROR << 16;
679 		goto exit_loopback_cmd;
680 	}
681 
682 	memcpy(mbox_cmd, &bsg_req->rqst_data.h_vendor.vendor_cmd[1],
683 	       sizeof(uint32_t) * MBOX_REG_COUNT);
684 
685 	if (is_qla8032(ha) || is_qla8042(ha)) {
686 		status = qla4_83xx_pre_loopback_config(ha, mbox_cmd);
687 		if (status != QLA_SUCCESS) {
688 			bsg_reply->result = DID_ERROR << 16;
689 			goto exit_loopback_cmd;
690 		}
691 
692 		status = qla4_83xx_wait_for_loopback_config_comp(ha,
693 								 wait_for_link);
694 		if (status != QLA_SUCCESS) {
695 			bsg_reply->result = DID_TIME_OUT << 16;
696 			goto restore;
697 		}
698 	}
699 
700 	DEBUG2(ql4_printk(KERN_INFO, ha,
701 			  "%s: mbox_cmd: %08X %08X %08X %08X %08X %08X %08X %08X\n",
702 			  __func__, mbox_cmd[0], mbox_cmd[1], mbox_cmd[2],
703 			  mbox_cmd[3], mbox_cmd[4], mbox_cmd[5], mbox_cmd[6],
704 			  mbox_cmd[7]));
705 
706 	status = qla4xxx_mailbox_command(ha, MBOX_REG_COUNT, 8, &mbox_cmd[0],
707 				&mbox_sts[0]);
708 
709 	if (status == QLA_SUCCESS)
710 		bsg_reply->result = DID_OK << 16;
711 	else
712 		bsg_reply->result = DID_ERROR << 16;
713 
714 	DEBUG2(ql4_printk(KERN_INFO, ha,
715 			  "%s: mbox_sts: %08X %08X %08X %08X %08X %08X %08X %08X\n",
716 			  __func__, mbox_sts[0], mbox_sts[1], mbox_sts[2],
717 			  mbox_sts[3], mbox_sts[4], mbox_sts[5], mbox_sts[6],
718 			  mbox_sts[7]));
719 
720 	/* Send mbox_sts to application */
721 	bsg_job->reply_len = sizeof(struct iscsi_bsg_reply) + sizeof(mbox_sts);
722 	rsp_ptr = ((uint8_t *)bsg_reply) + sizeof(struct iscsi_bsg_reply);
723 	memcpy(rsp_ptr, mbox_sts, sizeof(mbox_sts));
724 restore:
725 	if (is_qla8032(ha) || is_qla8042(ha)) {
726 		status = qla4_83xx_post_loopback_config(ha, mbox_cmd);
727 		if (status != QLA_SUCCESS) {
728 			bsg_reply->result = DID_ERROR << 16;
729 			goto exit_loopback_cmd;
730 		}
731 
732 		/* for pre_loopback_config() wait for LINK UP only
733 		 * if PHY LINK is UP */
734 		if (!(ha->addl_fw_state & FW_ADDSTATE_LINK_UP))
735 			wait_for_link = 0;
736 
737 		status = qla4_83xx_wait_for_loopback_config_comp(ha,
738 								 wait_for_link);
739 		if (status != QLA_SUCCESS) {
740 			bsg_reply->result = DID_TIME_OUT << 16;
741 			goto exit_loopback_cmd;
742 		}
743 	}
744 exit_loopback_cmd:
745 	DEBUG2(ql4_printk(KERN_INFO, ha,
746 			  "%s: bsg_reply->result = x%x, status = %s\n",
747 			  __func__, bsg_reply->result, STATUS(status)));
748 	bsg_job_done(bsg_job, bsg_reply->result,
749 		     bsg_reply->reply_payload_rcv_len);
750 }
751 
752 static int qla4xxx_execute_diag_test(struct bsg_job *bsg_job)
753 {
754 	struct Scsi_Host *host = iscsi_job_to_shost(bsg_job);
755 	struct scsi_qla_host *ha = to_qla_host(host);
756 	struct iscsi_bsg_request *bsg_req = bsg_job->request;
757 	uint32_t diag_cmd;
758 	int rval = -EINVAL;
759 
760 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: in\n", __func__));
761 
762 	diag_cmd = bsg_req->rqst_data.h_vendor.vendor_cmd[1];
763 	if (diag_cmd == MBOX_CMD_DIAG_TEST) {
764 		switch (bsg_req->rqst_data.h_vendor.vendor_cmd[2]) {
765 		case QL_DIAG_CMD_TEST_DDR_SIZE:
766 		case QL_DIAG_CMD_TEST_DDR_RW:
767 		case QL_DIAG_CMD_TEST_ONCHIP_MEM_RW:
768 		case QL_DIAG_CMD_TEST_NVRAM:
769 		case QL_DIAG_CMD_TEST_FLASH_ROM:
770 		case QL_DIAG_CMD_TEST_DMA_XFER:
771 		case QL_DIAG_CMD_SELF_DDR_RW:
772 		case QL_DIAG_CMD_SELF_ONCHIP_MEM_RW:
773 			/* Execute diag test for adapter RAM/FLASH */
774 			ql4xxx_execute_diag_cmd(bsg_job);
775 			/* Always return success as we want to sent bsg_reply
776 			 * to Application */
777 			rval = QLA_SUCCESS;
778 			break;
779 
780 		case QL_DIAG_CMD_TEST_INT_LOOPBACK:
781 		case QL_DIAG_CMD_TEST_EXT_LOOPBACK:
782 			/* Execute diag test for Network */
783 			qla4xxx_execute_diag_loopback_cmd(bsg_job);
784 			/* Always return success as we want to sent bsg_reply
785 			 * to Application */
786 			rval = QLA_SUCCESS;
787 			break;
788 		default:
789 			ql4_printk(KERN_ERR, ha, "%s: Invalid diag test: 0x%x\n",
790 				   __func__,
791 				   bsg_req->rqst_data.h_vendor.vendor_cmd[2]);
792 		}
793 	} else if ((diag_cmd == MBOX_CMD_SET_LED_CONFIG) ||
794 		   (diag_cmd == MBOX_CMD_GET_LED_CONFIG)) {
795 		ql4xxx_execute_diag_cmd(bsg_job);
796 		rval = QLA_SUCCESS;
797 	} else {
798 		ql4_printk(KERN_ERR, ha, "%s: Invalid diag cmd: 0x%x\n",
799 			   __func__, diag_cmd);
800 	}
801 
802 	return rval;
803 }
804 
805 /**
806  * qla4xxx_process_vendor_specific - handle vendor specific bsg request
807  * @bsg_job: iscsi_bsg_job to handle
808  **/
809 int qla4xxx_process_vendor_specific(struct bsg_job *bsg_job)
810 {
811 	struct iscsi_bsg_reply *bsg_reply = bsg_job->reply;
812 	struct iscsi_bsg_request *bsg_req = bsg_job->request;
813 	struct Scsi_Host *host = iscsi_job_to_shost(bsg_job);
814 	struct scsi_qla_host *ha = to_qla_host(host);
815 
816 	switch (bsg_req->rqst_data.h_vendor.vendor_cmd[0]) {
817 	case QLISCSI_VND_READ_FLASH:
818 		return qla4xxx_read_flash(bsg_job);
819 
820 	case QLISCSI_VND_UPDATE_FLASH:
821 		return qla4xxx_update_flash(bsg_job);
822 
823 	case QLISCSI_VND_GET_ACB_STATE:
824 		return qla4xxx_get_acb_state(bsg_job);
825 
826 	case QLISCSI_VND_READ_NVRAM:
827 		return qla4xxx_read_nvram(bsg_job);
828 
829 	case QLISCSI_VND_UPDATE_NVRAM:
830 		return qla4xxx_update_nvram(bsg_job);
831 
832 	case QLISCSI_VND_RESTORE_DEFAULTS:
833 		return qla4xxx_restore_defaults(bsg_job);
834 
835 	case QLISCSI_VND_GET_ACB:
836 		return qla4xxx_bsg_get_acb(bsg_job);
837 
838 	case QLISCSI_VND_DIAG_TEST:
839 		return qla4xxx_execute_diag_test(bsg_job);
840 
841 	default:
842 		ql4_printk(KERN_ERR, ha, "%s: invalid BSG vendor command: "
843 			   "0x%x\n", __func__, bsg_req->msgcode);
844 		bsg_reply->result = (DID_ERROR << 16);
845 		bsg_reply->reply_payload_rcv_len = 0;
846 		bsg_job_done(bsg_job, bsg_reply->result,
847 			     bsg_reply->reply_payload_rcv_len);
848 		return -ENOSYS;
849 	}
850 }
851 
852 /**
853  * qla4xxx_bsg_request - handle bsg request from ISCSI transport
854  * @bsg_job: iscsi_bsg_job to handle
855  */
856 int qla4xxx_bsg_request(struct bsg_job *bsg_job)
857 {
858 	struct iscsi_bsg_request *bsg_req = bsg_job->request;
859 	struct Scsi_Host *host = iscsi_job_to_shost(bsg_job);
860 	struct scsi_qla_host *ha = to_qla_host(host);
861 
862 	switch (bsg_req->msgcode) {
863 	case ISCSI_BSG_HST_VENDOR:
864 		return qla4xxx_process_vendor_specific(bsg_job);
865 
866 	default:
867 		ql4_printk(KERN_ERR, ha, "%s: invalid BSG command: 0x%x\n",
868 			   __func__, bsg_req->msgcode);
869 	}
870 
871 	return -ENOSYS;
872 }
873