xref: /linux/drivers/scsi/qla2xxx/qla_bsg.c (revision ca55b2fef3a9373fcfc30f82fd26bc7fccbda732)
1 /*
2  * QLogic Fibre Channel HBA Driver
3  * Copyright (c)  2003-2014 QLogic Corporation
4  *
5  * See LICENSE.qla2xxx for copyright and licensing details.
6  */
7 #include "qla_def.h"
8 
9 #include <linux/kthread.h>
10 #include <linux/vmalloc.h>
11 #include <linux/delay.h>
12 
13 /* BSG support for ELS/CT pass through */
14 void
15 qla2x00_bsg_job_done(void *data, void *ptr, int res)
16 {
17 	srb_t *sp = (srb_t *)ptr;
18 	struct scsi_qla_host *vha = (scsi_qla_host_t *)data;
19 	struct fc_bsg_job *bsg_job = sp->u.bsg_job;
20 
21 	bsg_job->reply->result = res;
22 	bsg_job->job_done(bsg_job);
23 	sp->free(vha, sp);
24 }
25 
26 void
27 qla2x00_bsg_sp_free(void *data, void *ptr)
28 {
29 	srb_t *sp = (srb_t *)ptr;
30 	struct scsi_qla_host *vha = sp->fcport->vha;
31 	struct fc_bsg_job *bsg_job = sp->u.bsg_job;
32 	struct qla_hw_data *ha = vha->hw;
33 	struct qla_mt_iocb_rqst_fx00 *piocb_rqst;
34 
35 	if (sp->type == SRB_FXIOCB_BCMD) {
36 		piocb_rqst = (struct qla_mt_iocb_rqst_fx00 *)
37 		    &bsg_job->request->rqst_data.h_vendor.vendor_cmd[1];
38 
39 		if (piocb_rqst->flags & SRB_FXDISC_REQ_DMA_VALID)
40 			dma_unmap_sg(&ha->pdev->dev,
41 			    bsg_job->request_payload.sg_list,
42 			    bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
43 
44 		if (piocb_rqst->flags & SRB_FXDISC_RESP_DMA_VALID)
45 			dma_unmap_sg(&ha->pdev->dev,
46 			    bsg_job->reply_payload.sg_list,
47 			    bsg_job->reply_payload.sg_cnt, DMA_FROM_DEVICE);
48 	} else {
49 		dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list,
50 		    bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
51 
52 		dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list,
53 		    bsg_job->reply_payload.sg_cnt, DMA_FROM_DEVICE);
54 	}
55 
56 	if (sp->type == SRB_CT_CMD ||
57 	    sp->type == SRB_FXIOCB_BCMD ||
58 	    sp->type == SRB_ELS_CMD_HST)
59 		kfree(sp->fcport);
60 	qla2x00_rel_sp(vha, sp);
61 }
62 
63 int
64 qla24xx_fcp_prio_cfg_valid(scsi_qla_host_t *vha,
65 	struct qla_fcp_prio_cfg *pri_cfg, uint8_t flag)
66 {
67 	int i, ret, num_valid;
68 	uint8_t *bcode;
69 	struct qla_fcp_prio_entry *pri_entry;
70 	uint32_t *bcode_val_ptr, bcode_val;
71 
72 	ret = 1;
73 	num_valid = 0;
74 	bcode = (uint8_t *)pri_cfg;
75 	bcode_val_ptr = (uint32_t *)pri_cfg;
76 	bcode_val = (uint32_t)(*bcode_val_ptr);
77 
78 	if (bcode_val == 0xFFFFFFFF) {
79 		/* No FCP Priority config data in flash */
80 		ql_dbg(ql_dbg_user, vha, 0x7051,
81 		    "No FCP Priority config data.\n");
82 		return 0;
83 	}
84 
85 	if (bcode[0] != 'H' || bcode[1] != 'Q' || bcode[2] != 'O' ||
86 			bcode[3] != 'S') {
87 		/* Invalid FCP priority data header*/
88 		ql_dbg(ql_dbg_user, vha, 0x7052,
89 		    "Invalid FCP Priority data header. bcode=0x%x.\n",
90 		    bcode_val);
91 		return 0;
92 	}
93 	if (flag != 1)
94 		return ret;
95 
96 	pri_entry = &pri_cfg->entry[0];
97 	for (i = 0; i < pri_cfg->num_entries; i++) {
98 		if (pri_entry->flags & FCP_PRIO_ENTRY_TAG_VALID)
99 			num_valid++;
100 		pri_entry++;
101 	}
102 
103 	if (num_valid == 0) {
104 		/* No valid FCP priority data entries */
105 		ql_dbg(ql_dbg_user, vha, 0x7053,
106 		    "No valid FCP Priority data entries.\n");
107 		ret = 0;
108 	} else {
109 		/* FCP priority data is valid */
110 		ql_dbg(ql_dbg_user, vha, 0x7054,
111 		    "Valid FCP priority data. num entries = %d.\n",
112 		    num_valid);
113 	}
114 
115 	return ret;
116 }
117 
118 static int
119 qla24xx_proc_fcp_prio_cfg_cmd(struct fc_bsg_job *bsg_job)
120 {
121 	struct Scsi_Host *host = bsg_job->shost;
122 	scsi_qla_host_t *vha = shost_priv(host);
123 	struct qla_hw_data *ha = vha->hw;
124 	int ret = 0;
125 	uint32_t len;
126 	uint32_t oper;
127 
128 	if (!(IS_QLA24XX_TYPE(ha) || IS_QLA25XX(ha) || IS_P3P_TYPE(ha))) {
129 		ret = -EINVAL;
130 		goto exit_fcp_prio_cfg;
131 	}
132 
133 	/* Get the sub command */
134 	oper = bsg_job->request->rqst_data.h_vendor.vendor_cmd[1];
135 
136 	/* Only set config is allowed if config memory is not allocated */
137 	if (!ha->fcp_prio_cfg && (oper != QLFC_FCP_PRIO_SET_CONFIG)) {
138 		ret = -EINVAL;
139 		goto exit_fcp_prio_cfg;
140 	}
141 	switch (oper) {
142 	case QLFC_FCP_PRIO_DISABLE:
143 		if (ha->flags.fcp_prio_enabled) {
144 			ha->flags.fcp_prio_enabled = 0;
145 			ha->fcp_prio_cfg->attributes &=
146 				~FCP_PRIO_ATTR_ENABLE;
147 			qla24xx_update_all_fcp_prio(vha);
148 			bsg_job->reply->result = DID_OK;
149 		} else {
150 			ret = -EINVAL;
151 			bsg_job->reply->result = (DID_ERROR << 16);
152 			goto exit_fcp_prio_cfg;
153 		}
154 		break;
155 
156 	case QLFC_FCP_PRIO_ENABLE:
157 		if (!ha->flags.fcp_prio_enabled) {
158 			if (ha->fcp_prio_cfg) {
159 				ha->flags.fcp_prio_enabled = 1;
160 				ha->fcp_prio_cfg->attributes |=
161 				    FCP_PRIO_ATTR_ENABLE;
162 				qla24xx_update_all_fcp_prio(vha);
163 				bsg_job->reply->result = DID_OK;
164 			} else {
165 				ret = -EINVAL;
166 				bsg_job->reply->result = (DID_ERROR << 16);
167 				goto exit_fcp_prio_cfg;
168 			}
169 		}
170 		break;
171 
172 	case QLFC_FCP_PRIO_GET_CONFIG:
173 		len = bsg_job->reply_payload.payload_len;
174 		if (!len || len > FCP_PRIO_CFG_SIZE) {
175 			ret = -EINVAL;
176 			bsg_job->reply->result = (DID_ERROR << 16);
177 			goto exit_fcp_prio_cfg;
178 		}
179 
180 		bsg_job->reply->result = DID_OK;
181 		bsg_job->reply->reply_payload_rcv_len =
182 			sg_copy_from_buffer(
183 			bsg_job->reply_payload.sg_list,
184 			bsg_job->reply_payload.sg_cnt, ha->fcp_prio_cfg,
185 			len);
186 
187 		break;
188 
189 	case QLFC_FCP_PRIO_SET_CONFIG:
190 		len = bsg_job->request_payload.payload_len;
191 		if (!len || len > FCP_PRIO_CFG_SIZE) {
192 			bsg_job->reply->result = (DID_ERROR << 16);
193 			ret = -EINVAL;
194 			goto exit_fcp_prio_cfg;
195 		}
196 
197 		if (!ha->fcp_prio_cfg) {
198 			ha->fcp_prio_cfg = vmalloc(FCP_PRIO_CFG_SIZE);
199 			if (!ha->fcp_prio_cfg) {
200 				ql_log(ql_log_warn, vha, 0x7050,
201 				    "Unable to allocate memory for fcp prio "
202 				    "config data (%x).\n", FCP_PRIO_CFG_SIZE);
203 				bsg_job->reply->result = (DID_ERROR << 16);
204 				ret = -ENOMEM;
205 				goto exit_fcp_prio_cfg;
206 			}
207 		}
208 
209 		memset(ha->fcp_prio_cfg, 0, FCP_PRIO_CFG_SIZE);
210 		sg_copy_to_buffer(bsg_job->request_payload.sg_list,
211 		bsg_job->request_payload.sg_cnt, ha->fcp_prio_cfg,
212 			FCP_PRIO_CFG_SIZE);
213 
214 		/* validate fcp priority data */
215 
216 		if (!qla24xx_fcp_prio_cfg_valid(vha,
217 		    (struct qla_fcp_prio_cfg *) ha->fcp_prio_cfg, 1)) {
218 			bsg_job->reply->result = (DID_ERROR << 16);
219 			ret = -EINVAL;
220 			/* If buffer was invalidatic int
221 			 * fcp_prio_cfg is of no use
222 			 */
223 			vfree(ha->fcp_prio_cfg);
224 			ha->fcp_prio_cfg = NULL;
225 			goto exit_fcp_prio_cfg;
226 		}
227 
228 		ha->flags.fcp_prio_enabled = 0;
229 		if (ha->fcp_prio_cfg->attributes & FCP_PRIO_ATTR_ENABLE)
230 			ha->flags.fcp_prio_enabled = 1;
231 		qla24xx_update_all_fcp_prio(vha);
232 		bsg_job->reply->result = DID_OK;
233 		break;
234 	default:
235 		ret = -EINVAL;
236 		break;
237 	}
238 exit_fcp_prio_cfg:
239 	if (!ret)
240 		bsg_job->job_done(bsg_job);
241 	return ret;
242 }
243 
244 static int
245 qla2x00_process_els(struct fc_bsg_job *bsg_job)
246 {
247 	struct fc_rport *rport;
248 	fc_port_t *fcport = NULL;
249 	struct Scsi_Host *host;
250 	scsi_qla_host_t *vha;
251 	struct qla_hw_data *ha;
252 	srb_t *sp;
253 	const char *type;
254 	int req_sg_cnt, rsp_sg_cnt;
255 	int rval =  (DRIVER_ERROR << 16);
256 	uint16_t nextlid = 0;
257 
258 	if (bsg_job->request->msgcode == FC_BSG_RPT_ELS) {
259 		rport = bsg_job->rport;
260 		fcport = *(fc_port_t **) rport->dd_data;
261 		host = rport_to_shost(rport);
262 		vha = shost_priv(host);
263 		ha = vha->hw;
264 		type = "FC_BSG_RPT_ELS";
265 	} else {
266 		host = bsg_job->shost;
267 		vha = shost_priv(host);
268 		ha = vha->hw;
269 		type = "FC_BSG_HST_ELS_NOLOGIN";
270 	}
271 
272 	if (!vha->flags.online) {
273 		ql_log(ql_log_warn, vha, 0x7005, "Host not online.\n");
274 		rval = -EIO;
275 		goto done;
276 	}
277 
278 	/* pass through is supported only for ISP 4Gb or higher */
279 	if (!IS_FWI2_CAPABLE(ha)) {
280 		ql_dbg(ql_dbg_user, vha, 0x7001,
281 		    "ELS passthru not supported for ISP23xx based adapters.\n");
282 		rval = -EPERM;
283 		goto done;
284 	}
285 
286 	/*  Multiple SG's are not supported for ELS requests */
287 	if (bsg_job->request_payload.sg_cnt > 1 ||
288 		bsg_job->reply_payload.sg_cnt > 1) {
289 		ql_dbg(ql_dbg_user, vha, 0x7002,
290 		    "Multiple SG's are not suppored for ELS requests, "
291 		    "request_sg_cnt=%x reply_sg_cnt=%x.\n",
292 		    bsg_job->request_payload.sg_cnt,
293 		    bsg_job->reply_payload.sg_cnt);
294 		rval = -EPERM;
295 		goto done;
296 	}
297 
298 	/* ELS request for rport */
299 	if (bsg_job->request->msgcode == FC_BSG_RPT_ELS) {
300 		/* make sure the rport is logged in,
301 		 * if not perform fabric login
302 		 */
303 		if (qla2x00_fabric_login(vha, fcport, &nextlid)) {
304 			ql_dbg(ql_dbg_user, vha, 0x7003,
305 			    "Failed to login port %06X for ELS passthru.\n",
306 			    fcport->d_id.b24);
307 			rval = -EIO;
308 			goto done;
309 		}
310 	} else {
311 		/* Allocate a dummy fcport structure, since functions
312 		 * preparing the IOCB and mailbox command retrieves port
313 		 * specific information from fcport structure. For Host based
314 		 * ELS commands there will be no fcport structure allocated
315 		 */
316 		fcport = qla2x00_alloc_fcport(vha, GFP_KERNEL);
317 		if (!fcport) {
318 			rval = -ENOMEM;
319 			goto done;
320 		}
321 
322 		/* Initialize all required  fields of fcport */
323 		fcport->vha = vha;
324 		fcport->d_id.b.al_pa =
325 			bsg_job->request->rqst_data.h_els.port_id[0];
326 		fcport->d_id.b.area =
327 			bsg_job->request->rqst_data.h_els.port_id[1];
328 		fcport->d_id.b.domain =
329 			bsg_job->request->rqst_data.h_els.port_id[2];
330 		fcport->loop_id =
331 			(fcport->d_id.b.al_pa == 0xFD) ?
332 			NPH_FABRIC_CONTROLLER : NPH_F_PORT;
333 	}
334 
335 	req_sg_cnt =
336 		dma_map_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list,
337 		bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
338 	if (!req_sg_cnt) {
339 		rval = -ENOMEM;
340 		goto done_free_fcport;
341 	}
342 
343 	rsp_sg_cnt = dma_map_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list,
344 		bsg_job->reply_payload.sg_cnt, DMA_FROM_DEVICE);
345         if (!rsp_sg_cnt) {
346 		rval = -ENOMEM;
347 		goto done_free_fcport;
348 	}
349 
350 	if ((req_sg_cnt !=  bsg_job->request_payload.sg_cnt) ||
351 		(rsp_sg_cnt != bsg_job->reply_payload.sg_cnt)) {
352 		ql_log(ql_log_warn, vha, 0x7008,
353 		    "dma mapping resulted in different sg counts, "
354 		    "request_sg_cnt: %x dma_request_sg_cnt:%x reply_sg_cnt:%x "
355 		    "dma_reply_sg_cnt:%x.\n", bsg_job->request_payload.sg_cnt,
356 		    req_sg_cnt, bsg_job->reply_payload.sg_cnt, rsp_sg_cnt);
357 		rval = -EAGAIN;
358 		goto done_unmap_sg;
359 	}
360 
361 	/* Alloc SRB structure */
362 	sp = qla2x00_get_sp(vha, fcport, GFP_KERNEL);
363 	if (!sp) {
364 		rval = -ENOMEM;
365 		goto done_unmap_sg;
366 	}
367 
368 	sp->type =
369 		(bsg_job->request->msgcode == FC_BSG_RPT_ELS ?
370 		SRB_ELS_CMD_RPT : SRB_ELS_CMD_HST);
371 	sp->name =
372 		(bsg_job->request->msgcode == FC_BSG_RPT_ELS ?
373 		"bsg_els_rpt" : "bsg_els_hst");
374 	sp->u.bsg_job = bsg_job;
375 	sp->free = qla2x00_bsg_sp_free;
376 	sp->done = qla2x00_bsg_job_done;
377 
378 	ql_dbg(ql_dbg_user, vha, 0x700a,
379 	    "bsg rqst type: %s els type: %x - loop-id=%x "
380 	    "portid=%-2x%02x%02x.\n", type,
381 	    bsg_job->request->rqst_data.h_els.command_code, fcport->loop_id,
382 	    fcport->d_id.b.domain, fcport->d_id.b.area, fcport->d_id.b.al_pa);
383 
384 	rval = qla2x00_start_sp(sp);
385 	if (rval != QLA_SUCCESS) {
386 		ql_log(ql_log_warn, vha, 0x700e,
387 		    "qla2x00_start_sp failed = %d\n", rval);
388 		qla2x00_rel_sp(vha, sp);
389 		rval = -EIO;
390 		goto done_unmap_sg;
391 	}
392 	return rval;
393 
394 done_unmap_sg:
395 	dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list,
396 		bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
397 	dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list,
398 		bsg_job->reply_payload.sg_cnt, DMA_FROM_DEVICE);
399 	goto done_free_fcport;
400 
401 done_free_fcport:
402 	if (bsg_job->request->msgcode == FC_BSG_RPT_ELS)
403 		kfree(fcport);
404 done:
405 	return rval;
406 }
407 
408 static inline uint16_t
409 qla24xx_calc_ct_iocbs(uint16_t dsds)
410 {
411 	uint16_t iocbs;
412 
413 	iocbs = 1;
414 	if (dsds > 2) {
415 		iocbs += (dsds - 2) / 5;
416 		if ((dsds - 2) % 5)
417 			iocbs++;
418 	}
419 	return iocbs;
420 }
421 
422 static int
423 qla2x00_process_ct(struct fc_bsg_job *bsg_job)
424 {
425 	srb_t *sp;
426 	struct Scsi_Host *host = bsg_job->shost;
427 	scsi_qla_host_t *vha = shost_priv(host);
428 	struct qla_hw_data *ha = vha->hw;
429 	int rval = (DRIVER_ERROR << 16);
430 	int req_sg_cnt, rsp_sg_cnt;
431 	uint16_t loop_id;
432 	struct fc_port *fcport;
433 	char  *type = "FC_BSG_HST_CT";
434 
435 	req_sg_cnt =
436 		dma_map_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list,
437 			bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
438 	if (!req_sg_cnt) {
439 		ql_log(ql_log_warn, vha, 0x700f,
440 		    "dma_map_sg return %d for request\n", req_sg_cnt);
441 		rval = -ENOMEM;
442 		goto done;
443 	}
444 
445 	rsp_sg_cnt = dma_map_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list,
446 		bsg_job->reply_payload.sg_cnt, DMA_FROM_DEVICE);
447 	if (!rsp_sg_cnt) {
448 		ql_log(ql_log_warn, vha, 0x7010,
449 		    "dma_map_sg return %d for reply\n", rsp_sg_cnt);
450 		rval = -ENOMEM;
451 		goto done;
452 	}
453 
454 	if ((req_sg_cnt !=  bsg_job->request_payload.sg_cnt) ||
455 	    (rsp_sg_cnt != bsg_job->reply_payload.sg_cnt)) {
456 		ql_log(ql_log_warn, vha, 0x7011,
457 		    "request_sg_cnt: %x dma_request_sg_cnt: %x reply_sg_cnt:%x "
458 		    "dma_reply_sg_cnt: %x\n", bsg_job->request_payload.sg_cnt,
459 		    req_sg_cnt, bsg_job->reply_payload.sg_cnt, rsp_sg_cnt);
460 		rval = -EAGAIN;
461 		goto done_unmap_sg;
462 	}
463 
464 	if (!vha->flags.online) {
465 		ql_log(ql_log_warn, vha, 0x7012,
466 		    "Host is not online.\n");
467 		rval = -EIO;
468 		goto done_unmap_sg;
469 	}
470 
471 	loop_id =
472 		(bsg_job->request->rqst_data.h_ct.preamble_word1 & 0xFF000000)
473 			>> 24;
474 	switch (loop_id) {
475 	case 0xFC:
476 		loop_id = cpu_to_le16(NPH_SNS);
477 		break;
478 	case 0xFA:
479 		loop_id = vha->mgmt_svr_loop_id;
480 		break;
481 	default:
482 		ql_dbg(ql_dbg_user, vha, 0x7013,
483 		    "Unknown loop id: %x.\n", loop_id);
484 		rval = -EINVAL;
485 		goto done_unmap_sg;
486 	}
487 
488 	/* Allocate a dummy fcport structure, since functions preparing the
489 	 * IOCB and mailbox command retrieves port specific information
490 	 * from fcport structure. For Host based ELS commands there will be
491 	 * no fcport structure allocated
492 	 */
493 	fcport = qla2x00_alloc_fcport(vha, GFP_KERNEL);
494 	if (!fcport) {
495 		ql_log(ql_log_warn, vha, 0x7014,
496 		    "Failed to allocate fcport.\n");
497 		rval = -ENOMEM;
498 		goto done_unmap_sg;
499 	}
500 
501 	/* Initialize all required  fields of fcport */
502 	fcport->vha = vha;
503 	fcport->d_id.b.al_pa = bsg_job->request->rqst_data.h_ct.port_id[0];
504 	fcport->d_id.b.area = bsg_job->request->rqst_data.h_ct.port_id[1];
505 	fcport->d_id.b.domain = bsg_job->request->rqst_data.h_ct.port_id[2];
506 	fcport->loop_id = loop_id;
507 
508 	/* Alloc SRB structure */
509 	sp = qla2x00_get_sp(vha, fcport, GFP_KERNEL);
510 	if (!sp) {
511 		ql_log(ql_log_warn, vha, 0x7015,
512 		    "qla2x00_get_sp failed.\n");
513 		rval = -ENOMEM;
514 		goto done_free_fcport;
515 	}
516 
517 	sp->type = SRB_CT_CMD;
518 	sp->name = "bsg_ct";
519 	sp->iocbs = qla24xx_calc_ct_iocbs(req_sg_cnt + rsp_sg_cnt);
520 	sp->u.bsg_job = bsg_job;
521 	sp->free = qla2x00_bsg_sp_free;
522 	sp->done = qla2x00_bsg_job_done;
523 
524 	ql_dbg(ql_dbg_user, vha, 0x7016,
525 	    "bsg rqst type: %s else type: %x - "
526 	    "loop-id=%x portid=%02x%02x%02x.\n", type,
527 	    (bsg_job->request->rqst_data.h_ct.preamble_word2 >> 16),
528 	    fcport->loop_id, fcport->d_id.b.domain, fcport->d_id.b.area,
529 	    fcport->d_id.b.al_pa);
530 
531 	rval = qla2x00_start_sp(sp);
532 	if (rval != QLA_SUCCESS) {
533 		ql_log(ql_log_warn, vha, 0x7017,
534 		    "qla2x00_start_sp failed=%d.\n", rval);
535 		qla2x00_rel_sp(vha, sp);
536 		rval = -EIO;
537 		goto done_free_fcport;
538 	}
539 	return rval;
540 
541 done_free_fcport:
542 	kfree(fcport);
543 done_unmap_sg:
544 	dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list,
545 		bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
546 	dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list,
547 		bsg_job->reply_payload.sg_cnt, DMA_FROM_DEVICE);
548 done:
549 	return rval;
550 }
551 
552 /* Disable loopback mode */
553 static inline int
554 qla81xx_reset_loopback_mode(scsi_qla_host_t *vha, uint16_t *config,
555 			    int wait, int wait2)
556 {
557 	int ret = 0;
558 	int rval = 0;
559 	uint16_t new_config[4];
560 	struct qla_hw_data *ha = vha->hw;
561 
562 	if (!IS_QLA81XX(ha) && !IS_QLA8031(ha) && !IS_QLA8044(ha))
563 		goto done_reset_internal;
564 
565 	memset(new_config, 0 , sizeof(new_config));
566 	if ((config[0] & INTERNAL_LOOPBACK_MASK) >> 1 ==
567 	    ENABLE_INTERNAL_LOOPBACK ||
568 	    (config[0] & INTERNAL_LOOPBACK_MASK) >> 1 ==
569 	    ENABLE_EXTERNAL_LOOPBACK) {
570 		new_config[0] = config[0] & ~INTERNAL_LOOPBACK_MASK;
571 		ql_dbg(ql_dbg_user, vha, 0x70bf, "new_config[0]=%02x\n",
572 		    (new_config[0] & INTERNAL_LOOPBACK_MASK));
573 		memcpy(&new_config[1], &config[1], sizeof(uint16_t) * 3) ;
574 
575 		ha->notify_dcbx_comp = wait;
576 		ha->notify_lb_portup_comp = wait2;
577 
578 		ret = qla81xx_set_port_config(vha, new_config);
579 		if (ret != QLA_SUCCESS) {
580 			ql_log(ql_log_warn, vha, 0x7025,
581 			    "Set port config failed.\n");
582 			ha->notify_dcbx_comp = 0;
583 			ha->notify_lb_portup_comp = 0;
584 			rval = -EINVAL;
585 			goto done_reset_internal;
586 		}
587 
588 		/* Wait for DCBX complete event */
589 		if (wait && !wait_for_completion_timeout(&ha->dcbx_comp,
590 			(DCBX_COMP_TIMEOUT * HZ))) {
591 			ql_dbg(ql_dbg_user, vha, 0x7026,
592 			    "DCBX completion not received.\n");
593 			ha->notify_dcbx_comp = 0;
594 			ha->notify_lb_portup_comp = 0;
595 			rval = -EINVAL;
596 			goto done_reset_internal;
597 		} else
598 			ql_dbg(ql_dbg_user, vha, 0x7027,
599 			    "DCBX completion received.\n");
600 
601 		if (wait2 &&
602 		    !wait_for_completion_timeout(&ha->lb_portup_comp,
603 		    (LB_PORTUP_COMP_TIMEOUT * HZ))) {
604 			ql_dbg(ql_dbg_user, vha, 0x70c5,
605 			    "Port up completion not received.\n");
606 			ha->notify_lb_portup_comp = 0;
607 			rval = -EINVAL;
608 			goto done_reset_internal;
609 		} else
610 			ql_dbg(ql_dbg_user, vha, 0x70c6,
611 			    "Port up completion received.\n");
612 
613 		ha->notify_dcbx_comp = 0;
614 		ha->notify_lb_portup_comp = 0;
615 	}
616 done_reset_internal:
617 	return rval;
618 }
619 
620 /*
621  * Set the port configuration to enable the internal or external loopback
622  * depending on the loopback mode.
623  */
624 static inline int
625 qla81xx_set_loopback_mode(scsi_qla_host_t *vha, uint16_t *config,
626 	uint16_t *new_config, uint16_t mode)
627 {
628 	int ret = 0;
629 	int rval = 0;
630 	unsigned long rem_tmo = 0, current_tmo = 0;
631 	struct qla_hw_data *ha = vha->hw;
632 
633 	if (!IS_QLA81XX(ha) && !IS_QLA8031(ha) && !IS_QLA8044(ha))
634 		goto done_set_internal;
635 
636 	if (mode == INTERNAL_LOOPBACK)
637 		new_config[0] = config[0] | (ENABLE_INTERNAL_LOOPBACK << 1);
638 	else if (mode == EXTERNAL_LOOPBACK)
639 		new_config[0] = config[0] | (ENABLE_EXTERNAL_LOOPBACK << 1);
640 	ql_dbg(ql_dbg_user, vha, 0x70be,
641 	     "new_config[0]=%02x\n", (new_config[0] & INTERNAL_LOOPBACK_MASK));
642 
643 	memcpy(&new_config[1], &config[1], sizeof(uint16_t) * 3);
644 
645 	ha->notify_dcbx_comp = 1;
646 	ret = qla81xx_set_port_config(vha, new_config);
647 	if (ret != QLA_SUCCESS) {
648 		ql_log(ql_log_warn, vha, 0x7021,
649 		    "set port config failed.\n");
650 		ha->notify_dcbx_comp = 0;
651 		rval = -EINVAL;
652 		goto done_set_internal;
653 	}
654 
655 	/* Wait for DCBX complete event */
656 	current_tmo = DCBX_COMP_TIMEOUT * HZ;
657 	while (1) {
658 		rem_tmo = wait_for_completion_timeout(&ha->dcbx_comp,
659 		    current_tmo);
660 		if (!ha->idc_extend_tmo || rem_tmo) {
661 			ha->idc_extend_tmo = 0;
662 			break;
663 		}
664 		current_tmo = ha->idc_extend_tmo * HZ;
665 		ha->idc_extend_tmo = 0;
666 	}
667 
668 	if (!rem_tmo) {
669 		ql_dbg(ql_dbg_user, vha, 0x7022,
670 		    "DCBX completion not received.\n");
671 		ret = qla81xx_reset_loopback_mode(vha, new_config, 0, 0);
672 		/*
673 		 * If the reset of the loopback mode doesn't work take a FCoE
674 		 * dump and reset the chip.
675 		 */
676 		if (ret) {
677 			ha->isp_ops->fw_dump(vha, 0);
678 			set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
679 		}
680 		rval = -EINVAL;
681 	} else {
682 		if (ha->flags.idc_compl_status) {
683 			ql_dbg(ql_dbg_user, vha, 0x70c3,
684 			    "Bad status in IDC Completion AEN\n");
685 			rval = -EINVAL;
686 			ha->flags.idc_compl_status = 0;
687 		} else
688 			ql_dbg(ql_dbg_user, vha, 0x7023,
689 			    "DCBX completion received.\n");
690 	}
691 
692 	ha->notify_dcbx_comp = 0;
693 	ha->idc_extend_tmo = 0;
694 
695 done_set_internal:
696 	return rval;
697 }
698 
699 static int
700 qla2x00_process_loopback(struct fc_bsg_job *bsg_job)
701 {
702 	struct Scsi_Host *host = bsg_job->shost;
703 	scsi_qla_host_t *vha = shost_priv(host);
704 	struct qla_hw_data *ha = vha->hw;
705 	int rval;
706 	uint8_t command_sent;
707 	char *type;
708 	struct msg_echo_lb elreq;
709 	uint16_t response[MAILBOX_REGISTER_COUNT];
710 	uint16_t config[4], new_config[4];
711 	uint8_t *fw_sts_ptr;
712 	uint8_t *req_data = NULL;
713 	dma_addr_t req_data_dma;
714 	uint32_t req_data_len;
715 	uint8_t *rsp_data = NULL;
716 	dma_addr_t rsp_data_dma;
717 	uint32_t rsp_data_len;
718 
719 	if (!vha->flags.online) {
720 		ql_log(ql_log_warn, vha, 0x7019, "Host is not online.\n");
721 		return -EIO;
722 	}
723 
724 	elreq.req_sg_cnt = dma_map_sg(&ha->pdev->dev,
725 		bsg_job->request_payload.sg_list, bsg_job->request_payload.sg_cnt,
726 		DMA_TO_DEVICE);
727 
728 	if (!elreq.req_sg_cnt) {
729 		ql_log(ql_log_warn, vha, 0x701a,
730 		    "dma_map_sg returned %d for request.\n", elreq.req_sg_cnt);
731 		return -ENOMEM;
732 	}
733 
734 	elreq.rsp_sg_cnt = dma_map_sg(&ha->pdev->dev,
735 		bsg_job->reply_payload.sg_list, bsg_job->reply_payload.sg_cnt,
736 		DMA_FROM_DEVICE);
737 
738 	if (!elreq.rsp_sg_cnt) {
739 		ql_log(ql_log_warn, vha, 0x701b,
740 		    "dma_map_sg returned %d for reply.\n", elreq.rsp_sg_cnt);
741 		rval = -ENOMEM;
742 		goto done_unmap_req_sg;
743 	}
744 
745 	if ((elreq.req_sg_cnt !=  bsg_job->request_payload.sg_cnt) ||
746 		(elreq.rsp_sg_cnt != bsg_job->reply_payload.sg_cnt)) {
747 		ql_log(ql_log_warn, vha, 0x701c,
748 		    "dma mapping resulted in different sg counts, "
749 		    "request_sg_cnt: %x dma_request_sg_cnt: %x "
750 		    "reply_sg_cnt: %x dma_reply_sg_cnt: %x.\n",
751 		    bsg_job->request_payload.sg_cnt, elreq.req_sg_cnt,
752 		    bsg_job->reply_payload.sg_cnt, elreq.rsp_sg_cnt);
753 		rval = -EAGAIN;
754 		goto done_unmap_sg;
755 	}
756 	req_data_len = rsp_data_len = bsg_job->request_payload.payload_len;
757 	req_data = dma_alloc_coherent(&ha->pdev->dev, req_data_len,
758 		&req_data_dma, GFP_KERNEL);
759 	if (!req_data) {
760 		ql_log(ql_log_warn, vha, 0x701d,
761 		    "dma alloc failed for req_data.\n");
762 		rval = -ENOMEM;
763 		goto done_unmap_sg;
764 	}
765 
766 	rsp_data = dma_alloc_coherent(&ha->pdev->dev, rsp_data_len,
767 		&rsp_data_dma, GFP_KERNEL);
768 	if (!rsp_data) {
769 		ql_log(ql_log_warn, vha, 0x7004,
770 		    "dma alloc failed for rsp_data.\n");
771 		rval = -ENOMEM;
772 		goto done_free_dma_req;
773 	}
774 
775 	/* Copy the request buffer in req_data now */
776 	sg_copy_to_buffer(bsg_job->request_payload.sg_list,
777 		bsg_job->request_payload.sg_cnt, req_data, req_data_len);
778 
779 	elreq.send_dma = req_data_dma;
780 	elreq.rcv_dma = rsp_data_dma;
781 	elreq.transfer_size = req_data_len;
782 
783 	elreq.options = bsg_job->request->rqst_data.h_vendor.vendor_cmd[1];
784 	elreq.iteration_count =
785 	    bsg_job->request->rqst_data.h_vendor.vendor_cmd[2];
786 
787 	if (atomic_read(&vha->loop_state) == LOOP_READY &&
788 	    (ha->current_topology == ISP_CFG_F ||
789 	    ((IS_QLA81XX(ha) || IS_QLA8031(ha) || IS_QLA8044(ha)) &&
790 	    le32_to_cpu(*(uint32_t *)req_data) == ELS_OPCODE_BYTE
791 	    && req_data_len == MAX_ELS_FRAME_PAYLOAD)) &&
792 		elreq.options == EXTERNAL_LOOPBACK) {
793 		type = "FC_BSG_HST_VENDOR_ECHO_DIAG";
794 		ql_dbg(ql_dbg_user, vha, 0x701e,
795 		    "BSG request type: %s.\n", type);
796 		command_sent = INT_DEF_LB_ECHO_CMD;
797 		rval = qla2x00_echo_test(vha, &elreq, response);
798 	} else {
799 		if (IS_QLA81XX(ha) || IS_QLA8031(ha) || IS_QLA8044(ha)) {
800 			memset(config, 0, sizeof(config));
801 			memset(new_config, 0, sizeof(new_config));
802 
803 			if (qla81xx_get_port_config(vha, config)) {
804 				ql_log(ql_log_warn, vha, 0x701f,
805 				    "Get port config failed.\n");
806 				rval = -EPERM;
807 				goto done_free_dma_rsp;
808 			}
809 
810 			if ((config[0] & INTERNAL_LOOPBACK_MASK) != 0) {
811 				ql_dbg(ql_dbg_user, vha, 0x70c4,
812 				    "Loopback operation already in "
813 				    "progress.\n");
814 				rval = -EAGAIN;
815 				goto done_free_dma_rsp;
816 			}
817 
818 			ql_dbg(ql_dbg_user, vha, 0x70c0,
819 			    "elreq.options=%04x\n", elreq.options);
820 
821 			if (elreq.options == EXTERNAL_LOOPBACK)
822 				if (IS_QLA8031(ha) || IS_QLA8044(ha))
823 					rval = qla81xx_set_loopback_mode(vha,
824 					    config, new_config, elreq.options);
825 				else
826 					rval = qla81xx_reset_loopback_mode(vha,
827 					    config, 1, 0);
828 			else
829 				rval = qla81xx_set_loopback_mode(vha, config,
830 				    new_config, elreq.options);
831 
832 			if (rval) {
833 				rval = -EPERM;
834 				goto done_free_dma_rsp;
835 			}
836 
837 			type = "FC_BSG_HST_VENDOR_LOOPBACK";
838 			ql_dbg(ql_dbg_user, vha, 0x7028,
839 			    "BSG request type: %s.\n", type);
840 
841 			command_sent = INT_DEF_LB_LOOPBACK_CMD;
842 			rval = qla2x00_loopback_test(vha, &elreq, response);
843 
844 			if (response[0] == MBS_COMMAND_ERROR &&
845 					response[1] == MBS_LB_RESET) {
846 				ql_log(ql_log_warn, vha, 0x7029,
847 				    "MBX command error, Aborting ISP.\n");
848 				set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
849 				qla2xxx_wake_dpc(vha);
850 				qla2x00_wait_for_chip_reset(vha);
851 				/* Also reset the MPI */
852 				if (IS_QLA81XX(ha)) {
853 					if (qla81xx_restart_mpi_firmware(vha) !=
854 					    QLA_SUCCESS) {
855 						ql_log(ql_log_warn, vha, 0x702a,
856 						    "MPI reset failed.\n");
857 					}
858 				}
859 
860 				rval = -EIO;
861 				goto done_free_dma_rsp;
862 			}
863 
864 			if (new_config[0]) {
865 				int ret;
866 
867 				/* Revert back to original port config
868 				 * Also clear internal loopback
869 				 */
870 				ret = qla81xx_reset_loopback_mode(vha,
871 				    new_config, 0, 1);
872 				if (ret) {
873 					/*
874 					 * If the reset of the loopback mode
875 					 * doesn't work take FCoE dump and then
876 					 * reset the chip.
877 					 */
878 					ha->isp_ops->fw_dump(vha, 0);
879 					set_bit(ISP_ABORT_NEEDED,
880 					    &vha->dpc_flags);
881 				}
882 
883 			}
884 
885 		} else {
886 			type = "FC_BSG_HST_VENDOR_LOOPBACK";
887 			ql_dbg(ql_dbg_user, vha, 0x702b,
888 			    "BSG request type: %s.\n", type);
889 			command_sent = INT_DEF_LB_LOOPBACK_CMD;
890 			rval = qla2x00_loopback_test(vha, &elreq, response);
891 		}
892 	}
893 
894 	if (rval) {
895 		ql_log(ql_log_warn, vha, 0x702c,
896 		    "Vendor request %s failed.\n", type);
897 
898 		rval = 0;
899 		bsg_job->reply->result = (DID_ERROR << 16);
900 		bsg_job->reply->reply_payload_rcv_len = 0;
901 	} else {
902 		ql_dbg(ql_dbg_user, vha, 0x702d,
903 		    "Vendor request %s completed.\n", type);
904 		bsg_job->reply->result = (DID_OK << 16);
905 		sg_copy_from_buffer(bsg_job->reply_payload.sg_list,
906 			bsg_job->reply_payload.sg_cnt, rsp_data,
907 			rsp_data_len);
908 	}
909 
910 	bsg_job->reply_len = sizeof(struct fc_bsg_reply) +
911 	    sizeof(response) + sizeof(uint8_t);
912 	fw_sts_ptr = ((uint8_t *)bsg_job->req->sense) +
913 	    sizeof(struct fc_bsg_reply);
914 	memcpy(fw_sts_ptr, response, sizeof(response));
915 	fw_sts_ptr += sizeof(response);
916 	*fw_sts_ptr = command_sent;
917 
918 done_free_dma_rsp:
919 	dma_free_coherent(&ha->pdev->dev, rsp_data_len,
920 		rsp_data, rsp_data_dma);
921 done_free_dma_req:
922 	dma_free_coherent(&ha->pdev->dev, req_data_len,
923 		req_data, req_data_dma);
924 done_unmap_sg:
925 	dma_unmap_sg(&ha->pdev->dev,
926 	    bsg_job->reply_payload.sg_list,
927 	    bsg_job->reply_payload.sg_cnt, DMA_FROM_DEVICE);
928 done_unmap_req_sg:
929 	dma_unmap_sg(&ha->pdev->dev,
930 	    bsg_job->request_payload.sg_list,
931 	    bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
932 	if (!rval)
933 		bsg_job->job_done(bsg_job);
934 	return rval;
935 }
936 
937 static int
938 qla84xx_reset(struct fc_bsg_job *bsg_job)
939 {
940 	struct Scsi_Host *host = bsg_job->shost;
941 	scsi_qla_host_t *vha = shost_priv(host);
942 	struct qla_hw_data *ha = vha->hw;
943 	int rval = 0;
944 	uint32_t flag;
945 
946 	if (!IS_QLA84XX(ha)) {
947 		ql_dbg(ql_dbg_user, vha, 0x702f, "Not 84xx, exiting.\n");
948 		return -EINVAL;
949 	}
950 
951 	flag = bsg_job->request->rqst_data.h_vendor.vendor_cmd[1];
952 
953 	rval = qla84xx_reset_chip(vha, flag == A84_ISSUE_RESET_DIAG_FW);
954 
955 	if (rval) {
956 		ql_log(ql_log_warn, vha, 0x7030,
957 		    "Vendor request 84xx reset failed.\n");
958 		rval = (DID_ERROR << 16);
959 
960 	} else {
961 		ql_dbg(ql_dbg_user, vha, 0x7031,
962 		    "Vendor request 84xx reset completed.\n");
963 		bsg_job->reply->result = DID_OK;
964 		bsg_job->job_done(bsg_job);
965 	}
966 
967 	return rval;
968 }
969 
970 static int
971 qla84xx_updatefw(struct fc_bsg_job *bsg_job)
972 {
973 	struct Scsi_Host *host = bsg_job->shost;
974 	scsi_qla_host_t *vha = shost_priv(host);
975 	struct qla_hw_data *ha = vha->hw;
976 	struct verify_chip_entry_84xx *mn = NULL;
977 	dma_addr_t mn_dma, fw_dma;
978 	void *fw_buf = NULL;
979 	int rval = 0;
980 	uint32_t sg_cnt;
981 	uint32_t data_len;
982 	uint16_t options;
983 	uint32_t flag;
984 	uint32_t fw_ver;
985 
986 	if (!IS_QLA84XX(ha)) {
987 		ql_dbg(ql_dbg_user, vha, 0x7032,
988 		    "Not 84xx, exiting.\n");
989 		return -EINVAL;
990 	}
991 
992 	sg_cnt = dma_map_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list,
993 		bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
994 	if (!sg_cnt) {
995 		ql_log(ql_log_warn, vha, 0x7033,
996 		    "dma_map_sg returned %d for request.\n", sg_cnt);
997 		return -ENOMEM;
998 	}
999 
1000 	if (sg_cnt != bsg_job->request_payload.sg_cnt) {
1001 		ql_log(ql_log_warn, vha, 0x7034,
1002 		    "DMA mapping resulted in different sg counts, "
1003 		    "request_sg_cnt: %x dma_request_sg_cnt: %x.\n",
1004 		    bsg_job->request_payload.sg_cnt, sg_cnt);
1005 		rval = -EAGAIN;
1006 		goto done_unmap_sg;
1007 	}
1008 
1009 	data_len = bsg_job->request_payload.payload_len;
1010 	fw_buf = dma_alloc_coherent(&ha->pdev->dev, data_len,
1011 		&fw_dma, GFP_KERNEL);
1012 	if (!fw_buf) {
1013 		ql_log(ql_log_warn, vha, 0x7035,
1014 		    "DMA alloc failed for fw_buf.\n");
1015 		rval = -ENOMEM;
1016 		goto done_unmap_sg;
1017 	}
1018 
1019 	sg_copy_to_buffer(bsg_job->request_payload.sg_list,
1020 		bsg_job->request_payload.sg_cnt, fw_buf, data_len);
1021 
1022 	mn = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &mn_dma);
1023 	if (!mn) {
1024 		ql_log(ql_log_warn, vha, 0x7036,
1025 		    "DMA alloc failed for fw buffer.\n");
1026 		rval = -ENOMEM;
1027 		goto done_free_fw_buf;
1028 	}
1029 
1030 	flag = bsg_job->request->rqst_data.h_vendor.vendor_cmd[1];
1031 	fw_ver = le32_to_cpu(*((uint32_t *)((uint32_t *)fw_buf + 2)));
1032 
1033 	memset(mn, 0, sizeof(struct access_chip_84xx));
1034 	mn->entry_type = VERIFY_CHIP_IOCB_TYPE;
1035 	mn->entry_count = 1;
1036 
1037 	options = VCO_FORCE_UPDATE | VCO_END_OF_DATA;
1038 	if (flag == A84_ISSUE_UPDATE_DIAGFW_CMD)
1039 		options |= VCO_DIAG_FW;
1040 
1041 	mn->options = cpu_to_le16(options);
1042 	mn->fw_ver =  cpu_to_le32(fw_ver);
1043 	mn->fw_size =  cpu_to_le32(data_len);
1044 	mn->fw_seq_size =  cpu_to_le32(data_len);
1045 	mn->dseg_address[0] = cpu_to_le32(LSD(fw_dma));
1046 	mn->dseg_address[1] = cpu_to_le32(MSD(fw_dma));
1047 	mn->dseg_length = cpu_to_le32(data_len);
1048 	mn->data_seg_cnt = cpu_to_le16(1);
1049 
1050 	rval = qla2x00_issue_iocb_timeout(vha, mn, mn_dma, 0, 120);
1051 
1052 	if (rval) {
1053 		ql_log(ql_log_warn, vha, 0x7037,
1054 		    "Vendor request 84xx updatefw failed.\n");
1055 
1056 		rval = (DID_ERROR << 16);
1057 	} else {
1058 		ql_dbg(ql_dbg_user, vha, 0x7038,
1059 		    "Vendor request 84xx updatefw completed.\n");
1060 
1061 		bsg_job->reply_len = sizeof(struct fc_bsg_reply);
1062 		bsg_job->reply->result = DID_OK;
1063 	}
1064 
1065 	dma_pool_free(ha->s_dma_pool, mn, mn_dma);
1066 
1067 done_free_fw_buf:
1068 	dma_free_coherent(&ha->pdev->dev, data_len, fw_buf, fw_dma);
1069 
1070 done_unmap_sg:
1071 	dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list,
1072 		bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
1073 
1074 	if (!rval)
1075 		bsg_job->job_done(bsg_job);
1076 	return rval;
1077 }
1078 
1079 static int
1080 qla84xx_mgmt_cmd(struct fc_bsg_job *bsg_job)
1081 {
1082 	struct Scsi_Host *host = bsg_job->shost;
1083 	scsi_qla_host_t *vha = shost_priv(host);
1084 	struct qla_hw_data *ha = vha->hw;
1085 	struct access_chip_84xx *mn = NULL;
1086 	dma_addr_t mn_dma, mgmt_dma;
1087 	void *mgmt_b = NULL;
1088 	int rval = 0;
1089 	struct qla_bsg_a84_mgmt *ql84_mgmt;
1090 	uint32_t sg_cnt;
1091 	uint32_t data_len = 0;
1092 	uint32_t dma_direction = DMA_NONE;
1093 
1094 	if (!IS_QLA84XX(ha)) {
1095 		ql_log(ql_log_warn, vha, 0x703a,
1096 		    "Not 84xx, exiting.\n");
1097 		return -EINVAL;
1098 	}
1099 
1100 	mn = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &mn_dma);
1101 	if (!mn) {
1102 		ql_log(ql_log_warn, vha, 0x703c,
1103 		    "DMA alloc failed for fw buffer.\n");
1104 		return -ENOMEM;
1105 	}
1106 
1107 	memset(mn, 0, sizeof(struct access_chip_84xx));
1108 	mn->entry_type = ACCESS_CHIP_IOCB_TYPE;
1109 	mn->entry_count = 1;
1110 	ql84_mgmt = (void *)bsg_job->request + sizeof(struct fc_bsg_request);
1111 	switch (ql84_mgmt->mgmt.cmd) {
1112 	case QLA84_MGMT_READ_MEM:
1113 	case QLA84_MGMT_GET_INFO:
1114 		sg_cnt = dma_map_sg(&ha->pdev->dev,
1115 			bsg_job->reply_payload.sg_list,
1116 			bsg_job->reply_payload.sg_cnt, DMA_FROM_DEVICE);
1117 		if (!sg_cnt) {
1118 			ql_log(ql_log_warn, vha, 0x703d,
1119 			    "dma_map_sg returned %d for reply.\n", sg_cnt);
1120 			rval = -ENOMEM;
1121 			goto exit_mgmt;
1122 		}
1123 
1124 		dma_direction = DMA_FROM_DEVICE;
1125 
1126 		if (sg_cnt != bsg_job->reply_payload.sg_cnt) {
1127 			ql_log(ql_log_warn, vha, 0x703e,
1128 			    "DMA mapping resulted in different sg counts, "
1129 			    "reply_sg_cnt: %x dma_reply_sg_cnt: %x.\n",
1130 			    bsg_job->reply_payload.sg_cnt, sg_cnt);
1131 			rval = -EAGAIN;
1132 			goto done_unmap_sg;
1133 		}
1134 
1135 		data_len = bsg_job->reply_payload.payload_len;
1136 
1137 		mgmt_b = dma_alloc_coherent(&ha->pdev->dev, data_len,
1138 		    &mgmt_dma, GFP_KERNEL);
1139 		if (!mgmt_b) {
1140 			ql_log(ql_log_warn, vha, 0x703f,
1141 			    "DMA alloc failed for mgmt_b.\n");
1142 			rval = -ENOMEM;
1143 			goto done_unmap_sg;
1144 		}
1145 
1146 		if (ql84_mgmt->mgmt.cmd == QLA84_MGMT_READ_MEM) {
1147 			mn->options = cpu_to_le16(ACO_DUMP_MEMORY);
1148 			mn->parameter1 =
1149 				cpu_to_le32(
1150 				ql84_mgmt->mgmt.mgmtp.u.mem.start_addr);
1151 
1152 		} else if (ql84_mgmt->mgmt.cmd == QLA84_MGMT_GET_INFO) {
1153 			mn->options = cpu_to_le16(ACO_REQUEST_INFO);
1154 			mn->parameter1 =
1155 				cpu_to_le32(ql84_mgmt->mgmt.mgmtp.u.info.type);
1156 
1157 			mn->parameter2 =
1158 				cpu_to_le32(
1159 				ql84_mgmt->mgmt.mgmtp.u.info.context);
1160 		}
1161 		break;
1162 
1163 	case QLA84_MGMT_WRITE_MEM:
1164 		sg_cnt = dma_map_sg(&ha->pdev->dev,
1165 			bsg_job->request_payload.sg_list,
1166 			bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
1167 
1168 		if (!sg_cnt) {
1169 			ql_log(ql_log_warn, vha, 0x7040,
1170 			    "dma_map_sg returned %d.\n", sg_cnt);
1171 			rval = -ENOMEM;
1172 			goto exit_mgmt;
1173 		}
1174 
1175 		dma_direction = DMA_TO_DEVICE;
1176 
1177 		if (sg_cnt != bsg_job->request_payload.sg_cnt) {
1178 			ql_log(ql_log_warn, vha, 0x7041,
1179 			    "DMA mapping resulted in different sg counts, "
1180 			    "request_sg_cnt: %x dma_request_sg_cnt: %x.\n",
1181 			    bsg_job->request_payload.sg_cnt, sg_cnt);
1182 			rval = -EAGAIN;
1183 			goto done_unmap_sg;
1184 		}
1185 
1186 		data_len = bsg_job->request_payload.payload_len;
1187 		mgmt_b = dma_alloc_coherent(&ha->pdev->dev, data_len,
1188 			&mgmt_dma, GFP_KERNEL);
1189 		if (!mgmt_b) {
1190 			ql_log(ql_log_warn, vha, 0x7042,
1191 			    "DMA alloc failed for mgmt_b.\n");
1192 			rval = -ENOMEM;
1193 			goto done_unmap_sg;
1194 		}
1195 
1196 		sg_copy_to_buffer(bsg_job->request_payload.sg_list,
1197 			bsg_job->request_payload.sg_cnt, mgmt_b, data_len);
1198 
1199 		mn->options = cpu_to_le16(ACO_LOAD_MEMORY);
1200 		mn->parameter1 =
1201 			cpu_to_le32(ql84_mgmt->mgmt.mgmtp.u.mem.start_addr);
1202 		break;
1203 
1204 	case QLA84_MGMT_CHNG_CONFIG:
1205 		mn->options = cpu_to_le16(ACO_CHANGE_CONFIG_PARAM);
1206 		mn->parameter1 =
1207 			cpu_to_le32(ql84_mgmt->mgmt.mgmtp.u.config.id);
1208 
1209 		mn->parameter2 =
1210 			cpu_to_le32(ql84_mgmt->mgmt.mgmtp.u.config.param0);
1211 
1212 		mn->parameter3 =
1213 			cpu_to_le32(ql84_mgmt->mgmt.mgmtp.u.config.param1);
1214 		break;
1215 
1216 	default:
1217 		rval = -EIO;
1218 		goto exit_mgmt;
1219 	}
1220 
1221 	if (ql84_mgmt->mgmt.cmd != QLA84_MGMT_CHNG_CONFIG) {
1222 		mn->total_byte_cnt = cpu_to_le32(ql84_mgmt->mgmt.len);
1223 		mn->dseg_count = cpu_to_le16(1);
1224 		mn->dseg_address[0] = cpu_to_le32(LSD(mgmt_dma));
1225 		mn->dseg_address[1] = cpu_to_le32(MSD(mgmt_dma));
1226 		mn->dseg_length = cpu_to_le32(ql84_mgmt->mgmt.len);
1227 	}
1228 
1229 	rval = qla2x00_issue_iocb(vha, mn, mn_dma, 0);
1230 
1231 	if (rval) {
1232 		ql_log(ql_log_warn, vha, 0x7043,
1233 		    "Vendor request 84xx mgmt failed.\n");
1234 
1235 		rval = (DID_ERROR << 16);
1236 
1237 	} else {
1238 		ql_dbg(ql_dbg_user, vha, 0x7044,
1239 		    "Vendor request 84xx mgmt completed.\n");
1240 
1241 		bsg_job->reply_len = sizeof(struct fc_bsg_reply);
1242 		bsg_job->reply->result = DID_OK;
1243 
1244 		if ((ql84_mgmt->mgmt.cmd == QLA84_MGMT_READ_MEM) ||
1245 			(ql84_mgmt->mgmt.cmd == QLA84_MGMT_GET_INFO)) {
1246 			bsg_job->reply->reply_payload_rcv_len =
1247 				bsg_job->reply_payload.payload_len;
1248 
1249 			sg_copy_from_buffer(bsg_job->reply_payload.sg_list,
1250 				bsg_job->reply_payload.sg_cnt, mgmt_b,
1251 				data_len);
1252 		}
1253 	}
1254 
1255 done_unmap_sg:
1256 	if (mgmt_b)
1257 		dma_free_coherent(&ha->pdev->dev, data_len, mgmt_b, mgmt_dma);
1258 
1259 	if (dma_direction == DMA_TO_DEVICE)
1260 		dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list,
1261 			bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
1262 	else if (dma_direction == DMA_FROM_DEVICE)
1263 		dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list,
1264 			bsg_job->reply_payload.sg_cnt, DMA_FROM_DEVICE);
1265 
1266 exit_mgmt:
1267 	dma_pool_free(ha->s_dma_pool, mn, mn_dma);
1268 
1269 	if (!rval)
1270 		bsg_job->job_done(bsg_job);
1271 	return rval;
1272 }
1273 
1274 static int
1275 qla24xx_iidma(struct fc_bsg_job *bsg_job)
1276 {
1277 	struct Scsi_Host *host = bsg_job->shost;
1278 	scsi_qla_host_t *vha = shost_priv(host);
1279 	int rval = 0;
1280 	struct qla_port_param *port_param = NULL;
1281 	fc_port_t *fcport = NULL;
1282 	int found = 0;
1283 	uint16_t mb[MAILBOX_REGISTER_COUNT];
1284 	uint8_t *rsp_ptr = NULL;
1285 
1286 	if (!IS_IIDMA_CAPABLE(vha->hw)) {
1287 		ql_log(ql_log_info, vha, 0x7046, "iiDMA not supported.\n");
1288 		return -EINVAL;
1289 	}
1290 
1291 	port_param = (void *)bsg_job->request + sizeof(struct fc_bsg_request);
1292 	if (port_param->fc_scsi_addr.dest_type != EXT_DEF_TYPE_WWPN) {
1293 		ql_log(ql_log_warn, vha, 0x7048,
1294 		    "Invalid destination type.\n");
1295 		return -EINVAL;
1296 	}
1297 
1298 	list_for_each_entry(fcport, &vha->vp_fcports, list) {
1299 		if (fcport->port_type != FCT_TARGET)
1300 			continue;
1301 
1302 		if (memcmp(port_param->fc_scsi_addr.dest_addr.wwpn,
1303 			fcport->port_name, sizeof(fcport->port_name)))
1304 			continue;
1305 
1306 		found = 1;
1307 		break;
1308 	}
1309 
1310 	if (!found) {
1311 		ql_log(ql_log_warn, vha, 0x7049,
1312 		    "Failed to find port.\n");
1313 		return -EINVAL;
1314 	}
1315 
1316 	if (atomic_read(&fcport->state) != FCS_ONLINE) {
1317 		ql_log(ql_log_warn, vha, 0x704a,
1318 		    "Port is not online.\n");
1319 		return -EINVAL;
1320 	}
1321 
1322 	if (fcport->flags & FCF_LOGIN_NEEDED) {
1323 		ql_log(ql_log_warn, vha, 0x704b,
1324 		    "Remote port not logged in flags = 0x%x.\n", fcport->flags);
1325 		return -EINVAL;
1326 	}
1327 
1328 	if (port_param->mode)
1329 		rval = qla2x00_set_idma_speed(vha, fcport->loop_id,
1330 			port_param->speed, mb);
1331 	else
1332 		rval = qla2x00_get_idma_speed(vha, fcport->loop_id,
1333 			&port_param->speed, mb);
1334 
1335 	if (rval) {
1336 		ql_log(ql_log_warn, vha, 0x704c,
1337 		    "iIDMA cmd failed for %8phN -- "
1338 		    "%04x %x %04x %04x.\n", fcport->port_name,
1339 		    rval, fcport->fp_speed, mb[0], mb[1]);
1340 		rval = (DID_ERROR << 16);
1341 	} else {
1342 		if (!port_param->mode) {
1343 			bsg_job->reply_len = sizeof(struct fc_bsg_reply) +
1344 				sizeof(struct qla_port_param);
1345 
1346 			rsp_ptr = ((uint8_t *)bsg_job->reply) +
1347 				sizeof(struct fc_bsg_reply);
1348 
1349 			memcpy(rsp_ptr, port_param,
1350 				sizeof(struct qla_port_param));
1351 		}
1352 
1353 		bsg_job->reply->result = DID_OK;
1354 		bsg_job->job_done(bsg_job);
1355 	}
1356 
1357 	return rval;
1358 }
1359 
1360 static int
1361 qla2x00_optrom_setup(struct fc_bsg_job *bsg_job, scsi_qla_host_t *vha,
1362 	uint8_t is_update)
1363 {
1364 	uint32_t start = 0;
1365 	int valid = 0;
1366 	struct qla_hw_data *ha = vha->hw;
1367 
1368 	if (unlikely(pci_channel_offline(ha->pdev)))
1369 		return -EINVAL;
1370 
1371 	start = bsg_job->request->rqst_data.h_vendor.vendor_cmd[1];
1372 	if (start > ha->optrom_size) {
1373 		ql_log(ql_log_warn, vha, 0x7055,
1374 		    "start %d > optrom_size %d.\n", start, ha->optrom_size);
1375 		return -EINVAL;
1376 	}
1377 
1378 	if (ha->optrom_state != QLA_SWAITING) {
1379 		ql_log(ql_log_info, vha, 0x7056,
1380 		    "optrom_state %d.\n", ha->optrom_state);
1381 		return -EBUSY;
1382 	}
1383 
1384 	ha->optrom_region_start = start;
1385 	ql_dbg(ql_dbg_user, vha, 0x7057, "is_update=%d.\n", is_update);
1386 	if (is_update) {
1387 		if (ha->optrom_size == OPTROM_SIZE_2300 && start == 0)
1388 			valid = 1;
1389 		else if (start == (ha->flt_region_boot * 4) ||
1390 		    start == (ha->flt_region_fw * 4))
1391 			valid = 1;
1392 		else if (IS_QLA24XX_TYPE(ha) || IS_QLA25XX(ha) ||
1393 		    IS_CNA_CAPABLE(ha) || IS_QLA2031(ha) || IS_QLA27XX(ha))
1394 			valid = 1;
1395 		if (!valid) {
1396 			ql_log(ql_log_warn, vha, 0x7058,
1397 			    "Invalid start region 0x%x/0x%x.\n", start,
1398 			    bsg_job->request_payload.payload_len);
1399 			return -EINVAL;
1400 		}
1401 
1402 		ha->optrom_region_size = start +
1403 		    bsg_job->request_payload.payload_len > ha->optrom_size ?
1404 		    ha->optrom_size - start :
1405 		    bsg_job->request_payload.payload_len;
1406 		ha->optrom_state = QLA_SWRITING;
1407 	} else {
1408 		ha->optrom_region_size = start +
1409 		    bsg_job->reply_payload.payload_len > ha->optrom_size ?
1410 		    ha->optrom_size - start :
1411 		    bsg_job->reply_payload.payload_len;
1412 		ha->optrom_state = QLA_SREADING;
1413 	}
1414 
1415 	ha->optrom_buffer = vmalloc(ha->optrom_region_size);
1416 	if (!ha->optrom_buffer) {
1417 		ql_log(ql_log_warn, vha, 0x7059,
1418 		    "Read: Unable to allocate memory for optrom retrieval "
1419 		    "(%x)\n", ha->optrom_region_size);
1420 
1421 		ha->optrom_state = QLA_SWAITING;
1422 		return -ENOMEM;
1423 	}
1424 
1425 	memset(ha->optrom_buffer, 0, ha->optrom_region_size);
1426 	return 0;
1427 }
1428 
1429 static int
1430 qla2x00_read_optrom(struct fc_bsg_job *bsg_job)
1431 {
1432 	struct Scsi_Host *host = bsg_job->shost;
1433 	scsi_qla_host_t *vha = shost_priv(host);
1434 	struct qla_hw_data *ha = vha->hw;
1435 	int rval = 0;
1436 
1437 	if (ha->flags.nic_core_reset_hdlr_active)
1438 		return -EBUSY;
1439 
1440 	mutex_lock(&ha->optrom_mutex);
1441 	rval = qla2x00_optrom_setup(bsg_job, vha, 0);
1442 	if (rval) {
1443 		mutex_unlock(&ha->optrom_mutex);
1444 		return rval;
1445 	}
1446 
1447 	ha->isp_ops->read_optrom(vha, ha->optrom_buffer,
1448 	    ha->optrom_region_start, ha->optrom_region_size);
1449 
1450 	sg_copy_from_buffer(bsg_job->reply_payload.sg_list,
1451 	    bsg_job->reply_payload.sg_cnt, ha->optrom_buffer,
1452 	    ha->optrom_region_size);
1453 
1454 	bsg_job->reply->reply_payload_rcv_len = ha->optrom_region_size;
1455 	bsg_job->reply->result = DID_OK;
1456 	vfree(ha->optrom_buffer);
1457 	ha->optrom_buffer = NULL;
1458 	ha->optrom_state = QLA_SWAITING;
1459 	mutex_unlock(&ha->optrom_mutex);
1460 	bsg_job->job_done(bsg_job);
1461 	return rval;
1462 }
1463 
1464 static int
1465 qla2x00_update_optrom(struct fc_bsg_job *bsg_job)
1466 {
1467 	struct Scsi_Host *host = bsg_job->shost;
1468 	scsi_qla_host_t *vha = shost_priv(host);
1469 	struct qla_hw_data *ha = vha->hw;
1470 	int rval = 0;
1471 
1472 	mutex_lock(&ha->optrom_mutex);
1473 	rval = qla2x00_optrom_setup(bsg_job, vha, 1);
1474 	if (rval) {
1475 		mutex_unlock(&ha->optrom_mutex);
1476 		return rval;
1477 	}
1478 
1479 	/* Set the isp82xx_no_md_cap not to capture minidump */
1480 	ha->flags.isp82xx_no_md_cap = 1;
1481 
1482 	sg_copy_to_buffer(bsg_job->request_payload.sg_list,
1483 	    bsg_job->request_payload.sg_cnt, ha->optrom_buffer,
1484 	    ha->optrom_region_size);
1485 
1486 	ha->isp_ops->write_optrom(vha, ha->optrom_buffer,
1487 	    ha->optrom_region_start, ha->optrom_region_size);
1488 
1489 	bsg_job->reply->result = DID_OK;
1490 	vfree(ha->optrom_buffer);
1491 	ha->optrom_buffer = NULL;
1492 	ha->optrom_state = QLA_SWAITING;
1493 	mutex_unlock(&ha->optrom_mutex);
1494 	bsg_job->job_done(bsg_job);
1495 	return rval;
1496 }
1497 
1498 static int
1499 qla2x00_update_fru_versions(struct fc_bsg_job *bsg_job)
1500 {
1501 	struct Scsi_Host *host = bsg_job->shost;
1502 	scsi_qla_host_t *vha = shost_priv(host);
1503 	struct qla_hw_data *ha = vha->hw;
1504 	int rval = 0;
1505 	uint8_t bsg[DMA_POOL_SIZE];
1506 	struct qla_image_version_list *list = (void *)bsg;
1507 	struct qla_image_version *image;
1508 	uint32_t count;
1509 	dma_addr_t sfp_dma;
1510 	void *sfp = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &sfp_dma);
1511 	if (!sfp) {
1512 		bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] =
1513 		    EXT_STATUS_NO_MEMORY;
1514 		goto done;
1515 	}
1516 
1517 	sg_copy_to_buffer(bsg_job->request_payload.sg_list,
1518 	    bsg_job->request_payload.sg_cnt, list, sizeof(bsg));
1519 
1520 	image = list->version;
1521 	count = list->count;
1522 	while (count--) {
1523 		memcpy(sfp, &image->field_info, sizeof(image->field_info));
1524 		rval = qla2x00_write_sfp(vha, sfp_dma, sfp,
1525 		    image->field_address.device, image->field_address.offset,
1526 		    sizeof(image->field_info), image->field_address.option);
1527 		if (rval) {
1528 			bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] =
1529 			    EXT_STATUS_MAILBOX;
1530 			goto dealloc;
1531 		}
1532 		image++;
1533 	}
1534 
1535 	bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] = 0;
1536 
1537 dealloc:
1538 	dma_pool_free(ha->s_dma_pool, sfp, sfp_dma);
1539 
1540 done:
1541 	bsg_job->reply_len = sizeof(struct fc_bsg_reply);
1542 	bsg_job->reply->result = DID_OK << 16;
1543 	bsg_job->job_done(bsg_job);
1544 
1545 	return 0;
1546 }
1547 
1548 static int
1549 qla2x00_read_fru_status(struct fc_bsg_job *bsg_job)
1550 {
1551 	struct Scsi_Host *host = bsg_job->shost;
1552 	scsi_qla_host_t *vha = shost_priv(host);
1553 	struct qla_hw_data *ha = vha->hw;
1554 	int rval = 0;
1555 	uint8_t bsg[DMA_POOL_SIZE];
1556 	struct qla_status_reg *sr = (void *)bsg;
1557 	dma_addr_t sfp_dma;
1558 	uint8_t *sfp = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &sfp_dma);
1559 	if (!sfp) {
1560 		bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] =
1561 		    EXT_STATUS_NO_MEMORY;
1562 		goto done;
1563 	}
1564 
1565 	sg_copy_to_buffer(bsg_job->request_payload.sg_list,
1566 	    bsg_job->request_payload.sg_cnt, sr, sizeof(*sr));
1567 
1568 	rval = qla2x00_read_sfp(vha, sfp_dma, sfp,
1569 	    sr->field_address.device, sr->field_address.offset,
1570 	    sizeof(sr->status_reg), sr->field_address.option);
1571 	sr->status_reg = *sfp;
1572 
1573 	if (rval) {
1574 		bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] =
1575 		    EXT_STATUS_MAILBOX;
1576 		goto dealloc;
1577 	}
1578 
1579 	sg_copy_from_buffer(bsg_job->reply_payload.sg_list,
1580 	    bsg_job->reply_payload.sg_cnt, sr, sizeof(*sr));
1581 
1582 	bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] = 0;
1583 
1584 dealloc:
1585 	dma_pool_free(ha->s_dma_pool, sfp, sfp_dma);
1586 
1587 done:
1588 	bsg_job->reply_len = sizeof(struct fc_bsg_reply);
1589 	bsg_job->reply->reply_payload_rcv_len = sizeof(*sr);
1590 	bsg_job->reply->result = DID_OK << 16;
1591 	bsg_job->job_done(bsg_job);
1592 
1593 	return 0;
1594 }
1595 
1596 static int
1597 qla2x00_write_fru_status(struct fc_bsg_job *bsg_job)
1598 {
1599 	struct Scsi_Host *host = bsg_job->shost;
1600 	scsi_qla_host_t *vha = shost_priv(host);
1601 	struct qla_hw_data *ha = vha->hw;
1602 	int rval = 0;
1603 	uint8_t bsg[DMA_POOL_SIZE];
1604 	struct qla_status_reg *sr = (void *)bsg;
1605 	dma_addr_t sfp_dma;
1606 	uint8_t *sfp = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &sfp_dma);
1607 	if (!sfp) {
1608 		bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] =
1609 		    EXT_STATUS_NO_MEMORY;
1610 		goto done;
1611 	}
1612 
1613 	sg_copy_to_buffer(bsg_job->request_payload.sg_list,
1614 	    bsg_job->request_payload.sg_cnt, sr, sizeof(*sr));
1615 
1616 	*sfp = sr->status_reg;
1617 	rval = qla2x00_write_sfp(vha, sfp_dma, sfp,
1618 	    sr->field_address.device, sr->field_address.offset,
1619 	    sizeof(sr->status_reg), sr->field_address.option);
1620 
1621 	if (rval) {
1622 		bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] =
1623 		    EXT_STATUS_MAILBOX;
1624 		goto dealloc;
1625 	}
1626 
1627 	bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] = 0;
1628 
1629 dealloc:
1630 	dma_pool_free(ha->s_dma_pool, sfp, sfp_dma);
1631 
1632 done:
1633 	bsg_job->reply_len = sizeof(struct fc_bsg_reply);
1634 	bsg_job->reply->result = DID_OK << 16;
1635 	bsg_job->job_done(bsg_job);
1636 
1637 	return 0;
1638 }
1639 
1640 static int
1641 qla2x00_write_i2c(struct fc_bsg_job *bsg_job)
1642 {
1643 	struct Scsi_Host *host = bsg_job->shost;
1644 	scsi_qla_host_t *vha = shost_priv(host);
1645 	struct qla_hw_data *ha = vha->hw;
1646 	int rval = 0;
1647 	uint8_t bsg[DMA_POOL_SIZE];
1648 	struct qla_i2c_access *i2c = (void *)bsg;
1649 	dma_addr_t sfp_dma;
1650 	uint8_t *sfp = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &sfp_dma);
1651 	if (!sfp) {
1652 		bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] =
1653 		    EXT_STATUS_NO_MEMORY;
1654 		goto done;
1655 	}
1656 
1657 	sg_copy_to_buffer(bsg_job->request_payload.sg_list,
1658 	    bsg_job->request_payload.sg_cnt, i2c, sizeof(*i2c));
1659 
1660 	memcpy(sfp, i2c->buffer, i2c->length);
1661 	rval = qla2x00_write_sfp(vha, sfp_dma, sfp,
1662 	    i2c->device, i2c->offset, i2c->length, i2c->option);
1663 
1664 	if (rval) {
1665 		bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] =
1666 		    EXT_STATUS_MAILBOX;
1667 		goto dealloc;
1668 	}
1669 
1670 	bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] = 0;
1671 
1672 dealloc:
1673 	dma_pool_free(ha->s_dma_pool, sfp, sfp_dma);
1674 
1675 done:
1676 	bsg_job->reply_len = sizeof(struct fc_bsg_reply);
1677 	bsg_job->reply->result = DID_OK << 16;
1678 	bsg_job->job_done(bsg_job);
1679 
1680 	return 0;
1681 }
1682 
1683 static int
1684 qla2x00_read_i2c(struct fc_bsg_job *bsg_job)
1685 {
1686 	struct Scsi_Host *host = bsg_job->shost;
1687 	scsi_qla_host_t *vha = shost_priv(host);
1688 	struct qla_hw_data *ha = vha->hw;
1689 	int rval = 0;
1690 	uint8_t bsg[DMA_POOL_SIZE];
1691 	struct qla_i2c_access *i2c = (void *)bsg;
1692 	dma_addr_t sfp_dma;
1693 	uint8_t *sfp = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &sfp_dma);
1694 	if (!sfp) {
1695 		bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] =
1696 		    EXT_STATUS_NO_MEMORY;
1697 		goto done;
1698 	}
1699 
1700 	sg_copy_to_buffer(bsg_job->request_payload.sg_list,
1701 	    bsg_job->request_payload.sg_cnt, i2c, sizeof(*i2c));
1702 
1703 	rval = qla2x00_read_sfp(vha, sfp_dma, sfp,
1704 		i2c->device, i2c->offset, i2c->length, i2c->option);
1705 
1706 	if (rval) {
1707 		bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] =
1708 		    EXT_STATUS_MAILBOX;
1709 		goto dealloc;
1710 	}
1711 
1712 	memcpy(i2c->buffer, sfp, i2c->length);
1713 	sg_copy_from_buffer(bsg_job->reply_payload.sg_list,
1714 	    bsg_job->reply_payload.sg_cnt, i2c, sizeof(*i2c));
1715 
1716 	bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] = 0;
1717 
1718 dealloc:
1719 	dma_pool_free(ha->s_dma_pool, sfp, sfp_dma);
1720 
1721 done:
1722 	bsg_job->reply_len = sizeof(struct fc_bsg_reply);
1723 	bsg_job->reply->reply_payload_rcv_len = sizeof(*i2c);
1724 	bsg_job->reply->result = DID_OK << 16;
1725 	bsg_job->job_done(bsg_job);
1726 
1727 	return 0;
1728 }
1729 
1730 static int
1731 qla24xx_process_bidir_cmd(struct fc_bsg_job *bsg_job)
1732 {
1733 	struct Scsi_Host *host = bsg_job->shost;
1734 	scsi_qla_host_t *vha = shost_priv(host);
1735 	struct qla_hw_data *ha = vha->hw;
1736 	uint32_t rval = EXT_STATUS_OK;
1737 	uint16_t req_sg_cnt = 0;
1738 	uint16_t rsp_sg_cnt = 0;
1739 	uint16_t nextlid = 0;
1740 	uint32_t tot_dsds;
1741 	srb_t *sp = NULL;
1742 	uint32_t req_data_len = 0;
1743 	uint32_t rsp_data_len = 0;
1744 
1745 	/* Check the type of the adapter */
1746 	if (!IS_BIDI_CAPABLE(ha)) {
1747 		ql_log(ql_log_warn, vha, 0x70a0,
1748 			"This adapter is not supported\n");
1749 		rval = EXT_STATUS_NOT_SUPPORTED;
1750 		goto done;
1751 	}
1752 
1753 	if (test_bit(ISP_ABORT_NEEDED, &vha->dpc_flags) ||
1754 		test_bit(ABORT_ISP_ACTIVE, &vha->dpc_flags) ||
1755 		test_bit(ISP_ABORT_RETRY, &vha->dpc_flags)) {
1756 		rval =  EXT_STATUS_BUSY;
1757 		goto done;
1758 	}
1759 
1760 	/* Check if host is online */
1761 	if (!vha->flags.online) {
1762 		ql_log(ql_log_warn, vha, 0x70a1,
1763 			"Host is not online\n");
1764 		rval = EXT_STATUS_DEVICE_OFFLINE;
1765 		goto done;
1766 	}
1767 
1768 	/* Check if cable is plugged in or not */
1769 	if (vha->device_flags & DFLG_NO_CABLE) {
1770 		ql_log(ql_log_warn, vha, 0x70a2,
1771 			"Cable is unplugged...\n");
1772 		rval = EXT_STATUS_INVALID_CFG;
1773 		goto done;
1774 	}
1775 
1776 	/* Check if the switch is connected or not */
1777 	if (ha->current_topology != ISP_CFG_F) {
1778 		ql_log(ql_log_warn, vha, 0x70a3,
1779 			"Host is not connected to the switch\n");
1780 		rval = EXT_STATUS_INVALID_CFG;
1781 		goto done;
1782 	}
1783 
1784 	/* Check if operating mode is P2P */
1785 	if (ha->operating_mode != P2P) {
1786 		ql_log(ql_log_warn, vha, 0x70a4,
1787 		    "Host is operating mode is not P2p\n");
1788 		rval = EXT_STATUS_INVALID_CFG;
1789 		goto done;
1790 	}
1791 
1792 	mutex_lock(&ha->selflogin_lock);
1793 	if (vha->self_login_loop_id == 0) {
1794 		/* Initialize all required  fields of fcport */
1795 		vha->bidir_fcport.vha = vha;
1796 		vha->bidir_fcport.d_id.b.al_pa = vha->d_id.b.al_pa;
1797 		vha->bidir_fcport.d_id.b.area = vha->d_id.b.area;
1798 		vha->bidir_fcport.d_id.b.domain = vha->d_id.b.domain;
1799 		vha->bidir_fcport.loop_id = vha->loop_id;
1800 
1801 		if (qla2x00_fabric_login(vha, &(vha->bidir_fcport), &nextlid)) {
1802 			ql_log(ql_log_warn, vha, 0x70a7,
1803 			    "Failed to login port %06X for bidirectional IOCB\n",
1804 			    vha->bidir_fcport.d_id.b24);
1805 			mutex_unlock(&ha->selflogin_lock);
1806 			rval = EXT_STATUS_MAILBOX;
1807 			goto done;
1808 		}
1809 		vha->self_login_loop_id = nextlid - 1;
1810 
1811 	}
1812 	/* Assign the self login loop id to fcport */
1813 	mutex_unlock(&ha->selflogin_lock);
1814 
1815 	vha->bidir_fcport.loop_id = vha->self_login_loop_id;
1816 
1817 	req_sg_cnt = dma_map_sg(&ha->pdev->dev,
1818 		bsg_job->request_payload.sg_list,
1819 		bsg_job->request_payload.sg_cnt,
1820 		DMA_TO_DEVICE);
1821 
1822 	if (!req_sg_cnt) {
1823 		rval = EXT_STATUS_NO_MEMORY;
1824 		goto done;
1825 	}
1826 
1827 	rsp_sg_cnt = dma_map_sg(&ha->pdev->dev,
1828 		bsg_job->reply_payload.sg_list, bsg_job->reply_payload.sg_cnt,
1829 		DMA_FROM_DEVICE);
1830 
1831 	if (!rsp_sg_cnt) {
1832 		rval = EXT_STATUS_NO_MEMORY;
1833 		goto done_unmap_req_sg;
1834 	}
1835 
1836 	if ((req_sg_cnt !=  bsg_job->request_payload.sg_cnt) ||
1837 		(rsp_sg_cnt != bsg_job->reply_payload.sg_cnt)) {
1838 		ql_dbg(ql_dbg_user, vha, 0x70a9,
1839 		    "Dma mapping resulted in different sg counts "
1840 		    "[request_sg_cnt: %x dma_request_sg_cnt: %x reply_sg_cnt: "
1841 		    "%x dma_reply_sg_cnt: %x]\n",
1842 		    bsg_job->request_payload.sg_cnt, req_sg_cnt,
1843 		    bsg_job->reply_payload.sg_cnt, rsp_sg_cnt);
1844 		rval = EXT_STATUS_NO_MEMORY;
1845 		goto done_unmap_sg;
1846 	}
1847 
1848 	if (req_data_len != rsp_data_len) {
1849 		rval = EXT_STATUS_BUSY;
1850 		ql_log(ql_log_warn, vha, 0x70aa,
1851 		    "req_data_len != rsp_data_len\n");
1852 		goto done_unmap_sg;
1853 	}
1854 
1855 	req_data_len = bsg_job->request_payload.payload_len;
1856 	rsp_data_len = bsg_job->reply_payload.payload_len;
1857 
1858 
1859 	/* Alloc SRB structure */
1860 	sp = qla2x00_get_sp(vha, &(vha->bidir_fcport), GFP_KERNEL);
1861 	if (!sp) {
1862 		ql_dbg(ql_dbg_user, vha, 0x70ac,
1863 		    "Alloc SRB structure failed\n");
1864 		rval = EXT_STATUS_NO_MEMORY;
1865 		goto done_unmap_sg;
1866 	}
1867 
1868 	/*Populate srb->ctx with bidir ctx*/
1869 	sp->u.bsg_job = bsg_job;
1870 	sp->free = qla2x00_bsg_sp_free;
1871 	sp->type = SRB_BIDI_CMD;
1872 	sp->done = qla2x00_bsg_job_done;
1873 
1874 	/* Add the read and write sg count */
1875 	tot_dsds = rsp_sg_cnt + req_sg_cnt;
1876 
1877 	rval = qla2x00_start_bidir(sp, vha, tot_dsds);
1878 	if (rval != EXT_STATUS_OK)
1879 		goto done_free_srb;
1880 	/* the bsg request  will be completed in the interrupt handler */
1881 	return rval;
1882 
1883 done_free_srb:
1884 	mempool_free(sp, ha->srb_mempool);
1885 done_unmap_sg:
1886 	dma_unmap_sg(&ha->pdev->dev,
1887 	    bsg_job->reply_payload.sg_list,
1888 	    bsg_job->reply_payload.sg_cnt, DMA_FROM_DEVICE);
1889 done_unmap_req_sg:
1890 	dma_unmap_sg(&ha->pdev->dev,
1891 	    bsg_job->request_payload.sg_list,
1892 	    bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
1893 done:
1894 
1895 	/* Return an error vendor specific response
1896 	 * and complete the bsg request
1897 	 */
1898 	bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] = rval;
1899 	bsg_job->reply_len = sizeof(struct fc_bsg_reply);
1900 	bsg_job->reply->reply_payload_rcv_len = 0;
1901 	bsg_job->reply->result = (DID_OK) << 16;
1902 	bsg_job->job_done(bsg_job);
1903 	/* Always return success, vendor rsp carries correct status */
1904 	return 0;
1905 }
1906 
1907 static int
1908 qlafx00_mgmt_cmd(struct fc_bsg_job *bsg_job)
1909 {
1910 	struct Scsi_Host *host = bsg_job->shost;
1911 	scsi_qla_host_t *vha = shost_priv(host);
1912 	struct qla_hw_data *ha = vha->hw;
1913 	int rval = (DRIVER_ERROR << 16);
1914 	struct qla_mt_iocb_rqst_fx00 *piocb_rqst;
1915 	srb_t *sp;
1916 	int req_sg_cnt = 0, rsp_sg_cnt = 0;
1917 	struct fc_port *fcport;
1918 	char  *type = "FC_BSG_HST_FX_MGMT";
1919 
1920 	/* Copy the IOCB specific information */
1921 	piocb_rqst = (struct qla_mt_iocb_rqst_fx00 *)
1922 	    &bsg_job->request->rqst_data.h_vendor.vendor_cmd[1];
1923 
1924 	/* Dump the vendor information */
1925 	ql_dump_buffer(ql_dbg_user + ql_dbg_verbose , vha, 0x70cf,
1926 	    (uint8_t *)piocb_rqst, sizeof(struct qla_mt_iocb_rqst_fx00));
1927 
1928 	if (!vha->flags.online) {
1929 		ql_log(ql_log_warn, vha, 0x70d0,
1930 		    "Host is not online.\n");
1931 		rval = -EIO;
1932 		goto done;
1933 	}
1934 
1935 	if (piocb_rqst->flags & SRB_FXDISC_REQ_DMA_VALID) {
1936 		req_sg_cnt = dma_map_sg(&ha->pdev->dev,
1937 		    bsg_job->request_payload.sg_list,
1938 		    bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
1939 		if (!req_sg_cnt) {
1940 			ql_log(ql_log_warn, vha, 0x70c7,
1941 			    "dma_map_sg return %d for request\n", req_sg_cnt);
1942 			rval = -ENOMEM;
1943 			goto done;
1944 		}
1945 	}
1946 
1947 	if (piocb_rqst->flags & SRB_FXDISC_RESP_DMA_VALID) {
1948 		rsp_sg_cnt = dma_map_sg(&ha->pdev->dev,
1949 		    bsg_job->reply_payload.sg_list,
1950 		    bsg_job->reply_payload.sg_cnt, DMA_FROM_DEVICE);
1951 		if (!rsp_sg_cnt) {
1952 			ql_log(ql_log_warn, vha, 0x70c8,
1953 			    "dma_map_sg return %d for reply\n", rsp_sg_cnt);
1954 			rval = -ENOMEM;
1955 			goto done_unmap_req_sg;
1956 		}
1957 	}
1958 
1959 	ql_dbg(ql_dbg_user, vha, 0x70c9,
1960 	    "request_sg_cnt: %x dma_request_sg_cnt: %x reply_sg_cnt:%x "
1961 	    "dma_reply_sg_cnt: %x\n", bsg_job->request_payload.sg_cnt,
1962 	    req_sg_cnt, bsg_job->reply_payload.sg_cnt, rsp_sg_cnt);
1963 
1964 	/* Allocate a dummy fcport structure, since functions preparing the
1965 	 * IOCB and mailbox command retrieves port specific information
1966 	 * from fcport structure. For Host based ELS commands there will be
1967 	 * no fcport structure allocated
1968 	 */
1969 	fcport = qla2x00_alloc_fcport(vha, GFP_KERNEL);
1970 	if (!fcport) {
1971 		ql_log(ql_log_warn, vha, 0x70ca,
1972 		    "Failed to allocate fcport.\n");
1973 		rval = -ENOMEM;
1974 		goto done_unmap_rsp_sg;
1975 	}
1976 
1977 	/* Alloc SRB structure */
1978 	sp = qla2x00_get_sp(vha, fcport, GFP_KERNEL);
1979 	if (!sp) {
1980 		ql_log(ql_log_warn, vha, 0x70cb,
1981 		    "qla2x00_get_sp failed.\n");
1982 		rval = -ENOMEM;
1983 		goto done_free_fcport;
1984 	}
1985 
1986 	/* Initialize all required  fields of fcport */
1987 	fcport->vha = vha;
1988 	fcport->loop_id = piocb_rqst->dataword;
1989 
1990 	sp->type = SRB_FXIOCB_BCMD;
1991 	sp->name = "bsg_fx_mgmt";
1992 	sp->iocbs = qla24xx_calc_ct_iocbs(req_sg_cnt + rsp_sg_cnt);
1993 	sp->u.bsg_job = bsg_job;
1994 	sp->free = qla2x00_bsg_sp_free;
1995 	sp->done = qla2x00_bsg_job_done;
1996 
1997 	ql_dbg(ql_dbg_user, vha, 0x70cc,
1998 	    "bsg rqst type: %s fx_mgmt_type: %x id=%x\n",
1999 	    type, piocb_rqst->func_type, fcport->loop_id);
2000 
2001 	rval = qla2x00_start_sp(sp);
2002 	if (rval != QLA_SUCCESS) {
2003 		ql_log(ql_log_warn, vha, 0x70cd,
2004 		    "qla2x00_start_sp failed=%d.\n", rval);
2005 		mempool_free(sp, ha->srb_mempool);
2006 		rval = -EIO;
2007 		goto done_free_fcport;
2008 	}
2009 	return rval;
2010 
2011 done_free_fcport:
2012 	kfree(fcport);
2013 
2014 done_unmap_rsp_sg:
2015 	if (piocb_rqst->flags & SRB_FXDISC_RESP_DMA_VALID)
2016 		dma_unmap_sg(&ha->pdev->dev,
2017 		    bsg_job->reply_payload.sg_list,
2018 		    bsg_job->reply_payload.sg_cnt, DMA_FROM_DEVICE);
2019 done_unmap_req_sg:
2020 	if (piocb_rqst->flags & SRB_FXDISC_REQ_DMA_VALID)
2021 		dma_unmap_sg(&ha->pdev->dev,
2022 		    bsg_job->request_payload.sg_list,
2023 		    bsg_job->request_payload.sg_cnt, DMA_TO_DEVICE);
2024 
2025 done:
2026 	return rval;
2027 }
2028 
2029 static int
2030 qla26xx_serdes_op(struct fc_bsg_job *bsg_job)
2031 {
2032 	struct Scsi_Host *host = bsg_job->shost;
2033 	scsi_qla_host_t *vha = shost_priv(host);
2034 	int rval = 0;
2035 	struct qla_serdes_reg sr;
2036 
2037 	memset(&sr, 0, sizeof(sr));
2038 
2039 	sg_copy_to_buffer(bsg_job->request_payload.sg_list,
2040 	    bsg_job->request_payload.sg_cnt, &sr, sizeof(sr));
2041 
2042 	switch (sr.cmd) {
2043 	case INT_SC_SERDES_WRITE_REG:
2044 		rval = qla2x00_write_serdes_word(vha, sr.addr, sr.val);
2045 		bsg_job->reply->reply_payload_rcv_len = 0;
2046 		break;
2047 	case INT_SC_SERDES_READ_REG:
2048 		rval = qla2x00_read_serdes_word(vha, sr.addr, &sr.val);
2049 		sg_copy_from_buffer(bsg_job->reply_payload.sg_list,
2050 		    bsg_job->reply_payload.sg_cnt, &sr, sizeof(sr));
2051 		bsg_job->reply->reply_payload_rcv_len = sizeof(sr);
2052 		break;
2053 	default:
2054 		ql_dbg(ql_dbg_user, vha, 0x708c,
2055 		    "Unknown serdes cmd %x.\n", sr.cmd);
2056 		rval = -EINVAL;
2057 		break;
2058 	}
2059 
2060 	bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] =
2061 	    rval ? EXT_STATUS_MAILBOX : 0;
2062 
2063 	bsg_job->reply_len = sizeof(struct fc_bsg_reply);
2064 	bsg_job->reply->result = DID_OK << 16;
2065 	bsg_job->job_done(bsg_job);
2066 	return 0;
2067 }
2068 
2069 static int
2070 qla8044_serdes_op(struct fc_bsg_job *bsg_job)
2071 {
2072 	struct Scsi_Host *host = bsg_job->shost;
2073 	scsi_qla_host_t *vha = shost_priv(host);
2074 	int rval = 0;
2075 	struct qla_serdes_reg_ex sr;
2076 
2077 	memset(&sr, 0, sizeof(sr));
2078 
2079 	sg_copy_to_buffer(bsg_job->request_payload.sg_list,
2080 	    bsg_job->request_payload.sg_cnt, &sr, sizeof(sr));
2081 
2082 	switch (sr.cmd) {
2083 	case INT_SC_SERDES_WRITE_REG:
2084 		rval = qla8044_write_serdes_word(vha, sr.addr, sr.val);
2085 		bsg_job->reply->reply_payload_rcv_len = 0;
2086 		break;
2087 	case INT_SC_SERDES_READ_REG:
2088 		rval = qla8044_read_serdes_word(vha, sr.addr, &sr.val);
2089 		sg_copy_from_buffer(bsg_job->reply_payload.sg_list,
2090 		    bsg_job->reply_payload.sg_cnt, &sr, sizeof(sr));
2091 		bsg_job->reply->reply_payload_rcv_len = sizeof(sr);
2092 		break;
2093 	default:
2094 		ql_dbg(ql_dbg_user, vha, 0x70cf,
2095 		    "Unknown serdes cmd %x.\n", sr.cmd);
2096 		rval = -EINVAL;
2097 		break;
2098 	}
2099 
2100 	bsg_job->reply->reply_data.vendor_reply.vendor_rsp[0] =
2101 	    rval ? EXT_STATUS_MAILBOX : 0;
2102 
2103 	bsg_job->reply_len = sizeof(struct fc_bsg_reply);
2104 	bsg_job->reply->result = DID_OK << 16;
2105 	bsg_job->job_done(bsg_job);
2106 	return 0;
2107 }
2108 
2109 static int
2110 qla2x00_process_vendor_specific(struct fc_bsg_job *bsg_job)
2111 {
2112 	switch (bsg_job->request->rqst_data.h_vendor.vendor_cmd[0]) {
2113 	case QL_VND_LOOPBACK:
2114 		return qla2x00_process_loopback(bsg_job);
2115 
2116 	case QL_VND_A84_RESET:
2117 		return qla84xx_reset(bsg_job);
2118 
2119 	case QL_VND_A84_UPDATE_FW:
2120 		return qla84xx_updatefw(bsg_job);
2121 
2122 	case QL_VND_A84_MGMT_CMD:
2123 		return qla84xx_mgmt_cmd(bsg_job);
2124 
2125 	case QL_VND_IIDMA:
2126 		return qla24xx_iidma(bsg_job);
2127 
2128 	case QL_VND_FCP_PRIO_CFG_CMD:
2129 		return qla24xx_proc_fcp_prio_cfg_cmd(bsg_job);
2130 
2131 	case QL_VND_READ_FLASH:
2132 		return qla2x00_read_optrom(bsg_job);
2133 
2134 	case QL_VND_UPDATE_FLASH:
2135 		return qla2x00_update_optrom(bsg_job);
2136 
2137 	case QL_VND_SET_FRU_VERSION:
2138 		return qla2x00_update_fru_versions(bsg_job);
2139 
2140 	case QL_VND_READ_FRU_STATUS:
2141 		return qla2x00_read_fru_status(bsg_job);
2142 
2143 	case QL_VND_WRITE_FRU_STATUS:
2144 		return qla2x00_write_fru_status(bsg_job);
2145 
2146 	case QL_VND_WRITE_I2C:
2147 		return qla2x00_write_i2c(bsg_job);
2148 
2149 	case QL_VND_READ_I2C:
2150 		return qla2x00_read_i2c(bsg_job);
2151 
2152 	case QL_VND_DIAG_IO_CMD:
2153 		return qla24xx_process_bidir_cmd(bsg_job);
2154 
2155 	case QL_VND_FX00_MGMT_CMD:
2156 		return qlafx00_mgmt_cmd(bsg_job);
2157 
2158 	case QL_VND_SERDES_OP:
2159 		return qla26xx_serdes_op(bsg_job);
2160 
2161 	case QL_VND_SERDES_OP_EX:
2162 		return qla8044_serdes_op(bsg_job);
2163 
2164 	default:
2165 		return -ENOSYS;
2166 	}
2167 }
2168 
2169 int
2170 qla24xx_bsg_request(struct fc_bsg_job *bsg_job)
2171 {
2172 	int ret = -EINVAL;
2173 	struct fc_rport *rport;
2174 	struct Scsi_Host *host;
2175 	scsi_qla_host_t *vha;
2176 
2177 	/* In case no data transferred. */
2178 	bsg_job->reply->reply_payload_rcv_len = 0;
2179 
2180 	if (bsg_job->request->msgcode == FC_BSG_RPT_ELS) {
2181 		rport = bsg_job->rport;
2182 		host = rport_to_shost(rport);
2183 		vha = shost_priv(host);
2184 	} else {
2185 		host = bsg_job->shost;
2186 		vha = shost_priv(host);
2187 	}
2188 
2189 	if (qla2x00_reset_active(vha)) {
2190 		ql_dbg(ql_dbg_user, vha, 0x709f,
2191 		    "BSG: ISP abort active/needed -- cmd=%d.\n",
2192 		    bsg_job->request->msgcode);
2193 		return -EBUSY;
2194 	}
2195 
2196 	ql_dbg(ql_dbg_user, vha, 0x7000,
2197 	    "Entered %s msgcode=0x%x.\n", __func__, bsg_job->request->msgcode);
2198 
2199 	switch (bsg_job->request->msgcode) {
2200 	case FC_BSG_RPT_ELS:
2201 	case FC_BSG_HST_ELS_NOLOGIN:
2202 		ret = qla2x00_process_els(bsg_job);
2203 		break;
2204 	case FC_BSG_HST_CT:
2205 		ret = qla2x00_process_ct(bsg_job);
2206 		break;
2207 	case FC_BSG_HST_VENDOR:
2208 		ret = qla2x00_process_vendor_specific(bsg_job);
2209 		break;
2210 	case FC_BSG_HST_ADD_RPORT:
2211 	case FC_BSG_HST_DEL_RPORT:
2212 	case FC_BSG_RPT_CT:
2213 	default:
2214 		ql_log(ql_log_warn, vha, 0x705a, "Unsupported BSG request.\n");
2215 		break;
2216 	}
2217 	return ret;
2218 }
2219 
2220 int
2221 qla24xx_bsg_timeout(struct fc_bsg_job *bsg_job)
2222 {
2223 	scsi_qla_host_t *vha = shost_priv(bsg_job->shost);
2224 	struct qla_hw_data *ha = vha->hw;
2225 	srb_t *sp;
2226 	int cnt, que;
2227 	unsigned long flags;
2228 	struct req_que *req;
2229 
2230 	/* find the bsg job from the active list of commands */
2231 	spin_lock_irqsave(&ha->hardware_lock, flags);
2232 	for (que = 0; que < ha->max_req_queues; que++) {
2233 		req = ha->req_q_map[que];
2234 		if (!req)
2235 			continue;
2236 
2237 		for (cnt = 1; cnt < req->num_outstanding_cmds; cnt++) {
2238 			sp = req->outstanding_cmds[cnt];
2239 			if (sp) {
2240 				if (((sp->type == SRB_CT_CMD) ||
2241 					(sp->type == SRB_ELS_CMD_HST) ||
2242 					(sp->type == SRB_FXIOCB_BCMD))
2243 					&& (sp->u.bsg_job == bsg_job)) {
2244 					req->outstanding_cmds[cnt] = NULL;
2245 					spin_unlock_irqrestore(&ha->hardware_lock, flags);
2246 					if (ha->isp_ops->abort_command(sp)) {
2247 						ql_log(ql_log_warn, vha, 0x7089,
2248 						    "mbx abort_command "
2249 						    "failed.\n");
2250 						bsg_job->req->errors =
2251 						bsg_job->reply->result = -EIO;
2252 					} else {
2253 						ql_dbg(ql_dbg_user, vha, 0x708a,
2254 						    "mbx abort_command "
2255 						    "success.\n");
2256 						bsg_job->req->errors =
2257 						bsg_job->reply->result = 0;
2258 					}
2259 					spin_lock_irqsave(&ha->hardware_lock, flags);
2260 					goto done;
2261 				}
2262 			}
2263 		}
2264 	}
2265 	spin_unlock_irqrestore(&ha->hardware_lock, flags);
2266 	ql_log(ql_log_info, vha, 0x708b, "SRB not found to abort.\n");
2267 	bsg_job->req->errors = bsg_job->reply->result = -ENXIO;
2268 	return 0;
2269 
2270 done:
2271 	spin_unlock_irqrestore(&ha->hardware_lock, flags);
2272 	sp->free(vha, sp);
2273 	return 0;
2274 }
2275