xref: /linux/drivers/s390/scsi/zfcp_dbf.c (revision f3a8b6645dc2e60d11f20c1c23afd964ff4e55ae)
1 /*
2  * zfcp device driver
3  *
4  * Debug traces for zfcp.
5  *
6  * Copyright IBM Corp. 2002, 2016
7  */
8 
9 #define KMSG_COMPONENT "zfcp"
10 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
11 
12 #include <linux/module.h>
13 #include <linux/ctype.h>
14 #include <linux/slab.h>
15 #include <asm/debug.h>
16 #include "zfcp_dbf.h"
17 #include "zfcp_ext.h"
18 #include "zfcp_fc.h"
19 
20 static u32 dbfsize = 4;
21 
22 module_param(dbfsize, uint, 0400);
23 MODULE_PARM_DESC(dbfsize,
24 		 "number of pages for each debug feature area (default 4)");
25 
26 static u32 dbflevel = 3;
27 
28 module_param(dbflevel, uint, 0400);
29 MODULE_PARM_DESC(dbflevel,
30 		 "log level for each debug feature area "
31 		 "(default 3, range 0..6)");
32 
33 static inline unsigned int zfcp_dbf_plen(unsigned int offset)
34 {
35 	return sizeof(struct zfcp_dbf_pay) + offset - ZFCP_DBF_PAY_MAX_REC;
36 }
37 
38 static inline
39 void zfcp_dbf_pl_write(struct zfcp_dbf *dbf, void *data, u16 length, char *area,
40 		       u64 req_id)
41 {
42 	struct zfcp_dbf_pay *pl = &dbf->pay_buf;
43 	u16 offset = 0, rec_length;
44 
45 	spin_lock(&dbf->pay_lock);
46 	memset(pl, 0, sizeof(*pl));
47 	pl->fsf_req_id = req_id;
48 	memcpy(pl->area, area, ZFCP_DBF_TAG_LEN);
49 
50 	while (offset < length) {
51 		rec_length = min((u16) ZFCP_DBF_PAY_MAX_REC,
52 				 (u16) (length - offset));
53 		memcpy(pl->data, data + offset, rec_length);
54 		debug_event(dbf->pay, 1, pl, zfcp_dbf_plen(rec_length));
55 
56 		offset += rec_length;
57 		pl->counter++;
58 	}
59 
60 	spin_unlock(&dbf->pay_lock);
61 }
62 
63 /**
64  * zfcp_dbf_hba_fsf_res - trace event for fsf responses
65  * @tag: tag indicating which kind of unsolicited status has been received
66  * @req: request for which a response was received
67  */
68 void zfcp_dbf_hba_fsf_res(char *tag, int level, struct zfcp_fsf_req *req)
69 {
70 	struct zfcp_dbf *dbf = req->adapter->dbf;
71 	struct fsf_qtcb_prefix *q_pref = &req->qtcb->prefix;
72 	struct fsf_qtcb_header *q_head = &req->qtcb->header;
73 	struct zfcp_dbf_hba *rec = &dbf->hba_buf;
74 	unsigned long flags;
75 
76 	spin_lock_irqsave(&dbf->hba_lock, flags);
77 	memset(rec, 0, sizeof(*rec));
78 
79 	memcpy(rec->tag, tag, ZFCP_DBF_TAG_LEN);
80 	rec->id = ZFCP_DBF_HBA_RES;
81 	rec->fsf_req_id = req->req_id;
82 	rec->fsf_req_status = req->status;
83 	rec->fsf_cmd = req->fsf_command;
84 	rec->fsf_seq_no = req->seq_no;
85 	rec->u.res.req_issued = req->issued;
86 	rec->u.res.prot_status = q_pref->prot_status;
87 	rec->u.res.fsf_status = q_head->fsf_status;
88 	rec->u.res.port_handle = q_head->port_handle;
89 	rec->u.res.lun_handle = q_head->lun_handle;
90 
91 	memcpy(rec->u.res.prot_status_qual, &q_pref->prot_status_qual,
92 	       FSF_PROT_STATUS_QUAL_SIZE);
93 	memcpy(rec->u.res.fsf_status_qual, &q_head->fsf_status_qual,
94 	       FSF_STATUS_QUALIFIER_SIZE);
95 
96 	if (req->fsf_command != FSF_QTCB_FCP_CMND) {
97 		rec->pl_len = q_head->log_length;
98 		zfcp_dbf_pl_write(dbf, (char *)q_pref + q_head->log_start,
99 				  rec->pl_len, "fsf_res", req->req_id);
100 	}
101 
102 	debug_event(dbf->hba, level, rec, sizeof(*rec));
103 	spin_unlock_irqrestore(&dbf->hba_lock, flags);
104 }
105 
106 /**
107  * zfcp_dbf_hba_fsf_uss - trace event for an unsolicited status buffer
108  * @tag: tag indicating which kind of unsolicited status has been received
109  * @req: request providing the unsolicited status
110  */
111 void zfcp_dbf_hba_fsf_uss(char *tag, struct zfcp_fsf_req *req)
112 {
113 	struct zfcp_dbf *dbf = req->adapter->dbf;
114 	struct fsf_status_read_buffer *srb = req->data;
115 	struct zfcp_dbf_hba *rec = &dbf->hba_buf;
116 	unsigned long flags;
117 
118 	spin_lock_irqsave(&dbf->hba_lock, flags);
119 	memset(rec, 0, sizeof(*rec));
120 
121 	memcpy(rec->tag, tag, ZFCP_DBF_TAG_LEN);
122 	rec->id = ZFCP_DBF_HBA_USS;
123 	rec->fsf_req_id = req->req_id;
124 	rec->fsf_req_status = req->status;
125 	rec->fsf_cmd = req->fsf_command;
126 
127 	if (!srb)
128 		goto log;
129 
130 	rec->u.uss.status_type = srb->status_type;
131 	rec->u.uss.status_subtype = srb->status_subtype;
132 	rec->u.uss.d_id = ntoh24(srb->d_id);
133 	rec->u.uss.lun = srb->fcp_lun;
134 	memcpy(&rec->u.uss.queue_designator, &srb->queue_designator,
135 	       sizeof(rec->u.uss.queue_designator));
136 
137 	/* status read buffer payload length */
138 	rec->pl_len = (!srb->length) ? 0 : srb->length -
139 			offsetof(struct fsf_status_read_buffer, payload);
140 
141 	if (rec->pl_len)
142 		zfcp_dbf_pl_write(dbf, srb->payload.data, rec->pl_len,
143 				  "fsf_uss", req->req_id);
144 log:
145 	debug_event(dbf->hba, 2, rec, sizeof(*rec));
146 	spin_unlock_irqrestore(&dbf->hba_lock, flags);
147 }
148 
149 /**
150  * zfcp_dbf_hba_bit_err - trace event for bit error conditions
151  * @tag: tag indicating which kind of unsolicited status has been received
152  * @req: request which caused the bit_error condition
153  */
154 void zfcp_dbf_hba_bit_err(char *tag, struct zfcp_fsf_req *req)
155 {
156 	struct zfcp_dbf *dbf = req->adapter->dbf;
157 	struct zfcp_dbf_hba *rec = &dbf->hba_buf;
158 	struct fsf_status_read_buffer *sr_buf = req->data;
159 	unsigned long flags;
160 
161 	spin_lock_irqsave(&dbf->hba_lock, flags);
162 	memset(rec, 0, sizeof(*rec));
163 
164 	memcpy(rec->tag, tag, ZFCP_DBF_TAG_LEN);
165 	rec->id = ZFCP_DBF_HBA_BIT;
166 	rec->fsf_req_id = req->req_id;
167 	rec->fsf_req_status = req->status;
168 	rec->fsf_cmd = req->fsf_command;
169 	memcpy(&rec->u.be, &sr_buf->payload.bit_error,
170 	       sizeof(struct fsf_bit_error_payload));
171 
172 	debug_event(dbf->hba, 1, rec, sizeof(*rec));
173 	spin_unlock_irqrestore(&dbf->hba_lock, flags);
174 }
175 
176 /**
177  * zfcp_dbf_hba_def_err - trace event for deferred error messages
178  * @adapter: pointer to struct zfcp_adapter
179  * @req_id: request id which caused the deferred error message
180  * @scount: number of sbals incl. the signaling sbal
181  * @pl: array of all involved sbals
182  */
183 void zfcp_dbf_hba_def_err(struct zfcp_adapter *adapter, u64 req_id, u16 scount,
184 			  void **pl)
185 {
186 	struct zfcp_dbf *dbf = adapter->dbf;
187 	struct zfcp_dbf_pay *payload = &dbf->pay_buf;
188 	unsigned long flags;
189 	u16 length;
190 
191 	if (!pl)
192 		return;
193 
194 	spin_lock_irqsave(&dbf->pay_lock, flags);
195 	memset(payload, 0, sizeof(*payload));
196 
197 	memcpy(payload->area, "def_err", 7);
198 	payload->fsf_req_id = req_id;
199 	payload->counter = 0;
200 	length = min((u16)sizeof(struct qdio_buffer),
201 		     (u16)ZFCP_DBF_PAY_MAX_REC);
202 
203 	while (payload->counter < scount && (char *)pl[payload->counter]) {
204 		memcpy(payload->data, (char *)pl[payload->counter], length);
205 		debug_event(dbf->pay, 1, payload, zfcp_dbf_plen(length));
206 		payload->counter++;
207 	}
208 
209 	spin_unlock_irqrestore(&dbf->pay_lock, flags);
210 }
211 
212 /**
213  * zfcp_dbf_hba_basic - trace event for basic adapter events
214  * @adapter: pointer to struct zfcp_adapter
215  */
216 void zfcp_dbf_hba_basic(char *tag, struct zfcp_adapter *adapter)
217 {
218 	struct zfcp_dbf *dbf = adapter->dbf;
219 	struct zfcp_dbf_hba *rec = &dbf->hba_buf;
220 	unsigned long flags;
221 
222 	spin_lock_irqsave(&dbf->hba_lock, flags);
223 	memset(rec, 0, sizeof(*rec));
224 
225 	memcpy(rec->tag, tag, ZFCP_DBF_TAG_LEN);
226 	rec->id = ZFCP_DBF_HBA_BASIC;
227 
228 	debug_event(dbf->hba, 1, rec, sizeof(*rec));
229 	spin_unlock_irqrestore(&dbf->hba_lock, flags);
230 }
231 
232 static void zfcp_dbf_set_common(struct zfcp_dbf_rec *rec,
233 				struct zfcp_adapter *adapter,
234 				struct zfcp_port *port,
235 				struct scsi_device *sdev)
236 {
237 	rec->adapter_status = atomic_read(&adapter->status);
238 	if (port) {
239 		rec->port_status = atomic_read(&port->status);
240 		rec->wwpn = port->wwpn;
241 		rec->d_id = port->d_id;
242 	}
243 	if (sdev) {
244 		rec->lun_status = atomic_read(&sdev_to_zfcp(sdev)->status);
245 		rec->lun = zfcp_scsi_dev_lun(sdev);
246 	} else
247 		rec->lun = ZFCP_DBF_INVALID_LUN;
248 }
249 
250 /**
251  * zfcp_dbf_rec_trig - trace event related to triggered recovery
252  * @tag: identifier for event
253  * @adapter: adapter on which the erp_action should run
254  * @port: remote port involved in the erp_action
255  * @sdev: scsi device involved in the erp_action
256  * @want: wanted erp_action
257  * @need: required erp_action
258  *
259  * The adapter->erp_lock has to be held.
260  */
261 void zfcp_dbf_rec_trig(char *tag, struct zfcp_adapter *adapter,
262 		       struct zfcp_port *port, struct scsi_device *sdev,
263 		       u8 want, u8 need)
264 {
265 	struct zfcp_dbf *dbf = adapter->dbf;
266 	struct zfcp_dbf_rec *rec = &dbf->rec_buf;
267 	struct list_head *entry;
268 	unsigned long flags;
269 
270 	spin_lock_irqsave(&dbf->rec_lock, flags);
271 	memset(rec, 0, sizeof(*rec));
272 
273 	rec->id = ZFCP_DBF_REC_TRIG;
274 	memcpy(rec->tag, tag, ZFCP_DBF_TAG_LEN);
275 	zfcp_dbf_set_common(rec, adapter, port, sdev);
276 
277 	list_for_each(entry, &adapter->erp_ready_head)
278 		rec->u.trig.ready++;
279 
280 	list_for_each(entry, &adapter->erp_running_head)
281 		rec->u.trig.running++;
282 
283 	rec->u.trig.want = want;
284 	rec->u.trig.need = need;
285 
286 	debug_event(dbf->rec, 1, rec, sizeof(*rec));
287 	spin_unlock_irqrestore(&dbf->rec_lock, flags);
288 }
289 
290 
291 /**
292  * zfcp_dbf_rec_run - trace event related to running recovery
293  * @tag: identifier for event
294  * @erp: erp_action running
295  */
296 void zfcp_dbf_rec_run(char *tag, struct zfcp_erp_action *erp)
297 {
298 	struct zfcp_dbf *dbf = erp->adapter->dbf;
299 	struct zfcp_dbf_rec *rec = &dbf->rec_buf;
300 	unsigned long flags;
301 
302 	spin_lock_irqsave(&dbf->rec_lock, flags);
303 	memset(rec, 0, sizeof(*rec));
304 
305 	rec->id = ZFCP_DBF_REC_RUN;
306 	memcpy(rec->tag, tag, ZFCP_DBF_TAG_LEN);
307 	zfcp_dbf_set_common(rec, erp->adapter, erp->port, erp->sdev);
308 
309 	rec->u.run.fsf_req_id = erp->fsf_req_id;
310 	rec->u.run.rec_status = erp->status;
311 	rec->u.run.rec_step = erp->step;
312 	rec->u.run.rec_action = erp->action;
313 
314 	if (erp->sdev)
315 		rec->u.run.rec_count =
316 			atomic_read(&sdev_to_zfcp(erp->sdev)->erp_counter);
317 	else if (erp->port)
318 		rec->u.run.rec_count = atomic_read(&erp->port->erp_counter);
319 	else
320 		rec->u.run.rec_count = atomic_read(&erp->adapter->erp_counter);
321 
322 	debug_event(dbf->rec, 1, rec, sizeof(*rec));
323 	spin_unlock_irqrestore(&dbf->rec_lock, flags);
324 }
325 
326 /**
327  * zfcp_dbf_rec_run_wka - trace wka port event with info like running recovery
328  * @tag: identifier for event
329  * @wka_port: well known address port
330  * @req_id: request ID to correlate with potential HBA trace record
331  */
332 void zfcp_dbf_rec_run_wka(char *tag, struct zfcp_fc_wka_port *wka_port,
333 			  u64 req_id)
334 {
335 	struct zfcp_dbf *dbf = wka_port->adapter->dbf;
336 	struct zfcp_dbf_rec *rec = &dbf->rec_buf;
337 	unsigned long flags;
338 
339 	spin_lock_irqsave(&dbf->rec_lock, flags);
340 	memset(rec, 0, sizeof(*rec));
341 
342 	rec->id = ZFCP_DBF_REC_RUN;
343 	memcpy(rec->tag, tag, ZFCP_DBF_TAG_LEN);
344 	rec->port_status = wka_port->status;
345 	rec->d_id = wka_port->d_id;
346 	rec->lun = ZFCP_DBF_INVALID_LUN;
347 
348 	rec->u.run.fsf_req_id = req_id;
349 	rec->u.run.rec_status = ~0;
350 	rec->u.run.rec_step = ~0;
351 	rec->u.run.rec_action = ~0;
352 	rec->u.run.rec_count = ~0;
353 
354 	debug_event(dbf->rec, 1, rec, sizeof(*rec));
355 	spin_unlock_irqrestore(&dbf->rec_lock, flags);
356 }
357 
358 static inline
359 void zfcp_dbf_san(char *tag, struct zfcp_dbf *dbf,
360 		  char *paytag, struct scatterlist *sg, u8 id, u16 len,
361 		  u64 req_id, u32 d_id, u16 cap_len)
362 {
363 	struct zfcp_dbf_san *rec = &dbf->san_buf;
364 	u16 rec_len;
365 	unsigned long flags;
366 	struct zfcp_dbf_pay *payload = &dbf->pay_buf;
367 	u16 pay_sum = 0;
368 
369 	spin_lock_irqsave(&dbf->san_lock, flags);
370 	memset(rec, 0, sizeof(*rec));
371 
372 	rec->id = id;
373 	rec->fsf_req_id = req_id;
374 	rec->d_id = d_id;
375 	memcpy(rec->tag, tag, ZFCP_DBF_TAG_LEN);
376 	rec->pl_len = len; /* full length even if we cap pay below */
377 	if (!sg)
378 		goto out;
379 	rec_len = min_t(unsigned int, sg->length, ZFCP_DBF_SAN_MAX_PAYLOAD);
380 	memcpy(rec->payload, sg_virt(sg), rec_len); /* part of 1st sg entry */
381 	if (len <= rec_len)
382 		goto out; /* skip pay record if full content in rec->payload */
383 
384 	/* if (len > rec_len):
385 	 * dump data up to cap_len ignoring small duplicate in rec->payload
386 	 */
387 	spin_lock(&dbf->pay_lock);
388 	memset(payload, 0, sizeof(*payload));
389 	memcpy(payload->area, paytag, ZFCP_DBF_TAG_LEN);
390 	payload->fsf_req_id = req_id;
391 	payload->counter = 0;
392 	for (; sg && pay_sum < cap_len; sg = sg_next(sg)) {
393 		u16 pay_len, offset = 0;
394 
395 		while (offset < sg->length && pay_sum < cap_len) {
396 			pay_len = min((u16)ZFCP_DBF_PAY_MAX_REC,
397 				      (u16)(sg->length - offset));
398 			/* cap_len <= pay_sum < cap_len+ZFCP_DBF_PAY_MAX_REC */
399 			memcpy(payload->data, sg_virt(sg) + offset, pay_len);
400 			debug_event(dbf->pay, 1, payload,
401 				    zfcp_dbf_plen(pay_len));
402 			payload->counter++;
403 			offset += pay_len;
404 			pay_sum += pay_len;
405 		}
406 	}
407 	spin_unlock(&dbf->pay_lock);
408 
409 out:
410 	debug_event(dbf->san, 1, rec, sizeof(*rec));
411 	spin_unlock_irqrestore(&dbf->san_lock, flags);
412 }
413 
414 /**
415  * zfcp_dbf_san_req - trace event for issued SAN request
416  * @tag: identifier for event
417  * @fsf_req: request containing issued CT data
418  * d_id: destination ID
419  */
420 void zfcp_dbf_san_req(char *tag, struct zfcp_fsf_req *fsf, u32 d_id)
421 {
422 	struct zfcp_dbf *dbf = fsf->adapter->dbf;
423 	struct zfcp_fsf_ct_els *ct_els = fsf->data;
424 	u16 length;
425 
426 	length = (u16)zfcp_qdio_real_bytes(ct_els->req);
427 	zfcp_dbf_san(tag, dbf, "san_req", ct_els->req, ZFCP_DBF_SAN_REQ,
428 		     length, fsf->req_id, d_id, length);
429 }
430 
431 static u16 zfcp_dbf_san_res_cap_len_if_gpn_ft(char *tag,
432 					      struct zfcp_fsf_req *fsf,
433 					      u16 len)
434 {
435 	struct zfcp_fsf_ct_els *ct_els = fsf->data;
436 	struct fc_ct_hdr *reqh = sg_virt(ct_els->req);
437 	struct fc_ns_gid_ft *reqn = (struct fc_ns_gid_ft *)(reqh + 1);
438 	struct scatterlist *resp_entry = ct_els->resp;
439 	struct fc_gpn_ft_resp *acc;
440 	int max_entries, x, last = 0;
441 
442 	if (!(memcmp(tag, "fsscth2", 7) == 0
443 	      && ct_els->d_id == FC_FID_DIR_SERV
444 	      && reqh->ct_rev == FC_CT_REV
445 	      && reqh->ct_in_id[0] == 0
446 	      && reqh->ct_in_id[1] == 0
447 	      && reqh->ct_in_id[2] == 0
448 	      && reqh->ct_fs_type == FC_FST_DIR
449 	      && reqh->ct_fs_subtype == FC_NS_SUBTYPE
450 	      && reqh->ct_options == 0
451 	      && reqh->_ct_resvd1 == 0
452 	      && reqh->ct_cmd == FC_NS_GPN_FT
453 	      /* reqh->ct_mr_size can vary so do not match but read below */
454 	      && reqh->_ct_resvd2 == 0
455 	      && reqh->ct_reason == 0
456 	      && reqh->ct_explan == 0
457 	      && reqh->ct_vendor == 0
458 	      && reqn->fn_resvd == 0
459 	      && reqn->fn_domain_id_scope == 0
460 	      && reqn->fn_area_id_scope == 0
461 	      && reqn->fn_fc4_type == FC_TYPE_FCP))
462 		return len; /* not GPN_FT response so do not cap */
463 
464 	acc = sg_virt(resp_entry);
465 	max_entries = (reqh->ct_mr_size * 4 / sizeof(struct fc_gpn_ft_resp))
466 		+ 1 /* zfcp_fc_scan_ports: bytes correct, entries off-by-one
467 		     * to account for header as 1st pseudo "entry" */;
468 
469 	/* the basic CT_IU preamble is the same size as one entry in the GPN_FT
470 	 * response, allowing us to skip special handling for it - just skip it
471 	 */
472 	for (x = 1; x < max_entries && !last; x++) {
473 		if (x % (ZFCP_FC_GPN_FT_ENT_PAGE + 1))
474 			acc++;
475 		else
476 			acc = sg_virt(++resp_entry);
477 
478 		last = acc->fp_flags & FC_NS_FID_LAST;
479 	}
480 	len = min(len, (u16)(x * sizeof(struct fc_gpn_ft_resp)));
481 	return len; /* cap after last entry */
482 }
483 
484 /**
485  * zfcp_dbf_san_res - trace event for received SAN request
486  * @tag: identifier for event
487  * @fsf_req: request containing issued CT data
488  */
489 void zfcp_dbf_san_res(char *tag, struct zfcp_fsf_req *fsf)
490 {
491 	struct zfcp_dbf *dbf = fsf->adapter->dbf;
492 	struct zfcp_fsf_ct_els *ct_els = fsf->data;
493 	u16 length;
494 
495 	length = (u16)zfcp_qdio_real_bytes(ct_els->resp);
496 	zfcp_dbf_san(tag, dbf, "san_res", ct_els->resp, ZFCP_DBF_SAN_RES,
497 		     length, fsf->req_id, ct_els->d_id,
498 		     zfcp_dbf_san_res_cap_len_if_gpn_ft(tag, fsf, length));
499 }
500 
501 /**
502  * zfcp_dbf_san_in_els - trace event for incoming ELS
503  * @tag: identifier for event
504  * @fsf_req: request containing issued CT data
505  */
506 void zfcp_dbf_san_in_els(char *tag, struct zfcp_fsf_req *fsf)
507 {
508 	struct zfcp_dbf *dbf = fsf->adapter->dbf;
509 	struct fsf_status_read_buffer *srb =
510 		(struct fsf_status_read_buffer *) fsf->data;
511 	u16 length;
512 	struct scatterlist sg;
513 
514 	length = (u16)(srb->length -
515 			offsetof(struct fsf_status_read_buffer, payload));
516 	sg_init_one(&sg, srb->payload.data, length);
517 	zfcp_dbf_san(tag, dbf, "san_els", &sg, ZFCP_DBF_SAN_ELS, length,
518 		     fsf->req_id, ntoh24(srb->d_id), length);
519 }
520 
521 /**
522  * zfcp_dbf_scsi - trace event for scsi commands
523  * @tag: identifier for event
524  * @sc: pointer to struct scsi_cmnd
525  * @fsf: pointer to struct zfcp_fsf_req
526  */
527 void zfcp_dbf_scsi(char *tag, int level, struct scsi_cmnd *sc,
528 		   struct zfcp_fsf_req *fsf)
529 {
530 	struct zfcp_adapter *adapter =
531 		(struct zfcp_adapter *) sc->device->host->hostdata[0];
532 	struct zfcp_dbf *dbf = adapter->dbf;
533 	struct zfcp_dbf_scsi *rec = &dbf->scsi_buf;
534 	struct fcp_resp_with_ext *fcp_rsp;
535 	struct fcp_resp_rsp_info *fcp_rsp_info;
536 	unsigned long flags;
537 
538 	spin_lock_irqsave(&dbf->scsi_lock, flags);
539 	memset(rec, 0, sizeof(*rec));
540 
541 	memcpy(rec->tag, tag, ZFCP_DBF_TAG_LEN);
542 	rec->id = ZFCP_DBF_SCSI_CMND;
543 	rec->scsi_result = sc->result;
544 	rec->scsi_retries = sc->retries;
545 	rec->scsi_allowed = sc->allowed;
546 	rec->scsi_id = sc->device->id;
547 	/* struct zfcp_dbf_scsi needs to be updated to handle 64bit LUNs */
548 	rec->scsi_lun = (u32)sc->device->lun;
549 	rec->host_scribble = (unsigned long)sc->host_scribble;
550 
551 	memcpy(rec->scsi_opcode, sc->cmnd,
552 	       min((int)sc->cmd_len, ZFCP_DBF_SCSI_OPCODE));
553 
554 	if (fsf) {
555 		rec->fsf_req_id = fsf->req_id;
556 		fcp_rsp = (struct fcp_resp_with_ext *)
557 				&(fsf->qtcb->bottom.io.fcp_rsp);
558 		memcpy(&rec->fcp_rsp, fcp_rsp, FCP_RESP_WITH_EXT);
559 		if (fcp_rsp->resp.fr_flags & FCP_RSP_LEN_VAL) {
560 			fcp_rsp_info = (struct fcp_resp_rsp_info *) &fcp_rsp[1];
561 			rec->fcp_rsp_info = fcp_rsp_info->rsp_code;
562 		}
563 		if (fcp_rsp->resp.fr_flags & FCP_SNS_LEN_VAL) {
564 			rec->pl_len = min((u16)SCSI_SENSE_BUFFERSIZE,
565 					  (u16)ZFCP_DBF_PAY_MAX_REC);
566 			zfcp_dbf_pl_write(dbf, sc->sense_buffer, rec->pl_len,
567 					  "fcp_sns", fsf->req_id);
568 		}
569 	}
570 
571 	debug_event(dbf->scsi, level, rec, sizeof(*rec));
572 	spin_unlock_irqrestore(&dbf->scsi_lock, flags);
573 }
574 
575 static debug_info_t *zfcp_dbf_reg(const char *name, int size, int rec_size)
576 {
577 	struct debug_info *d;
578 
579 	d = debug_register(name, size, 1, rec_size);
580 	if (!d)
581 		return NULL;
582 
583 	debug_register_view(d, &debug_hex_ascii_view);
584 	debug_set_level(d, dbflevel);
585 
586 	return d;
587 }
588 
589 static void zfcp_dbf_unregister(struct zfcp_dbf *dbf)
590 {
591 	if (!dbf)
592 		return;
593 
594 	debug_unregister(dbf->scsi);
595 	debug_unregister(dbf->san);
596 	debug_unregister(dbf->hba);
597 	debug_unregister(dbf->pay);
598 	debug_unregister(dbf->rec);
599 	kfree(dbf);
600 }
601 
602 /**
603  * zfcp_adapter_debug_register - registers debug feature for an adapter
604  * @adapter: pointer to adapter for which debug features should be registered
605  * return: -ENOMEM on error, 0 otherwise
606  */
607 int zfcp_dbf_adapter_register(struct zfcp_adapter *adapter)
608 {
609 	char name[DEBUG_MAX_NAME_LEN];
610 	struct zfcp_dbf *dbf;
611 
612 	dbf = kzalloc(sizeof(struct zfcp_dbf), GFP_KERNEL);
613 	if (!dbf)
614 		return -ENOMEM;
615 
616 	spin_lock_init(&dbf->pay_lock);
617 	spin_lock_init(&dbf->hba_lock);
618 	spin_lock_init(&dbf->san_lock);
619 	spin_lock_init(&dbf->scsi_lock);
620 	spin_lock_init(&dbf->rec_lock);
621 
622 	/* debug feature area which records recovery activity */
623 	sprintf(name, "zfcp_%s_rec", dev_name(&adapter->ccw_device->dev));
624 	dbf->rec = zfcp_dbf_reg(name, dbfsize, sizeof(struct zfcp_dbf_rec));
625 	if (!dbf->rec)
626 		goto err_out;
627 
628 	/* debug feature area which records HBA (FSF and QDIO) conditions */
629 	sprintf(name, "zfcp_%s_hba", dev_name(&adapter->ccw_device->dev));
630 	dbf->hba = zfcp_dbf_reg(name, dbfsize, sizeof(struct zfcp_dbf_hba));
631 	if (!dbf->hba)
632 		goto err_out;
633 
634 	/* debug feature area which records payload info */
635 	sprintf(name, "zfcp_%s_pay", dev_name(&adapter->ccw_device->dev));
636 	dbf->pay = zfcp_dbf_reg(name, dbfsize * 2, sizeof(struct zfcp_dbf_pay));
637 	if (!dbf->pay)
638 		goto err_out;
639 
640 	/* debug feature area which records SAN command failures and recovery */
641 	sprintf(name, "zfcp_%s_san", dev_name(&adapter->ccw_device->dev));
642 	dbf->san = zfcp_dbf_reg(name, dbfsize, sizeof(struct zfcp_dbf_san));
643 	if (!dbf->san)
644 		goto err_out;
645 
646 	/* debug feature area which records SCSI command failures and recovery */
647 	sprintf(name, "zfcp_%s_scsi", dev_name(&adapter->ccw_device->dev));
648 	dbf->scsi = zfcp_dbf_reg(name, dbfsize, sizeof(struct zfcp_dbf_scsi));
649 	if (!dbf->scsi)
650 		goto err_out;
651 
652 	adapter->dbf = dbf;
653 
654 	return 0;
655 err_out:
656 	zfcp_dbf_unregister(dbf);
657 	return -ENOMEM;
658 }
659 
660 /**
661  * zfcp_adapter_debug_unregister - unregisters debug feature for an adapter
662  * @adapter: pointer to adapter for which debug features should be unregistered
663  */
664 void zfcp_dbf_adapter_unregister(struct zfcp_adapter *adapter)
665 {
666 	struct zfcp_dbf *dbf = adapter->dbf;
667 
668 	adapter->dbf = NULL;
669 	zfcp_dbf_unregister(dbf);
670 }
671 
672