1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * NVM Express target device driver tracepoints
4 * Copyright (c) 2018 Johannes Thumshirn, SUSE Linux GmbH
5 */
6
7 #include <linux/unaligned.h>
8 #include "trace.h"
9
nvmet_trace_admin_identify(struct trace_seq * p,u8 * cdw10)10 static const char *nvmet_trace_admin_identify(struct trace_seq *p, u8 *cdw10)
11 {
12 const char *ret = trace_seq_buffer_ptr(p);
13 u8 cns = cdw10[0];
14 u16 ctrlid = get_unaligned_le16(cdw10 + 2);
15
16 trace_seq_printf(p, "cns=%u, ctrlid=%u", cns, ctrlid);
17 trace_seq_putc(p, 0);
18
19 return ret;
20 }
21
nvmet_trace_admin_get_features(struct trace_seq * p,u8 * cdw10)22 static const char *nvmet_trace_admin_get_features(struct trace_seq *p,
23 u8 *cdw10)
24 {
25 const char *ret = trace_seq_buffer_ptr(p);
26 u8 fid = cdw10[0];
27 u8 sel = cdw10[1] & 0x7;
28 u32 cdw11 = get_unaligned_le32(cdw10 + 4);
29
30 trace_seq_printf(p, "fid=0x%x, sel=0x%x, cdw11=0x%x", fid, sel, cdw11);
31 trace_seq_putc(p, 0);
32
33 return ret;
34 }
35
nvmet_trace_get_lba_status(struct trace_seq * p,u8 * cdw10)36 static const char *nvmet_trace_get_lba_status(struct trace_seq *p,
37 u8 *cdw10)
38 {
39 const char *ret = trace_seq_buffer_ptr(p);
40 u64 slba = get_unaligned_le64(cdw10);
41 u32 mndw = get_unaligned_le32(cdw10 + 8);
42 u16 rl = get_unaligned_le16(cdw10 + 12);
43 u8 atype = cdw10[15];
44
45 trace_seq_printf(p, "slba=0x%llx, mndw=0x%x, rl=0x%x, atype=%u",
46 slba, mndw, rl, atype);
47 trace_seq_putc(p, 0);
48
49 return ret;
50 }
51
nvmet_trace_admin_set_features(struct trace_seq * p,u8 * cdw10)52 static const char *nvmet_trace_admin_set_features(struct trace_seq *p,
53 u8 *cdw10)
54 {
55 const char *ret = trace_seq_buffer_ptr(p);
56 u8 fid = cdw10[0];
57 u8 sv = cdw10[3] & 0x8;
58 u32 cdw11 = get_unaligned_le32(cdw10 + 4);
59
60 trace_seq_printf(p, "fid=0x%x, sv=0x%x, cdw11=0x%x", fid, sv, cdw11);
61 trace_seq_putc(p, 0);
62
63 return ret;
64 }
65
nvmet_trace_read_write(struct trace_seq * p,u8 * cdw10)66 static const char *nvmet_trace_read_write(struct trace_seq *p, u8 *cdw10)
67 {
68 const char *ret = trace_seq_buffer_ptr(p);
69 u64 slba = get_unaligned_le64(cdw10);
70 u16 length = get_unaligned_le16(cdw10 + 8);
71 u16 control = get_unaligned_le16(cdw10 + 10);
72 u32 dsmgmt = get_unaligned_le32(cdw10 + 12);
73 u32 reftag = get_unaligned_le32(cdw10 + 16);
74
75 trace_seq_printf(p,
76 "slba=%llu, len=%u, ctrl=0x%x, dsmgmt=%u, reftag=%u",
77 slba, length, control, dsmgmt, reftag);
78 trace_seq_putc(p, 0);
79
80 return ret;
81 }
82
nvmet_trace_dsm(struct trace_seq * p,u8 * cdw10)83 static const char *nvmet_trace_dsm(struct trace_seq *p, u8 *cdw10)
84 {
85 const char *ret = trace_seq_buffer_ptr(p);
86
87 trace_seq_printf(p, "nr=%u, attributes=%u",
88 get_unaligned_le32(cdw10),
89 get_unaligned_le32(cdw10 + 4));
90 trace_seq_putc(p, 0);
91
92 return ret;
93 }
94
nvmet_trace_common(struct trace_seq * p,u8 * cdw10)95 static const char *nvmet_trace_common(struct trace_seq *p, u8 *cdw10)
96 {
97 const char *ret = trace_seq_buffer_ptr(p);
98
99 trace_seq_printf(p, "cdw10=%*ph", 24, cdw10);
100 trace_seq_putc(p, 0);
101
102 return ret;
103 }
104
nvmet_trace_parse_admin_cmd(struct trace_seq * p,u8 opcode,u8 * cdw10)105 const char *nvmet_trace_parse_admin_cmd(struct trace_seq *p,
106 u8 opcode, u8 *cdw10)
107 {
108 switch (opcode) {
109 case nvme_admin_identify:
110 return nvmet_trace_admin_identify(p, cdw10);
111 case nvme_admin_set_features:
112 return nvmet_trace_admin_set_features(p, cdw10);
113 case nvme_admin_get_features:
114 return nvmet_trace_admin_get_features(p, cdw10);
115 case nvme_admin_get_lba_status:
116 return nvmet_trace_get_lba_status(p, cdw10);
117 default:
118 return nvmet_trace_common(p, cdw10);
119 }
120 }
121
nvmet_trace_zone_mgmt_send(struct trace_seq * p,u8 * cdw10)122 static const char *nvmet_trace_zone_mgmt_send(struct trace_seq *p, u8 *cdw10)
123 {
124 static const char * const zsa_strs[] = {
125 [0x01] = "close zone",
126 [0x02] = "finish zone",
127 [0x03] = "open zone",
128 [0x04] = "reset zone",
129 [0x05] = "offline zone",
130 [0x10] = "set zone descriptor extension"
131 };
132 const char *ret = trace_seq_buffer_ptr(p);
133 u64 slba = get_unaligned_le64(cdw10);
134 const char *zsa_str;
135 u8 zsa = cdw10[12];
136 u8 all = cdw10[13];
137
138 if (zsa < ARRAY_SIZE(zsa_strs) && zsa_strs[zsa])
139 zsa_str = zsa_strs[zsa];
140 else
141 zsa_str = "reserved";
142
143 trace_seq_printf(p, "slba=%llu, zsa=%u:%s, all=%u",
144 slba, zsa, zsa_str, all);
145 trace_seq_putc(p, 0);
146
147 return ret;
148 }
149
nvmet_trace_zone_mgmt_recv(struct trace_seq * p,u8 * cdw10)150 static const char *nvmet_trace_zone_mgmt_recv(struct trace_seq *p, u8 *cdw10)
151 {
152 static const char * const zrasf_strs[] = {
153 [0x00] = "list all zones",
154 [0x01] = "list the zones in the ZSE: Empty state",
155 [0x02] = "list the zones in the ZSIO: Implicitly Opened state",
156 [0x03] = "list the zones in the ZSEO: Explicitly Opened state",
157 [0x04] = "list the zones in the ZSC: Closed state",
158 [0x05] = "list the zones in the ZSF: Full state",
159 [0x06] = "list the zones in the ZSRO: Read Only state",
160 [0x07] = "list the zones in the ZSO: Offline state",
161 [0x09] = "list the zones that have the zone attribute"
162 };
163 const char *ret = trace_seq_buffer_ptr(p);
164 u64 slba = get_unaligned_le64(cdw10);
165 u32 numd = get_unaligned_le32(&cdw10[8]);
166 u8 zra = cdw10[12];
167 u8 zrasf = cdw10[13];
168 const char *zrasf_str;
169 u8 pr = cdw10[14];
170
171 if (zrasf < ARRAY_SIZE(zrasf_strs) && zrasf_strs[zrasf])
172 zrasf_str = zrasf_strs[zrasf];
173 else
174 zrasf_str = "reserved";
175
176 trace_seq_printf(p, "slba=%llu, numd=%u, zra=%u, zrasf=%u:%s, pr=%u",
177 slba, numd, zra, zrasf, zrasf_str, pr);
178 trace_seq_putc(p, 0);
179
180 return ret;
181 }
182
nvmet_trace_parse_nvm_cmd(struct trace_seq * p,u8 opcode,u8 * cdw10)183 const char *nvmet_trace_parse_nvm_cmd(struct trace_seq *p,
184 u8 opcode, u8 *cdw10)
185 {
186 switch (opcode) {
187 case nvme_cmd_read:
188 case nvme_cmd_write:
189 case nvme_cmd_write_zeroes:
190 case nvme_cmd_zone_append:
191 return nvmet_trace_read_write(p, cdw10);
192 case nvme_cmd_dsm:
193 return nvmet_trace_dsm(p, cdw10);
194 case nvme_cmd_zone_mgmt_send:
195 return nvmet_trace_zone_mgmt_send(p, cdw10);
196 case nvme_cmd_zone_mgmt_recv:
197 return nvmet_trace_zone_mgmt_recv(p, cdw10);
198 default:
199 return nvmet_trace_common(p, cdw10);
200 }
201 }
202
nvmet_trace_fabrics_property_set(struct trace_seq * p,u8 * spc)203 static const char *nvmet_trace_fabrics_property_set(struct trace_seq *p,
204 u8 *spc)
205 {
206 const char *ret = trace_seq_buffer_ptr(p);
207 u8 attrib = spc[0];
208 u32 ofst = get_unaligned_le32(spc + 4);
209 u64 value = get_unaligned_le64(spc + 8);
210
211 trace_seq_printf(p, "attrib=%u, ofst=0x%x, value=0x%llx",
212 attrib, ofst, value);
213 trace_seq_putc(p, 0);
214 return ret;
215 }
216
nvmet_trace_fabrics_connect(struct trace_seq * p,u8 * spc)217 static const char *nvmet_trace_fabrics_connect(struct trace_seq *p,
218 u8 *spc)
219 {
220 const char *ret = trace_seq_buffer_ptr(p);
221 u16 recfmt = get_unaligned_le16(spc);
222 u16 qid = get_unaligned_le16(spc + 2);
223 u16 sqsize = get_unaligned_le16(spc + 4);
224 u8 cattr = spc[6];
225 u32 kato = get_unaligned_le32(spc + 8);
226
227 trace_seq_printf(p, "recfmt=%u, qid=%u, sqsize=%u, cattr=%u, kato=%u",
228 recfmt, qid, sqsize, cattr, kato);
229 trace_seq_putc(p, 0);
230 return ret;
231 }
232
nvmet_trace_fabrics_property_get(struct trace_seq * p,u8 * spc)233 static const char *nvmet_trace_fabrics_property_get(struct trace_seq *p,
234 u8 *spc)
235 {
236 const char *ret = trace_seq_buffer_ptr(p);
237 u8 attrib = spc[0];
238 u32 ofst = get_unaligned_le32(spc + 4);
239
240 trace_seq_printf(p, "attrib=%u, ofst=0x%x", attrib, ofst);
241 trace_seq_putc(p, 0);
242 return ret;
243 }
244
nvmet_trace_fabrics_auth_send(struct trace_seq * p,u8 * spc)245 static const char *nvmet_trace_fabrics_auth_send(struct trace_seq *p, u8 *spc)
246 {
247 const char *ret = trace_seq_buffer_ptr(p);
248 u8 spsp0 = spc[1];
249 u8 spsp1 = spc[2];
250 u8 secp = spc[3];
251 u32 tl = get_unaligned_le32(spc + 4);
252
253 trace_seq_printf(p, "spsp0=%02x, spsp1=%02x, secp=%02x, tl=%u",
254 spsp0, spsp1, secp, tl);
255 trace_seq_putc(p, 0);
256 return ret;
257 }
258
nvmet_trace_fabrics_auth_receive(struct trace_seq * p,u8 * spc)259 static const char *nvmet_trace_fabrics_auth_receive(struct trace_seq *p, u8 *spc)
260 {
261 const char *ret = trace_seq_buffer_ptr(p);
262 u8 spsp0 = spc[1];
263 u8 spsp1 = spc[2];
264 u8 secp = spc[3];
265 u32 al = get_unaligned_le32(spc + 4);
266
267 trace_seq_printf(p, "spsp0=%02x, spsp1=%02x, secp=%02x, al=%u",
268 spsp0, spsp1, secp, al);
269 trace_seq_putc(p, 0);
270 return ret;
271 }
272
nvmet_trace_fabrics_common(struct trace_seq * p,u8 * spc)273 static const char *nvmet_trace_fabrics_common(struct trace_seq *p, u8 *spc)
274 {
275 const char *ret = trace_seq_buffer_ptr(p);
276
277 trace_seq_printf(p, "specific=%*ph", 24, spc);
278 trace_seq_putc(p, 0);
279 return ret;
280 }
281
nvmet_trace_parse_fabrics_cmd(struct trace_seq * p,u8 fctype,u8 * spc)282 const char *nvmet_trace_parse_fabrics_cmd(struct trace_seq *p,
283 u8 fctype, u8 *spc)
284 {
285 switch (fctype) {
286 case nvme_fabrics_type_property_set:
287 return nvmet_trace_fabrics_property_set(p, spc);
288 case nvme_fabrics_type_connect:
289 return nvmet_trace_fabrics_connect(p, spc);
290 case nvme_fabrics_type_property_get:
291 return nvmet_trace_fabrics_property_get(p, spc);
292 case nvme_fabrics_type_auth_send:
293 return nvmet_trace_fabrics_auth_send(p, spc);
294 case nvme_fabrics_type_auth_receive:
295 return nvmet_trace_fabrics_auth_receive(p, spc);
296 default:
297 return nvmet_trace_fabrics_common(p, spc);
298 }
299 }
300
nvmet_trace_disk_name(struct trace_seq * p,char * name)301 const char *nvmet_trace_disk_name(struct trace_seq *p, char *name)
302 {
303 const char *ret = trace_seq_buffer_ptr(p);
304
305 if (*name)
306 trace_seq_printf(p, "disk=%s, ", name);
307 trace_seq_putc(p, 0);
308
309 return ret;
310 }
311
nvmet_trace_ctrl_id(struct trace_seq * p,u16 ctrl_id)312 const char *nvmet_trace_ctrl_id(struct trace_seq *p, u16 ctrl_id)
313 {
314 const char *ret = trace_seq_buffer_ptr(p);
315
316 /*
317 * XXX: We don't know the controller instance before executing the
318 * connect command itself because the connect command for the admin
319 * queue will not provide the cntlid which will be allocated in this
320 * command. In case of io queues, the controller instance will be
321 * mapped by the extra data of the connect command.
322 * If we can know the extra data of the connect command in this stage,
323 * we can update this print statement later.
324 */
325 if (ctrl_id)
326 trace_seq_printf(p, "%d", ctrl_id);
327 else
328 trace_seq_printf(p, "_");
329 trace_seq_putc(p, 0);
330
331 return ret;
332 }
333
334