xref: /linux/drivers/scsi/scsi_debug.c (revision c0e297dc61f8d4453e07afbea1fa8d0e67cd4a34)
1 /*
2  * vvvvvvvvvvvvvvvvvvvvvvv Original vvvvvvvvvvvvvvvvvvvvvvvvvvvvvvv
3  *  Copyright (C) 1992  Eric Youngdale
4  *  Simulate a host adapter with 2 disks attached.  Do a lot of checking
5  *  to make sure that we are not getting blocks mixed up, and PANIC if
6  *  anything out of the ordinary is seen.
7  * ^^^^^^^^^^^^^^^^^^^^^^^ Original ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
8  *
9  *  This version is more generic, simulating a variable number of disk
10  *  (or disk like devices) sharing a common amount of RAM. To be more
11  *  realistic, the simulated devices have the transport attributes of
12  *  SAS disks.
13  *
14  *
15  *  For documentation see http://sg.danny.cz/sg/sdebug26.html
16  *
17  *   D. Gilbert (dpg) work for Magneto-Optical device test [20010421]
18  *   dpg: work for devfs large number of disks [20010809]
19  *        forked for lk 2.5 series [20011216, 20020101]
20  *        use vmalloc() more inquiry+mode_sense [20020302]
21  *        add timers for delayed responses [20020721]
22  *   Patrick Mansfield <patmans@us.ibm.com> max_luns+scsi_level [20021031]
23  *   Mike Anderson <andmike@us.ibm.com> sysfs work [20021118]
24  *   dpg: change style of boot options to "scsi_debug.num_tgts=2" and
25  *        module options to "modprobe scsi_debug num_tgts=2" [20021221]
26  */
27 
28 #include <linux/module.h>
29 
30 #include <linux/kernel.h>
31 #include <linux/errno.h>
32 #include <linux/timer.h>
33 #include <linux/slab.h>
34 #include <linux/types.h>
35 #include <linux/string.h>
36 #include <linux/genhd.h>
37 #include <linux/fs.h>
38 #include <linux/init.h>
39 #include <linux/proc_fs.h>
40 #include <linux/vmalloc.h>
41 #include <linux/moduleparam.h>
42 #include <linux/scatterlist.h>
43 #include <linux/blkdev.h>
44 #include <linux/crc-t10dif.h>
45 #include <linux/spinlock.h>
46 #include <linux/interrupt.h>
47 #include <linux/atomic.h>
48 #include <linux/hrtimer.h>
49 
50 #include <net/checksum.h>
51 
52 #include <asm/unaligned.h>
53 
54 #include <scsi/scsi.h>
55 #include <scsi/scsi_cmnd.h>
56 #include <scsi/scsi_device.h>
57 #include <scsi/scsi_host.h>
58 #include <scsi/scsicam.h>
59 #include <scsi/scsi_eh.h>
60 #include <scsi/scsi_tcq.h>
61 #include <scsi/scsi_dbg.h>
62 
63 #include "sd.h"
64 #include "scsi_logging.h"
65 
66 #define SCSI_DEBUG_VERSION "1.85"
67 static const char *scsi_debug_version_date = "20141022";
68 
69 #define MY_NAME "scsi_debug"
70 
71 /* Additional Sense Code (ASC) */
72 #define NO_ADDITIONAL_SENSE 0x0
73 #define LOGICAL_UNIT_NOT_READY 0x4
74 #define LOGICAL_UNIT_COMMUNICATION_FAILURE 0x8
75 #define UNRECOVERED_READ_ERR 0x11
76 #define PARAMETER_LIST_LENGTH_ERR 0x1a
77 #define INVALID_OPCODE 0x20
78 #define LBA_OUT_OF_RANGE 0x21
79 #define INVALID_FIELD_IN_CDB 0x24
80 #define INVALID_FIELD_IN_PARAM_LIST 0x26
81 #define UA_RESET_ASC 0x29
82 #define UA_CHANGED_ASC 0x2a
83 #define TARGET_CHANGED_ASC 0x3f
84 #define LUNS_CHANGED_ASCQ 0x0e
85 #define INSUFF_RES_ASC 0x55
86 #define INSUFF_RES_ASCQ 0x3
87 #define POWER_ON_RESET_ASCQ 0x0
88 #define BUS_RESET_ASCQ 0x2	/* scsi bus reset occurred */
89 #define MODE_CHANGED_ASCQ 0x1	/* mode parameters changed */
90 #define CAPACITY_CHANGED_ASCQ 0x9
91 #define SAVING_PARAMS_UNSUP 0x39
92 #define TRANSPORT_PROBLEM 0x4b
93 #define THRESHOLD_EXCEEDED 0x5d
94 #define LOW_POWER_COND_ON 0x5e
95 #define MISCOMPARE_VERIFY_ASC 0x1d
96 #define MICROCODE_CHANGED_ASCQ 0x1	/* with TARGET_CHANGED_ASC */
97 #define MICROCODE_CHANGED_WO_RESET_ASCQ 0x16
98 
99 /* Additional Sense Code Qualifier (ASCQ) */
100 #define ACK_NAK_TO 0x3
101 
102 
103 /* Default values for driver parameters */
104 #define DEF_NUM_HOST   1
105 #define DEF_NUM_TGTS   1
106 #define DEF_MAX_LUNS   1
107 /* With these defaults, this driver will make 1 host with 1 target
108  * (id 0) containing 1 logical unit (lun 0). That is 1 device.
109  */
110 #define DEF_ATO 1
111 #define DEF_DELAY   1		/* if > 0 unit is a jiffy */
112 #define DEF_DEV_SIZE_MB   8
113 #define DEF_DIF 0
114 #define DEF_DIX 0
115 #define DEF_D_SENSE   0
116 #define DEF_EVERY_NTH   0
117 #define DEF_FAKE_RW	0
118 #define DEF_GUARD 0
119 #define DEF_HOST_LOCK 0
120 #define DEF_LBPU 0
121 #define DEF_LBPWS 0
122 #define DEF_LBPWS10 0
123 #define DEF_LBPRZ 1
124 #define DEF_LOWEST_ALIGNED 0
125 #define DEF_NDELAY   0		/* if > 0 unit is a nanosecond */
126 #define DEF_NO_LUN_0   0
127 #define DEF_NUM_PARTS   0
128 #define DEF_OPTS   0
129 #define DEF_OPT_BLKS 64
130 #define DEF_PHYSBLK_EXP 0
131 #define DEF_PTYPE   0
132 #define DEF_REMOVABLE false
133 #define DEF_SCSI_LEVEL   6    /* INQUIRY, byte2 [6->SPC-4] */
134 #define DEF_SECTOR_SIZE 512
135 #define DEF_UNMAP_ALIGNMENT 0
136 #define DEF_UNMAP_GRANULARITY 1
137 #define DEF_UNMAP_MAX_BLOCKS 0xFFFFFFFF
138 #define DEF_UNMAP_MAX_DESC 256
139 #define DEF_VIRTUAL_GB   0
140 #define DEF_VPD_USE_HOSTNO 1
141 #define DEF_WRITESAME_LENGTH 0xFFFF
142 #define DEF_STRICT 0
143 #define DELAY_OVERRIDDEN -9999
144 
145 /* bit mask values for scsi_debug_opts */
146 #define SCSI_DEBUG_OPT_NOISE   1
147 #define SCSI_DEBUG_OPT_MEDIUM_ERR   2
148 #define SCSI_DEBUG_OPT_TIMEOUT   4
149 #define SCSI_DEBUG_OPT_RECOVERED_ERR   8
150 #define SCSI_DEBUG_OPT_TRANSPORT_ERR   16
151 #define SCSI_DEBUG_OPT_DIF_ERR   32
152 #define SCSI_DEBUG_OPT_DIX_ERR   64
153 #define SCSI_DEBUG_OPT_MAC_TIMEOUT  128
154 #define SCSI_DEBUG_OPT_SHORT_TRANSFER	0x100
155 #define SCSI_DEBUG_OPT_Q_NOISE	0x200
156 #define SCSI_DEBUG_OPT_ALL_TSF	0x400
157 #define SCSI_DEBUG_OPT_RARE_TSF	0x800
158 #define SCSI_DEBUG_OPT_N_WCE	0x1000
159 #define SCSI_DEBUG_OPT_RESET_NOISE 0x2000
160 #define SCSI_DEBUG_OPT_NO_CDB_NOISE 0x4000
161 #define SCSI_DEBUG_OPT_ALL_NOISE (0x1 | 0x200 | 0x2000)
162 /* When "every_nth" > 0 then modulo "every_nth" commands:
163  *   - a no response is simulated if SCSI_DEBUG_OPT_TIMEOUT is set
164  *   - a RECOVERED_ERROR is simulated on successful read and write
165  *     commands if SCSI_DEBUG_OPT_RECOVERED_ERR is set.
166  *   - a TRANSPORT_ERROR is simulated on successful read and write
167  *     commands if SCSI_DEBUG_OPT_TRANSPORT_ERR is set.
168  *
169  * When "every_nth" < 0 then after "- every_nth" commands:
170  *   - a no response is simulated if SCSI_DEBUG_OPT_TIMEOUT is set
171  *   - a RECOVERED_ERROR is simulated on successful read and write
172  *     commands if SCSI_DEBUG_OPT_RECOVERED_ERR is set.
173  *   - a TRANSPORT_ERROR is simulated on successful read and write
174  *     commands if SCSI_DEBUG_OPT_TRANSPORT_ERR is set.
175  * This will continue until some other action occurs (e.g. the user
176  * writing a new value (other than -1 or 1) to every_nth via sysfs).
177  */
178 
179 /* As indicated in SAM-5 and SPC-4 Unit Attentions (UAs)are returned in
180  * priority order. In the subset implemented here lower numbers have higher
181  * priority. The UA numbers should be a sequence starting from 0 with
182  * SDEBUG_NUM_UAS being 1 higher than the highest numbered UA. */
183 #define SDEBUG_UA_POR 0		/* Power on, reset, or bus device reset */
184 #define SDEBUG_UA_BUS_RESET 1
185 #define SDEBUG_UA_MODE_CHANGED 2
186 #define SDEBUG_UA_CAPACITY_CHANGED 3
187 #define SDEBUG_UA_LUNS_CHANGED 4
188 #define SDEBUG_UA_MICROCODE_CHANGED 5	/* simulate firmware change */
189 #define SDEBUG_UA_MICROCODE_CHANGED_WO_RESET 6
190 #define SDEBUG_NUM_UAS 7
191 
192 /* for check_readiness() */
193 #define UAS_ONLY 1	/* check for UAs only */
194 #define UAS_TUR 0	/* if no UAs then check if media access possible */
195 
196 /* when 1==SCSI_DEBUG_OPT_MEDIUM_ERR, a medium error is simulated at this
197  * sector on read commands: */
198 #define OPT_MEDIUM_ERR_ADDR   0x1234 /* that's sector 4660 in decimal */
199 #define OPT_MEDIUM_ERR_NUM    10     /* number of consecutive medium errs */
200 
201 /* If REPORT LUNS has luns >= 256 it can choose "flat space" (value 1)
202  * or "peripheral device" addressing (value 0) */
203 #define SAM2_LUN_ADDRESS_METHOD 0
204 #define SAM2_WLUN_REPORT_LUNS 0xc101
205 
206 /* SCSI_DEBUG_CANQUEUE is the maximum number of commands that can be queued
207  * (for response) at one time. Can be reduced by max_queue option. Command
208  * responses are not queued when delay=0 and ndelay=0. The per-device
209  * DEF_CMD_PER_LUN can be changed via sysfs:
210  * /sys/class/scsi_device/<h:c:t:l>/device/queue_depth but cannot exceed
211  * SCSI_DEBUG_CANQUEUE. */
212 #define SCSI_DEBUG_CANQUEUE_WORDS  9	/* a WORD is bits in a long */
213 #define SCSI_DEBUG_CANQUEUE  (SCSI_DEBUG_CANQUEUE_WORDS * BITS_PER_LONG)
214 #define DEF_CMD_PER_LUN  255
215 
216 #if DEF_CMD_PER_LUN > SCSI_DEBUG_CANQUEUE
217 #warning "Expect DEF_CMD_PER_LUN <= SCSI_DEBUG_CANQUEUE"
218 #endif
219 
220 /* SCSI opcodes (first byte of cdb) mapped onto these indexes */
221 enum sdeb_opcode_index {
222 	SDEB_I_INVALID_OPCODE =	0,
223 	SDEB_I_INQUIRY = 1,
224 	SDEB_I_REPORT_LUNS = 2,
225 	SDEB_I_REQUEST_SENSE = 3,
226 	SDEB_I_TEST_UNIT_READY = 4,
227 	SDEB_I_MODE_SENSE = 5,		/* 6, 10 */
228 	SDEB_I_MODE_SELECT = 6,		/* 6, 10 */
229 	SDEB_I_LOG_SENSE = 7,
230 	SDEB_I_READ_CAPACITY = 8,	/* 10; 16 is in SA_IN(16) */
231 	SDEB_I_READ = 9,		/* 6, 10, 12, 16 */
232 	SDEB_I_WRITE = 10,		/* 6, 10, 12, 16 */
233 	SDEB_I_START_STOP = 11,
234 	SDEB_I_SERV_ACT_IN = 12,	/* 12, 16 */
235 	SDEB_I_SERV_ACT_OUT = 13,	/* 12, 16 */
236 	SDEB_I_MAINT_IN = 14,
237 	SDEB_I_MAINT_OUT = 15,
238 	SDEB_I_VERIFY = 16,		/* 10 only */
239 	SDEB_I_VARIABLE_LEN = 17,
240 	SDEB_I_RESERVE = 18,		/* 6, 10 */
241 	SDEB_I_RELEASE = 19,		/* 6, 10 */
242 	SDEB_I_ALLOW_REMOVAL = 20,	/* PREVENT ALLOW MEDIUM REMOVAL */
243 	SDEB_I_REZERO_UNIT = 21,	/* REWIND in SSC */
244 	SDEB_I_ATA_PT = 22,		/* 12, 16 */
245 	SDEB_I_SEND_DIAG = 23,
246 	SDEB_I_UNMAP = 24,
247 	SDEB_I_XDWRITEREAD = 25,	/* 10 only */
248 	SDEB_I_WRITE_BUFFER = 26,
249 	SDEB_I_WRITE_SAME = 27,		/* 10, 16 */
250 	SDEB_I_SYNC_CACHE = 28,		/* 10 only */
251 	SDEB_I_COMP_WRITE = 29,
252 	SDEB_I_LAST_ELEMENT = 30,	/* keep this last */
253 };
254 
255 static const unsigned char opcode_ind_arr[256] = {
256 /* 0x0; 0x0->0x1f: 6 byte cdbs */
257 	SDEB_I_TEST_UNIT_READY, SDEB_I_REZERO_UNIT, 0, SDEB_I_REQUEST_SENSE,
258 	    0, 0, 0, 0,
259 	SDEB_I_READ, 0, SDEB_I_WRITE, 0, 0, 0, 0, 0,
260 	0, 0, SDEB_I_INQUIRY, 0, 0, SDEB_I_MODE_SELECT, SDEB_I_RESERVE,
261 	    SDEB_I_RELEASE,
262 	0, 0, SDEB_I_MODE_SENSE, SDEB_I_START_STOP, 0, SDEB_I_SEND_DIAG,
263 	    SDEB_I_ALLOW_REMOVAL, 0,
264 /* 0x20; 0x20->0x3f: 10 byte cdbs */
265 	0, 0, 0, 0, 0, SDEB_I_READ_CAPACITY, 0, 0,
266 	SDEB_I_READ, 0, SDEB_I_WRITE, 0, 0, 0, 0, SDEB_I_VERIFY,
267 	0, 0, 0, 0, 0, SDEB_I_SYNC_CACHE, 0, 0,
268 	0, 0, 0, SDEB_I_WRITE_BUFFER, 0, 0, 0, 0,
269 /* 0x40; 0x40->0x5f: 10 byte cdbs */
270 	0, SDEB_I_WRITE_SAME, SDEB_I_UNMAP, 0, 0, 0, 0, 0,
271 	0, 0, 0, 0, 0, SDEB_I_LOG_SENSE, 0, 0,
272 	0, 0, 0, SDEB_I_XDWRITEREAD, 0, SDEB_I_MODE_SELECT, SDEB_I_RESERVE,
273 	    SDEB_I_RELEASE,
274 	0, 0, SDEB_I_MODE_SENSE, 0, 0, 0, 0, 0,
275 /* 0x60; 0x60->0x7d are reserved */
276 	0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
277 	0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
278 	0, SDEB_I_VARIABLE_LEN,
279 /* 0x80; 0x80->0x9f: 16 byte cdbs */
280 	0, 0, 0, 0, 0, SDEB_I_ATA_PT, 0, 0,
281 	SDEB_I_READ, SDEB_I_COMP_WRITE, SDEB_I_WRITE, 0, 0, 0, 0, 0,
282 	0, 0, 0, SDEB_I_WRITE_SAME, 0, 0, 0, 0,
283 	0, 0, 0, 0, 0, 0, SDEB_I_SERV_ACT_IN, SDEB_I_SERV_ACT_OUT,
284 /* 0xa0; 0xa0->0xbf: 12 byte cdbs */
285 	SDEB_I_REPORT_LUNS, SDEB_I_ATA_PT, 0, SDEB_I_MAINT_IN,
286 	     SDEB_I_MAINT_OUT, 0, 0, 0,
287 	SDEB_I_READ, SDEB_I_SERV_ACT_OUT, SDEB_I_WRITE, SDEB_I_SERV_ACT_IN,
288 	     0, 0, 0, 0,
289 	0, 0, 0, 0, 0, 0, 0, 0,
290 	0, 0, 0, 0, 0, 0, 0, 0,
291 /* 0xc0; 0xc0->0xff: vendor specific */
292 	0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
293 	0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
294 	0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
295 	0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
296 };
297 
298 #define F_D_IN			1
299 #define F_D_OUT			2
300 #define F_D_OUT_MAYBE		4	/* WRITE SAME, NDOB bit */
301 #define F_D_UNKN		8
302 #define F_RL_WLUN_OK		0x10
303 #define F_SKIP_UA		0x20
304 #define F_DELAY_OVERR		0x40
305 #define F_SA_LOW		0x80	/* cdb byte 1, bits 4 to 0 */
306 #define F_SA_HIGH		0x100	/* as used by variable length cdbs */
307 #define F_INV_OP		0x200
308 #define F_FAKE_RW		0x400
309 #define F_M_ACCESS		0x800	/* media access */
310 
311 #define FF_RESPOND (F_RL_WLUN_OK | F_SKIP_UA | F_DELAY_OVERR)
312 #define FF_DIRECT_IO (F_M_ACCESS | F_FAKE_RW)
313 #define FF_SA (F_SA_HIGH | F_SA_LOW)
314 
315 struct sdebug_dev_info;
316 static int resp_inquiry(struct scsi_cmnd *, struct sdebug_dev_info *);
317 static int resp_report_luns(struct scsi_cmnd *, struct sdebug_dev_info *);
318 static int resp_requests(struct scsi_cmnd *, struct sdebug_dev_info *);
319 static int resp_mode_sense(struct scsi_cmnd *, struct sdebug_dev_info *);
320 static int resp_mode_select(struct scsi_cmnd *, struct sdebug_dev_info *);
321 static int resp_log_sense(struct scsi_cmnd *, struct sdebug_dev_info *);
322 static int resp_readcap(struct scsi_cmnd *, struct sdebug_dev_info *);
323 static int resp_read_dt0(struct scsi_cmnd *, struct sdebug_dev_info *);
324 static int resp_write_dt0(struct scsi_cmnd *, struct sdebug_dev_info *);
325 static int resp_start_stop(struct scsi_cmnd *, struct sdebug_dev_info *);
326 static int resp_readcap16(struct scsi_cmnd *, struct sdebug_dev_info *);
327 static int resp_get_lba_status(struct scsi_cmnd *, struct sdebug_dev_info *);
328 static int resp_report_tgtpgs(struct scsi_cmnd *, struct sdebug_dev_info *);
329 static int resp_unmap(struct scsi_cmnd *, struct sdebug_dev_info *);
330 static int resp_rsup_opcodes(struct scsi_cmnd *, struct sdebug_dev_info *);
331 static int resp_rsup_tmfs(struct scsi_cmnd *, struct sdebug_dev_info *);
332 static int resp_write_same_10(struct scsi_cmnd *, struct sdebug_dev_info *);
333 static int resp_write_same_16(struct scsi_cmnd *, struct sdebug_dev_info *);
334 static int resp_xdwriteread_10(struct scsi_cmnd *, struct sdebug_dev_info *);
335 static int resp_comp_write(struct scsi_cmnd *, struct sdebug_dev_info *);
336 static int resp_write_buffer(struct scsi_cmnd *, struct sdebug_dev_info *);
337 
338 struct opcode_info_t {
339 	u8 num_attached;	/* 0 if this is it (i.e. a leaf); use 0xff
340 				 * for terminating element */
341 	u8 opcode;		/* if num_attached > 0, preferred */
342 	u16 sa;			/* service action */
343 	u32 flags;		/* OR-ed set of SDEB_F_* */
344 	int (*pfp)(struct scsi_cmnd *, struct sdebug_dev_info *);
345 	const struct opcode_info_t *arrp;  /* num_attached elements or NULL */
346 	u8 len_mask[16];	/* len=len_mask[0], then mask for cdb[1]... */
347 				/* ignore cdb bytes after position 15 */
348 };
349 
350 static const struct opcode_info_t msense_iarr[1] = {
351 	{0, 0x1a, 0, F_D_IN, NULL, NULL,
352 	    {6,  0xe8, 0xff, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
353 };
354 
355 static const struct opcode_info_t mselect_iarr[1] = {
356 	{0, 0x15, 0, F_D_OUT, NULL, NULL,
357 	    {6,  0xf1, 0, 0, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
358 };
359 
360 static const struct opcode_info_t read_iarr[3] = {
361 	{0, 0x28, 0, F_D_IN | FF_DIRECT_IO, resp_read_dt0, NULL,/* READ(10) */
362 	    {10,  0xff, 0xff, 0xff, 0xff, 0xff, 0x1f, 0xff, 0xff, 0xc7, 0, 0,
363 	     0, 0, 0, 0} },
364 	{0, 0x8, 0, F_D_IN | FF_DIRECT_IO, resp_read_dt0, NULL, /* READ(6) */
365 	    {6,  0xff, 0xff, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
366 	{0, 0xa8, 0, F_D_IN | FF_DIRECT_IO, resp_read_dt0, NULL,/* READ(12) */
367 	    {12,  0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0x9f,
368 	     0xc7, 0, 0, 0, 0} },
369 };
370 
371 static const struct opcode_info_t write_iarr[3] = {
372 	{0, 0x2a, 0, F_D_OUT | FF_DIRECT_IO, resp_write_dt0, NULL,   /* 10 */
373 	    {10,  0xfb, 0xff, 0xff, 0xff, 0xff, 0x1f, 0xff, 0xff, 0xc7, 0, 0,
374 	     0, 0, 0, 0} },
375 	{0, 0xa, 0, F_D_OUT | FF_DIRECT_IO, resp_write_dt0, NULL,    /* 6 */
376 	    {6,  0xff, 0xff, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
377 	{0, 0xaa, 0, F_D_OUT | FF_DIRECT_IO, resp_write_dt0, NULL,   /* 12 */
378 	    {12,  0xfb, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0x9f,
379 	     0xc7, 0, 0, 0, 0} },
380 };
381 
382 static const struct opcode_info_t sa_in_iarr[1] = {
383 	{0, 0x9e, 0x12, F_SA_LOW | F_D_IN, resp_get_lba_status, NULL,
384 	    {16,  0x12, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
385 	     0xff, 0xff, 0xff, 0, 0xc7} },
386 };
387 
388 static const struct opcode_info_t vl_iarr[1] = {	/* VARIABLE LENGTH */
389 	{0, 0x7f, 0xb, F_SA_HIGH | F_D_OUT | FF_DIRECT_IO, resp_write_dt0,
390 	    NULL, {32,  0xc7, 0, 0, 0, 0, 0x1f, 0x18, 0x0, 0xb, 0xfa,
391 		   0, 0xff, 0xff, 0xff, 0xff} },	/* WRITE(32) */
392 };
393 
394 static const struct opcode_info_t maint_in_iarr[2] = {
395 	{0, 0xa3, 0xc, F_SA_LOW | F_D_IN, resp_rsup_opcodes, NULL,
396 	    {12,  0xc, 0x87, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0,
397 	     0xc7, 0, 0, 0, 0} },
398 	{0, 0xa3, 0xd, F_SA_LOW | F_D_IN, resp_rsup_tmfs, NULL,
399 	    {12,  0xd, 0x80, 0, 0, 0, 0xff, 0xff, 0xff, 0xff, 0, 0xc7, 0, 0,
400 	     0, 0} },
401 };
402 
403 static const struct opcode_info_t write_same_iarr[1] = {
404 	{0, 0x93, 0, F_D_OUT_MAYBE | FF_DIRECT_IO, resp_write_same_16, NULL,
405 	    {16,  0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
406 	     0xff, 0xff, 0xff, 0x1f, 0xc7} },
407 };
408 
409 static const struct opcode_info_t reserve_iarr[1] = {
410 	{0, 0x16, 0, F_D_OUT, NULL, NULL,	/* RESERVE(6) */
411 	    {6,  0x1f, 0xff, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
412 };
413 
414 static const struct opcode_info_t release_iarr[1] = {
415 	{0, 0x17, 0, F_D_OUT, NULL, NULL,	/* RELEASE(6) */
416 	    {6,  0x1f, 0xff, 0, 0, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
417 };
418 
419 
420 /* This array is accessed via SDEB_I_* values. Make sure all are mapped,
421  * plus the terminating elements for logic that scans this table such as
422  * REPORT SUPPORTED OPERATION CODES. */
423 static const struct opcode_info_t opcode_info_arr[SDEB_I_LAST_ELEMENT + 1] = {
424 /* 0 */
425 	{0, 0, 0, F_INV_OP | FF_RESPOND, NULL, NULL,
426 	    {0,  0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
427 	{0, 0x12, 0, FF_RESPOND | F_D_IN, resp_inquiry, NULL,
428 	    {6,  0xe3, 0xff, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
429 	{0, 0xa0, 0, FF_RESPOND | F_D_IN, resp_report_luns, NULL,
430 	    {12,  0xe3, 0xff, 0, 0, 0, 0xff, 0xff, 0xff, 0xff, 0, 0xc7, 0, 0,
431 	     0, 0} },
432 	{0, 0x3, 0, FF_RESPOND | F_D_IN, resp_requests, NULL,
433 	    {6,  0xe1, 0, 0, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
434 	{0, 0x0, 0, F_M_ACCESS | F_RL_WLUN_OK, NULL, NULL,/* TEST UNIT READY */
435 	    {6,  0, 0, 0, 0, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
436 	{1, 0x5a, 0, F_D_IN, resp_mode_sense, msense_iarr,
437 	    {10,  0xf8, 0xff, 0xff, 0, 0, 0, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0,
438 	     0} },
439 	{1, 0x55, 0, F_D_OUT, resp_mode_select, mselect_iarr,
440 	    {10,  0xf1, 0, 0, 0, 0, 0, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0} },
441 	{0, 0x4d, 0, F_D_IN, resp_log_sense, NULL,
442 	    {10,  0xe3, 0xff, 0xff, 0, 0xff, 0xff, 0xff, 0xff, 0xc7, 0, 0, 0,
443 	     0, 0, 0} },
444 	{0, 0x25, 0, F_D_IN, resp_readcap, NULL,
445 	    {10,  0xe1, 0xff, 0xff, 0xff, 0xff, 0, 0, 0x1, 0xc7, 0, 0, 0, 0,
446 	     0, 0} },
447 	{3, 0x88, 0, F_D_IN | FF_DIRECT_IO, resp_read_dt0, read_iarr,
448 	    {16,  0xfe, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
449 	     0xff, 0xff, 0xff, 0x9f, 0xc7} },		/* READ(16) */
450 /* 10 */
451 	{3, 0x8a, 0, F_D_OUT | FF_DIRECT_IO, resp_write_dt0, write_iarr,
452 	    {16,  0xfa, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
453 	     0xff, 0xff, 0xff, 0x9f, 0xc7} },		/* WRITE(16) */
454 	{0, 0x1b, 0, 0, resp_start_stop, NULL,		/* START STOP UNIT */
455 	    {6,  0x1, 0, 0xf, 0xf7, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
456 	{1, 0x9e, 0x10, F_SA_LOW | F_D_IN, resp_readcap16, sa_in_iarr,
457 	    {16,  0x10, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
458 	     0xff, 0xff, 0xff, 0x1, 0xc7} },	/* READ CAPACITY(16) */
459 	{0, 0, 0, F_INV_OP | FF_RESPOND, NULL, NULL, /* SA OUT */
460 	    {0,  0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
461 	{2, 0xa3, 0xa, F_SA_LOW | F_D_IN, resp_report_tgtpgs, maint_in_iarr,
462 	    {12,  0xea, 0, 0, 0, 0, 0xff, 0xff, 0xff, 0xff, 0, 0xc7, 0, 0, 0,
463 	     0} },
464 	{0, 0, 0, F_INV_OP | FF_RESPOND, NULL, NULL, /* MAINT OUT */
465 	    {0,  0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
466 	{0, 0, 0, F_INV_OP | FF_RESPOND, NULL, NULL, /* VERIFY */
467 	    {0,  0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
468 	{1, 0x7f, 0x9, F_SA_HIGH | F_D_IN | FF_DIRECT_IO, resp_read_dt0,
469 	    vl_iarr, {32,  0xc7, 0, 0, 0, 0, 0x1f, 0x18, 0x0, 0x9, 0xfe, 0,
470 		      0xff, 0xff, 0xff, 0xff} },/* VARIABLE LENGTH, READ(32) */
471 	{1, 0x56, 0, F_D_OUT, NULL, reserve_iarr, /* RESERVE(10) */
472 	    {10,  0xff, 0xff, 0xff, 0, 0, 0, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0,
473 	     0} },
474 	{1, 0x57, 0, F_D_OUT, NULL, release_iarr, /* RELEASE(10) */
475 	    {10,  0x13, 0xff, 0xff, 0, 0, 0, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0,
476 	     0} },
477 /* 20 */
478 	{0, 0, 0, F_INV_OP | FF_RESPOND, NULL, NULL, /* ALLOW REMOVAL */
479 	    {0,  0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
480 	{0, 0x1, 0, 0, resp_start_stop, NULL, /* REWIND ?? */
481 	    {6,  0x1, 0, 0, 0, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
482 	{0, 0, 0, F_INV_OP | FF_RESPOND, NULL, NULL, /* ATA_PT */
483 	    {0,  0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
484 	{0, 0x1d, F_D_OUT, 0, NULL, NULL,	/* SEND DIAGNOSTIC */
485 	    {6,  0xf7, 0, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
486 	{0, 0x42, 0, F_D_OUT | FF_DIRECT_IO, resp_unmap, NULL, /* UNMAP */
487 	    {10,  0x1, 0, 0, 0, 0, 0x1f, 0xff, 0xff, 0xc7, 0, 0, 0, 0, 0, 0} },
488 	{0, 0x53, 0, F_D_IN | F_D_OUT | FF_DIRECT_IO, resp_xdwriteread_10,
489 	    NULL, {10,  0xff, 0xff, 0xff, 0xff, 0xff, 0x1f, 0xff, 0xff, 0xc7,
490 		   0, 0, 0, 0, 0, 0} },
491 	{0, 0x3b, 0, F_D_OUT_MAYBE, resp_write_buffer, NULL,
492 	    {10,  0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xc7, 0, 0,
493 	     0, 0, 0, 0} },			/* WRITE_BUFFER */
494 	{1, 0x41, 0, F_D_OUT_MAYBE | FF_DIRECT_IO, resp_write_same_10,
495 	    write_same_iarr, {10,  0xff, 0xff, 0xff, 0xff, 0xff, 0x1f, 0xff,
496 			      0xff, 0xc7, 0, 0, 0, 0, 0, 0} },
497 	{0, 0x35, 0, F_DELAY_OVERR | FF_DIRECT_IO, NULL, NULL, /* SYNC_CACHE */
498 	    {10,  0x7, 0xff, 0xff, 0xff, 0xff, 0x1f, 0xff, 0xff, 0xc7, 0, 0,
499 	     0, 0, 0, 0} },
500 	{0, 0x89, 0, F_D_OUT | FF_DIRECT_IO, resp_comp_write, NULL,
501 	    {16,  0xf8, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0, 0,
502 	     0, 0xff, 0x1f, 0xc7} },		/* COMPARE AND WRITE */
503 
504 /* 30 */
505 	{0xff, 0, 0, 0, NULL, NULL,		/* terminating element */
506 	    {0,  0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} },
507 };
508 
509 struct sdebug_scmd_extra_t {
510 	bool inj_recovered;
511 	bool inj_transport;
512 	bool inj_dif;
513 	bool inj_dix;
514 	bool inj_short;
515 };
516 
517 static int scsi_debug_add_host = DEF_NUM_HOST;
518 static int scsi_debug_ato = DEF_ATO;
519 static int scsi_debug_delay = DEF_DELAY;
520 static int scsi_debug_dev_size_mb = DEF_DEV_SIZE_MB;
521 static int scsi_debug_dif = DEF_DIF;
522 static int scsi_debug_dix = DEF_DIX;
523 static int scsi_debug_dsense = DEF_D_SENSE;
524 static int scsi_debug_every_nth = DEF_EVERY_NTH;
525 static int scsi_debug_fake_rw = DEF_FAKE_RW;
526 static unsigned int scsi_debug_guard = DEF_GUARD;
527 static int scsi_debug_lowest_aligned = DEF_LOWEST_ALIGNED;
528 static int scsi_debug_max_luns = DEF_MAX_LUNS;
529 static int scsi_debug_max_queue = SCSI_DEBUG_CANQUEUE;
530 static atomic_t retired_max_queue;	/* if > 0 then was prior max_queue */
531 static int scsi_debug_ndelay = DEF_NDELAY;
532 static int scsi_debug_no_lun_0 = DEF_NO_LUN_0;
533 static int scsi_debug_no_uld = 0;
534 static int scsi_debug_num_parts = DEF_NUM_PARTS;
535 static int scsi_debug_num_tgts = DEF_NUM_TGTS; /* targets per host */
536 static int scsi_debug_opt_blks = DEF_OPT_BLKS;
537 static int scsi_debug_opts = DEF_OPTS;
538 static int scsi_debug_physblk_exp = DEF_PHYSBLK_EXP;
539 static int scsi_debug_ptype = DEF_PTYPE; /* SCSI peripheral type (0==disk) */
540 static int scsi_debug_scsi_level = DEF_SCSI_LEVEL;
541 static int scsi_debug_sector_size = DEF_SECTOR_SIZE;
542 static int scsi_debug_virtual_gb = DEF_VIRTUAL_GB;
543 static int scsi_debug_vpd_use_hostno = DEF_VPD_USE_HOSTNO;
544 static unsigned int scsi_debug_lbpu = DEF_LBPU;
545 static unsigned int scsi_debug_lbpws = DEF_LBPWS;
546 static unsigned int scsi_debug_lbpws10 = DEF_LBPWS10;
547 static unsigned int scsi_debug_lbprz = DEF_LBPRZ;
548 static unsigned int scsi_debug_unmap_alignment = DEF_UNMAP_ALIGNMENT;
549 static unsigned int scsi_debug_unmap_granularity = DEF_UNMAP_GRANULARITY;
550 static unsigned int scsi_debug_unmap_max_blocks = DEF_UNMAP_MAX_BLOCKS;
551 static unsigned int scsi_debug_unmap_max_desc = DEF_UNMAP_MAX_DESC;
552 static unsigned int scsi_debug_write_same_length = DEF_WRITESAME_LENGTH;
553 static bool scsi_debug_removable = DEF_REMOVABLE;
554 static bool scsi_debug_clustering;
555 static bool scsi_debug_host_lock = DEF_HOST_LOCK;
556 static bool scsi_debug_strict = DEF_STRICT;
557 static bool sdebug_any_injecting_opt;
558 
559 static atomic_t sdebug_cmnd_count;
560 static atomic_t sdebug_completions;
561 static atomic_t sdebug_a_tsf;		/* counter of 'almost' TSFs */
562 
563 #define DEV_READONLY(TGT)      (0)
564 
565 static unsigned int sdebug_store_sectors;
566 static sector_t sdebug_capacity;	/* in sectors */
567 
568 /* old BIOS stuff, kernel may get rid of them but some mode sense pages
569    may still need them */
570 static int sdebug_heads;		/* heads per disk */
571 static int sdebug_cylinders_per;	/* cylinders per surface */
572 static int sdebug_sectors_per;		/* sectors per cylinder */
573 
574 #define SDEBUG_MAX_PARTS 4
575 
576 #define SCSI_DEBUG_MAX_CMD_LEN 32
577 
578 static unsigned int scsi_debug_lbp(void)
579 {
580 	return ((0 == scsi_debug_fake_rw) &&
581 		(scsi_debug_lbpu | scsi_debug_lbpws | scsi_debug_lbpws10));
582 }
583 
584 struct sdebug_dev_info {
585 	struct list_head dev_list;
586 	unsigned int channel;
587 	unsigned int target;
588 	u64 lun;
589 	struct sdebug_host_info *sdbg_host;
590 	unsigned long uas_bm[1];
591 	atomic_t num_in_q;
592 	char stopped;		/* TODO: should be atomic */
593 	bool used;
594 };
595 
596 struct sdebug_host_info {
597 	struct list_head host_list;
598 	struct Scsi_Host *shost;
599 	struct device dev;
600 	struct list_head dev_info_list;
601 };
602 
603 #define to_sdebug_host(d)	\
604 	container_of(d, struct sdebug_host_info, dev)
605 
606 static LIST_HEAD(sdebug_host_list);
607 static DEFINE_SPINLOCK(sdebug_host_list_lock);
608 
609 
610 struct sdebug_hrtimer {		/* ... is derived from hrtimer */
611 	struct hrtimer hrt;	/* must be first element */
612 	int qa_indx;
613 };
614 
615 struct sdebug_queued_cmd {
616 	/* in_use flagged by a bit in queued_in_use_bm[] */
617 	struct timer_list *cmnd_timerp;
618 	struct tasklet_struct *tletp;
619 	struct sdebug_hrtimer *sd_hrtp;
620 	struct scsi_cmnd * a_cmnd;
621 };
622 static struct sdebug_queued_cmd queued_arr[SCSI_DEBUG_CANQUEUE];
623 static unsigned long queued_in_use_bm[SCSI_DEBUG_CANQUEUE_WORDS];
624 
625 
626 static unsigned char * fake_storep;	/* ramdisk storage */
627 static struct sd_dif_tuple *dif_storep;	/* protection info */
628 static void *map_storep;		/* provisioning map */
629 
630 static unsigned long map_size;
631 static int num_aborts;
632 static int num_dev_resets;
633 static int num_target_resets;
634 static int num_bus_resets;
635 static int num_host_resets;
636 static int dix_writes;
637 static int dix_reads;
638 static int dif_errors;
639 
640 static DEFINE_SPINLOCK(queued_arr_lock);
641 static DEFINE_RWLOCK(atomic_rw);
642 
643 static char sdebug_proc_name[] = MY_NAME;
644 static const char *my_name = MY_NAME;
645 
646 static struct bus_type pseudo_lld_bus;
647 
648 static struct device_driver sdebug_driverfs_driver = {
649 	.name 		= sdebug_proc_name,
650 	.bus		= &pseudo_lld_bus,
651 };
652 
653 static const int check_condition_result =
654 		(DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION;
655 
656 static const int illegal_condition_result =
657 	(DRIVER_SENSE << 24) | (DID_ABORT << 16) | SAM_STAT_CHECK_CONDITION;
658 
659 static const int device_qfull_result =
660 	(DID_OK << 16) | (COMMAND_COMPLETE << 8) | SAM_STAT_TASK_SET_FULL;
661 
662 static unsigned char caching_pg[] = {0x8, 18, 0x14, 0, 0xff, 0xff, 0, 0,
663 				     0xff, 0xff, 0xff, 0xff, 0x80, 0x14, 0, 0,
664 				     0, 0, 0, 0};
665 static unsigned char ctrl_m_pg[] = {0xa, 10, 2, 0, 0, 0, 0, 0,
666 				    0, 0, 0x2, 0x4b};
667 static unsigned char iec_m_pg[] = {0x1c, 0xa, 0x08, 0, 0, 0, 0, 0,
668 			           0, 0, 0x0, 0x0};
669 
670 static void *fake_store(unsigned long long lba)
671 {
672 	lba = do_div(lba, sdebug_store_sectors);
673 
674 	return fake_storep + lba * scsi_debug_sector_size;
675 }
676 
677 static struct sd_dif_tuple *dif_store(sector_t sector)
678 {
679 	sector = do_div(sector, sdebug_store_sectors);
680 
681 	return dif_storep + sector;
682 }
683 
684 static int sdebug_add_adapter(void);
685 static void sdebug_remove_adapter(void);
686 
687 static void sdebug_max_tgts_luns(void)
688 {
689 	struct sdebug_host_info *sdbg_host;
690 	struct Scsi_Host *hpnt;
691 
692 	spin_lock(&sdebug_host_list_lock);
693 	list_for_each_entry(sdbg_host, &sdebug_host_list, host_list) {
694 		hpnt = sdbg_host->shost;
695 		if ((hpnt->this_id >= 0) &&
696 		    (scsi_debug_num_tgts > hpnt->this_id))
697 			hpnt->max_id = scsi_debug_num_tgts + 1;
698 		else
699 			hpnt->max_id = scsi_debug_num_tgts;
700 		/* scsi_debug_max_luns; */
701 		hpnt->max_lun = SAM2_WLUN_REPORT_LUNS;
702 	}
703 	spin_unlock(&sdebug_host_list_lock);
704 }
705 
706 enum sdeb_cmd_data {SDEB_IN_DATA = 0, SDEB_IN_CDB = 1};
707 
708 /* Set in_bit to -1 to indicate no bit position of invalid field */
709 static void
710 mk_sense_invalid_fld(struct scsi_cmnd *scp, enum sdeb_cmd_data c_d,
711 		     int in_byte, int in_bit)
712 {
713 	unsigned char *sbuff;
714 	u8 sks[4];
715 	int sl, asc;
716 
717 	sbuff = scp->sense_buffer;
718 	if (!sbuff) {
719 		sdev_printk(KERN_ERR, scp->device,
720 			    "%s: sense_buffer is NULL\n", __func__);
721 		return;
722 	}
723 	asc = c_d ? INVALID_FIELD_IN_CDB : INVALID_FIELD_IN_PARAM_LIST;
724 	memset(sbuff, 0, SCSI_SENSE_BUFFERSIZE);
725 	scsi_build_sense_buffer(scsi_debug_dsense, sbuff, ILLEGAL_REQUEST,
726 				asc, 0);
727 	memset(sks, 0, sizeof(sks));
728 	sks[0] = 0x80;
729 	if (c_d)
730 		sks[0] |= 0x40;
731 	if (in_bit >= 0) {
732 		sks[0] |= 0x8;
733 		sks[0] |= 0x7 & in_bit;
734 	}
735 	put_unaligned_be16(in_byte, sks + 1);
736 	if (scsi_debug_dsense) {
737 		sl = sbuff[7] + 8;
738 		sbuff[7] = sl;
739 		sbuff[sl] = 0x2;
740 		sbuff[sl + 1] = 0x6;
741 		memcpy(sbuff + sl + 4, sks, 3);
742 	} else
743 		memcpy(sbuff + 15, sks, 3);
744 	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
745 		sdev_printk(KERN_INFO, scp->device, "%s:  [sense_key,asc,ascq"
746 			    "]: [0x5,0x%x,0x0] %c byte=%d, bit=%d\n",
747 			    my_name, asc, c_d ? 'C' : 'D', in_byte, in_bit);
748 }
749 
750 static void mk_sense_buffer(struct scsi_cmnd *scp, int key, int asc, int asq)
751 {
752 	unsigned char *sbuff;
753 
754 	sbuff = scp->sense_buffer;
755 	if (!sbuff) {
756 		sdev_printk(KERN_ERR, scp->device,
757 			    "%s: sense_buffer is NULL\n", __func__);
758 		return;
759 	}
760 	memset(sbuff, 0, SCSI_SENSE_BUFFERSIZE);
761 
762 	scsi_build_sense_buffer(scsi_debug_dsense, sbuff, key, asc, asq);
763 
764 	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
765 		sdev_printk(KERN_INFO, scp->device,
766 			    "%s:  [sense_key,asc,ascq]: [0x%x,0x%x,0x%x]\n",
767 			    my_name, key, asc, asq);
768 }
769 
770 static void
771 mk_sense_invalid_opcode(struct scsi_cmnd *scp)
772 {
773 	mk_sense_buffer(scp, ILLEGAL_REQUEST, INVALID_OPCODE, 0);
774 }
775 
776 static int scsi_debug_ioctl(struct scsi_device *dev, int cmd, void __user *arg)
777 {
778 	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts) {
779 		if (0x1261 == cmd)
780 			sdev_printk(KERN_INFO, dev,
781 				    "%s: BLKFLSBUF [0x1261]\n", __func__);
782 		else if (0x5331 == cmd)
783 			sdev_printk(KERN_INFO, dev,
784 				    "%s: CDROM_GET_CAPABILITY [0x5331]\n",
785 				    __func__);
786 		else
787 			sdev_printk(KERN_INFO, dev, "%s: cmd=0x%x\n",
788 				    __func__, cmd);
789 	}
790 	return -EINVAL;
791 	/* return -ENOTTY; // correct return but upsets fdisk */
792 }
793 
794 static void clear_luns_changed_on_target(struct sdebug_dev_info *devip)
795 {
796 	struct sdebug_host_info *sdhp;
797 	struct sdebug_dev_info *dp;
798 
799 	spin_lock(&sdebug_host_list_lock);
800 	list_for_each_entry(sdhp, &sdebug_host_list, host_list) {
801 		list_for_each_entry(dp, &sdhp->dev_info_list, dev_list) {
802 			if ((devip->sdbg_host == dp->sdbg_host) &&
803 			    (devip->target == dp->target))
804 				clear_bit(SDEBUG_UA_LUNS_CHANGED, dp->uas_bm);
805 		}
806 	}
807 	spin_unlock(&sdebug_host_list_lock);
808 }
809 
810 static int check_readiness(struct scsi_cmnd *SCpnt, int uas_only,
811 			   struct sdebug_dev_info * devip)
812 {
813 	int k;
814 	bool debug = !!(SCSI_DEBUG_OPT_NOISE & scsi_debug_opts);
815 
816 	k = find_first_bit(devip->uas_bm, SDEBUG_NUM_UAS);
817 	if (k != SDEBUG_NUM_UAS) {
818 		const char *cp = NULL;
819 
820 		switch (k) {
821 		case SDEBUG_UA_POR:
822 			mk_sense_buffer(SCpnt, UNIT_ATTENTION,
823 					UA_RESET_ASC, POWER_ON_RESET_ASCQ);
824 			if (debug)
825 				cp = "power on reset";
826 			break;
827 		case SDEBUG_UA_BUS_RESET:
828 			mk_sense_buffer(SCpnt, UNIT_ATTENTION,
829 					UA_RESET_ASC, BUS_RESET_ASCQ);
830 			if (debug)
831 				cp = "bus reset";
832 			break;
833 		case SDEBUG_UA_MODE_CHANGED:
834 			mk_sense_buffer(SCpnt, UNIT_ATTENTION,
835 					UA_CHANGED_ASC, MODE_CHANGED_ASCQ);
836 			if (debug)
837 				cp = "mode parameters changed";
838 			break;
839 		case SDEBUG_UA_CAPACITY_CHANGED:
840 			mk_sense_buffer(SCpnt, UNIT_ATTENTION,
841 					UA_CHANGED_ASC, CAPACITY_CHANGED_ASCQ);
842 			if (debug)
843 				cp = "capacity data changed";
844 			break;
845 		case SDEBUG_UA_MICROCODE_CHANGED:
846 			mk_sense_buffer(SCpnt, UNIT_ATTENTION,
847 				 TARGET_CHANGED_ASC, MICROCODE_CHANGED_ASCQ);
848 			if (debug)
849 				cp = "microcode has been changed";
850 			break;
851 		case SDEBUG_UA_MICROCODE_CHANGED_WO_RESET:
852 			mk_sense_buffer(SCpnt, UNIT_ATTENTION,
853 					TARGET_CHANGED_ASC,
854 					MICROCODE_CHANGED_WO_RESET_ASCQ);
855 			if (debug)
856 				cp = "microcode has been changed without reset";
857 			break;
858 		case SDEBUG_UA_LUNS_CHANGED:
859 			/*
860 			 * SPC-3 behavior is to report a UNIT ATTENTION with
861 			 * ASC/ASCQ REPORTED LUNS DATA HAS CHANGED on every LUN
862 			 * on the target, until a REPORT LUNS command is
863 			 * received.  SPC-4 behavior is to report it only once.
864 			 * NOTE:  scsi_debug_scsi_level does not use the same
865 			 * values as struct scsi_device->scsi_level.
866 			 */
867 			if (scsi_debug_scsi_level >= 6)	/* SPC-4 and above */
868 				clear_luns_changed_on_target(devip);
869 			mk_sense_buffer(SCpnt, UNIT_ATTENTION,
870 					TARGET_CHANGED_ASC,
871 					LUNS_CHANGED_ASCQ);
872 			if (debug)
873 				cp = "reported luns data has changed";
874 			break;
875 		default:
876 			pr_warn("%s: unexpected unit attention code=%d\n",
877 				__func__, k);
878 			if (debug)
879 				cp = "unknown";
880 			break;
881 		}
882 		clear_bit(k, devip->uas_bm);
883 		if (debug)
884 			sdev_printk(KERN_INFO, SCpnt->device,
885 				   "%s reports: Unit attention: %s\n",
886 				   my_name, cp);
887 		return check_condition_result;
888 	}
889 	if ((UAS_TUR == uas_only) && devip->stopped) {
890 		mk_sense_buffer(SCpnt, NOT_READY, LOGICAL_UNIT_NOT_READY,
891 				0x2);
892 		if (debug)
893 			sdev_printk(KERN_INFO, SCpnt->device,
894 				    "%s reports: Not ready: %s\n", my_name,
895 				    "initializing command required");
896 		return check_condition_result;
897 	}
898 	return 0;
899 }
900 
901 /* Returns 0 if ok else (DID_ERROR << 16). Sets scp->resid . */
902 static int fill_from_dev_buffer(struct scsi_cmnd *scp, unsigned char *arr,
903 				int arr_len)
904 {
905 	int act_len;
906 	struct scsi_data_buffer *sdb = scsi_in(scp);
907 
908 	if (!sdb->length)
909 		return 0;
910 	if (!(scsi_bidi_cmnd(scp) || scp->sc_data_direction == DMA_FROM_DEVICE))
911 		return (DID_ERROR << 16);
912 
913 	act_len = sg_copy_from_buffer(sdb->table.sgl, sdb->table.nents,
914 				      arr, arr_len);
915 	sdb->resid = scsi_bufflen(scp) - act_len;
916 
917 	return 0;
918 }
919 
920 /* Returns number of bytes fetched into 'arr' or -1 if error. */
921 static int fetch_to_dev_buffer(struct scsi_cmnd *scp, unsigned char *arr,
922 			       int arr_len)
923 {
924 	if (!scsi_bufflen(scp))
925 		return 0;
926 	if (!(scsi_bidi_cmnd(scp) || scp->sc_data_direction == DMA_TO_DEVICE))
927 		return -1;
928 
929 	return scsi_sg_copy_to_buffer(scp, arr, arr_len);
930 }
931 
932 
933 static const char * inq_vendor_id = "Linux   ";
934 static const char * inq_product_id = "scsi_debug      ";
935 static const char *inq_product_rev = "0184";	/* version less '.' */
936 
937 /* Device identification VPD page. Returns number of bytes placed in arr */
938 static int inquiry_evpd_83(unsigned char * arr, int port_group_id,
939 			   int target_dev_id, int dev_id_num,
940 			   const char * dev_id_str,
941 			   int dev_id_str_len)
942 {
943 	int num, port_a;
944 	char b[32];
945 
946 	port_a = target_dev_id + 1;
947 	/* T10 vendor identifier field format (faked) */
948 	arr[0] = 0x2;	/* ASCII */
949 	arr[1] = 0x1;
950 	arr[2] = 0x0;
951 	memcpy(&arr[4], inq_vendor_id, 8);
952 	memcpy(&arr[12], inq_product_id, 16);
953 	memcpy(&arr[28], dev_id_str, dev_id_str_len);
954 	num = 8 + 16 + dev_id_str_len;
955 	arr[3] = num;
956 	num += 4;
957 	if (dev_id_num >= 0) {
958 		/* NAA-5, Logical unit identifier (binary) */
959 		arr[num++] = 0x1;	/* binary (not necessarily sas) */
960 		arr[num++] = 0x3;	/* PIV=0, lu, naa */
961 		arr[num++] = 0x0;
962 		arr[num++] = 0x8;
963 		arr[num++] = 0x53;  /* naa-5 ieee company id=0x333333 (fake) */
964 		arr[num++] = 0x33;
965 		arr[num++] = 0x33;
966 		arr[num++] = 0x30;
967 		arr[num++] = (dev_id_num >> 24);
968 		arr[num++] = (dev_id_num >> 16) & 0xff;
969 		arr[num++] = (dev_id_num >> 8) & 0xff;
970 		arr[num++] = dev_id_num & 0xff;
971 		/* Target relative port number */
972 		arr[num++] = 0x61;	/* proto=sas, binary */
973 		arr[num++] = 0x94;	/* PIV=1, target port, rel port */
974 		arr[num++] = 0x0;	/* reserved */
975 		arr[num++] = 0x4;	/* length */
976 		arr[num++] = 0x0;	/* reserved */
977 		arr[num++] = 0x0;	/* reserved */
978 		arr[num++] = 0x0;
979 		arr[num++] = 0x1;	/* relative port A */
980 	}
981 	/* NAA-5, Target port identifier */
982 	arr[num++] = 0x61;	/* proto=sas, binary */
983 	arr[num++] = 0x93;	/* piv=1, target port, naa */
984 	arr[num++] = 0x0;
985 	arr[num++] = 0x8;
986 	arr[num++] = 0x52;	/* naa-5, company id=0x222222 (fake) */
987 	arr[num++] = 0x22;
988 	arr[num++] = 0x22;
989 	arr[num++] = 0x20;
990 	arr[num++] = (port_a >> 24);
991 	arr[num++] = (port_a >> 16) & 0xff;
992 	arr[num++] = (port_a >> 8) & 0xff;
993 	arr[num++] = port_a & 0xff;
994 	/* NAA-5, Target port group identifier */
995 	arr[num++] = 0x61;	/* proto=sas, binary */
996 	arr[num++] = 0x95;	/* piv=1, target port group id */
997 	arr[num++] = 0x0;
998 	arr[num++] = 0x4;
999 	arr[num++] = 0;
1000 	arr[num++] = 0;
1001 	arr[num++] = (port_group_id >> 8) & 0xff;
1002 	arr[num++] = port_group_id & 0xff;
1003 	/* NAA-5, Target device identifier */
1004 	arr[num++] = 0x61;	/* proto=sas, binary */
1005 	arr[num++] = 0xa3;	/* piv=1, target device, naa */
1006 	arr[num++] = 0x0;
1007 	arr[num++] = 0x8;
1008 	arr[num++] = 0x52;	/* naa-5, company id=0x222222 (fake) */
1009 	arr[num++] = 0x22;
1010 	arr[num++] = 0x22;
1011 	arr[num++] = 0x20;
1012 	arr[num++] = (target_dev_id >> 24);
1013 	arr[num++] = (target_dev_id >> 16) & 0xff;
1014 	arr[num++] = (target_dev_id >> 8) & 0xff;
1015 	arr[num++] = target_dev_id & 0xff;
1016 	/* SCSI name string: Target device identifier */
1017 	arr[num++] = 0x63;	/* proto=sas, UTF-8 */
1018 	arr[num++] = 0xa8;	/* piv=1, target device, SCSI name string */
1019 	arr[num++] = 0x0;
1020 	arr[num++] = 24;
1021 	memcpy(arr + num, "naa.52222220", 12);
1022 	num += 12;
1023 	snprintf(b, sizeof(b), "%08X", target_dev_id);
1024 	memcpy(arr + num, b, 8);
1025 	num += 8;
1026 	memset(arr + num, 0, 4);
1027 	num += 4;
1028 	return num;
1029 }
1030 
1031 
1032 static unsigned char vpd84_data[] = {
1033 /* from 4th byte */ 0x22,0x22,0x22,0x0,0xbb,0x0,
1034     0x22,0x22,0x22,0x0,0xbb,0x1,
1035     0x22,0x22,0x22,0x0,0xbb,0x2,
1036 };
1037 
1038 /*  Software interface identification VPD page */
1039 static int inquiry_evpd_84(unsigned char * arr)
1040 {
1041 	memcpy(arr, vpd84_data, sizeof(vpd84_data));
1042 	return sizeof(vpd84_data);
1043 }
1044 
1045 /* Management network addresses VPD page */
1046 static int inquiry_evpd_85(unsigned char * arr)
1047 {
1048 	int num = 0;
1049 	const char * na1 = "https://www.kernel.org/config";
1050 	const char * na2 = "http://www.kernel.org/log";
1051 	int plen, olen;
1052 
1053 	arr[num++] = 0x1;	/* lu, storage config */
1054 	arr[num++] = 0x0;	/* reserved */
1055 	arr[num++] = 0x0;
1056 	olen = strlen(na1);
1057 	plen = olen + 1;
1058 	if (plen % 4)
1059 		plen = ((plen / 4) + 1) * 4;
1060 	arr[num++] = plen;	/* length, null termianted, padded */
1061 	memcpy(arr + num, na1, olen);
1062 	memset(arr + num + olen, 0, plen - olen);
1063 	num += plen;
1064 
1065 	arr[num++] = 0x4;	/* lu, logging */
1066 	arr[num++] = 0x0;	/* reserved */
1067 	arr[num++] = 0x0;
1068 	olen = strlen(na2);
1069 	plen = olen + 1;
1070 	if (plen % 4)
1071 		plen = ((plen / 4) + 1) * 4;
1072 	arr[num++] = plen;	/* length, null terminated, padded */
1073 	memcpy(arr + num, na2, olen);
1074 	memset(arr + num + olen, 0, plen - olen);
1075 	num += plen;
1076 
1077 	return num;
1078 }
1079 
1080 /* SCSI ports VPD page */
1081 static int inquiry_evpd_88(unsigned char * arr, int target_dev_id)
1082 {
1083 	int num = 0;
1084 	int port_a, port_b;
1085 
1086 	port_a = target_dev_id + 1;
1087 	port_b = port_a + 1;
1088 	arr[num++] = 0x0;	/* reserved */
1089 	arr[num++] = 0x0;	/* reserved */
1090 	arr[num++] = 0x0;
1091 	arr[num++] = 0x1;	/* relative port 1 (primary) */
1092 	memset(arr + num, 0, 6);
1093 	num += 6;
1094 	arr[num++] = 0x0;
1095 	arr[num++] = 12;	/* length tp descriptor */
1096 	/* naa-5 target port identifier (A) */
1097 	arr[num++] = 0x61;	/* proto=sas, binary */
1098 	arr[num++] = 0x93;	/* PIV=1, target port, NAA */
1099 	arr[num++] = 0x0;	/* reserved */
1100 	arr[num++] = 0x8;	/* length */
1101 	arr[num++] = 0x52;	/* NAA-5, company_id=0x222222 (fake) */
1102 	arr[num++] = 0x22;
1103 	arr[num++] = 0x22;
1104 	arr[num++] = 0x20;
1105 	arr[num++] = (port_a >> 24);
1106 	arr[num++] = (port_a >> 16) & 0xff;
1107 	arr[num++] = (port_a >> 8) & 0xff;
1108 	arr[num++] = port_a & 0xff;
1109 
1110 	arr[num++] = 0x0;	/* reserved */
1111 	arr[num++] = 0x0;	/* reserved */
1112 	arr[num++] = 0x0;
1113 	arr[num++] = 0x2;	/* relative port 2 (secondary) */
1114 	memset(arr + num, 0, 6);
1115 	num += 6;
1116 	arr[num++] = 0x0;
1117 	arr[num++] = 12;	/* length tp descriptor */
1118 	/* naa-5 target port identifier (B) */
1119 	arr[num++] = 0x61;	/* proto=sas, binary */
1120 	arr[num++] = 0x93;	/* PIV=1, target port, NAA */
1121 	arr[num++] = 0x0;	/* reserved */
1122 	arr[num++] = 0x8;	/* length */
1123 	arr[num++] = 0x52;	/* NAA-5, company_id=0x222222 (fake) */
1124 	arr[num++] = 0x22;
1125 	arr[num++] = 0x22;
1126 	arr[num++] = 0x20;
1127 	arr[num++] = (port_b >> 24);
1128 	arr[num++] = (port_b >> 16) & 0xff;
1129 	arr[num++] = (port_b >> 8) & 0xff;
1130 	arr[num++] = port_b & 0xff;
1131 
1132 	return num;
1133 }
1134 
1135 
1136 static unsigned char vpd89_data[] = {
1137 /* from 4th byte */ 0,0,0,0,
1138 'l','i','n','u','x',' ',' ',' ',
1139 'S','A','T',' ','s','c','s','i','_','d','e','b','u','g',' ',' ',
1140 '1','2','3','4',
1141 0x34,0,0,0,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,
1142 0xec,0,0,0,
1143 0x5a,0xc,0xff,0x3f,0x37,0xc8,0x10,0,0,0,0,0,0x3f,0,0,0,
1144 0,0,0,0,0x58,0x58,0x58,0x58,0x58,0x58,0x58,0x58,0x20,0x20,0x20,0x20,
1145 0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0,0,0,0x40,0x4,0,0x2e,0x33,
1146 0x38,0x31,0x20,0x20,0x20,0x20,0x54,0x53,0x38,0x33,0x30,0x30,0x33,0x31,
1147 0x53,0x41,
1148 0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,
1149 0x20,0x20,
1150 0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,0x20,
1151 0x10,0x80,
1152 0,0,0,0x2f,0,0,0,0x2,0,0x2,0x7,0,0xff,0xff,0x1,0,
1153 0x3f,0,0xc1,0xff,0x3e,0,0x10,0x1,0xb0,0xf8,0x50,0x9,0,0,0x7,0,
1154 0x3,0,0x78,0,0x78,0,0xf0,0,0x78,0,0,0,0,0,0,0,
1155 0,0,0,0,0,0,0,0,0x2,0,0,0,0,0,0,0,
1156 0x7e,0,0x1b,0,0x6b,0x34,0x1,0x7d,0x3,0x40,0x69,0x34,0x1,0x3c,0x3,0x40,
1157 0x7f,0x40,0,0,0,0,0xfe,0xfe,0,0,0,0,0,0xfe,0,0,
1158 0,0,0,0,0,0,0,0,0xb0,0xf8,0x50,0x9,0,0,0,0,
1159 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1160 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1161 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1162 0x1,0,0xb0,0xf8,0x50,0x9,0xb0,0xf8,0x50,0x9,0x20,0x20,0x2,0,0xb6,0x42,
1163 0,0x80,0x8a,0,0x6,0x3c,0xa,0x3c,0xff,0xff,0xc6,0x7,0,0x1,0,0x8,
1164 0xf0,0xf,0,0x10,0x2,0,0x30,0,0,0,0,0,0,0,0x6,0xfe,
1165 0,0,0x2,0,0x50,0,0x8a,0,0x4f,0x95,0,0,0x21,0,0xb,0,
1166 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1167 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1168 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1169 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1170 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1171 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1172 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1173 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1174 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1175 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1176 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1177 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0xa5,0x51,
1178 };
1179 
1180 /* ATA Information VPD page */
1181 static int inquiry_evpd_89(unsigned char * arr)
1182 {
1183 	memcpy(arr, vpd89_data, sizeof(vpd89_data));
1184 	return sizeof(vpd89_data);
1185 }
1186 
1187 
1188 static unsigned char vpdb0_data[] = {
1189 	/* from 4th byte */ 0,0,0,4, 0,0,0x4,0, 0,0,0,64,
1190 	0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1191 	0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1192 	0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
1193 };
1194 
1195 /* Block limits VPD page (SBC-3) */
1196 static int inquiry_evpd_b0(unsigned char * arr)
1197 {
1198 	unsigned int gran;
1199 
1200 	memcpy(arr, vpdb0_data, sizeof(vpdb0_data));
1201 
1202 	/* Optimal transfer length granularity */
1203 	gran = 1 << scsi_debug_physblk_exp;
1204 	arr[2] = (gran >> 8) & 0xff;
1205 	arr[3] = gran & 0xff;
1206 
1207 	/* Maximum Transfer Length */
1208 	if (sdebug_store_sectors > 0x400) {
1209 		arr[4] = (sdebug_store_sectors >> 24) & 0xff;
1210 		arr[5] = (sdebug_store_sectors >> 16) & 0xff;
1211 		arr[6] = (sdebug_store_sectors >> 8) & 0xff;
1212 		arr[7] = sdebug_store_sectors & 0xff;
1213 	}
1214 
1215 	/* Optimal Transfer Length */
1216 	put_unaligned_be32(scsi_debug_opt_blks, &arr[8]);
1217 
1218 	if (scsi_debug_lbpu) {
1219 		/* Maximum Unmap LBA Count */
1220 		put_unaligned_be32(scsi_debug_unmap_max_blocks, &arr[16]);
1221 
1222 		/* Maximum Unmap Block Descriptor Count */
1223 		put_unaligned_be32(scsi_debug_unmap_max_desc, &arr[20]);
1224 	}
1225 
1226 	/* Unmap Granularity Alignment */
1227 	if (scsi_debug_unmap_alignment) {
1228 		put_unaligned_be32(scsi_debug_unmap_alignment, &arr[28]);
1229 		arr[28] |= 0x80; /* UGAVALID */
1230 	}
1231 
1232 	/* Optimal Unmap Granularity */
1233 	put_unaligned_be32(scsi_debug_unmap_granularity, &arr[24]);
1234 
1235 	/* Maximum WRITE SAME Length */
1236 	put_unaligned_be64(scsi_debug_write_same_length, &arr[32]);
1237 
1238 	return 0x3c; /* Mandatory page length for Logical Block Provisioning */
1239 
1240 	return sizeof(vpdb0_data);
1241 }
1242 
1243 /* Block device characteristics VPD page (SBC-3) */
1244 static int inquiry_evpd_b1(unsigned char *arr)
1245 {
1246 	memset(arr, 0, 0x3c);
1247 	arr[0] = 0;
1248 	arr[1] = 1;	/* non rotating medium (e.g. solid state) */
1249 	arr[2] = 0;
1250 	arr[3] = 5;	/* less than 1.8" */
1251 
1252 	return 0x3c;
1253 }
1254 
1255 /* Logical block provisioning VPD page (SBC-3) */
1256 static int inquiry_evpd_b2(unsigned char *arr)
1257 {
1258 	memset(arr, 0, 0x4);
1259 	arr[0] = 0;			/* threshold exponent */
1260 
1261 	if (scsi_debug_lbpu)
1262 		arr[1] = 1 << 7;
1263 
1264 	if (scsi_debug_lbpws)
1265 		arr[1] |= 1 << 6;
1266 
1267 	if (scsi_debug_lbpws10)
1268 		arr[1] |= 1 << 5;
1269 
1270 	if (scsi_debug_lbprz)
1271 		arr[1] |= 1 << 2;
1272 
1273 	return 0x4;
1274 }
1275 
1276 #define SDEBUG_LONG_INQ_SZ 96
1277 #define SDEBUG_MAX_INQ_ARR_SZ 584
1278 
1279 static int resp_inquiry(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
1280 {
1281 	unsigned char pq_pdt;
1282 	unsigned char * arr;
1283 	unsigned char *cmd = scp->cmnd;
1284 	int alloc_len, n, ret;
1285 	bool have_wlun;
1286 
1287 	alloc_len = (cmd[3] << 8) + cmd[4];
1288 	arr = kzalloc(SDEBUG_MAX_INQ_ARR_SZ, GFP_ATOMIC);
1289 	if (! arr)
1290 		return DID_REQUEUE << 16;
1291 	have_wlun = (scp->device->lun == SAM2_WLUN_REPORT_LUNS);
1292 	if (have_wlun)
1293 		pq_pdt = 0x1e;	/* present, wlun */
1294 	else if (scsi_debug_no_lun_0 && (0 == devip->lun))
1295 		pq_pdt = 0x7f;	/* not present, no device type */
1296 	else
1297 		pq_pdt = (scsi_debug_ptype & 0x1f);
1298 	arr[0] = pq_pdt;
1299 	if (0x2 & cmd[1]) {  /* CMDDT bit set */
1300 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 1, 1);
1301 		kfree(arr);
1302 		return check_condition_result;
1303 	} else if (0x1 & cmd[1]) {  /* EVPD bit set */
1304 		int lu_id_num, port_group_id, target_dev_id, len;
1305 		char lu_id_str[6];
1306 		int host_no = devip->sdbg_host->shost->host_no;
1307 
1308 		port_group_id = (((host_no + 1) & 0x7f) << 8) +
1309 		    (devip->channel & 0x7f);
1310 		if (0 == scsi_debug_vpd_use_hostno)
1311 			host_no = 0;
1312 		lu_id_num = have_wlun ? -1 : (((host_no + 1) * 2000) +
1313 			    (devip->target * 1000) + devip->lun);
1314 		target_dev_id = ((host_no + 1) * 2000) +
1315 				 (devip->target * 1000) - 3;
1316 		len = scnprintf(lu_id_str, 6, "%d", lu_id_num);
1317 		if (0 == cmd[2]) { /* supported vital product data pages */
1318 			arr[1] = cmd[2];	/*sanity */
1319 			n = 4;
1320 			arr[n++] = 0x0;   /* this page */
1321 			arr[n++] = 0x80;  /* unit serial number */
1322 			arr[n++] = 0x83;  /* device identification */
1323 			arr[n++] = 0x84;  /* software interface ident. */
1324 			arr[n++] = 0x85;  /* management network addresses */
1325 			arr[n++] = 0x86;  /* extended inquiry */
1326 			arr[n++] = 0x87;  /* mode page policy */
1327 			arr[n++] = 0x88;  /* SCSI ports */
1328 			arr[n++] = 0x89;  /* ATA information */
1329 			arr[n++] = 0xb0;  /* Block limits (SBC) */
1330 			arr[n++] = 0xb1;  /* Block characteristics (SBC) */
1331 			if (scsi_debug_lbp()) /* Logical Block Prov. (SBC) */
1332 				arr[n++] = 0xb2;
1333 			arr[3] = n - 4;	  /* number of supported VPD pages */
1334 		} else if (0x80 == cmd[2]) { /* unit serial number */
1335 			arr[1] = cmd[2];	/*sanity */
1336 			arr[3] = len;
1337 			memcpy(&arr[4], lu_id_str, len);
1338 		} else if (0x83 == cmd[2]) { /* device identification */
1339 			arr[1] = cmd[2];	/*sanity */
1340 			arr[3] = inquiry_evpd_83(&arr[4], port_group_id,
1341 						 target_dev_id, lu_id_num,
1342 						 lu_id_str, len);
1343 		} else if (0x84 == cmd[2]) { /* Software interface ident. */
1344 			arr[1] = cmd[2];	/*sanity */
1345 			arr[3] = inquiry_evpd_84(&arr[4]);
1346 		} else if (0x85 == cmd[2]) { /* Management network addresses */
1347 			arr[1] = cmd[2];	/*sanity */
1348 			arr[3] = inquiry_evpd_85(&arr[4]);
1349 		} else if (0x86 == cmd[2]) { /* extended inquiry */
1350 			arr[1] = cmd[2];	/*sanity */
1351 			arr[3] = 0x3c;	/* number of following entries */
1352 			if (scsi_debug_dif == SD_DIF_TYPE3_PROTECTION)
1353 				arr[4] = 0x4;	/* SPT: GRD_CHK:1 */
1354 			else if (scsi_debug_dif)
1355 				arr[4] = 0x5;   /* SPT: GRD_CHK:1, REF_CHK:1 */
1356 			else
1357 				arr[4] = 0x0;   /* no protection stuff */
1358 			arr[5] = 0x7;   /* head of q, ordered + simple q's */
1359 		} else if (0x87 == cmd[2]) { /* mode page policy */
1360 			arr[1] = cmd[2];	/*sanity */
1361 			arr[3] = 0x8;	/* number of following entries */
1362 			arr[4] = 0x2;	/* disconnect-reconnect mp */
1363 			arr[6] = 0x80;	/* mlus, shared */
1364 			arr[8] = 0x18;	 /* protocol specific lu */
1365 			arr[10] = 0x82;	 /* mlus, per initiator port */
1366 		} else if (0x88 == cmd[2]) { /* SCSI Ports */
1367 			arr[1] = cmd[2];	/*sanity */
1368 			arr[3] = inquiry_evpd_88(&arr[4], target_dev_id);
1369 		} else if (0x89 == cmd[2]) { /* ATA information */
1370 			arr[1] = cmd[2];        /*sanity */
1371 			n = inquiry_evpd_89(&arr[4]);
1372 			arr[2] = (n >> 8);
1373 			arr[3] = (n & 0xff);
1374 		} else if (0xb0 == cmd[2]) { /* Block limits (SBC) */
1375 			arr[1] = cmd[2];        /*sanity */
1376 			arr[3] = inquiry_evpd_b0(&arr[4]);
1377 		} else if (0xb1 == cmd[2]) { /* Block characteristics (SBC) */
1378 			arr[1] = cmd[2];        /*sanity */
1379 			arr[3] = inquiry_evpd_b1(&arr[4]);
1380 		} else if (0xb2 == cmd[2]) { /* Logical Block Prov. (SBC) */
1381 			arr[1] = cmd[2];        /*sanity */
1382 			arr[3] = inquiry_evpd_b2(&arr[4]);
1383 		} else {
1384 			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 2, -1);
1385 			kfree(arr);
1386 			return check_condition_result;
1387 		}
1388 		len = min(((arr[2] << 8) + arr[3]) + 4, alloc_len);
1389 		ret = fill_from_dev_buffer(scp, arr,
1390 			    min(len, SDEBUG_MAX_INQ_ARR_SZ));
1391 		kfree(arr);
1392 		return ret;
1393 	}
1394 	/* drops through here for a standard inquiry */
1395 	arr[1] = scsi_debug_removable ? 0x80 : 0;	/* Removable disk */
1396 	arr[2] = scsi_debug_scsi_level;
1397 	arr[3] = 2;    /* response_data_format==2 */
1398 	arr[4] = SDEBUG_LONG_INQ_SZ - 5;
1399 	arr[5] = scsi_debug_dif ? 1 : 0; /* PROTECT bit */
1400 	if (0 == scsi_debug_vpd_use_hostno)
1401 		arr[5] = 0x10; /* claim: implicit TGPS */
1402 	arr[6] = 0x10; /* claim: MultiP */
1403 	/* arr[6] |= 0x40; ... claim: EncServ (enclosure services) */
1404 	arr[7] = 0xa; /* claim: LINKED + CMDQUE */
1405 	memcpy(&arr[8], inq_vendor_id, 8);
1406 	memcpy(&arr[16], inq_product_id, 16);
1407 	memcpy(&arr[32], inq_product_rev, 4);
1408 	/* version descriptors (2 bytes each) follow */
1409 	arr[58] = 0x0; arr[59] = 0xa2;  /* SAM-5 rev 4 */
1410 	arr[60] = 0x4; arr[61] = 0x68;  /* SPC-4 rev 37 */
1411 	n = 62;
1412 	if (scsi_debug_ptype == 0) {
1413 		arr[n++] = 0x4; arr[n++] = 0xc5; /* SBC-4 rev 36 */
1414 	} else if (scsi_debug_ptype == 1) {
1415 		arr[n++] = 0x5; arr[n++] = 0x25; /* SSC-4 rev 3 */
1416 	}
1417 	arr[n++] = 0x20; arr[n++] = 0xe6;  /* SPL-3 rev 7 */
1418 	ret = fill_from_dev_buffer(scp, arr,
1419 			    min(alloc_len, SDEBUG_LONG_INQ_SZ));
1420 	kfree(arr);
1421 	return ret;
1422 }
1423 
1424 static int resp_requests(struct scsi_cmnd * scp,
1425 			 struct sdebug_dev_info * devip)
1426 {
1427 	unsigned char * sbuff;
1428 	unsigned char *cmd = scp->cmnd;
1429 	unsigned char arr[SCSI_SENSE_BUFFERSIZE];
1430 	bool dsense, want_dsense;
1431 	int len = 18;
1432 
1433 	memset(arr, 0, sizeof(arr));
1434 	dsense = !!(cmd[1] & 1);
1435 	want_dsense = dsense || scsi_debug_dsense;
1436 	sbuff = scp->sense_buffer;
1437 	if ((iec_m_pg[2] & 0x4) && (6 == (iec_m_pg[3] & 0xf))) {
1438 		if (dsense) {
1439 			arr[0] = 0x72;
1440 			arr[1] = 0x0;		/* NO_SENSE in sense_key */
1441 			arr[2] = THRESHOLD_EXCEEDED;
1442 			arr[3] = 0xff;		/* TEST set and MRIE==6 */
1443 			len = 8;
1444 		} else {
1445 			arr[0] = 0x70;
1446 			arr[2] = 0x0;		/* NO_SENSE in sense_key */
1447 			arr[7] = 0xa;   	/* 18 byte sense buffer */
1448 			arr[12] = THRESHOLD_EXCEEDED;
1449 			arr[13] = 0xff;		/* TEST set and MRIE==6 */
1450 		}
1451 	} else {
1452 		memcpy(arr, sbuff, SCSI_SENSE_BUFFERSIZE);
1453 		if (arr[0] >= 0x70 && dsense == scsi_debug_dsense)
1454 			;	/* have sense and formats match */
1455 		else if (arr[0] <= 0x70) {
1456 			if (dsense) {
1457 				memset(arr, 0, 8);
1458 				arr[0] = 0x72;
1459 				len = 8;
1460 			} else {
1461 				memset(arr, 0, 18);
1462 				arr[0] = 0x70;
1463 				arr[7] = 0xa;
1464 			}
1465 		} else if (dsense) {
1466 			memset(arr, 0, 8);
1467 			arr[0] = 0x72;
1468 			arr[1] = sbuff[2];     /* sense key */
1469 			arr[2] = sbuff[12];    /* asc */
1470 			arr[3] = sbuff[13];    /* ascq */
1471 			len = 8;
1472 		} else {
1473 			memset(arr, 0, 18);
1474 			arr[0] = 0x70;
1475 			arr[2] = sbuff[1];
1476 			arr[7] = 0xa;
1477 			arr[12] = sbuff[1];
1478 			arr[13] = sbuff[3];
1479 		}
1480 
1481 	}
1482 	mk_sense_buffer(scp, 0, NO_ADDITIONAL_SENSE, 0);
1483 	return fill_from_dev_buffer(scp, arr, len);
1484 }
1485 
1486 static int resp_start_stop(struct scsi_cmnd * scp,
1487 			   struct sdebug_dev_info * devip)
1488 {
1489 	unsigned char *cmd = scp->cmnd;
1490 	int power_cond, start;
1491 
1492 	power_cond = (cmd[4] & 0xf0) >> 4;
1493 	if (power_cond) {
1494 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 4, 7);
1495 		return check_condition_result;
1496 	}
1497 	start = cmd[4] & 1;
1498 	if (start == devip->stopped)
1499 		devip->stopped = !start;
1500 	return 0;
1501 }
1502 
1503 static sector_t get_sdebug_capacity(void)
1504 {
1505 	if (scsi_debug_virtual_gb > 0)
1506 		return (sector_t)scsi_debug_virtual_gb *
1507 			(1073741824 / scsi_debug_sector_size);
1508 	else
1509 		return sdebug_store_sectors;
1510 }
1511 
1512 #define SDEBUG_READCAP_ARR_SZ 8
1513 static int resp_readcap(struct scsi_cmnd * scp,
1514 			struct sdebug_dev_info * devip)
1515 {
1516 	unsigned char arr[SDEBUG_READCAP_ARR_SZ];
1517 	unsigned int capac;
1518 
1519 	/* following just in case virtual_gb changed */
1520 	sdebug_capacity = get_sdebug_capacity();
1521 	memset(arr, 0, SDEBUG_READCAP_ARR_SZ);
1522 	if (sdebug_capacity < 0xffffffff) {
1523 		capac = (unsigned int)sdebug_capacity - 1;
1524 		arr[0] = (capac >> 24);
1525 		arr[1] = (capac >> 16) & 0xff;
1526 		arr[2] = (capac >> 8) & 0xff;
1527 		arr[3] = capac & 0xff;
1528 	} else {
1529 		arr[0] = 0xff;
1530 		arr[1] = 0xff;
1531 		arr[2] = 0xff;
1532 		arr[3] = 0xff;
1533 	}
1534 	arr[6] = (scsi_debug_sector_size >> 8) & 0xff;
1535 	arr[7] = scsi_debug_sector_size & 0xff;
1536 	return fill_from_dev_buffer(scp, arr, SDEBUG_READCAP_ARR_SZ);
1537 }
1538 
1539 #define SDEBUG_READCAP16_ARR_SZ 32
1540 static int resp_readcap16(struct scsi_cmnd * scp,
1541 			  struct sdebug_dev_info * devip)
1542 {
1543 	unsigned char *cmd = scp->cmnd;
1544 	unsigned char arr[SDEBUG_READCAP16_ARR_SZ];
1545 	unsigned long long capac;
1546 	int k, alloc_len;
1547 
1548 	alloc_len = ((cmd[10] << 24) + (cmd[11] << 16) + (cmd[12] << 8)
1549 		     + cmd[13]);
1550 	/* following just in case virtual_gb changed */
1551 	sdebug_capacity = get_sdebug_capacity();
1552 	memset(arr, 0, SDEBUG_READCAP16_ARR_SZ);
1553 	capac = sdebug_capacity - 1;
1554 	for (k = 0; k < 8; ++k, capac >>= 8)
1555 		arr[7 - k] = capac & 0xff;
1556 	arr[8] = (scsi_debug_sector_size >> 24) & 0xff;
1557 	arr[9] = (scsi_debug_sector_size >> 16) & 0xff;
1558 	arr[10] = (scsi_debug_sector_size >> 8) & 0xff;
1559 	arr[11] = scsi_debug_sector_size & 0xff;
1560 	arr[13] = scsi_debug_physblk_exp & 0xf;
1561 	arr[14] = (scsi_debug_lowest_aligned >> 8) & 0x3f;
1562 
1563 	if (scsi_debug_lbp()) {
1564 		arr[14] |= 0x80; /* LBPME */
1565 		if (scsi_debug_lbprz)
1566 			arr[14] |= 0x40; /* LBPRZ */
1567 	}
1568 
1569 	arr[15] = scsi_debug_lowest_aligned & 0xff;
1570 
1571 	if (scsi_debug_dif) {
1572 		arr[12] = (scsi_debug_dif - 1) << 1; /* P_TYPE */
1573 		arr[12] |= 1; /* PROT_EN */
1574 	}
1575 
1576 	return fill_from_dev_buffer(scp, arr,
1577 				    min(alloc_len, SDEBUG_READCAP16_ARR_SZ));
1578 }
1579 
1580 #define SDEBUG_MAX_TGTPGS_ARR_SZ 1412
1581 
1582 static int resp_report_tgtpgs(struct scsi_cmnd * scp,
1583 			      struct sdebug_dev_info * devip)
1584 {
1585 	unsigned char *cmd = scp->cmnd;
1586 	unsigned char * arr;
1587 	int host_no = devip->sdbg_host->shost->host_no;
1588 	int n, ret, alen, rlen;
1589 	int port_group_a, port_group_b, port_a, port_b;
1590 
1591 	alen = ((cmd[6] << 24) + (cmd[7] << 16) + (cmd[8] << 8)
1592 		+ cmd[9]);
1593 
1594 	arr = kzalloc(SDEBUG_MAX_TGTPGS_ARR_SZ, GFP_ATOMIC);
1595 	if (! arr)
1596 		return DID_REQUEUE << 16;
1597 	/*
1598 	 * EVPD page 0x88 states we have two ports, one
1599 	 * real and a fake port with no device connected.
1600 	 * So we create two port groups with one port each
1601 	 * and set the group with port B to unavailable.
1602 	 */
1603 	port_a = 0x1; /* relative port A */
1604 	port_b = 0x2; /* relative port B */
1605 	port_group_a = (((host_no + 1) & 0x7f) << 8) +
1606 	    (devip->channel & 0x7f);
1607 	port_group_b = (((host_no + 1) & 0x7f) << 8) +
1608 	    (devip->channel & 0x7f) + 0x80;
1609 
1610 	/*
1611 	 * The asymmetric access state is cycled according to the host_id.
1612 	 */
1613 	n = 4;
1614 	if (0 == scsi_debug_vpd_use_hostno) {
1615 	    arr[n++] = host_no % 3; /* Asymm access state */
1616 	    arr[n++] = 0x0F; /* claim: all states are supported */
1617 	} else {
1618 	    arr[n++] = 0x0; /* Active/Optimized path */
1619 	    arr[n++] = 0x01; /* claim: only support active/optimized paths */
1620 	}
1621 	arr[n++] = (port_group_a >> 8) & 0xff;
1622 	arr[n++] = port_group_a & 0xff;
1623 	arr[n++] = 0;    /* Reserved */
1624 	arr[n++] = 0;    /* Status code */
1625 	arr[n++] = 0;    /* Vendor unique */
1626 	arr[n++] = 0x1;  /* One port per group */
1627 	arr[n++] = 0;    /* Reserved */
1628 	arr[n++] = 0;    /* Reserved */
1629 	arr[n++] = (port_a >> 8) & 0xff;
1630 	arr[n++] = port_a & 0xff;
1631 	arr[n++] = 3;    /* Port unavailable */
1632 	arr[n++] = 0x08; /* claim: only unavailalbe paths are supported */
1633 	arr[n++] = (port_group_b >> 8) & 0xff;
1634 	arr[n++] = port_group_b & 0xff;
1635 	arr[n++] = 0;    /* Reserved */
1636 	arr[n++] = 0;    /* Status code */
1637 	arr[n++] = 0;    /* Vendor unique */
1638 	arr[n++] = 0x1;  /* One port per group */
1639 	arr[n++] = 0;    /* Reserved */
1640 	arr[n++] = 0;    /* Reserved */
1641 	arr[n++] = (port_b >> 8) & 0xff;
1642 	arr[n++] = port_b & 0xff;
1643 
1644 	rlen = n - 4;
1645 	arr[0] = (rlen >> 24) & 0xff;
1646 	arr[1] = (rlen >> 16) & 0xff;
1647 	arr[2] = (rlen >> 8) & 0xff;
1648 	arr[3] = rlen & 0xff;
1649 
1650 	/*
1651 	 * Return the smallest value of either
1652 	 * - The allocated length
1653 	 * - The constructed command length
1654 	 * - The maximum array size
1655 	 */
1656 	rlen = min(alen,n);
1657 	ret = fill_from_dev_buffer(scp, arr,
1658 				   min(rlen, SDEBUG_MAX_TGTPGS_ARR_SZ));
1659 	kfree(arr);
1660 	return ret;
1661 }
1662 
1663 static int
1664 resp_rsup_opcodes(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
1665 {
1666 	bool rctd;
1667 	u8 reporting_opts, req_opcode, sdeb_i, supp;
1668 	u16 req_sa, u;
1669 	u32 alloc_len, a_len;
1670 	int k, offset, len, errsts, count, bump, na;
1671 	const struct opcode_info_t *oip;
1672 	const struct opcode_info_t *r_oip;
1673 	u8 *arr;
1674 	u8 *cmd = scp->cmnd;
1675 
1676 	rctd = !!(cmd[2] & 0x80);
1677 	reporting_opts = cmd[2] & 0x7;
1678 	req_opcode = cmd[3];
1679 	req_sa = get_unaligned_be16(cmd + 4);
1680 	alloc_len = get_unaligned_be32(cmd + 6);
1681 	if (alloc_len < 4 || alloc_len > 0xffff) {
1682 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 6, -1);
1683 		return check_condition_result;
1684 	}
1685 	if (alloc_len > 8192)
1686 		a_len = 8192;
1687 	else
1688 		a_len = alloc_len;
1689 	arr = kzalloc((a_len < 256) ? 320 : a_len + 64, GFP_ATOMIC);
1690 	if (NULL == arr) {
1691 		mk_sense_buffer(scp, ILLEGAL_REQUEST, INSUFF_RES_ASC,
1692 				INSUFF_RES_ASCQ);
1693 		return check_condition_result;
1694 	}
1695 	switch (reporting_opts) {
1696 	case 0:	/* all commands */
1697 		/* count number of commands */
1698 		for (count = 0, oip = opcode_info_arr;
1699 		     oip->num_attached != 0xff; ++oip) {
1700 			if (F_INV_OP & oip->flags)
1701 				continue;
1702 			count += (oip->num_attached + 1);
1703 		}
1704 		bump = rctd ? 20 : 8;
1705 		put_unaligned_be32(count * bump, arr);
1706 		for (offset = 4, oip = opcode_info_arr;
1707 		     oip->num_attached != 0xff && offset < a_len; ++oip) {
1708 			if (F_INV_OP & oip->flags)
1709 				continue;
1710 			na = oip->num_attached;
1711 			arr[offset] = oip->opcode;
1712 			put_unaligned_be16(oip->sa, arr + offset + 2);
1713 			if (rctd)
1714 				arr[offset + 5] |= 0x2;
1715 			if (FF_SA & oip->flags)
1716 				arr[offset + 5] |= 0x1;
1717 			put_unaligned_be16(oip->len_mask[0], arr + offset + 6);
1718 			if (rctd)
1719 				put_unaligned_be16(0xa, arr + offset + 8);
1720 			r_oip = oip;
1721 			for (k = 0, oip = oip->arrp; k < na; ++k, ++oip) {
1722 				if (F_INV_OP & oip->flags)
1723 					continue;
1724 				offset += bump;
1725 				arr[offset] = oip->opcode;
1726 				put_unaligned_be16(oip->sa, arr + offset + 2);
1727 				if (rctd)
1728 					arr[offset + 5] |= 0x2;
1729 				if (FF_SA & oip->flags)
1730 					arr[offset + 5] |= 0x1;
1731 				put_unaligned_be16(oip->len_mask[0],
1732 						   arr + offset + 6);
1733 				if (rctd)
1734 					put_unaligned_be16(0xa,
1735 							   arr + offset + 8);
1736 			}
1737 			oip = r_oip;
1738 			offset += bump;
1739 		}
1740 		break;
1741 	case 1:	/* one command: opcode only */
1742 	case 2:	/* one command: opcode plus service action */
1743 	case 3:	/* one command: if sa==0 then opcode only else opcode+sa */
1744 		sdeb_i = opcode_ind_arr[req_opcode];
1745 		oip = &opcode_info_arr[sdeb_i];
1746 		if (F_INV_OP & oip->flags) {
1747 			supp = 1;
1748 			offset = 4;
1749 		} else {
1750 			if (1 == reporting_opts) {
1751 				if (FF_SA & oip->flags) {
1752 					mk_sense_invalid_fld(scp, SDEB_IN_CDB,
1753 							     2, 2);
1754 					kfree(arr);
1755 					return check_condition_result;
1756 				}
1757 				req_sa = 0;
1758 			} else if (2 == reporting_opts &&
1759 				   0 == (FF_SA & oip->flags)) {
1760 				mk_sense_invalid_fld(scp, SDEB_IN_CDB, 4, -1);
1761 				kfree(arr);	/* point at requested sa */
1762 				return check_condition_result;
1763 			}
1764 			if (0 == (FF_SA & oip->flags) &&
1765 			    req_opcode == oip->opcode)
1766 				supp = 3;
1767 			else if (0 == (FF_SA & oip->flags)) {
1768 				na = oip->num_attached;
1769 				for (k = 0, oip = oip->arrp; k < na;
1770 				     ++k, ++oip) {
1771 					if (req_opcode == oip->opcode)
1772 						break;
1773 				}
1774 				supp = (k >= na) ? 1 : 3;
1775 			} else if (req_sa != oip->sa) {
1776 				na = oip->num_attached;
1777 				for (k = 0, oip = oip->arrp; k < na;
1778 				     ++k, ++oip) {
1779 					if (req_sa == oip->sa)
1780 						break;
1781 				}
1782 				supp = (k >= na) ? 1 : 3;
1783 			} else
1784 				supp = 3;
1785 			if (3 == supp) {
1786 				u = oip->len_mask[0];
1787 				put_unaligned_be16(u, arr + 2);
1788 				arr[4] = oip->opcode;
1789 				for (k = 1; k < u; ++k)
1790 					arr[4 + k] = (k < 16) ?
1791 						 oip->len_mask[k] : 0xff;
1792 				offset = 4 + u;
1793 			} else
1794 				offset = 4;
1795 		}
1796 		arr[1] = (rctd ? 0x80 : 0) | supp;
1797 		if (rctd) {
1798 			put_unaligned_be16(0xa, arr + offset);
1799 			offset += 12;
1800 		}
1801 		break;
1802 	default:
1803 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 2, 2);
1804 		kfree(arr);
1805 		return check_condition_result;
1806 	}
1807 	offset = (offset < a_len) ? offset : a_len;
1808 	len = (offset < alloc_len) ? offset : alloc_len;
1809 	errsts = fill_from_dev_buffer(scp, arr, len);
1810 	kfree(arr);
1811 	return errsts;
1812 }
1813 
1814 static int
1815 resp_rsup_tmfs(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
1816 {
1817 	bool repd;
1818 	u32 alloc_len, len;
1819 	u8 arr[16];
1820 	u8 *cmd = scp->cmnd;
1821 
1822 	memset(arr, 0, sizeof(arr));
1823 	repd = !!(cmd[2] & 0x80);
1824 	alloc_len = get_unaligned_be32(cmd + 6);
1825 	if (alloc_len < 4) {
1826 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 6, -1);
1827 		return check_condition_result;
1828 	}
1829 	arr[0] = 0xc8;		/* ATS | ATSS | LURS */
1830 	arr[1] = 0x1;		/* ITNRS */
1831 	if (repd) {
1832 		arr[3] = 0xc;
1833 		len = 16;
1834 	} else
1835 		len = 4;
1836 
1837 	len = (len < alloc_len) ? len : alloc_len;
1838 	return fill_from_dev_buffer(scp, arr, len);
1839 }
1840 
1841 /* <<Following mode page info copied from ST318451LW>> */
1842 
1843 static int resp_err_recov_pg(unsigned char * p, int pcontrol, int target)
1844 {	/* Read-Write Error Recovery page for mode_sense */
1845 	unsigned char err_recov_pg[] = {0x1, 0xa, 0xc0, 11, 240, 0, 0, 0,
1846 					5, 0, 0xff, 0xff};
1847 
1848 	memcpy(p, err_recov_pg, sizeof(err_recov_pg));
1849 	if (1 == pcontrol)
1850 		memset(p + 2, 0, sizeof(err_recov_pg) - 2);
1851 	return sizeof(err_recov_pg);
1852 }
1853 
1854 static int resp_disconnect_pg(unsigned char * p, int pcontrol, int target)
1855 { 	/* Disconnect-Reconnect page for mode_sense */
1856 	unsigned char disconnect_pg[] = {0x2, 0xe, 128, 128, 0, 10, 0, 0,
1857 					 0, 0, 0, 0, 0, 0, 0, 0};
1858 
1859 	memcpy(p, disconnect_pg, sizeof(disconnect_pg));
1860 	if (1 == pcontrol)
1861 		memset(p + 2, 0, sizeof(disconnect_pg) - 2);
1862 	return sizeof(disconnect_pg);
1863 }
1864 
1865 static int resp_format_pg(unsigned char * p, int pcontrol, int target)
1866 {       /* Format device page for mode_sense */
1867 	unsigned char format_pg[] = {0x3, 0x16, 0, 0, 0, 0, 0, 0,
1868 				     0, 0, 0, 0, 0, 0, 0, 0,
1869 				     0, 0, 0, 0, 0x40, 0, 0, 0};
1870 
1871 	memcpy(p, format_pg, sizeof(format_pg));
1872 	p[10] = (sdebug_sectors_per >> 8) & 0xff;
1873 	p[11] = sdebug_sectors_per & 0xff;
1874 	p[12] = (scsi_debug_sector_size >> 8) & 0xff;
1875 	p[13] = scsi_debug_sector_size & 0xff;
1876 	if (scsi_debug_removable)
1877 		p[20] |= 0x20; /* should agree with INQUIRY */
1878 	if (1 == pcontrol)
1879 		memset(p + 2, 0, sizeof(format_pg) - 2);
1880 	return sizeof(format_pg);
1881 }
1882 
1883 static int resp_caching_pg(unsigned char * p, int pcontrol, int target)
1884 { 	/* Caching page for mode_sense */
1885 	unsigned char ch_caching_pg[] = {/* 0x8, 18, */ 0x4, 0, 0, 0, 0, 0,
1886 		0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0};
1887 	unsigned char d_caching_pg[] = {0x8, 18, 0x14, 0, 0xff, 0xff, 0, 0,
1888 		0xff, 0xff, 0xff, 0xff, 0x80, 0x14, 0, 0,     0, 0, 0, 0};
1889 
1890 	if (SCSI_DEBUG_OPT_N_WCE & scsi_debug_opts)
1891 		caching_pg[2] &= ~0x4;	/* set WCE=0 (default WCE=1) */
1892 	memcpy(p, caching_pg, sizeof(caching_pg));
1893 	if (1 == pcontrol)
1894 		memcpy(p + 2, ch_caching_pg, sizeof(ch_caching_pg));
1895 	else if (2 == pcontrol)
1896 		memcpy(p, d_caching_pg, sizeof(d_caching_pg));
1897 	return sizeof(caching_pg);
1898 }
1899 
1900 static int resp_ctrl_m_pg(unsigned char * p, int pcontrol, int target)
1901 { 	/* Control mode page for mode_sense */
1902 	unsigned char ch_ctrl_m_pg[] = {/* 0xa, 10, */ 0x6, 0, 0, 0, 0, 0,
1903 				        0, 0, 0, 0};
1904 	unsigned char d_ctrl_m_pg[] = {0xa, 10, 2, 0, 0, 0, 0, 0,
1905 				     0, 0, 0x2, 0x4b};
1906 
1907 	if (scsi_debug_dsense)
1908 		ctrl_m_pg[2] |= 0x4;
1909 	else
1910 		ctrl_m_pg[2] &= ~0x4;
1911 
1912 	if (scsi_debug_ato)
1913 		ctrl_m_pg[5] |= 0x80; /* ATO=1 */
1914 
1915 	memcpy(p, ctrl_m_pg, sizeof(ctrl_m_pg));
1916 	if (1 == pcontrol)
1917 		memcpy(p + 2, ch_ctrl_m_pg, sizeof(ch_ctrl_m_pg));
1918 	else if (2 == pcontrol)
1919 		memcpy(p, d_ctrl_m_pg, sizeof(d_ctrl_m_pg));
1920 	return sizeof(ctrl_m_pg);
1921 }
1922 
1923 
1924 static int resp_iec_m_pg(unsigned char * p, int pcontrol, int target)
1925 {	/* Informational Exceptions control mode page for mode_sense */
1926 	unsigned char ch_iec_m_pg[] = {/* 0x1c, 0xa, */ 0x4, 0xf, 0, 0, 0, 0,
1927 				       0, 0, 0x0, 0x0};
1928 	unsigned char d_iec_m_pg[] = {0x1c, 0xa, 0x08, 0, 0, 0, 0, 0,
1929 				      0, 0, 0x0, 0x0};
1930 
1931 	memcpy(p, iec_m_pg, sizeof(iec_m_pg));
1932 	if (1 == pcontrol)
1933 		memcpy(p + 2, ch_iec_m_pg, sizeof(ch_iec_m_pg));
1934 	else if (2 == pcontrol)
1935 		memcpy(p, d_iec_m_pg, sizeof(d_iec_m_pg));
1936 	return sizeof(iec_m_pg);
1937 }
1938 
1939 static int resp_sas_sf_m_pg(unsigned char * p, int pcontrol, int target)
1940 {	/* SAS SSP mode page - short format for mode_sense */
1941 	unsigned char sas_sf_m_pg[] = {0x19, 0x6,
1942 		0x6, 0x0, 0x7, 0xd0, 0x0, 0x0};
1943 
1944 	memcpy(p, sas_sf_m_pg, sizeof(sas_sf_m_pg));
1945 	if (1 == pcontrol)
1946 		memset(p + 2, 0, sizeof(sas_sf_m_pg) - 2);
1947 	return sizeof(sas_sf_m_pg);
1948 }
1949 
1950 
1951 static int resp_sas_pcd_m_spg(unsigned char * p, int pcontrol, int target,
1952 			      int target_dev_id)
1953 {	/* SAS phy control and discover mode page for mode_sense */
1954 	unsigned char sas_pcd_m_pg[] = {0x59, 0x1, 0, 0x64, 0, 0x6, 0, 2,
1955 		    0, 0, 0, 0, 0x10, 0x9, 0x8, 0x0,
1956 		    0x52, 0x22, 0x22, 0x20, 0x0, 0x0, 0x0, 0x0,
1957 		    0x51, 0x11, 0x11, 0x10, 0x0, 0x0, 0x0, 0x1,
1958 		    0x2, 0, 0, 0, 0, 0, 0, 0,
1959 		    0x88, 0x99, 0, 0, 0, 0, 0, 0,
1960 		    0, 0, 0, 0, 0, 0, 0, 0,
1961 		    0, 1, 0, 0, 0x10, 0x9, 0x8, 0x0,
1962 		    0x52, 0x22, 0x22, 0x20, 0x0, 0x0, 0x0, 0x0,
1963 		    0x51, 0x11, 0x11, 0x10, 0x0, 0x0, 0x0, 0x1,
1964 		    0x3, 0, 0, 0, 0, 0, 0, 0,
1965 		    0x88, 0x99, 0, 0, 0, 0, 0, 0,
1966 		    0, 0, 0, 0, 0, 0, 0, 0,
1967 		};
1968 	int port_a, port_b;
1969 
1970 	port_a = target_dev_id + 1;
1971 	port_b = port_a + 1;
1972 	memcpy(p, sas_pcd_m_pg, sizeof(sas_pcd_m_pg));
1973 	p[20] = (port_a >> 24);
1974 	p[21] = (port_a >> 16) & 0xff;
1975 	p[22] = (port_a >> 8) & 0xff;
1976 	p[23] = port_a & 0xff;
1977 	p[48 + 20] = (port_b >> 24);
1978 	p[48 + 21] = (port_b >> 16) & 0xff;
1979 	p[48 + 22] = (port_b >> 8) & 0xff;
1980 	p[48 + 23] = port_b & 0xff;
1981 	if (1 == pcontrol)
1982 		memset(p + 4, 0, sizeof(sas_pcd_m_pg) - 4);
1983 	return sizeof(sas_pcd_m_pg);
1984 }
1985 
1986 static int resp_sas_sha_m_spg(unsigned char * p, int pcontrol)
1987 {	/* SAS SSP shared protocol specific port mode subpage */
1988 	unsigned char sas_sha_m_pg[] = {0x59, 0x2, 0, 0xc, 0, 0x6, 0x10, 0,
1989 		    0, 0, 0, 0, 0, 0, 0, 0,
1990 		};
1991 
1992 	memcpy(p, sas_sha_m_pg, sizeof(sas_sha_m_pg));
1993 	if (1 == pcontrol)
1994 		memset(p + 4, 0, sizeof(sas_sha_m_pg) - 4);
1995 	return sizeof(sas_sha_m_pg);
1996 }
1997 
1998 #define SDEBUG_MAX_MSENSE_SZ 256
1999 
2000 static int
2001 resp_mode_sense(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
2002 {
2003 	unsigned char dbd, llbaa;
2004 	int pcontrol, pcode, subpcode, bd_len;
2005 	unsigned char dev_spec;
2006 	int k, alloc_len, msense_6, offset, len, target_dev_id;
2007 	int target = scp->device->id;
2008 	unsigned char * ap;
2009 	unsigned char arr[SDEBUG_MAX_MSENSE_SZ];
2010 	unsigned char *cmd = scp->cmnd;
2011 
2012 	dbd = !!(cmd[1] & 0x8);
2013 	pcontrol = (cmd[2] & 0xc0) >> 6;
2014 	pcode = cmd[2] & 0x3f;
2015 	subpcode = cmd[3];
2016 	msense_6 = (MODE_SENSE == cmd[0]);
2017 	llbaa = msense_6 ? 0 : !!(cmd[1] & 0x10);
2018 	if ((0 == scsi_debug_ptype) && (0 == dbd))
2019 		bd_len = llbaa ? 16 : 8;
2020 	else
2021 		bd_len = 0;
2022 	alloc_len = msense_6 ? cmd[4] : ((cmd[7] << 8) | cmd[8]);
2023 	memset(arr, 0, SDEBUG_MAX_MSENSE_SZ);
2024 	if (0x3 == pcontrol) {  /* Saving values not supported */
2025 		mk_sense_buffer(scp, ILLEGAL_REQUEST, SAVING_PARAMS_UNSUP, 0);
2026 		return check_condition_result;
2027 	}
2028 	target_dev_id = ((devip->sdbg_host->shost->host_no + 1) * 2000) +
2029 			(devip->target * 1000) - 3;
2030 	/* set DPOFUA bit for disks */
2031 	if (0 == scsi_debug_ptype)
2032 		dev_spec = (DEV_READONLY(target) ? 0x80 : 0x0) | 0x10;
2033 	else
2034 		dev_spec = 0x0;
2035 	if (msense_6) {
2036 		arr[2] = dev_spec;
2037 		arr[3] = bd_len;
2038 		offset = 4;
2039 	} else {
2040 		arr[3] = dev_spec;
2041 		if (16 == bd_len)
2042 			arr[4] = 0x1;	/* set LONGLBA bit */
2043 		arr[7] = bd_len;	/* assume 255 or less */
2044 		offset = 8;
2045 	}
2046 	ap = arr + offset;
2047 	if ((bd_len > 0) && (!sdebug_capacity))
2048 		sdebug_capacity = get_sdebug_capacity();
2049 
2050 	if (8 == bd_len) {
2051 		if (sdebug_capacity > 0xfffffffe) {
2052 			ap[0] = 0xff;
2053 			ap[1] = 0xff;
2054 			ap[2] = 0xff;
2055 			ap[3] = 0xff;
2056 		} else {
2057 			ap[0] = (sdebug_capacity >> 24) & 0xff;
2058 			ap[1] = (sdebug_capacity >> 16) & 0xff;
2059 			ap[2] = (sdebug_capacity >> 8) & 0xff;
2060 			ap[3] = sdebug_capacity & 0xff;
2061 		}
2062 		ap[6] = (scsi_debug_sector_size >> 8) & 0xff;
2063 		ap[7] = scsi_debug_sector_size & 0xff;
2064 		offset += bd_len;
2065 		ap = arr + offset;
2066 	} else if (16 == bd_len) {
2067 		unsigned long long capac = sdebug_capacity;
2068 
2069         	for (k = 0; k < 8; ++k, capac >>= 8)
2070                 	ap[7 - k] = capac & 0xff;
2071 		ap[12] = (scsi_debug_sector_size >> 24) & 0xff;
2072 		ap[13] = (scsi_debug_sector_size >> 16) & 0xff;
2073 		ap[14] = (scsi_debug_sector_size >> 8) & 0xff;
2074 		ap[15] = scsi_debug_sector_size & 0xff;
2075 		offset += bd_len;
2076 		ap = arr + offset;
2077 	}
2078 
2079 	if ((subpcode > 0x0) && (subpcode < 0xff) && (0x19 != pcode)) {
2080 		/* TODO: Control Extension page */
2081 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 3, -1);
2082 		return check_condition_result;
2083 	}
2084 	switch (pcode) {
2085 	case 0x1:	/* Read-Write error recovery page, direct access */
2086 		len = resp_err_recov_pg(ap, pcontrol, target);
2087 		offset += len;
2088 		break;
2089 	case 0x2:	/* Disconnect-Reconnect page, all devices */
2090 		len = resp_disconnect_pg(ap, pcontrol, target);
2091 		offset += len;
2092 		break;
2093         case 0x3:       /* Format device page, direct access */
2094                 len = resp_format_pg(ap, pcontrol, target);
2095                 offset += len;
2096                 break;
2097 	case 0x8:	/* Caching page, direct access */
2098 		len = resp_caching_pg(ap, pcontrol, target);
2099 		offset += len;
2100 		break;
2101 	case 0xa:	/* Control Mode page, all devices */
2102 		len = resp_ctrl_m_pg(ap, pcontrol, target);
2103 		offset += len;
2104 		break;
2105 	case 0x19:	/* if spc==1 then sas phy, control+discover */
2106 		if ((subpcode > 0x2) && (subpcode < 0xff)) {
2107 			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 3, -1);
2108 			return check_condition_result;
2109 	        }
2110 		len = 0;
2111 		if ((0x0 == subpcode) || (0xff == subpcode))
2112 			len += resp_sas_sf_m_pg(ap + len, pcontrol, target);
2113 		if ((0x1 == subpcode) || (0xff == subpcode))
2114 			len += resp_sas_pcd_m_spg(ap + len, pcontrol, target,
2115 						  target_dev_id);
2116 		if ((0x2 == subpcode) || (0xff == subpcode))
2117 			len += resp_sas_sha_m_spg(ap + len, pcontrol);
2118 		offset += len;
2119 		break;
2120 	case 0x1c:	/* Informational Exceptions Mode page, all devices */
2121 		len = resp_iec_m_pg(ap, pcontrol, target);
2122 		offset += len;
2123 		break;
2124 	case 0x3f:	/* Read all Mode pages */
2125 		if ((0 == subpcode) || (0xff == subpcode)) {
2126 			len = resp_err_recov_pg(ap, pcontrol, target);
2127 			len += resp_disconnect_pg(ap + len, pcontrol, target);
2128 			len += resp_format_pg(ap + len, pcontrol, target);
2129 			len += resp_caching_pg(ap + len, pcontrol, target);
2130 			len += resp_ctrl_m_pg(ap + len, pcontrol, target);
2131 			len += resp_sas_sf_m_pg(ap + len, pcontrol, target);
2132 			if (0xff == subpcode) {
2133 				len += resp_sas_pcd_m_spg(ap + len, pcontrol,
2134 						  target, target_dev_id);
2135 				len += resp_sas_sha_m_spg(ap + len, pcontrol);
2136 			}
2137 			len += resp_iec_m_pg(ap + len, pcontrol, target);
2138 		} else {
2139 			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 3, -1);
2140 			return check_condition_result;
2141                 }
2142 		offset += len;
2143 		break;
2144 	default:
2145 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 2, 5);
2146 		return check_condition_result;
2147 	}
2148 	if (msense_6)
2149 		arr[0] = offset - 1;
2150 	else {
2151 		arr[0] = ((offset - 2) >> 8) & 0xff;
2152 		arr[1] = (offset - 2) & 0xff;
2153 	}
2154 	return fill_from_dev_buffer(scp, arr, min(alloc_len, offset));
2155 }
2156 
2157 #define SDEBUG_MAX_MSELECT_SZ 512
2158 
2159 static int
2160 resp_mode_select(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
2161 {
2162 	int pf, sp, ps, md_len, bd_len, off, spf, pg_len;
2163 	int param_len, res, mpage;
2164 	unsigned char arr[SDEBUG_MAX_MSELECT_SZ];
2165 	unsigned char *cmd = scp->cmnd;
2166 	int mselect6 = (MODE_SELECT == cmd[0]);
2167 
2168 	memset(arr, 0, sizeof(arr));
2169 	pf = cmd[1] & 0x10;
2170 	sp = cmd[1] & 0x1;
2171 	param_len = mselect6 ? cmd[4] : ((cmd[7] << 8) + cmd[8]);
2172 	if ((0 == pf) || sp || (param_len > SDEBUG_MAX_MSELECT_SZ)) {
2173 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, mselect6 ? 4 : 7, -1);
2174 		return check_condition_result;
2175 	}
2176         res = fetch_to_dev_buffer(scp, arr, param_len);
2177         if (-1 == res)
2178                 return (DID_ERROR << 16);
2179         else if ((res < param_len) &&
2180                  (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts))
2181 		sdev_printk(KERN_INFO, scp->device,
2182 			    "%s: cdb indicated=%d, IO sent=%d bytes\n",
2183 			    __func__, param_len, res);
2184 	md_len = mselect6 ? (arr[0] + 1) : ((arr[0] << 8) + arr[1] + 2);
2185 	bd_len = mselect6 ? arr[3] : ((arr[6] << 8) + arr[7]);
2186 	if (md_len > 2) {
2187 		mk_sense_invalid_fld(scp, SDEB_IN_DATA, 0, -1);
2188 		return check_condition_result;
2189 	}
2190 	off = bd_len + (mselect6 ? 4 : 8);
2191 	mpage = arr[off] & 0x3f;
2192 	ps = !!(arr[off] & 0x80);
2193 	if (ps) {
2194 		mk_sense_invalid_fld(scp, SDEB_IN_DATA, off, 7);
2195 		return check_condition_result;
2196 	}
2197 	spf = !!(arr[off] & 0x40);
2198 	pg_len = spf ? ((arr[off + 2] << 8) + arr[off + 3] + 4) :
2199 		       (arr[off + 1] + 2);
2200 	if ((pg_len + off) > param_len) {
2201 		mk_sense_buffer(scp, ILLEGAL_REQUEST,
2202 				PARAMETER_LIST_LENGTH_ERR, 0);
2203 		return check_condition_result;
2204 	}
2205 	switch (mpage) {
2206 	case 0x8:      /* Caching Mode page */
2207 		if (caching_pg[1] == arr[off + 1]) {
2208 			memcpy(caching_pg + 2, arr + off + 2,
2209 			       sizeof(caching_pg) - 2);
2210 			goto set_mode_changed_ua;
2211 		}
2212 		break;
2213 	case 0xa:      /* Control Mode page */
2214 		if (ctrl_m_pg[1] == arr[off + 1]) {
2215 			memcpy(ctrl_m_pg + 2, arr + off + 2,
2216 			       sizeof(ctrl_m_pg) - 2);
2217 			scsi_debug_dsense = !!(ctrl_m_pg[2] & 0x4);
2218 			goto set_mode_changed_ua;
2219 		}
2220 		break;
2221 	case 0x1c:      /* Informational Exceptions Mode page */
2222 		if (iec_m_pg[1] == arr[off + 1]) {
2223 			memcpy(iec_m_pg + 2, arr + off + 2,
2224 			       sizeof(iec_m_pg) - 2);
2225 			goto set_mode_changed_ua;
2226 		}
2227 		break;
2228 	default:
2229 		break;
2230 	}
2231 	mk_sense_invalid_fld(scp, SDEB_IN_DATA, off, 5);
2232 	return check_condition_result;
2233 set_mode_changed_ua:
2234 	set_bit(SDEBUG_UA_MODE_CHANGED, devip->uas_bm);
2235 	return 0;
2236 }
2237 
2238 static int resp_temp_l_pg(unsigned char * arr)
2239 {
2240 	unsigned char temp_l_pg[] = {0x0, 0x0, 0x3, 0x2, 0x0, 38,
2241 				     0x0, 0x1, 0x3, 0x2, 0x0, 65,
2242 		};
2243 
2244         memcpy(arr, temp_l_pg, sizeof(temp_l_pg));
2245         return sizeof(temp_l_pg);
2246 }
2247 
2248 static int resp_ie_l_pg(unsigned char * arr)
2249 {
2250 	unsigned char ie_l_pg[] = {0x0, 0x0, 0x3, 0x3, 0x0, 0x0, 38,
2251 		};
2252 
2253         memcpy(arr, ie_l_pg, sizeof(ie_l_pg));
2254 	if (iec_m_pg[2] & 0x4) {	/* TEST bit set */
2255 		arr[4] = THRESHOLD_EXCEEDED;
2256 		arr[5] = 0xff;
2257 	}
2258         return sizeof(ie_l_pg);
2259 }
2260 
2261 #define SDEBUG_MAX_LSENSE_SZ 512
2262 
2263 static int resp_log_sense(struct scsi_cmnd * scp,
2264                           struct sdebug_dev_info * devip)
2265 {
2266 	int ppc, sp, pcontrol, pcode, subpcode, alloc_len, len, n;
2267 	unsigned char arr[SDEBUG_MAX_LSENSE_SZ];
2268 	unsigned char *cmd = scp->cmnd;
2269 
2270 	memset(arr, 0, sizeof(arr));
2271 	ppc = cmd[1] & 0x2;
2272 	sp = cmd[1] & 0x1;
2273 	if (ppc || sp) {
2274 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 1, ppc ? 1 : 0);
2275 		return check_condition_result;
2276 	}
2277 	pcontrol = (cmd[2] & 0xc0) >> 6;
2278 	pcode = cmd[2] & 0x3f;
2279 	subpcode = cmd[3] & 0xff;
2280 	alloc_len = (cmd[7] << 8) + cmd[8];
2281 	arr[0] = pcode;
2282 	if (0 == subpcode) {
2283 		switch (pcode) {
2284 		case 0x0:	/* Supported log pages log page */
2285 			n = 4;
2286 			arr[n++] = 0x0;		/* this page */
2287 			arr[n++] = 0xd;		/* Temperature */
2288 			arr[n++] = 0x2f;	/* Informational exceptions */
2289 			arr[3] = n - 4;
2290 			break;
2291 		case 0xd:	/* Temperature log page */
2292 			arr[3] = resp_temp_l_pg(arr + 4);
2293 			break;
2294 		case 0x2f:	/* Informational exceptions log page */
2295 			arr[3] = resp_ie_l_pg(arr + 4);
2296 			break;
2297 		default:
2298 			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 2, 5);
2299 			return check_condition_result;
2300 		}
2301 	} else if (0xff == subpcode) {
2302 		arr[0] |= 0x40;
2303 		arr[1] = subpcode;
2304 		switch (pcode) {
2305 		case 0x0:	/* Supported log pages and subpages log page */
2306 			n = 4;
2307 			arr[n++] = 0x0;
2308 			arr[n++] = 0x0;		/* 0,0 page */
2309 			arr[n++] = 0x0;
2310 			arr[n++] = 0xff;	/* this page */
2311 			arr[n++] = 0xd;
2312 			arr[n++] = 0x0;		/* Temperature */
2313 			arr[n++] = 0x2f;
2314 			arr[n++] = 0x0;	/* Informational exceptions */
2315 			arr[3] = n - 4;
2316 			break;
2317 		case 0xd:	/* Temperature subpages */
2318 			n = 4;
2319 			arr[n++] = 0xd;
2320 			arr[n++] = 0x0;		/* Temperature */
2321 			arr[3] = n - 4;
2322 			break;
2323 		case 0x2f:	/* Informational exceptions subpages */
2324 			n = 4;
2325 			arr[n++] = 0x2f;
2326 			arr[n++] = 0x0;		/* Informational exceptions */
2327 			arr[3] = n - 4;
2328 			break;
2329 		default:
2330 			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 2, 5);
2331 			return check_condition_result;
2332 		}
2333 	} else {
2334 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 3, -1);
2335 		return check_condition_result;
2336 	}
2337 	len = min(((arr[2] << 8) + arr[3]) + 4, alloc_len);
2338 	return fill_from_dev_buffer(scp, arr,
2339 		    min(len, SDEBUG_MAX_INQ_ARR_SZ));
2340 }
2341 
2342 static int check_device_access_params(struct scsi_cmnd *scp,
2343 				      unsigned long long lba, unsigned int num)
2344 {
2345 	if (lba + num > sdebug_capacity) {
2346 		mk_sense_buffer(scp, ILLEGAL_REQUEST, LBA_OUT_OF_RANGE, 0);
2347 		return check_condition_result;
2348 	}
2349 	/* transfer length excessive (tie in to block limits VPD page) */
2350 	if (num > sdebug_store_sectors) {
2351 		/* needs work to find which cdb byte 'num' comes from */
2352 		mk_sense_buffer(scp, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB, 0);
2353 		return check_condition_result;
2354 	}
2355 	return 0;
2356 }
2357 
2358 /* Returns number of bytes copied or -1 if error. */
2359 static int
2360 do_device_access(struct scsi_cmnd *scmd, u64 lba, u32 num, bool do_write)
2361 {
2362 	int ret;
2363 	u64 block, rest = 0;
2364 	struct scsi_data_buffer *sdb;
2365 	enum dma_data_direction dir;
2366 
2367 	if (do_write) {
2368 		sdb = scsi_out(scmd);
2369 		dir = DMA_TO_DEVICE;
2370 	} else {
2371 		sdb = scsi_in(scmd);
2372 		dir = DMA_FROM_DEVICE;
2373 	}
2374 
2375 	if (!sdb->length)
2376 		return 0;
2377 	if (!(scsi_bidi_cmnd(scmd) || scmd->sc_data_direction == dir))
2378 		return -1;
2379 
2380 	block = do_div(lba, sdebug_store_sectors);
2381 	if (block + num > sdebug_store_sectors)
2382 		rest = block + num - sdebug_store_sectors;
2383 
2384 	ret = sg_copy_buffer(sdb->table.sgl, sdb->table.nents,
2385 		   fake_storep + (block * scsi_debug_sector_size),
2386 		   (num - rest) * scsi_debug_sector_size, 0, do_write);
2387 	if (ret != (num - rest) * scsi_debug_sector_size)
2388 		return ret;
2389 
2390 	if (rest) {
2391 		ret += sg_copy_buffer(sdb->table.sgl, sdb->table.nents,
2392 			    fake_storep, rest * scsi_debug_sector_size,
2393 			    (num - rest) * scsi_debug_sector_size, do_write);
2394 	}
2395 
2396 	return ret;
2397 }
2398 
2399 /* If fake_store(lba,num) compares equal to arr(num), then copy top half of
2400  * arr into fake_store(lba,num) and return true. If comparison fails then
2401  * return false. */
2402 static bool
2403 comp_write_worker(u64 lba, u32 num, const u8 *arr)
2404 {
2405 	bool res;
2406 	u64 block, rest = 0;
2407 	u32 store_blks = sdebug_store_sectors;
2408 	u32 lb_size = scsi_debug_sector_size;
2409 
2410 	block = do_div(lba, store_blks);
2411 	if (block + num > store_blks)
2412 		rest = block + num - store_blks;
2413 
2414 	res = !memcmp(fake_storep + (block * lb_size), arr,
2415 		      (num - rest) * lb_size);
2416 	if (!res)
2417 		return res;
2418 	if (rest)
2419 		res = memcmp(fake_storep, arr + ((num - rest) * lb_size),
2420 			     rest * lb_size);
2421 	if (!res)
2422 		return res;
2423 	arr += num * lb_size;
2424 	memcpy(fake_storep + (block * lb_size), arr, (num - rest) * lb_size);
2425 	if (rest)
2426 		memcpy(fake_storep, arr + ((num - rest) * lb_size),
2427 		       rest * lb_size);
2428 	return res;
2429 }
2430 
2431 static __be16 dif_compute_csum(const void *buf, int len)
2432 {
2433 	__be16 csum;
2434 
2435 	if (scsi_debug_guard)
2436 		csum = (__force __be16)ip_compute_csum(buf, len);
2437 	else
2438 		csum = cpu_to_be16(crc_t10dif(buf, len));
2439 
2440 	return csum;
2441 }
2442 
2443 static int dif_verify(struct sd_dif_tuple *sdt, const void *data,
2444 		      sector_t sector, u32 ei_lba)
2445 {
2446 	__be16 csum = dif_compute_csum(data, scsi_debug_sector_size);
2447 
2448 	if (sdt->guard_tag != csum) {
2449 		pr_err("%s: GUARD check failed on sector %lu rcvd 0x%04x, data 0x%04x\n",
2450 			__func__,
2451 			(unsigned long)sector,
2452 			be16_to_cpu(sdt->guard_tag),
2453 			be16_to_cpu(csum));
2454 		return 0x01;
2455 	}
2456 	if (scsi_debug_dif == SD_DIF_TYPE1_PROTECTION &&
2457 	    be32_to_cpu(sdt->ref_tag) != (sector & 0xffffffff)) {
2458 		pr_err("%s: REF check failed on sector %lu\n",
2459 			__func__, (unsigned long)sector);
2460 		return 0x03;
2461 	}
2462 	if (scsi_debug_dif == SD_DIF_TYPE2_PROTECTION &&
2463 	    be32_to_cpu(sdt->ref_tag) != ei_lba) {
2464 		pr_err("%s: REF check failed on sector %lu\n",
2465 			__func__, (unsigned long)sector);
2466 		return 0x03;
2467 	}
2468 	return 0;
2469 }
2470 
2471 static void dif_copy_prot(struct scsi_cmnd *SCpnt, sector_t sector,
2472 			  unsigned int sectors, bool read)
2473 {
2474 	size_t resid;
2475 	void *paddr;
2476 	const void *dif_store_end = dif_storep + sdebug_store_sectors;
2477 	struct sg_mapping_iter miter;
2478 
2479 	/* Bytes of protection data to copy into sgl */
2480 	resid = sectors * sizeof(*dif_storep);
2481 
2482 	sg_miter_start(&miter, scsi_prot_sglist(SCpnt),
2483 			scsi_prot_sg_count(SCpnt), SG_MITER_ATOMIC |
2484 			(read ? SG_MITER_TO_SG : SG_MITER_FROM_SG));
2485 
2486 	while (sg_miter_next(&miter) && resid > 0) {
2487 		size_t len = min(miter.length, resid);
2488 		void *start = dif_store(sector);
2489 		size_t rest = 0;
2490 
2491 		if (dif_store_end < start + len)
2492 			rest = start + len - dif_store_end;
2493 
2494 		paddr = miter.addr;
2495 
2496 		if (read)
2497 			memcpy(paddr, start, len - rest);
2498 		else
2499 			memcpy(start, paddr, len - rest);
2500 
2501 		if (rest) {
2502 			if (read)
2503 				memcpy(paddr + len - rest, dif_storep, rest);
2504 			else
2505 				memcpy(dif_storep, paddr + len - rest, rest);
2506 		}
2507 
2508 		sector += len / sizeof(*dif_storep);
2509 		resid -= len;
2510 	}
2511 	sg_miter_stop(&miter);
2512 }
2513 
2514 static int prot_verify_read(struct scsi_cmnd *SCpnt, sector_t start_sec,
2515 			    unsigned int sectors, u32 ei_lba)
2516 {
2517 	unsigned int i;
2518 	struct sd_dif_tuple *sdt;
2519 	sector_t sector;
2520 
2521 	for (i = 0; i < sectors; i++, ei_lba++) {
2522 		int ret;
2523 
2524 		sector = start_sec + i;
2525 		sdt = dif_store(sector);
2526 
2527 		if (sdt->app_tag == cpu_to_be16(0xffff))
2528 			continue;
2529 
2530 		ret = dif_verify(sdt, fake_store(sector), sector, ei_lba);
2531 		if (ret) {
2532 			dif_errors++;
2533 			return ret;
2534 		}
2535 	}
2536 
2537 	dif_copy_prot(SCpnt, start_sec, sectors, true);
2538 	dix_reads++;
2539 
2540 	return 0;
2541 }
2542 
2543 static int
2544 resp_read_dt0(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
2545 {
2546 	u8 *cmd = scp->cmnd;
2547 	u64 lba;
2548 	u32 num;
2549 	u32 ei_lba;
2550 	unsigned long iflags;
2551 	int ret;
2552 	bool check_prot;
2553 
2554 	switch (cmd[0]) {
2555 	case READ_16:
2556 		ei_lba = 0;
2557 		lba = get_unaligned_be64(cmd + 2);
2558 		num = get_unaligned_be32(cmd + 10);
2559 		check_prot = true;
2560 		break;
2561 	case READ_10:
2562 		ei_lba = 0;
2563 		lba = get_unaligned_be32(cmd + 2);
2564 		num = get_unaligned_be16(cmd + 7);
2565 		check_prot = true;
2566 		break;
2567 	case READ_6:
2568 		ei_lba = 0;
2569 		lba = (u32)cmd[3] | (u32)cmd[2] << 8 |
2570 		      (u32)(cmd[1] & 0x1f) << 16;
2571 		num = (0 == cmd[4]) ? 256 : cmd[4];
2572 		check_prot = true;
2573 		break;
2574 	case READ_12:
2575 		ei_lba = 0;
2576 		lba = get_unaligned_be32(cmd + 2);
2577 		num = get_unaligned_be32(cmd + 6);
2578 		check_prot = true;
2579 		break;
2580 	case XDWRITEREAD_10:
2581 		ei_lba = 0;
2582 		lba = get_unaligned_be32(cmd + 2);
2583 		num = get_unaligned_be16(cmd + 7);
2584 		check_prot = false;
2585 		break;
2586 	default:	/* assume READ(32) */
2587 		lba = get_unaligned_be64(cmd + 12);
2588 		ei_lba = get_unaligned_be32(cmd + 20);
2589 		num = get_unaligned_be32(cmd + 28);
2590 		check_prot = false;
2591 		break;
2592 	}
2593 	if (check_prot) {
2594 		if (scsi_debug_dif == SD_DIF_TYPE2_PROTECTION &&
2595 		    (cmd[1] & 0xe0)) {
2596 			mk_sense_invalid_opcode(scp);
2597 			return check_condition_result;
2598 		}
2599 		if ((scsi_debug_dif == SD_DIF_TYPE1_PROTECTION ||
2600 		     scsi_debug_dif == SD_DIF_TYPE3_PROTECTION) &&
2601 		    (cmd[1] & 0xe0) == 0)
2602 			sdev_printk(KERN_ERR, scp->device, "Unprotected RD "
2603 				    "to DIF device\n");
2604 	}
2605 	if (sdebug_any_injecting_opt) {
2606 		struct sdebug_scmd_extra_t *ep = scsi_cmd_priv(scp);
2607 
2608 		if (ep->inj_short)
2609 			num /= 2;
2610 	}
2611 
2612 	/* inline check_device_access_params() */
2613 	if (lba + num > sdebug_capacity) {
2614 		mk_sense_buffer(scp, ILLEGAL_REQUEST, LBA_OUT_OF_RANGE, 0);
2615 		return check_condition_result;
2616 	}
2617 	/* transfer length excessive (tie in to block limits VPD page) */
2618 	if (num > sdebug_store_sectors) {
2619 		/* needs work to find which cdb byte 'num' comes from */
2620 		mk_sense_buffer(scp, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB, 0);
2621 		return check_condition_result;
2622 	}
2623 
2624 	if ((SCSI_DEBUG_OPT_MEDIUM_ERR & scsi_debug_opts) &&
2625 	    (lba <= (OPT_MEDIUM_ERR_ADDR + OPT_MEDIUM_ERR_NUM - 1)) &&
2626 	    ((lba + num) > OPT_MEDIUM_ERR_ADDR)) {
2627 		/* claim unrecoverable read error */
2628 		mk_sense_buffer(scp, MEDIUM_ERROR, UNRECOVERED_READ_ERR, 0);
2629 		/* set info field and valid bit for fixed descriptor */
2630 		if (0x70 == (scp->sense_buffer[0] & 0x7f)) {
2631 			scp->sense_buffer[0] |= 0x80;	/* Valid bit */
2632 			ret = (lba < OPT_MEDIUM_ERR_ADDR)
2633 			      ? OPT_MEDIUM_ERR_ADDR : (int)lba;
2634 			put_unaligned_be32(ret, scp->sense_buffer + 3);
2635 		}
2636 		scsi_set_resid(scp, scsi_bufflen(scp));
2637 		return check_condition_result;
2638 	}
2639 
2640 	read_lock_irqsave(&atomic_rw, iflags);
2641 
2642 	/* DIX + T10 DIF */
2643 	if (scsi_debug_dix && scsi_prot_sg_count(scp)) {
2644 		int prot_ret = prot_verify_read(scp, lba, num, ei_lba);
2645 
2646 		if (prot_ret) {
2647 			read_unlock_irqrestore(&atomic_rw, iflags);
2648 			mk_sense_buffer(scp, ABORTED_COMMAND, 0x10, prot_ret);
2649 			return illegal_condition_result;
2650 		}
2651 	}
2652 
2653 	ret = do_device_access(scp, lba, num, false);
2654 	read_unlock_irqrestore(&atomic_rw, iflags);
2655 	if (ret == -1)
2656 		return DID_ERROR << 16;
2657 
2658 	scsi_in(scp)->resid = scsi_bufflen(scp) - ret;
2659 
2660 	if (sdebug_any_injecting_opt) {
2661 		struct sdebug_scmd_extra_t *ep = scsi_cmd_priv(scp);
2662 
2663 		if (ep->inj_recovered) {
2664 			mk_sense_buffer(scp, RECOVERED_ERROR,
2665 					THRESHOLD_EXCEEDED, 0);
2666 			return check_condition_result;
2667 		} else if (ep->inj_transport) {
2668 			mk_sense_buffer(scp, ABORTED_COMMAND,
2669 					TRANSPORT_PROBLEM, ACK_NAK_TO);
2670 			return check_condition_result;
2671 		} else if (ep->inj_dif) {
2672 			/* Logical block guard check failed */
2673 			mk_sense_buffer(scp, ABORTED_COMMAND, 0x10, 1);
2674 			return illegal_condition_result;
2675 		} else if (ep->inj_dix) {
2676 			mk_sense_buffer(scp, ILLEGAL_REQUEST, 0x10, 1);
2677 			return illegal_condition_result;
2678 		}
2679 	}
2680 	return 0;
2681 }
2682 
2683 void dump_sector(unsigned char *buf, int len)
2684 {
2685 	int i, j, n;
2686 
2687 	pr_err(">>> Sector Dump <<<\n");
2688 	for (i = 0 ; i < len ; i += 16) {
2689 		char b[128];
2690 
2691 		for (j = 0, n = 0; j < 16; j++) {
2692 			unsigned char c = buf[i+j];
2693 
2694 			if (c >= 0x20 && c < 0x7e)
2695 				n += scnprintf(b + n, sizeof(b) - n,
2696 					       " %c ", buf[i+j]);
2697 			else
2698 				n += scnprintf(b + n, sizeof(b) - n,
2699 					       "%02x ", buf[i+j]);
2700 		}
2701 		pr_err("%04d: %s\n", i, b);
2702 	}
2703 }
2704 
2705 static int prot_verify_write(struct scsi_cmnd *SCpnt, sector_t start_sec,
2706 			     unsigned int sectors, u32 ei_lba)
2707 {
2708 	int ret;
2709 	struct sd_dif_tuple *sdt;
2710 	void *daddr;
2711 	sector_t sector = start_sec;
2712 	int ppage_offset;
2713 	int dpage_offset;
2714 	struct sg_mapping_iter diter;
2715 	struct sg_mapping_iter piter;
2716 
2717 	BUG_ON(scsi_sg_count(SCpnt) == 0);
2718 	BUG_ON(scsi_prot_sg_count(SCpnt) == 0);
2719 
2720 	sg_miter_start(&piter, scsi_prot_sglist(SCpnt),
2721 			scsi_prot_sg_count(SCpnt),
2722 			SG_MITER_ATOMIC | SG_MITER_FROM_SG);
2723 	sg_miter_start(&diter, scsi_sglist(SCpnt), scsi_sg_count(SCpnt),
2724 			SG_MITER_ATOMIC | SG_MITER_FROM_SG);
2725 
2726 	/* For each protection page */
2727 	while (sg_miter_next(&piter)) {
2728 		dpage_offset = 0;
2729 		if (WARN_ON(!sg_miter_next(&diter))) {
2730 			ret = 0x01;
2731 			goto out;
2732 		}
2733 
2734 		for (ppage_offset = 0; ppage_offset < piter.length;
2735 		     ppage_offset += sizeof(struct sd_dif_tuple)) {
2736 			/* If we're at the end of the current
2737 			 * data page advance to the next one
2738 			 */
2739 			if (dpage_offset >= diter.length) {
2740 				if (WARN_ON(!sg_miter_next(&diter))) {
2741 					ret = 0x01;
2742 					goto out;
2743 				}
2744 				dpage_offset = 0;
2745 			}
2746 
2747 			sdt = piter.addr + ppage_offset;
2748 			daddr = diter.addr + dpage_offset;
2749 
2750 			ret = dif_verify(sdt, daddr, sector, ei_lba);
2751 			if (ret) {
2752 				dump_sector(daddr, scsi_debug_sector_size);
2753 				goto out;
2754 			}
2755 
2756 			sector++;
2757 			ei_lba++;
2758 			dpage_offset += scsi_debug_sector_size;
2759 		}
2760 		diter.consumed = dpage_offset;
2761 		sg_miter_stop(&diter);
2762 	}
2763 	sg_miter_stop(&piter);
2764 
2765 	dif_copy_prot(SCpnt, start_sec, sectors, false);
2766 	dix_writes++;
2767 
2768 	return 0;
2769 
2770 out:
2771 	dif_errors++;
2772 	sg_miter_stop(&diter);
2773 	sg_miter_stop(&piter);
2774 	return ret;
2775 }
2776 
2777 static unsigned long lba_to_map_index(sector_t lba)
2778 {
2779 	if (scsi_debug_unmap_alignment) {
2780 		lba += scsi_debug_unmap_granularity -
2781 			scsi_debug_unmap_alignment;
2782 	}
2783 	do_div(lba, scsi_debug_unmap_granularity);
2784 
2785 	return lba;
2786 }
2787 
2788 static sector_t map_index_to_lba(unsigned long index)
2789 {
2790 	sector_t lba = index * scsi_debug_unmap_granularity;
2791 
2792 	if (scsi_debug_unmap_alignment) {
2793 		lba -= scsi_debug_unmap_granularity -
2794 			scsi_debug_unmap_alignment;
2795 	}
2796 
2797 	return lba;
2798 }
2799 
2800 static unsigned int map_state(sector_t lba, unsigned int *num)
2801 {
2802 	sector_t end;
2803 	unsigned int mapped;
2804 	unsigned long index;
2805 	unsigned long next;
2806 
2807 	index = lba_to_map_index(lba);
2808 	mapped = test_bit(index, map_storep);
2809 
2810 	if (mapped)
2811 		next = find_next_zero_bit(map_storep, map_size, index);
2812 	else
2813 		next = find_next_bit(map_storep, map_size, index);
2814 
2815 	end = min_t(sector_t, sdebug_store_sectors,  map_index_to_lba(next));
2816 	*num = end - lba;
2817 
2818 	return mapped;
2819 }
2820 
2821 static void map_region(sector_t lba, unsigned int len)
2822 {
2823 	sector_t end = lba + len;
2824 
2825 	while (lba < end) {
2826 		unsigned long index = lba_to_map_index(lba);
2827 
2828 		if (index < map_size)
2829 			set_bit(index, map_storep);
2830 
2831 		lba = map_index_to_lba(index + 1);
2832 	}
2833 }
2834 
2835 static void unmap_region(sector_t lba, unsigned int len)
2836 {
2837 	sector_t end = lba + len;
2838 
2839 	while (lba < end) {
2840 		unsigned long index = lba_to_map_index(lba);
2841 
2842 		if (lba == map_index_to_lba(index) &&
2843 		    lba + scsi_debug_unmap_granularity <= end &&
2844 		    index < map_size) {
2845 			clear_bit(index, map_storep);
2846 			if (scsi_debug_lbprz) {
2847 				memset(fake_storep +
2848 				       lba * scsi_debug_sector_size, 0,
2849 				       scsi_debug_sector_size *
2850 				       scsi_debug_unmap_granularity);
2851 			}
2852 			if (dif_storep) {
2853 				memset(dif_storep + lba, 0xff,
2854 				       sizeof(*dif_storep) *
2855 				       scsi_debug_unmap_granularity);
2856 			}
2857 		}
2858 		lba = map_index_to_lba(index + 1);
2859 	}
2860 }
2861 
2862 static int
2863 resp_write_dt0(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
2864 {
2865 	u8 *cmd = scp->cmnd;
2866 	u64 lba;
2867 	u32 num;
2868 	u32 ei_lba;
2869 	unsigned long iflags;
2870 	int ret;
2871 	bool check_prot;
2872 
2873 	switch (cmd[0]) {
2874 	case WRITE_16:
2875 		ei_lba = 0;
2876 		lba = get_unaligned_be64(cmd + 2);
2877 		num = get_unaligned_be32(cmd + 10);
2878 		check_prot = true;
2879 		break;
2880 	case WRITE_10:
2881 		ei_lba = 0;
2882 		lba = get_unaligned_be32(cmd + 2);
2883 		num = get_unaligned_be16(cmd + 7);
2884 		check_prot = true;
2885 		break;
2886 	case WRITE_6:
2887 		ei_lba = 0;
2888 		lba = (u32)cmd[3] | (u32)cmd[2] << 8 |
2889 		      (u32)(cmd[1] & 0x1f) << 16;
2890 		num = (0 == cmd[4]) ? 256 : cmd[4];
2891 		check_prot = true;
2892 		break;
2893 	case WRITE_12:
2894 		ei_lba = 0;
2895 		lba = get_unaligned_be32(cmd + 2);
2896 		num = get_unaligned_be32(cmd + 6);
2897 		check_prot = true;
2898 		break;
2899 	case 0x53:	/* XDWRITEREAD(10) */
2900 		ei_lba = 0;
2901 		lba = get_unaligned_be32(cmd + 2);
2902 		num = get_unaligned_be16(cmd + 7);
2903 		check_prot = false;
2904 		break;
2905 	default:	/* assume WRITE(32) */
2906 		lba = get_unaligned_be64(cmd + 12);
2907 		ei_lba = get_unaligned_be32(cmd + 20);
2908 		num = get_unaligned_be32(cmd + 28);
2909 		check_prot = false;
2910 		break;
2911 	}
2912 	if (check_prot) {
2913 		if (scsi_debug_dif == SD_DIF_TYPE2_PROTECTION &&
2914 		    (cmd[1] & 0xe0)) {
2915 			mk_sense_invalid_opcode(scp);
2916 			return check_condition_result;
2917 		}
2918 		if ((scsi_debug_dif == SD_DIF_TYPE1_PROTECTION ||
2919 		     scsi_debug_dif == SD_DIF_TYPE3_PROTECTION) &&
2920 		    (cmd[1] & 0xe0) == 0)
2921 			sdev_printk(KERN_ERR, scp->device, "Unprotected WR "
2922 				    "to DIF device\n");
2923 	}
2924 
2925 	/* inline check_device_access_params() */
2926 	if (lba + num > sdebug_capacity) {
2927 		mk_sense_buffer(scp, ILLEGAL_REQUEST, LBA_OUT_OF_RANGE, 0);
2928 		return check_condition_result;
2929 	}
2930 	/* transfer length excessive (tie in to block limits VPD page) */
2931 	if (num > sdebug_store_sectors) {
2932 		/* needs work to find which cdb byte 'num' comes from */
2933 		mk_sense_buffer(scp, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB, 0);
2934 		return check_condition_result;
2935 	}
2936 
2937 	write_lock_irqsave(&atomic_rw, iflags);
2938 
2939 	/* DIX + T10 DIF */
2940 	if (scsi_debug_dix && scsi_prot_sg_count(scp)) {
2941 		int prot_ret = prot_verify_write(scp, lba, num, ei_lba);
2942 
2943 		if (prot_ret) {
2944 			write_unlock_irqrestore(&atomic_rw, iflags);
2945 			mk_sense_buffer(scp, ILLEGAL_REQUEST, 0x10, prot_ret);
2946 			return illegal_condition_result;
2947 		}
2948 	}
2949 
2950 	ret = do_device_access(scp, lba, num, true);
2951 	if (scsi_debug_lbp())
2952 		map_region(lba, num);
2953 	write_unlock_irqrestore(&atomic_rw, iflags);
2954 	if (-1 == ret)
2955 		return (DID_ERROR << 16);
2956 	else if ((ret < (num * scsi_debug_sector_size)) &&
2957 		 (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts))
2958 		sdev_printk(KERN_INFO, scp->device,
2959 			    "%s: write: cdb indicated=%u, IO sent=%d bytes\n",
2960 			    my_name, num * scsi_debug_sector_size, ret);
2961 
2962 	if (sdebug_any_injecting_opt) {
2963 		struct sdebug_scmd_extra_t *ep = scsi_cmd_priv(scp);
2964 
2965 		if (ep->inj_recovered) {
2966 			mk_sense_buffer(scp, RECOVERED_ERROR,
2967 					THRESHOLD_EXCEEDED, 0);
2968 			return check_condition_result;
2969 		} else if (ep->inj_dif) {
2970 			/* Logical block guard check failed */
2971 			mk_sense_buffer(scp, ABORTED_COMMAND, 0x10, 1);
2972 			return illegal_condition_result;
2973 		} else if (ep->inj_dix) {
2974 			mk_sense_buffer(scp, ILLEGAL_REQUEST, 0x10, 1);
2975 			return illegal_condition_result;
2976 		}
2977 	}
2978 	return 0;
2979 }
2980 
2981 static int
2982 resp_write_same(struct scsi_cmnd *scp, u64 lba, u32 num, u32 ei_lba,
2983 		bool unmap, bool ndob)
2984 {
2985 	unsigned long iflags;
2986 	unsigned long long i;
2987 	int ret;
2988 
2989 	ret = check_device_access_params(scp, lba, num);
2990 	if (ret)
2991 		return ret;
2992 
2993 	write_lock_irqsave(&atomic_rw, iflags);
2994 
2995 	if (unmap && scsi_debug_lbp()) {
2996 		unmap_region(lba, num);
2997 		goto out;
2998 	}
2999 
3000 	/* if ndob then zero 1 logical block, else fetch 1 logical block */
3001 	if (ndob) {
3002 		memset(fake_storep + (lba * scsi_debug_sector_size), 0,
3003 		       scsi_debug_sector_size);
3004 		ret = 0;
3005 	} else
3006 		ret = fetch_to_dev_buffer(scp, fake_storep +
3007 					       (lba * scsi_debug_sector_size),
3008 					  scsi_debug_sector_size);
3009 
3010 	if (-1 == ret) {
3011 		write_unlock_irqrestore(&atomic_rw, iflags);
3012 		return (DID_ERROR << 16);
3013 	} else if ((ret < (num * scsi_debug_sector_size)) &&
3014 		 (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts))
3015 		sdev_printk(KERN_INFO, scp->device,
3016 			    "%s: %s: cdb indicated=%u, IO sent=%d bytes\n",
3017 			    my_name, "write same",
3018 			    num * scsi_debug_sector_size, ret);
3019 
3020 	/* Copy first sector to remaining blocks */
3021 	for (i = 1 ; i < num ; i++)
3022 		memcpy(fake_storep + ((lba + i) * scsi_debug_sector_size),
3023 		       fake_storep + (lba * scsi_debug_sector_size),
3024 		       scsi_debug_sector_size);
3025 
3026 	if (scsi_debug_lbp())
3027 		map_region(lba, num);
3028 out:
3029 	write_unlock_irqrestore(&atomic_rw, iflags);
3030 
3031 	return 0;
3032 }
3033 
3034 static int
3035 resp_write_same_10(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
3036 {
3037 	u8 *cmd = scp->cmnd;
3038 	u32 lba;
3039 	u16 num;
3040 	u32 ei_lba = 0;
3041 	bool unmap = false;
3042 
3043 	if (cmd[1] & 0x8) {
3044 		if (scsi_debug_lbpws10 == 0) {
3045 			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 1, 3);
3046 			return check_condition_result;
3047 		} else
3048 			unmap = true;
3049 	}
3050 	lba = get_unaligned_be32(cmd + 2);
3051 	num = get_unaligned_be16(cmd + 7);
3052 	if (num > scsi_debug_write_same_length) {
3053 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 7, -1);
3054 		return check_condition_result;
3055 	}
3056 	return resp_write_same(scp, lba, num, ei_lba, unmap, false);
3057 }
3058 
3059 static int
3060 resp_write_same_16(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
3061 {
3062 	u8 *cmd = scp->cmnd;
3063 	u64 lba;
3064 	u32 num;
3065 	u32 ei_lba = 0;
3066 	bool unmap = false;
3067 	bool ndob = false;
3068 
3069 	if (cmd[1] & 0x8) {	/* UNMAP */
3070 		if (scsi_debug_lbpws == 0) {
3071 			mk_sense_invalid_fld(scp, SDEB_IN_CDB, 1, 3);
3072 			return check_condition_result;
3073 		} else
3074 			unmap = true;
3075 	}
3076 	if (cmd[1] & 0x1)  /* NDOB (no data-out buffer, assumes zeroes) */
3077 		ndob = true;
3078 	lba = get_unaligned_be64(cmd + 2);
3079 	num = get_unaligned_be32(cmd + 10);
3080 	if (num > scsi_debug_write_same_length) {
3081 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 10, -1);
3082 		return check_condition_result;
3083 	}
3084 	return resp_write_same(scp, lba, num, ei_lba, unmap, ndob);
3085 }
3086 
3087 /* Note the mode field is in the same position as the (lower) service action
3088  * field. For the Report supported operation codes command, SPC-4 suggests
3089  * each mode of this command should be reported separately; for future. */
3090 static int
3091 resp_write_buffer(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
3092 {
3093 	u8 *cmd = scp->cmnd;
3094 	struct scsi_device *sdp = scp->device;
3095 	struct sdebug_dev_info *dp;
3096 	u8 mode;
3097 
3098 	mode = cmd[1] & 0x1f;
3099 	switch (mode) {
3100 	case 0x4:	/* download microcode (MC) and activate (ACT) */
3101 		/* set UAs on this device only */
3102 		set_bit(SDEBUG_UA_BUS_RESET, devip->uas_bm);
3103 		set_bit(SDEBUG_UA_MICROCODE_CHANGED, devip->uas_bm);
3104 		break;
3105 	case 0x5:	/* download MC, save and ACT */
3106 		set_bit(SDEBUG_UA_MICROCODE_CHANGED_WO_RESET, devip->uas_bm);
3107 		break;
3108 	case 0x6:	/* download MC with offsets and ACT */
3109 		/* set UAs on most devices (LUs) in this target */
3110 		list_for_each_entry(dp,
3111 				    &devip->sdbg_host->dev_info_list,
3112 				    dev_list)
3113 			if (dp->target == sdp->id) {
3114 				set_bit(SDEBUG_UA_BUS_RESET, dp->uas_bm);
3115 				if (devip != dp)
3116 					set_bit(SDEBUG_UA_MICROCODE_CHANGED,
3117 						dp->uas_bm);
3118 			}
3119 		break;
3120 	case 0x7:	/* download MC with offsets, save, and ACT */
3121 		/* set UA on all devices (LUs) in this target */
3122 		list_for_each_entry(dp,
3123 				    &devip->sdbg_host->dev_info_list,
3124 				    dev_list)
3125 			if (dp->target == sdp->id)
3126 				set_bit(SDEBUG_UA_MICROCODE_CHANGED_WO_RESET,
3127 					dp->uas_bm);
3128 		break;
3129 	default:
3130 		/* do nothing for this command for other mode values */
3131 		break;
3132 	}
3133 	return 0;
3134 }
3135 
3136 static int
3137 resp_comp_write(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
3138 {
3139 	u8 *cmd = scp->cmnd;
3140 	u8 *arr;
3141 	u8 *fake_storep_hold;
3142 	u64 lba;
3143 	u32 dnum;
3144 	u32 lb_size = scsi_debug_sector_size;
3145 	u8 num;
3146 	unsigned long iflags;
3147 	int ret;
3148 	int retval = 0;
3149 
3150 	lba = get_unaligned_be64(cmd + 2);
3151 	num = cmd[13];		/* 1 to a maximum of 255 logical blocks */
3152 	if (0 == num)
3153 		return 0;	/* degenerate case, not an error */
3154 	if (scsi_debug_dif == SD_DIF_TYPE2_PROTECTION &&
3155 	    (cmd[1] & 0xe0)) {
3156 		mk_sense_invalid_opcode(scp);
3157 		return check_condition_result;
3158 	}
3159 	if ((scsi_debug_dif == SD_DIF_TYPE1_PROTECTION ||
3160 	     scsi_debug_dif == SD_DIF_TYPE3_PROTECTION) &&
3161 	    (cmd[1] & 0xe0) == 0)
3162 		sdev_printk(KERN_ERR, scp->device, "Unprotected WR "
3163 			    "to DIF device\n");
3164 
3165 	/* inline check_device_access_params() */
3166 	if (lba + num > sdebug_capacity) {
3167 		mk_sense_buffer(scp, ILLEGAL_REQUEST, LBA_OUT_OF_RANGE, 0);
3168 		return check_condition_result;
3169 	}
3170 	/* transfer length excessive (tie in to block limits VPD page) */
3171 	if (num > sdebug_store_sectors) {
3172 		/* needs work to find which cdb byte 'num' comes from */
3173 		mk_sense_buffer(scp, ILLEGAL_REQUEST, INVALID_FIELD_IN_CDB, 0);
3174 		return check_condition_result;
3175 	}
3176 	dnum = 2 * num;
3177 	arr = kzalloc(dnum * lb_size, GFP_ATOMIC);
3178 	if (NULL == arr) {
3179 		mk_sense_buffer(scp, ILLEGAL_REQUEST, INSUFF_RES_ASC,
3180 				INSUFF_RES_ASCQ);
3181 		return check_condition_result;
3182 	}
3183 
3184 	write_lock_irqsave(&atomic_rw, iflags);
3185 
3186 	/* trick do_device_access() to fetch both compare and write buffers
3187 	 * from data-in into arr. Safe (atomic) since write_lock held. */
3188 	fake_storep_hold = fake_storep;
3189 	fake_storep = arr;
3190 	ret = do_device_access(scp, 0, dnum, true);
3191 	fake_storep = fake_storep_hold;
3192 	if (ret == -1) {
3193 		retval = DID_ERROR << 16;
3194 		goto cleanup;
3195 	} else if ((ret < (dnum * lb_size)) &&
3196 		 (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts))
3197 		sdev_printk(KERN_INFO, scp->device, "%s: compare_write: cdb "
3198 			    "indicated=%u, IO sent=%d bytes\n", my_name,
3199 			    dnum * lb_size, ret);
3200 	if (!comp_write_worker(lba, num, arr)) {
3201 		mk_sense_buffer(scp, MISCOMPARE, MISCOMPARE_VERIFY_ASC, 0);
3202 		retval = check_condition_result;
3203 		goto cleanup;
3204 	}
3205 	if (scsi_debug_lbp())
3206 		map_region(lba, num);
3207 cleanup:
3208 	write_unlock_irqrestore(&atomic_rw, iflags);
3209 	kfree(arr);
3210 	return retval;
3211 }
3212 
3213 struct unmap_block_desc {
3214 	__be64	lba;
3215 	__be32	blocks;
3216 	__be32	__reserved;
3217 };
3218 
3219 static int
3220 resp_unmap(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
3221 {
3222 	unsigned char *buf;
3223 	struct unmap_block_desc *desc;
3224 	unsigned int i, payload_len, descriptors;
3225 	int ret;
3226 	unsigned long iflags;
3227 
3228 
3229 	if (!scsi_debug_lbp())
3230 		return 0;	/* fib and say its done */
3231 	payload_len = get_unaligned_be16(scp->cmnd + 7);
3232 	BUG_ON(scsi_bufflen(scp) != payload_len);
3233 
3234 	descriptors = (payload_len - 8) / 16;
3235 	if (descriptors > scsi_debug_unmap_max_desc) {
3236 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, 7, -1);
3237 		return check_condition_result;
3238 	}
3239 
3240 	buf = kmalloc(scsi_bufflen(scp), GFP_ATOMIC);
3241 	if (!buf) {
3242 		mk_sense_buffer(scp, ILLEGAL_REQUEST, INSUFF_RES_ASC,
3243 				INSUFF_RES_ASCQ);
3244 		return check_condition_result;
3245 	}
3246 
3247 	scsi_sg_copy_to_buffer(scp, buf, scsi_bufflen(scp));
3248 
3249 	BUG_ON(get_unaligned_be16(&buf[0]) != payload_len - 2);
3250 	BUG_ON(get_unaligned_be16(&buf[2]) != descriptors * 16);
3251 
3252 	desc = (void *)&buf[8];
3253 
3254 	write_lock_irqsave(&atomic_rw, iflags);
3255 
3256 	for (i = 0 ; i < descriptors ; i++) {
3257 		unsigned long long lba = get_unaligned_be64(&desc[i].lba);
3258 		unsigned int num = get_unaligned_be32(&desc[i].blocks);
3259 
3260 		ret = check_device_access_params(scp, lba, num);
3261 		if (ret)
3262 			goto out;
3263 
3264 		unmap_region(lba, num);
3265 	}
3266 
3267 	ret = 0;
3268 
3269 out:
3270 	write_unlock_irqrestore(&atomic_rw, iflags);
3271 	kfree(buf);
3272 
3273 	return ret;
3274 }
3275 
3276 #define SDEBUG_GET_LBA_STATUS_LEN 32
3277 
3278 static int
3279 resp_get_lba_status(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
3280 {
3281 	u8 *cmd = scp->cmnd;
3282 	u64 lba;
3283 	u32 alloc_len, mapped, num;
3284 	u8 arr[SDEBUG_GET_LBA_STATUS_LEN];
3285 	int ret;
3286 
3287 	lba = get_unaligned_be64(cmd + 2);
3288 	alloc_len = get_unaligned_be32(cmd + 10);
3289 
3290 	if (alloc_len < 24)
3291 		return 0;
3292 
3293 	ret = check_device_access_params(scp, lba, 1);
3294 	if (ret)
3295 		return ret;
3296 
3297 	if (scsi_debug_lbp())
3298 		mapped = map_state(lba, &num);
3299 	else {
3300 		mapped = 1;
3301 		/* following just in case virtual_gb changed */
3302 		sdebug_capacity = get_sdebug_capacity();
3303 		if (sdebug_capacity - lba <= 0xffffffff)
3304 			num = sdebug_capacity - lba;
3305 		else
3306 			num = 0xffffffff;
3307 	}
3308 
3309 	memset(arr, 0, SDEBUG_GET_LBA_STATUS_LEN);
3310 	put_unaligned_be32(20, arr);		/* Parameter Data Length */
3311 	put_unaligned_be64(lba, arr + 8);	/* LBA */
3312 	put_unaligned_be32(num, arr + 16);	/* Number of blocks */
3313 	arr[20] = !mapped;		/* prov_stat=0: mapped; 1: dealloc */
3314 
3315 	return fill_from_dev_buffer(scp, arr, SDEBUG_GET_LBA_STATUS_LEN);
3316 }
3317 
3318 #define SDEBUG_RLUN_ARR_SZ 256
3319 
3320 static int resp_report_luns(struct scsi_cmnd * scp,
3321 			    struct sdebug_dev_info * devip)
3322 {
3323 	unsigned int alloc_len;
3324 	int lun_cnt, i, upper, num, n, want_wlun, shortish;
3325 	u64 lun;
3326 	unsigned char *cmd = scp->cmnd;
3327 	int select_report = (int)cmd[2];
3328 	struct scsi_lun *one_lun;
3329 	unsigned char arr[SDEBUG_RLUN_ARR_SZ];
3330 	unsigned char * max_addr;
3331 
3332 	clear_luns_changed_on_target(devip);
3333 	alloc_len = cmd[9] + (cmd[8] << 8) + (cmd[7] << 16) + (cmd[6] << 24);
3334 	shortish = (alloc_len < 4);
3335 	if (shortish || (select_report > 2)) {
3336 		mk_sense_invalid_fld(scp, SDEB_IN_CDB, shortish ? 6 : 2, -1);
3337 		return check_condition_result;
3338 	}
3339 	/* can produce response with up to 16k luns (lun 0 to lun 16383) */
3340 	memset(arr, 0, SDEBUG_RLUN_ARR_SZ);
3341 	lun_cnt = scsi_debug_max_luns;
3342 	if (1 == select_report)
3343 		lun_cnt = 0;
3344 	else if (scsi_debug_no_lun_0 && (lun_cnt > 0))
3345 		--lun_cnt;
3346 	want_wlun = (select_report > 0) ? 1 : 0;
3347 	num = lun_cnt + want_wlun;
3348 	arr[2] = ((sizeof(struct scsi_lun) * num) >> 8) & 0xff;
3349 	arr[3] = (sizeof(struct scsi_lun) * num) & 0xff;
3350 	n = min((int)((SDEBUG_RLUN_ARR_SZ - 8) /
3351 			    sizeof(struct scsi_lun)), num);
3352 	if (n < num) {
3353 		want_wlun = 0;
3354 		lun_cnt = n;
3355 	}
3356 	one_lun = (struct scsi_lun *) &arr[8];
3357 	max_addr = arr + SDEBUG_RLUN_ARR_SZ;
3358 	for (i = 0, lun = (scsi_debug_no_lun_0 ? 1 : 0);
3359              ((i < lun_cnt) && ((unsigned char *)(one_lun + i) < max_addr));
3360 	     i++, lun++) {
3361 		upper = (lun >> 8) & 0x3f;
3362 		if (upper)
3363 			one_lun[i].scsi_lun[0] =
3364 			    (upper | (SAM2_LUN_ADDRESS_METHOD << 6));
3365 		one_lun[i].scsi_lun[1] = lun & 0xff;
3366 	}
3367 	if (want_wlun) {
3368 		one_lun[i].scsi_lun[0] = (SAM2_WLUN_REPORT_LUNS >> 8) & 0xff;
3369 		one_lun[i].scsi_lun[1] = SAM2_WLUN_REPORT_LUNS & 0xff;
3370 		i++;
3371 	}
3372 	alloc_len = (unsigned char *)(one_lun + i) - arr;
3373 	return fill_from_dev_buffer(scp, arr,
3374 				    min((int)alloc_len, SDEBUG_RLUN_ARR_SZ));
3375 }
3376 
3377 static int resp_xdwriteread(struct scsi_cmnd *scp, unsigned long long lba,
3378 			    unsigned int num, struct sdebug_dev_info *devip)
3379 {
3380 	int j;
3381 	unsigned char *kaddr, *buf;
3382 	unsigned int offset;
3383 	struct scsi_data_buffer *sdb = scsi_in(scp);
3384 	struct sg_mapping_iter miter;
3385 
3386 	/* better not to use temporary buffer. */
3387 	buf = kmalloc(scsi_bufflen(scp), GFP_ATOMIC);
3388 	if (!buf) {
3389 		mk_sense_buffer(scp, ILLEGAL_REQUEST, INSUFF_RES_ASC,
3390 				INSUFF_RES_ASCQ);
3391 		return check_condition_result;
3392 	}
3393 
3394 	scsi_sg_copy_to_buffer(scp, buf, scsi_bufflen(scp));
3395 
3396 	offset = 0;
3397 	sg_miter_start(&miter, sdb->table.sgl, sdb->table.nents,
3398 			SG_MITER_ATOMIC | SG_MITER_TO_SG);
3399 
3400 	while (sg_miter_next(&miter)) {
3401 		kaddr = miter.addr;
3402 		for (j = 0; j < miter.length; j++)
3403 			*(kaddr + j) ^= *(buf + offset + j);
3404 
3405 		offset += miter.length;
3406 	}
3407 	sg_miter_stop(&miter);
3408 	kfree(buf);
3409 
3410 	return 0;
3411 }
3412 
3413 static int
3414 resp_xdwriteread_10(struct scsi_cmnd *scp, struct sdebug_dev_info *devip)
3415 {
3416 	u8 *cmd = scp->cmnd;
3417 	u64 lba;
3418 	u32 num;
3419 	int errsts;
3420 
3421 	if (!scsi_bidi_cmnd(scp)) {
3422 		mk_sense_buffer(scp, ILLEGAL_REQUEST, INSUFF_RES_ASC,
3423 				INSUFF_RES_ASCQ);
3424 		return check_condition_result;
3425 	}
3426 	errsts = resp_read_dt0(scp, devip);
3427 	if (errsts)
3428 		return errsts;
3429 	if (!(cmd[1] & 0x4)) {		/* DISABLE_WRITE is not set */
3430 		errsts = resp_write_dt0(scp, devip);
3431 		if (errsts)
3432 			return errsts;
3433 	}
3434 	lba = get_unaligned_be32(cmd + 2);
3435 	num = get_unaligned_be16(cmd + 7);
3436 	return resp_xdwriteread(scp, lba, num, devip);
3437 }
3438 
3439 /* When timer or tasklet goes off this function is called. */
3440 static void sdebug_q_cmd_complete(unsigned long indx)
3441 {
3442 	int qa_indx;
3443 	int retiring = 0;
3444 	unsigned long iflags;
3445 	struct sdebug_queued_cmd *sqcp;
3446 	struct scsi_cmnd *scp;
3447 	struct sdebug_dev_info *devip;
3448 
3449 	atomic_inc(&sdebug_completions);
3450 	qa_indx = indx;
3451 	if ((qa_indx < 0) || (qa_indx >= SCSI_DEBUG_CANQUEUE)) {
3452 		pr_err("%s: wild qa_indx=%d\n", __func__, qa_indx);
3453 		return;
3454 	}
3455 	spin_lock_irqsave(&queued_arr_lock, iflags);
3456 	sqcp = &queued_arr[qa_indx];
3457 	scp = sqcp->a_cmnd;
3458 	if (NULL == scp) {
3459 		spin_unlock_irqrestore(&queued_arr_lock, iflags);
3460 		pr_err("%s: scp is NULL\n", __func__);
3461 		return;
3462 	}
3463 	devip = (struct sdebug_dev_info *)scp->device->hostdata;
3464 	if (devip)
3465 		atomic_dec(&devip->num_in_q);
3466 	else
3467 		pr_err("%s: devip=NULL\n", __func__);
3468 	if (atomic_read(&retired_max_queue) > 0)
3469 		retiring = 1;
3470 
3471 	sqcp->a_cmnd = NULL;
3472 	if (!test_and_clear_bit(qa_indx, queued_in_use_bm)) {
3473 		spin_unlock_irqrestore(&queued_arr_lock, iflags);
3474 		pr_err("%s: Unexpected completion\n", __func__);
3475 		return;
3476 	}
3477 
3478 	if (unlikely(retiring)) {	/* user has reduced max_queue */
3479 		int k, retval;
3480 
3481 		retval = atomic_read(&retired_max_queue);
3482 		if (qa_indx >= retval) {
3483 			spin_unlock_irqrestore(&queued_arr_lock, iflags);
3484 			pr_err("%s: index %d too large\n", __func__, retval);
3485 			return;
3486 		}
3487 		k = find_last_bit(queued_in_use_bm, retval);
3488 		if ((k < scsi_debug_max_queue) || (k == retval))
3489 			atomic_set(&retired_max_queue, 0);
3490 		else
3491 			atomic_set(&retired_max_queue, k + 1);
3492 	}
3493 	spin_unlock_irqrestore(&queued_arr_lock, iflags);
3494 	scp->scsi_done(scp); /* callback to mid level */
3495 }
3496 
3497 /* When high resolution timer goes off this function is called. */
3498 static enum hrtimer_restart
3499 sdebug_q_cmd_hrt_complete(struct hrtimer *timer)
3500 {
3501 	int qa_indx;
3502 	int retiring = 0;
3503 	unsigned long iflags;
3504 	struct sdebug_hrtimer *sd_hrtp = (struct sdebug_hrtimer *)timer;
3505 	struct sdebug_queued_cmd *sqcp;
3506 	struct scsi_cmnd *scp;
3507 	struct sdebug_dev_info *devip;
3508 
3509 	atomic_inc(&sdebug_completions);
3510 	qa_indx = sd_hrtp->qa_indx;
3511 	if ((qa_indx < 0) || (qa_indx >= SCSI_DEBUG_CANQUEUE)) {
3512 		pr_err("%s: wild qa_indx=%d\n", __func__, qa_indx);
3513 		goto the_end;
3514 	}
3515 	spin_lock_irqsave(&queued_arr_lock, iflags);
3516 	sqcp = &queued_arr[qa_indx];
3517 	scp = sqcp->a_cmnd;
3518 	if (NULL == scp) {
3519 		spin_unlock_irqrestore(&queued_arr_lock, iflags);
3520 		pr_err("%s: scp is NULL\n", __func__);
3521 		goto the_end;
3522 	}
3523 	devip = (struct sdebug_dev_info *)scp->device->hostdata;
3524 	if (devip)
3525 		atomic_dec(&devip->num_in_q);
3526 	else
3527 		pr_err("%s: devip=NULL\n", __func__);
3528 	if (atomic_read(&retired_max_queue) > 0)
3529 		retiring = 1;
3530 
3531 	sqcp->a_cmnd = NULL;
3532 	if (!test_and_clear_bit(qa_indx, queued_in_use_bm)) {
3533 		spin_unlock_irqrestore(&queued_arr_lock, iflags);
3534 		pr_err("%s: Unexpected completion\n", __func__);
3535 		goto the_end;
3536 	}
3537 
3538 	if (unlikely(retiring)) {	/* user has reduced max_queue */
3539 		int k, retval;
3540 
3541 		retval = atomic_read(&retired_max_queue);
3542 		if (qa_indx >= retval) {
3543 			spin_unlock_irqrestore(&queued_arr_lock, iflags);
3544 			pr_err("%s: index %d too large\n", __func__, retval);
3545 			goto the_end;
3546 		}
3547 		k = find_last_bit(queued_in_use_bm, retval);
3548 		if ((k < scsi_debug_max_queue) || (k == retval))
3549 			atomic_set(&retired_max_queue, 0);
3550 		else
3551 			atomic_set(&retired_max_queue, k + 1);
3552 	}
3553 	spin_unlock_irqrestore(&queued_arr_lock, iflags);
3554 	scp->scsi_done(scp); /* callback to mid level */
3555 the_end:
3556 	return HRTIMER_NORESTART;
3557 }
3558 
3559 static struct sdebug_dev_info *
3560 sdebug_device_create(struct sdebug_host_info *sdbg_host, gfp_t flags)
3561 {
3562 	struct sdebug_dev_info *devip;
3563 
3564 	devip = kzalloc(sizeof(*devip), flags);
3565 	if (devip) {
3566 		devip->sdbg_host = sdbg_host;
3567 		list_add_tail(&devip->dev_list, &sdbg_host->dev_info_list);
3568 	}
3569 	return devip;
3570 }
3571 
3572 static struct sdebug_dev_info * devInfoReg(struct scsi_device * sdev)
3573 {
3574 	struct sdebug_host_info * sdbg_host;
3575 	struct sdebug_dev_info * open_devip = NULL;
3576 	struct sdebug_dev_info * devip =
3577 			(struct sdebug_dev_info *)sdev->hostdata;
3578 
3579 	if (devip)
3580 		return devip;
3581 	sdbg_host = *(struct sdebug_host_info **)shost_priv(sdev->host);
3582 	if (!sdbg_host) {
3583 		pr_err("%s: Host info NULL\n", __func__);
3584 		return NULL;
3585         }
3586 	list_for_each_entry(devip, &sdbg_host->dev_info_list, dev_list) {
3587 		if ((devip->used) && (devip->channel == sdev->channel) &&
3588                     (devip->target == sdev->id) &&
3589                     (devip->lun == sdev->lun))
3590                         return devip;
3591 		else {
3592 			if ((!devip->used) && (!open_devip))
3593 				open_devip = devip;
3594 		}
3595 	}
3596 	if (!open_devip) { /* try and make a new one */
3597 		open_devip = sdebug_device_create(sdbg_host, GFP_ATOMIC);
3598 		if (!open_devip) {
3599 			printk(KERN_ERR "%s: out of memory at line %d\n",
3600 				__func__, __LINE__);
3601 			return NULL;
3602 		}
3603 	}
3604 
3605 	open_devip->channel = sdev->channel;
3606 	open_devip->target = sdev->id;
3607 	open_devip->lun = sdev->lun;
3608 	open_devip->sdbg_host = sdbg_host;
3609 	atomic_set(&open_devip->num_in_q, 0);
3610 	set_bit(SDEBUG_UA_POR, open_devip->uas_bm);
3611 	open_devip->used = true;
3612 	return open_devip;
3613 }
3614 
3615 static int scsi_debug_slave_alloc(struct scsi_device *sdp)
3616 {
3617 	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
3618 		printk(KERN_INFO "scsi_debug: slave_alloc <%u %u %u %llu>\n",
3619 		       sdp->host->host_no, sdp->channel, sdp->id, sdp->lun);
3620 	queue_flag_set_unlocked(QUEUE_FLAG_BIDI, sdp->request_queue);
3621 	return 0;
3622 }
3623 
3624 static int scsi_debug_slave_configure(struct scsi_device *sdp)
3625 {
3626 	struct sdebug_dev_info *devip;
3627 
3628 	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
3629 		printk(KERN_INFO "scsi_debug: slave_configure <%u %u %u %llu>\n",
3630 		       sdp->host->host_no, sdp->channel, sdp->id, sdp->lun);
3631 	if (sdp->host->max_cmd_len != SCSI_DEBUG_MAX_CMD_LEN)
3632 		sdp->host->max_cmd_len = SCSI_DEBUG_MAX_CMD_LEN;
3633 	devip = devInfoReg(sdp);
3634 	if (NULL == devip)
3635 		return 1;	/* no resources, will be marked offline */
3636 	sdp->hostdata = devip;
3637 	blk_queue_max_segment_size(sdp->request_queue, -1U);
3638 	if (scsi_debug_no_uld)
3639 		sdp->no_uld_attach = 1;
3640 	return 0;
3641 }
3642 
3643 static void scsi_debug_slave_destroy(struct scsi_device *sdp)
3644 {
3645 	struct sdebug_dev_info *devip =
3646 		(struct sdebug_dev_info *)sdp->hostdata;
3647 
3648 	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts)
3649 		printk(KERN_INFO "scsi_debug: slave_destroy <%u %u %u %llu>\n",
3650 		       sdp->host->host_no, sdp->channel, sdp->id, sdp->lun);
3651 	if (devip) {
3652 		/* make this slot available for re-use */
3653 		devip->used = false;
3654 		sdp->hostdata = NULL;
3655 	}
3656 }
3657 
3658 /* Returns 1 if cmnd found (deletes its timer or tasklet), else returns 0 */
3659 static int stop_queued_cmnd(struct scsi_cmnd *cmnd)
3660 {
3661 	unsigned long iflags;
3662 	int k, qmax, r_qmax;
3663 	struct sdebug_queued_cmd *sqcp;
3664 	struct sdebug_dev_info *devip;
3665 
3666 	spin_lock_irqsave(&queued_arr_lock, iflags);
3667 	qmax = scsi_debug_max_queue;
3668 	r_qmax = atomic_read(&retired_max_queue);
3669 	if (r_qmax > qmax)
3670 		qmax = r_qmax;
3671 	for (k = 0; k < qmax; ++k) {
3672 		if (test_bit(k, queued_in_use_bm)) {
3673 			sqcp = &queued_arr[k];
3674 			if (cmnd == sqcp->a_cmnd) {
3675 				devip = (struct sdebug_dev_info *)
3676 					cmnd->device->hostdata;
3677 				if (devip)
3678 					atomic_dec(&devip->num_in_q);
3679 				sqcp->a_cmnd = NULL;
3680 				spin_unlock_irqrestore(&queued_arr_lock,
3681 						       iflags);
3682 				if (scsi_debug_ndelay > 0) {
3683 					if (sqcp->sd_hrtp)
3684 						hrtimer_cancel(
3685 							&sqcp->sd_hrtp->hrt);
3686 				} else if (scsi_debug_delay > 0) {
3687 					if (sqcp->cmnd_timerp)
3688 						del_timer_sync(
3689 							sqcp->cmnd_timerp);
3690 				} else if (scsi_debug_delay < 0) {
3691 					if (sqcp->tletp)
3692 						tasklet_kill(sqcp->tletp);
3693 				}
3694 				clear_bit(k, queued_in_use_bm);
3695 				return 1;
3696 			}
3697 		}
3698 	}
3699 	spin_unlock_irqrestore(&queued_arr_lock, iflags);
3700 	return 0;
3701 }
3702 
3703 /* Deletes (stops) timers or tasklets of all queued commands */
3704 static void stop_all_queued(void)
3705 {
3706 	unsigned long iflags;
3707 	int k;
3708 	struct sdebug_queued_cmd *sqcp;
3709 	struct sdebug_dev_info *devip;
3710 
3711 	spin_lock_irqsave(&queued_arr_lock, iflags);
3712 	for (k = 0; k < SCSI_DEBUG_CANQUEUE; ++k) {
3713 		if (test_bit(k, queued_in_use_bm)) {
3714 			sqcp = &queued_arr[k];
3715 			if (sqcp->a_cmnd) {
3716 				devip = (struct sdebug_dev_info *)
3717 					sqcp->a_cmnd->device->hostdata;
3718 				if (devip)
3719 					atomic_dec(&devip->num_in_q);
3720 				sqcp->a_cmnd = NULL;
3721 				spin_unlock_irqrestore(&queued_arr_lock,
3722 						       iflags);
3723 				if (scsi_debug_ndelay > 0) {
3724 					if (sqcp->sd_hrtp)
3725 						hrtimer_cancel(
3726 							&sqcp->sd_hrtp->hrt);
3727 				} else if (scsi_debug_delay > 0) {
3728 					if (sqcp->cmnd_timerp)
3729 						del_timer_sync(
3730 							sqcp->cmnd_timerp);
3731 				} else if (scsi_debug_delay < 0) {
3732 					if (sqcp->tletp)
3733 						tasklet_kill(sqcp->tletp);
3734 				}
3735 				clear_bit(k, queued_in_use_bm);
3736 				spin_lock_irqsave(&queued_arr_lock, iflags);
3737 			}
3738 		}
3739 	}
3740 	spin_unlock_irqrestore(&queued_arr_lock, iflags);
3741 }
3742 
3743 /* Free queued command memory on heap */
3744 static void free_all_queued(void)
3745 {
3746 	unsigned long iflags;
3747 	int k;
3748 	struct sdebug_queued_cmd *sqcp;
3749 
3750 	spin_lock_irqsave(&queued_arr_lock, iflags);
3751 	for (k = 0; k < SCSI_DEBUG_CANQUEUE; ++k) {
3752 		sqcp = &queued_arr[k];
3753 		kfree(sqcp->cmnd_timerp);
3754 		sqcp->cmnd_timerp = NULL;
3755 		kfree(sqcp->tletp);
3756 		sqcp->tletp = NULL;
3757 		kfree(sqcp->sd_hrtp);
3758 		sqcp->sd_hrtp = NULL;
3759 	}
3760 	spin_unlock_irqrestore(&queued_arr_lock, iflags);
3761 }
3762 
3763 static int scsi_debug_abort(struct scsi_cmnd *SCpnt)
3764 {
3765 	++num_aborts;
3766 	if (SCpnt) {
3767 		if (SCpnt->device &&
3768 		    (SCSI_DEBUG_OPT_ALL_NOISE & scsi_debug_opts))
3769 			sdev_printk(KERN_INFO, SCpnt->device, "%s\n",
3770 				    __func__);
3771 		stop_queued_cmnd(SCpnt);
3772 	}
3773 	return SUCCESS;
3774 }
3775 
3776 static int scsi_debug_device_reset(struct scsi_cmnd * SCpnt)
3777 {
3778 	struct sdebug_dev_info * devip;
3779 
3780 	++num_dev_resets;
3781 	if (SCpnt && SCpnt->device) {
3782 		struct scsi_device *sdp = SCpnt->device;
3783 
3784 		if (SCSI_DEBUG_OPT_ALL_NOISE & scsi_debug_opts)
3785 			sdev_printk(KERN_INFO, sdp, "%s\n", __func__);
3786 		devip = devInfoReg(sdp);
3787 		if (devip)
3788 			set_bit(SDEBUG_UA_POR, devip->uas_bm);
3789 	}
3790 	return SUCCESS;
3791 }
3792 
3793 static int scsi_debug_target_reset(struct scsi_cmnd *SCpnt)
3794 {
3795 	struct sdebug_host_info *sdbg_host;
3796 	struct sdebug_dev_info *devip;
3797 	struct scsi_device *sdp;
3798 	struct Scsi_Host *hp;
3799 	int k = 0;
3800 
3801 	++num_target_resets;
3802 	if (!SCpnt)
3803 		goto lie;
3804 	sdp = SCpnt->device;
3805 	if (!sdp)
3806 		goto lie;
3807 	if (SCSI_DEBUG_OPT_ALL_NOISE & scsi_debug_opts)
3808 		sdev_printk(KERN_INFO, sdp, "%s\n", __func__);
3809 	hp = sdp->host;
3810 	if (!hp)
3811 		goto lie;
3812 	sdbg_host = *(struct sdebug_host_info **)shost_priv(hp);
3813 	if (sdbg_host) {
3814 		list_for_each_entry(devip,
3815 				    &sdbg_host->dev_info_list,
3816 				    dev_list)
3817 			if (devip->target == sdp->id) {
3818 				set_bit(SDEBUG_UA_BUS_RESET, devip->uas_bm);
3819 				++k;
3820 			}
3821 	}
3822 	if (SCSI_DEBUG_OPT_RESET_NOISE & scsi_debug_opts)
3823 		sdev_printk(KERN_INFO, sdp,
3824 			    "%s: %d device(s) found in target\n", __func__, k);
3825 lie:
3826 	return SUCCESS;
3827 }
3828 
3829 static int scsi_debug_bus_reset(struct scsi_cmnd * SCpnt)
3830 {
3831 	struct sdebug_host_info *sdbg_host;
3832 	struct sdebug_dev_info *devip;
3833         struct scsi_device * sdp;
3834         struct Scsi_Host * hp;
3835 	int k = 0;
3836 
3837 	++num_bus_resets;
3838 	if (!(SCpnt && SCpnt->device))
3839 		goto lie;
3840 	sdp = SCpnt->device;
3841 	if (SCSI_DEBUG_OPT_ALL_NOISE & scsi_debug_opts)
3842 		sdev_printk(KERN_INFO, sdp, "%s\n", __func__);
3843 	hp = sdp->host;
3844 	if (hp) {
3845 		sdbg_host = *(struct sdebug_host_info **)shost_priv(hp);
3846 		if (sdbg_host) {
3847 			list_for_each_entry(devip,
3848                                             &sdbg_host->dev_info_list,
3849 					    dev_list) {
3850 				set_bit(SDEBUG_UA_BUS_RESET, devip->uas_bm);
3851 				++k;
3852 			}
3853 		}
3854 	}
3855 	if (SCSI_DEBUG_OPT_RESET_NOISE & scsi_debug_opts)
3856 		sdev_printk(KERN_INFO, sdp,
3857 			    "%s: %d device(s) found in host\n", __func__, k);
3858 lie:
3859 	return SUCCESS;
3860 }
3861 
3862 static int scsi_debug_host_reset(struct scsi_cmnd * SCpnt)
3863 {
3864 	struct sdebug_host_info * sdbg_host;
3865 	struct sdebug_dev_info *devip;
3866 	int k = 0;
3867 
3868 	++num_host_resets;
3869 	if ((SCpnt->device) && (SCSI_DEBUG_OPT_ALL_NOISE & scsi_debug_opts))
3870 		sdev_printk(KERN_INFO, SCpnt->device, "%s\n", __func__);
3871         spin_lock(&sdebug_host_list_lock);
3872         list_for_each_entry(sdbg_host, &sdebug_host_list, host_list) {
3873 		list_for_each_entry(devip, &sdbg_host->dev_info_list,
3874 				    dev_list) {
3875 			set_bit(SDEBUG_UA_BUS_RESET, devip->uas_bm);
3876 			++k;
3877 		}
3878         }
3879         spin_unlock(&sdebug_host_list_lock);
3880 	stop_all_queued();
3881 	if (SCSI_DEBUG_OPT_RESET_NOISE & scsi_debug_opts)
3882 		sdev_printk(KERN_INFO, SCpnt->device,
3883 			    "%s: %d device(s) found\n", __func__, k);
3884 	return SUCCESS;
3885 }
3886 
3887 static void __init sdebug_build_parts(unsigned char *ramp,
3888 				      unsigned long store_size)
3889 {
3890 	struct partition * pp;
3891 	int starts[SDEBUG_MAX_PARTS + 2];
3892 	int sectors_per_part, num_sectors, k;
3893 	int heads_by_sects, start_sec, end_sec;
3894 
3895 	/* assume partition table already zeroed */
3896 	if ((scsi_debug_num_parts < 1) || (store_size < 1048576))
3897 		return;
3898 	if (scsi_debug_num_parts > SDEBUG_MAX_PARTS) {
3899 		scsi_debug_num_parts = SDEBUG_MAX_PARTS;
3900 		pr_warn("%s: reducing partitions to %d\n", __func__,
3901 			SDEBUG_MAX_PARTS);
3902 	}
3903 	num_sectors = (int)sdebug_store_sectors;
3904 	sectors_per_part = (num_sectors - sdebug_sectors_per)
3905 			   / scsi_debug_num_parts;
3906 	heads_by_sects = sdebug_heads * sdebug_sectors_per;
3907         starts[0] = sdebug_sectors_per;
3908 	for (k = 1; k < scsi_debug_num_parts; ++k)
3909 		starts[k] = ((k * sectors_per_part) / heads_by_sects)
3910 			    * heads_by_sects;
3911 	starts[scsi_debug_num_parts] = num_sectors;
3912 	starts[scsi_debug_num_parts + 1] = 0;
3913 
3914 	ramp[510] = 0x55;	/* magic partition markings */
3915 	ramp[511] = 0xAA;
3916 	pp = (struct partition *)(ramp + 0x1be);
3917 	for (k = 0; starts[k + 1]; ++k, ++pp) {
3918 		start_sec = starts[k];
3919 		end_sec = starts[k + 1] - 1;
3920 		pp->boot_ind = 0;
3921 
3922 		pp->cyl = start_sec / heads_by_sects;
3923 		pp->head = (start_sec - (pp->cyl * heads_by_sects))
3924 			   / sdebug_sectors_per;
3925 		pp->sector = (start_sec % sdebug_sectors_per) + 1;
3926 
3927 		pp->end_cyl = end_sec / heads_by_sects;
3928 		pp->end_head = (end_sec - (pp->end_cyl * heads_by_sects))
3929 			       / sdebug_sectors_per;
3930 		pp->end_sector = (end_sec % sdebug_sectors_per) + 1;
3931 
3932 		pp->start_sect = cpu_to_le32(start_sec);
3933 		pp->nr_sects = cpu_to_le32(end_sec - start_sec + 1);
3934 		pp->sys_ind = 0x83;	/* plain Linux partition */
3935 	}
3936 }
3937 
3938 static int
3939 schedule_resp(struct scsi_cmnd *cmnd, struct sdebug_dev_info *devip,
3940 	      int scsi_result, int delta_jiff)
3941 {
3942 	unsigned long iflags;
3943 	int k, num_in_q, qdepth, inject;
3944 	struct sdebug_queued_cmd *sqcp = NULL;
3945 	struct scsi_device *sdp = cmnd->device;
3946 
3947 	if (NULL == cmnd || NULL == devip) {
3948 		pr_warn("%s: called with NULL cmnd or devip pointer\n",
3949 			__func__);
3950 		/* no particularly good error to report back */
3951 		return SCSI_MLQUEUE_HOST_BUSY;
3952 	}
3953 	if ((scsi_result) && (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts))
3954 		sdev_printk(KERN_INFO, sdp, "%s: non-zero result=0x%x\n",
3955 			    __func__, scsi_result);
3956 	if (delta_jiff == 0)
3957 		goto respond_in_thread;
3958 
3959 	/* schedule the response at a later time if resources permit */
3960 	spin_lock_irqsave(&queued_arr_lock, iflags);
3961 	num_in_q = atomic_read(&devip->num_in_q);
3962 	qdepth = cmnd->device->queue_depth;
3963 	inject = 0;
3964 	if ((qdepth > 0) && (num_in_q >= qdepth)) {
3965 		if (scsi_result) {
3966 			spin_unlock_irqrestore(&queued_arr_lock, iflags);
3967 			goto respond_in_thread;
3968 		} else
3969 			scsi_result = device_qfull_result;
3970 	} else if ((scsi_debug_every_nth != 0) &&
3971 		   (SCSI_DEBUG_OPT_RARE_TSF & scsi_debug_opts) &&
3972 		   (scsi_result == 0)) {
3973 		if ((num_in_q == (qdepth - 1)) &&
3974 		    (atomic_inc_return(&sdebug_a_tsf) >=
3975 		     abs(scsi_debug_every_nth))) {
3976 			atomic_set(&sdebug_a_tsf, 0);
3977 			inject = 1;
3978 			scsi_result = device_qfull_result;
3979 		}
3980 	}
3981 
3982 	k = find_first_zero_bit(queued_in_use_bm, scsi_debug_max_queue);
3983 	if (k >= scsi_debug_max_queue) {
3984 		spin_unlock_irqrestore(&queued_arr_lock, iflags);
3985 		if (scsi_result)
3986 			goto respond_in_thread;
3987 		else if (SCSI_DEBUG_OPT_ALL_TSF & scsi_debug_opts)
3988 			scsi_result = device_qfull_result;
3989 		if (SCSI_DEBUG_OPT_Q_NOISE & scsi_debug_opts)
3990 			sdev_printk(KERN_INFO, sdp,
3991 				    "%s: max_queue=%d exceeded, %s\n",
3992 				    __func__, scsi_debug_max_queue,
3993 				    (scsi_result ?  "status: TASK SET FULL" :
3994 						    "report: host busy"));
3995 		if (scsi_result)
3996 			goto respond_in_thread;
3997 		else
3998 			return SCSI_MLQUEUE_HOST_BUSY;
3999 	}
4000 	__set_bit(k, queued_in_use_bm);
4001 	atomic_inc(&devip->num_in_q);
4002 	sqcp = &queued_arr[k];
4003 	sqcp->a_cmnd = cmnd;
4004 	cmnd->result = scsi_result;
4005 	spin_unlock_irqrestore(&queued_arr_lock, iflags);
4006 	if (delta_jiff > 0) {
4007 		if (NULL == sqcp->cmnd_timerp) {
4008 			sqcp->cmnd_timerp = kmalloc(sizeof(struct timer_list),
4009 						    GFP_ATOMIC);
4010 			if (NULL == sqcp->cmnd_timerp)
4011 				return SCSI_MLQUEUE_HOST_BUSY;
4012 			init_timer(sqcp->cmnd_timerp);
4013 		}
4014 		sqcp->cmnd_timerp->function = sdebug_q_cmd_complete;
4015 		sqcp->cmnd_timerp->data = k;
4016 		sqcp->cmnd_timerp->expires = get_jiffies_64() + delta_jiff;
4017 		add_timer(sqcp->cmnd_timerp);
4018 	} else if (scsi_debug_ndelay > 0) {
4019 		ktime_t kt = ktime_set(0, scsi_debug_ndelay);
4020 		struct sdebug_hrtimer *sd_hp = sqcp->sd_hrtp;
4021 
4022 		if (NULL == sd_hp) {
4023 			sd_hp = kmalloc(sizeof(*sd_hp), GFP_ATOMIC);
4024 			if (NULL == sd_hp)
4025 				return SCSI_MLQUEUE_HOST_BUSY;
4026 			sqcp->sd_hrtp = sd_hp;
4027 			hrtimer_init(&sd_hp->hrt, CLOCK_MONOTONIC,
4028 				     HRTIMER_MODE_REL);
4029 			sd_hp->hrt.function = sdebug_q_cmd_hrt_complete;
4030 			sd_hp->qa_indx = k;
4031 		}
4032 		hrtimer_start(&sd_hp->hrt, kt, HRTIMER_MODE_REL);
4033 	} else {	/* delay < 0 */
4034 		if (NULL == sqcp->tletp) {
4035 			sqcp->tletp = kmalloc(sizeof(*sqcp->tletp),
4036 					      GFP_ATOMIC);
4037 			if (NULL == sqcp->tletp)
4038 				return SCSI_MLQUEUE_HOST_BUSY;
4039 			tasklet_init(sqcp->tletp,
4040 				     sdebug_q_cmd_complete, k);
4041 		}
4042 		if (-1 == delta_jiff)
4043 			tasklet_hi_schedule(sqcp->tletp);
4044 		else
4045 			tasklet_schedule(sqcp->tletp);
4046 	}
4047 	if ((SCSI_DEBUG_OPT_Q_NOISE & scsi_debug_opts) &&
4048 	    (scsi_result == device_qfull_result))
4049 		sdev_printk(KERN_INFO, sdp,
4050 			    "%s: num_in_q=%d +1, %s%s\n", __func__,
4051 			    num_in_q, (inject ? "<inject> " : ""),
4052 			    "status: TASK SET FULL");
4053 	return 0;
4054 
4055 respond_in_thread:	/* call back to mid-layer using invocation thread */
4056 	cmnd->result = scsi_result;
4057 	cmnd->scsi_done(cmnd);
4058 	return 0;
4059 }
4060 
4061 /* Note: The following macros create attribute files in the
4062    /sys/module/scsi_debug/parameters directory. Unfortunately this
4063    driver is unaware of a change and cannot trigger auxiliary actions
4064    as it can when the corresponding attribute in the
4065    /sys/bus/pseudo/drivers/scsi_debug directory is changed.
4066  */
4067 module_param_named(add_host, scsi_debug_add_host, int, S_IRUGO | S_IWUSR);
4068 module_param_named(ato, scsi_debug_ato, int, S_IRUGO);
4069 module_param_named(clustering, scsi_debug_clustering, bool, S_IRUGO | S_IWUSR);
4070 module_param_named(delay, scsi_debug_delay, int, S_IRUGO | S_IWUSR);
4071 module_param_named(dev_size_mb, scsi_debug_dev_size_mb, int, S_IRUGO);
4072 module_param_named(dif, scsi_debug_dif, int, S_IRUGO);
4073 module_param_named(dix, scsi_debug_dix, int, S_IRUGO);
4074 module_param_named(dsense, scsi_debug_dsense, int, S_IRUGO | S_IWUSR);
4075 module_param_named(every_nth, scsi_debug_every_nth, int, S_IRUGO | S_IWUSR);
4076 module_param_named(fake_rw, scsi_debug_fake_rw, int, S_IRUGO | S_IWUSR);
4077 module_param_named(guard, scsi_debug_guard, uint, S_IRUGO);
4078 module_param_named(host_lock, scsi_debug_host_lock, bool, S_IRUGO | S_IWUSR);
4079 module_param_named(lbpu, scsi_debug_lbpu, int, S_IRUGO);
4080 module_param_named(lbpws, scsi_debug_lbpws, int, S_IRUGO);
4081 module_param_named(lbpws10, scsi_debug_lbpws10, int, S_IRUGO);
4082 module_param_named(lbprz, scsi_debug_lbprz, int, S_IRUGO);
4083 module_param_named(lowest_aligned, scsi_debug_lowest_aligned, int, S_IRUGO);
4084 module_param_named(max_luns, scsi_debug_max_luns, int, S_IRUGO | S_IWUSR);
4085 module_param_named(max_queue, scsi_debug_max_queue, int, S_IRUGO | S_IWUSR);
4086 module_param_named(ndelay, scsi_debug_ndelay, int, S_IRUGO | S_IWUSR);
4087 module_param_named(no_lun_0, scsi_debug_no_lun_0, int, S_IRUGO | S_IWUSR);
4088 module_param_named(no_uld, scsi_debug_no_uld, int, S_IRUGO);
4089 module_param_named(num_parts, scsi_debug_num_parts, int, S_IRUGO);
4090 module_param_named(num_tgts, scsi_debug_num_tgts, int, S_IRUGO | S_IWUSR);
4091 module_param_named(opt_blks, scsi_debug_opt_blks, int, S_IRUGO);
4092 module_param_named(opts, scsi_debug_opts, int, S_IRUGO | S_IWUSR);
4093 module_param_named(physblk_exp, scsi_debug_physblk_exp, int, S_IRUGO);
4094 module_param_named(ptype, scsi_debug_ptype, int, S_IRUGO | S_IWUSR);
4095 module_param_named(removable, scsi_debug_removable, bool, S_IRUGO | S_IWUSR);
4096 module_param_named(scsi_level, scsi_debug_scsi_level, int, S_IRUGO);
4097 module_param_named(sector_size, scsi_debug_sector_size, int, S_IRUGO);
4098 module_param_named(strict, scsi_debug_strict, bool, S_IRUGO | S_IWUSR);
4099 module_param_named(unmap_alignment, scsi_debug_unmap_alignment, int, S_IRUGO);
4100 module_param_named(unmap_granularity, scsi_debug_unmap_granularity, int, S_IRUGO);
4101 module_param_named(unmap_max_blocks, scsi_debug_unmap_max_blocks, int, S_IRUGO);
4102 module_param_named(unmap_max_desc, scsi_debug_unmap_max_desc, int, S_IRUGO);
4103 module_param_named(virtual_gb, scsi_debug_virtual_gb, int, S_IRUGO | S_IWUSR);
4104 module_param_named(vpd_use_hostno, scsi_debug_vpd_use_hostno, int,
4105 		   S_IRUGO | S_IWUSR);
4106 module_param_named(write_same_length, scsi_debug_write_same_length, int,
4107 		   S_IRUGO | S_IWUSR);
4108 
4109 MODULE_AUTHOR("Eric Youngdale + Douglas Gilbert");
4110 MODULE_DESCRIPTION("SCSI debug adapter driver");
4111 MODULE_LICENSE("GPL");
4112 MODULE_VERSION(SCSI_DEBUG_VERSION);
4113 
4114 MODULE_PARM_DESC(add_host, "0..127 hosts allowed(def=1)");
4115 MODULE_PARM_DESC(ato, "application tag ownership: 0=disk 1=host (def=1)");
4116 MODULE_PARM_DESC(clustering, "when set enables larger transfers (def=0)");
4117 MODULE_PARM_DESC(delay, "response delay (def=1 jiffy); 0:imm, -1,-2:tiny");
4118 MODULE_PARM_DESC(dev_size_mb, "size in MiB of ram shared by devs(def=8)");
4119 MODULE_PARM_DESC(dif, "data integrity field type: 0-3 (def=0)");
4120 MODULE_PARM_DESC(dix, "data integrity extensions mask (def=0)");
4121 MODULE_PARM_DESC(dsense, "use descriptor sense format(def=0 -> fixed)");
4122 MODULE_PARM_DESC(every_nth, "timeout every nth command(def=0)");
4123 MODULE_PARM_DESC(fake_rw, "fake reads/writes instead of copying (def=0)");
4124 MODULE_PARM_DESC(guard, "protection checksum: 0=crc, 1=ip (def=0)");
4125 MODULE_PARM_DESC(host_lock, "use host_lock around all commands (def=0)");
4126 MODULE_PARM_DESC(lbpu, "enable LBP, support UNMAP command (def=0)");
4127 MODULE_PARM_DESC(lbpws, "enable LBP, support WRITE SAME(16) with UNMAP bit (def=0)");
4128 MODULE_PARM_DESC(lbpws10, "enable LBP, support WRITE SAME(10) with UNMAP bit (def=0)");
4129 MODULE_PARM_DESC(lbprz, "unmapped blocks return 0 on read (def=1)");
4130 MODULE_PARM_DESC(lowest_aligned, "lowest aligned lba (def=0)");
4131 MODULE_PARM_DESC(max_luns, "number of LUNs per target to simulate(def=1)");
4132 MODULE_PARM_DESC(max_queue, "max number of queued commands (1 to max(def))");
4133 MODULE_PARM_DESC(ndelay, "response delay in nanoseconds (def=0 -> ignore)");
4134 MODULE_PARM_DESC(no_lun_0, "no LU number 0 (def=0 -> have lun 0)");
4135 MODULE_PARM_DESC(no_uld, "stop ULD (e.g. sd driver) attaching (def=0))");
4136 MODULE_PARM_DESC(num_parts, "number of partitions(def=0)");
4137 MODULE_PARM_DESC(num_tgts, "number of targets per host to simulate(def=1)");
4138 MODULE_PARM_DESC(opt_blks, "optimal transfer length in block (def=64)");
4139 MODULE_PARM_DESC(opts, "1->noise, 2->medium_err, 4->timeout, 8->recovered_err... (def=0)");
4140 MODULE_PARM_DESC(physblk_exp, "physical block exponent (def=0)");
4141 MODULE_PARM_DESC(ptype, "SCSI peripheral type(def=0[disk])");
4142 MODULE_PARM_DESC(removable, "claim to have removable media (def=0)");
4143 MODULE_PARM_DESC(scsi_level, "SCSI level to simulate(def=6[SPC-4])");
4144 MODULE_PARM_DESC(sector_size, "logical block size in bytes (def=512)");
4145 MODULE_PARM_DESC(strict, "stricter checks: reserved field in cdb (def=0)");
4146 MODULE_PARM_DESC(unmap_alignment, "lowest aligned thin provisioning lba (def=0)");
4147 MODULE_PARM_DESC(unmap_granularity, "thin provisioning granularity in blocks (def=1)");
4148 MODULE_PARM_DESC(unmap_max_blocks, "max # of blocks can be unmapped in one cmd (def=0xffffffff)");
4149 MODULE_PARM_DESC(unmap_max_desc, "max # of ranges that can be unmapped in one cmd (def=256)");
4150 MODULE_PARM_DESC(virtual_gb, "virtual gigabyte (GiB) size (def=0 -> use dev_size_mb)");
4151 MODULE_PARM_DESC(vpd_use_hostno, "0 -> dev ids ignore hostno (def=1 -> unique dev ids)");
4152 MODULE_PARM_DESC(write_same_length, "Maximum blocks per WRITE SAME cmd (def=0xffff)");
4153 
4154 static char sdebug_info[256];
4155 
4156 static const char * scsi_debug_info(struct Scsi_Host * shp)
4157 {
4158 	sprintf(sdebug_info, "scsi_debug, version %s [%s], "
4159 		"dev_size_mb=%d, opts=0x%x", SCSI_DEBUG_VERSION,
4160 		scsi_debug_version_date, scsi_debug_dev_size_mb,
4161 		scsi_debug_opts);
4162 	return sdebug_info;
4163 }
4164 
4165 /* 'echo <val> > /proc/scsi/scsi_debug/<host_id>' writes to opts */
4166 static int scsi_debug_write_info(struct Scsi_Host *host, char *buffer, int length)
4167 {
4168 	char arr[16];
4169 	int opts;
4170 	int minLen = length > 15 ? 15 : length;
4171 
4172 	if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO))
4173 		return -EACCES;
4174 	memcpy(arr, buffer, minLen);
4175 	arr[minLen] = '\0';
4176 	if (1 != sscanf(arr, "%d", &opts))
4177 		return -EINVAL;
4178 	scsi_debug_opts = opts;
4179 	if (scsi_debug_every_nth != 0)
4180 		atomic_set(&sdebug_cmnd_count, 0);
4181 	return length;
4182 }
4183 
4184 /* Output seen with 'cat /proc/scsi/scsi_debug/<host_id>'. It will be the
4185  * same for each scsi_debug host (if more than one). Some of the counters
4186  * output are not atomics so might be inaccurate in a busy system. */
4187 static int scsi_debug_show_info(struct seq_file *m, struct Scsi_Host *host)
4188 {
4189 	int f, l;
4190 	char b[32];
4191 
4192 	if (scsi_debug_every_nth > 0)
4193 		snprintf(b, sizeof(b), " (curr:%d)",
4194 			 ((SCSI_DEBUG_OPT_RARE_TSF & scsi_debug_opts) ?
4195 				atomic_read(&sdebug_a_tsf) :
4196 				atomic_read(&sdebug_cmnd_count)));
4197 	else
4198 		b[0] = '\0';
4199 
4200 	seq_printf(m, "scsi_debug adapter driver, version %s [%s]\n"
4201 		"num_tgts=%d, shared (ram) size=%d MB, opts=0x%x, "
4202 		"every_nth=%d%s\n"
4203 		"delay=%d, ndelay=%d, max_luns=%d, q_completions=%d\n"
4204 		"sector_size=%d bytes, cylinders=%d, heads=%d, sectors=%d\n"
4205 		"command aborts=%d; RESETs: device=%d, target=%d, bus=%d, "
4206 		"host=%d\ndix_reads=%d dix_writes=%d dif_errors=%d "
4207 		"usec_in_jiffy=%lu\n",
4208 		SCSI_DEBUG_VERSION, scsi_debug_version_date,
4209 		scsi_debug_num_tgts, scsi_debug_dev_size_mb, scsi_debug_opts,
4210 		scsi_debug_every_nth, b, scsi_debug_delay, scsi_debug_ndelay,
4211 		scsi_debug_max_luns, atomic_read(&sdebug_completions),
4212 		scsi_debug_sector_size, sdebug_cylinders_per, sdebug_heads,
4213 		sdebug_sectors_per, num_aborts, num_dev_resets,
4214 		num_target_resets, num_bus_resets, num_host_resets,
4215 		dix_reads, dix_writes, dif_errors, TICK_NSEC / 1000);
4216 
4217 	f = find_first_bit(queued_in_use_bm, scsi_debug_max_queue);
4218 	if (f != scsi_debug_max_queue) {
4219 		l = find_last_bit(queued_in_use_bm, scsi_debug_max_queue);
4220 		seq_printf(m, "   %s BUSY: first,last bits set: %d,%d\n",
4221 			   "queued_in_use_bm", f, l);
4222 	}
4223 	return 0;
4224 }
4225 
4226 static ssize_t delay_show(struct device_driver *ddp, char *buf)
4227 {
4228         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_delay);
4229 }
4230 /* Returns -EBUSY if delay is being changed and commands are queued */
4231 static ssize_t delay_store(struct device_driver *ddp, const char *buf,
4232 			   size_t count)
4233 {
4234 	int delay, res;
4235 
4236 	if ((count > 0) && (1 == sscanf(buf, "%d", &delay))) {
4237 		res = count;
4238 		if (scsi_debug_delay != delay) {
4239 			unsigned long iflags;
4240 			int k;
4241 
4242 			spin_lock_irqsave(&queued_arr_lock, iflags);
4243 			k = find_first_bit(queued_in_use_bm,
4244 					   scsi_debug_max_queue);
4245 			if (k != scsi_debug_max_queue)
4246 				res = -EBUSY;	/* have queued commands */
4247 			else {
4248 				scsi_debug_delay = delay;
4249 				scsi_debug_ndelay = 0;
4250 			}
4251 			spin_unlock_irqrestore(&queued_arr_lock, iflags);
4252 		}
4253 		return res;
4254 	}
4255 	return -EINVAL;
4256 }
4257 static DRIVER_ATTR_RW(delay);
4258 
4259 static ssize_t ndelay_show(struct device_driver *ddp, char *buf)
4260 {
4261 	return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_ndelay);
4262 }
4263 /* Returns -EBUSY if ndelay is being changed and commands are queued */
4264 /* If > 0 and accepted then scsi_debug_delay is set to DELAY_OVERRIDDEN */
4265 static ssize_t ndelay_store(struct device_driver *ddp, const char *buf,
4266 			   size_t count)
4267 {
4268 	unsigned long iflags;
4269 	int ndelay, res, k;
4270 
4271 	if ((count > 0) && (1 == sscanf(buf, "%d", &ndelay)) &&
4272 	    (ndelay >= 0) && (ndelay < 1000000000)) {
4273 		res = count;
4274 		if (scsi_debug_ndelay != ndelay) {
4275 			spin_lock_irqsave(&queued_arr_lock, iflags);
4276 			k = find_first_bit(queued_in_use_bm,
4277 					   scsi_debug_max_queue);
4278 			if (k != scsi_debug_max_queue)
4279 				res = -EBUSY;	/* have queued commands */
4280 			else {
4281 				scsi_debug_ndelay = ndelay;
4282 				scsi_debug_delay = ndelay ? DELAY_OVERRIDDEN
4283 							  : DEF_DELAY;
4284 			}
4285 			spin_unlock_irqrestore(&queued_arr_lock, iflags);
4286 		}
4287 		return res;
4288 	}
4289 	return -EINVAL;
4290 }
4291 static DRIVER_ATTR_RW(ndelay);
4292 
4293 static ssize_t opts_show(struct device_driver *ddp, char *buf)
4294 {
4295         return scnprintf(buf, PAGE_SIZE, "0x%x\n", scsi_debug_opts);
4296 }
4297 
4298 static ssize_t opts_store(struct device_driver *ddp, const char *buf,
4299 			  size_t count)
4300 {
4301         int opts;
4302 	char work[20];
4303 
4304         if (1 == sscanf(buf, "%10s", work)) {
4305 		if (0 == strncasecmp(work,"0x", 2)) {
4306 			if (1 == sscanf(&work[2], "%x", &opts))
4307 				goto opts_done;
4308 		} else {
4309 			if (1 == sscanf(work, "%d", &opts))
4310 				goto opts_done;
4311 		}
4312 	}
4313 	return -EINVAL;
4314 opts_done:
4315 	scsi_debug_opts = opts;
4316 	if (SCSI_DEBUG_OPT_RECOVERED_ERR & opts)
4317 		sdebug_any_injecting_opt = true;
4318 	else if (SCSI_DEBUG_OPT_TRANSPORT_ERR & opts)
4319 		sdebug_any_injecting_opt = true;
4320 	else if (SCSI_DEBUG_OPT_DIF_ERR & opts)
4321 		sdebug_any_injecting_opt = true;
4322 	else if (SCSI_DEBUG_OPT_DIX_ERR & opts)
4323 		sdebug_any_injecting_opt = true;
4324 	else if (SCSI_DEBUG_OPT_SHORT_TRANSFER & opts)
4325 		sdebug_any_injecting_opt = true;
4326 	atomic_set(&sdebug_cmnd_count, 0);
4327 	atomic_set(&sdebug_a_tsf, 0);
4328 	return count;
4329 }
4330 static DRIVER_ATTR_RW(opts);
4331 
4332 static ssize_t ptype_show(struct device_driver *ddp, char *buf)
4333 {
4334         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_ptype);
4335 }
4336 static ssize_t ptype_store(struct device_driver *ddp, const char *buf,
4337 			   size_t count)
4338 {
4339         int n;
4340 
4341 	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4342 		scsi_debug_ptype = n;
4343 		return count;
4344 	}
4345 	return -EINVAL;
4346 }
4347 static DRIVER_ATTR_RW(ptype);
4348 
4349 static ssize_t dsense_show(struct device_driver *ddp, char *buf)
4350 {
4351         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_dsense);
4352 }
4353 static ssize_t dsense_store(struct device_driver *ddp, const char *buf,
4354 			    size_t count)
4355 {
4356         int n;
4357 
4358 	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4359 		scsi_debug_dsense = n;
4360 		return count;
4361 	}
4362 	return -EINVAL;
4363 }
4364 static DRIVER_ATTR_RW(dsense);
4365 
4366 static ssize_t fake_rw_show(struct device_driver *ddp, char *buf)
4367 {
4368         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_fake_rw);
4369 }
4370 static ssize_t fake_rw_store(struct device_driver *ddp, const char *buf,
4371 			     size_t count)
4372 {
4373         int n;
4374 
4375 	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4376 		n = (n > 0);
4377 		scsi_debug_fake_rw = (scsi_debug_fake_rw > 0);
4378 		if (scsi_debug_fake_rw != n) {
4379 			if ((0 == n) && (NULL == fake_storep)) {
4380 				unsigned long sz =
4381 					(unsigned long)scsi_debug_dev_size_mb *
4382 					1048576;
4383 
4384 				fake_storep = vmalloc(sz);
4385 				if (NULL == fake_storep) {
4386 					pr_err("%s: out of memory, 9\n",
4387 					       __func__);
4388 					return -ENOMEM;
4389 				}
4390 				memset(fake_storep, 0, sz);
4391 			}
4392 			scsi_debug_fake_rw = n;
4393 		}
4394 		return count;
4395 	}
4396 	return -EINVAL;
4397 }
4398 static DRIVER_ATTR_RW(fake_rw);
4399 
4400 static ssize_t no_lun_0_show(struct device_driver *ddp, char *buf)
4401 {
4402         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_no_lun_0);
4403 }
4404 static ssize_t no_lun_0_store(struct device_driver *ddp, const char *buf,
4405 			      size_t count)
4406 {
4407         int n;
4408 
4409 	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4410 		scsi_debug_no_lun_0 = n;
4411 		return count;
4412 	}
4413 	return -EINVAL;
4414 }
4415 static DRIVER_ATTR_RW(no_lun_0);
4416 
4417 static ssize_t num_tgts_show(struct device_driver *ddp, char *buf)
4418 {
4419         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_num_tgts);
4420 }
4421 static ssize_t num_tgts_store(struct device_driver *ddp, const char *buf,
4422 			      size_t count)
4423 {
4424         int n;
4425 
4426 	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4427 		scsi_debug_num_tgts = n;
4428 		sdebug_max_tgts_luns();
4429 		return count;
4430 	}
4431 	return -EINVAL;
4432 }
4433 static DRIVER_ATTR_RW(num_tgts);
4434 
4435 static ssize_t dev_size_mb_show(struct device_driver *ddp, char *buf)
4436 {
4437         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_dev_size_mb);
4438 }
4439 static DRIVER_ATTR_RO(dev_size_mb);
4440 
4441 static ssize_t num_parts_show(struct device_driver *ddp, char *buf)
4442 {
4443         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_num_parts);
4444 }
4445 static DRIVER_ATTR_RO(num_parts);
4446 
4447 static ssize_t every_nth_show(struct device_driver *ddp, char *buf)
4448 {
4449         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_every_nth);
4450 }
4451 static ssize_t every_nth_store(struct device_driver *ddp, const char *buf,
4452 			       size_t count)
4453 {
4454         int nth;
4455 
4456 	if ((count > 0) && (1 == sscanf(buf, "%d", &nth))) {
4457 		scsi_debug_every_nth = nth;
4458 		atomic_set(&sdebug_cmnd_count, 0);
4459 		return count;
4460 	}
4461 	return -EINVAL;
4462 }
4463 static DRIVER_ATTR_RW(every_nth);
4464 
4465 static ssize_t max_luns_show(struct device_driver *ddp, char *buf)
4466 {
4467         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_max_luns);
4468 }
4469 static ssize_t max_luns_store(struct device_driver *ddp, const char *buf,
4470 			      size_t count)
4471 {
4472         int n;
4473 	bool changed;
4474 
4475 	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4476 		changed = (scsi_debug_max_luns != n);
4477 		scsi_debug_max_luns = n;
4478 		sdebug_max_tgts_luns();
4479 		if (changed && (scsi_debug_scsi_level >= 5)) {	/* >= SPC-3 */
4480 			struct sdebug_host_info *sdhp;
4481 			struct sdebug_dev_info *dp;
4482 
4483 			spin_lock(&sdebug_host_list_lock);
4484 			list_for_each_entry(sdhp, &sdebug_host_list,
4485 					    host_list) {
4486 				list_for_each_entry(dp, &sdhp->dev_info_list,
4487 						    dev_list) {
4488 					set_bit(SDEBUG_UA_LUNS_CHANGED,
4489 						dp->uas_bm);
4490 				}
4491 			}
4492 			spin_unlock(&sdebug_host_list_lock);
4493 		}
4494 		return count;
4495 	}
4496 	return -EINVAL;
4497 }
4498 static DRIVER_ATTR_RW(max_luns);
4499 
4500 static ssize_t max_queue_show(struct device_driver *ddp, char *buf)
4501 {
4502         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_max_queue);
4503 }
4504 /* N.B. max_queue can be changed while there are queued commands. In flight
4505  * commands beyond the new max_queue will be completed. */
4506 static ssize_t max_queue_store(struct device_driver *ddp, const char *buf,
4507 			       size_t count)
4508 {
4509 	unsigned long iflags;
4510 	int n, k;
4511 
4512 	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n > 0) &&
4513 	    (n <= SCSI_DEBUG_CANQUEUE)) {
4514 		spin_lock_irqsave(&queued_arr_lock, iflags);
4515 		k = find_last_bit(queued_in_use_bm, SCSI_DEBUG_CANQUEUE);
4516 		scsi_debug_max_queue = n;
4517 		if (SCSI_DEBUG_CANQUEUE == k)
4518 			atomic_set(&retired_max_queue, 0);
4519 		else if (k >= n)
4520 			atomic_set(&retired_max_queue, k + 1);
4521 		else
4522 			atomic_set(&retired_max_queue, 0);
4523 		spin_unlock_irqrestore(&queued_arr_lock, iflags);
4524 		return count;
4525 	}
4526 	return -EINVAL;
4527 }
4528 static DRIVER_ATTR_RW(max_queue);
4529 
4530 static ssize_t no_uld_show(struct device_driver *ddp, char *buf)
4531 {
4532         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_no_uld);
4533 }
4534 static DRIVER_ATTR_RO(no_uld);
4535 
4536 static ssize_t scsi_level_show(struct device_driver *ddp, char *buf)
4537 {
4538         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_scsi_level);
4539 }
4540 static DRIVER_ATTR_RO(scsi_level);
4541 
4542 static ssize_t virtual_gb_show(struct device_driver *ddp, char *buf)
4543 {
4544         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_virtual_gb);
4545 }
4546 static ssize_t virtual_gb_store(struct device_driver *ddp, const char *buf,
4547 				size_t count)
4548 {
4549         int n;
4550 	bool changed;
4551 
4552 	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4553 		changed = (scsi_debug_virtual_gb != n);
4554 		scsi_debug_virtual_gb = n;
4555 		sdebug_capacity = get_sdebug_capacity();
4556 		if (changed) {
4557 			struct sdebug_host_info *sdhp;
4558 			struct sdebug_dev_info *dp;
4559 
4560 			spin_lock(&sdebug_host_list_lock);
4561 			list_for_each_entry(sdhp, &sdebug_host_list,
4562 					    host_list) {
4563 				list_for_each_entry(dp, &sdhp->dev_info_list,
4564 						    dev_list) {
4565 					set_bit(SDEBUG_UA_CAPACITY_CHANGED,
4566 						dp->uas_bm);
4567 				}
4568 			}
4569 			spin_unlock(&sdebug_host_list_lock);
4570 		}
4571 		return count;
4572 	}
4573 	return -EINVAL;
4574 }
4575 static DRIVER_ATTR_RW(virtual_gb);
4576 
4577 static ssize_t add_host_show(struct device_driver *ddp, char *buf)
4578 {
4579         return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_add_host);
4580 }
4581 
4582 static ssize_t add_host_store(struct device_driver *ddp, const char *buf,
4583 			      size_t count)
4584 {
4585 	int delta_hosts;
4586 
4587 	if (sscanf(buf, "%d", &delta_hosts) != 1)
4588 		return -EINVAL;
4589 	if (delta_hosts > 0) {
4590 		do {
4591 			sdebug_add_adapter();
4592 		} while (--delta_hosts);
4593 	} else if (delta_hosts < 0) {
4594 		do {
4595 			sdebug_remove_adapter();
4596 		} while (++delta_hosts);
4597 	}
4598 	return count;
4599 }
4600 static DRIVER_ATTR_RW(add_host);
4601 
4602 static ssize_t vpd_use_hostno_show(struct device_driver *ddp, char *buf)
4603 {
4604 	return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_vpd_use_hostno);
4605 }
4606 static ssize_t vpd_use_hostno_store(struct device_driver *ddp, const char *buf,
4607 				    size_t count)
4608 {
4609 	int n;
4610 
4611 	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4612 		scsi_debug_vpd_use_hostno = n;
4613 		return count;
4614 	}
4615 	return -EINVAL;
4616 }
4617 static DRIVER_ATTR_RW(vpd_use_hostno);
4618 
4619 static ssize_t sector_size_show(struct device_driver *ddp, char *buf)
4620 {
4621 	return scnprintf(buf, PAGE_SIZE, "%u\n", scsi_debug_sector_size);
4622 }
4623 static DRIVER_ATTR_RO(sector_size);
4624 
4625 static ssize_t dix_show(struct device_driver *ddp, char *buf)
4626 {
4627 	return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_dix);
4628 }
4629 static DRIVER_ATTR_RO(dix);
4630 
4631 static ssize_t dif_show(struct device_driver *ddp, char *buf)
4632 {
4633 	return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_dif);
4634 }
4635 static DRIVER_ATTR_RO(dif);
4636 
4637 static ssize_t guard_show(struct device_driver *ddp, char *buf)
4638 {
4639 	return scnprintf(buf, PAGE_SIZE, "%u\n", scsi_debug_guard);
4640 }
4641 static DRIVER_ATTR_RO(guard);
4642 
4643 static ssize_t ato_show(struct device_driver *ddp, char *buf)
4644 {
4645 	return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_ato);
4646 }
4647 static DRIVER_ATTR_RO(ato);
4648 
4649 static ssize_t map_show(struct device_driver *ddp, char *buf)
4650 {
4651 	ssize_t count;
4652 
4653 	if (!scsi_debug_lbp())
4654 		return scnprintf(buf, PAGE_SIZE, "0-%u\n",
4655 				 sdebug_store_sectors);
4656 
4657 	count = scnprintf(buf, PAGE_SIZE - 1, "%*pbl",
4658 			  (int)map_size, map_storep);
4659 	buf[count++] = '\n';
4660 	buf[count] = '\0';
4661 
4662 	return count;
4663 }
4664 static DRIVER_ATTR_RO(map);
4665 
4666 static ssize_t removable_show(struct device_driver *ddp, char *buf)
4667 {
4668 	return scnprintf(buf, PAGE_SIZE, "%d\n", scsi_debug_removable ? 1 : 0);
4669 }
4670 static ssize_t removable_store(struct device_driver *ddp, const char *buf,
4671 			       size_t count)
4672 {
4673 	int n;
4674 
4675 	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4676 		scsi_debug_removable = (n > 0);
4677 		return count;
4678 	}
4679 	return -EINVAL;
4680 }
4681 static DRIVER_ATTR_RW(removable);
4682 
4683 static ssize_t host_lock_show(struct device_driver *ddp, char *buf)
4684 {
4685 	return scnprintf(buf, PAGE_SIZE, "%d\n", !!scsi_debug_host_lock);
4686 }
4687 /* Returns -EBUSY if host_lock is being changed and commands are queued */
4688 static ssize_t host_lock_store(struct device_driver *ddp, const char *buf,
4689 			       size_t count)
4690 {
4691 	int n, res;
4692 
4693 	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4694 		bool new_host_lock = (n > 0);
4695 
4696 		res = count;
4697 		if (new_host_lock != scsi_debug_host_lock) {
4698 			unsigned long iflags;
4699 			int k;
4700 
4701 			spin_lock_irqsave(&queued_arr_lock, iflags);
4702 			k = find_first_bit(queued_in_use_bm,
4703 					   scsi_debug_max_queue);
4704 			if (k != scsi_debug_max_queue)
4705 				res = -EBUSY;	/* have queued commands */
4706 			else
4707 				scsi_debug_host_lock = new_host_lock;
4708 			spin_unlock_irqrestore(&queued_arr_lock, iflags);
4709 		}
4710 		return res;
4711 	}
4712 	return -EINVAL;
4713 }
4714 static DRIVER_ATTR_RW(host_lock);
4715 
4716 static ssize_t strict_show(struct device_driver *ddp, char *buf)
4717 {
4718 	return scnprintf(buf, PAGE_SIZE, "%d\n", !!scsi_debug_strict);
4719 }
4720 static ssize_t strict_store(struct device_driver *ddp, const char *buf,
4721 			    size_t count)
4722 {
4723 	int n;
4724 
4725 	if ((count > 0) && (1 == sscanf(buf, "%d", &n)) && (n >= 0)) {
4726 		scsi_debug_strict = (n > 0);
4727 		return count;
4728 	}
4729 	return -EINVAL;
4730 }
4731 static DRIVER_ATTR_RW(strict);
4732 
4733 
4734 /* Note: The following array creates attribute files in the
4735    /sys/bus/pseudo/drivers/scsi_debug directory. The advantage of these
4736    files (over those found in the /sys/module/scsi_debug/parameters
4737    directory) is that auxiliary actions can be triggered when an attribute
4738    is changed. For example see: sdebug_add_host_store() above.
4739  */
4740 
4741 static struct attribute *sdebug_drv_attrs[] = {
4742 	&driver_attr_delay.attr,
4743 	&driver_attr_opts.attr,
4744 	&driver_attr_ptype.attr,
4745 	&driver_attr_dsense.attr,
4746 	&driver_attr_fake_rw.attr,
4747 	&driver_attr_no_lun_0.attr,
4748 	&driver_attr_num_tgts.attr,
4749 	&driver_attr_dev_size_mb.attr,
4750 	&driver_attr_num_parts.attr,
4751 	&driver_attr_every_nth.attr,
4752 	&driver_attr_max_luns.attr,
4753 	&driver_attr_max_queue.attr,
4754 	&driver_attr_no_uld.attr,
4755 	&driver_attr_scsi_level.attr,
4756 	&driver_attr_virtual_gb.attr,
4757 	&driver_attr_add_host.attr,
4758 	&driver_attr_vpd_use_hostno.attr,
4759 	&driver_attr_sector_size.attr,
4760 	&driver_attr_dix.attr,
4761 	&driver_attr_dif.attr,
4762 	&driver_attr_guard.attr,
4763 	&driver_attr_ato.attr,
4764 	&driver_attr_map.attr,
4765 	&driver_attr_removable.attr,
4766 	&driver_attr_host_lock.attr,
4767 	&driver_attr_ndelay.attr,
4768 	&driver_attr_strict.attr,
4769 	NULL,
4770 };
4771 ATTRIBUTE_GROUPS(sdebug_drv);
4772 
4773 static struct device *pseudo_primary;
4774 
4775 static int __init scsi_debug_init(void)
4776 {
4777 	unsigned long sz;
4778 	int host_to_add;
4779 	int k;
4780 	int ret;
4781 
4782 	atomic_set(&sdebug_cmnd_count, 0);
4783 	atomic_set(&sdebug_completions, 0);
4784 	atomic_set(&retired_max_queue, 0);
4785 
4786 	if (scsi_debug_ndelay >= 1000000000) {
4787 		pr_warn("%s: ndelay must be less than 1 second, ignored\n",
4788 			__func__);
4789 		scsi_debug_ndelay = 0;
4790 	} else if (scsi_debug_ndelay > 0)
4791 		scsi_debug_delay = DELAY_OVERRIDDEN;
4792 
4793 	switch (scsi_debug_sector_size) {
4794 	case  512:
4795 	case 1024:
4796 	case 2048:
4797 	case 4096:
4798 		break;
4799 	default:
4800 		pr_err("%s: invalid sector_size %d\n", __func__,
4801 		       scsi_debug_sector_size);
4802 		return -EINVAL;
4803 	}
4804 
4805 	switch (scsi_debug_dif) {
4806 
4807 	case SD_DIF_TYPE0_PROTECTION:
4808 	case SD_DIF_TYPE1_PROTECTION:
4809 	case SD_DIF_TYPE2_PROTECTION:
4810 	case SD_DIF_TYPE3_PROTECTION:
4811 		break;
4812 
4813 	default:
4814 		pr_err("%s: dif must be 0, 1, 2 or 3\n", __func__);
4815 		return -EINVAL;
4816 	}
4817 
4818 	if (scsi_debug_guard > 1) {
4819 		pr_err("%s: guard must be 0 or 1\n", __func__);
4820 		return -EINVAL;
4821 	}
4822 
4823 	if (scsi_debug_ato > 1) {
4824 		pr_err("%s: ato must be 0 or 1\n", __func__);
4825 		return -EINVAL;
4826 	}
4827 
4828 	if (scsi_debug_physblk_exp > 15) {
4829 		pr_err("%s: invalid physblk_exp %u\n", __func__,
4830 		       scsi_debug_physblk_exp);
4831 		return -EINVAL;
4832 	}
4833 
4834 	if (scsi_debug_lowest_aligned > 0x3fff) {
4835 		pr_err("%s: lowest_aligned too big: %u\n", __func__,
4836 		       scsi_debug_lowest_aligned);
4837 		return -EINVAL;
4838 	}
4839 
4840 	if (scsi_debug_dev_size_mb < 1)
4841 		scsi_debug_dev_size_mb = 1;  /* force minimum 1 MB ramdisk */
4842 	sz = (unsigned long)scsi_debug_dev_size_mb * 1048576;
4843 	sdebug_store_sectors = sz / scsi_debug_sector_size;
4844 	sdebug_capacity = get_sdebug_capacity();
4845 
4846 	/* play around with geometry, don't waste too much on track 0 */
4847 	sdebug_heads = 8;
4848 	sdebug_sectors_per = 32;
4849 	if (scsi_debug_dev_size_mb >= 16)
4850 		sdebug_heads = 32;
4851 	else if (scsi_debug_dev_size_mb >= 256)
4852 		sdebug_heads = 64;
4853 	sdebug_cylinders_per = (unsigned long)sdebug_capacity /
4854 			       (sdebug_sectors_per * sdebug_heads);
4855 	if (sdebug_cylinders_per >= 1024) {
4856 		/* other LLDs do this; implies >= 1GB ram disk ... */
4857 		sdebug_heads = 255;
4858 		sdebug_sectors_per = 63;
4859 		sdebug_cylinders_per = (unsigned long)sdebug_capacity /
4860 			       (sdebug_sectors_per * sdebug_heads);
4861 	}
4862 
4863 	if (0 == scsi_debug_fake_rw) {
4864 		fake_storep = vmalloc(sz);
4865 		if (NULL == fake_storep) {
4866 			pr_err("%s: out of memory, 1\n", __func__);
4867 			return -ENOMEM;
4868 		}
4869 		memset(fake_storep, 0, sz);
4870 		if (scsi_debug_num_parts > 0)
4871 			sdebug_build_parts(fake_storep, sz);
4872 	}
4873 
4874 	if (scsi_debug_dix) {
4875 		int dif_size;
4876 
4877 		dif_size = sdebug_store_sectors * sizeof(struct sd_dif_tuple);
4878 		dif_storep = vmalloc(dif_size);
4879 
4880 		pr_err("%s: dif_storep %u bytes @ %p\n", __func__, dif_size,
4881 			dif_storep);
4882 
4883 		if (dif_storep == NULL) {
4884 			pr_err("%s: out of mem. (DIX)\n", __func__);
4885 			ret = -ENOMEM;
4886 			goto free_vm;
4887 		}
4888 
4889 		memset(dif_storep, 0xff, dif_size);
4890 	}
4891 
4892 	/* Logical Block Provisioning */
4893 	if (scsi_debug_lbp()) {
4894 		scsi_debug_unmap_max_blocks =
4895 			clamp(scsi_debug_unmap_max_blocks, 0U, 0xffffffffU);
4896 
4897 		scsi_debug_unmap_max_desc =
4898 			clamp(scsi_debug_unmap_max_desc, 0U, 256U);
4899 
4900 		scsi_debug_unmap_granularity =
4901 			clamp(scsi_debug_unmap_granularity, 1U, 0xffffffffU);
4902 
4903 		if (scsi_debug_unmap_alignment &&
4904 		    scsi_debug_unmap_granularity <=
4905 		    scsi_debug_unmap_alignment) {
4906 			pr_err("%s: ERR: unmap_granularity <= unmap_alignment\n",
4907 			       __func__);
4908 			return -EINVAL;
4909 		}
4910 
4911 		map_size = lba_to_map_index(sdebug_store_sectors - 1) + 1;
4912 		map_storep = vmalloc(BITS_TO_LONGS(map_size) * sizeof(long));
4913 
4914 		pr_info("%s: %lu provisioning blocks\n", __func__, map_size);
4915 
4916 		if (map_storep == NULL) {
4917 			pr_err("%s: out of mem. (MAP)\n", __func__);
4918 			ret = -ENOMEM;
4919 			goto free_vm;
4920 		}
4921 
4922 		bitmap_zero(map_storep, map_size);
4923 
4924 		/* Map first 1KB for partition table */
4925 		if (scsi_debug_num_parts)
4926 			map_region(0, 2);
4927 	}
4928 
4929 	pseudo_primary = root_device_register("pseudo_0");
4930 	if (IS_ERR(pseudo_primary)) {
4931 		pr_warn("%s: root_device_register() error\n", __func__);
4932 		ret = PTR_ERR(pseudo_primary);
4933 		goto free_vm;
4934 	}
4935 	ret = bus_register(&pseudo_lld_bus);
4936 	if (ret < 0) {
4937 		pr_warn("%s: bus_register error: %d\n", __func__, ret);
4938 		goto dev_unreg;
4939 	}
4940 	ret = driver_register(&sdebug_driverfs_driver);
4941 	if (ret < 0) {
4942 		pr_warn("%s: driver_register error: %d\n", __func__, ret);
4943 		goto bus_unreg;
4944 	}
4945 
4946 	host_to_add = scsi_debug_add_host;
4947         scsi_debug_add_host = 0;
4948 
4949         for (k = 0; k < host_to_add; k++) {
4950                 if (sdebug_add_adapter()) {
4951 			pr_err("%s: sdebug_add_adapter failed k=%d\n",
4952 				__func__, k);
4953                         break;
4954                 }
4955         }
4956 
4957 	if (SCSI_DEBUG_OPT_NOISE & scsi_debug_opts) {
4958 		pr_info("%s: built %d host(s)\n", __func__,
4959 			scsi_debug_add_host);
4960 	}
4961 	return 0;
4962 
4963 bus_unreg:
4964 	bus_unregister(&pseudo_lld_bus);
4965 dev_unreg:
4966 	root_device_unregister(pseudo_primary);
4967 free_vm:
4968 	if (map_storep)
4969 		vfree(map_storep);
4970 	if (dif_storep)
4971 		vfree(dif_storep);
4972 	vfree(fake_storep);
4973 
4974 	return ret;
4975 }
4976 
4977 static void __exit scsi_debug_exit(void)
4978 {
4979 	int k = scsi_debug_add_host;
4980 
4981 	stop_all_queued();
4982 	free_all_queued();
4983 	for (; k; k--)
4984 		sdebug_remove_adapter();
4985 	driver_unregister(&sdebug_driverfs_driver);
4986 	bus_unregister(&pseudo_lld_bus);
4987 	root_device_unregister(pseudo_primary);
4988 
4989 	if (dif_storep)
4990 		vfree(dif_storep);
4991 
4992 	vfree(fake_storep);
4993 }
4994 
4995 device_initcall(scsi_debug_init);
4996 module_exit(scsi_debug_exit);
4997 
4998 static void sdebug_release_adapter(struct device * dev)
4999 {
5000         struct sdebug_host_info *sdbg_host;
5001 
5002 	sdbg_host = to_sdebug_host(dev);
5003         kfree(sdbg_host);
5004 }
5005 
5006 static int sdebug_add_adapter(void)
5007 {
5008 	int k, devs_per_host;
5009         int error = 0;
5010         struct sdebug_host_info *sdbg_host;
5011 	struct sdebug_dev_info *sdbg_devinfo, *tmp;
5012 
5013         sdbg_host = kzalloc(sizeof(*sdbg_host),GFP_KERNEL);
5014         if (NULL == sdbg_host) {
5015                 printk(KERN_ERR "%s: out of memory at line %d\n",
5016                        __func__, __LINE__);
5017                 return -ENOMEM;
5018         }
5019 
5020         INIT_LIST_HEAD(&sdbg_host->dev_info_list);
5021 
5022 	devs_per_host = scsi_debug_num_tgts * scsi_debug_max_luns;
5023         for (k = 0; k < devs_per_host; k++) {
5024 		sdbg_devinfo = sdebug_device_create(sdbg_host, GFP_KERNEL);
5025 		if (!sdbg_devinfo) {
5026                         printk(KERN_ERR "%s: out of memory at line %d\n",
5027                                __func__, __LINE__);
5028                         error = -ENOMEM;
5029 			goto clean;
5030                 }
5031         }
5032 
5033         spin_lock(&sdebug_host_list_lock);
5034         list_add_tail(&sdbg_host->host_list, &sdebug_host_list);
5035         spin_unlock(&sdebug_host_list_lock);
5036 
5037         sdbg_host->dev.bus = &pseudo_lld_bus;
5038         sdbg_host->dev.parent = pseudo_primary;
5039         sdbg_host->dev.release = &sdebug_release_adapter;
5040         dev_set_name(&sdbg_host->dev, "adapter%d", scsi_debug_add_host);
5041 
5042         error = device_register(&sdbg_host->dev);
5043 
5044         if (error)
5045 		goto clean;
5046 
5047 	++scsi_debug_add_host;
5048         return error;
5049 
5050 clean:
5051 	list_for_each_entry_safe(sdbg_devinfo, tmp, &sdbg_host->dev_info_list,
5052 				 dev_list) {
5053 		list_del(&sdbg_devinfo->dev_list);
5054 		kfree(sdbg_devinfo);
5055 	}
5056 
5057 	kfree(sdbg_host);
5058         return error;
5059 }
5060 
5061 static void sdebug_remove_adapter(void)
5062 {
5063         struct sdebug_host_info * sdbg_host = NULL;
5064 
5065         spin_lock(&sdebug_host_list_lock);
5066         if (!list_empty(&sdebug_host_list)) {
5067                 sdbg_host = list_entry(sdebug_host_list.prev,
5068                                        struct sdebug_host_info, host_list);
5069 		list_del(&sdbg_host->host_list);
5070 	}
5071         spin_unlock(&sdebug_host_list_lock);
5072 
5073 	if (!sdbg_host)
5074 		return;
5075 
5076         device_unregister(&sdbg_host->dev);
5077         --scsi_debug_add_host;
5078 }
5079 
5080 static int
5081 sdebug_change_qdepth(struct scsi_device *sdev, int qdepth)
5082 {
5083 	int num_in_q = 0;
5084 	unsigned long iflags;
5085 	struct sdebug_dev_info *devip;
5086 
5087 	spin_lock_irqsave(&queued_arr_lock, iflags);
5088 	devip = (struct sdebug_dev_info *)sdev->hostdata;
5089 	if (NULL == devip) {
5090 		spin_unlock_irqrestore(&queued_arr_lock, iflags);
5091 		return	-ENODEV;
5092 	}
5093 	num_in_q = atomic_read(&devip->num_in_q);
5094 	spin_unlock_irqrestore(&queued_arr_lock, iflags);
5095 
5096 	if (qdepth < 1)
5097 		qdepth = 1;
5098 	/* allow to exceed max host queued_arr elements for testing */
5099 	if (qdepth > SCSI_DEBUG_CANQUEUE + 10)
5100 		qdepth = SCSI_DEBUG_CANQUEUE + 10;
5101 	scsi_change_queue_depth(sdev, qdepth);
5102 
5103 	if (SCSI_DEBUG_OPT_Q_NOISE & scsi_debug_opts) {
5104 		sdev_printk(KERN_INFO, sdev,
5105 			    "%s: qdepth=%d, num_in_q=%d\n",
5106 			    __func__, qdepth, num_in_q);
5107 	}
5108 	return sdev->queue_depth;
5109 }
5110 
5111 static int
5112 check_inject(struct scsi_cmnd *scp)
5113 {
5114 	struct sdebug_scmd_extra_t *ep = scsi_cmd_priv(scp);
5115 
5116 	memset(ep, 0, sizeof(struct sdebug_scmd_extra_t));
5117 
5118 	if (atomic_inc_return(&sdebug_cmnd_count) >=
5119 	    abs(scsi_debug_every_nth)) {
5120 		atomic_set(&sdebug_cmnd_count, 0);
5121 		if (scsi_debug_every_nth < -1)
5122 			scsi_debug_every_nth = -1;
5123 		if (SCSI_DEBUG_OPT_TIMEOUT & scsi_debug_opts)
5124 			return 1; /* ignore command causing timeout */
5125 		else if (SCSI_DEBUG_OPT_MAC_TIMEOUT & scsi_debug_opts &&
5126 			 scsi_medium_access_command(scp))
5127 			return 1; /* time out reads and writes */
5128 		if (sdebug_any_injecting_opt) {
5129 			int opts = scsi_debug_opts;
5130 
5131 			if (SCSI_DEBUG_OPT_RECOVERED_ERR & opts)
5132 				ep->inj_recovered = true;
5133 			else if (SCSI_DEBUG_OPT_TRANSPORT_ERR & opts)
5134 				ep->inj_transport = true;
5135 			else if (SCSI_DEBUG_OPT_DIF_ERR & opts)
5136 				ep->inj_dif = true;
5137 			else if (SCSI_DEBUG_OPT_DIX_ERR & opts)
5138 				ep->inj_dix = true;
5139 			else if (SCSI_DEBUG_OPT_SHORT_TRANSFER & opts)
5140 				ep->inj_short = true;
5141 		}
5142 	}
5143 	return 0;
5144 }
5145 
5146 static int
5147 scsi_debug_queuecommand(struct scsi_cmnd *scp)
5148 {
5149 	u8 sdeb_i;
5150 	struct scsi_device *sdp = scp->device;
5151 	const struct opcode_info_t *oip;
5152 	const struct opcode_info_t *r_oip;
5153 	struct sdebug_dev_info *devip;
5154 	u8 *cmd = scp->cmnd;
5155 	int (*r_pfp)(struct scsi_cmnd *, struct sdebug_dev_info *);
5156 	int k, na;
5157 	int errsts = 0;
5158 	int errsts_no_connect = DID_NO_CONNECT << 16;
5159 	u32 flags;
5160 	u16 sa;
5161 	u8 opcode = cmd[0];
5162 	bool has_wlun_rl;
5163 	bool debug = !!(SCSI_DEBUG_OPT_NOISE & scsi_debug_opts);
5164 
5165 	scsi_set_resid(scp, 0);
5166 	if (debug && !(SCSI_DEBUG_OPT_NO_CDB_NOISE & scsi_debug_opts)) {
5167 		char b[120];
5168 		int n, len, sb;
5169 
5170 		len = scp->cmd_len;
5171 		sb = (int)sizeof(b);
5172 		if (len > 32)
5173 			strcpy(b, "too long, over 32 bytes");
5174 		else {
5175 			for (k = 0, n = 0; k < len && n < sb; ++k)
5176 				n += scnprintf(b + n, sb - n, "%02x ",
5177 					       (u32)cmd[k]);
5178 		}
5179 		sdev_printk(KERN_INFO, sdp, "%s: cmd %s\n", my_name, b);
5180 	}
5181 	has_wlun_rl = (sdp->lun == SAM2_WLUN_REPORT_LUNS);
5182 	if ((sdp->lun >= scsi_debug_max_luns) && !has_wlun_rl)
5183 		return schedule_resp(scp, NULL, errsts_no_connect, 0);
5184 
5185 	sdeb_i = opcode_ind_arr[opcode];	/* fully mapped */
5186 	oip = &opcode_info_arr[sdeb_i];		/* safe if table consistent */
5187 	devip = (struct sdebug_dev_info *)sdp->hostdata;
5188 	if (!devip) {
5189 		devip = devInfoReg(sdp);
5190 		if (NULL == devip)
5191 			return schedule_resp(scp, NULL, errsts_no_connect, 0);
5192 	}
5193 	na = oip->num_attached;
5194 	r_pfp = oip->pfp;
5195 	if (na) {	/* multiple commands with this opcode */
5196 		r_oip = oip;
5197 		if (FF_SA & r_oip->flags) {
5198 			if (F_SA_LOW & oip->flags)
5199 				sa = 0x1f & cmd[1];
5200 			else
5201 				sa = get_unaligned_be16(cmd + 8);
5202 			for (k = 0; k <= na; oip = r_oip->arrp + k++) {
5203 				if (opcode == oip->opcode && sa == oip->sa)
5204 					break;
5205 			}
5206 		} else {   /* since no service action only check opcode */
5207 			for (k = 0; k <= na; oip = r_oip->arrp + k++) {
5208 				if (opcode == oip->opcode)
5209 					break;
5210 			}
5211 		}
5212 		if (k > na) {
5213 			if (F_SA_LOW & r_oip->flags)
5214 				mk_sense_invalid_fld(scp, SDEB_IN_CDB, 1, 4);
5215 			else if (F_SA_HIGH & r_oip->flags)
5216 				mk_sense_invalid_fld(scp, SDEB_IN_CDB, 8, 7);
5217 			else
5218 				mk_sense_invalid_opcode(scp);
5219 			goto check_cond;
5220 		}
5221 	}	/* else (when na==0) we assume the oip is a match */
5222 	flags = oip->flags;
5223 	if (F_INV_OP & flags) {
5224 		mk_sense_invalid_opcode(scp);
5225 		goto check_cond;
5226 	}
5227 	if (has_wlun_rl && !(F_RL_WLUN_OK & flags)) {
5228 		if (debug)
5229 			sdev_printk(KERN_INFO, sdp, "scsi_debug: Opcode: "
5230 				    "0x%x not supported for wlun\n", opcode);
5231 		mk_sense_invalid_opcode(scp);
5232 		goto check_cond;
5233 	}
5234 	if (scsi_debug_strict) {	/* check cdb against mask */
5235 		u8 rem;
5236 		int j;
5237 
5238 		for (k = 1; k < oip->len_mask[0] && k < 16; ++k) {
5239 			rem = ~oip->len_mask[k] & cmd[k];
5240 			if (rem) {
5241 				for (j = 7; j >= 0; --j, rem <<= 1) {
5242 					if (0x80 & rem)
5243 						break;
5244 				}
5245 				mk_sense_invalid_fld(scp, SDEB_IN_CDB, k, j);
5246 				goto check_cond;
5247 			}
5248 		}
5249 	}
5250 	if (!(F_SKIP_UA & flags) &&
5251 	    SDEBUG_NUM_UAS != find_first_bit(devip->uas_bm, SDEBUG_NUM_UAS)) {
5252 		errsts = check_readiness(scp, UAS_ONLY, devip);
5253 		if (errsts)
5254 			goto check_cond;
5255 	}
5256 	if ((F_M_ACCESS & flags) && devip->stopped) {
5257 		mk_sense_buffer(scp, NOT_READY, LOGICAL_UNIT_NOT_READY, 0x2);
5258 		if (debug)
5259 			sdev_printk(KERN_INFO, sdp, "%s reports: Not ready: "
5260 				    "%s\n", my_name, "initializing command "
5261 				    "required");
5262 		errsts = check_condition_result;
5263 		goto fini;
5264 	}
5265 	if (scsi_debug_fake_rw && (F_FAKE_RW & flags))
5266 		goto fini;
5267 	if (scsi_debug_every_nth) {
5268 		if (check_inject(scp))
5269 			return 0;	/* ignore command: make trouble */
5270 	}
5271 	if (oip->pfp)	/* if this command has a resp_* function, call it */
5272 		errsts = oip->pfp(scp, devip);
5273 	else if (r_pfp)	/* if leaf function ptr NULL, try the root's */
5274 		errsts = r_pfp(scp, devip);
5275 
5276 fini:
5277 	return schedule_resp(scp, devip, errsts,
5278 			     ((F_DELAY_OVERR & flags) ? 0 : scsi_debug_delay));
5279 check_cond:
5280 	return schedule_resp(scp, devip, check_condition_result, 0);
5281 }
5282 
5283 static int
5284 sdebug_queuecommand_lock_or_not(struct Scsi_Host *shost, struct scsi_cmnd *cmd)
5285 {
5286 	if (scsi_debug_host_lock) {
5287 		unsigned long iflags;
5288 		int rc;
5289 
5290 		spin_lock_irqsave(shost->host_lock, iflags);
5291 		rc = scsi_debug_queuecommand(cmd);
5292 		spin_unlock_irqrestore(shost->host_lock, iflags);
5293 		return rc;
5294 	} else
5295 		return scsi_debug_queuecommand(cmd);
5296 }
5297 
5298 static struct scsi_host_template sdebug_driver_template = {
5299 	.show_info =		scsi_debug_show_info,
5300 	.write_info =		scsi_debug_write_info,
5301 	.proc_name =		sdebug_proc_name,
5302 	.name =			"SCSI DEBUG",
5303 	.info =			scsi_debug_info,
5304 	.slave_alloc =		scsi_debug_slave_alloc,
5305 	.slave_configure =	scsi_debug_slave_configure,
5306 	.slave_destroy =	scsi_debug_slave_destroy,
5307 	.ioctl =		scsi_debug_ioctl,
5308 	.queuecommand =		sdebug_queuecommand_lock_or_not,
5309 	.change_queue_depth =	sdebug_change_qdepth,
5310 	.eh_abort_handler =	scsi_debug_abort,
5311 	.eh_device_reset_handler = scsi_debug_device_reset,
5312 	.eh_target_reset_handler = scsi_debug_target_reset,
5313 	.eh_bus_reset_handler = scsi_debug_bus_reset,
5314 	.eh_host_reset_handler = scsi_debug_host_reset,
5315 	.can_queue =		SCSI_DEBUG_CANQUEUE,
5316 	.this_id =		7,
5317 	.sg_tablesize =		SCSI_MAX_SG_CHAIN_SEGMENTS,
5318 	.cmd_per_lun =		DEF_CMD_PER_LUN,
5319 	.max_sectors =		-1U,
5320 	.use_clustering = 	DISABLE_CLUSTERING,
5321 	.module =		THIS_MODULE,
5322 	.track_queue_depth =	1,
5323 	.cmd_size =		sizeof(struct sdebug_scmd_extra_t),
5324 };
5325 
5326 static int sdebug_driver_probe(struct device * dev)
5327 {
5328 	int error = 0;
5329 	int opts;
5330 	struct sdebug_host_info *sdbg_host;
5331 	struct Scsi_Host *hpnt;
5332 	int host_prot;
5333 
5334 	sdbg_host = to_sdebug_host(dev);
5335 
5336 	sdebug_driver_template.can_queue = scsi_debug_max_queue;
5337 	if (scsi_debug_clustering)
5338 		sdebug_driver_template.use_clustering = ENABLE_CLUSTERING;
5339 	hpnt = scsi_host_alloc(&sdebug_driver_template, sizeof(sdbg_host));
5340 	if (NULL == hpnt) {
5341 		pr_err("%s: scsi_host_alloc failed\n", __func__);
5342 		error = -ENODEV;
5343 		return error;
5344 	}
5345 
5346         sdbg_host->shost = hpnt;
5347 	*((struct sdebug_host_info **)hpnt->hostdata) = sdbg_host;
5348 	if ((hpnt->this_id >= 0) && (scsi_debug_num_tgts > hpnt->this_id))
5349 		hpnt->max_id = scsi_debug_num_tgts + 1;
5350 	else
5351 		hpnt->max_id = scsi_debug_num_tgts;
5352 	hpnt->max_lun = SAM2_WLUN_REPORT_LUNS;	/* = scsi_debug_max_luns; */
5353 
5354 	host_prot = 0;
5355 
5356 	switch (scsi_debug_dif) {
5357 
5358 	case SD_DIF_TYPE1_PROTECTION:
5359 		host_prot = SHOST_DIF_TYPE1_PROTECTION;
5360 		if (scsi_debug_dix)
5361 			host_prot |= SHOST_DIX_TYPE1_PROTECTION;
5362 		break;
5363 
5364 	case SD_DIF_TYPE2_PROTECTION:
5365 		host_prot = SHOST_DIF_TYPE2_PROTECTION;
5366 		if (scsi_debug_dix)
5367 			host_prot |= SHOST_DIX_TYPE2_PROTECTION;
5368 		break;
5369 
5370 	case SD_DIF_TYPE3_PROTECTION:
5371 		host_prot = SHOST_DIF_TYPE3_PROTECTION;
5372 		if (scsi_debug_dix)
5373 			host_prot |= SHOST_DIX_TYPE3_PROTECTION;
5374 		break;
5375 
5376 	default:
5377 		if (scsi_debug_dix)
5378 			host_prot |= SHOST_DIX_TYPE0_PROTECTION;
5379 		break;
5380 	}
5381 
5382 	scsi_host_set_prot(hpnt, host_prot);
5383 
5384 	printk(KERN_INFO "scsi_debug: host protection%s%s%s%s%s%s%s\n",
5385 	       (host_prot & SHOST_DIF_TYPE1_PROTECTION) ? " DIF1" : "",
5386 	       (host_prot & SHOST_DIF_TYPE2_PROTECTION) ? " DIF2" : "",
5387 	       (host_prot & SHOST_DIF_TYPE3_PROTECTION) ? " DIF3" : "",
5388 	       (host_prot & SHOST_DIX_TYPE0_PROTECTION) ? " DIX0" : "",
5389 	       (host_prot & SHOST_DIX_TYPE1_PROTECTION) ? " DIX1" : "",
5390 	       (host_prot & SHOST_DIX_TYPE2_PROTECTION) ? " DIX2" : "",
5391 	       (host_prot & SHOST_DIX_TYPE3_PROTECTION) ? " DIX3" : "");
5392 
5393 	if (scsi_debug_guard == 1)
5394 		scsi_host_set_guard(hpnt, SHOST_DIX_GUARD_IP);
5395 	else
5396 		scsi_host_set_guard(hpnt, SHOST_DIX_GUARD_CRC);
5397 
5398 	opts = scsi_debug_opts;
5399 	if (SCSI_DEBUG_OPT_RECOVERED_ERR & opts)
5400 		sdebug_any_injecting_opt = true;
5401 	else if (SCSI_DEBUG_OPT_TRANSPORT_ERR & opts)
5402 		sdebug_any_injecting_opt = true;
5403 	else if (SCSI_DEBUG_OPT_DIF_ERR & opts)
5404 		sdebug_any_injecting_opt = true;
5405 	else if (SCSI_DEBUG_OPT_DIX_ERR & opts)
5406 		sdebug_any_injecting_opt = true;
5407 	else if (SCSI_DEBUG_OPT_SHORT_TRANSFER & opts)
5408 		sdebug_any_injecting_opt = true;
5409 
5410         error = scsi_add_host(hpnt, &sdbg_host->dev);
5411         if (error) {
5412                 printk(KERN_ERR "%s: scsi_add_host failed\n", __func__);
5413                 error = -ENODEV;
5414 		scsi_host_put(hpnt);
5415         } else
5416 		scsi_scan_host(hpnt);
5417 
5418 	return error;
5419 }
5420 
5421 static int sdebug_driver_remove(struct device * dev)
5422 {
5423         struct sdebug_host_info *sdbg_host;
5424 	struct sdebug_dev_info *sdbg_devinfo, *tmp;
5425 
5426 	sdbg_host = to_sdebug_host(dev);
5427 
5428 	if (!sdbg_host) {
5429 		printk(KERN_ERR "%s: Unable to locate host info\n",
5430 		       __func__);
5431 		return -ENODEV;
5432 	}
5433 
5434         scsi_remove_host(sdbg_host->shost);
5435 
5436 	list_for_each_entry_safe(sdbg_devinfo, tmp, &sdbg_host->dev_info_list,
5437 				 dev_list) {
5438                 list_del(&sdbg_devinfo->dev_list);
5439                 kfree(sdbg_devinfo);
5440         }
5441 
5442         scsi_host_put(sdbg_host->shost);
5443         return 0;
5444 }
5445 
5446 static int pseudo_lld_bus_match(struct device *dev,
5447 				struct device_driver *dev_driver)
5448 {
5449 	return 1;
5450 }
5451 
5452 static struct bus_type pseudo_lld_bus = {
5453 	.name = "pseudo",
5454 	.match = pseudo_lld_bus_match,
5455 	.probe = sdebug_driver_probe,
5456 	.remove = sdebug_driver_remove,
5457 	.drv_groups = sdebug_drv_groups,
5458 };
5459