xref: /linux/drivers/md/dm-flakey.c (revision a36e9f5cfe9eb3a1dce8769c7058251c42705357)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright (C) 2003 Sistina Software (UK) Limited.
4  * Copyright (C) 2004, 2010-2011 Red Hat, Inc. All rights reserved.
5  *
6  * This file is released under the GPL.
7  */
8 
9 #include <linux/device-mapper.h>
10 
11 #include <linux/module.h>
12 #include <linux/init.h>
13 #include <linux/blkdev.h>
14 #include <linux/bio.h>
15 #include <linux/slab.h>
16 
17 #define DM_MSG_PREFIX "flakey"
18 
19 #define PROBABILITY_BASE	1000000000
20 
21 #define all_corrupt_bio_flags_match(bio, fc)	\
22 	(((bio)->bi_opf & (fc)->corrupt_bio_flags) == (fc)->corrupt_bio_flags)
23 
24 /*
25  * Flakey: Used for testing only, simulates intermittent,
26  * catastrophic device failure.
27  */
28 struct flakey_c {
29 	struct dm_dev *dev;
30 	unsigned long start_time;
31 	sector_t start;
32 	unsigned int up_interval;
33 	unsigned int down_interval;
34 	unsigned long flags;
35 	unsigned int corrupt_bio_byte;
36 	unsigned int corrupt_bio_rw;
37 	unsigned int corrupt_bio_value;
38 	blk_opf_t corrupt_bio_flags;
39 	unsigned int random_read_corrupt;
40 	unsigned int random_write_corrupt;
41 };
42 
43 enum feature_flag_bits {
44 	ERROR_READS,
45 	DROP_WRITES,
46 	ERROR_WRITES
47 };
48 
49 struct per_bio_data {
50 	bool bio_submitted;
51 };
52 
53 static int parse_features(struct dm_arg_set *as, struct flakey_c *fc,
54 			  struct dm_target *ti)
55 {
56 	int r;
57 	unsigned int argc;
58 	const char *arg_name;
59 
60 	static const struct dm_arg _args[] = {
61 		{0, 11, "Invalid number of feature args"},
62 		{1, UINT_MAX, "Invalid corrupt bio byte"},
63 		{0, 255, "Invalid corrupt value to write into bio byte (0-255)"},
64 		{0, UINT_MAX, "Invalid corrupt bio flags mask"},
65 		{0, PROBABILITY_BASE, "Invalid random corrupt argument"},
66 	};
67 
68 	/* No feature arguments supplied. */
69 	if (!as->argc)
70 		return 0;
71 
72 	r = dm_read_arg_group(_args, as, &argc, &ti->error);
73 	if (r)
74 		return r;
75 
76 	while (argc) {
77 		arg_name = dm_shift_arg(as);
78 		argc--;
79 
80 		if (!arg_name) {
81 			ti->error = "Insufficient feature arguments";
82 			return -EINVAL;
83 		}
84 
85 		/*
86 		 * error_reads
87 		 */
88 		if (!strcasecmp(arg_name, "error_reads")) {
89 			if (test_and_set_bit(ERROR_READS, &fc->flags)) {
90 				ti->error = "Feature error_reads duplicated";
91 				return -EINVAL;
92 			}
93 			continue;
94 		}
95 
96 		/*
97 		 * drop_writes
98 		 */
99 		if (!strcasecmp(arg_name, "drop_writes")) {
100 			if (test_and_set_bit(DROP_WRITES, &fc->flags)) {
101 				ti->error = "Feature drop_writes duplicated";
102 				return -EINVAL;
103 			} else if (test_bit(ERROR_WRITES, &fc->flags)) {
104 				ti->error = "Feature drop_writes conflicts with feature error_writes";
105 				return -EINVAL;
106 			}
107 
108 			continue;
109 		}
110 
111 		/*
112 		 * error_writes
113 		 */
114 		if (!strcasecmp(arg_name, "error_writes")) {
115 			if (test_and_set_bit(ERROR_WRITES, &fc->flags)) {
116 				ti->error = "Feature error_writes duplicated";
117 				return -EINVAL;
118 
119 			} else if (test_bit(DROP_WRITES, &fc->flags)) {
120 				ti->error = "Feature error_writes conflicts with feature drop_writes";
121 				return -EINVAL;
122 			}
123 
124 			continue;
125 		}
126 
127 		/*
128 		 * corrupt_bio_byte <Nth_byte> <direction> <value> <bio_flags>
129 		 */
130 		if (!strcasecmp(arg_name, "corrupt_bio_byte")) {
131 			if (!argc) {
132 				ti->error = "Feature corrupt_bio_byte requires parameters";
133 				return -EINVAL;
134 			}
135 
136 			r = dm_read_arg(_args + 1, as, &fc->corrupt_bio_byte, &ti->error);
137 			if (r)
138 				return r;
139 			argc--;
140 
141 			/*
142 			 * Direction r or w?
143 			 */
144 			arg_name = dm_shift_arg(as);
145 			if (arg_name && !strcasecmp(arg_name, "w"))
146 				fc->corrupt_bio_rw = WRITE;
147 			else if (arg_name && !strcasecmp(arg_name, "r"))
148 				fc->corrupt_bio_rw = READ;
149 			else {
150 				ti->error = "Invalid corrupt bio direction (r or w)";
151 				return -EINVAL;
152 			}
153 			argc--;
154 
155 			/*
156 			 * Value of byte (0-255) to write in place of correct one.
157 			 */
158 			r = dm_read_arg(_args + 2, as, &fc->corrupt_bio_value, &ti->error);
159 			if (r)
160 				return r;
161 			argc--;
162 
163 			/*
164 			 * Only corrupt bios with these flags set.
165 			 */
166 			BUILD_BUG_ON(sizeof(fc->corrupt_bio_flags) !=
167 				     sizeof(unsigned int));
168 			r = dm_read_arg(_args + 3, as,
169 				(__force unsigned int *)&fc->corrupt_bio_flags,
170 				&ti->error);
171 			if (r)
172 				return r;
173 			argc--;
174 
175 			continue;
176 		}
177 
178 		if (!strcasecmp(arg_name, "random_read_corrupt")) {
179 			if (!argc) {
180 				ti->error = "Feature random_read_corrupt requires a parameter";
181 				return -EINVAL;
182 			}
183 			r = dm_read_arg(_args + 4, as, &fc->random_read_corrupt, &ti->error);
184 			if (r)
185 				return r;
186 			argc--;
187 
188 			continue;
189 		}
190 
191 		if (!strcasecmp(arg_name, "random_write_corrupt")) {
192 			if (!argc) {
193 				ti->error = "Feature random_write_corrupt requires a parameter";
194 				return -EINVAL;
195 			}
196 			r = dm_read_arg(_args + 4, as, &fc->random_write_corrupt, &ti->error);
197 			if (r)
198 				return r;
199 			argc--;
200 
201 			continue;
202 		}
203 
204 		ti->error = "Unrecognised flakey feature requested";
205 		return -EINVAL;
206 	}
207 
208 	if (test_bit(DROP_WRITES, &fc->flags) && (fc->corrupt_bio_rw == WRITE)) {
209 		ti->error = "drop_writes is incompatible with corrupt_bio_byte with the WRITE flag set";
210 		return -EINVAL;
211 
212 	} else if (test_bit(ERROR_WRITES, &fc->flags) && (fc->corrupt_bio_rw == WRITE)) {
213 		ti->error = "error_writes is incompatible with corrupt_bio_byte with the WRITE flag set";
214 		return -EINVAL;
215 	}
216 
217 	if (!fc->corrupt_bio_byte && !test_bit(ERROR_READS, &fc->flags) &&
218 	    !test_bit(DROP_WRITES, &fc->flags) && !test_bit(ERROR_WRITES, &fc->flags) &&
219 	    !fc->random_read_corrupt && !fc->random_write_corrupt) {
220 		set_bit(ERROR_WRITES, &fc->flags);
221 		set_bit(ERROR_READS, &fc->flags);
222 	}
223 
224 	return 0;
225 }
226 
227 /*
228  * Construct a flakey mapping:
229  * <dev_path> <offset> <up interval> <down interval> [<#feature args> [<arg>]*]
230  *
231  *   Feature args:
232  *     [drop_writes]
233  *     [corrupt_bio_byte <Nth_byte> <direction> <value> <bio_flags>]
234  *
235  *   Nth_byte starts from 1 for the first byte.
236  *   Direction is r for READ or w for WRITE.
237  *   bio_flags is ignored if 0.
238  */
239 static int flakey_ctr(struct dm_target *ti, unsigned int argc, char **argv)
240 {
241 	static const struct dm_arg _args[] = {
242 		{0, UINT_MAX, "Invalid up interval"},
243 		{0, UINT_MAX, "Invalid down interval"},
244 	};
245 
246 	int r;
247 	struct flakey_c *fc;
248 	unsigned long long tmpll;
249 	struct dm_arg_set as;
250 	const char *devname;
251 	char dummy;
252 
253 	as.argc = argc;
254 	as.argv = argv;
255 
256 	if (argc < 4) {
257 		ti->error = "Invalid argument count";
258 		return -EINVAL;
259 	}
260 
261 	fc = kzalloc(sizeof(*fc), GFP_KERNEL);
262 	if (!fc) {
263 		ti->error = "Cannot allocate context";
264 		return -ENOMEM;
265 	}
266 	fc->start_time = jiffies;
267 
268 	devname = dm_shift_arg(&as);
269 
270 	r = -EINVAL;
271 	if (sscanf(dm_shift_arg(&as), "%llu%c", &tmpll, &dummy) != 1 || tmpll != (sector_t)tmpll) {
272 		ti->error = "Invalid device sector";
273 		goto bad;
274 	}
275 	fc->start = tmpll;
276 
277 	r = dm_read_arg(_args, &as, &fc->up_interval, &ti->error);
278 	if (r)
279 		goto bad;
280 
281 	r = dm_read_arg(_args, &as, &fc->down_interval, &ti->error);
282 	if (r)
283 		goto bad;
284 
285 	if (!(fc->up_interval + fc->down_interval)) {
286 		ti->error = "Total (up + down) interval is zero";
287 		r = -EINVAL;
288 		goto bad;
289 	}
290 
291 	if (fc->up_interval + fc->down_interval < fc->up_interval) {
292 		ti->error = "Interval overflow";
293 		r = -EINVAL;
294 		goto bad;
295 	}
296 
297 	r = parse_features(&as, fc, ti);
298 	if (r)
299 		goto bad;
300 
301 	r = dm_get_device(ti, devname, dm_table_get_mode(ti->table), &fc->dev);
302 	if (r) {
303 		ti->error = "Device lookup failed";
304 		goto bad;
305 	}
306 
307 	ti->num_flush_bios = 1;
308 	ti->num_discard_bios = 1;
309 	ti->per_io_data_size = sizeof(struct per_bio_data);
310 	ti->private = fc;
311 	return 0;
312 
313 bad:
314 	kfree(fc);
315 	return r;
316 }
317 
318 static void flakey_dtr(struct dm_target *ti)
319 {
320 	struct flakey_c *fc = ti->private;
321 
322 	dm_put_device(ti, fc->dev);
323 	kfree(fc);
324 }
325 
326 static sector_t flakey_map_sector(struct dm_target *ti, sector_t bi_sector)
327 {
328 	struct flakey_c *fc = ti->private;
329 
330 	return fc->start + dm_target_offset(ti, bi_sector);
331 }
332 
333 static void flakey_map_bio(struct dm_target *ti, struct bio *bio)
334 {
335 	struct flakey_c *fc = ti->private;
336 
337 	bio_set_dev(bio, fc->dev->bdev);
338 	bio->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector);
339 }
340 
341 static void corrupt_bio_common(struct bio *bio, unsigned int corrupt_bio_byte,
342 			       unsigned char corrupt_bio_value)
343 {
344 	struct bvec_iter iter;
345 	struct bio_vec bvec;
346 
347 	/*
348 	 * Overwrite the Nth byte of the bio's data, on whichever page
349 	 * it falls.
350 	 */
351 	bio_for_each_segment(bvec, bio, iter) {
352 		if (bio_iter_len(bio, iter) > corrupt_bio_byte) {
353 			unsigned char *segment = bvec_kmap_local(&bvec);
354 			segment[corrupt_bio_byte] = corrupt_bio_value;
355 			kunmap_local(segment);
356 			DMDEBUG("Corrupting data bio=%p by writing %u to byte %u "
357 				"(rw=%c bi_opf=%u bi_sector=%llu size=%u)\n",
358 				bio, corrupt_bio_value, corrupt_bio_byte,
359 				(bio_data_dir(bio) == WRITE) ? 'w' : 'r', bio->bi_opf,
360 				(unsigned long long)bio->bi_iter.bi_sector,
361 				bio->bi_iter.bi_size);
362 			break;
363 		}
364 		corrupt_bio_byte -= bio_iter_len(bio, iter);
365 	}
366 }
367 
368 static void corrupt_bio_data(struct bio *bio, struct flakey_c *fc)
369 {
370 	unsigned int corrupt_bio_byte = fc->corrupt_bio_byte - 1;
371 
372 	if (!bio_has_data(bio))
373 		return;
374 
375 	corrupt_bio_common(bio, corrupt_bio_byte, fc->corrupt_bio_value);
376 }
377 
378 static void corrupt_bio_random(struct bio *bio)
379 {
380 	unsigned int corrupt_byte;
381 	unsigned char corrupt_value;
382 
383 	if (!bio_has_data(bio))
384 		return;
385 
386 	corrupt_byte = get_random_u32() % bio->bi_iter.bi_size;
387 	corrupt_value = get_random_u8();
388 
389 	corrupt_bio_common(bio, corrupt_byte, corrupt_value);
390 }
391 
392 static void clone_free(struct bio *clone)
393 {
394 	struct folio_iter fi;
395 
396 	if (clone->bi_vcnt > 0) { /* bio_for_each_folio_all crashes with an empty bio */
397 		bio_for_each_folio_all(fi, clone)
398 			folio_put(fi.folio);
399 	}
400 
401 	bio_uninit(clone);
402 	kfree(clone);
403 }
404 
405 static void clone_endio(struct bio *clone)
406 {
407 	struct bio *bio = clone->bi_private;
408 	bio->bi_status = clone->bi_status;
409 	clone_free(clone);
410 	bio_endio(bio);
411 }
412 
413 static struct bio *clone_bio(struct dm_target *ti, struct flakey_c *fc, struct bio *bio)
414 {
415 	struct bio *clone;
416 	unsigned size, remaining_size, nr_iovecs, order;
417 	struct bvec_iter iter = bio->bi_iter;
418 
419 	if (unlikely(bio->bi_iter.bi_size > UIO_MAXIOV << PAGE_SHIFT))
420 		dm_accept_partial_bio(bio, UIO_MAXIOV << PAGE_SHIFT >> SECTOR_SHIFT);
421 
422 	size = bio->bi_iter.bi_size;
423 	nr_iovecs = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
424 
425 	clone = bio_kmalloc(nr_iovecs, GFP_NOIO | __GFP_NORETRY | __GFP_NOWARN);
426 	if (!clone)
427 		return NULL;
428 
429 	bio_init(clone, fc->dev->bdev, bio->bi_inline_vecs, nr_iovecs, bio->bi_opf);
430 
431 	clone->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector);
432 	clone->bi_private = bio;
433 	clone->bi_end_io = clone_endio;
434 
435 	remaining_size = size;
436 
437 	order = MAX_PAGE_ORDER;
438 	while (remaining_size) {
439 		struct page *pages;
440 		unsigned size_to_add, to_copy;
441 		unsigned char *virt;
442 		unsigned remaining_order = __fls((remaining_size + PAGE_SIZE - 1) >> PAGE_SHIFT);
443 		order = min(order, remaining_order);
444 
445 retry_alloc_pages:
446 		pages = alloc_pages(GFP_NOIO | __GFP_NORETRY | __GFP_NOWARN | __GFP_COMP, order);
447 		if (unlikely(!pages)) {
448 			if (order) {
449 				order--;
450 				goto retry_alloc_pages;
451 			}
452 			clone_free(clone);
453 			return NULL;
454 		}
455 		size_to_add = min((unsigned)PAGE_SIZE << order, remaining_size);
456 
457 		virt = page_to_virt(pages);
458 		to_copy = size_to_add;
459 		do {
460 			struct bio_vec bvec = bvec_iter_bvec(bio->bi_io_vec, iter);
461 			unsigned this_step = min(bvec.bv_len, to_copy);
462 			void *map = bvec_kmap_local(&bvec);
463 			memcpy(virt, map, this_step);
464 			kunmap_local(map);
465 
466 			bvec_iter_advance(bio->bi_io_vec, &iter, this_step);
467 			to_copy -= this_step;
468 			virt += this_step;
469 		} while (to_copy);
470 
471 		__bio_add_page(clone, pages, size_to_add, 0);
472 		remaining_size -= size_to_add;
473 	}
474 
475 	return clone;
476 }
477 
478 static int flakey_map(struct dm_target *ti, struct bio *bio)
479 {
480 	struct flakey_c *fc = ti->private;
481 	unsigned int elapsed;
482 	struct per_bio_data *pb = dm_per_bio_data(bio, sizeof(struct per_bio_data));
483 
484 	pb->bio_submitted = false;
485 
486 	if (op_is_zone_mgmt(bio_op(bio)))
487 		goto map_bio;
488 
489 	/* Are we alive ? */
490 	elapsed = (jiffies - fc->start_time) / HZ;
491 	if (elapsed % (fc->up_interval + fc->down_interval) >= fc->up_interval) {
492 		bool corrupt_fixed, corrupt_random;
493 		/*
494 		 * Flag this bio as submitted while down.
495 		 */
496 		pb->bio_submitted = true;
497 
498 		/*
499 		 * Error reads if neither corrupt_bio_byte or drop_writes or error_writes are set.
500 		 * Otherwise, flakey_end_io() will decide if the reads should be modified.
501 		 */
502 		if (bio_data_dir(bio) == READ) {
503 			if (test_bit(ERROR_READS, &fc->flags))
504 				return DM_MAPIO_KILL;
505 			goto map_bio;
506 		}
507 
508 		/*
509 		 * Drop or error writes?
510 		 */
511 		if (test_bit(DROP_WRITES, &fc->flags)) {
512 			bio_endio(bio);
513 			return DM_MAPIO_SUBMITTED;
514 		} else if (test_bit(ERROR_WRITES, &fc->flags)) {
515 			bio_io_error(bio);
516 			return DM_MAPIO_SUBMITTED;
517 		}
518 
519 		/*
520 		 * Corrupt matching writes.
521 		 */
522 		corrupt_fixed = false;
523 		corrupt_random = false;
524 		if (fc->corrupt_bio_byte && fc->corrupt_bio_rw == WRITE) {
525 			if (all_corrupt_bio_flags_match(bio, fc))
526 				corrupt_fixed = true;
527 		}
528 		if (fc->random_write_corrupt) {
529 			u64 rnd = get_random_u64();
530 			u32 rem = do_div(rnd, PROBABILITY_BASE);
531 			if (rem < fc->random_write_corrupt)
532 				corrupt_random = true;
533 		}
534 		if (corrupt_fixed || corrupt_random) {
535 			struct bio *clone = clone_bio(ti, fc, bio);
536 			if (clone) {
537 				if (corrupt_fixed)
538 					corrupt_bio_data(clone, fc);
539 				if (corrupt_random)
540 					corrupt_bio_random(clone);
541 				submit_bio(clone);
542 				return DM_MAPIO_SUBMITTED;
543 			}
544 		}
545 	}
546 
547 map_bio:
548 	flakey_map_bio(ti, bio);
549 
550 	return DM_MAPIO_REMAPPED;
551 }
552 
553 static int flakey_end_io(struct dm_target *ti, struct bio *bio,
554 			 blk_status_t *error)
555 {
556 	struct flakey_c *fc = ti->private;
557 	struct per_bio_data *pb = dm_per_bio_data(bio, sizeof(struct per_bio_data));
558 
559 	if (op_is_zone_mgmt(bio_op(bio)))
560 		return DM_ENDIO_DONE;
561 
562 	if (!*error && pb->bio_submitted && (bio_data_dir(bio) == READ)) {
563 		if (fc->corrupt_bio_byte) {
564 			if ((fc->corrupt_bio_rw == READ) &&
565 			    all_corrupt_bio_flags_match(bio, fc)) {
566 				/*
567 				 * Corrupt successful matching READs while in down state.
568 				 */
569 				corrupt_bio_data(bio, fc);
570 			}
571 		}
572 		if (fc->random_read_corrupt) {
573 			u64 rnd = get_random_u64();
574 			u32 rem = do_div(rnd, PROBABILITY_BASE);
575 			if (rem < fc->random_read_corrupt)
576 				corrupt_bio_random(bio);
577 		}
578 		if (test_bit(ERROR_READS, &fc->flags)) {
579 			/*
580 			 * Error read during the down_interval if drop_writes
581 			 * and error_writes were not configured.
582 			 */
583 			*error = BLK_STS_IOERR;
584 		}
585 	}
586 
587 	return DM_ENDIO_DONE;
588 }
589 
590 static void flakey_status(struct dm_target *ti, status_type_t type,
591 			  unsigned int status_flags, char *result, unsigned int maxlen)
592 {
593 	unsigned int sz = 0;
594 	struct flakey_c *fc = ti->private;
595 	unsigned int error_reads, drop_writes, error_writes;
596 
597 	switch (type) {
598 	case STATUSTYPE_INFO:
599 		result[0] = '\0';
600 		break;
601 
602 	case STATUSTYPE_TABLE:
603 		DMEMIT("%s %llu %u %u", fc->dev->name,
604 		       (unsigned long long)fc->start, fc->up_interval,
605 		       fc->down_interval);
606 
607 		error_reads = test_bit(ERROR_READS, &fc->flags);
608 		drop_writes = test_bit(DROP_WRITES, &fc->flags);
609 		error_writes = test_bit(ERROR_WRITES, &fc->flags);
610 		DMEMIT(" %u", error_reads + drop_writes + error_writes +
611 			(fc->corrupt_bio_byte > 0) * 5 +
612 			(fc->random_read_corrupt > 0) * 2 +
613 			(fc->random_write_corrupt > 0) * 2);
614 
615 		if (error_reads)
616 			DMEMIT(" error_reads");
617 		if (drop_writes)
618 			DMEMIT(" drop_writes");
619 		else if (error_writes)
620 			DMEMIT(" error_writes");
621 
622 		if (fc->corrupt_bio_byte)
623 			DMEMIT(" corrupt_bio_byte %u %c %u %u",
624 			       fc->corrupt_bio_byte,
625 			       (fc->corrupt_bio_rw == WRITE) ? 'w' : 'r',
626 			       fc->corrupt_bio_value, fc->corrupt_bio_flags);
627 
628 		if (fc->random_read_corrupt > 0)
629 			DMEMIT(" random_read_corrupt %u", fc->random_read_corrupt);
630 		if (fc->random_write_corrupt > 0)
631 			DMEMIT(" random_write_corrupt %u", fc->random_write_corrupt);
632 
633 		break;
634 
635 	case STATUSTYPE_IMA:
636 		result[0] = '\0';
637 		break;
638 	}
639 }
640 
641 static int flakey_prepare_ioctl(struct dm_target *ti, struct block_device **bdev)
642 {
643 	struct flakey_c *fc = ti->private;
644 
645 	*bdev = fc->dev->bdev;
646 
647 	/*
648 	 * Only pass ioctls through if the device sizes match exactly.
649 	 */
650 	if (fc->start || ti->len != bdev_nr_sectors((*bdev)))
651 		return 1;
652 	return 0;
653 }
654 
655 #ifdef CONFIG_BLK_DEV_ZONED
656 static int flakey_report_zones(struct dm_target *ti,
657 		struct dm_report_zones_args *args, unsigned int nr_zones)
658 {
659 	struct flakey_c *fc = ti->private;
660 
661 	return dm_report_zones(fc->dev->bdev, fc->start,
662 			       flakey_map_sector(ti, args->next_sector),
663 			       args, nr_zones);
664 }
665 #else
666 #define flakey_report_zones NULL
667 #endif
668 
669 static int flakey_iterate_devices(struct dm_target *ti, iterate_devices_callout_fn fn, void *data)
670 {
671 	struct flakey_c *fc = ti->private;
672 
673 	return fn(ti, fc->dev, fc->start, ti->len, data);
674 }
675 
676 static struct target_type flakey_target = {
677 	.name   = "flakey",
678 	.version = {1, 5, 0},
679 	.features = DM_TARGET_ZONED_HM | DM_TARGET_PASSES_CRYPTO,
680 	.report_zones = flakey_report_zones,
681 	.module = THIS_MODULE,
682 	.ctr    = flakey_ctr,
683 	.dtr    = flakey_dtr,
684 	.map    = flakey_map,
685 	.end_io = flakey_end_io,
686 	.status = flakey_status,
687 	.prepare_ioctl = flakey_prepare_ioctl,
688 	.iterate_devices = flakey_iterate_devices,
689 };
690 module_dm(flakey);
691 
692 MODULE_DESCRIPTION(DM_NAME " flakey target");
693 MODULE_AUTHOR("Joe Thornber <dm-devel@lists.linux.dev>");
694 MODULE_LICENSE("GPL");
695