xref: /titanic_50/usr/src/uts/common/fs/zfs/zvol.c (revision 700c902c445eb3882848aaddc19d13638818cfd6)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2006 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 #pragma ident	"%Z%%M%	%I%	%E% SMI"
27 
28 /*
29  * ZFS volume emulation driver.
30  *
31  * Makes a DMU object look like a volume of arbitrary size, up to 2^64 bytes.
32  * Volumes are accessed through the symbolic links named:
33  *
34  * /dev/zvol/dsk/<pool_name>/<dataset_name>
35  * /dev/zvol/rdsk/<pool_name>/<dataset_name>
36  *
37  * These links are created by the ZFS-specific devfsadm link generator.
38  * Volumes are persistent through reboot.  No user command needs to be
39  * run before opening and using a device.
40  */
41 
42 #include <sys/types.h>
43 #include <sys/param.h>
44 #include <sys/errno.h>
45 #include <sys/aio_req.h>
46 #include <sys/uio.h>
47 #include <sys/buf.h>
48 #include <sys/modctl.h>
49 #include <sys/open.h>
50 #include <sys/kmem.h>
51 #include <sys/conf.h>
52 #include <sys/cmn_err.h>
53 #include <sys/stat.h>
54 #include <sys/zap.h>
55 #include <sys/spa.h>
56 #include <sys/zio.h>
57 #include <sys/dsl_prop.h>
58 #include <sys/dkio.h>
59 #include <sys/efi_partition.h>
60 #include <sys/byteorder.h>
61 #include <sys/pathname.h>
62 #include <sys/ddi.h>
63 #include <sys/sunddi.h>
64 #include <sys/crc32.h>
65 #include <sys/dirent.h>
66 #include <sys/policy.h>
67 #include <sys/fs/zfs.h>
68 #include <sys/zfs_ioctl.h>
69 #include <sys/mkdev.h>
70 #include <sys/zil.h>
71 #include <sys/refcount.h>
72 
73 #include "zfs_namecheck.h"
74 
75 #define	ZVOL_OBJ		1ULL
76 #define	ZVOL_ZAP_OBJ		2ULL
77 
78 static void *zvol_state;
79 
80 /*
81  * This lock protects the zvol_state structure from being modified
82  * while it's being used, e.g. an open that comes in before a create
83  * finishes.  It also protects temporary opens of the dataset so that,
84  * e.g., an open doesn't get a spurious EBUSY.
85  */
86 static kmutex_t zvol_state_lock;
87 static uint32_t zvol_minors;
88 
89 /*
90  * The in-core state of each volume.
91  */
92 typedef struct zvol_state {
93 	char		zv_name[MAXPATHLEN]; /* pool/dd name */
94 	uint64_t	zv_volsize;	/* amount of space we advertise */
95 	minor_t		zv_minor;	/* minor number */
96 	uint8_t		zv_min_bs;	/* minimum addressable block shift */
97 	uint8_t		zv_readonly;	/* hard readonly; like write-protect */
98 	objset_t	*zv_objset;	/* objset handle */
99 	uint32_t	zv_mode;	/* DS_MODE_* flags at open time */
100 	uint32_t	zv_open_count[OTYPCNT];	/* open counts */
101 	uint32_t	zv_total_opens;	/* total open count */
102 	zilog_t		*zv_zilog;	/* ZIL handle */
103 	uint64_t	zv_txg_assign;	/* txg to assign during ZIL replay */
104 	krwlock_t	zv_dslock;	/* dmu_sync() rwlock */
105 } zvol_state_t;
106 
107 static void
108 zvol_size_changed(zvol_state_t *zv, dev_t dev)
109 {
110 	dev = makedevice(getmajor(dev), zv->zv_minor);
111 
112 	VERIFY(ddi_prop_update_int64(dev, zfs_dip,
113 	    "Size", zv->zv_volsize) == DDI_SUCCESS);
114 	VERIFY(ddi_prop_update_int64(dev, zfs_dip,
115 	    "Nblocks", lbtodb(zv->zv_volsize)) == DDI_SUCCESS);
116 }
117 
118 int
119 zvol_check_volsize(zfs_cmd_t *zc, uint64_t blocksize)
120 {
121 	if (zc->zc_volsize == 0)
122 		return (EINVAL);
123 
124 	if (zc->zc_volsize % blocksize != 0)
125 		return (EINVAL);
126 
127 #ifdef _ILP32
128 	if (zc->zc_volsize - 1 > SPEC_MAXOFFSET_T)
129 		return (EOVERFLOW);
130 #endif
131 	return (0);
132 }
133 
134 int
135 zvol_check_volblocksize(zfs_cmd_t *zc)
136 {
137 	if (zc->zc_volblocksize < SPA_MINBLOCKSIZE ||
138 	    zc->zc_volblocksize > SPA_MAXBLOCKSIZE ||
139 	    !ISP2(zc->zc_volblocksize))
140 		return (EDOM);
141 
142 	return (0);
143 }
144 
145 static void
146 zvol_readonly_changed_cb(void *arg, uint64_t newval)
147 {
148 	zvol_state_t *zv = arg;
149 
150 	zv->zv_readonly = (uint8_t)newval;
151 }
152 
153 int
154 zvol_get_stats(zfs_cmd_t *zc, objset_t *os)
155 {
156 	int error;
157 	dmu_object_info_t doi;
158 
159 	error = zap_lookup(os, ZVOL_ZAP_OBJ, "size", 8, 1, &zc->zc_volsize);
160 
161 	if (error)
162 		return (error);
163 
164 	error = dmu_object_info(os, ZVOL_OBJ, &doi);
165 
166 	if (error == 0)
167 		zc->zc_volblocksize = doi.doi_data_block_size;
168 
169 	return (error);
170 }
171 
172 /*
173  * Find a free minor number.
174  */
175 static minor_t
176 zvol_minor_alloc(void)
177 {
178 	minor_t minor;
179 
180 	ASSERT(MUTEX_HELD(&zvol_state_lock));
181 
182 	for (minor = 1; minor <= ZVOL_MAX_MINOR; minor++)
183 		if (ddi_get_soft_state(zvol_state, minor) == NULL)
184 			return (minor);
185 
186 	return (0);
187 }
188 
189 static zvol_state_t *
190 zvol_minor_lookup(char *name)
191 {
192 	minor_t minor;
193 	zvol_state_t *zv;
194 
195 	ASSERT(MUTEX_HELD(&zvol_state_lock));
196 
197 	for (minor = 1; minor <= ZVOL_MAX_MINOR; minor++) {
198 		zv = ddi_get_soft_state(zvol_state, minor);
199 		if (zv == NULL)
200 			continue;
201 		if (strcmp(zv->zv_name, name) == 0)
202 			break;
203 	}
204 
205 	return (zv);
206 }
207 
208 void
209 zvol_create_cb(objset_t *os, void *arg, dmu_tx_t *tx)
210 {
211 	zfs_cmd_t *zc = arg;
212 	int error;
213 
214 	error = dmu_object_claim(os, ZVOL_OBJ, DMU_OT_ZVOL, zc->zc_volblocksize,
215 	    DMU_OT_NONE, 0, tx);
216 	ASSERT(error == 0);
217 
218 	error = zap_create_claim(os, ZVOL_ZAP_OBJ, DMU_OT_ZVOL_PROP,
219 	    DMU_OT_NONE, 0, tx);
220 	ASSERT(error == 0);
221 
222 	error = zap_update(os, ZVOL_ZAP_OBJ, "size", 8, 1, &zc->zc_volsize, tx);
223 	ASSERT(error == 0);
224 }
225 
226 /*
227  * Replay a TX_WRITE ZIL transaction that didn't get committed
228  * after a system failure
229  */
230 static int
231 zvol_replay_write(zvol_state_t *zv, lr_write_t *lr, boolean_t byteswap)
232 {
233 	objset_t *os = zv->zv_objset;
234 	char *data = (char *)(lr + 1);	/* data follows lr_write_t */
235 	uint64_t off = lr->lr_offset;
236 	uint64_t len = lr->lr_length;
237 	dmu_tx_t *tx;
238 	int error;
239 
240 	if (byteswap)
241 		byteswap_uint64_array(lr, sizeof (*lr));
242 
243 	tx = dmu_tx_create(os);
244 	dmu_tx_hold_write(tx, ZVOL_OBJ, off, len);
245 	error = dmu_tx_assign(tx, zv->zv_txg_assign);
246 	if (error) {
247 		dmu_tx_abort(tx);
248 	} else {
249 		dmu_write(os, ZVOL_OBJ, off, len, data, tx);
250 		dmu_tx_commit(tx);
251 	}
252 
253 	return (error);
254 }
255 
256 /* ARGSUSED */
257 static int
258 zvol_replay_err(zvol_state_t *zv, lr_t *lr, boolean_t byteswap)
259 {
260 	return (ENOTSUP);
261 }
262 
263 /*
264  * Callback vectors for replaying records.
265  * Only TX_WRITE is needed for zvol.
266  */
267 zil_replay_func_t *zvol_replay_vector[TX_MAX_TYPE] = {
268 	zvol_replay_err,	/* 0 no such transaction type */
269 	zvol_replay_err,	/* TX_CREATE */
270 	zvol_replay_err,	/* TX_MKDIR */
271 	zvol_replay_err,	/* TX_MKXATTR */
272 	zvol_replay_err,	/* TX_SYMLINK */
273 	zvol_replay_err,	/* TX_REMOVE */
274 	zvol_replay_err,	/* TX_RMDIR */
275 	zvol_replay_err,	/* TX_LINK */
276 	zvol_replay_err,	/* TX_RENAME */
277 	zvol_replay_write,	/* TX_WRITE */
278 	zvol_replay_err,	/* TX_TRUNCATE */
279 	zvol_replay_err,	/* TX_SETATTR */
280 	zvol_replay_err,	/* TX_ACL */
281 };
282 
283 /*
284  * Create a minor node for the specified volume.
285  */
286 int
287 zvol_create_minor(zfs_cmd_t *zc)
288 {
289 	char *name = zc->zc_name;
290 	dev_t dev = zc->zc_dev;
291 	zvol_state_t *zv;
292 	objset_t *os;
293 	uint64_t volsize;
294 	minor_t minor = 0;
295 	struct pathname linkpath;
296 	int ds_mode = DS_MODE_PRIMARY;
297 	vnode_t *vp = NULL;
298 	char *devpath;
299 	size_t devpathlen = strlen(ZVOL_FULL_DEV_DIR) + 1 + strlen(name) + 1;
300 	char chrbuf[30], blkbuf[30];
301 	int error;
302 
303 	mutex_enter(&zvol_state_lock);
304 
305 	if ((zv = zvol_minor_lookup(name)) != NULL) {
306 		mutex_exit(&zvol_state_lock);
307 		return (EEXIST);
308 	}
309 
310 	if (strchr(name, '@') != 0)
311 		ds_mode |= DS_MODE_READONLY;
312 
313 	error = dmu_objset_open(name, DMU_OST_ZVOL, ds_mode, &os);
314 
315 	if (error) {
316 		mutex_exit(&zvol_state_lock);
317 		return (error);
318 	}
319 
320 	error = zap_lookup(os, ZVOL_ZAP_OBJ, "size", 8, 1, &volsize);
321 
322 	if (error) {
323 		dmu_objset_close(os);
324 		mutex_exit(&zvol_state_lock);
325 		return (error);
326 	}
327 
328 	/*
329 	 * If there's an existing /dev/zvol symlink, try to use the
330 	 * same minor number we used last time.
331 	 */
332 	devpath = kmem_alloc(devpathlen, KM_SLEEP);
333 
334 	(void) sprintf(devpath, "%s/%s", ZVOL_FULL_DEV_DIR, name);
335 
336 	error = lookupname(devpath, UIO_SYSSPACE, NO_FOLLOW, NULL, &vp);
337 
338 	kmem_free(devpath, devpathlen);
339 
340 	if (error == 0 && vp->v_type != VLNK)
341 		error = EINVAL;
342 
343 	if (error == 0) {
344 		pn_alloc(&linkpath);
345 		error = pn_getsymlink(vp, &linkpath, kcred);
346 		if (error == 0) {
347 			char *ms = strstr(linkpath.pn_path, ZVOL_PSEUDO_DEV);
348 			if (ms != NULL) {
349 				ms += strlen(ZVOL_PSEUDO_DEV);
350 				minor = stoi(&ms);
351 			}
352 		}
353 		pn_free(&linkpath);
354 	}
355 
356 	if (vp != NULL)
357 		VN_RELE(vp);
358 
359 	/*
360 	 * If we found a minor but it's already in use, we must pick a new one.
361 	 */
362 	if (minor != 0 && ddi_get_soft_state(zvol_state, minor) != NULL)
363 		minor = 0;
364 
365 	if (minor == 0)
366 		minor = zvol_minor_alloc();
367 
368 	if (minor == 0) {
369 		dmu_objset_close(os);
370 		mutex_exit(&zvol_state_lock);
371 		return (ENXIO);
372 	}
373 
374 	if (ddi_soft_state_zalloc(zvol_state, minor) != DDI_SUCCESS) {
375 		dmu_objset_close(os);
376 		mutex_exit(&zvol_state_lock);
377 		return (EAGAIN);
378 	}
379 
380 	(void) ddi_prop_update_string(minor, zfs_dip, ZVOL_PROP_NAME, name);
381 
382 	(void) sprintf(chrbuf, "%uc,raw", minor);
383 
384 	if (ddi_create_minor_node(zfs_dip, chrbuf, S_IFCHR,
385 	    minor, DDI_PSEUDO, 0) == DDI_FAILURE) {
386 		ddi_soft_state_free(zvol_state, minor);
387 		dmu_objset_close(os);
388 		mutex_exit(&zvol_state_lock);
389 		return (EAGAIN);
390 	}
391 
392 	(void) sprintf(blkbuf, "%uc", minor);
393 
394 	if (ddi_create_minor_node(zfs_dip, blkbuf, S_IFBLK,
395 	    minor, DDI_PSEUDO, 0) == DDI_FAILURE) {
396 		ddi_remove_minor_node(zfs_dip, chrbuf);
397 		ddi_soft_state_free(zvol_state, minor);
398 		dmu_objset_close(os);
399 		mutex_exit(&zvol_state_lock);
400 		return (EAGAIN);
401 	}
402 
403 	zv = ddi_get_soft_state(zvol_state, minor);
404 
405 	(void) strcpy(zv->zv_name, name);
406 	zv->zv_min_bs = DEV_BSHIFT;
407 	zv->zv_minor = minor;
408 	zv->zv_volsize = volsize;
409 	zv->zv_objset = os;
410 	zv->zv_mode = ds_mode;
411 	zv->zv_zilog = zil_open(os, NULL);
412 
413 	rw_init(&zv->zv_dslock, NULL, RW_DEFAULT, NULL);
414 
415 	zil_replay(os, zv, &zv->zv_txg_assign, zvol_replay_vector, NULL);
416 
417 	zvol_size_changed(zv, dev);
418 
419 	/* XXX this should handle the possible i/o error */
420 	VERIFY(dsl_prop_register(dmu_objset_ds(zv->zv_objset),
421 	    "readonly", zvol_readonly_changed_cb, zv) == 0);
422 
423 	zvol_minors++;
424 
425 	mutex_exit(&zvol_state_lock);
426 
427 	return (0);
428 }
429 
430 /*
431  * Remove minor node for the specified volume.
432  */
433 int
434 zvol_remove_minor(zfs_cmd_t *zc)
435 {
436 	zvol_state_t *zv;
437 	char namebuf[30];
438 
439 	mutex_enter(&zvol_state_lock);
440 
441 	if ((zv = zvol_minor_lookup(zc->zc_name)) == NULL) {
442 		mutex_exit(&zvol_state_lock);
443 		return (ENXIO);
444 	}
445 
446 	if (zv->zv_total_opens != 0) {
447 		mutex_exit(&zvol_state_lock);
448 		return (EBUSY);
449 	}
450 
451 	(void) sprintf(namebuf, "%uc,raw", zv->zv_minor);
452 	ddi_remove_minor_node(zfs_dip, namebuf);
453 
454 	(void) sprintf(namebuf, "%uc", zv->zv_minor);
455 	ddi_remove_minor_node(zfs_dip, namebuf);
456 
457 	VERIFY(dsl_prop_unregister(dmu_objset_ds(zv->zv_objset),
458 	    "readonly", zvol_readonly_changed_cb, zv) == 0);
459 
460 	zil_close(zv->zv_zilog);
461 	zv->zv_zilog = NULL;
462 	dmu_objset_close(zv->zv_objset);
463 	zv->zv_objset = NULL;
464 
465 	ddi_soft_state_free(zvol_state, zv->zv_minor);
466 
467 	zvol_minors--;
468 
469 	mutex_exit(&zvol_state_lock);
470 
471 	return (0);
472 }
473 
474 int
475 zvol_set_volsize(zfs_cmd_t *zc)
476 {
477 	zvol_state_t *zv;
478 	dev_t dev = zc->zc_dev;
479 	dmu_tx_t *tx;
480 	int error;
481 	dmu_object_info_t doi;
482 
483 	mutex_enter(&zvol_state_lock);
484 
485 	if ((zv = zvol_minor_lookup(zc->zc_name)) == NULL) {
486 		mutex_exit(&zvol_state_lock);
487 		return (ENXIO);
488 	}
489 
490 	if ((error = dmu_object_info(zv->zv_objset, ZVOL_OBJ, &doi)) != 0 ||
491 	    (error = zvol_check_volsize(zc, doi.doi_data_block_size)) != 0) {
492 		mutex_exit(&zvol_state_lock);
493 		return (error);
494 	}
495 
496 	if (zv->zv_readonly || (zv->zv_mode & DS_MODE_READONLY)) {
497 		mutex_exit(&zvol_state_lock);
498 		return (EROFS);
499 	}
500 
501 	tx = dmu_tx_create(zv->zv_objset);
502 	dmu_tx_hold_zap(tx, ZVOL_ZAP_OBJ, TRUE, NULL);
503 	dmu_tx_hold_free(tx, ZVOL_OBJ, zc->zc_volsize, DMU_OBJECT_END);
504 	error = dmu_tx_assign(tx, TXG_WAIT);
505 	if (error) {
506 		dmu_tx_abort(tx);
507 		mutex_exit(&zvol_state_lock);
508 		return (error);
509 	}
510 
511 	error = zap_update(zv->zv_objset, ZVOL_ZAP_OBJ, "size", 8, 1,
512 	    &zc->zc_volsize, tx);
513 	if (error == 0) {
514 		error = dmu_free_range(zv->zv_objset, ZVOL_OBJ, zc->zc_volsize,
515 		    DMU_OBJECT_END, tx);
516 	}
517 
518 	dmu_tx_commit(tx);
519 
520 	if (error == 0) {
521 		zv->zv_volsize = zc->zc_volsize;
522 		zvol_size_changed(zv, dev);
523 	}
524 
525 	mutex_exit(&zvol_state_lock);
526 
527 	return (error);
528 }
529 
530 int
531 zvol_set_volblocksize(zfs_cmd_t *zc)
532 {
533 	zvol_state_t *zv;
534 	dmu_tx_t *tx;
535 	int error;
536 
537 	mutex_enter(&zvol_state_lock);
538 
539 	if ((zv = zvol_minor_lookup(zc->zc_name)) == NULL) {
540 		mutex_exit(&zvol_state_lock);
541 		return (ENXIO);
542 	}
543 
544 	if (zv->zv_readonly || (zv->zv_mode & DS_MODE_READONLY)) {
545 		mutex_exit(&zvol_state_lock);
546 		return (EROFS);
547 	}
548 
549 	tx = dmu_tx_create(zv->zv_objset);
550 	dmu_tx_hold_bonus(tx, ZVOL_OBJ);
551 	error = dmu_tx_assign(tx, TXG_WAIT);
552 	if (error) {
553 		dmu_tx_abort(tx);
554 	} else {
555 		error = dmu_object_set_blocksize(zv->zv_objset, ZVOL_OBJ,
556 		    zc->zc_volblocksize, 0, tx);
557 		if (error == ENOTSUP)
558 			error = EBUSY;
559 		dmu_tx_commit(tx);
560 	}
561 
562 	mutex_exit(&zvol_state_lock);
563 
564 	return (error);
565 }
566 
567 /*ARGSUSED*/
568 int
569 zvol_open(dev_t *devp, int flag, int otyp, cred_t *cr)
570 {
571 	minor_t minor = getminor(*devp);
572 	zvol_state_t *zv;
573 
574 	if (minor == 0)			/* This is the control device */
575 		return (0);
576 
577 	mutex_enter(&zvol_state_lock);
578 
579 	zv = ddi_get_soft_state(zvol_state, minor);
580 	if (zv == NULL) {
581 		mutex_exit(&zvol_state_lock);
582 		return (ENXIO);
583 	}
584 
585 	ASSERT(zv->zv_objset != NULL);
586 
587 	if ((flag & FWRITE) &&
588 	    (zv->zv_readonly || (zv->zv_mode & DS_MODE_READONLY))) {
589 		mutex_exit(&zvol_state_lock);
590 		return (EROFS);
591 	}
592 
593 	if (zv->zv_open_count[otyp] == 0 || otyp == OTYP_LYR) {
594 		zv->zv_open_count[otyp]++;
595 		zv->zv_total_opens++;
596 	}
597 
598 	mutex_exit(&zvol_state_lock);
599 
600 	return (0);
601 }
602 
603 /*ARGSUSED*/
604 int
605 zvol_close(dev_t dev, int flag, int otyp, cred_t *cr)
606 {
607 	minor_t minor = getminor(dev);
608 	zvol_state_t *zv;
609 
610 	if (minor == 0)		/* This is the control device */
611 		return (0);
612 
613 	mutex_enter(&zvol_state_lock);
614 
615 	zv = ddi_get_soft_state(zvol_state, minor);
616 	if (zv == NULL) {
617 		mutex_exit(&zvol_state_lock);
618 		return (ENXIO);
619 	}
620 
621 	/*
622 	 * The next statement is a workaround for the following DDI bug:
623 	 * 6343604 specfs race: multiple "last-close" of the same device
624 	 */
625 	if (zv->zv_total_opens == 0) {
626 		mutex_exit(&zvol_state_lock);
627 		return (0);
628 	}
629 
630 	/*
631 	 * If the open count is zero, this is a spurious close.
632 	 * That indicates a bug in the kernel / DDI framework.
633 	 */
634 	ASSERT(zv->zv_open_count[otyp] != 0);
635 	ASSERT(zv->zv_total_opens != 0);
636 
637 	/*
638 	 * You may get multiple opens, but only one close.
639 	 */
640 	zv->zv_open_count[otyp]--;
641 	zv->zv_total_opens--;
642 
643 	mutex_exit(&zvol_state_lock);
644 
645 	return (0);
646 }
647 
648 /*
649  * Create and return an immediate write ZIL transaction.
650  */
651 itx_t *
652 zvol_immediate_itx(offset_t off, ssize_t len, char *addr)
653 {
654 	itx_t *itx;
655 	lr_write_t *lr;
656 
657 	itx = zil_itx_create(TX_WRITE, sizeof (*lr) + len);
658 	lr = (lr_write_t *)&itx->itx_lr;
659 	lr->lr_foid = ZVOL_OBJ;
660 	lr->lr_offset = off;
661 	lr->lr_length = len;
662 	lr->lr_blkoff = 0;
663 	BP_ZERO(&lr->lr_blkptr);
664 	bcopy(addr, (char *)itx + offsetof(itx_t, itx_lr) +
665 	    sizeof (*lr), len);
666 	itx->itx_wr_state = WR_COPIED;
667 	return (itx);
668 }
669 
670 /*
671  * zvol_log_write() handles synchronous writes using TX_WRITE ZIL transactions.
672  *
673  * We store data in the log buffers if it's small enough.
674  * Otherwise we flush the data out via dmu_sync().
675  */
676 ssize_t zvol_immediate_write_sz = 65536;
677 
678 int
679 zvol_log_write(zvol_state_t *zv, dmu_tx_t *tx, offset_t off, ssize_t len,
680     char *addr)
681 {
682 	dmu_object_info_t doi;
683 	ssize_t nbytes;
684 	itx_t *itx;
685 	lr_write_t *lr;
686 	objset_t *os;
687 	dmu_buf_t *db;
688 	uint64_t txg;
689 	uint64_t boff;
690 	int error;
691 	uint32_t blocksize;
692 
693 	/* handle common case */
694 	if (len <= zvol_immediate_write_sz) {
695 		itx = zvol_immediate_itx(off, len, addr);
696 		(void) zil_itx_assign(zv->zv_zilog, itx, tx);
697 		return (0);
698 	}
699 
700 	txg = dmu_tx_get_txg(tx);
701 	os = zv->zv_objset;
702 
703 	/*
704 	 * We need to dmu_sync() each block in the range.
705 	 * For this we need the blocksize.
706 	 */
707 	error = dmu_object_info(os, ZVOL_OBJ, &doi);
708 	if (error)
709 		return (error);
710 	blocksize = doi.doi_data_block_size;
711 
712 	/*
713 	 * We need to immediate write or dmu_sync() each block in the range.
714 	 */
715 	while (len) {
716 		nbytes = MIN(len, blocksize - P2PHASE(off, blocksize));
717 		if (nbytes <= zvol_immediate_write_sz) {
718 			itx = zvol_immediate_itx(off, nbytes, addr);
719 		} else {
720 			boff =  P2ALIGN_TYPED(off, blocksize, uint64_t);
721 			itx = zil_itx_create(TX_WRITE, sizeof (*lr));
722 			lr = (lr_write_t *)&itx->itx_lr;
723 			lr->lr_foid = ZVOL_OBJ;
724 			lr->lr_offset = off;
725 			lr->lr_length = nbytes;
726 			lr->lr_blkoff = off - boff;
727 			BP_ZERO(&lr->lr_blkptr);
728 
729 			/* XXX - we should do these IOs in parallel */
730 			VERIFY(0 == dmu_buf_hold(os, ZVOL_OBJ, boff,
731 			    FTAG, &db));
732 			ASSERT(boff == db->db_offset);
733 			error = dmu_sync(NULL, db, &lr->lr_blkptr,
734 			    txg, NULL, NULL);
735 			dmu_buf_rele(db, FTAG);
736 			if (error) {
737 				kmem_free(itx, offsetof(itx_t, itx_lr));
738 				return (error);
739 			}
740 			itx->itx_wr_state = WR_COPIED;
741 		}
742 		(void) zil_itx_assign(zv->zv_zilog, itx, tx);
743 		len -= nbytes;
744 		off += nbytes;
745 	}
746 	return (0);
747 }
748 
749 int
750 zvol_strategy(buf_t *bp)
751 {
752 	zvol_state_t *zv = ddi_get_soft_state(zvol_state, getminor(bp->b_edev));
753 	uint64_t off, volsize;
754 	size_t size, resid;
755 	char *addr;
756 	objset_t *os;
757 	int error = 0;
758 	int sync;
759 	int reading;
760 	int txg_sync_needed = B_FALSE;
761 
762 	if (zv == NULL) {
763 		bioerror(bp, ENXIO);
764 		biodone(bp);
765 		return (0);
766 	}
767 
768 	if (getminor(bp->b_edev) == 0) {
769 		bioerror(bp, EINVAL);
770 		biodone(bp);
771 		return (0);
772 	}
773 
774 	if (zv->zv_readonly && !(bp->b_flags & B_READ)) {
775 		bioerror(bp, EROFS);
776 		biodone(bp);
777 		return (0);
778 	}
779 
780 	off = ldbtob(bp->b_blkno);
781 	volsize = zv->zv_volsize;
782 
783 	os = zv->zv_objset;
784 	ASSERT(os != NULL);
785 	sync = !(bp->b_flags & B_ASYNC) && !(zil_disable);
786 
787 	bp_mapin(bp);
788 	addr = bp->b_un.b_addr;
789 	resid = bp->b_bcount;
790 
791 	/*
792 	 * There must be no buffer changes when doing a dmu_sync() because
793 	 * we can't change the data whilst calculating the checksum.
794 	 * A better approach than a per zvol rwlock would be to lock ranges.
795 	 */
796 	reading = bp->b_flags & B_READ;
797 	if (reading || resid <= zvol_immediate_write_sz)
798 		rw_enter(&zv->zv_dslock, RW_READER);
799 	else
800 		rw_enter(&zv->zv_dslock, RW_WRITER);
801 
802 	while (resid != 0 && off < volsize) {
803 
804 		size = MIN(resid, 1UL << 20);	/* cap at 1MB per tx */
805 
806 		if (size > volsize - off)	/* don't write past the end */
807 			size = volsize - off;
808 
809 		if (reading) {
810 			error = dmu_read(os, ZVOL_OBJ, off, size, addr);
811 		} else {
812 			dmu_tx_t *tx = dmu_tx_create(os);
813 			dmu_tx_hold_write(tx, ZVOL_OBJ, off, size);
814 			error = dmu_tx_assign(tx, TXG_WAIT);
815 			if (error) {
816 				dmu_tx_abort(tx);
817 			} else {
818 				dmu_write(os, ZVOL_OBJ, off, size, addr, tx);
819 				if (sync) {
820 					/* use the ZIL to commit this write */
821 					if (zvol_log_write(zv, tx, off, size,
822 					    addr) != 0) {
823 						txg_sync_needed = B_TRUE;
824 					}
825 				}
826 				dmu_tx_commit(tx);
827 			}
828 		}
829 		if (error)
830 			break;
831 		off += size;
832 		addr += size;
833 		resid -= size;
834 	}
835 	rw_exit(&zv->zv_dslock);
836 
837 	if ((bp->b_resid = resid) == bp->b_bcount)
838 		bioerror(bp, off > volsize ? EINVAL : error);
839 
840 	biodone(bp);
841 
842 	if (sync) {
843 		if (txg_sync_needed)
844 			txg_wait_synced(dmu_objset_pool(os), 0);
845 		else
846 			zil_commit(zv->zv_zilog, UINT64_MAX, 0);
847 	}
848 
849 	return (0);
850 }
851 
852 /*ARGSUSED*/
853 int
854 zvol_read(dev_t dev, uio_t *uiop, cred_t *cr)
855 {
856 	return (physio(zvol_strategy, NULL, dev, B_READ, minphys, uiop));
857 }
858 
859 /*ARGSUSED*/
860 int
861 zvol_write(dev_t dev, uio_t *uiop, cred_t *cr)
862 {
863 	return (physio(zvol_strategy, NULL, dev, B_WRITE, minphys, uiop));
864 }
865 
866 /*ARGSUSED*/
867 int
868 zvol_aread(dev_t dev, struct aio_req *aio, cred_t *cr)
869 {
870 	return (aphysio(zvol_strategy, anocancel, dev, B_READ, minphys, aio));
871 }
872 
873 /*ARGSUSED*/
874 int
875 zvol_awrite(dev_t dev, struct aio_req *aio, cred_t *cr)
876 {
877 	return (aphysio(zvol_strategy, anocancel, dev, B_WRITE, minphys, aio));
878 }
879 
880 /*
881  * Dirtbag ioctls to support mkfs(1M) for UFS filesystems.  See dkio(7I).
882  */
883 /*ARGSUSED*/
884 int
885 zvol_ioctl(dev_t dev, int cmd, intptr_t arg, int flag, cred_t *cr, int *rvalp)
886 {
887 	zvol_state_t *zv;
888 	struct dk_cinfo dkc;
889 	struct dk_minfo dkm;
890 	dk_efi_t efi;
891 	efi_gpt_t gpt;
892 	efi_gpe_t gpe;
893 	struct uuid uuid = EFI_RESERVED;
894 	uint32_t crc;
895 	int error = 0;
896 
897 	mutex_enter(&zvol_state_lock);
898 
899 	zv = ddi_get_soft_state(zvol_state, getminor(dev));
900 
901 	if (zv == NULL) {
902 		mutex_exit(&zvol_state_lock);
903 		return (ENXIO);
904 	}
905 
906 	switch (cmd) {
907 
908 	case DKIOCINFO:
909 		bzero(&dkc, sizeof (dkc));
910 		(void) strcpy(dkc.dki_cname, "zvol");
911 		(void) strcpy(dkc.dki_dname, "zvol");
912 		dkc.dki_ctype = DKC_UNKNOWN;
913 		dkc.dki_maxtransfer = 1 << (SPA_MAXBLOCKSHIFT - zv->zv_min_bs);
914 		mutex_exit(&zvol_state_lock);
915 		if (ddi_copyout(&dkc, (void *)arg, sizeof (dkc), flag))
916 			error = EFAULT;
917 		return (error);
918 
919 	case DKIOCGMEDIAINFO:
920 		bzero(&dkm, sizeof (dkm));
921 		dkm.dki_lbsize = 1U << zv->zv_min_bs;
922 		dkm.dki_capacity = zv->zv_volsize >> zv->zv_min_bs;
923 		dkm.dki_media_type = DK_UNKNOWN;
924 		mutex_exit(&zvol_state_lock);
925 		if (ddi_copyout(&dkm, (void *)arg, sizeof (dkm), flag))
926 			error = EFAULT;
927 		return (error);
928 
929 	case DKIOCGETEFI:
930 		if (ddi_copyin((void *)arg, &efi, sizeof (dk_efi_t), flag)) {
931 			mutex_exit(&zvol_state_lock);
932 			return (EFAULT);
933 		}
934 
935 		bzero(&gpt, sizeof (gpt));
936 		bzero(&gpe, sizeof (gpe));
937 
938 		efi.dki_data = (void *)(uintptr_t)efi.dki_data_64;
939 
940 		if (efi.dki_length < sizeof (gpt) + sizeof (gpe)) {
941 			mutex_exit(&zvol_state_lock);
942 			return (EINVAL);
943 		}
944 
945 		efi.dki_length = sizeof (gpt) + sizeof (gpe);
946 
947 		gpt.efi_gpt_Signature = LE_64(EFI_SIGNATURE);
948 		gpt.efi_gpt_Revision = LE_32(EFI_VERSION_CURRENT);
949 		gpt.efi_gpt_HeaderSize = LE_32(sizeof (gpt));
950 		gpt.efi_gpt_FirstUsableLBA = LE_64(0ULL);
951 		gpt.efi_gpt_LastUsableLBA =
952 		    LE_64((zv->zv_volsize >> zv->zv_min_bs) - 1);
953 		gpt.efi_gpt_NumberOfPartitionEntries = LE_32(1);
954 		gpt.efi_gpt_SizeOfPartitionEntry = LE_32(sizeof (gpe));
955 
956 		UUID_LE_CONVERT(gpe.efi_gpe_PartitionTypeGUID, uuid);
957 		gpe.efi_gpe_StartingLBA = gpt.efi_gpt_FirstUsableLBA;
958 		gpe.efi_gpe_EndingLBA = gpt.efi_gpt_LastUsableLBA;
959 
960 		CRC32(crc, &gpe, sizeof (gpe), -1U, crc32_table);
961 		gpt.efi_gpt_PartitionEntryArrayCRC32 = LE_32(~crc);
962 
963 		CRC32(crc, &gpt, sizeof (gpt), -1U, crc32_table);
964 		gpt.efi_gpt_HeaderCRC32 = LE_32(~crc);
965 
966 		mutex_exit(&zvol_state_lock);
967 		if (ddi_copyout(&gpt, efi.dki_data, sizeof (gpt), flag) ||
968 		    ddi_copyout(&gpe, efi.dki_data + 1, sizeof (gpe), flag))
969 			error = EFAULT;
970 		return (error);
971 
972 	default:
973 		error = ENOTSUP;
974 		break;
975 
976 	}
977 	mutex_exit(&zvol_state_lock);
978 	return (error);
979 }
980 
981 int
982 zvol_busy(void)
983 {
984 	return (zvol_minors != 0);
985 }
986 
987 void
988 zvol_init(void)
989 {
990 	VERIFY(ddi_soft_state_init(&zvol_state, sizeof (zvol_state_t), 1) == 0);
991 	mutex_init(&zvol_state_lock, NULL, MUTEX_DEFAULT, NULL);
992 }
993 
994 void
995 zvol_fini(void)
996 {
997 	mutex_destroy(&zvol_state_lock);
998 	ddi_soft_state_fini(&zvol_state);
999 }
1000