xref: /titanic_41/usr/src/uts/common/io/lofi.c (revision f8919bdadda3ebb97bd55cc14a16e0271ed57615)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2007 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 #pragma ident	"%Z%%M%	%I%	%E% SMI"
27 
28 /*
29  * lofi (loopback file) driver - allows you to attach a file to a device,
30  * which can then be accessed through that device. The simple model is that
31  * you tell lofi to open a file, and then use the block device you get as
32  * you would any block device. lofi translates access to the block device
33  * into I/O on the underlying file. This is mostly useful for
34  * mounting images of filesystems.
35  *
36  * lofi is controlled through /dev/lofictl - this is the only device exported
37  * during attach, and is minor number 0. lofiadm communicates with lofi through
38  * ioctls on this device. When a file is attached to lofi, block and character
39  * devices are exported in /dev/lofi and /dev/rlofi. Currently, these devices
40  * are identified by their minor number, and the minor number is also used
41  * as the name in /dev/lofi. If we ever decide to support virtual disks,
42  * we'll have to divide the minor number space to identify fdisk partitions
43  * and slices, and the name will then be the minor number shifted down a
44  * few bits. Minor devices are tracked with state structures handled with
45  * ddi_soft_state(9F) for simplicity.
46  *
47  * A file attached to lofi is opened when attached and not closed until
48  * explicitly detached from lofi. This seems more sensible than deferring
49  * the open until the /dev/lofi device is opened, for a number of reasons.
50  * One is that any failure is likely to be noticed by the person (or script)
51  * running lofiadm. Another is that it would be a security problem if the
52  * file was replaced by another one after being added but before being opened.
53  *
54  * The only hard part about lofi is the ioctls. In order to support things
55  * like 'newfs' on a lofi device, it needs to support certain disk ioctls.
56  * So it has to fake disk geometry and partition information. More may need
57  * to be faked if your favorite utility doesn't work and you think it should
58  * (fdformat doesn't work because it really wants to know the type of floppy
59  * controller to talk to, and that didn't seem easy to fake. Or possibly even
60  * necessary, since we have mkfs_pcfs now).
61  *
62  * Normally, a lofi device cannot be detached if it is open (i.e. busy).  To
63  * support simulation of hotplug events, an optional force flag is provided.
64  * If a lofi device is open when a force detach is requested, then the
65  * underlying file is closed and any subsequent operations return EIO.  When the
66  * device is closed for the last time, it will be cleaned up at that time.  In
67  * addition, the DKIOCSTATE ioctl will return DKIO_DEV_GONE when the device is
68  * detached but not removed.
69  *
70  * Known problems:
71  *
72  *	UFS logging. Mounting a UFS filesystem image "logging"
73  *	works for basic copy testing but wedges during a build of ON through
74  *	that image. Some deadlock in lufs holding the log mutex and then
75  *	getting stuck on a buf. So for now, don't do that.
76  *
77  *	Direct I/O. Since the filesystem data is being cached in the buffer
78  *	cache, _and_ again in the underlying filesystem, it's tempting to
79  *	enable direct I/O on the underlying file. Don't, because that deadlocks.
80  *	I think to fix the cache-twice problem we might need filesystem support.
81  *
82  *	lofi on itself. The simple lock strategy (lofi_lock) precludes this
83  *	because you'll be in lofi_ioctl, holding the lock when you open the
84  *	file, which, if it's lofi, will grab lofi_lock. We prevent this for
85  *	now, though not using ddi_soft_state(9F) would make it possible to
86  *	do. Though it would still be silly.
87  *
88  * Interesting things to do:
89  *
90  *	Allow multiple files for each device. A poor-man's metadisk, basically.
91  *
92  *	Pass-through ioctls on block devices. You can (though it's not
93  *	documented), give lofi a block device as a file name. Then we shouldn't
94  *	need to fake a geometry. But this is also silly unless you're replacing
95  *	metadisk.
96  *
97  *	Encryption. tpm would like this. Apparently Windows 2000 has it, and
98  *	so does Linux.
99  */
100 
101 #include <sys/types.h>
102 #include <netinet/in.h>
103 #include <sys/sysmacros.h>
104 #include <sys/cmn_err.h>
105 #include <sys/uio.h>
106 #include <sys/kmem.h>
107 #include <sys/cred.h>
108 #include <sys/mman.h>
109 #include <sys/errno.h>
110 #include <sys/aio_req.h>
111 #include <sys/stat.h>
112 #include <sys/file.h>
113 #include <sys/modctl.h>
114 #include <sys/conf.h>
115 #include <sys/debug.h>
116 #include <sys/vnode.h>
117 #include <sys/lofi.h>
118 #include <sys/fcntl.h>
119 #include <sys/pathname.h>
120 #include <sys/filio.h>
121 #include <sys/fdio.h>
122 #include <sys/open.h>
123 #include <sys/disp.h>
124 #include <vm/seg_map.h>
125 #include <sys/ddi.h>
126 #include <sys/sunddi.h>
127 #include <sys/zmod.h>
128 
129 #define	NBLOCKS_PROP_NAME	"Nblocks"
130 #define	SIZE_PROP_NAME		"Size"
131 
132 static dev_info_t *lofi_dip;
133 static void	*lofi_statep;
134 static kmutex_t lofi_lock;		/* state lock */
135 
136 /*
137  * Because lofi_taskq_nthreads limits the actual swamping of the device, the
138  * maxalloc parameter (lofi_taskq_maxalloc) should be tuned conservatively
139  * high.  If we want to be assured that the underlying device is always busy,
140  * we must be sure that the number of bytes enqueued when the number of
141  * enqueued tasks exceeds maxalloc is sufficient to keep the device busy for
142  * the duration of the sleep time in taskq_ent_alloc().  That is, lofi should
143  * set maxalloc to be the maximum throughput (in bytes per second) of the
144  * underlying device divided by the minimum I/O size.  We assume a realistic
145  * maximum throughput of one hundred megabytes per second; we set maxalloc on
146  * the lofi task queue to be 104857600 divided by DEV_BSIZE.
147  */
148 static int lofi_taskq_maxalloc = 104857600 / DEV_BSIZE;
149 static int lofi_taskq_nthreads = 4;	/* # of taskq threads per device */
150 
151 uint32_t lofi_max_files = LOFI_MAX_FILES;
152 
153 static int gzip_decompress(void *src, size_t srclen, void *dst,
154 	size_t *destlen, int level);
155 
156 lofi_compress_info_t lofi_compress_table[LOFI_COMPRESS_FUNCTIONS] = {
157 	{gzip_decompress,	NULL,	6,	"gzip"}, /* default */
158 	{gzip_decompress,	NULL,	6,	"gzip-6"},
159 	{gzip_decompress,	NULL,	9,	"gzip-9"}
160 };
161 
162 static int
163 lofi_busy(void)
164 {
165 	minor_t	minor;
166 
167 	/*
168 	 * We need to make sure no mappings exist - mod_remove won't
169 	 * help because the device isn't open.
170 	 */
171 	mutex_enter(&lofi_lock);
172 	for (minor = 1; minor <= lofi_max_files; minor++) {
173 		if (ddi_get_soft_state(lofi_statep, minor) != NULL) {
174 			mutex_exit(&lofi_lock);
175 			return (EBUSY);
176 		}
177 	}
178 	mutex_exit(&lofi_lock);
179 	return (0);
180 }
181 
182 static int
183 is_opened(struct lofi_state *lsp)
184 {
185 	ASSERT(mutex_owned(&lofi_lock));
186 	return (lsp->ls_chr_open || lsp->ls_blk_open || lsp->ls_lyr_open_count);
187 }
188 
189 static int
190 mark_opened(struct lofi_state *lsp, int otyp)
191 {
192 	ASSERT(mutex_owned(&lofi_lock));
193 	switch (otyp) {
194 	case OTYP_CHR:
195 		lsp->ls_chr_open = 1;
196 		break;
197 	case OTYP_BLK:
198 		lsp->ls_blk_open = 1;
199 		break;
200 	case OTYP_LYR:
201 		lsp->ls_lyr_open_count++;
202 		break;
203 	default:
204 		return (-1);
205 	}
206 	return (0);
207 }
208 
209 static void
210 mark_closed(struct lofi_state *lsp, int otyp)
211 {
212 	ASSERT(mutex_owned(&lofi_lock));
213 	switch (otyp) {
214 	case OTYP_CHR:
215 		lsp->ls_chr_open = 0;
216 		break;
217 	case OTYP_BLK:
218 		lsp->ls_blk_open = 0;
219 		break;
220 	case OTYP_LYR:
221 		lsp->ls_lyr_open_count--;
222 		break;
223 	default:
224 		break;
225 	}
226 }
227 
228 static void
229 lofi_free_handle(dev_t dev, minor_t minor, struct lofi_state *lsp,
230     cred_t *credp)
231 {
232 	dev_t	newdev;
233 	char	namebuf[50];
234 
235 	if (lsp->ls_vp) {
236 		(void) VOP_CLOSE(lsp->ls_vp, lsp->ls_openflag,
237 		    1, 0, credp, NULL);
238 		VN_RELE(lsp->ls_vp);
239 		lsp->ls_vp = NULL;
240 	}
241 
242 	newdev = makedevice(getmajor(dev), minor);
243 	(void) ddi_prop_remove(newdev, lofi_dip, SIZE_PROP_NAME);
244 	(void) ddi_prop_remove(newdev, lofi_dip, NBLOCKS_PROP_NAME);
245 
246 	(void) snprintf(namebuf, sizeof (namebuf), "%d", minor);
247 	ddi_remove_minor_node(lofi_dip, namebuf);
248 	(void) snprintf(namebuf, sizeof (namebuf), "%d,raw", minor);
249 	ddi_remove_minor_node(lofi_dip, namebuf);
250 
251 	kmem_free(lsp->ls_filename, lsp->ls_filename_sz);
252 	taskq_destroy(lsp->ls_taskq);
253 	if (lsp->ls_kstat) {
254 		kstat_delete(lsp->ls_kstat);
255 		mutex_destroy(&lsp->ls_kstat_lock);
256 	}
257 	ddi_soft_state_free(lofi_statep, minor);
258 }
259 
260 /*ARGSUSED*/
261 static int
262 lofi_open(dev_t *devp, int flag, int otyp, struct cred *credp)
263 {
264 	minor_t	minor;
265 	struct lofi_state *lsp;
266 
267 	mutex_enter(&lofi_lock);
268 	minor = getminor(*devp);
269 	if (minor == 0) {
270 		/* master control device */
271 		/* must be opened exclusively */
272 		if (((flag & FEXCL) != FEXCL) || (otyp != OTYP_CHR)) {
273 			mutex_exit(&lofi_lock);
274 			return (EINVAL);
275 		}
276 		lsp = ddi_get_soft_state(lofi_statep, 0);
277 		if (lsp == NULL) {
278 			mutex_exit(&lofi_lock);
279 			return (ENXIO);
280 		}
281 		if (is_opened(lsp)) {
282 			mutex_exit(&lofi_lock);
283 			return (EBUSY);
284 		}
285 		(void) mark_opened(lsp, OTYP_CHR);
286 		mutex_exit(&lofi_lock);
287 		return (0);
288 	}
289 
290 	/* otherwise, the mapping should already exist */
291 	lsp = ddi_get_soft_state(lofi_statep, minor);
292 	if (lsp == NULL) {
293 		mutex_exit(&lofi_lock);
294 		return (EINVAL);
295 	}
296 
297 	if (lsp->ls_vp == NULL) {
298 		mutex_exit(&lofi_lock);
299 		return (ENXIO);
300 	}
301 
302 	if (mark_opened(lsp, otyp) == -1) {
303 		mutex_exit(&lofi_lock);
304 		return (EINVAL);
305 	}
306 
307 	mutex_exit(&lofi_lock);
308 	return (0);
309 }
310 
311 /*ARGSUSED*/
312 static int
313 lofi_close(dev_t dev, int flag, int otyp, struct cred *credp)
314 {
315 	minor_t	minor;
316 	struct lofi_state *lsp;
317 
318 	mutex_enter(&lofi_lock);
319 	minor = getminor(dev);
320 	lsp = ddi_get_soft_state(lofi_statep, minor);
321 	if (lsp == NULL) {
322 		mutex_exit(&lofi_lock);
323 		return (EINVAL);
324 	}
325 	mark_closed(lsp, otyp);
326 
327 	/*
328 	 * If we have forcibly closed the underlying device, and this is the
329 	 * last close, then tear down the rest of the device.
330 	 */
331 	if (minor != 0 && lsp->ls_vp == NULL && !is_opened(lsp))
332 		lofi_free_handle(dev, minor, lsp, credp);
333 	mutex_exit(&lofi_lock);
334 	return (0);
335 }
336 
337 static int
338 lofi_mapped_rdwr(caddr_t bufaddr, offset_t offset, struct buf *bp,
339 	struct lofi_state *lsp)
340 {
341 	int error;
342 	offset_t alignedoffset, mapoffset;
343 	size_t	xfersize;
344 	int	isread;
345 	int 	smflags;
346 	caddr_t	mapaddr;
347 	size_t	len;
348 	enum seg_rw srw;
349 
350 	/*
351 	 * segmap always gives us an 8K (MAXBSIZE) chunk, aligned on
352 	 * an 8K boundary, but the buf transfer address may not be
353 	 * aligned on more than a 512-byte boundary (we don't enforce
354 	 * that even though we could). This matters since the initial
355 	 * part of the transfer may not start at offset 0 within the
356 	 * segmap'd chunk. So we have to compensate for that with
357 	 * 'mapoffset'. Subsequent chunks always start off at the
358 	 * beginning, and the last is capped by b_resid
359 	 */
360 	mapoffset = offset & MAXBOFFSET;
361 	alignedoffset = offset - mapoffset;
362 	bp->b_resid = bp->b_bcount;
363 	isread = bp->b_flags & B_READ;
364 	srw = isread ? S_READ : S_WRITE;
365 	do {
366 		xfersize = MIN(lsp->ls_vp_comp_size - offset,
367 		    MIN(MAXBSIZE - mapoffset, bp->b_resid));
368 		len = roundup(mapoffset + xfersize, PAGESIZE);
369 		mapaddr = segmap_getmapflt(segkmap, lsp->ls_vp,
370 		    alignedoffset, MAXBSIZE, 1, srw);
371 		/*
372 		 * Now fault in the pages. This lets us check
373 		 * for errors before we reference mapaddr and
374 		 * try to resolve the fault in bcopy (which would
375 		 * panic instead). And this can easily happen,
376 		 * particularly if you've lofi'd a file over NFS
377 		 * and someone deletes the file on the server.
378 		 */
379 		error = segmap_fault(kas.a_hat, segkmap, mapaddr,
380 		    len, F_SOFTLOCK, srw);
381 		if (error) {
382 			(void) segmap_release(segkmap, mapaddr, 0);
383 			if (FC_CODE(error) == FC_OBJERR)
384 				error = FC_ERRNO(error);
385 			else
386 				error = EIO;
387 			break;
388 		}
389 		smflags = 0;
390 		if (isread) {
391 			smflags |= SM_FREE;
392 			/*
393 			 * If we're reading an entire page starting
394 			 * at a page boundary, there's a good chance
395 			 * we won't need it again. Put it on the
396 			 * head of the freelist.
397 			 */
398 			if (mapoffset == 0 && xfersize == PAGESIZE)
399 				smflags |= SM_DONTNEED;
400 			bcopy(mapaddr + mapoffset, bufaddr, xfersize);
401 		} else {
402 			smflags |= SM_WRITE;
403 			bcopy(bufaddr, mapaddr + mapoffset, xfersize);
404 		}
405 		bp->b_resid -= xfersize;
406 		bufaddr += xfersize;
407 		offset += xfersize;
408 		(void) segmap_fault(kas.a_hat, segkmap, mapaddr,
409 		    len, F_SOFTUNLOCK, srw);
410 		error = segmap_release(segkmap, mapaddr, smflags);
411 		/* only the first map may start partial */
412 		mapoffset = 0;
413 		alignedoffset += MAXBSIZE;
414 	} while ((error == 0) && (bp->b_resid > 0) &&
415 	    (offset < lsp->ls_vp_comp_size));
416 
417 	return (error);
418 }
419 
420 /*ARGSUSED*/
421 static int gzip_decompress(void *src, size_t srclen, void *dst,
422     size_t *dstlen, int level)
423 {
424 	ASSERT(*dstlen >= srclen);
425 
426 	if (z_uncompress(dst, dstlen, src, srclen) != Z_OK)
427 		return (-1);
428 	return (0);
429 }
430 
431 /*
432  * This is basically what strategy used to be before we found we
433  * needed task queues.
434  */
435 static void
436 lofi_strategy_task(void *arg)
437 {
438 	struct buf *bp = (struct buf *)arg;
439 	int error;
440 	struct lofi_state *lsp;
441 	uint64_t sblkno, eblkno, cmpbytes;
442 	offset_t offset, sblkoff, eblkoff;
443 	u_offset_t salign, ealign;
444 	u_offset_t sdiff;
445 	uint32_t comp_data_sz;
446 	caddr_t bufaddr;
447 	unsigned char *compressed_seg = NULL, *cmpbuf;
448 	unsigned char *uncompressed_seg = NULL;
449 	lofi_compress_info_t *li;
450 	size_t oblkcount, xfersize;
451 	unsigned long seglen;
452 
453 	lsp = ddi_get_soft_state(lofi_statep, getminor(bp->b_edev));
454 	if (lsp->ls_kstat) {
455 		mutex_enter(lsp->ls_kstat->ks_lock);
456 		kstat_waitq_to_runq(KSTAT_IO_PTR(lsp->ls_kstat));
457 		mutex_exit(lsp->ls_kstat->ks_lock);
458 	}
459 	bp_mapin(bp);
460 	bufaddr = bp->b_un.b_addr;
461 	offset = bp->b_lblkno * DEV_BSIZE;	/* offset within file */
462 
463 	/*
464 	 * We used to always use vn_rdwr here, but we cannot do that because
465 	 * we might decide to read or write from the the underlying
466 	 * file during this call, which would be a deadlock because
467 	 * we have the rw_lock. So instead we page, unless it's not
468 	 * mapable or it's a character device.
469 	 */
470 	if (lsp->ls_vp == NULL || lsp->ls_vp_closereq) {
471 		error = EIO;
472 	} else if (((lsp->ls_vp->v_flag & VNOMAP) == 0) &&
473 	    (lsp->ls_vp->v_type != VCHR)) {
474 		uint64_t i;
475 
476 		/*
477 		 * Handle uncompressed files with a regular read
478 		 */
479 		if (lsp->ls_uncomp_seg_sz == 0) {
480 			error = lofi_mapped_rdwr(bufaddr, offset, bp, lsp);
481 			goto done;
482 		}
483 
484 		/*
485 		 * From here on we're dealing primarily with compressed files
486 		 */
487 
488 		/*
489 		 * Compressed files can only be read from and
490 		 * not written to
491 		 */
492 		if (!(bp->b_flags & B_READ)) {
493 			bp->b_resid = bp->b_bcount;
494 			error = EROFS;
495 			goto done;
496 		}
497 
498 		ASSERT(lsp->ls_comp_algorithm_index >= 0);
499 		li = &lofi_compress_table[lsp->ls_comp_algorithm_index];
500 		/*
501 		 * Compute starting and ending compressed segment numbers
502 		 * We use only bitwise operations avoiding division and
503 		 * modulus because we enforce the compression segment size
504 		 * to a power of 2
505 		 */
506 		sblkno = offset >> lsp->ls_comp_seg_shift;
507 		sblkoff = offset & (lsp->ls_uncomp_seg_sz - 1);
508 		eblkno = (offset + bp->b_bcount) >> lsp->ls_comp_seg_shift;
509 		eblkoff = (offset + bp->b_bcount) & (lsp->ls_uncomp_seg_sz - 1);
510 
511 		/*
512 		 * Align start offset to block boundary for segmap
513 		 */
514 		salign = lsp->ls_comp_seg_index[sblkno];
515 		sdiff = salign & (DEV_BSIZE - 1);
516 		salign -= sdiff;
517 		if (eblkno >= (lsp->ls_comp_index_sz - 1)) {
518 			/*
519 			 * We're dealing with the last segment of
520 			 * the compressed file -- the size of this
521 			 * segment *may not* be the same as the
522 			 * segment size for the file
523 			 */
524 			eblkoff = (offset + bp->b_bcount) &
525 			    (lsp->ls_uncomp_last_seg_sz - 1);
526 			ealign = lsp->ls_vp_comp_size;
527 		} else {
528 			ealign = lsp->ls_comp_seg_index[eblkno + 1];
529 		}
530 
531 		/*
532 		 * Preserve original request paramaters
533 		 */
534 		oblkcount = bp->b_bcount;
535 
536 		/*
537 		 * Assign the calculated parameters
538 		 */
539 		comp_data_sz = ealign - salign;
540 		bp->b_bcount = comp_data_sz;
541 
542 		/*
543 		 * Allocate fixed size memory blocks to hold compressed
544 		 * segments and one uncompressed segment since we
545 		 * uncompress segments one at a time
546 		 */
547 		compressed_seg = kmem_alloc(bp->b_bcount, KM_SLEEP);
548 		uncompressed_seg = kmem_alloc(lsp->ls_uncomp_seg_sz, KM_SLEEP);
549 		/*
550 		 * Map in the calculated number of blocks
551 		 */
552 		error = lofi_mapped_rdwr((caddr_t)compressed_seg, salign,
553 		    bp, lsp);
554 
555 		bp->b_bcount = oblkcount;
556 		bp->b_resid = oblkcount;
557 		if (error != 0)
558 			goto done;
559 
560 		/*
561 		 * We have the compressed blocks, now uncompress them
562 		 */
563 		cmpbuf = compressed_seg + sdiff;
564 		for (i = sblkno; i < (eblkno + 1) && i < lsp->ls_comp_index_sz;
565 		    i++) {
566 			/*
567 			 * Each of the segment index entries contains
568 			 * the starting block number for that segment.
569 			 * The number of compressed bytes in a segment
570 			 * is thus the difference between the starting
571 			 * block number of this segment and the starting
572 			 * block number of the next segment.
573 			 */
574 			if ((i == eblkno) &&
575 			    (i == lsp->ls_comp_index_sz - 1)) {
576 				cmpbytes = lsp->ls_vp_comp_size -
577 				    lsp->ls_comp_seg_index[i];
578 			} else {
579 				cmpbytes = lsp->ls_comp_seg_index[i + 1] -
580 				    lsp->ls_comp_seg_index[i];
581 			}
582 
583 			/*
584 			 * The first byte in a compressed segment is a flag
585 			 * that indicates whether this segment is compressed
586 			 * at all
587 			 */
588 			if (*cmpbuf == UNCOMPRESSED) {
589 				bcopy((cmpbuf + SEGHDR), uncompressed_seg,
590 				    (cmpbytes - SEGHDR));
591 			} else {
592 				seglen = lsp->ls_uncomp_seg_sz;
593 
594 				if (li->l_decompress((cmpbuf + SEGHDR),
595 				    (cmpbytes - SEGHDR), uncompressed_seg,
596 				    &seglen, li->l_level) != 0) {
597 					error = EIO;
598 					goto done;
599 				}
600 			}
601 
602 			/*
603 			 * Determine how much uncompressed data we
604 			 * have to copy and copy it
605 			 */
606 			xfersize = lsp->ls_uncomp_seg_sz - sblkoff;
607 			if (i == eblkno) {
608 				if (i == (lsp->ls_comp_index_sz - 1))
609 					xfersize -= (lsp->ls_uncomp_last_seg_sz
610 					    - eblkoff);
611 				else
612 					xfersize -=
613 					    (lsp->ls_uncomp_seg_sz - eblkoff);
614 			}
615 
616 			bcopy((uncompressed_seg + sblkoff), bufaddr, xfersize);
617 
618 			cmpbuf += cmpbytes;
619 			bufaddr += xfersize;
620 			bp->b_resid -= xfersize;
621 			sblkoff = 0;
622 
623 			if (bp->b_resid == 0)
624 				break;
625 		}
626 	} else {
627 		ssize_t	resid;
628 		enum uio_rw rw;
629 
630 		if (bp->b_flags & B_READ)
631 			rw = UIO_READ;
632 		else
633 			rw = UIO_WRITE;
634 		error = vn_rdwr(rw, lsp->ls_vp, bufaddr, bp->b_bcount,
635 		    offset, UIO_SYSSPACE, 0, RLIM64_INFINITY, kcred, &resid);
636 		bp->b_resid = resid;
637 	}
638 
639 done:
640 	if (compressed_seg != NULL)
641 		kmem_free(compressed_seg, comp_data_sz);
642 	if (uncompressed_seg != NULL)
643 		kmem_free(uncompressed_seg, lsp->ls_uncomp_seg_sz);
644 
645 	if (lsp->ls_kstat) {
646 		size_t n_done = bp->b_bcount - bp->b_resid;
647 		kstat_io_t *kioptr;
648 
649 		mutex_enter(lsp->ls_kstat->ks_lock);
650 		kioptr = KSTAT_IO_PTR(lsp->ls_kstat);
651 		if (bp->b_flags & B_READ) {
652 			kioptr->nread += n_done;
653 			kioptr->reads++;
654 		} else {
655 			kioptr->nwritten += n_done;
656 			kioptr->writes++;
657 		}
658 		kstat_runq_exit(kioptr);
659 		mutex_exit(lsp->ls_kstat->ks_lock);
660 	}
661 
662 	mutex_enter(&lsp->ls_vp_lock);
663 	if (--lsp->ls_vp_iocount == 0)
664 		cv_broadcast(&lsp->ls_vp_cv);
665 	mutex_exit(&lsp->ls_vp_lock);
666 
667 	bioerror(bp, error);
668 	biodone(bp);
669 }
670 
671 static int
672 lofi_strategy(struct buf *bp)
673 {
674 	struct lofi_state *lsp;
675 	offset_t	offset;
676 
677 	/*
678 	 * We cannot just do I/O here, because the current thread
679 	 * _might_ end up back in here because the underlying filesystem
680 	 * wants a buffer, which eventually gets into bio_recycle and
681 	 * might call into lofi to write out a delayed-write buffer.
682 	 * This is bad if the filesystem above lofi is the same as below.
683 	 *
684 	 * We could come up with a complex strategy using threads to
685 	 * do the I/O asynchronously, or we could use task queues. task
686 	 * queues were incredibly easy so they win.
687 	 */
688 	lsp = ddi_get_soft_state(lofi_statep, getminor(bp->b_edev));
689 	mutex_enter(&lsp->ls_vp_lock);
690 	if (lsp->ls_vp == NULL || lsp->ls_vp_closereq) {
691 		bioerror(bp, EIO);
692 		biodone(bp);
693 		mutex_exit(&lsp->ls_vp_lock);
694 		return (0);
695 	}
696 
697 	offset = bp->b_lblkno * DEV_BSIZE;	/* offset within file */
698 	if (offset == lsp->ls_vp_size) {
699 		/* EOF */
700 		if ((bp->b_flags & B_READ) != 0) {
701 			bp->b_resid = bp->b_bcount;
702 			bioerror(bp, 0);
703 		} else {
704 			/* writes should fail */
705 			bioerror(bp, ENXIO);
706 		}
707 		biodone(bp);
708 		mutex_exit(&lsp->ls_vp_lock);
709 		return (0);
710 	}
711 	if (offset > lsp->ls_vp_size) {
712 		bioerror(bp, ENXIO);
713 		biodone(bp);
714 		mutex_exit(&lsp->ls_vp_lock);
715 		return (0);
716 	}
717 	lsp->ls_vp_iocount++;
718 	mutex_exit(&lsp->ls_vp_lock);
719 
720 	if (lsp->ls_kstat) {
721 		mutex_enter(lsp->ls_kstat->ks_lock);
722 		kstat_waitq_enter(KSTAT_IO_PTR(lsp->ls_kstat));
723 		mutex_exit(lsp->ls_kstat->ks_lock);
724 	}
725 	(void) taskq_dispatch(lsp->ls_taskq, lofi_strategy_task, bp, KM_SLEEP);
726 	return (0);
727 }
728 
729 /*ARGSUSED2*/
730 static int
731 lofi_read(dev_t dev, struct uio *uio, struct cred *credp)
732 {
733 	if (getminor(dev) == 0)
734 		return (EINVAL);
735 	return (physio(lofi_strategy, NULL, dev, B_READ, minphys, uio));
736 }
737 
738 /*ARGSUSED2*/
739 static int
740 lofi_write(dev_t dev, struct uio *uio, struct cred *credp)
741 {
742 	if (getminor(dev) == 0)
743 		return (EINVAL);
744 	return (physio(lofi_strategy, NULL, dev, B_WRITE, minphys, uio));
745 }
746 
747 /*ARGSUSED2*/
748 static int
749 lofi_aread(dev_t dev, struct aio_req *aio, struct cred *credp)
750 {
751 	if (getminor(dev) == 0)
752 		return (EINVAL);
753 	return (aphysio(lofi_strategy, anocancel, dev, B_READ, minphys, aio));
754 }
755 
756 /*ARGSUSED2*/
757 static int
758 lofi_awrite(dev_t dev, struct aio_req *aio, struct cred *credp)
759 {
760 	if (getminor(dev) == 0)
761 		return (EINVAL);
762 	return (aphysio(lofi_strategy, anocancel, dev, B_WRITE, minphys, aio));
763 }
764 
765 /*ARGSUSED*/
766 static int
767 lofi_info(dev_info_t *dip, ddi_info_cmd_t infocmd, void *arg, void **result)
768 {
769 	switch (infocmd) {
770 	case DDI_INFO_DEVT2DEVINFO:
771 		*result = lofi_dip;
772 		return (DDI_SUCCESS);
773 	case DDI_INFO_DEVT2INSTANCE:
774 		*result = 0;
775 		return (DDI_SUCCESS);
776 	}
777 	return (DDI_FAILURE);
778 }
779 
780 static int
781 lofi_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
782 {
783 	int	error;
784 
785 	if (cmd != DDI_ATTACH)
786 		return (DDI_FAILURE);
787 	error = ddi_soft_state_zalloc(lofi_statep, 0);
788 	if (error == DDI_FAILURE) {
789 		return (DDI_FAILURE);
790 	}
791 	error = ddi_create_minor_node(dip, LOFI_CTL_NODE, S_IFCHR, 0,
792 	    DDI_PSEUDO, NULL);
793 	if (error == DDI_FAILURE) {
794 		ddi_soft_state_free(lofi_statep, 0);
795 		return (DDI_FAILURE);
796 	}
797 	/* driver handles kernel-issued IOCTLs */
798 	if (ddi_prop_create(DDI_DEV_T_NONE, dip, DDI_PROP_CANSLEEP,
799 	    DDI_KERNEL_IOCTL, NULL, 0) != DDI_PROP_SUCCESS) {
800 		ddi_remove_minor_node(dip, NULL);
801 		ddi_soft_state_free(lofi_statep, 0);
802 		return (DDI_FAILURE);
803 	}
804 	lofi_dip = dip;
805 	ddi_report_dev(dip);
806 	return (DDI_SUCCESS);
807 }
808 
809 static int
810 lofi_detach(dev_info_t *dip, ddi_detach_cmd_t cmd)
811 {
812 	if (cmd != DDI_DETACH)
813 		return (DDI_FAILURE);
814 	if (lofi_busy())
815 		return (DDI_FAILURE);
816 	lofi_dip = NULL;
817 	ddi_remove_minor_node(dip, NULL);
818 	ddi_prop_remove_all(dip);
819 	ddi_soft_state_free(lofi_statep, 0);
820 	return (DDI_SUCCESS);
821 }
822 
823 /*
824  * These two just simplify the rest of the ioctls that need to copyin/out
825  * the lofi_ioctl structure.
826  */
827 struct lofi_ioctl *
828 copy_in_lofi_ioctl(const struct lofi_ioctl *ulip, int flag)
829 {
830 	struct lofi_ioctl *klip;
831 	int	error;
832 
833 	klip = kmem_alloc(sizeof (struct lofi_ioctl), KM_SLEEP);
834 	error = ddi_copyin(ulip, klip, sizeof (struct lofi_ioctl), flag);
835 	if (error) {
836 		kmem_free(klip, sizeof (struct lofi_ioctl));
837 		return (NULL);
838 	}
839 
840 	/* make sure filename is always null-terminated */
841 	klip->li_filename[MAXPATHLEN] = '\0';
842 
843 	/* validate minor number */
844 	if (klip->li_minor > lofi_max_files) {
845 		kmem_free(klip, sizeof (struct lofi_ioctl));
846 		return (NULL);
847 	}
848 	return (klip);
849 }
850 
851 int
852 copy_out_lofi_ioctl(const struct lofi_ioctl *klip, struct lofi_ioctl *ulip,
853 	int flag)
854 {
855 	int	error;
856 
857 	error = ddi_copyout(klip, ulip, sizeof (struct lofi_ioctl), flag);
858 	if (error)
859 		return (EFAULT);
860 	return (0);
861 }
862 
863 void
864 free_lofi_ioctl(struct lofi_ioctl *klip)
865 {
866 	kmem_free(klip, sizeof (struct lofi_ioctl));
867 }
868 
869 /*
870  * Return the minor number 'filename' is mapped to, if it is.
871  */
872 static int
873 file_to_minor(char *filename)
874 {
875 	minor_t	minor;
876 	struct lofi_state *lsp;
877 
878 	ASSERT(mutex_owned(&lofi_lock));
879 	for (minor = 1; minor <= lofi_max_files; minor++) {
880 		lsp = ddi_get_soft_state(lofi_statep, minor);
881 		if (lsp == NULL)
882 			continue;
883 		if (strcmp(lsp->ls_filename, filename) == 0)
884 			return (minor);
885 	}
886 	return (0);
887 }
888 
889 /*
890  * lofiadm does some validation, but since Joe Random (or crashme) could
891  * do our ioctls, we need to do some validation too.
892  */
893 static int
894 valid_filename(const char *filename)
895 {
896 	static char *blkprefix = "/dev/" LOFI_BLOCK_NAME "/";
897 	static char *charprefix = "/dev/" LOFI_CHAR_NAME "/";
898 
899 	/* must be absolute path */
900 	if (filename[0] != '/')
901 		return (0);
902 	/* must not be lofi */
903 	if (strncmp(filename, blkprefix, strlen(blkprefix)) == 0)
904 		return (0);
905 	if (strncmp(filename, charprefix, strlen(charprefix)) == 0)
906 		return (0);
907 	return (1);
908 }
909 
910 /*
911  * Fakes up a disk geometry, and one big partition, based on the size
912  * of the file. This is needed because we allow newfs'ing the device,
913  * and newfs will do several disk ioctls to figure out the geometry and
914  * partition information. It uses that information to determine the parameters
915  * to pass to mkfs. Geometry is pretty much irrelevant these days, but we
916  * have to support it.
917  */
918 static void
919 fake_disk_geometry(struct lofi_state *lsp)
920 {
921 	/* dk_geom - see dkio(7I) */
922 	/*
923 	 * dkg_ncyl _could_ be set to one here (one big cylinder with gobs
924 	 * of sectors), but that breaks programs like fdisk which want to
925 	 * partition a disk by cylinder. With one cylinder, you can't create
926 	 * an fdisk partition and put pcfs on it for testing (hard to pick
927 	 * a number between one and one).
928 	 *
929 	 * The cheezy floppy test is an attempt to not have too few cylinders
930 	 * for a small file, or so many on a big file that you waste space
931 	 * for backup superblocks or cylinder group structures.
932 	 */
933 	if (lsp->ls_vp_size < (2 * 1024 * 1024)) /* floppy? */
934 		lsp->ls_dkg.dkg_ncyl = lsp->ls_vp_size / (100 * 1024);
935 	else
936 		lsp->ls_dkg.dkg_ncyl = lsp->ls_vp_size / (300 * 1024);
937 	/* in case file file is < 100k */
938 	if (lsp->ls_dkg.dkg_ncyl == 0)
939 		lsp->ls_dkg.dkg_ncyl = 1;
940 	lsp->ls_dkg.dkg_acyl = 0;
941 	lsp->ls_dkg.dkg_bcyl = 0;
942 	lsp->ls_dkg.dkg_nhead = 1;
943 	lsp->ls_dkg.dkg_obs1 = 0;
944 	lsp->ls_dkg.dkg_intrlv = 0;
945 	lsp->ls_dkg.dkg_obs2 = 0;
946 	lsp->ls_dkg.dkg_obs3 = 0;
947 	lsp->ls_dkg.dkg_apc = 0;
948 	lsp->ls_dkg.dkg_rpm = 7200;
949 	lsp->ls_dkg.dkg_pcyl = lsp->ls_dkg.dkg_ncyl + lsp->ls_dkg.dkg_acyl;
950 	lsp->ls_dkg.dkg_nsect = lsp->ls_vp_size /
951 	    (DEV_BSIZE * lsp->ls_dkg.dkg_ncyl);
952 	lsp->ls_dkg.dkg_write_reinstruct = 0;
953 	lsp->ls_dkg.dkg_read_reinstruct = 0;
954 
955 	/* vtoc - see dkio(7I) */
956 	bzero(&lsp->ls_vtoc, sizeof (struct vtoc));
957 	lsp->ls_vtoc.v_sanity = VTOC_SANE;
958 	lsp->ls_vtoc.v_version = V_VERSION;
959 	bcopy(LOFI_DRIVER_NAME, lsp->ls_vtoc.v_volume, 7);
960 	lsp->ls_vtoc.v_sectorsz = DEV_BSIZE;
961 	lsp->ls_vtoc.v_nparts = 1;
962 	lsp->ls_vtoc.v_part[0].p_tag = V_UNASSIGNED;
963 
964 	/*
965 	 * A compressed file is read-only, other files can
966 	 * be read-write
967 	 */
968 	if (lsp->ls_uncomp_seg_sz > 0) {
969 		lsp->ls_vtoc.v_part[0].p_flag = V_UNMNT | V_RONLY;
970 	} else {
971 		lsp->ls_vtoc.v_part[0].p_flag = V_UNMNT;
972 	}
973 	lsp->ls_vtoc.v_part[0].p_start = (daddr_t)0;
974 	/*
975 	 * The partition size cannot just be the number of sectors, because
976 	 * that might not end on a cylinder boundary. And if that's the case,
977 	 * newfs/mkfs will print a scary warning. So just figure the size
978 	 * based on the number of cylinders and sectors/cylinder.
979 	 */
980 	lsp->ls_vtoc.v_part[0].p_size = lsp->ls_dkg.dkg_pcyl *
981 	    lsp->ls_dkg.dkg_nsect * lsp->ls_dkg.dkg_nhead;
982 
983 	/* dk_cinfo - see dkio(7I) */
984 	bzero(&lsp->ls_ci, sizeof (struct dk_cinfo));
985 	(void) strcpy(lsp->ls_ci.dki_cname, LOFI_DRIVER_NAME);
986 	lsp->ls_ci.dki_ctype = DKC_MD;
987 	lsp->ls_ci.dki_flags = 0;
988 	lsp->ls_ci.dki_cnum = 0;
989 	lsp->ls_ci.dki_addr = 0;
990 	lsp->ls_ci.dki_space = 0;
991 	lsp->ls_ci.dki_prio = 0;
992 	lsp->ls_ci.dki_vec = 0;
993 	(void) strcpy(lsp->ls_ci.dki_dname, LOFI_DRIVER_NAME);
994 	lsp->ls_ci.dki_unit = 0;
995 	lsp->ls_ci.dki_slave = 0;
996 	lsp->ls_ci.dki_partition = 0;
997 	/*
998 	 * newfs uses this to set maxcontig. Must not be < 16, or it
999 	 * will be 0 when newfs multiplies it by DEV_BSIZE and divides
1000 	 * it by the block size. Then tunefs doesn't work because
1001 	 * maxcontig is 0.
1002 	 */
1003 	lsp->ls_ci.dki_maxtransfer = 16;
1004 }
1005 
1006 /*
1007  * map in a compressed file
1008  *
1009  * Read in the header and the index that follows.
1010  *
1011  * The header is as follows -
1012  *
1013  * Signature (name of the compression algorithm)
1014  * Compression segment size (a multiple of 512)
1015  * Number of index entries
1016  * Size of the last block
1017  * The array containing the index entries
1018  *
1019  * The header information is always stored in
1020  * network byte order on disk.
1021  */
1022 static int
1023 lofi_map_compressed_file(struct lofi_state *lsp, char *buf)
1024 {
1025 	uint32_t index_sz, header_len, i;
1026 	ssize_t	resid;
1027 	enum uio_rw rw;
1028 	char *tbuf = buf;
1029 	int error;
1030 
1031 	/* The signature has already been read */
1032 	tbuf += sizeof (lsp->ls_comp_algorithm);
1033 	bcopy(tbuf, &(lsp->ls_uncomp_seg_sz), sizeof (lsp->ls_uncomp_seg_sz));
1034 	lsp->ls_uncomp_seg_sz = ntohl(lsp->ls_uncomp_seg_sz);
1035 
1036 	/*
1037 	 * The compressed segment size must be a power of 2
1038 	 */
1039 	if (lsp->ls_uncomp_seg_sz % 2)
1040 		return (EINVAL);
1041 
1042 	for (i = 0; !((lsp->ls_uncomp_seg_sz >> i) & 1); i++)
1043 		;
1044 
1045 	lsp->ls_comp_seg_shift = i;
1046 
1047 	tbuf += sizeof (lsp->ls_uncomp_seg_sz);
1048 	bcopy(tbuf, &(lsp->ls_comp_index_sz), sizeof (lsp->ls_comp_index_sz));
1049 	lsp->ls_comp_index_sz = ntohl(lsp->ls_comp_index_sz);
1050 
1051 	tbuf += sizeof (lsp->ls_comp_index_sz);
1052 	bcopy(tbuf, &(lsp->ls_uncomp_last_seg_sz),
1053 	    sizeof (lsp->ls_uncomp_last_seg_sz));
1054 	lsp->ls_uncomp_last_seg_sz = ntohl(lsp->ls_uncomp_last_seg_sz);
1055 
1056 	/*
1057 	 * Compute the total size of the uncompressed data
1058 	 * for use in fake_disk_geometry and other calculations.
1059 	 * Disk geometry has to be faked with respect to the
1060 	 * actual uncompressed data size rather than the
1061 	 * compressed file size.
1062 	 */
1063 	lsp->ls_vp_size = (lsp->ls_comp_index_sz - 2) * lsp->ls_uncomp_seg_sz
1064 	    + lsp->ls_uncomp_last_seg_sz;
1065 
1066 	/*
1067 	 * Index size is rounded up to a 512 byte boundary for ease
1068 	 * of segmapping
1069 	 */
1070 	index_sz = sizeof (*lsp->ls_comp_seg_index) * lsp->ls_comp_index_sz;
1071 	header_len = sizeof (lsp->ls_comp_algorithm) +
1072 	    sizeof (lsp->ls_uncomp_seg_sz) +
1073 	    sizeof (lsp->ls_comp_index_sz) +
1074 	    sizeof (lsp->ls_uncomp_last_seg_sz);
1075 	lsp->ls_comp_offbase = header_len + index_sz;
1076 
1077 	index_sz += header_len;
1078 	index_sz = roundup(index_sz, DEV_BSIZE);
1079 
1080 	lsp->ls_comp_index_data = kmem_alloc(index_sz, KM_SLEEP);
1081 	lsp->ls_comp_index_data_sz = index_sz;
1082 
1083 	/*
1084 	 * Read in the index -- this has a side-effect
1085 	 * of reading in the header as well
1086 	 */
1087 	rw = UIO_READ;
1088 	error = vn_rdwr(rw, lsp->ls_vp, lsp->ls_comp_index_data, index_sz,
1089 	    0, UIO_SYSSPACE, 0, RLIM64_INFINITY, kcred, &resid);
1090 
1091 	if (error != 0)
1092 		return (error);
1093 
1094 	/* Skip the header, this is where the index really begins */
1095 	lsp->ls_comp_seg_index =
1096 	    /*LINTED*/
1097 	    (uint64_t *)(lsp->ls_comp_index_data + header_len);
1098 
1099 	/*
1100 	 * Now recompute offsets in the index to account for
1101 	 * the header length
1102 	 */
1103 	for (i = 0; i < lsp->ls_comp_index_sz; i++) {
1104 		lsp->ls_comp_seg_index[i] = lsp->ls_comp_offbase +
1105 		    BE_64(lsp->ls_comp_seg_index[i]);
1106 	}
1107 
1108 	return (error);
1109 }
1110 
1111 /*
1112  * Check to see if the passed in signature is a valid
1113  * one. If it is valid, return the index into
1114  * lofi_compress_table.
1115  *
1116  * Return -1 if it is invalid
1117  */
1118 static int lofi_compress_select(char *signature)
1119 {
1120 	int i;
1121 
1122 	for (i = 0; i < LOFI_COMPRESS_FUNCTIONS; i++) {
1123 		if (strcmp(lofi_compress_table[i].l_name, signature) == 0)
1124 			return (i);
1125 	}
1126 
1127 	return (-1);
1128 }
1129 
1130 /*
1131  * map a file to a minor number. Return the minor number.
1132  */
1133 static int
1134 lofi_map_file(dev_t dev, struct lofi_ioctl *ulip, int pickminor,
1135     int *rvalp, struct cred *credp, int ioctl_flag)
1136 {
1137 	minor_t	newminor;
1138 	struct lofi_state *lsp;
1139 	struct lofi_ioctl *klip;
1140 	int	error;
1141 	struct vnode *vp;
1142 	int64_t	Nblocks_prop_val;
1143 	int64_t	Size_prop_val;
1144 	int	compress_index;
1145 	vattr_t	vattr;
1146 	int	flag;
1147 	enum vtype v_type;
1148 	int zalloced = 0;
1149 	dev_t	newdev;
1150 	char	namebuf[50];
1151 	char 	buf[DEV_BSIZE];
1152 	char 	*tbuf;
1153 	ssize_t	resid;
1154 	enum uio_rw rw;
1155 
1156 	klip = copy_in_lofi_ioctl(ulip, ioctl_flag);
1157 	if (klip == NULL)
1158 		return (EFAULT);
1159 
1160 	mutex_enter(&lofi_lock);
1161 
1162 	if (!valid_filename(klip->li_filename)) {
1163 		error = EINVAL;
1164 		goto out;
1165 	}
1166 
1167 	if (file_to_minor(klip->li_filename) != 0) {
1168 		error = EBUSY;
1169 		goto out;
1170 	}
1171 
1172 	if (pickminor) {
1173 		/* Find a free one */
1174 		for (newminor = 1; newminor <= lofi_max_files; newminor++)
1175 			if (ddi_get_soft_state(lofi_statep, newminor) == NULL)
1176 				break;
1177 		if (newminor >= lofi_max_files) {
1178 			error = EAGAIN;
1179 			goto out;
1180 		}
1181 	} else {
1182 		newminor = klip->li_minor;
1183 		if (ddi_get_soft_state(lofi_statep, newminor) != NULL) {
1184 			error = EEXIST;
1185 			goto out;
1186 		}
1187 	}
1188 
1189 	/* make sure it's valid */
1190 	error = lookupname(klip->li_filename, UIO_SYSSPACE, FOLLOW,
1191 	    NULLVPP, &vp);
1192 	if (error) {
1193 		goto out;
1194 	}
1195 	v_type = vp->v_type;
1196 	VN_RELE(vp);
1197 	if (!V_ISLOFIABLE(v_type)) {
1198 		error = EINVAL;
1199 		goto out;
1200 	}
1201 	flag = FREAD | FWRITE | FOFFMAX | FEXCL;
1202 	error = vn_open(klip->li_filename, UIO_SYSSPACE, flag, 0, &vp, 0, 0);
1203 	if (error) {
1204 		/* try read-only */
1205 		flag &= ~FWRITE;
1206 		error = vn_open(klip->li_filename, UIO_SYSSPACE, flag, 0,
1207 		    &vp, 0, 0);
1208 		if (error) {
1209 			goto out;
1210 		}
1211 	}
1212 	vattr.va_mask = AT_SIZE;
1213 	error = VOP_GETATTR(vp, &vattr, 0, credp, NULL);
1214 	if (error) {
1215 		goto closeout;
1216 	}
1217 	/* the file needs to be a multiple of the block size */
1218 	if ((vattr.va_size % DEV_BSIZE) != 0) {
1219 		error = EINVAL;
1220 		goto closeout;
1221 	}
1222 	newdev = makedevice(getmajor(dev), newminor);
1223 	Size_prop_val = vattr.va_size;
1224 	if ((ddi_prop_update_int64(newdev, lofi_dip,
1225 	    SIZE_PROP_NAME, Size_prop_val)) != DDI_PROP_SUCCESS) {
1226 		error = EINVAL;
1227 		goto closeout;
1228 	}
1229 	Nblocks_prop_val = vattr.va_size / DEV_BSIZE;
1230 	if ((ddi_prop_update_int64(newdev, lofi_dip,
1231 	    NBLOCKS_PROP_NAME, Nblocks_prop_val)) != DDI_PROP_SUCCESS) {
1232 		error = EINVAL;
1233 		goto propout;
1234 	}
1235 	error = ddi_soft_state_zalloc(lofi_statep, newminor);
1236 	if (error == DDI_FAILURE) {
1237 		error = ENOMEM;
1238 		goto propout;
1239 	}
1240 	zalloced = 1;
1241 	(void) snprintf(namebuf, sizeof (namebuf), "%d", newminor);
1242 	(void) ddi_create_minor_node(lofi_dip, namebuf, S_IFBLK, newminor,
1243 	    DDI_PSEUDO, NULL);
1244 	if (error != DDI_SUCCESS) {
1245 		error = ENXIO;
1246 		goto propout;
1247 	}
1248 	(void) snprintf(namebuf, sizeof (namebuf), "%d,raw", newminor);
1249 	error = ddi_create_minor_node(lofi_dip, namebuf, S_IFCHR, newminor,
1250 	    DDI_PSEUDO, NULL);
1251 	if (error != DDI_SUCCESS) {
1252 		/* remove block node */
1253 		(void) snprintf(namebuf, sizeof (namebuf), "%d", newminor);
1254 		ddi_remove_minor_node(lofi_dip, namebuf);
1255 		error = ENXIO;
1256 		goto propout;
1257 	}
1258 	lsp = ddi_get_soft_state(lofi_statep, newminor);
1259 	lsp->ls_filename_sz = strlen(klip->li_filename) + 1;
1260 	lsp->ls_filename = kmem_alloc(lsp->ls_filename_sz, KM_SLEEP);
1261 	(void) snprintf(namebuf, sizeof (namebuf), "%s_taskq_%d",
1262 	    LOFI_DRIVER_NAME, newminor);
1263 	lsp->ls_taskq = taskq_create(namebuf, lofi_taskq_nthreads,
1264 	    minclsyspri, 1, lofi_taskq_maxalloc, 0);
1265 	lsp->ls_kstat = kstat_create(LOFI_DRIVER_NAME, newminor,
1266 	    NULL, "disk", KSTAT_TYPE_IO, 1, 0);
1267 	if (lsp->ls_kstat) {
1268 		mutex_init(&lsp->ls_kstat_lock, NULL, MUTEX_DRIVER, NULL);
1269 		lsp->ls_kstat->ks_lock = &lsp->ls_kstat_lock;
1270 		kstat_install(lsp->ls_kstat);
1271 	}
1272 	cv_init(&lsp->ls_vp_cv, NULL, CV_DRIVER, NULL);
1273 	mutex_init(&lsp->ls_vp_lock, NULL, MUTEX_DRIVER, NULL);
1274 
1275 	/*
1276 	 * save open mode so file can be closed properly and vnode counts
1277 	 * updated correctly.
1278 	 */
1279 	lsp->ls_openflag = flag;
1280 
1281 	/*
1282 	 * Try to handle stacked lofs vnodes.
1283 	 */
1284 	if (vp->v_type == VREG) {
1285 		if (VOP_REALVP(vp, &lsp->ls_vp, NULL) != 0) {
1286 			lsp->ls_vp = vp;
1287 		} else {
1288 			/*
1289 			 * Even though vp was obtained via vn_open(), we
1290 			 * can't call vn_close() on it, since lofs will
1291 			 * pass the VOP_CLOSE() on down to the realvp
1292 			 * (which we are about to use). Hence we merely
1293 			 * drop the reference to the lofs vnode and hold
1294 			 * the realvp so things behave as if we've
1295 			 * opened the realvp without any interaction
1296 			 * with lofs.
1297 			 */
1298 			VN_HOLD(lsp->ls_vp);
1299 			VN_RELE(vp);
1300 		}
1301 	} else {
1302 		lsp->ls_vp = vp;
1303 	}
1304 	lsp->ls_vp_size = vattr.va_size;
1305 	(void) strcpy(lsp->ls_filename, klip->li_filename);
1306 	if (rvalp)
1307 		*rvalp = (int)newminor;
1308 	klip->li_minor = newminor;
1309 
1310 	/*
1311 	 * Read the file signature to check if it is compressed.
1312 	 * 'rw' is set to read since only reads are allowed to
1313 	 * a compressed file.
1314 	 */
1315 	rw = UIO_READ;
1316 	error = vn_rdwr(rw, lsp->ls_vp, buf, DEV_BSIZE, 0, UIO_SYSSPACE,
1317 	    0, RLIM64_INFINITY, kcred, &resid);
1318 
1319 	if (error != 0)
1320 		goto propout;
1321 
1322 	tbuf = buf;
1323 	lsp->ls_uncomp_seg_sz = 0;
1324 	lsp->ls_vp_comp_size = lsp->ls_vp_size;
1325 	lsp->ls_comp_algorithm[0] = '\0';
1326 
1327 	compress_index = lofi_compress_select(tbuf);
1328 	if (compress_index != -1) {
1329 		lsp->ls_comp_algorithm_index = compress_index;
1330 		(void) strlcpy(lsp->ls_comp_algorithm,
1331 		    lofi_compress_table[compress_index].l_name,
1332 		    sizeof (lsp->ls_comp_algorithm));
1333 		error = lofi_map_compressed_file(lsp, buf);
1334 		if (error != 0)
1335 			goto propout;
1336 
1337 		/* update DDI properties */
1338 		Size_prop_val = lsp->ls_vp_size;
1339 		if ((ddi_prop_update_int64(newdev, lofi_dip, SIZE_PROP_NAME,
1340 		    Size_prop_val)) != DDI_PROP_SUCCESS) {
1341 			error = EINVAL;
1342 			goto propout;
1343 		}
1344 
1345 		Nblocks_prop_val = lsp->ls_vp_size / DEV_BSIZE;
1346 		if ((ddi_prop_update_int64(newdev, lofi_dip, NBLOCKS_PROP_NAME,
1347 		    Nblocks_prop_val)) != DDI_PROP_SUCCESS) {
1348 			error = EINVAL;
1349 			goto propout;
1350 		}
1351 	}
1352 
1353 	fake_disk_geometry(lsp);
1354 	mutex_exit(&lofi_lock);
1355 	(void) copy_out_lofi_ioctl(klip, ulip, ioctl_flag);
1356 	free_lofi_ioctl(klip);
1357 	return (0);
1358 
1359 propout:
1360 	(void) ddi_prop_remove(newdev, lofi_dip, SIZE_PROP_NAME);
1361 	(void) ddi_prop_remove(newdev, lofi_dip, NBLOCKS_PROP_NAME);
1362 closeout:
1363 	(void) VOP_CLOSE(vp, flag, 1, 0, credp, NULL);
1364 	VN_RELE(vp);
1365 out:
1366 	if (zalloced)
1367 		ddi_soft_state_free(lofi_statep, newminor);
1368 	mutex_exit(&lofi_lock);
1369 	free_lofi_ioctl(klip);
1370 	return (error);
1371 }
1372 
1373 /*
1374  * unmap a file.
1375  */
1376 static int
1377 lofi_unmap_file(dev_t dev, struct lofi_ioctl *ulip, int byfilename,
1378     struct cred *credp, int ioctl_flag)
1379 {
1380 	struct lofi_state *lsp;
1381 	struct lofi_ioctl *klip;
1382 	minor_t	minor;
1383 
1384 	klip = copy_in_lofi_ioctl(ulip, ioctl_flag);
1385 	if (klip == NULL)
1386 		return (EFAULT);
1387 
1388 	mutex_enter(&lofi_lock);
1389 	if (byfilename) {
1390 		minor = file_to_minor(klip->li_filename);
1391 	} else {
1392 		minor = klip->li_minor;
1393 	}
1394 	if (minor == 0) {
1395 		mutex_exit(&lofi_lock);
1396 		free_lofi_ioctl(klip);
1397 		return (ENXIO);
1398 	}
1399 	lsp = ddi_get_soft_state(lofi_statep, minor);
1400 	if (lsp == NULL || lsp->ls_vp == NULL) {
1401 		mutex_exit(&lofi_lock);
1402 		free_lofi_ioctl(klip);
1403 		return (ENXIO);
1404 	}
1405 
1406 	if (is_opened(lsp)) {
1407 		/*
1408 		 * If the 'force' flag is set, then we forcibly close the
1409 		 * underlying file.  Subsequent operations will fail, and the
1410 		 * DKIOCSTATE ioctl will return DKIO_DEV_GONE.  When the device
1411 		 * is last closed, the device will be cleaned up appropriately.
1412 		 *
1413 		 * This is complicated by the fact that we may have outstanding
1414 		 * dispatched I/Os.  Rather than having a single mutex to
1415 		 * serialize all I/O, we keep a count of the number of
1416 		 * outstanding I/O requests, as well as a flag to indicate that
1417 		 * no new I/Os should be dispatched.  We set the flag, wait for
1418 		 * the number of outstanding I/Os to reach 0, and then close the
1419 		 * underlying vnode.
1420 		 */
1421 		if (klip->li_force) {
1422 			mutex_enter(&lsp->ls_vp_lock);
1423 			lsp->ls_vp_closereq = B_TRUE;
1424 			while (lsp->ls_vp_iocount > 0)
1425 				cv_wait(&lsp->ls_vp_cv, &lsp->ls_vp_lock);
1426 			(void) VOP_CLOSE(lsp->ls_vp, lsp->ls_openflag, 1, 0,
1427 			    credp, NULL);
1428 			VN_RELE(lsp->ls_vp);
1429 			lsp->ls_vp = NULL;
1430 			cv_broadcast(&lsp->ls_vp_cv);
1431 			mutex_exit(&lsp->ls_vp_lock);
1432 			mutex_exit(&lofi_lock);
1433 			klip->li_minor = minor;
1434 			(void) copy_out_lofi_ioctl(klip, ulip, ioctl_flag);
1435 			free_lofi_ioctl(klip);
1436 			return (0);
1437 		}
1438 		mutex_exit(&lofi_lock);
1439 		free_lofi_ioctl(klip);
1440 		return (EBUSY);
1441 	}
1442 
1443 	if (lsp->ls_uncomp_seg_sz > 0) {
1444 		kmem_free(lsp->ls_comp_index_data, lsp->ls_comp_index_data_sz);
1445 		lsp->ls_uncomp_seg_sz = 0;
1446 	}
1447 
1448 	lofi_free_handle(dev, minor, lsp, credp);
1449 
1450 	klip->li_minor = minor;
1451 	mutex_exit(&lofi_lock);
1452 	(void) copy_out_lofi_ioctl(klip, ulip, ioctl_flag);
1453 	free_lofi_ioctl(klip);
1454 	return (0);
1455 }
1456 
1457 /*
1458  * get the filename given the minor number, or the minor number given
1459  * the name.
1460  */
1461 /*ARGSUSED*/
1462 static int
1463 lofi_get_info(dev_t dev, struct lofi_ioctl *ulip, int which,
1464     struct cred *credp, int ioctl_flag)
1465 {
1466 	struct lofi_state *lsp;
1467 	struct lofi_ioctl *klip;
1468 	int	error;
1469 	minor_t	minor;
1470 
1471 	klip = copy_in_lofi_ioctl(ulip, ioctl_flag);
1472 	if (klip == NULL)
1473 		return (EFAULT);
1474 
1475 	switch (which) {
1476 	case LOFI_GET_FILENAME:
1477 		minor = klip->li_minor;
1478 		if (minor == 0) {
1479 			free_lofi_ioctl(klip);
1480 			return (EINVAL);
1481 		}
1482 
1483 		mutex_enter(&lofi_lock);
1484 		lsp = ddi_get_soft_state(lofi_statep, minor);
1485 		if (lsp == NULL) {
1486 			mutex_exit(&lofi_lock);
1487 			free_lofi_ioctl(klip);
1488 			return (ENXIO);
1489 		}
1490 		(void) strcpy(klip->li_filename, lsp->ls_filename);
1491 		(void) strlcpy(klip->li_algorithm, lsp->ls_comp_algorithm,
1492 		    sizeof (klip->li_algorithm));
1493 		mutex_exit(&lofi_lock);
1494 		error = copy_out_lofi_ioctl(klip, ulip, ioctl_flag);
1495 		free_lofi_ioctl(klip);
1496 		return (error);
1497 	case LOFI_GET_MINOR:
1498 		mutex_enter(&lofi_lock);
1499 		klip->li_minor = file_to_minor(klip->li_filename);
1500 		mutex_exit(&lofi_lock);
1501 		if (klip->li_minor == 0) {
1502 			free_lofi_ioctl(klip);
1503 			return (ENOENT);
1504 		}
1505 		error = copy_out_lofi_ioctl(klip, ulip, ioctl_flag);
1506 		free_lofi_ioctl(klip);
1507 		return (error);
1508 	case LOFI_CHECK_COMPRESSED:
1509 		mutex_enter(&lofi_lock);
1510 		klip->li_minor = file_to_minor(klip->li_filename);
1511 		mutex_exit(&lofi_lock);
1512 		if (klip->li_minor == 0) {
1513 			free_lofi_ioctl(klip);
1514 			return (ENOENT);
1515 		}
1516 		mutex_enter(&lofi_lock);
1517 		lsp = ddi_get_soft_state(lofi_statep, klip->li_minor);
1518 		if (lsp == NULL) {
1519 			mutex_exit(&lofi_lock);
1520 			free_lofi_ioctl(klip);
1521 			return (ENXIO);
1522 		}
1523 		ASSERT(strcmp(klip->li_filename, lsp->ls_filename) == 0);
1524 
1525 		(void) strlcpy(klip->li_algorithm, lsp->ls_comp_algorithm,
1526 		    sizeof (klip->li_algorithm));
1527 		mutex_exit(&lofi_lock);
1528 		error = copy_out_lofi_ioctl(klip, ulip, ioctl_flag);
1529 		free_lofi_ioctl(klip);
1530 		return (error);
1531 	default:
1532 		free_lofi_ioctl(klip);
1533 		return (EINVAL);
1534 	}
1535 
1536 }
1537 
1538 static int
1539 lofi_ioctl(dev_t dev, int cmd, intptr_t arg, int flag, cred_t *credp,
1540     int *rvalp)
1541 {
1542 	int	error;
1543 	enum dkio_state dkstate;
1544 	struct lofi_state *lsp;
1545 	minor_t	minor;
1546 
1547 #ifdef lint
1548 	credp = credp;
1549 #endif
1550 
1551 	minor = getminor(dev);
1552 	/* lofi ioctls only apply to the master device */
1553 	if (minor == 0) {
1554 		struct lofi_ioctl *lip = (struct lofi_ioctl *)arg;
1555 
1556 		/*
1557 		 * the query command only need read-access - i.e., normal
1558 		 * users are allowed to do those on the ctl device as
1559 		 * long as they can open it read-only.
1560 		 */
1561 		switch (cmd) {
1562 		case LOFI_MAP_FILE:
1563 			if ((flag & FWRITE) == 0)
1564 				return (EPERM);
1565 			return (lofi_map_file(dev, lip, 1, rvalp, credp, flag));
1566 		case LOFI_MAP_FILE_MINOR:
1567 			if ((flag & FWRITE) == 0)
1568 				return (EPERM);
1569 			return (lofi_map_file(dev, lip, 0, rvalp, credp, flag));
1570 		case LOFI_UNMAP_FILE:
1571 			if ((flag & FWRITE) == 0)
1572 				return (EPERM);
1573 			return (lofi_unmap_file(dev, lip, 1, credp, flag));
1574 		case LOFI_UNMAP_FILE_MINOR:
1575 			if ((flag & FWRITE) == 0)
1576 				return (EPERM);
1577 			return (lofi_unmap_file(dev, lip, 0, credp, flag));
1578 		case LOFI_GET_FILENAME:
1579 			return (lofi_get_info(dev, lip, LOFI_GET_FILENAME,
1580 			    credp, flag));
1581 		case LOFI_GET_MINOR:
1582 			return (lofi_get_info(dev, lip, LOFI_GET_MINOR,
1583 			    credp, flag));
1584 		case LOFI_GET_MAXMINOR:
1585 			error = ddi_copyout(&lofi_max_files, &lip->li_minor,
1586 			    sizeof (lofi_max_files), flag);
1587 			if (error)
1588 				return (EFAULT);
1589 			return (0);
1590 		case LOFI_CHECK_COMPRESSED:
1591 			return (lofi_get_info(dev, lip, LOFI_CHECK_COMPRESSED,
1592 			    credp, flag));
1593 		default:
1594 			break;
1595 		}
1596 	}
1597 
1598 	lsp = ddi_get_soft_state(lofi_statep, minor);
1599 	if (lsp == NULL)
1600 		return (ENXIO);
1601 
1602 	/*
1603 	 * We explicitly allow DKIOCSTATE, but all other ioctls should fail with
1604 	 * EIO as if the device was no longer present.
1605 	 */
1606 	if (lsp->ls_vp == NULL && cmd != DKIOCSTATE)
1607 		return (EIO);
1608 
1609 	/* these are for faking out utilities like newfs */
1610 	switch (cmd) {
1611 	case DKIOCGVTOC:
1612 		switch (ddi_model_convert_from(flag & FMODELS)) {
1613 		case DDI_MODEL_ILP32: {
1614 			struct vtoc32 vtoc32;
1615 
1616 			vtoctovtoc32(lsp->ls_vtoc, vtoc32);
1617 			if (ddi_copyout(&vtoc32, (void *)arg,
1618 			    sizeof (struct vtoc32), flag))
1619 				return (EFAULT);
1620 				break;
1621 			}
1622 
1623 		case DDI_MODEL_NONE:
1624 			if (ddi_copyout(&lsp->ls_vtoc, (void *)arg,
1625 			    sizeof (struct vtoc), flag))
1626 				return (EFAULT);
1627 			break;
1628 		}
1629 		return (0);
1630 	case DKIOCINFO:
1631 		error = ddi_copyout(&lsp->ls_ci, (void *)arg,
1632 		    sizeof (struct dk_cinfo), flag);
1633 		if (error)
1634 			return (EFAULT);
1635 		return (0);
1636 	case DKIOCG_VIRTGEOM:
1637 	case DKIOCG_PHYGEOM:
1638 	case DKIOCGGEOM:
1639 		error = ddi_copyout(&lsp->ls_dkg, (void *)arg,
1640 		    sizeof (struct dk_geom), flag);
1641 		if (error)
1642 			return (EFAULT);
1643 		return (0);
1644 	case DKIOCSTATE:
1645 		/*
1646 		 * Normally, lofi devices are always in the INSERTED state.  If
1647 		 * a device is forcefully unmapped, then the device transitions
1648 		 * to the DKIO_DEV_GONE state.
1649 		 */
1650 		if (ddi_copyin((void *)arg, &dkstate, sizeof (dkstate),
1651 		    flag) != 0)
1652 			return (EFAULT);
1653 
1654 		mutex_enter(&lsp->ls_vp_lock);
1655 		while ((dkstate == DKIO_INSERTED && lsp->ls_vp != NULL) ||
1656 		    (dkstate == DKIO_DEV_GONE && lsp->ls_vp == NULL)) {
1657 			/*
1658 			 * By virtue of having the device open, we know that
1659 			 * 'lsp' will remain valid when we return.
1660 			 */
1661 			if (!cv_wait_sig(&lsp->ls_vp_cv,
1662 			    &lsp->ls_vp_lock)) {
1663 				mutex_exit(&lsp->ls_vp_lock);
1664 				return (EINTR);
1665 			}
1666 		}
1667 
1668 		dkstate = (lsp->ls_vp != NULL ? DKIO_INSERTED : DKIO_DEV_GONE);
1669 		mutex_exit(&lsp->ls_vp_lock);
1670 
1671 		if (ddi_copyout(&dkstate, (void *)arg,
1672 		    sizeof (dkstate), flag) != 0)
1673 			return (EFAULT);
1674 		return (0);
1675 	default:
1676 		return (ENOTTY);
1677 	}
1678 }
1679 
1680 static struct cb_ops lofi_cb_ops = {
1681 	lofi_open,		/* open */
1682 	lofi_close,		/* close */
1683 	lofi_strategy,		/* strategy */
1684 	nodev,			/* print */
1685 	nodev,			/* dump */
1686 	lofi_read,		/* read */
1687 	lofi_write,		/* write */
1688 	lofi_ioctl,		/* ioctl */
1689 	nodev,			/* devmap */
1690 	nodev,			/* mmap */
1691 	nodev,			/* segmap */
1692 	nochpoll,		/* poll */
1693 	ddi_prop_op,		/* prop_op */
1694 	0,			/* streamtab  */
1695 	D_64BIT | D_NEW | D_MP,	/* Driver compatibility flag */
1696 	CB_REV,
1697 	lofi_aread,
1698 	lofi_awrite
1699 };
1700 
1701 static struct dev_ops lofi_ops = {
1702 	DEVO_REV,		/* devo_rev, */
1703 	0,			/* refcnt  */
1704 	lofi_info,		/* info */
1705 	nulldev,		/* identify */
1706 	nulldev,		/* probe */
1707 	lofi_attach,		/* attach */
1708 	lofi_detach,		/* detach */
1709 	nodev,			/* reset */
1710 	&lofi_cb_ops,		/* driver operations */
1711 	NULL			/* no bus operations */
1712 };
1713 
1714 static struct modldrv modldrv = {
1715 	&mod_driverops,
1716 	"loopback file driver (%I%)",
1717 	&lofi_ops,
1718 };
1719 
1720 static struct modlinkage modlinkage = {
1721 	MODREV_1,
1722 	&modldrv,
1723 	NULL
1724 };
1725 
1726 int
1727 _init(void)
1728 {
1729 	int error;
1730 
1731 	error = ddi_soft_state_init(&lofi_statep,
1732 	    sizeof (struct lofi_state), 0);
1733 	if (error)
1734 		return (error);
1735 
1736 	mutex_init(&lofi_lock, NULL, MUTEX_DRIVER, NULL);
1737 	error = mod_install(&modlinkage);
1738 	if (error) {
1739 		mutex_destroy(&lofi_lock);
1740 		ddi_soft_state_fini(&lofi_statep);
1741 	}
1742 
1743 	return (error);
1744 }
1745 
1746 int
1747 _fini(void)
1748 {
1749 	int	error;
1750 
1751 	if (lofi_busy())
1752 		return (EBUSY);
1753 
1754 	error = mod_remove(&modlinkage);
1755 	if (error)
1756 		return (error);
1757 
1758 	mutex_destroy(&lofi_lock);
1759 	ddi_soft_state_fini(&lofi_statep);
1760 
1761 	return (error);
1762 }
1763 
1764 int
1765 _info(struct modinfo *modinfop)
1766 {
1767 	return (mod_info(&modlinkage, modinfop));
1768 }
1769