1 /*- 2 * Copyright (c) 2004 Max Khon 3 * Copyright (c) 2014 Juniper Networks, Inc. 4 * All rights reserved. 5 * 6 * Redistribution and use in source and binary forms, with or without 7 * modification, are permitted provided that the following conditions 8 * are met: 9 * 1. Redistributions of source code must retain the above copyright 10 * notice, this list of conditions and the following disclaimer. 11 * 2. Redistributions in binary form must reproduce the above copyright 12 * notice, this list of conditions and the following disclaimer in the 13 * documentation and/or other materials provided with the distribution. 14 * 15 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 16 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 17 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 18 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 19 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 20 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 21 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 22 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 23 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 24 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 25 * SUCH DAMAGE. 26 */ 27 28 #include <sys/cdefs.h> 29 __FBSDID("$FreeBSD$"); 30 31 #include <sys/param.h> 32 #include <sys/bio.h> 33 #include <sys/endian.h> 34 #include <sys/errno.h> 35 #include <sys/kernel.h> 36 #include <sys/lock.h> 37 #include <sys/mutex.h> 38 #include <sys/malloc.h> 39 #include <sys/sysctl.h> 40 #include <sys/systm.h> 41 #include <sys/zlib.h> 42 43 #include <geom/geom.h> 44 45 FEATURE(geom_uzip, "GEOM uzip read-only compressed disks support"); 46 47 #undef GEOM_UZIP_DEBUG 48 #ifdef GEOM_UZIP_DEBUG 49 #define DPRINTF(a) printf a 50 #else 51 #define DPRINTF(a) 52 #endif 53 54 static MALLOC_DEFINE(M_GEOM_UZIP, "geom_uzip", "GEOM UZIP data structures"); 55 56 #define UZIP_CLASS_NAME "UZIP" 57 58 /* 59 * Maximum allowed valid block size (to prevent foot-shooting) 60 */ 61 #define MAX_BLKSZ (MAXPHYS - MAXPHYS / 1000 - 12) 62 63 /* 64 * Integer values (block size, number of blocks, offsets) 65 * are stored in big-endian (network) order on disk and struct cloop_header 66 * and in native order in struct g_uzip_softc 67 */ 68 69 #define CLOOP_MAGIC_LEN 128 70 static char CLOOP_MAGIC_START[] = "#!/bin/sh\n"; 71 72 struct cloop_header { 73 char magic[CLOOP_MAGIC_LEN]; /* cloop magic */ 74 uint32_t blksz; /* block size */ 75 uint32_t nblocks; /* number of blocks */ 76 }; 77 78 struct g_uzip_softc { 79 uint32_t blksz; /* block size */ 80 uint32_t nblocks; /* number of blocks */ 81 uint64_t *offsets; 82 83 struct mtx last_mtx; 84 uint32_t last_blk; /* last blk no */ 85 char *last_buf; /* last blk data */ 86 int req_total; /* total requests */ 87 int req_cached; /* cached requests */ 88 }; 89 90 static void g_uzip_done(struct bio *bp); 91 92 static void 93 g_uzip_softc_free(struct g_uzip_softc *sc, struct g_geom *gp) 94 { 95 96 if (gp != NULL) { 97 DPRINTF(("%s: %d requests, %d cached\n", 98 gp->name, sc->req_total, sc->req_cached)); 99 } 100 if (sc->offsets != NULL) { 101 free(sc->offsets, M_GEOM_UZIP); 102 sc->offsets = NULL; 103 } 104 mtx_destroy(&sc->last_mtx); 105 free(sc->last_buf, M_GEOM_UZIP); 106 free(sc, M_GEOM_UZIP); 107 } 108 109 static void * 110 z_alloc(void *nil, u_int type, u_int size) 111 { 112 void *ptr; 113 114 ptr = malloc(type * size, M_GEOM_UZIP, M_NOWAIT); 115 116 return (ptr); 117 } 118 119 static void 120 z_free(void *nil, void *ptr) 121 { 122 123 free(ptr, M_GEOM_UZIP); 124 } 125 126 static int 127 g_uzip_cached(struct g_geom *gp, struct bio *bp) 128 { 129 struct g_uzip_softc *sc; 130 off_t ofs; 131 size_t blk, blkofs, usz; 132 133 sc = gp->softc; 134 ofs = bp->bio_offset + bp->bio_completed; 135 blk = ofs / sc->blksz; 136 mtx_lock(&sc->last_mtx); 137 if (blk == sc->last_blk) { 138 blkofs = ofs % sc->blksz; 139 usz = sc->blksz - blkofs; 140 if (bp->bio_resid < usz) 141 usz = bp->bio_resid; 142 memcpy(bp->bio_data + bp->bio_completed, sc->last_buf + blkofs, 143 usz); 144 sc->req_cached++; 145 mtx_unlock(&sc->last_mtx); 146 147 DPRINTF(("%s/%s: %p: offset=%jd: got %jd bytes from cache\n", 148 __func__, gp->name, bp, (intmax_t)ofs, (intmax_t)usz)); 149 150 bp->bio_completed += usz; 151 bp->bio_resid -= usz; 152 153 if (bp->bio_resid == 0) { 154 g_io_deliver(bp, 0); 155 return (1); 156 } 157 } else 158 mtx_unlock(&sc->last_mtx); 159 160 return (0); 161 } 162 163 static int 164 g_uzip_request(struct g_geom *gp, struct bio *bp) 165 { 166 struct g_uzip_softc *sc; 167 struct bio *bp2; 168 struct g_consumer *cp; 169 struct g_provider *pp; 170 off_t ofs; 171 size_t start_blk, end_blk; 172 173 if (g_uzip_cached(gp, bp) != 0) 174 return (1); 175 176 sc = gp->softc; 177 178 bp2 = g_clone_bio(bp); 179 if (bp2 == NULL) { 180 g_io_deliver(bp, ENOMEM); 181 return (1); 182 } 183 bp2->bio_done = g_uzip_done; 184 185 cp = LIST_FIRST(&gp->consumer); 186 pp = cp->provider; 187 188 ofs = bp->bio_offset + bp->bio_completed; 189 start_blk = ofs / sc->blksz; 190 KASSERT(start_blk < sc->nblocks, ("start_blk out of range")); 191 end_blk = (ofs + bp->bio_resid + sc->blksz - 1) / sc->blksz; 192 KASSERT(end_blk <= sc->nblocks, ("end_blk out of range")); 193 194 DPRINTF(("%s/%s: %p: start=%u (%jd), end=%u (%jd)\n", 195 __func__, gp->name, bp, 196 (u_int)start_blk, (intmax_t)sc->offsets[start_blk], 197 (u_int)end_blk, (intmax_t)sc->offsets[end_blk])); 198 199 bp2->bio_offset = sc->offsets[start_blk] - 200 sc->offsets[start_blk] % pp->sectorsize; 201 while (1) { 202 bp2->bio_length = sc->offsets[end_blk] - bp2->bio_offset; 203 bp2->bio_length = (bp2->bio_length + pp->sectorsize - 1) / 204 pp->sectorsize * pp->sectorsize; 205 if (bp2->bio_length <= MAXPHYS) 206 break; 207 208 end_blk--; 209 } 210 211 bp2->bio_data = malloc(bp2->bio_length, M_GEOM_UZIP, M_NOWAIT); 212 if (bp2->bio_data == NULL) { 213 g_destroy_bio(bp2); 214 g_io_deliver(bp, ENOMEM); 215 return (1); 216 } 217 218 DPRINTF(("%s/%s: %p: reading %jd bytes from offset %jd\n", 219 __func__, gp->name, bp, 220 (intmax_t)bp2->bio_length, (intmax_t)bp2->bio_offset)); 221 222 g_io_request(bp2, cp); 223 return (0); 224 } 225 226 static void 227 g_uzip_done(struct bio *bp) 228 { 229 z_stream zs; 230 struct bio *bp2; 231 struct g_provider *pp; 232 struct g_consumer *cp; 233 struct g_geom *gp; 234 struct g_uzip_softc *sc; 235 char *data, *data2; 236 off_t ofs; 237 size_t blk, blkofs, len, ulen; 238 239 bp2 = bp->bio_parent; 240 gp = bp2->bio_to->geom; 241 sc = gp->softc; 242 243 cp = LIST_FIRST(&gp->consumer); 244 pp = cp->provider; 245 246 bp2->bio_error = bp->bio_error; 247 if (bp2->bio_error != 0) 248 goto done; 249 250 /* Make sure there's forward progress. */ 251 if (bp->bio_completed == 0) { 252 bp2->bio_error = ECANCELED; 253 goto done; 254 } 255 256 zs.zalloc = z_alloc; 257 zs.zfree = z_free; 258 if (inflateInit(&zs) != Z_OK) { 259 bp2->bio_error = EILSEQ; 260 goto done; 261 } 262 263 ofs = bp2->bio_offset + bp2->bio_completed; 264 blk = ofs / sc->blksz; 265 blkofs = ofs % sc->blksz; 266 data = bp->bio_data + sc->offsets[blk] % pp->sectorsize; 267 data2 = bp2->bio_data + bp2->bio_completed; 268 while (bp->bio_completed && bp2->bio_resid) { 269 ulen = MIN(sc->blksz - blkofs, bp2->bio_resid); 270 len = sc->offsets[blk + 1] - sc->offsets[blk]; 271 DPRINTF(("%s/%s: %p/%ju: data2=%p, ulen=%u, data=%p, len=%u\n", 272 __func__, gp->name, gp, bp->bio_completed, 273 data2, (u_int)ulen, data, (u_int)len)); 274 if (len == 0) { 275 /* All zero block: no cache update */ 276 bzero(data2, ulen); 277 } else if (len <= bp->bio_completed) { 278 zs.next_in = data; 279 zs.avail_in = len; 280 zs.next_out = sc->last_buf; 281 zs.avail_out = sc->blksz; 282 mtx_lock(&sc->last_mtx); 283 if (inflate(&zs, Z_FINISH) != Z_STREAM_END) { 284 sc->last_blk = -1; 285 mtx_unlock(&sc->last_mtx); 286 inflateEnd(&zs); 287 bp2->bio_error = EILSEQ; 288 goto done; 289 } 290 sc->last_blk = blk; 291 memcpy(data2, sc->last_buf + blkofs, ulen); 292 mtx_unlock(&sc->last_mtx); 293 if (inflateReset(&zs) != Z_OK) { 294 inflateEnd(&zs); 295 bp2->bio_error = EILSEQ; 296 goto done; 297 } 298 data += len; 299 } else 300 break; 301 302 data2 += ulen; 303 bp2->bio_completed += ulen; 304 bp2->bio_resid -= ulen; 305 bp->bio_completed -= len; 306 blkofs = 0; 307 blk++; 308 } 309 310 if (inflateEnd(&zs) != Z_OK) 311 bp2->bio_error = EILSEQ; 312 313 done: 314 /* Finish processing the request. */ 315 free(bp->bio_data, M_GEOM_UZIP); 316 g_destroy_bio(bp); 317 if (bp2->bio_error != 0 || bp2->bio_resid == 0) 318 g_io_deliver(bp2, bp2->bio_error); 319 else 320 g_uzip_request(gp, bp2); 321 } 322 323 static void 324 g_uzip_start(struct bio *bp) 325 { 326 struct g_provider *pp; 327 struct g_geom *gp; 328 struct g_uzip_softc *sc; 329 330 pp = bp->bio_to; 331 gp = pp->geom; 332 333 DPRINTF(("%s/%s: %p: cmd=%d, offset=%jd, length=%jd, buffer=%p\n", 334 __func__, gp->name, bp, bp->bio_cmd, (intmax_t)bp->bio_offset, 335 (intmax_t)bp->bio_length, bp->bio_data)); 336 337 sc = gp->softc; 338 sc->req_total++; 339 340 if (bp->bio_cmd != BIO_READ) { 341 g_io_deliver(bp, EOPNOTSUPP); 342 return; 343 } 344 345 bp->bio_resid = bp->bio_length; 346 bp->bio_completed = 0; 347 348 g_uzip_request(gp, bp); 349 } 350 351 static void 352 g_uzip_orphan(struct g_consumer *cp) 353 { 354 struct g_geom *gp; 355 356 g_trace(G_T_TOPOLOGY, "%s(%p/%s)", __func__, cp, cp->provider->name); 357 g_topology_assert(); 358 359 gp = cp->geom; 360 g_uzip_softc_free(gp->softc, gp); 361 gp->softc = NULL; 362 g_wither_geom(gp, ENXIO); 363 } 364 365 static int 366 g_uzip_access(struct g_provider *pp, int dr, int dw, int de) 367 { 368 struct g_geom *gp; 369 struct g_consumer *cp; 370 371 gp = pp->geom; 372 cp = LIST_FIRST(&gp->consumer); 373 KASSERT (cp != NULL, ("g_uzip_access but no consumer")); 374 375 if (cp->acw + dw > 0) 376 return (EROFS); 377 378 return (g_access(cp, dr, dw, de)); 379 } 380 381 static void 382 g_uzip_spoiled(struct g_consumer *cp) 383 { 384 struct g_geom *gp; 385 386 gp = cp->geom; 387 g_trace(G_T_TOPOLOGY, "%s(%p/%s)", __func__, cp, gp->name); 388 g_topology_assert(); 389 390 g_uzip_softc_free(gp->softc, gp); 391 gp->softc = NULL; 392 g_wither_geom(gp, ENXIO); 393 } 394 395 static struct g_geom * 396 g_uzip_taste(struct g_class *mp, struct g_provider *pp, int flags) 397 { 398 int error; 399 uint32_t i, total_offsets, offsets_read, blk; 400 void *buf; 401 struct cloop_header *header; 402 struct g_consumer *cp; 403 struct g_geom *gp; 404 struct g_provider *pp2; 405 struct g_uzip_softc *sc; 406 407 g_trace(G_T_TOPOLOGY, "%s(%s,%s)", __func__, mp->name, pp->name); 408 g_topology_assert(); 409 410 /* Skip providers that are already open for writing. */ 411 if (pp->acw > 0) 412 return (NULL); 413 414 buf = NULL; 415 416 /* 417 * Create geom instance. 418 */ 419 gp = g_new_geomf(mp, "%s.uzip", pp->name); 420 cp = g_new_consumer(gp); 421 error = g_attach(cp, pp); 422 if (error == 0) 423 error = g_access(cp, 1, 0, 0); 424 if (error) { 425 g_detach(cp); 426 g_destroy_consumer(cp); 427 g_destroy_geom(gp); 428 return (NULL); 429 } 430 g_topology_unlock(); 431 432 /* 433 * Read cloop header, look for CLOOP magic, perform 434 * other validity checks. 435 */ 436 DPRINTF(("%s: media sectorsize %u, mediasize %jd\n", 437 gp->name, pp->sectorsize, (intmax_t)pp->mediasize)); 438 buf = g_read_data(cp, 0, pp->sectorsize, NULL); 439 if (buf == NULL) 440 goto err; 441 header = (struct cloop_header *) buf; 442 if (strncmp(header->magic, CLOOP_MAGIC_START, 443 sizeof(CLOOP_MAGIC_START) - 1) != 0) { 444 DPRINTF(("%s: no CLOOP magic\n", gp->name)); 445 goto err; 446 } 447 if (header->magic[0x0b] != 'V' || header->magic[0x0c] < '2') { 448 DPRINTF(("%s: image version too old\n", gp->name)); 449 goto err; 450 } 451 452 /* 453 * Initialize softc and read offsets. 454 */ 455 sc = malloc(sizeof(*sc), M_GEOM_UZIP, M_WAITOK | M_ZERO); 456 gp->softc = sc; 457 sc->blksz = ntohl(header->blksz); 458 sc->nblocks = ntohl(header->nblocks); 459 if (sc->blksz % 512 != 0) { 460 printf("%s: block size (%u) should be multiple of 512.\n", 461 gp->name, sc->blksz); 462 goto err; 463 } 464 if (sc->blksz > MAX_BLKSZ) { 465 printf("%s: block size (%u) should not be larger than %d.\n", 466 gp->name, sc->blksz, MAX_BLKSZ); 467 } 468 total_offsets = sc->nblocks + 1; 469 if (sizeof(struct cloop_header) + 470 total_offsets * sizeof(uint64_t) > pp->mediasize) { 471 printf("%s: media too small for %u blocks\n", 472 gp->name, sc->nblocks); 473 goto err; 474 } 475 sc->offsets = malloc( 476 total_offsets * sizeof(uint64_t), M_GEOM_UZIP, M_WAITOK); 477 offsets_read = MIN(total_offsets, 478 (pp->sectorsize - sizeof(*header)) / sizeof(uint64_t)); 479 for (i = 0; i < offsets_read; i++) 480 sc->offsets[i] = be64toh(((uint64_t *) (header + 1))[i]); 481 DPRINTF(("%s: %u offsets in the first sector\n", 482 gp->name, offsets_read)); 483 for (blk = 1; offsets_read < total_offsets; blk++) { 484 uint32_t nread; 485 486 free(buf, M_GEOM); 487 buf = g_read_data( 488 cp, blk * pp->sectorsize, pp->sectorsize, NULL); 489 if (buf == NULL) 490 goto err; 491 nread = MIN(total_offsets - offsets_read, 492 pp->sectorsize / sizeof(uint64_t)); 493 DPRINTF(("%s: %u offsets read from sector %d\n", 494 gp->name, nread, blk)); 495 for (i = 0; i < nread; i++) { 496 sc->offsets[offsets_read + i] = 497 be64toh(((uint64_t *) buf)[i]); 498 } 499 offsets_read += nread; 500 } 501 free(buf, M_GEOM); 502 DPRINTF(("%s: done reading offsets\n", gp->name)); 503 mtx_init(&sc->last_mtx, "geom_uzip cache", NULL, MTX_DEF); 504 sc->last_blk = -1; 505 sc->last_buf = malloc(sc->blksz, M_GEOM_UZIP, M_WAITOK); 506 sc->req_total = 0; 507 sc->req_cached = 0; 508 509 g_topology_lock(); 510 pp2 = g_new_providerf(gp, "%s", gp->name); 511 pp2->sectorsize = 512; 512 pp2->mediasize = (off_t)sc->nblocks * sc->blksz; 513 pp2->stripesize = pp->stripesize; 514 pp2->stripeoffset = pp->stripeoffset; 515 g_error_provider(pp2, 0); 516 g_access(cp, -1, 0, 0); 517 518 DPRINTF(("%s: taste ok (%d, %jd), (%d, %d), %x\n", 519 gp->name, 520 pp2->sectorsize, (intmax_t)pp2->mediasize, 521 pp2->stripeoffset, pp2->stripesize, pp2->flags)); 522 DPRINTF(("%s: %u x %u blocks\n", gp->name, sc->nblocks, sc->blksz)); 523 return (gp); 524 525 err: 526 g_topology_lock(); 527 g_access(cp, -1, 0, 0); 528 if (buf != NULL) 529 free(buf, M_GEOM); 530 if (gp->softc != NULL) { 531 g_uzip_softc_free(gp->softc, NULL); 532 gp->softc = NULL; 533 } 534 g_detach(cp); 535 g_destroy_consumer(cp); 536 g_destroy_geom(gp); 537 538 return (NULL); 539 } 540 541 static int 542 g_uzip_destroy_geom(struct gctl_req *req, struct g_class *mp, struct g_geom *gp) 543 { 544 struct g_provider *pp; 545 546 g_trace(G_T_TOPOLOGY, "%s(%s, %s)", __func__, mp->name, gp->name); 547 g_topology_assert(); 548 549 if (gp->softc == NULL) { 550 DPRINTF(("%s(%s): gp->softc == NULL\n", __func__, gp->name)); 551 return (ENXIO); 552 } 553 554 KASSERT(gp != NULL, ("NULL geom")); 555 pp = LIST_FIRST(&gp->provider); 556 KASSERT(pp != NULL, ("NULL provider")); 557 if (pp->acr > 0 || pp->acw > 0 || pp->ace > 0) 558 return (EBUSY); 559 560 g_uzip_softc_free(gp->softc, gp); 561 gp->softc = NULL; 562 g_wither_geom(gp, ENXIO); 563 564 return (0); 565 } 566 567 static struct g_class g_uzip_class = { 568 .name = UZIP_CLASS_NAME, 569 .version = G_VERSION, 570 .taste = g_uzip_taste, 571 .destroy_geom = g_uzip_destroy_geom, 572 573 .start = g_uzip_start, 574 .orphan = g_uzip_orphan, 575 .access = g_uzip_access, 576 .spoiled = g_uzip_spoiled, 577 }; 578 579 DECLARE_GEOM_CLASS(g_uzip_class, g_uzip); 580 MODULE_DEPEND(g_uzip, zlib, 1, 1, 1); 581