1 /*- 2 * Implementation of SCSI Direct Access Peripheral driver for CAM. 3 * 4 * Copyright (c) 1997 Justin T. Gibbs. 5 * All rights reserved. 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice, this list of conditions, and the following disclaimer, 12 * without modification, immediately at the beginning of the file. 13 * 2. The name of the author may not be used to endorse or promote products 14 * derived from this software without specific prior written permission. 15 * 16 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 17 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 18 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 19 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR 20 * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 21 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 22 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 23 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 24 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 25 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 26 * SUCH DAMAGE. 27 */ 28 29 #include <sys/cdefs.h> 30 __FBSDID("$FreeBSD$"); 31 32 #include <sys/param.h> 33 34 #ifdef _KERNEL 35 #include <sys/systm.h> 36 #include <sys/kernel.h> 37 #include <sys/bio.h> 38 #include <sys/sysctl.h> 39 #include <sys/taskqueue.h> 40 #include <sys/lock.h> 41 #include <sys/mutex.h> 42 #include <sys/conf.h> 43 #include <sys/devicestat.h> 44 #include <sys/eventhandler.h> 45 #include <sys/malloc.h> 46 #include <sys/cons.h> 47 #include <sys/endian.h> 48 #include <sys/proc.h> 49 #include <sys/sbuf.h> 50 #include <geom/geom.h> 51 #include <geom/geom_disk.h> 52 #endif /* _KERNEL */ 53 54 #ifndef _KERNEL 55 #include <stdio.h> 56 #include <string.h> 57 #endif /* _KERNEL */ 58 59 #include <cam/cam.h> 60 #include <cam/cam_ccb.h> 61 #include <cam/cam_periph.h> 62 #include <cam/cam_xpt_periph.h> 63 #include <cam/cam_sim.h> 64 #include <cam/cam_iosched.h> 65 66 #include <cam/scsi/scsi_message.h> 67 #include <cam/scsi/scsi_da.h> 68 69 #ifdef _KERNEL 70 /* 71 * Note that there are probe ordering dependencies here. The order isn't 72 * controlled by this enumeration, but by explicit state transitions in 73 * dastart() and dadone(). Here are some of the dependencies: 74 * 75 * 1. RC should come first, before RC16, unless there is evidence that RC16 76 * is supported. 77 * 2. BDC needs to come before any of the ATA probes, or the ZONE probe. 78 * 3. The ATA probes should go in this order: 79 * ATA -> LOGDIR -> IDDIR -> SUP -> ATA_ZONE 80 */ 81 typedef enum { 82 DA_STATE_PROBE_RC, 83 DA_STATE_PROBE_RC16, 84 DA_STATE_PROBE_LBP, 85 DA_STATE_PROBE_BLK_LIMITS, 86 DA_STATE_PROBE_BDC, 87 DA_STATE_PROBE_ATA, 88 DA_STATE_PROBE_ATA_LOGDIR, 89 DA_STATE_PROBE_ATA_IDDIR, 90 DA_STATE_PROBE_ATA_SUP, 91 DA_STATE_PROBE_ATA_ZONE, 92 DA_STATE_PROBE_ZONE, 93 DA_STATE_NORMAL 94 } da_state; 95 96 typedef enum { 97 DA_FLAG_PACK_INVALID = 0x000001, 98 DA_FLAG_NEW_PACK = 0x000002, 99 DA_FLAG_PACK_LOCKED = 0x000004, 100 DA_FLAG_PACK_REMOVABLE = 0x000008, 101 DA_FLAG_NEED_OTAG = 0x000020, 102 DA_FLAG_WAS_OTAG = 0x000040, 103 DA_FLAG_RETRY_UA = 0x000080, 104 DA_FLAG_OPEN = 0x000100, 105 DA_FLAG_SCTX_INIT = 0x000200, 106 DA_FLAG_CAN_RC16 = 0x000400, 107 DA_FLAG_PROBED = 0x000800, 108 DA_FLAG_DIRTY = 0x001000, 109 DA_FLAG_ANNOUNCED = 0x002000, 110 DA_FLAG_CAN_ATA_DMA = 0x004000, 111 DA_FLAG_CAN_ATA_LOG = 0x008000, 112 DA_FLAG_CAN_ATA_IDLOG = 0x010000, 113 DA_FLAG_CAN_ATA_SUPCAP = 0x020000, 114 DA_FLAG_CAN_ATA_ZONE = 0x040000 115 } da_flags; 116 117 typedef enum { 118 DA_Q_NONE = 0x00, 119 DA_Q_NO_SYNC_CACHE = 0x01, 120 DA_Q_NO_6_BYTE = 0x02, 121 DA_Q_NO_PREVENT = 0x04, 122 DA_Q_4K = 0x08, 123 DA_Q_NO_RC16 = 0x10, 124 DA_Q_NO_UNMAP = 0x20, 125 DA_Q_RETRY_BUSY = 0x40, 126 DA_Q_SMR_DM = 0x80, 127 DA_Q_STRICT_UNMAP = 0x100 128 } da_quirks; 129 130 #define DA_Q_BIT_STRING \ 131 "\020" \ 132 "\001NO_SYNC_CACHE" \ 133 "\002NO_6_BYTE" \ 134 "\003NO_PREVENT" \ 135 "\0044K" \ 136 "\005NO_RC16" \ 137 "\006NO_UNMAP" \ 138 "\007RETRY_BUSY" \ 139 "\010SMR_DM" \ 140 "\011STRICT_UNMAP" 141 142 typedef enum { 143 DA_CCB_PROBE_RC = 0x01, 144 DA_CCB_PROBE_RC16 = 0x02, 145 DA_CCB_PROBE_LBP = 0x03, 146 DA_CCB_PROBE_BLK_LIMITS = 0x04, 147 DA_CCB_PROBE_BDC = 0x05, 148 DA_CCB_PROBE_ATA = 0x06, 149 DA_CCB_BUFFER_IO = 0x07, 150 DA_CCB_DUMP = 0x0A, 151 DA_CCB_DELETE = 0x0B, 152 DA_CCB_TUR = 0x0C, 153 DA_CCB_PROBE_ZONE = 0x0D, 154 DA_CCB_PROBE_ATA_LOGDIR = 0x0E, 155 DA_CCB_PROBE_ATA_IDDIR = 0x0F, 156 DA_CCB_PROBE_ATA_SUP = 0x10, 157 DA_CCB_PROBE_ATA_ZONE = 0x11, 158 DA_CCB_TYPE_MASK = 0x1F, 159 DA_CCB_RETRY_UA = 0x20 160 } da_ccb_state; 161 162 /* 163 * Order here is important for method choice 164 * 165 * We prefer ATA_TRIM as tests run against a Sandforce 2281 SSD attached to 166 * LSI 2008 (mps) controller (FW: v12, Drv: v14) resulted 20% quicker deletes 167 * using ATA_TRIM than the corresponding UNMAP results for a real world mysql 168 * import taking 5mins. 169 * 170 */ 171 typedef enum { 172 DA_DELETE_NONE, 173 DA_DELETE_DISABLE, 174 DA_DELETE_ATA_TRIM, 175 DA_DELETE_UNMAP, 176 DA_DELETE_WS16, 177 DA_DELETE_WS10, 178 DA_DELETE_ZERO, 179 DA_DELETE_MIN = DA_DELETE_ATA_TRIM, 180 DA_DELETE_MAX = DA_DELETE_ZERO 181 } da_delete_methods; 182 183 /* 184 * For SCSI, host managed drives show up as a separate device type. For 185 * ATA, host managed drives also have a different device signature. 186 * XXX KDM figure out the ATA host managed signature. 187 */ 188 typedef enum { 189 DA_ZONE_NONE = 0x00, 190 DA_ZONE_DRIVE_MANAGED = 0x01, 191 DA_ZONE_HOST_AWARE = 0x02, 192 DA_ZONE_HOST_MANAGED = 0x03 193 } da_zone_mode; 194 195 /* 196 * We distinguish between these interface cases in addition to the drive type: 197 * o ATA drive behind a SCSI translation layer that knows about ZBC/ZAC 198 * o ATA drive behind a SCSI translation layer that does not know about 199 * ZBC/ZAC, and so needs to be managed via ATA passthrough. In this 200 * case, we would need to share the ATA code with the ada(4) driver. 201 * o SCSI drive. 202 */ 203 typedef enum { 204 DA_ZONE_IF_SCSI, 205 DA_ZONE_IF_ATA_PASS, 206 DA_ZONE_IF_ATA_SAT, 207 } da_zone_interface; 208 209 typedef enum { 210 DA_ZONE_FLAG_RZ_SUP = 0x0001, 211 DA_ZONE_FLAG_OPEN_SUP = 0x0002, 212 DA_ZONE_FLAG_CLOSE_SUP = 0x0004, 213 DA_ZONE_FLAG_FINISH_SUP = 0x0008, 214 DA_ZONE_FLAG_RWP_SUP = 0x0010, 215 DA_ZONE_FLAG_SUP_MASK = (DA_ZONE_FLAG_RZ_SUP | 216 DA_ZONE_FLAG_OPEN_SUP | 217 DA_ZONE_FLAG_CLOSE_SUP | 218 DA_ZONE_FLAG_FINISH_SUP | 219 DA_ZONE_FLAG_RWP_SUP), 220 DA_ZONE_FLAG_URSWRZ = 0x0020, 221 DA_ZONE_FLAG_OPT_SEQ_SET = 0x0040, 222 DA_ZONE_FLAG_OPT_NONSEQ_SET = 0x0080, 223 DA_ZONE_FLAG_MAX_SEQ_SET = 0x0100, 224 DA_ZONE_FLAG_SET_MASK = (DA_ZONE_FLAG_OPT_SEQ_SET | 225 DA_ZONE_FLAG_OPT_NONSEQ_SET | 226 DA_ZONE_FLAG_MAX_SEQ_SET) 227 } da_zone_flags; 228 229 static struct da_zone_desc { 230 da_zone_flags value; 231 const char *desc; 232 } da_zone_desc_table[] = { 233 {DA_ZONE_FLAG_RZ_SUP, "Report Zones" }, 234 {DA_ZONE_FLAG_OPEN_SUP, "Open" }, 235 {DA_ZONE_FLAG_CLOSE_SUP, "Close" }, 236 {DA_ZONE_FLAG_FINISH_SUP, "Finish" }, 237 {DA_ZONE_FLAG_RWP_SUP, "Reset Write Pointer" }, 238 }; 239 240 typedef void da_delete_func_t (struct cam_periph *periph, union ccb *ccb, 241 struct bio *bp); 242 static da_delete_func_t da_delete_trim; 243 static da_delete_func_t da_delete_unmap; 244 static da_delete_func_t da_delete_ws; 245 246 static const void * da_delete_functions[] = { 247 NULL, 248 NULL, 249 da_delete_trim, 250 da_delete_unmap, 251 da_delete_ws, 252 da_delete_ws, 253 da_delete_ws 254 }; 255 256 static const char *da_delete_method_names[] = 257 { "NONE", "DISABLE", "ATA_TRIM", "UNMAP", "WS16", "WS10", "ZERO" }; 258 static const char *da_delete_method_desc[] = 259 { "NONE", "DISABLED", "ATA TRIM", "UNMAP", "WRITE SAME(16) with UNMAP", 260 "WRITE SAME(10) with UNMAP", "ZERO" }; 261 262 /* Offsets into our private area for storing information */ 263 #define ccb_state ppriv_field0 264 #define ccb_bp ppriv_ptr1 265 266 struct disk_params { 267 u_int8_t heads; 268 u_int32_t cylinders; 269 u_int8_t secs_per_track; 270 u_int32_t secsize; /* Number of bytes/sector */ 271 u_int64_t sectors; /* total number sectors */ 272 u_int stripesize; 273 u_int stripeoffset; 274 }; 275 276 #define UNMAP_RANGE_MAX 0xffffffff 277 #define UNMAP_HEAD_SIZE 8 278 #define UNMAP_RANGE_SIZE 16 279 #define UNMAP_MAX_RANGES 2048 /* Protocol Max is 4095 */ 280 #define UNMAP_BUF_SIZE ((UNMAP_MAX_RANGES * UNMAP_RANGE_SIZE) + \ 281 UNMAP_HEAD_SIZE) 282 283 #define WS10_MAX_BLKS 0xffff 284 #define WS16_MAX_BLKS 0xffffffff 285 #define ATA_TRIM_MAX_RANGES ((UNMAP_BUF_SIZE / \ 286 (ATA_DSM_RANGE_SIZE * ATA_DSM_BLK_SIZE)) * ATA_DSM_BLK_SIZE) 287 288 #define DA_WORK_TUR (1 << 16) 289 290 struct da_softc { 291 struct cam_iosched_softc *cam_iosched; 292 struct bio_queue_head delete_run_queue; 293 LIST_HEAD(, ccb_hdr) pending_ccbs; 294 int refcount; /* Active xpt_action() calls */ 295 da_state state; 296 da_flags flags; 297 da_quirks quirks; 298 int minimum_cmd_size; 299 int error_inject; 300 int trim_max_ranges; 301 int delete_available; /* Delete methods possibly available */ 302 da_zone_mode zone_mode; 303 da_zone_interface zone_interface; 304 da_zone_flags zone_flags; 305 struct ata_gp_log_dir ata_logdir; 306 int valid_logdir_len; 307 struct ata_identify_log_pages ata_iddir; 308 int valid_iddir_len; 309 uint64_t optimal_seq_zones; 310 uint64_t optimal_nonseq_zones; 311 uint64_t max_seq_zones; 312 u_int maxio; 313 uint32_t unmap_max_ranges; 314 uint32_t unmap_max_lba; /* Max LBAs in UNMAP req */ 315 uint32_t unmap_gran; 316 uint32_t unmap_gran_align; 317 uint64_t ws_max_blks; 318 da_delete_methods delete_method_pref; 319 da_delete_methods delete_method; 320 da_delete_func_t *delete_func; 321 int unmappedio; 322 int rotating; 323 struct disk_params params; 324 struct disk *disk; 325 union ccb saved_ccb; 326 struct task sysctl_task; 327 struct sysctl_ctx_list sysctl_ctx; 328 struct sysctl_oid *sysctl_tree; 329 struct callout sendordered_c; 330 uint64_t wwpn; 331 uint8_t unmap_buf[UNMAP_BUF_SIZE]; 332 struct scsi_read_capacity_data_long rcaplong; 333 struct callout mediapoll_c; 334 #ifdef CAM_IO_STATS 335 struct sysctl_ctx_list sysctl_stats_ctx; 336 struct sysctl_oid *sysctl_stats_tree; 337 u_int errors; 338 u_int timeouts; 339 u_int invalidations; 340 #endif 341 }; 342 343 #define dadeleteflag(softc, delete_method, enable) \ 344 if (enable) { \ 345 softc->delete_available |= (1 << delete_method); \ 346 } else { \ 347 softc->delete_available &= ~(1 << delete_method); \ 348 } 349 350 struct da_quirk_entry { 351 struct scsi_inquiry_pattern inq_pat; 352 da_quirks quirks; 353 }; 354 355 static const char quantum[] = "QUANTUM"; 356 static const char microp[] = "MICROP"; 357 358 static struct da_quirk_entry da_quirk_table[] = 359 { 360 /* SPI, FC devices */ 361 { 362 /* 363 * Fujitsu M2513A MO drives. 364 * Tested devices: M2513A2 firmware versions 1200 & 1300. 365 * (dip switch selects whether T_DIRECT or T_OPTICAL device) 366 * Reported by: W.Scholten <whs@xs4all.nl> 367 */ 368 {T_DIRECT, SIP_MEDIA_REMOVABLE, "FUJITSU", "M2513A", "*"}, 369 /*quirks*/ DA_Q_NO_SYNC_CACHE 370 }, 371 { 372 /* See above. */ 373 {T_OPTICAL, SIP_MEDIA_REMOVABLE, "FUJITSU", "M2513A", "*"}, 374 /*quirks*/ DA_Q_NO_SYNC_CACHE 375 }, 376 { 377 /* 378 * This particular Fujitsu drive doesn't like the 379 * synchronize cache command. 380 * Reported by: Tom Jackson <toj@gorilla.net> 381 */ 382 {T_DIRECT, SIP_MEDIA_FIXED, "FUJITSU", "M2954*", "*"}, 383 /*quirks*/ DA_Q_NO_SYNC_CACHE 384 }, 385 { 386 /* 387 * This drive doesn't like the synchronize cache command 388 * either. Reported by: Matthew Jacob <mjacob@feral.com> 389 * in NetBSD PR kern/6027, August 24, 1998. 390 */ 391 {T_DIRECT, SIP_MEDIA_FIXED, microp, "2217*", "*"}, 392 /*quirks*/ DA_Q_NO_SYNC_CACHE 393 }, 394 { 395 /* 396 * This drive doesn't like the synchronize cache command 397 * either. Reported by: Hellmuth Michaelis (hm@kts.org) 398 * (PR 8882). 399 */ 400 {T_DIRECT, SIP_MEDIA_FIXED, microp, "2112*", "*"}, 401 /*quirks*/ DA_Q_NO_SYNC_CACHE 402 }, 403 { 404 /* 405 * Doesn't like the synchronize cache command. 406 * Reported by: Blaz Zupan <blaz@gold.amis.net> 407 */ 408 {T_DIRECT, SIP_MEDIA_FIXED, "NEC", "D3847*", "*"}, 409 /*quirks*/ DA_Q_NO_SYNC_CACHE 410 }, 411 { 412 /* 413 * Doesn't like the synchronize cache command. 414 * Reported by: Blaz Zupan <blaz@gold.amis.net> 415 */ 416 {T_DIRECT, SIP_MEDIA_FIXED, quantum, "MAVERICK 540S", "*"}, 417 /*quirks*/ DA_Q_NO_SYNC_CACHE 418 }, 419 { 420 /* 421 * Doesn't like the synchronize cache command. 422 */ 423 {T_DIRECT, SIP_MEDIA_FIXED, quantum, "LPS525S", "*"}, 424 /*quirks*/ DA_Q_NO_SYNC_CACHE 425 }, 426 { 427 /* 428 * Doesn't like the synchronize cache command. 429 * Reported by: walter@pelissero.de 430 */ 431 {T_DIRECT, SIP_MEDIA_FIXED, quantum, "LPS540S", "*"}, 432 /*quirks*/ DA_Q_NO_SYNC_CACHE 433 }, 434 { 435 /* 436 * Doesn't work correctly with 6 byte reads/writes. 437 * Returns illegal request, and points to byte 9 of the 438 * 6-byte CDB. 439 * Reported by: Adam McDougall <bsdx@spawnet.com> 440 */ 441 {T_DIRECT, SIP_MEDIA_FIXED, quantum, "VIKING 4*", "*"}, 442 /*quirks*/ DA_Q_NO_6_BYTE 443 }, 444 { 445 /* See above. */ 446 {T_DIRECT, SIP_MEDIA_FIXED, quantum, "VIKING 2*", "*"}, 447 /*quirks*/ DA_Q_NO_6_BYTE 448 }, 449 { 450 /* 451 * Doesn't like the synchronize cache command. 452 * Reported by: walter@pelissero.de 453 */ 454 {T_DIRECT, SIP_MEDIA_FIXED, "CONNER", "CP3500*", "*"}, 455 /*quirks*/ DA_Q_NO_SYNC_CACHE 456 }, 457 { 458 /* 459 * The CISS RAID controllers do not support SYNC_CACHE 460 */ 461 {T_DIRECT, SIP_MEDIA_FIXED, "COMPAQ", "RAID*", "*"}, 462 /*quirks*/ DA_Q_NO_SYNC_CACHE 463 }, 464 { 465 /* 466 * The STEC SSDs sometimes hang on UNMAP. 467 */ 468 {T_DIRECT, SIP_MEDIA_FIXED, "STEC", "*", "*"}, 469 /*quirks*/ DA_Q_NO_UNMAP 470 }, 471 { 472 /* 473 * VMware returns BUSY status when storage has transient 474 * connectivity problems, so better wait. 475 * Also VMware returns odd errors on misaligned UNMAPs. 476 */ 477 {T_DIRECT, SIP_MEDIA_FIXED, "VMware*", "*", "*"}, 478 /*quirks*/ DA_Q_RETRY_BUSY | DA_Q_STRICT_UNMAP 479 }, 480 /* USB mass storage devices supported by umass(4) */ 481 { 482 /* 483 * EXATELECOM (Sigmatel) i-Bead 100/105 USB Flash MP3 Player 484 * PR: kern/51675 485 */ 486 {T_DIRECT, SIP_MEDIA_REMOVABLE, "EXATEL", "i-BEAD10*", "*"}, 487 /*quirks*/ DA_Q_NO_SYNC_CACHE 488 }, 489 { 490 /* 491 * Power Quotient Int. (PQI) USB flash key 492 * PR: kern/53067 493 */ 494 {T_DIRECT, SIP_MEDIA_REMOVABLE, "Generic*", "USB Flash Disk*", 495 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 496 }, 497 { 498 /* 499 * Creative Nomad MUVO mp3 player (USB) 500 * PR: kern/53094 501 */ 502 {T_DIRECT, SIP_MEDIA_REMOVABLE, "CREATIVE", "NOMAD_MUVO", "*"}, 503 /*quirks*/ DA_Q_NO_SYNC_CACHE|DA_Q_NO_PREVENT 504 }, 505 { 506 /* 507 * Jungsoft NEXDISK USB flash key 508 * PR: kern/54737 509 */ 510 {T_DIRECT, SIP_MEDIA_REMOVABLE, "JUNGSOFT", "NEXDISK*", "*"}, 511 /*quirks*/ DA_Q_NO_SYNC_CACHE 512 }, 513 { 514 /* 515 * FreeDik USB Mini Data Drive 516 * PR: kern/54786 517 */ 518 {T_DIRECT, SIP_MEDIA_REMOVABLE, "FreeDik*", "Mini Data Drive", 519 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 520 }, 521 { 522 /* 523 * Sigmatel USB Flash MP3 Player 524 * PR: kern/57046 525 */ 526 {T_DIRECT, SIP_MEDIA_REMOVABLE, "SigmaTel", "MSCN", "*"}, 527 /*quirks*/ DA_Q_NO_SYNC_CACHE|DA_Q_NO_PREVENT 528 }, 529 { 530 /* 531 * Neuros USB Digital Audio Computer 532 * PR: kern/63645 533 */ 534 {T_DIRECT, SIP_MEDIA_REMOVABLE, "NEUROS", "dig. audio comp.", 535 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 536 }, 537 { 538 /* 539 * SEAGRAND NP-900 MP3 Player 540 * PR: kern/64563 541 */ 542 {T_DIRECT, SIP_MEDIA_REMOVABLE, "SEAGRAND", "NP-900*", "*"}, 543 /*quirks*/ DA_Q_NO_SYNC_CACHE|DA_Q_NO_PREVENT 544 }, 545 { 546 /* 547 * iRiver iFP MP3 player (with UMS Firmware) 548 * PR: kern/54881, i386/63941, kern/66124 549 */ 550 {T_DIRECT, SIP_MEDIA_REMOVABLE, "iRiver", "iFP*", "*"}, 551 /*quirks*/ DA_Q_NO_SYNC_CACHE 552 }, 553 { 554 /* 555 * Frontier Labs NEX IA+ Digital Audio Player, rev 1.10/0.01 556 * PR: kern/70158 557 */ 558 {T_DIRECT, SIP_MEDIA_REMOVABLE, "FL" , "Nex*", "*"}, 559 /*quirks*/ DA_Q_NO_SYNC_CACHE 560 }, 561 { 562 /* 563 * ZICPlay USB MP3 Player with FM 564 * PR: kern/75057 565 */ 566 {T_DIRECT, SIP_MEDIA_REMOVABLE, "ACTIONS*" , "USB DISK*", "*"}, 567 /*quirks*/ DA_Q_NO_SYNC_CACHE 568 }, 569 { 570 /* 571 * TEAC USB floppy mechanisms 572 */ 573 {T_DIRECT, SIP_MEDIA_REMOVABLE, "TEAC" , "FD-05*", "*"}, 574 /*quirks*/ DA_Q_NO_SYNC_CACHE 575 }, 576 { 577 /* 578 * Kingston DataTraveler II+ USB Pen-Drive. 579 * Reported by: Pawel Jakub Dawidek <pjd@FreeBSD.org> 580 */ 581 {T_DIRECT, SIP_MEDIA_REMOVABLE, "Kingston" , "DataTraveler II+", 582 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 583 }, 584 { 585 /* 586 * USB DISK Pro PMAP 587 * Reported by: jhs 588 * PR: usb/96381 589 */ 590 {T_DIRECT, SIP_MEDIA_REMOVABLE, " ", "USB DISK Pro", "PMAP"}, 591 /*quirks*/ DA_Q_NO_SYNC_CACHE 592 }, 593 { 594 /* 595 * Motorola E398 Mobile Phone (TransFlash memory card). 596 * Reported by: Wojciech A. Koszek <dunstan@FreeBSD.czest.pl> 597 * PR: usb/89889 598 */ 599 {T_DIRECT, SIP_MEDIA_REMOVABLE, "Motorola" , "Motorola Phone", 600 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 601 }, 602 { 603 /* 604 * Qware BeatZkey! Pro 605 * PR: usb/79164 606 */ 607 {T_DIRECT, SIP_MEDIA_REMOVABLE, "GENERIC", "USB DISK DEVICE", 608 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 609 }, 610 { 611 /* 612 * Time DPA20B 1GB MP3 Player 613 * PR: usb/81846 614 */ 615 {T_DIRECT, SIP_MEDIA_REMOVABLE, "USB2.0*", "(FS) FLASH DISK*", 616 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 617 }, 618 { 619 /* 620 * Samsung USB key 128Mb 621 * PR: usb/90081 622 */ 623 {T_DIRECT, SIP_MEDIA_REMOVABLE, "USB-DISK", "FreeDik-FlashUsb", 624 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 625 }, 626 { 627 /* 628 * Kingston DataTraveler 2.0 USB Flash memory. 629 * PR: usb/89196 630 */ 631 {T_DIRECT, SIP_MEDIA_REMOVABLE, "Kingston", "DataTraveler 2.0", 632 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 633 }, 634 { 635 /* 636 * Creative MUVO Slim mp3 player (USB) 637 * PR: usb/86131 638 */ 639 {T_DIRECT, SIP_MEDIA_REMOVABLE, "CREATIVE", "MuVo Slim", 640 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE|DA_Q_NO_PREVENT 641 }, 642 { 643 /* 644 * United MP5512 Portable MP3 Player (2-in-1 USB DISK/MP3) 645 * PR: usb/80487 646 */ 647 {T_DIRECT, SIP_MEDIA_REMOVABLE, "Generic*", "MUSIC DISK", 648 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 649 }, 650 { 651 /* 652 * SanDisk Micro Cruzer 128MB 653 * PR: usb/75970 654 */ 655 {T_DIRECT, SIP_MEDIA_REMOVABLE, "SanDisk" , "Micro Cruzer", 656 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 657 }, 658 { 659 /* 660 * TOSHIBA TransMemory USB sticks 661 * PR: kern/94660 662 */ 663 {T_DIRECT, SIP_MEDIA_REMOVABLE, "TOSHIBA", "TransMemory", 664 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 665 }, 666 { 667 /* 668 * PNY USB 3.0 Flash Drives 669 */ 670 {T_DIRECT, SIP_MEDIA_REMOVABLE, "PNY", "USB 3.0 FD*", 671 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE | DA_Q_NO_RC16 672 }, 673 { 674 /* 675 * PNY USB Flash keys 676 * PR: usb/75578, usb/72344, usb/65436 677 */ 678 {T_DIRECT, SIP_MEDIA_REMOVABLE, "*" , "USB DISK*", 679 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 680 }, 681 { 682 /* 683 * Genesys 6-in-1 Card Reader 684 * PR: usb/94647 685 */ 686 {T_DIRECT, SIP_MEDIA_REMOVABLE, "Generic*", "STORAGE DEVICE*", 687 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 688 }, 689 { 690 /* 691 * Rekam Digital CAMERA 692 * PR: usb/98713 693 */ 694 {T_DIRECT, SIP_MEDIA_REMOVABLE, "CAMERA*", "4MP-9J6*", 695 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 696 }, 697 { 698 /* 699 * iRiver H10 MP3 player 700 * PR: usb/102547 701 */ 702 {T_DIRECT, SIP_MEDIA_REMOVABLE, "iriver", "H10*", 703 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 704 }, 705 { 706 /* 707 * iRiver U10 MP3 player 708 * PR: usb/92306 709 */ 710 {T_DIRECT, SIP_MEDIA_REMOVABLE, "iriver", "U10*", 711 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 712 }, 713 { 714 /* 715 * X-Micro Flash Disk 716 * PR: usb/96901 717 */ 718 {T_DIRECT, SIP_MEDIA_REMOVABLE, "X-Micro", "Flash Disk", 719 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 720 }, 721 { 722 /* 723 * EasyMP3 EM732X USB 2.0 Flash MP3 Player 724 * PR: usb/96546 725 */ 726 {T_DIRECT, SIP_MEDIA_REMOVABLE, "EM732X", "MP3 Player*", 727 "1.00"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 728 }, 729 { 730 /* 731 * Denver MP3 player 732 * PR: usb/107101 733 */ 734 {T_DIRECT, SIP_MEDIA_REMOVABLE, "DENVER", "MP3 PLAYER", 735 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 736 }, 737 { 738 /* 739 * Philips USB Key Audio KEY013 740 * PR: usb/68412 741 */ 742 {T_DIRECT, SIP_MEDIA_REMOVABLE, "PHILIPS", "Key*", "*"}, 743 /*quirks*/ DA_Q_NO_SYNC_CACHE | DA_Q_NO_PREVENT 744 }, 745 { 746 /* 747 * JNC MP3 Player 748 * PR: usb/94439 749 */ 750 {T_DIRECT, SIP_MEDIA_REMOVABLE, "JNC*" , "MP3 Player*", 751 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 752 }, 753 { 754 /* 755 * SAMSUNG MP0402H 756 * PR: usb/108427 757 */ 758 {T_DIRECT, SIP_MEDIA_FIXED, "SAMSUNG", "MP0402H", "*"}, 759 /*quirks*/ DA_Q_NO_SYNC_CACHE 760 }, 761 { 762 /* 763 * I/O Magic USB flash - Giga Bank 764 * PR: usb/108810 765 */ 766 {T_DIRECT, SIP_MEDIA_FIXED, "GS-Magic", "stor*", "*"}, 767 /*quirks*/ DA_Q_NO_SYNC_CACHE 768 }, 769 { 770 /* 771 * JoyFly 128mb USB Flash Drive 772 * PR: 96133 773 */ 774 {T_DIRECT, SIP_MEDIA_REMOVABLE, "USB 2.0", "Flash Disk*", 775 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 776 }, 777 { 778 /* 779 * ChipsBnk usb stick 780 * PR: 103702 781 */ 782 {T_DIRECT, SIP_MEDIA_REMOVABLE, "ChipsBnk", "USB*", 783 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 784 }, 785 { 786 /* 787 * Storcase (Kingston) InfoStation IFS FC2/SATA-R 201A 788 * PR: 129858 789 */ 790 {T_DIRECT, SIP_MEDIA_FIXED, "IFS", "FC2/SATA-R*", 791 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 792 }, 793 { 794 /* 795 * Samsung YP-U3 mp3-player 796 * PR: 125398 797 */ 798 {T_DIRECT, SIP_MEDIA_REMOVABLE, "Samsung", "YP-U3", 799 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 800 }, 801 { 802 {T_DIRECT, SIP_MEDIA_REMOVABLE, "Netac", "OnlyDisk*", 803 "2000"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 804 }, 805 { 806 /* 807 * Sony Cyber-Shot DSC cameras 808 * PR: usb/137035 809 */ 810 {T_DIRECT, SIP_MEDIA_REMOVABLE, "Sony", "Sony DSC", "*"}, 811 /*quirks*/ DA_Q_NO_SYNC_CACHE | DA_Q_NO_PREVENT 812 }, 813 { 814 {T_DIRECT, SIP_MEDIA_REMOVABLE, "Kingston", "DataTraveler G3", 815 "1.00"}, /*quirks*/ DA_Q_NO_PREVENT 816 }, 817 { 818 /* At least several Transcent USB sticks lie on RC16. */ 819 {T_DIRECT, SIP_MEDIA_REMOVABLE, "JetFlash", "Transcend*", 820 "*"}, /*quirks*/ DA_Q_NO_RC16 821 }, 822 { 823 /* 824 * I-O Data USB Flash Disk 825 * PR: usb/211716 826 */ 827 {T_DIRECT, SIP_MEDIA_REMOVABLE, "I-O DATA", "USB Flash Disk*", 828 "*"}, /*quirks*/ DA_Q_NO_RC16 829 }, 830 /* ATA/SATA devices over SAS/USB/... */ 831 { 832 /* Hitachi Advanced Format (4k) drives */ 833 { T_DIRECT, SIP_MEDIA_FIXED, "Hitachi", "H??????????E3*", "*" }, 834 /*quirks*/DA_Q_4K 835 }, 836 { 837 /* Micron Advanced Format (4k) drives */ 838 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "Micron 5100 MTFDDAK*", "*" }, 839 /*quirks*/DA_Q_4K 840 }, 841 { 842 /* Samsung Advanced Format (4k) drives */ 843 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "SAMSUNG HD155UI*", "*" }, 844 /*quirks*/DA_Q_4K 845 }, 846 { 847 /* Samsung Advanced Format (4k) drives */ 848 { T_DIRECT, SIP_MEDIA_FIXED, "SAMSUNG", "HD155UI*", "*" }, 849 /*quirks*/DA_Q_4K 850 }, 851 { 852 /* Samsung Advanced Format (4k) drives */ 853 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "SAMSUNG HD204UI*", "*" }, 854 /*quirks*/DA_Q_4K 855 }, 856 { 857 /* Samsung Advanced Format (4k) drives */ 858 { T_DIRECT, SIP_MEDIA_FIXED, "SAMSUNG", "HD204UI*", "*" }, 859 /*quirks*/DA_Q_4K 860 }, 861 { 862 /* Seagate Barracuda Green Advanced Format (4k) drives */ 863 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "ST????DL*", "*" }, 864 /*quirks*/DA_Q_4K 865 }, 866 { 867 /* Seagate Barracuda Green Advanced Format (4k) drives */ 868 { T_DIRECT, SIP_MEDIA_FIXED, "ST????DL", "*", "*" }, 869 /*quirks*/DA_Q_4K 870 }, 871 { 872 /* Seagate Barracuda Green Advanced Format (4k) drives */ 873 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "ST???DM*", "*" }, 874 /*quirks*/DA_Q_4K 875 }, 876 { 877 /* Seagate Barracuda Green Advanced Format (4k) drives */ 878 { T_DIRECT, SIP_MEDIA_FIXED, "ST???DM*", "*", "*" }, 879 /*quirks*/DA_Q_4K 880 }, 881 { 882 /* Seagate Barracuda Green Advanced Format (4k) drives */ 883 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "ST????DM*", "*" }, 884 /*quirks*/DA_Q_4K 885 }, 886 { 887 /* Seagate Barracuda Green Advanced Format (4k) drives */ 888 { T_DIRECT, SIP_MEDIA_FIXED, "ST????DM", "*", "*" }, 889 /*quirks*/DA_Q_4K 890 }, 891 { 892 /* Seagate Momentus Advanced Format (4k) drives */ 893 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "ST9500423AS*", "*" }, 894 /*quirks*/DA_Q_4K 895 }, 896 { 897 /* Seagate Momentus Advanced Format (4k) drives */ 898 { T_DIRECT, SIP_MEDIA_FIXED, "ST950042", "3AS*", "*" }, 899 /*quirks*/DA_Q_4K 900 }, 901 { 902 /* Seagate Momentus Advanced Format (4k) drives */ 903 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "ST9500424AS*", "*" }, 904 /*quirks*/DA_Q_4K 905 }, 906 { 907 /* Seagate Momentus Advanced Format (4k) drives */ 908 { T_DIRECT, SIP_MEDIA_FIXED, "ST950042", "4AS*", "*" }, 909 /*quirks*/DA_Q_4K 910 }, 911 { 912 /* Seagate Momentus Advanced Format (4k) drives */ 913 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "ST9640423AS*", "*" }, 914 /*quirks*/DA_Q_4K 915 }, 916 { 917 /* Seagate Momentus Advanced Format (4k) drives */ 918 { T_DIRECT, SIP_MEDIA_FIXED, "ST964042", "3AS*", "*" }, 919 /*quirks*/DA_Q_4K 920 }, 921 { 922 /* Seagate Momentus Advanced Format (4k) drives */ 923 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "ST9640424AS*", "*" }, 924 /*quirks*/DA_Q_4K 925 }, 926 { 927 /* Seagate Momentus Advanced Format (4k) drives */ 928 { T_DIRECT, SIP_MEDIA_FIXED, "ST964042", "4AS*", "*" }, 929 /*quirks*/DA_Q_4K 930 }, 931 { 932 /* Seagate Momentus Advanced Format (4k) drives */ 933 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "ST9750420AS*", "*" }, 934 /*quirks*/DA_Q_4K 935 }, 936 { 937 /* Seagate Momentus Advanced Format (4k) drives */ 938 { T_DIRECT, SIP_MEDIA_FIXED, "ST975042", "0AS*", "*" }, 939 /*quirks*/DA_Q_4K 940 }, 941 { 942 /* Seagate Momentus Advanced Format (4k) drives */ 943 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "ST9750422AS*", "*" }, 944 /*quirks*/DA_Q_4K 945 }, 946 { 947 /* Seagate Momentus Advanced Format (4k) drives */ 948 { T_DIRECT, SIP_MEDIA_FIXED, "ST975042", "2AS*", "*" }, 949 /*quirks*/DA_Q_4K 950 }, 951 { 952 /* Seagate Momentus Advanced Format (4k) drives */ 953 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "ST9750423AS*", "*" }, 954 /*quirks*/DA_Q_4K 955 }, 956 { 957 /* Seagate Momentus Advanced Format (4k) drives */ 958 { T_DIRECT, SIP_MEDIA_FIXED, "ST975042", "3AS*", "*" }, 959 /*quirks*/DA_Q_4K 960 }, 961 { 962 /* Seagate Momentus Thin Advanced Format (4k) drives */ 963 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "ST???LT*", "*" }, 964 /*quirks*/DA_Q_4K 965 }, 966 { 967 /* Seagate Momentus Thin Advanced Format (4k) drives */ 968 { T_DIRECT, SIP_MEDIA_FIXED, "ST???LT*", "*", "*" }, 969 /*quirks*/DA_Q_4K 970 }, 971 { 972 /* WDC Caviar Green Advanced Format (4k) drives */ 973 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "WDC WD????RS*", "*" }, 974 /*quirks*/DA_Q_4K 975 }, 976 { 977 /* WDC Caviar Green Advanced Format (4k) drives */ 978 { T_DIRECT, SIP_MEDIA_FIXED, "WDC WD??", "??RS*", "*" }, 979 /*quirks*/DA_Q_4K 980 }, 981 { 982 /* WDC Caviar Green Advanced Format (4k) drives */ 983 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "WDC WD????RX*", "*" }, 984 /*quirks*/DA_Q_4K 985 }, 986 { 987 /* WDC Caviar Green Advanced Format (4k) drives */ 988 { T_DIRECT, SIP_MEDIA_FIXED, "WDC WD??", "??RX*", "*" }, 989 /*quirks*/DA_Q_4K 990 }, 991 { 992 /* WDC Caviar Green Advanced Format (4k) drives */ 993 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "WDC WD??????RS*", "*" }, 994 /*quirks*/DA_Q_4K 995 }, 996 { 997 /* WDC Caviar Green Advanced Format (4k) drives */ 998 { T_DIRECT, SIP_MEDIA_FIXED, "WDC WD??", "????RS*", "*" }, 999 /*quirks*/DA_Q_4K 1000 }, 1001 { 1002 /* WDC Caviar Green Advanced Format (4k) drives */ 1003 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "WDC WD??????RX*", "*" }, 1004 /*quirks*/DA_Q_4K 1005 }, 1006 { 1007 /* WDC Caviar Green Advanced Format (4k) drives */ 1008 { T_DIRECT, SIP_MEDIA_FIXED, "WDC WD??", "????RX*", "*" }, 1009 /*quirks*/DA_Q_4K 1010 }, 1011 { 1012 /* WDC Scorpio Black Advanced Format (4k) drives */ 1013 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "WDC WD???PKT*", "*" }, 1014 /*quirks*/DA_Q_4K 1015 }, 1016 { 1017 /* WDC Scorpio Black Advanced Format (4k) drives */ 1018 { T_DIRECT, SIP_MEDIA_FIXED, "WDC WD??", "?PKT*", "*" }, 1019 /*quirks*/DA_Q_4K 1020 }, 1021 { 1022 /* WDC Scorpio Black Advanced Format (4k) drives */ 1023 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "WDC WD?????PKT*", "*" }, 1024 /*quirks*/DA_Q_4K 1025 }, 1026 { 1027 /* WDC Scorpio Black Advanced Format (4k) drives */ 1028 { T_DIRECT, SIP_MEDIA_FIXED, "WDC WD??", "???PKT*", "*" }, 1029 /*quirks*/DA_Q_4K 1030 }, 1031 { 1032 /* WDC Scorpio Blue Advanced Format (4k) drives */ 1033 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "WDC WD???PVT*", "*" }, 1034 /*quirks*/DA_Q_4K 1035 }, 1036 { 1037 /* WDC Scorpio Blue Advanced Format (4k) drives */ 1038 { T_DIRECT, SIP_MEDIA_FIXED, "WDC WD??", "?PVT*", "*" }, 1039 /*quirks*/DA_Q_4K 1040 }, 1041 { 1042 /* WDC Scorpio Blue Advanced Format (4k) drives */ 1043 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "WDC WD?????PVT*", "*" }, 1044 /*quirks*/DA_Q_4K 1045 }, 1046 { 1047 /* WDC Scorpio Blue Advanced Format (4k) drives */ 1048 { T_DIRECT, SIP_MEDIA_FIXED, "WDC WD??", "???PVT*", "*" }, 1049 /*quirks*/DA_Q_4K 1050 }, 1051 { 1052 /* 1053 * Olympus FE-210 camera 1054 */ 1055 {T_DIRECT, SIP_MEDIA_REMOVABLE, "OLYMPUS", "FE210*", 1056 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 1057 }, 1058 { 1059 /* 1060 * LG UP3S MP3 player 1061 */ 1062 {T_DIRECT, SIP_MEDIA_REMOVABLE, "LG", "UP3S", 1063 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 1064 }, 1065 { 1066 /* 1067 * Laser MP3-2GA13 MP3 player 1068 */ 1069 {T_DIRECT, SIP_MEDIA_REMOVABLE, "USB 2.0", "(HS) Flash Disk", 1070 "*"}, /*quirks*/ DA_Q_NO_SYNC_CACHE 1071 }, 1072 { 1073 /* 1074 * LaCie external 250GB Hard drive des by Porsche 1075 * Submitted by: Ben Stuyts <ben@altesco.nl> 1076 * PR: 121474 1077 */ 1078 {T_DIRECT, SIP_MEDIA_FIXED, "SAMSUNG", "HM250JI", "*"}, 1079 /*quirks*/ DA_Q_NO_SYNC_CACHE 1080 }, 1081 /* SATA SSDs */ 1082 { 1083 /* 1084 * Corsair Force 2 SSDs 1085 * 4k optimised & trim only works in 4k requests + 4k aligned 1086 */ 1087 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "Corsair CSSD-F*", "*" }, 1088 /*quirks*/DA_Q_4K 1089 }, 1090 { 1091 /* 1092 * Corsair Force 3 SSDs 1093 * 4k optimised & trim only works in 4k requests + 4k aligned 1094 */ 1095 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "Corsair Force 3*", "*" }, 1096 /*quirks*/DA_Q_4K 1097 }, 1098 { 1099 /* 1100 * Corsair Neutron GTX SSDs 1101 * 4k optimised & trim only works in 4k requests + 4k aligned 1102 */ 1103 { T_DIRECT, SIP_MEDIA_FIXED, "*", "Corsair Neutron GTX*", "*" }, 1104 /*quirks*/DA_Q_4K 1105 }, 1106 { 1107 /* 1108 * Corsair Force GT & GS SSDs 1109 * 4k optimised & trim only works in 4k requests + 4k aligned 1110 */ 1111 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "Corsair Force G*", "*" }, 1112 /*quirks*/DA_Q_4K 1113 }, 1114 { 1115 /* 1116 * Crucial M4 SSDs 1117 * 4k optimised & trim only works in 4k requests + 4k aligned 1118 */ 1119 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "M4-CT???M4SSD2*", "*" }, 1120 /*quirks*/DA_Q_4K 1121 }, 1122 { 1123 /* 1124 * Crucial RealSSD C300 SSDs 1125 * 4k optimised 1126 */ 1127 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "C300-CTFDDAC???MAG*", 1128 "*" }, /*quirks*/DA_Q_4K 1129 }, 1130 { 1131 /* 1132 * Intel 320 Series SSDs 1133 * 4k optimised & trim only works in 4k requests + 4k aligned 1134 */ 1135 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "INTEL SSDSA2CW*", "*" }, 1136 /*quirks*/DA_Q_4K 1137 }, 1138 { 1139 /* 1140 * Intel 330 Series SSDs 1141 * 4k optimised & trim only works in 4k requests + 4k aligned 1142 */ 1143 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "INTEL SSDSC2CT*", "*" }, 1144 /*quirks*/DA_Q_4K 1145 }, 1146 { 1147 /* 1148 * Intel 510 Series SSDs 1149 * 4k optimised & trim only works in 4k requests + 4k aligned 1150 */ 1151 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "INTEL SSDSC2MH*", "*" }, 1152 /*quirks*/DA_Q_4K 1153 }, 1154 { 1155 /* 1156 * Intel 520 Series SSDs 1157 * 4k optimised & trim only works in 4k requests + 4k aligned 1158 */ 1159 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "INTEL SSDSC2BW*", "*" }, 1160 /*quirks*/DA_Q_4K 1161 }, 1162 { 1163 /* 1164 * Intel S3610 Series SSDs 1165 * 4k optimised & trim only works in 4k requests + 4k aligned 1166 */ 1167 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "INTEL SSDSC2BX*", "*" }, 1168 /*quirks*/DA_Q_4K 1169 }, 1170 { 1171 /* 1172 * Intel X25-M Series SSDs 1173 * 4k optimised & trim only works in 4k requests + 4k aligned 1174 */ 1175 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "INTEL SSDSA2M*", "*" }, 1176 /*quirks*/DA_Q_4K 1177 }, 1178 { 1179 /* 1180 * Kingston E100 Series SSDs 1181 * 4k optimised & trim only works in 4k requests + 4k aligned 1182 */ 1183 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "KINGSTON SE100S3*", "*" }, 1184 /*quirks*/DA_Q_4K 1185 }, 1186 { 1187 /* 1188 * Kingston HyperX 3k SSDs 1189 * 4k optimised & trim only works in 4k requests + 4k aligned 1190 */ 1191 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "KINGSTON SH103S3*", "*" }, 1192 /*quirks*/DA_Q_4K 1193 }, 1194 { 1195 /* 1196 * Marvell SSDs (entry taken from OpenSolaris) 1197 * 4k optimised & trim only works in 4k requests + 4k aligned 1198 */ 1199 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "MARVELL SD88SA02*", "*" }, 1200 /*quirks*/DA_Q_4K 1201 }, 1202 { 1203 /* 1204 * OCZ Agility 2 SSDs 1205 * 4k optimised & trim only works in 4k requests + 4k aligned 1206 */ 1207 { T_DIRECT, SIP_MEDIA_FIXED, "*", "OCZ-AGILITY2*", "*" }, 1208 /*quirks*/DA_Q_4K 1209 }, 1210 { 1211 /* 1212 * OCZ Agility 3 SSDs 1213 * 4k optimised & trim only works in 4k requests + 4k aligned 1214 */ 1215 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "OCZ-AGILITY3*", "*" }, 1216 /*quirks*/DA_Q_4K 1217 }, 1218 { 1219 /* 1220 * OCZ Deneva R Series SSDs 1221 * 4k optimised & trim only works in 4k requests + 4k aligned 1222 */ 1223 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "DENRSTE251M45*", "*" }, 1224 /*quirks*/DA_Q_4K 1225 }, 1226 { 1227 /* 1228 * OCZ Vertex 2 SSDs (inc pro series) 1229 * 4k optimised & trim only works in 4k requests + 4k aligned 1230 */ 1231 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "OCZ?VERTEX2*", "*" }, 1232 /*quirks*/DA_Q_4K 1233 }, 1234 { 1235 /* 1236 * OCZ Vertex 3 SSDs 1237 * 4k optimised & trim only works in 4k requests + 4k aligned 1238 */ 1239 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "OCZ-VERTEX3*", "*" }, 1240 /*quirks*/DA_Q_4K 1241 }, 1242 { 1243 /* 1244 * OCZ Vertex 4 SSDs 1245 * 4k optimised & trim only works in 4k requests + 4k aligned 1246 */ 1247 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "OCZ-VERTEX4*", "*" }, 1248 /*quirks*/DA_Q_4K 1249 }, 1250 { 1251 /* 1252 * Samsung 830 Series SSDs 1253 * 4k optimised & trim only works in 4k requests + 4k aligned 1254 */ 1255 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "SAMSUNG SSD 830 Series*", "*" }, 1256 /*quirks*/DA_Q_4K 1257 }, 1258 { 1259 /* 1260 * Samsung 840 SSDs 1261 * 4k optimised & trim only works in 4k requests + 4k aligned 1262 */ 1263 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "Samsung SSD 840*", "*" }, 1264 /*quirks*/DA_Q_4K 1265 }, 1266 { 1267 /* 1268 * Samsung 850 SSDs 1269 * 4k optimised & trim only works in 4k requests + 4k aligned 1270 */ 1271 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "Samsung SSD 850*", "*" }, 1272 /*quirks*/DA_Q_4K 1273 }, 1274 { 1275 /* 1276 * Samsung 843T Series SSDs (MZ7WD*) 1277 * Samsung PM851 Series SSDs (MZ7TE*) 1278 * Samsung PM853T Series SSDs (MZ7GE*) 1279 * Samsung SM863 Series SSDs (MZ7KM*) 1280 * 4k optimised 1281 */ 1282 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "SAMSUNG MZ7*", "*" }, 1283 /*quirks*/DA_Q_4K 1284 }, 1285 { 1286 /* 1287 * SuperTalent TeraDrive CT SSDs 1288 * 4k optimised & trim only works in 4k requests + 4k aligned 1289 */ 1290 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "FTM??CT25H*", "*" }, 1291 /*quirks*/DA_Q_4K 1292 }, 1293 { 1294 /* 1295 * XceedIOPS SATA SSDs 1296 * 4k optimised 1297 */ 1298 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "SG9XCS2D*", "*" }, 1299 /*quirks*/DA_Q_4K 1300 }, 1301 { 1302 /* 1303 * Hama Innostor USB-Stick 1304 */ 1305 { T_DIRECT, SIP_MEDIA_REMOVABLE, "Innostor", "Innostor*", "*" }, 1306 /*quirks*/DA_Q_NO_RC16 1307 }, 1308 { 1309 /* 1310 * Seagate Lamarr 8TB Shingled Magnetic Recording (SMR) 1311 * Drive Managed SATA hard drive. This drive doesn't report 1312 * in firmware that it is a drive managed SMR drive. 1313 */ 1314 { T_DIRECT, SIP_MEDIA_FIXED, "ATA", "ST8000AS0002*", "*" }, 1315 /*quirks*/DA_Q_SMR_DM 1316 }, 1317 { 1318 /* 1319 * MX-ES USB Drive by Mach Xtreme 1320 */ 1321 { T_DIRECT, SIP_MEDIA_REMOVABLE, "MX", "MXUB3*", "*"}, 1322 /*quirks*/DA_Q_NO_RC16 1323 }, 1324 }; 1325 1326 static disk_strategy_t dastrategy; 1327 static dumper_t dadump; 1328 static periph_init_t dainit; 1329 static void daasync(void *callback_arg, u_int32_t code, 1330 struct cam_path *path, void *arg); 1331 static void dasysctlinit(void *context, int pending); 1332 static int dasysctlsofttimeout(SYSCTL_HANDLER_ARGS); 1333 static int dacmdsizesysctl(SYSCTL_HANDLER_ARGS); 1334 static int dadeletemethodsysctl(SYSCTL_HANDLER_ARGS); 1335 static int dazonemodesysctl(SYSCTL_HANDLER_ARGS); 1336 static int dazonesupsysctl(SYSCTL_HANDLER_ARGS); 1337 static int dadeletemaxsysctl(SYSCTL_HANDLER_ARGS); 1338 static void dadeletemethodset(struct da_softc *softc, 1339 da_delete_methods delete_method); 1340 static off_t dadeletemaxsize(struct da_softc *softc, 1341 da_delete_methods delete_method); 1342 static void dadeletemethodchoose(struct da_softc *softc, 1343 da_delete_methods default_method); 1344 static void daprobedone(struct cam_periph *periph, union ccb *ccb); 1345 1346 static periph_ctor_t daregister; 1347 static periph_dtor_t dacleanup; 1348 static periph_start_t dastart; 1349 static periph_oninv_t daoninvalidate; 1350 static void dazonedone(struct cam_periph *periph, union ccb *ccb); 1351 static void dadone(struct cam_periph *periph, 1352 union ccb *done_ccb); 1353 static int daerror(union ccb *ccb, u_int32_t cam_flags, 1354 u_int32_t sense_flags); 1355 static void daprevent(struct cam_periph *periph, int action); 1356 static void dareprobe(struct cam_periph *periph); 1357 static void dasetgeom(struct cam_periph *periph, uint32_t block_len, 1358 uint64_t maxsector, 1359 struct scsi_read_capacity_data_long *rcaplong, 1360 size_t rcap_size); 1361 static timeout_t dasendorderedtag; 1362 static void dashutdown(void *arg, int howto); 1363 static timeout_t damediapoll; 1364 1365 #ifndef DA_DEFAULT_POLL_PERIOD 1366 #define DA_DEFAULT_POLL_PERIOD 3 1367 #endif 1368 1369 #ifndef DA_DEFAULT_TIMEOUT 1370 #define DA_DEFAULT_TIMEOUT 60 /* Timeout in seconds */ 1371 #endif 1372 1373 #ifndef DA_DEFAULT_SOFTTIMEOUT 1374 #define DA_DEFAULT_SOFTTIMEOUT 0 1375 #endif 1376 1377 #ifndef DA_DEFAULT_RETRY 1378 #define DA_DEFAULT_RETRY 4 1379 #endif 1380 1381 #ifndef DA_DEFAULT_SEND_ORDERED 1382 #define DA_DEFAULT_SEND_ORDERED 1 1383 #endif 1384 1385 static int da_poll_period = DA_DEFAULT_POLL_PERIOD; 1386 static int da_retry_count = DA_DEFAULT_RETRY; 1387 static int da_default_timeout = DA_DEFAULT_TIMEOUT; 1388 static sbintime_t da_default_softtimeout = DA_DEFAULT_SOFTTIMEOUT; 1389 static int da_send_ordered = DA_DEFAULT_SEND_ORDERED; 1390 1391 static SYSCTL_NODE(_kern_cam, OID_AUTO, da, CTLFLAG_RD, 0, 1392 "CAM Direct Access Disk driver"); 1393 SYSCTL_INT(_kern_cam_da, OID_AUTO, poll_period, CTLFLAG_RWTUN, 1394 &da_poll_period, 0, "Media polling period in seconds"); 1395 SYSCTL_INT(_kern_cam_da, OID_AUTO, retry_count, CTLFLAG_RWTUN, 1396 &da_retry_count, 0, "Normal I/O retry count"); 1397 SYSCTL_INT(_kern_cam_da, OID_AUTO, default_timeout, CTLFLAG_RWTUN, 1398 &da_default_timeout, 0, "Normal I/O timeout (in seconds)"); 1399 SYSCTL_INT(_kern_cam_da, OID_AUTO, send_ordered, CTLFLAG_RWTUN, 1400 &da_send_ordered, 0, "Send Ordered Tags"); 1401 1402 SYSCTL_PROC(_kern_cam_da, OID_AUTO, default_softtimeout, 1403 CTLTYPE_UINT | CTLFLAG_RW, NULL, 0, dasysctlsofttimeout, "I", 1404 "Soft I/O timeout (ms)"); 1405 TUNABLE_INT64("kern.cam.da.default_softtimeout", &da_default_softtimeout); 1406 1407 /* 1408 * DA_ORDEREDTAG_INTERVAL determines how often, relative 1409 * to the default timeout, we check to see whether an ordered 1410 * tagged transaction is appropriate to prevent simple tag 1411 * starvation. Since we'd like to ensure that there is at least 1412 * 1/2 of the timeout length left for a starved transaction to 1413 * complete after we've sent an ordered tag, we must poll at least 1414 * four times in every timeout period. This takes care of the worst 1415 * case where a starved transaction starts during an interval that 1416 * meets the requirement "don't send an ordered tag" test so it takes 1417 * us two intervals to determine that a tag must be sent. 1418 */ 1419 #ifndef DA_ORDEREDTAG_INTERVAL 1420 #define DA_ORDEREDTAG_INTERVAL 4 1421 #endif 1422 1423 static struct periph_driver dadriver = 1424 { 1425 dainit, "da", 1426 TAILQ_HEAD_INITIALIZER(dadriver.units), /* generation */ 0 1427 }; 1428 1429 PERIPHDRIVER_DECLARE(da, dadriver); 1430 1431 static MALLOC_DEFINE(M_SCSIDA, "scsi_da", "scsi_da buffers"); 1432 1433 static int 1434 daopen(struct disk *dp) 1435 { 1436 struct cam_periph *periph; 1437 struct da_softc *softc; 1438 int error; 1439 1440 periph = (struct cam_periph *)dp->d_drv1; 1441 if (cam_periph_acquire(periph) != CAM_REQ_CMP) { 1442 return (ENXIO); 1443 } 1444 1445 cam_periph_lock(periph); 1446 if ((error = cam_periph_hold(periph, PRIBIO|PCATCH)) != 0) { 1447 cam_periph_unlock(periph); 1448 cam_periph_release(periph); 1449 return (error); 1450 } 1451 1452 CAM_DEBUG(periph->path, CAM_DEBUG_TRACE | CAM_DEBUG_PERIPH, 1453 ("daopen\n")); 1454 1455 softc = (struct da_softc *)periph->softc; 1456 dareprobe(periph); 1457 1458 /* Wait for the disk size update. */ 1459 error = cam_periph_sleep(periph, &softc->disk->d_mediasize, PRIBIO, 1460 "dareprobe", 0); 1461 if (error != 0) 1462 xpt_print(periph->path, "unable to retrieve capacity data\n"); 1463 1464 if (periph->flags & CAM_PERIPH_INVALID) 1465 error = ENXIO; 1466 1467 if (error == 0 && (softc->flags & DA_FLAG_PACK_REMOVABLE) != 0 && 1468 (softc->quirks & DA_Q_NO_PREVENT) == 0) 1469 daprevent(periph, PR_PREVENT); 1470 1471 if (error == 0) { 1472 softc->flags &= ~DA_FLAG_PACK_INVALID; 1473 softc->flags |= DA_FLAG_OPEN; 1474 } 1475 1476 cam_periph_unhold(periph); 1477 cam_periph_unlock(periph); 1478 1479 if (error != 0) 1480 cam_periph_release(periph); 1481 1482 return (error); 1483 } 1484 1485 static int 1486 daclose(struct disk *dp) 1487 { 1488 struct cam_periph *periph; 1489 struct da_softc *softc; 1490 union ccb *ccb; 1491 int error; 1492 1493 periph = (struct cam_periph *)dp->d_drv1; 1494 softc = (struct da_softc *)periph->softc; 1495 cam_periph_lock(periph); 1496 CAM_DEBUG(periph->path, CAM_DEBUG_TRACE | CAM_DEBUG_PERIPH, 1497 ("daclose\n")); 1498 1499 if (cam_periph_hold(periph, PRIBIO) == 0) { 1500 1501 /* Flush disk cache. */ 1502 if ((softc->flags & DA_FLAG_DIRTY) != 0 && 1503 (softc->quirks & DA_Q_NO_SYNC_CACHE) == 0 && 1504 (softc->flags & DA_FLAG_PACK_INVALID) == 0) { 1505 ccb = cam_periph_getccb(periph, CAM_PRIORITY_NORMAL); 1506 scsi_synchronize_cache(&ccb->csio, /*retries*/1, 1507 /*cbfcnp*/dadone, MSG_SIMPLE_Q_TAG, 1508 /*begin_lba*/0, /*lb_count*/0, SSD_FULL_SIZE, 1509 5 * 60 * 1000); 1510 error = cam_periph_runccb(ccb, daerror, /*cam_flags*/0, 1511 /*sense_flags*/SF_RETRY_UA | SF_QUIET_IR, 1512 softc->disk->d_devstat); 1513 softc->flags &= ~DA_FLAG_DIRTY; 1514 xpt_release_ccb(ccb); 1515 } 1516 1517 /* Allow medium removal. */ 1518 if ((softc->flags & DA_FLAG_PACK_REMOVABLE) != 0 && 1519 (softc->quirks & DA_Q_NO_PREVENT) == 0) 1520 daprevent(periph, PR_ALLOW); 1521 1522 cam_periph_unhold(periph); 1523 } 1524 1525 /* 1526 * If we've got removeable media, mark the blocksize as 1527 * unavailable, since it could change when new media is 1528 * inserted. 1529 */ 1530 if ((softc->flags & DA_FLAG_PACK_REMOVABLE) != 0) 1531 softc->disk->d_devstat->flags |= DEVSTAT_BS_UNAVAILABLE; 1532 1533 softc->flags &= ~DA_FLAG_OPEN; 1534 while (softc->refcount != 0) 1535 cam_periph_sleep(periph, &softc->refcount, PRIBIO, "daclose", 1); 1536 cam_periph_unlock(periph); 1537 cam_periph_release(periph); 1538 return (0); 1539 } 1540 1541 static void 1542 daschedule(struct cam_periph *periph) 1543 { 1544 struct da_softc *softc = (struct da_softc *)periph->softc; 1545 1546 if (softc->state != DA_STATE_NORMAL) 1547 return; 1548 1549 cam_iosched_schedule(softc->cam_iosched, periph); 1550 } 1551 1552 /* 1553 * Actually translate the requested transfer into one the physical driver 1554 * can understand. The transfer is described by a buf and will include 1555 * only one physical transfer. 1556 */ 1557 static void 1558 dastrategy(struct bio *bp) 1559 { 1560 struct cam_periph *periph; 1561 struct da_softc *softc; 1562 1563 periph = (struct cam_periph *)bp->bio_disk->d_drv1; 1564 softc = (struct da_softc *)periph->softc; 1565 1566 cam_periph_lock(periph); 1567 1568 /* 1569 * If the device has been made invalid, error out 1570 */ 1571 if ((softc->flags & DA_FLAG_PACK_INVALID)) { 1572 cam_periph_unlock(periph); 1573 biofinish(bp, NULL, ENXIO); 1574 return; 1575 } 1576 1577 CAM_DEBUG(periph->path, CAM_DEBUG_TRACE, ("dastrategy(%p)\n", bp)); 1578 1579 /* 1580 * Zone commands must be ordered, because they can depend on the 1581 * effects of previously issued commands, and they may affect 1582 * commands after them. 1583 */ 1584 if (bp->bio_cmd == BIO_ZONE) 1585 bp->bio_flags |= BIO_ORDERED; 1586 1587 /* 1588 * Place it in the queue of disk activities for this disk 1589 */ 1590 cam_iosched_queue_work(softc->cam_iosched, bp); 1591 1592 /* 1593 * Schedule ourselves for performing the work. 1594 */ 1595 daschedule(periph); 1596 cam_periph_unlock(periph); 1597 1598 return; 1599 } 1600 1601 static int 1602 dadump(void *arg, void *virtual, vm_offset_t physical, off_t offset, size_t length) 1603 { 1604 struct cam_periph *periph; 1605 struct da_softc *softc; 1606 u_int secsize; 1607 struct ccb_scsiio csio; 1608 struct disk *dp; 1609 int error = 0; 1610 1611 dp = arg; 1612 periph = dp->d_drv1; 1613 softc = (struct da_softc *)periph->softc; 1614 cam_periph_lock(periph); 1615 secsize = softc->params.secsize; 1616 1617 if ((softc->flags & DA_FLAG_PACK_INVALID) != 0) { 1618 cam_periph_unlock(periph); 1619 return (ENXIO); 1620 } 1621 1622 if (length > 0) { 1623 xpt_setup_ccb(&csio.ccb_h, periph->path, CAM_PRIORITY_NORMAL); 1624 csio.ccb_h.ccb_state = DA_CCB_DUMP; 1625 scsi_read_write(&csio, 1626 /*retries*/0, 1627 dadone, 1628 MSG_ORDERED_Q_TAG, 1629 /*read*/SCSI_RW_WRITE, 1630 /*byte2*/0, 1631 /*minimum_cmd_size*/ softc->minimum_cmd_size, 1632 offset / secsize, 1633 length / secsize, 1634 /*data_ptr*/(u_int8_t *) virtual, 1635 /*dxfer_len*/length, 1636 /*sense_len*/SSD_FULL_SIZE, 1637 da_default_timeout * 1000); 1638 xpt_polled_action((union ccb *)&csio); 1639 1640 error = cam_periph_error((union ccb *)&csio, 1641 0, SF_NO_RECOVERY | SF_NO_RETRY, NULL); 1642 if ((csio.ccb_h.status & CAM_DEV_QFRZN) != 0) 1643 cam_release_devq(csio.ccb_h.path, /*relsim_flags*/0, 1644 /*reduction*/0, /*timeout*/0, /*getcount_only*/0); 1645 if (error != 0) 1646 printf("Aborting dump due to I/O error.\n"); 1647 cam_periph_unlock(periph); 1648 return (error); 1649 } 1650 1651 /* 1652 * Sync the disk cache contents to the physical media. 1653 */ 1654 if ((softc->quirks & DA_Q_NO_SYNC_CACHE) == 0) { 1655 1656 xpt_setup_ccb(&csio.ccb_h, periph->path, CAM_PRIORITY_NORMAL); 1657 csio.ccb_h.ccb_state = DA_CCB_DUMP; 1658 scsi_synchronize_cache(&csio, 1659 /*retries*/0, 1660 /*cbfcnp*/dadone, 1661 MSG_SIMPLE_Q_TAG, 1662 /*begin_lba*/0,/* Cover the whole disk */ 1663 /*lb_count*/0, 1664 SSD_FULL_SIZE, 1665 5 * 1000); 1666 xpt_polled_action((union ccb *)&csio); 1667 1668 error = cam_periph_error((union ccb *)&csio, 1669 0, SF_NO_RECOVERY | SF_NO_RETRY | SF_QUIET_IR, NULL); 1670 if ((csio.ccb_h.status & CAM_DEV_QFRZN) != 0) 1671 cam_release_devq(csio.ccb_h.path, /*relsim_flags*/0, 1672 /*reduction*/0, /*timeout*/0, /*getcount_only*/0); 1673 if (error != 0) 1674 xpt_print(periph->path, "Synchronize cache failed\n"); 1675 } 1676 cam_periph_unlock(periph); 1677 return (error); 1678 } 1679 1680 static int 1681 dagetattr(struct bio *bp) 1682 { 1683 int ret; 1684 struct cam_periph *periph; 1685 1686 periph = (struct cam_periph *)bp->bio_disk->d_drv1; 1687 cam_periph_lock(periph); 1688 ret = xpt_getattr(bp->bio_data, bp->bio_length, bp->bio_attribute, 1689 periph->path); 1690 cam_periph_unlock(periph); 1691 if (ret == 0) 1692 bp->bio_completed = bp->bio_length; 1693 return ret; 1694 } 1695 1696 static void 1697 dainit(void) 1698 { 1699 cam_status status; 1700 1701 /* 1702 * Install a global async callback. This callback will 1703 * receive async callbacks like "new device found". 1704 */ 1705 status = xpt_register_async(AC_FOUND_DEVICE, daasync, NULL, NULL); 1706 1707 if (status != CAM_REQ_CMP) { 1708 printf("da: Failed to attach master async callback " 1709 "due to status 0x%x!\n", status); 1710 } else if (da_send_ordered) { 1711 1712 /* Register our shutdown event handler */ 1713 if ((EVENTHANDLER_REGISTER(shutdown_post_sync, dashutdown, 1714 NULL, SHUTDOWN_PRI_DEFAULT)) == NULL) 1715 printf("dainit: shutdown event registration failed!\n"); 1716 } 1717 } 1718 1719 /* 1720 * Callback from GEOM, called when it has finished cleaning up its 1721 * resources. 1722 */ 1723 static void 1724 dadiskgonecb(struct disk *dp) 1725 { 1726 struct cam_periph *periph; 1727 1728 periph = (struct cam_periph *)dp->d_drv1; 1729 cam_periph_release(periph); 1730 } 1731 1732 static void 1733 daoninvalidate(struct cam_periph *periph) 1734 { 1735 struct da_softc *softc; 1736 1737 softc = (struct da_softc *)periph->softc; 1738 1739 /* 1740 * De-register any async callbacks. 1741 */ 1742 xpt_register_async(0, daasync, periph, periph->path); 1743 1744 softc->flags |= DA_FLAG_PACK_INVALID; 1745 #ifdef CAM_IO_STATS 1746 softc->invalidations++; 1747 #endif 1748 1749 /* 1750 * Return all queued I/O with ENXIO. 1751 * XXX Handle any transactions queued to the card 1752 * with XPT_ABORT_CCB. 1753 */ 1754 cam_iosched_flush(softc->cam_iosched, NULL, ENXIO); 1755 1756 /* 1757 * Tell GEOM that we've gone away, we'll get a callback when it is 1758 * done cleaning up its resources. 1759 */ 1760 disk_gone(softc->disk); 1761 } 1762 1763 static void 1764 dacleanup(struct cam_periph *periph) 1765 { 1766 struct da_softc *softc; 1767 1768 softc = (struct da_softc *)periph->softc; 1769 1770 cam_periph_unlock(periph); 1771 1772 cam_iosched_fini(softc->cam_iosched); 1773 1774 /* 1775 * If we can't free the sysctl tree, oh well... 1776 */ 1777 if ((softc->flags & DA_FLAG_SCTX_INIT) != 0) { 1778 #ifdef CAM_IO_STATS 1779 if (sysctl_ctx_free(&softc->sysctl_stats_ctx) != 0) 1780 xpt_print(periph->path, 1781 "can't remove sysctl stats context\n"); 1782 #endif 1783 if (sysctl_ctx_free(&softc->sysctl_ctx) != 0) 1784 xpt_print(periph->path, 1785 "can't remove sysctl context\n"); 1786 } 1787 1788 callout_drain(&softc->mediapoll_c); 1789 disk_destroy(softc->disk); 1790 callout_drain(&softc->sendordered_c); 1791 free(softc, M_DEVBUF); 1792 cam_periph_lock(periph); 1793 } 1794 1795 static void 1796 daasync(void *callback_arg, u_int32_t code, 1797 struct cam_path *path, void *arg) 1798 { 1799 struct cam_periph *periph; 1800 struct da_softc *softc; 1801 1802 periph = (struct cam_periph *)callback_arg; 1803 switch (code) { 1804 case AC_FOUND_DEVICE: 1805 { 1806 struct ccb_getdev *cgd; 1807 cam_status status; 1808 1809 cgd = (struct ccb_getdev *)arg; 1810 if (cgd == NULL) 1811 break; 1812 1813 if (cgd->protocol != PROTO_SCSI) 1814 break; 1815 if (SID_QUAL(&cgd->inq_data) != SID_QUAL_LU_CONNECTED) 1816 break; 1817 if (SID_TYPE(&cgd->inq_data) != T_DIRECT 1818 && SID_TYPE(&cgd->inq_data) != T_RBC 1819 && SID_TYPE(&cgd->inq_data) != T_OPTICAL 1820 && SID_TYPE(&cgd->inq_data) != T_ZBC_HM) 1821 break; 1822 1823 /* 1824 * Allocate a peripheral instance for 1825 * this device and start the probe 1826 * process. 1827 */ 1828 status = cam_periph_alloc(daregister, daoninvalidate, 1829 dacleanup, dastart, 1830 "da", CAM_PERIPH_BIO, 1831 path, daasync, 1832 AC_FOUND_DEVICE, cgd); 1833 1834 if (status != CAM_REQ_CMP 1835 && status != CAM_REQ_INPROG) 1836 printf("daasync: Unable to attach to new device " 1837 "due to status 0x%x\n", status); 1838 return; 1839 } 1840 case AC_ADVINFO_CHANGED: 1841 { 1842 uintptr_t buftype; 1843 1844 buftype = (uintptr_t)arg; 1845 if (buftype == CDAI_TYPE_PHYS_PATH) { 1846 struct da_softc *softc; 1847 1848 softc = periph->softc; 1849 disk_attr_changed(softc->disk, "GEOM::physpath", 1850 M_NOWAIT); 1851 } 1852 break; 1853 } 1854 case AC_UNIT_ATTENTION: 1855 { 1856 union ccb *ccb; 1857 int error_code, sense_key, asc, ascq; 1858 1859 softc = (struct da_softc *)periph->softc; 1860 ccb = (union ccb *)arg; 1861 1862 /* 1863 * Handle all UNIT ATTENTIONs except our own, 1864 * as they will be handled by daerror(). 1865 */ 1866 if (xpt_path_periph(ccb->ccb_h.path) != periph && 1867 scsi_extract_sense_ccb(ccb, 1868 &error_code, &sense_key, &asc, &ascq)) { 1869 if (asc == 0x2A && ascq == 0x09) { 1870 xpt_print(ccb->ccb_h.path, 1871 "Capacity data has changed\n"); 1872 softc->flags &= ~DA_FLAG_PROBED; 1873 dareprobe(periph); 1874 } else if (asc == 0x28 && ascq == 0x00) { 1875 softc->flags &= ~DA_FLAG_PROBED; 1876 disk_media_changed(softc->disk, M_NOWAIT); 1877 } else if (asc == 0x3F && ascq == 0x03) { 1878 xpt_print(ccb->ccb_h.path, 1879 "INQUIRY data has changed\n"); 1880 softc->flags &= ~DA_FLAG_PROBED; 1881 dareprobe(periph); 1882 } 1883 } 1884 cam_periph_async(periph, code, path, arg); 1885 break; 1886 } 1887 case AC_SCSI_AEN: 1888 softc = (struct da_softc *)periph->softc; 1889 if (!cam_iosched_has_work_flags(softc->cam_iosched, DA_WORK_TUR)) { 1890 if (cam_periph_acquire(periph) == CAM_REQ_CMP) { 1891 cam_iosched_set_work_flags(softc->cam_iosched, DA_WORK_TUR); 1892 daschedule(periph); 1893 } 1894 } 1895 /* FALLTHROUGH */ 1896 case AC_SENT_BDR: 1897 case AC_BUS_RESET: 1898 { 1899 struct ccb_hdr *ccbh; 1900 1901 softc = (struct da_softc *)periph->softc; 1902 /* 1903 * Don't fail on the expected unit attention 1904 * that will occur. 1905 */ 1906 softc->flags |= DA_FLAG_RETRY_UA; 1907 LIST_FOREACH(ccbh, &softc->pending_ccbs, periph_links.le) 1908 ccbh->ccb_state |= DA_CCB_RETRY_UA; 1909 break; 1910 } 1911 case AC_INQ_CHANGED: 1912 softc = (struct da_softc *)periph->softc; 1913 softc->flags &= ~DA_FLAG_PROBED; 1914 dareprobe(periph); 1915 break; 1916 default: 1917 break; 1918 } 1919 cam_periph_async(periph, code, path, arg); 1920 } 1921 1922 static void 1923 dasysctlinit(void *context, int pending) 1924 { 1925 struct cam_periph *periph; 1926 struct da_softc *softc; 1927 char tmpstr[80], tmpstr2[80]; 1928 struct ccb_trans_settings cts; 1929 1930 periph = (struct cam_periph *)context; 1931 /* 1932 * periph was held for us when this task was enqueued 1933 */ 1934 if (periph->flags & CAM_PERIPH_INVALID) { 1935 cam_periph_release(periph); 1936 return; 1937 } 1938 1939 softc = (struct da_softc *)periph->softc; 1940 snprintf(tmpstr, sizeof(tmpstr), "CAM DA unit %d", periph->unit_number); 1941 snprintf(tmpstr2, sizeof(tmpstr2), "%d", periph->unit_number); 1942 1943 sysctl_ctx_init(&softc->sysctl_ctx); 1944 softc->flags |= DA_FLAG_SCTX_INIT; 1945 softc->sysctl_tree = SYSCTL_ADD_NODE_WITH_LABEL(&softc->sysctl_ctx, 1946 SYSCTL_STATIC_CHILDREN(_kern_cam_da), OID_AUTO, tmpstr2, 1947 CTLFLAG_RD, 0, tmpstr, "device_index"); 1948 if (softc->sysctl_tree == NULL) { 1949 printf("dasysctlinit: unable to allocate sysctl tree\n"); 1950 cam_periph_release(periph); 1951 return; 1952 } 1953 1954 /* 1955 * Now register the sysctl handler, so the user can change the value on 1956 * the fly. 1957 */ 1958 SYSCTL_ADD_PROC(&softc->sysctl_ctx, SYSCTL_CHILDREN(softc->sysctl_tree), 1959 OID_AUTO, "delete_method", CTLTYPE_STRING | CTLFLAG_RWTUN, 1960 softc, 0, dadeletemethodsysctl, "A", 1961 "BIO_DELETE execution method"); 1962 SYSCTL_ADD_PROC(&softc->sysctl_ctx, SYSCTL_CHILDREN(softc->sysctl_tree), 1963 OID_AUTO, "delete_max", CTLTYPE_U64 | CTLFLAG_RW, 1964 softc, 0, dadeletemaxsysctl, "Q", 1965 "Maximum BIO_DELETE size"); 1966 SYSCTL_ADD_PROC(&softc->sysctl_ctx, SYSCTL_CHILDREN(softc->sysctl_tree), 1967 OID_AUTO, "minimum_cmd_size", CTLTYPE_INT | CTLFLAG_RW, 1968 &softc->minimum_cmd_size, 0, dacmdsizesysctl, "I", 1969 "Minimum CDB size"); 1970 1971 SYSCTL_ADD_PROC(&softc->sysctl_ctx, SYSCTL_CHILDREN(softc->sysctl_tree), 1972 OID_AUTO, "zone_mode", CTLTYPE_STRING | CTLFLAG_RD, 1973 softc, 0, dazonemodesysctl, "A", 1974 "Zone Mode"); 1975 SYSCTL_ADD_PROC(&softc->sysctl_ctx, SYSCTL_CHILDREN(softc->sysctl_tree), 1976 OID_AUTO, "zone_support", CTLTYPE_STRING | CTLFLAG_RD, 1977 softc, 0, dazonesupsysctl, "A", 1978 "Zone Support"); 1979 SYSCTL_ADD_UQUAD(&softc->sysctl_ctx, 1980 SYSCTL_CHILDREN(softc->sysctl_tree), OID_AUTO, 1981 "optimal_seq_zones", CTLFLAG_RD, &softc->optimal_seq_zones, 1982 "Optimal Number of Open Sequential Write Preferred Zones"); 1983 SYSCTL_ADD_UQUAD(&softc->sysctl_ctx, 1984 SYSCTL_CHILDREN(softc->sysctl_tree), OID_AUTO, 1985 "optimal_nonseq_zones", CTLFLAG_RD, 1986 &softc->optimal_nonseq_zones, 1987 "Optimal Number of Non-Sequentially Written Sequential Write " 1988 "Preferred Zones"); 1989 SYSCTL_ADD_UQUAD(&softc->sysctl_ctx, 1990 SYSCTL_CHILDREN(softc->sysctl_tree), OID_AUTO, 1991 "max_seq_zones", CTLFLAG_RD, &softc->max_seq_zones, 1992 "Maximum Number of Open Sequential Write Required Zones"); 1993 1994 SYSCTL_ADD_INT(&softc->sysctl_ctx, 1995 SYSCTL_CHILDREN(softc->sysctl_tree), 1996 OID_AUTO, 1997 "error_inject", 1998 CTLFLAG_RW, 1999 &softc->error_inject, 2000 0, 2001 "error_inject leaf"); 2002 2003 SYSCTL_ADD_INT(&softc->sysctl_ctx, 2004 SYSCTL_CHILDREN(softc->sysctl_tree), 2005 OID_AUTO, 2006 "unmapped_io", 2007 CTLFLAG_RD, 2008 &softc->unmappedio, 2009 0, 2010 "Unmapped I/O leaf"); 2011 2012 SYSCTL_ADD_INT(&softc->sysctl_ctx, 2013 SYSCTL_CHILDREN(softc->sysctl_tree), 2014 OID_AUTO, 2015 "rotating", 2016 CTLFLAG_RD, 2017 &softc->rotating, 2018 0, 2019 "Rotating media"); 2020 2021 /* 2022 * Add some addressing info. 2023 */ 2024 memset(&cts, 0, sizeof (cts)); 2025 xpt_setup_ccb(&cts.ccb_h, periph->path, CAM_PRIORITY_NONE); 2026 cts.ccb_h.func_code = XPT_GET_TRAN_SETTINGS; 2027 cts.type = CTS_TYPE_CURRENT_SETTINGS; 2028 cam_periph_lock(periph); 2029 xpt_action((union ccb *)&cts); 2030 cam_periph_unlock(periph); 2031 if (cts.ccb_h.status != CAM_REQ_CMP) { 2032 cam_periph_release(periph); 2033 return; 2034 } 2035 if (cts.protocol == PROTO_SCSI && cts.transport == XPORT_FC) { 2036 struct ccb_trans_settings_fc *fc = &cts.xport_specific.fc; 2037 if (fc->valid & CTS_FC_VALID_WWPN) { 2038 softc->wwpn = fc->wwpn; 2039 SYSCTL_ADD_UQUAD(&softc->sysctl_ctx, 2040 SYSCTL_CHILDREN(softc->sysctl_tree), 2041 OID_AUTO, "wwpn", CTLFLAG_RD, 2042 &softc->wwpn, "World Wide Port Name"); 2043 } 2044 } 2045 2046 #ifdef CAM_IO_STATS 2047 /* 2048 * Now add some useful stats. 2049 * XXX These should live in cam_periph and be common to all periphs 2050 */ 2051 softc->sysctl_stats_tree = SYSCTL_ADD_NODE(&softc->sysctl_stats_ctx, 2052 SYSCTL_CHILDREN(softc->sysctl_tree), OID_AUTO, "stats", 2053 CTLFLAG_RD, 0, "Statistics"); 2054 SYSCTL_ADD_INT(&softc->sysctl_stats_ctx, 2055 SYSCTL_CHILDREN(softc->sysctl_stats_tree), 2056 OID_AUTO, 2057 "errors", 2058 CTLFLAG_RD, 2059 &softc->errors, 2060 0, 2061 "Transport errors reported by the SIM"); 2062 SYSCTL_ADD_INT(&softc->sysctl_stats_ctx, 2063 SYSCTL_CHILDREN(softc->sysctl_stats_tree), 2064 OID_AUTO, 2065 "timeouts", 2066 CTLFLAG_RD, 2067 &softc->timeouts, 2068 0, 2069 "Device timeouts reported by the SIM"); 2070 SYSCTL_ADD_INT(&softc->sysctl_stats_ctx, 2071 SYSCTL_CHILDREN(softc->sysctl_stats_tree), 2072 OID_AUTO, 2073 "pack_invalidations", 2074 CTLFLAG_RD, 2075 &softc->invalidations, 2076 0, 2077 "Device pack invalidations"); 2078 #endif 2079 2080 cam_iosched_sysctl_init(softc->cam_iosched, &softc->sysctl_ctx, 2081 softc->sysctl_tree); 2082 2083 cam_periph_release(periph); 2084 } 2085 2086 static int 2087 dadeletemaxsysctl(SYSCTL_HANDLER_ARGS) 2088 { 2089 int error; 2090 uint64_t value; 2091 struct da_softc *softc; 2092 2093 softc = (struct da_softc *)arg1; 2094 2095 value = softc->disk->d_delmaxsize; 2096 error = sysctl_handle_64(oidp, &value, 0, req); 2097 if ((error != 0) || (req->newptr == NULL)) 2098 return (error); 2099 2100 /* only accept values smaller than the calculated value */ 2101 if (value > dadeletemaxsize(softc, softc->delete_method)) { 2102 return (EINVAL); 2103 } 2104 softc->disk->d_delmaxsize = value; 2105 2106 return (0); 2107 } 2108 2109 static int 2110 dacmdsizesysctl(SYSCTL_HANDLER_ARGS) 2111 { 2112 int error, value; 2113 2114 value = *(int *)arg1; 2115 2116 error = sysctl_handle_int(oidp, &value, 0, req); 2117 2118 if ((error != 0) 2119 || (req->newptr == NULL)) 2120 return (error); 2121 2122 /* 2123 * Acceptable values here are 6, 10, 12 or 16. 2124 */ 2125 if (value < 6) 2126 value = 6; 2127 else if ((value > 6) 2128 && (value <= 10)) 2129 value = 10; 2130 else if ((value > 10) 2131 && (value <= 12)) 2132 value = 12; 2133 else if (value > 12) 2134 value = 16; 2135 2136 *(int *)arg1 = value; 2137 2138 return (0); 2139 } 2140 2141 static int 2142 dasysctlsofttimeout(SYSCTL_HANDLER_ARGS) 2143 { 2144 sbintime_t value; 2145 int error; 2146 2147 value = da_default_softtimeout / SBT_1MS; 2148 2149 error = sysctl_handle_int(oidp, (int *)&value, 0, req); 2150 if ((error != 0) || (req->newptr == NULL)) 2151 return (error); 2152 2153 /* XXX Should clip this to a reasonable level */ 2154 if (value > da_default_timeout * 1000) 2155 return (EINVAL); 2156 2157 da_default_softtimeout = value * SBT_1MS; 2158 return (0); 2159 } 2160 2161 static void 2162 dadeletemethodset(struct da_softc *softc, da_delete_methods delete_method) 2163 { 2164 2165 softc->delete_method = delete_method; 2166 softc->disk->d_delmaxsize = dadeletemaxsize(softc, delete_method); 2167 softc->delete_func = da_delete_functions[delete_method]; 2168 2169 if (softc->delete_method > DA_DELETE_DISABLE) 2170 softc->disk->d_flags |= DISKFLAG_CANDELETE; 2171 else 2172 softc->disk->d_flags &= ~DISKFLAG_CANDELETE; 2173 } 2174 2175 static off_t 2176 dadeletemaxsize(struct da_softc *softc, da_delete_methods delete_method) 2177 { 2178 off_t sectors; 2179 2180 switch(delete_method) { 2181 case DA_DELETE_UNMAP: 2182 sectors = (off_t)softc->unmap_max_lba; 2183 break; 2184 case DA_DELETE_ATA_TRIM: 2185 sectors = (off_t)ATA_DSM_RANGE_MAX * softc->trim_max_ranges; 2186 break; 2187 case DA_DELETE_WS16: 2188 sectors = omin(softc->ws_max_blks, WS16_MAX_BLKS); 2189 break; 2190 case DA_DELETE_ZERO: 2191 case DA_DELETE_WS10: 2192 sectors = omin(softc->ws_max_blks, WS10_MAX_BLKS); 2193 break; 2194 default: 2195 return 0; 2196 } 2197 2198 return (off_t)softc->params.secsize * 2199 omin(sectors, softc->params.sectors); 2200 } 2201 2202 static void 2203 daprobedone(struct cam_periph *periph, union ccb *ccb) 2204 { 2205 struct da_softc *softc; 2206 2207 softc = (struct da_softc *)periph->softc; 2208 2209 dadeletemethodchoose(softc, DA_DELETE_NONE); 2210 2211 if (bootverbose && (softc->flags & DA_FLAG_ANNOUNCED) == 0) { 2212 char buf[80]; 2213 int i, sep; 2214 2215 snprintf(buf, sizeof(buf), "Delete methods: <"); 2216 sep = 0; 2217 for (i = 0; i <= DA_DELETE_MAX; i++) { 2218 if ((softc->delete_available & (1 << i)) == 0 && 2219 i != softc->delete_method) 2220 continue; 2221 if (sep) 2222 strlcat(buf, ",", sizeof(buf)); 2223 strlcat(buf, da_delete_method_names[i], 2224 sizeof(buf)); 2225 if (i == softc->delete_method) 2226 strlcat(buf, "(*)", sizeof(buf)); 2227 sep = 1; 2228 } 2229 strlcat(buf, ">", sizeof(buf)); 2230 printf("%s%d: %s\n", periph->periph_name, 2231 periph->unit_number, buf); 2232 } 2233 2234 /* 2235 * Since our peripheral may be invalidated by an error 2236 * above or an external event, we must release our CCB 2237 * before releasing the probe lock on the peripheral. 2238 * The peripheral will only go away once the last lock 2239 * is removed, and we need it around for the CCB release 2240 * operation. 2241 */ 2242 xpt_release_ccb(ccb); 2243 softc->state = DA_STATE_NORMAL; 2244 softc->flags |= DA_FLAG_PROBED; 2245 daschedule(periph); 2246 wakeup(&softc->disk->d_mediasize); 2247 if ((softc->flags & DA_FLAG_ANNOUNCED) == 0) { 2248 softc->flags |= DA_FLAG_ANNOUNCED; 2249 cam_periph_unhold(periph); 2250 } else 2251 cam_periph_release_locked(periph); 2252 } 2253 2254 static void 2255 dadeletemethodchoose(struct da_softc *softc, da_delete_methods default_method) 2256 { 2257 int i, methods; 2258 2259 /* If available, prefer the method requested by user. */ 2260 i = softc->delete_method_pref; 2261 methods = softc->delete_available | (1 << DA_DELETE_DISABLE); 2262 if (methods & (1 << i)) { 2263 dadeletemethodset(softc, i); 2264 return; 2265 } 2266 2267 /* Use the pre-defined order to choose the best performing delete. */ 2268 for (i = DA_DELETE_MIN; i <= DA_DELETE_MAX; i++) { 2269 if (i == DA_DELETE_ZERO) 2270 continue; 2271 if (softc->delete_available & (1 << i)) { 2272 dadeletemethodset(softc, i); 2273 return; 2274 } 2275 } 2276 2277 /* Fallback to default. */ 2278 dadeletemethodset(softc, default_method); 2279 } 2280 2281 static int 2282 dadeletemethodsysctl(SYSCTL_HANDLER_ARGS) 2283 { 2284 char buf[16]; 2285 const char *p; 2286 struct da_softc *softc; 2287 int i, error, methods, value; 2288 2289 softc = (struct da_softc *)arg1; 2290 2291 value = softc->delete_method; 2292 if (value < 0 || value > DA_DELETE_MAX) 2293 p = "UNKNOWN"; 2294 else 2295 p = da_delete_method_names[value]; 2296 strncpy(buf, p, sizeof(buf)); 2297 error = sysctl_handle_string(oidp, buf, sizeof(buf), req); 2298 if (error != 0 || req->newptr == NULL) 2299 return (error); 2300 methods = softc->delete_available | (1 << DA_DELETE_DISABLE); 2301 for (i = 0; i <= DA_DELETE_MAX; i++) { 2302 if (strcmp(buf, da_delete_method_names[i]) == 0) 2303 break; 2304 } 2305 if (i > DA_DELETE_MAX) 2306 return (EINVAL); 2307 softc->delete_method_pref = i; 2308 dadeletemethodchoose(softc, DA_DELETE_NONE); 2309 return (0); 2310 } 2311 2312 static int 2313 dazonemodesysctl(SYSCTL_HANDLER_ARGS) 2314 { 2315 char tmpbuf[40]; 2316 struct da_softc *softc; 2317 int error; 2318 2319 softc = (struct da_softc *)arg1; 2320 2321 switch (softc->zone_mode) { 2322 case DA_ZONE_DRIVE_MANAGED: 2323 snprintf(tmpbuf, sizeof(tmpbuf), "Drive Managed"); 2324 break; 2325 case DA_ZONE_HOST_AWARE: 2326 snprintf(tmpbuf, sizeof(tmpbuf), "Host Aware"); 2327 break; 2328 case DA_ZONE_HOST_MANAGED: 2329 snprintf(tmpbuf, sizeof(tmpbuf), "Host Managed"); 2330 break; 2331 case DA_ZONE_NONE: 2332 default: 2333 snprintf(tmpbuf, sizeof(tmpbuf), "Not Zoned"); 2334 break; 2335 } 2336 2337 error = sysctl_handle_string(oidp, tmpbuf, sizeof(tmpbuf), req); 2338 2339 return (error); 2340 } 2341 2342 static int 2343 dazonesupsysctl(SYSCTL_HANDLER_ARGS) 2344 { 2345 char tmpbuf[180]; 2346 struct da_softc *softc; 2347 struct sbuf sb; 2348 int error, first; 2349 unsigned int i; 2350 2351 softc = (struct da_softc *)arg1; 2352 2353 error = 0; 2354 first = 1; 2355 sbuf_new(&sb, tmpbuf, sizeof(tmpbuf), 0); 2356 2357 for (i = 0; i < sizeof(da_zone_desc_table) / 2358 sizeof(da_zone_desc_table[0]); i++) { 2359 if (softc->zone_flags & da_zone_desc_table[i].value) { 2360 if (first == 0) 2361 sbuf_printf(&sb, ", "); 2362 else 2363 first = 0; 2364 sbuf_cat(&sb, da_zone_desc_table[i].desc); 2365 } 2366 } 2367 2368 if (first == 1) 2369 sbuf_printf(&sb, "None"); 2370 2371 sbuf_finish(&sb); 2372 2373 error = sysctl_handle_string(oidp, sbuf_data(&sb), sbuf_len(&sb), req); 2374 2375 return (error); 2376 } 2377 2378 static cam_status 2379 daregister(struct cam_periph *periph, void *arg) 2380 { 2381 struct da_softc *softc; 2382 struct ccb_pathinq cpi; 2383 struct ccb_getdev *cgd; 2384 char tmpstr[80]; 2385 caddr_t match; 2386 2387 cgd = (struct ccb_getdev *)arg; 2388 if (cgd == NULL) { 2389 printf("daregister: no getdev CCB, can't register device\n"); 2390 return(CAM_REQ_CMP_ERR); 2391 } 2392 2393 softc = (struct da_softc *)malloc(sizeof(*softc), M_DEVBUF, 2394 M_NOWAIT|M_ZERO); 2395 2396 if (softc == NULL) { 2397 printf("daregister: Unable to probe new device. " 2398 "Unable to allocate softc\n"); 2399 return(CAM_REQ_CMP_ERR); 2400 } 2401 2402 if (cam_iosched_init(&softc->cam_iosched, periph) != 0) { 2403 printf("daregister: Unable to probe new device. " 2404 "Unable to allocate iosched memory\n"); 2405 free(softc, M_DEVBUF); 2406 return(CAM_REQ_CMP_ERR); 2407 } 2408 2409 LIST_INIT(&softc->pending_ccbs); 2410 softc->state = DA_STATE_PROBE_RC; 2411 bioq_init(&softc->delete_run_queue); 2412 if (SID_IS_REMOVABLE(&cgd->inq_data)) 2413 softc->flags |= DA_FLAG_PACK_REMOVABLE; 2414 softc->unmap_max_ranges = UNMAP_MAX_RANGES; 2415 softc->unmap_max_lba = UNMAP_RANGE_MAX; 2416 softc->unmap_gran = 0; 2417 softc->unmap_gran_align = 0; 2418 softc->ws_max_blks = WS16_MAX_BLKS; 2419 softc->trim_max_ranges = ATA_TRIM_MAX_RANGES; 2420 softc->rotating = 1; 2421 2422 periph->softc = softc; 2423 2424 /* 2425 * See if this device has any quirks. 2426 */ 2427 match = cam_quirkmatch((caddr_t)&cgd->inq_data, 2428 (caddr_t)da_quirk_table, 2429 nitems(da_quirk_table), 2430 sizeof(*da_quirk_table), scsi_inquiry_match); 2431 2432 if (match != NULL) 2433 softc->quirks = ((struct da_quirk_entry *)match)->quirks; 2434 else 2435 softc->quirks = DA_Q_NONE; 2436 2437 /* Check if the SIM does not want 6 byte commands */ 2438 bzero(&cpi, sizeof(cpi)); 2439 xpt_setup_ccb(&cpi.ccb_h, periph->path, CAM_PRIORITY_NORMAL); 2440 cpi.ccb_h.func_code = XPT_PATH_INQ; 2441 xpt_action((union ccb *)&cpi); 2442 if (cpi.ccb_h.status == CAM_REQ_CMP && (cpi.hba_misc & PIM_NO_6_BYTE)) 2443 softc->quirks |= DA_Q_NO_6_BYTE; 2444 2445 if (SID_TYPE(&cgd->inq_data) == T_ZBC_HM) 2446 softc->zone_mode = DA_ZONE_HOST_MANAGED; 2447 else if (softc->quirks & DA_Q_SMR_DM) 2448 softc->zone_mode = DA_ZONE_DRIVE_MANAGED; 2449 else 2450 softc->zone_mode = DA_ZONE_NONE; 2451 2452 if (softc->zone_mode != DA_ZONE_NONE) { 2453 if (scsi_vpd_supported_page(periph, SVPD_ATA_INFORMATION)) { 2454 if (scsi_vpd_supported_page(periph, SVPD_ZONED_BDC)) 2455 softc->zone_interface = DA_ZONE_IF_ATA_SAT; 2456 else 2457 softc->zone_interface = DA_ZONE_IF_ATA_PASS; 2458 } else 2459 softc->zone_interface = DA_ZONE_IF_SCSI; 2460 } 2461 2462 TASK_INIT(&softc->sysctl_task, 0, dasysctlinit, periph); 2463 2464 /* 2465 * Take an exclusive refcount on the periph while dastart is called 2466 * to finish the probe. The reference will be dropped in dadone at 2467 * the end of probe. 2468 */ 2469 (void)cam_periph_hold(periph, PRIBIO); 2470 2471 /* 2472 * Schedule a periodic event to occasionally send an 2473 * ordered tag to a device. 2474 */ 2475 callout_init_mtx(&softc->sendordered_c, cam_periph_mtx(periph), 0); 2476 callout_reset(&softc->sendordered_c, 2477 (da_default_timeout * hz) / DA_ORDEREDTAG_INTERVAL, 2478 dasendorderedtag, softc); 2479 2480 cam_periph_unlock(periph); 2481 /* 2482 * RBC devices don't have to support READ(6), only READ(10). 2483 */ 2484 if (softc->quirks & DA_Q_NO_6_BYTE || SID_TYPE(&cgd->inq_data) == T_RBC) 2485 softc->minimum_cmd_size = 10; 2486 else 2487 softc->minimum_cmd_size = 6; 2488 2489 /* 2490 * Load the user's default, if any. 2491 */ 2492 snprintf(tmpstr, sizeof(tmpstr), "kern.cam.da.%d.minimum_cmd_size", 2493 periph->unit_number); 2494 TUNABLE_INT_FETCH(tmpstr, &softc->minimum_cmd_size); 2495 2496 /* 2497 * 6, 10, 12 and 16 are the currently permissible values. 2498 */ 2499 if (softc->minimum_cmd_size < 6) 2500 softc->minimum_cmd_size = 6; 2501 else if ((softc->minimum_cmd_size > 6) 2502 && (softc->minimum_cmd_size <= 10)) 2503 softc->minimum_cmd_size = 10; 2504 else if ((softc->minimum_cmd_size > 10) 2505 && (softc->minimum_cmd_size <= 12)) 2506 softc->minimum_cmd_size = 12; 2507 else if (softc->minimum_cmd_size > 12) 2508 softc->minimum_cmd_size = 16; 2509 2510 /* Predict whether device may support READ CAPACITY(16). */ 2511 if (SID_ANSI_REV(&cgd->inq_data) >= SCSI_REV_SPC3 && 2512 (softc->quirks & DA_Q_NO_RC16) == 0) { 2513 softc->flags |= DA_FLAG_CAN_RC16; 2514 softc->state = DA_STATE_PROBE_RC16; 2515 } 2516 2517 /* 2518 * Register this media as a disk. 2519 */ 2520 softc->disk = disk_alloc(); 2521 softc->disk->d_devstat = devstat_new_entry(periph->periph_name, 2522 periph->unit_number, 0, 2523 DEVSTAT_BS_UNAVAILABLE, 2524 SID_TYPE(&cgd->inq_data) | 2525 XPORT_DEVSTAT_TYPE(cpi.transport), 2526 DEVSTAT_PRIORITY_DISK); 2527 softc->disk->d_open = daopen; 2528 softc->disk->d_close = daclose; 2529 softc->disk->d_strategy = dastrategy; 2530 softc->disk->d_dump = dadump; 2531 softc->disk->d_getattr = dagetattr; 2532 softc->disk->d_gone = dadiskgonecb; 2533 softc->disk->d_name = "da"; 2534 softc->disk->d_drv1 = periph; 2535 if (cpi.maxio == 0) 2536 softc->maxio = DFLTPHYS; /* traditional default */ 2537 else if (cpi.maxio > MAXPHYS) 2538 softc->maxio = MAXPHYS; /* for safety */ 2539 else 2540 softc->maxio = cpi.maxio; 2541 softc->disk->d_maxsize = softc->maxio; 2542 softc->disk->d_unit = periph->unit_number; 2543 softc->disk->d_flags = DISKFLAG_DIRECT_COMPLETION | DISKFLAG_CANZONE; 2544 if ((softc->quirks & DA_Q_NO_SYNC_CACHE) == 0) 2545 softc->disk->d_flags |= DISKFLAG_CANFLUSHCACHE; 2546 if ((cpi.hba_misc & PIM_UNMAPPED) != 0) { 2547 softc->unmappedio = 1; 2548 softc->disk->d_flags |= DISKFLAG_UNMAPPED_BIO; 2549 } 2550 cam_strvis(softc->disk->d_descr, cgd->inq_data.vendor, 2551 sizeof(cgd->inq_data.vendor), sizeof(softc->disk->d_descr)); 2552 strlcat(softc->disk->d_descr, " ", sizeof(softc->disk->d_descr)); 2553 cam_strvis(&softc->disk->d_descr[strlen(softc->disk->d_descr)], 2554 cgd->inq_data.product, sizeof(cgd->inq_data.product), 2555 sizeof(softc->disk->d_descr) - strlen(softc->disk->d_descr)); 2556 softc->disk->d_hba_vendor = cpi.hba_vendor; 2557 softc->disk->d_hba_device = cpi.hba_device; 2558 softc->disk->d_hba_subvendor = cpi.hba_subvendor; 2559 softc->disk->d_hba_subdevice = cpi.hba_subdevice; 2560 2561 /* 2562 * Acquire a reference to the periph before we register with GEOM. 2563 * We'll release this reference once GEOM calls us back (via 2564 * dadiskgonecb()) telling us that our provider has been freed. 2565 */ 2566 if (cam_periph_acquire(periph) != CAM_REQ_CMP) { 2567 xpt_print(periph->path, "%s: lost periph during " 2568 "registration!\n", __func__); 2569 cam_periph_lock(periph); 2570 return (CAM_REQ_CMP_ERR); 2571 } 2572 2573 disk_create(softc->disk, DISK_VERSION); 2574 cam_periph_lock(periph); 2575 2576 /* 2577 * Add async callbacks for events of interest. 2578 * I don't bother checking if this fails as, 2579 * in most cases, the system will function just 2580 * fine without them and the only alternative 2581 * would be to not attach the device on failure. 2582 */ 2583 xpt_register_async(AC_SENT_BDR | AC_BUS_RESET | AC_LOST_DEVICE | 2584 AC_ADVINFO_CHANGED | AC_SCSI_AEN | AC_UNIT_ATTENTION | 2585 AC_INQ_CHANGED, daasync, periph, periph->path); 2586 2587 /* 2588 * Emit an attribute changed notification just in case 2589 * physical path information arrived before our async 2590 * event handler was registered, but after anyone attaching 2591 * to our disk device polled it. 2592 */ 2593 disk_attr_changed(softc->disk, "GEOM::physpath", M_NOWAIT); 2594 2595 /* 2596 * Schedule a periodic media polling events. 2597 */ 2598 callout_init_mtx(&softc->mediapoll_c, cam_periph_mtx(periph), 0); 2599 if ((softc->flags & DA_FLAG_PACK_REMOVABLE) && 2600 (cgd->inq_flags & SID_AEN) == 0 && 2601 da_poll_period != 0) 2602 callout_reset(&softc->mediapoll_c, da_poll_period * hz, 2603 damediapoll, periph); 2604 2605 xpt_schedule(periph, CAM_PRIORITY_DEV); 2606 2607 return(CAM_REQ_CMP); 2608 } 2609 2610 static int 2611 da_zone_bio_to_scsi(int disk_zone_cmd) 2612 { 2613 switch (disk_zone_cmd) { 2614 case DISK_ZONE_OPEN: 2615 return ZBC_OUT_SA_OPEN; 2616 case DISK_ZONE_CLOSE: 2617 return ZBC_OUT_SA_CLOSE; 2618 case DISK_ZONE_FINISH: 2619 return ZBC_OUT_SA_FINISH; 2620 case DISK_ZONE_RWP: 2621 return ZBC_OUT_SA_RWP; 2622 } 2623 2624 return -1; 2625 } 2626 2627 static int 2628 da_zone_cmd(struct cam_periph *periph, union ccb *ccb, struct bio *bp, 2629 int *queue_ccb) 2630 { 2631 struct da_softc *softc; 2632 int error; 2633 2634 error = 0; 2635 2636 if (bp->bio_cmd != BIO_ZONE) { 2637 error = EINVAL; 2638 goto bailout; 2639 } 2640 2641 softc = periph->softc; 2642 2643 switch (bp->bio_zone.zone_cmd) { 2644 case DISK_ZONE_OPEN: 2645 case DISK_ZONE_CLOSE: 2646 case DISK_ZONE_FINISH: 2647 case DISK_ZONE_RWP: { 2648 int zone_flags; 2649 int zone_sa; 2650 uint64_t lba; 2651 2652 zone_sa = da_zone_bio_to_scsi(bp->bio_zone.zone_cmd); 2653 if (zone_sa == -1) { 2654 xpt_print(periph->path, "Cannot translate zone " 2655 "cmd %#x to SCSI\n", bp->bio_zone.zone_cmd); 2656 error = EINVAL; 2657 goto bailout; 2658 } 2659 2660 zone_flags = 0; 2661 lba = bp->bio_zone.zone_params.rwp.id; 2662 2663 if (bp->bio_zone.zone_params.rwp.flags & 2664 DISK_ZONE_RWP_FLAG_ALL) 2665 zone_flags |= ZBC_OUT_ALL; 2666 2667 if (softc->zone_interface != DA_ZONE_IF_ATA_PASS) { 2668 scsi_zbc_out(&ccb->csio, 2669 /*retries*/ da_retry_count, 2670 /*cbfcnp*/ dadone, 2671 /*tag_action*/ MSG_SIMPLE_Q_TAG, 2672 /*service_action*/ zone_sa, 2673 /*zone_id*/ lba, 2674 /*zone_flags*/ zone_flags, 2675 /*data_ptr*/ NULL, 2676 /*dxfer_len*/ 0, 2677 /*sense_len*/ SSD_FULL_SIZE, 2678 /*timeout*/ da_default_timeout * 1000); 2679 } else { 2680 /* 2681 * Note that in this case, even though we can 2682 * technically use NCQ, we don't bother for several 2683 * reasons: 2684 * 1. It hasn't been tested on a SAT layer that 2685 * supports it. This is new as of SAT-4. 2686 * 2. Even when there is a SAT layer that supports 2687 * it, that SAT layer will also probably support 2688 * ZBC -> ZAC translation, since they are both 2689 * in the SAT-4 spec. 2690 * 3. Translation will likely be preferable to ATA 2691 * passthrough. LSI / Avago at least single 2692 * steps ATA passthrough commands in the HBA, 2693 * regardless of protocol, so unless that 2694 * changes, there is a performance penalty for 2695 * doing ATA passthrough no matter whether 2696 * you're using NCQ/FPDMA, DMA or PIO. 2697 * 4. It requires a 32-byte CDB, which at least at 2698 * this point in CAM requires a CDB pointer, which 2699 * would require us to allocate an additional bit 2700 * of storage separate from the CCB. 2701 */ 2702 error = scsi_ata_zac_mgmt_out(&ccb->csio, 2703 /*retries*/ da_retry_count, 2704 /*cbfcnp*/ dadone, 2705 /*tag_action*/ MSG_SIMPLE_Q_TAG, 2706 /*use_ncq*/ 0, 2707 /*zm_action*/ zone_sa, 2708 /*zone_id*/ lba, 2709 /*zone_flags*/ zone_flags, 2710 /*data_ptr*/ NULL, 2711 /*dxfer_len*/ 0, 2712 /*cdb_storage*/ NULL, 2713 /*cdb_storage_len*/ 0, 2714 /*sense_len*/ SSD_FULL_SIZE, 2715 /*timeout*/ da_default_timeout * 1000); 2716 if (error != 0) { 2717 error = EINVAL; 2718 xpt_print(periph->path, 2719 "scsi_ata_zac_mgmt_out() returned an " 2720 "error!"); 2721 goto bailout; 2722 } 2723 } 2724 *queue_ccb = 1; 2725 2726 break; 2727 } 2728 case DISK_ZONE_REPORT_ZONES: { 2729 uint8_t *rz_ptr; 2730 uint32_t num_entries, alloc_size; 2731 struct disk_zone_report *rep; 2732 2733 rep = &bp->bio_zone.zone_params.report; 2734 2735 num_entries = rep->entries_allocated; 2736 if (num_entries == 0) { 2737 xpt_print(periph->path, "No entries allocated for " 2738 "Report Zones request\n"); 2739 error = EINVAL; 2740 goto bailout; 2741 } 2742 alloc_size = sizeof(struct scsi_report_zones_hdr) + 2743 (sizeof(struct scsi_report_zones_desc) * num_entries); 2744 alloc_size = min(alloc_size, softc->disk->d_maxsize); 2745 rz_ptr = malloc(alloc_size, M_SCSIDA, M_NOWAIT | M_ZERO); 2746 if (rz_ptr == NULL) { 2747 xpt_print(periph->path, "Unable to allocate memory " 2748 "for Report Zones request\n"); 2749 error = ENOMEM; 2750 goto bailout; 2751 } 2752 2753 if (softc->zone_interface != DA_ZONE_IF_ATA_PASS) { 2754 scsi_zbc_in(&ccb->csio, 2755 /*retries*/ da_retry_count, 2756 /*cbcfnp*/ dadone, 2757 /*tag_action*/ MSG_SIMPLE_Q_TAG, 2758 /*service_action*/ ZBC_IN_SA_REPORT_ZONES, 2759 /*zone_start_lba*/ rep->starting_id, 2760 /*zone_options*/ rep->rep_options, 2761 /*data_ptr*/ rz_ptr, 2762 /*dxfer_len*/ alloc_size, 2763 /*sense_len*/ SSD_FULL_SIZE, 2764 /*timeout*/ da_default_timeout * 1000); 2765 } else { 2766 /* 2767 * Note that in this case, even though we can 2768 * technically use NCQ, we don't bother for several 2769 * reasons: 2770 * 1. It hasn't been tested on a SAT layer that 2771 * supports it. This is new as of SAT-4. 2772 * 2. Even when there is a SAT layer that supports 2773 * it, that SAT layer will also probably support 2774 * ZBC -> ZAC translation, since they are both 2775 * in the SAT-4 spec. 2776 * 3. Translation will likely be preferable to ATA 2777 * passthrough. LSI / Avago at least single 2778 * steps ATA passthrough commands in the HBA, 2779 * regardless of protocol, so unless that 2780 * changes, there is a performance penalty for 2781 * doing ATA passthrough no matter whether 2782 * you're using NCQ/FPDMA, DMA or PIO. 2783 * 4. It requires a 32-byte CDB, which at least at 2784 * this point in CAM requires a CDB pointer, which 2785 * would require us to allocate an additional bit 2786 * of storage separate from the CCB. 2787 */ 2788 error = scsi_ata_zac_mgmt_in(&ccb->csio, 2789 /*retries*/ da_retry_count, 2790 /*cbcfnp*/ dadone, 2791 /*tag_action*/ MSG_SIMPLE_Q_TAG, 2792 /*use_ncq*/ 0, 2793 /*zm_action*/ ATA_ZM_REPORT_ZONES, 2794 /*zone_id*/ rep->starting_id, 2795 /*zone_flags*/ rep->rep_options, 2796 /*data_ptr*/ rz_ptr, 2797 /*dxfer_len*/ alloc_size, 2798 /*cdb_storage*/ NULL, 2799 /*cdb_storage_len*/ 0, 2800 /*sense_len*/ SSD_FULL_SIZE, 2801 /*timeout*/ da_default_timeout * 1000); 2802 if (error != 0) { 2803 error = EINVAL; 2804 xpt_print(periph->path, 2805 "scsi_ata_zac_mgmt_in() returned an " 2806 "error!"); 2807 goto bailout; 2808 } 2809 } 2810 2811 /* 2812 * For BIO_ZONE, this isn't normally needed. However, it 2813 * is used by devstat_end_transaction_bio() to determine 2814 * how much data was transferred. 2815 */ 2816 /* 2817 * XXX KDM we have a problem. But I'm not sure how to fix 2818 * it. devstat uses bio_bcount - bio_resid to calculate 2819 * the amount of data transferred. The GEOM disk code 2820 * uses bio_length - bio_resid to calculate the amount of 2821 * data in bio_completed. We have different structure 2822 * sizes above and below the ada(4) driver. So, if we 2823 * use the sizes above, the amount transferred won't be 2824 * quite accurate for devstat. If we use different sizes 2825 * for bio_bcount and bio_length (above and below 2826 * respectively), then the residual needs to match one or 2827 * the other. Everything is calculated after the bio 2828 * leaves the driver, so changing the values around isn't 2829 * really an option. For now, just set the count to the 2830 * passed in length. This means that the calculations 2831 * above (e.g. bio_completed) will be correct, but the 2832 * amount of data reported to devstat will be slightly 2833 * under or overstated. 2834 */ 2835 bp->bio_bcount = bp->bio_length; 2836 2837 *queue_ccb = 1; 2838 2839 break; 2840 } 2841 case DISK_ZONE_GET_PARAMS: { 2842 struct disk_zone_disk_params *params; 2843 2844 params = &bp->bio_zone.zone_params.disk_params; 2845 bzero(params, sizeof(*params)); 2846 2847 switch (softc->zone_mode) { 2848 case DA_ZONE_DRIVE_MANAGED: 2849 params->zone_mode = DISK_ZONE_MODE_DRIVE_MANAGED; 2850 break; 2851 case DA_ZONE_HOST_AWARE: 2852 params->zone_mode = DISK_ZONE_MODE_HOST_AWARE; 2853 break; 2854 case DA_ZONE_HOST_MANAGED: 2855 params->zone_mode = DISK_ZONE_MODE_HOST_MANAGED; 2856 break; 2857 default: 2858 case DA_ZONE_NONE: 2859 params->zone_mode = DISK_ZONE_MODE_NONE; 2860 break; 2861 } 2862 2863 if (softc->zone_flags & DA_ZONE_FLAG_URSWRZ) 2864 params->flags |= DISK_ZONE_DISK_URSWRZ; 2865 2866 if (softc->zone_flags & DA_ZONE_FLAG_OPT_SEQ_SET) { 2867 params->optimal_seq_zones = softc->optimal_seq_zones; 2868 params->flags |= DISK_ZONE_OPT_SEQ_SET; 2869 } 2870 2871 if (softc->zone_flags & DA_ZONE_FLAG_OPT_NONSEQ_SET) { 2872 params->optimal_nonseq_zones = 2873 softc->optimal_nonseq_zones; 2874 params->flags |= DISK_ZONE_OPT_NONSEQ_SET; 2875 } 2876 2877 if (softc->zone_flags & DA_ZONE_FLAG_MAX_SEQ_SET) { 2878 params->max_seq_zones = softc->max_seq_zones; 2879 params->flags |= DISK_ZONE_MAX_SEQ_SET; 2880 } 2881 if (softc->zone_flags & DA_ZONE_FLAG_RZ_SUP) 2882 params->flags |= DISK_ZONE_RZ_SUP; 2883 2884 if (softc->zone_flags & DA_ZONE_FLAG_OPEN_SUP) 2885 params->flags |= DISK_ZONE_OPEN_SUP; 2886 2887 if (softc->zone_flags & DA_ZONE_FLAG_CLOSE_SUP) 2888 params->flags |= DISK_ZONE_CLOSE_SUP; 2889 2890 if (softc->zone_flags & DA_ZONE_FLAG_FINISH_SUP) 2891 params->flags |= DISK_ZONE_FINISH_SUP; 2892 2893 if (softc->zone_flags & DA_ZONE_FLAG_RWP_SUP) 2894 params->flags |= DISK_ZONE_RWP_SUP; 2895 break; 2896 } 2897 default: 2898 break; 2899 } 2900 bailout: 2901 return (error); 2902 } 2903 2904 static void 2905 dastart(struct cam_periph *periph, union ccb *start_ccb) 2906 { 2907 struct da_softc *softc; 2908 2909 softc = (struct da_softc *)periph->softc; 2910 2911 CAM_DEBUG(periph->path, CAM_DEBUG_TRACE, ("dastart\n")); 2912 2913 skipstate: 2914 switch (softc->state) { 2915 case DA_STATE_NORMAL: 2916 { 2917 struct bio *bp; 2918 uint8_t tag_code; 2919 2920 more: 2921 bp = cam_iosched_next_bio(softc->cam_iosched); 2922 if (bp == NULL) { 2923 if (cam_iosched_has_work_flags(softc->cam_iosched, DA_WORK_TUR)) { 2924 cam_iosched_clr_work_flags(softc->cam_iosched, DA_WORK_TUR); 2925 scsi_test_unit_ready(&start_ccb->csio, 2926 /*retries*/ da_retry_count, 2927 dadone, 2928 MSG_SIMPLE_Q_TAG, 2929 SSD_FULL_SIZE, 2930 da_default_timeout * 1000); 2931 start_ccb->ccb_h.ccb_bp = NULL; 2932 start_ccb->ccb_h.ccb_state = DA_CCB_TUR; 2933 xpt_action(start_ccb); 2934 } else 2935 xpt_release_ccb(start_ccb); 2936 break; 2937 } 2938 2939 if (bp->bio_cmd == BIO_DELETE) { 2940 if (softc->delete_func != NULL) { 2941 softc->delete_func(periph, start_ccb, bp); 2942 goto out; 2943 } else { 2944 /* Not sure this is possible, but failsafe by lying and saying "sure, done." */ 2945 biofinish(bp, NULL, 0); 2946 goto more; 2947 } 2948 } 2949 2950 if (cam_iosched_has_work_flags(softc->cam_iosched, DA_WORK_TUR)) { 2951 cam_iosched_clr_work_flags(softc->cam_iosched, DA_WORK_TUR); 2952 cam_periph_release_locked(periph); /* XXX is this still valid? I think so but unverified */ 2953 } 2954 2955 if ((bp->bio_flags & BIO_ORDERED) != 0 || 2956 (softc->flags & DA_FLAG_NEED_OTAG) != 0) { 2957 softc->flags &= ~DA_FLAG_NEED_OTAG; 2958 softc->flags |= DA_FLAG_WAS_OTAG; 2959 tag_code = MSG_ORDERED_Q_TAG; 2960 } else { 2961 tag_code = MSG_SIMPLE_Q_TAG; 2962 } 2963 2964 switch (bp->bio_cmd) { 2965 case BIO_WRITE: 2966 case BIO_READ: 2967 { 2968 void *data_ptr; 2969 int rw_op; 2970 2971 biotrack(bp, __func__); 2972 2973 if (bp->bio_cmd == BIO_WRITE) { 2974 softc->flags |= DA_FLAG_DIRTY; 2975 rw_op = SCSI_RW_WRITE; 2976 } else { 2977 rw_op = SCSI_RW_READ; 2978 } 2979 2980 data_ptr = bp->bio_data; 2981 if ((bp->bio_flags & (BIO_UNMAPPED|BIO_VLIST)) != 0) { 2982 rw_op |= SCSI_RW_BIO; 2983 data_ptr = bp; 2984 } 2985 2986 scsi_read_write(&start_ccb->csio, 2987 /*retries*/da_retry_count, 2988 /*cbfcnp*/dadone, 2989 /*tag_action*/tag_code, 2990 rw_op, 2991 /*byte2*/0, 2992 softc->minimum_cmd_size, 2993 /*lba*/bp->bio_pblkno, 2994 /*block_count*/bp->bio_bcount / 2995 softc->params.secsize, 2996 data_ptr, 2997 /*dxfer_len*/ bp->bio_bcount, 2998 /*sense_len*/SSD_FULL_SIZE, 2999 da_default_timeout * 1000); 3000 #if defined(BUF_TRACKING) || defined(FULL_BUF_TRACKING) 3001 start_ccb->csio.bio = bp; 3002 #endif 3003 break; 3004 } 3005 case BIO_FLUSH: 3006 /* 3007 * BIO_FLUSH doesn't currently communicate 3008 * range data, so we synchronize the cache 3009 * over the whole disk. We also force 3010 * ordered tag semantics the flush applies 3011 * to all previously queued I/O. 3012 */ 3013 scsi_synchronize_cache(&start_ccb->csio, 3014 /*retries*/1, 3015 /*cbfcnp*/dadone, 3016 MSG_ORDERED_Q_TAG, 3017 /*begin_lba*/0, 3018 /*lb_count*/0, 3019 SSD_FULL_SIZE, 3020 da_default_timeout*1000); 3021 break; 3022 case BIO_ZONE: { 3023 int error, queue_ccb; 3024 3025 queue_ccb = 0; 3026 3027 error = da_zone_cmd(periph, start_ccb, bp,&queue_ccb); 3028 if ((error != 0) 3029 || (queue_ccb == 0)) { 3030 biofinish(bp, NULL, error); 3031 xpt_release_ccb(start_ccb); 3032 return; 3033 } 3034 break; 3035 } 3036 } 3037 start_ccb->ccb_h.ccb_state = DA_CCB_BUFFER_IO; 3038 start_ccb->ccb_h.flags |= CAM_UNLOCKED; 3039 start_ccb->ccb_h.softtimeout = sbttotv(da_default_softtimeout); 3040 3041 out: 3042 LIST_INSERT_HEAD(&softc->pending_ccbs, 3043 &start_ccb->ccb_h, periph_links.le); 3044 3045 /* We expect a unit attention from this device */ 3046 if ((softc->flags & DA_FLAG_RETRY_UA) != 0) { 3047 start_ccb->ccb_h.ccb_state |= DA_CCB_RETRY_UA; 3048 softc->flags &= ~DA_FLAG_RETRY_UA; 3049 } 3050 3051 start_ccb->ccb_h.ccb_bp = bp; 3052 softc->refcount++; 3053 cam_periph_unlock(periph); 3054 xpt_action(start_ccb); 3055 cam_periph_lock(periph); 3056 softc->refcount--; 3057 3058 /* May have more work to do, so ensure we stay scheduled */ 3059 daschedule(periph); 3060 break; 3061 } 3062 case DA_STATE_PROBE_RC: 3063 { 3064 struct scsi_read_capacity_data *rcap; 3065 3066 rcap = (struct scsi_read_capacity_data *) 3067 malloc(sizeof(*rcap), M_SCSIDA, M_NOWAIT|M_ZERO); 3068 if (rcap == NULL) { 3069 printf("dastart: Couldn't malloc read_capacity data\n"); 3070 /* da_free_periph??? */ 3071 break; 3072 } 3073 scsi_read_capacity(&start_ccb->csio, 3074 /*retries*/da_retry_count, 3075 dadone, 3076 MSG_SIMPLE_Q_TAG, 3077 rcap, 3078 SSD_FULL_SIZE, 3079 /*timeout*/5000); 3080 start_ccb->ccb_h.ccb_bp = NULL; 3081 start_ccb->ccb_h.ccb_state = DA_CCB_PROBE_RC; 3082 xpt_action(start_ccb); 3083 break; 3084 } 3085 case DA_STATE_PROBE_RC16: 3086 { 3087 struct scsi_read_capacity_data_long *rcaplong; 3088 3089 rcaplong = (struct scsi_read_capacity_data_long *) 3090 malloc(sizeof(*rcaplong), M_SCSIDA, M_NOWAIT|M_ZERO); 3091 if (rcaplong == NULL) { 3092 printf("dastart: Couldn't malloc read_capacity data\n"); 3093 /* da_free_periph??? */ 3094 break; 3095 } 3096 scsi_read_capacity_16(&start_ccb->csio, 3097 /*retries*/ da_retry_count, 3098 /*cbfcnp*/ dadone, 3099 /*tag_action*/ MSG_SIMPLE_Q_TAG, 3100 /*lba*/ 0, 3101 /*reladr*/ 0, 3102 /*pmi*/ 0, 3103 /*rcap_buf*/ (uint8_t *)rcaplong, 3104 /*rcap_buf_len*/ sizeof(*rcaplong), 3105 /*sense_len*/ SSD_FULL_SIZE, 3106 /*timeout*/ da_default_timeout * 1000); 3107 start_ccb->ccb_h.ccb_bp = NULL; 3108 start_ccb->ccb_h.ccb_state = DA_CCB_PROBE_RC16; 3109 xpt_action(start_ccb); 3110 break; 3111 } 3112 case DA_STATE_PROBE_LBP: 3113 { 3114 struct scsi_vpd_logical_block_prov *lbp; 3115 3116 if (!scsi_vpd_supported_page(periph, SVPD_LBP)) { 3117 /* 3118 * If we get here we don't support any SBC-3 delete 3119 * methods with UNMAP as the Logical Block Provisioning 3120 * VPD page support is required for devices which 3121 * support it according to T10/1799-D Revision 31 3122 * however older revisions of the spec don't mandate 3123 * this so we currently don't remove these methods 3124 * from the available set. 3125 */ 3126 softc->state = DA_STATE_PROBE_BLK_LIMITS; 3127 goto skipstate; 3128 } 3129 3130 lbp = (struct scsi_vpd_logical_block_prov *) 3131 malloc(sizeof(*lbp), M_SCSIDA, M_NOWAIT|M_ZERO); 3132 3133 if (lbp == NULL) { 3134 printf("dastart: Couldn't malloc lbp data\n"); 3135 /* da_free_periph??? */ 3136 break; 3137 } 3138 3139 scsi_inquiry(&start_ccb->csio, 3140 /*retries*/da_retry_count, 3141 /*cbfcnp*/dadone, 3142 /*tag_action*/MSG_SIMPLE_Q_TAG, 3143 /*inq_buf*/(u_int8_t *)lbp, 3144 /*inq_len*/sizeof(*lbp), 3145 /*evpd*/TRUE, 3146 /*page_code*/SVPD_LBP, 3147 /*sense_len*/SSD_MIN_SIZE, 3148 /*timeout*/da_default_timeout * 1000); 3149 start_ccb->ccb_h.ccb_bp = NULL; 3150 start_ccb->ccb_h.ccb_state = DA_CCB_PROBE_LBP; 3151 xpt_action(start_ccb); 3152 break; 3153 } 3154 case DA_STATE_PROBE_BLK_LIMITS: 3155 { 3156 struct scsi_vpd_block_limits *block_limits; 3157 3158 if (!scsi_vpd_supported_page(periph, SVPD_BLOCK_LIMITS)) { 3159 /* Not supported skip to next probe */ 3160 softc->state = DA_STATE_PROBE_BDC; 3161 goto skipstate; 3162 } 3163 3164 block_limits = (struct scsi_vpd_block_limits *) 3165 malloc(sizeof(*block_limits), M_SCSIDA, M_NOWAIT|M_ZERO); 3166 3167 if (block_limits == NULL) { 3168 printf("dastart: Couldn't malloc block_limits data\n"); 3169 /* da_free_periph??? */ 3170 break; 3171 } 3172 3173 scsi_inquiry(&start_ccb->csio, 3174 /*retries*/da_retry_count, 3175 /*cbfcnp*/dadone, 3176 /*tag_action*/MSG_SIMPLE_Q_TAG, 3177 /*inq_buf*/(u_int8_t *)block_limits, 3178 /*inq_len*/sizeof(*block_limits), 3179 /*evpd*/TRUE, 3180 /*page_code*/SVPD_BLOCK_LIMITS, 3181 /*sense_len*/SSD_MIN_SIZE, 3182 /*timeout*/da_default_timeout * 1000); 3183 start_ccb->ccb_h.ccb_bp = NULL; 3184 start_ccb->ccb_h.ccb_state = DA_CCB_PROBE_BLK_LIMITS; 3185 xpt_action(start_ccb); 3186 break; 3187 } 3188 case DA_STATE_PROBE_BDC: 3189 { 3190 struct scsi_vpd_block_characteristics *bdc; 3191 3192 if (!scsi_vpd_supported_page(periph, SVPD_BDC)) { 3193 softc->state = DA_STATE_PROBE_ATA; 3194 goto skipstate; 3195 } 3196 3197 bdc = (struct scsi_vpd_block_characteristics *) 3198 malloc(sizeof(*bdc), M_SCSIDA, M_NOWAIT|M_ZERO); 3199 3200 if (bdc == NULL) { 3201 printf("dastart: Couldn't malloc bdc data\n"); 3202 /* da_free_periph??? */ 3203 break; 3204 } 3205 3206 scsi_inquiry(&start_ccb->csio, 3207 /*retries*/da_retry_count, 3208 /*cbfcnp*/dadone, 3209 /*tag_action*/MSG_SIMPLE_Q_TAG, 3210 /*inq_buf*/(u_int8_t *)bdc, 3211 /*inq_len*/sizeof(*bdc), 3212 /*evpd*/TRUE, 3213 /*page_code*/SVPD_BDC, 3214 /*sense_len*/SSD_MIN_SIZE, 3215 /*timeout*/da_default_timeout * 1000); 3216 start_ccb->ccb_h.ccb_bp = NULL; 3217 start_ccb->ccb_h.ccb_state = DA_CCB_PROBE_BDC; 3218 xpt_action(start_ccb); 3219 break; 3220 } 3221 case DA_STATE_PROBE_ATA: 3222 { 3223 struct ata_params *ata_params; 3224 3225 if (!scsi_vpd_supported_page(periph, SVPD_ATA_INFORMATION)) { 3226 if ((softc->zone_mode == DA_ZONE_HOST_AWARE) 3227 || (softc->zone_mode == DA_ZONE_HOST_MANAGED)) { 3228 /* 3229 * Note that if the ATA VPD page isn't 3230 * supported, we aren't talking to an ATA 3231 * device anyway. Support for that VPD 3232 * page is mandatory for SCSI to ATA (SAT) 3233 * translation layers. 3234 */ 3235 softc->state = DA_STATE_PROBE_ZONE; 3236 goto skipstate; 3237 } 3238 daprobedone(periph, start_ccb); 3239 break; 3240 } 3241 3242 ata_params = (struct ata_params*) 3243 malloc(sizeof(*ata_params), M_SCSIDA,M_NOWAIT|M_ZERO); 3244 3245 if (ata_params == NULL) { 3246 xpt_print(periph->path, "Couldn't malloc ata_params " 3247 "data\n"); 3248 /* da_free_periph??? */ 3249 break; 3250 } 3251 3252 scsi_ata_identify(&start_ccb->csio, 3253 /*retries*/da_retry_count, 3254 /*cbfcnp*/dadone, 3255 /*tag_action*/MSG_SIMPLE_Q_TAG, 3256 /*data_ptr*/(u_int8_t *)ata_params, 3257 /*dxfer_len*/sizeof(*ata_params), 3258 /*sense_len*/SSD_FULL_SIZE, 3259 /*timeout*/da_default_timeout * 1000); 3260 start_ccb->ccb_h.ccb_bp = NULL; 3261 start_ccb->ccb_h.ccb_state = DA_CCB_PROBE_ATA; 3262 xpt_action(start_ccb); 3263 break; 3264 } 3265 case DA_STATE_PROBE_ATA_LOGDIR: 3266 { 3267 struct ata_gp_log_dir *log_dir; 3268 int retval; 3269 3270 retval = 0; 3271 3272 if ((softc->flags & DA_FLAG_CAN_ATA_LOG) == 0) { 3273 /* 3274 * If we don't have log support, not much point in 3275 * trying to probe zone support. 3276 */ 3277 daprobedone(periph, start_ccb); 3278 break; 3279 } 3280 3281 /* 3282 * If we have an ATA device (the SCSI ATA Information VPD 3283 * page should be present and the ATA identify should have 3284 * succeeded) and it supports logs, ask for the log directory. 3285 */ 3286 3287 log_dir = malloc(sizeof(*log_dir), M_SCSIDA, M_NOWAIT|M_ZERO); 3288 if (log_dir == NULL) { 3289 xpt_print(periph->path, "Couldn't malloc log_dir " 3290 "data\n"); 3291 daprobedone(periph, start_ccb); 3292 break; 3293 } 3294 3295 retval = scsi_ata_read_log(&start_ccb->csio, 3296 /*retries*/ da_retry_count, 3297 /*cbfcnp*/ dadone, 3298 /*tag_action*/ MSG_SIMPLE_Q_TAG, 3299 /*log_address*/ ATA_LOG_DIRECTORY, 3300 /*page_number*/ 0, 3301 /*block_count*/ 1, 3302 /*protocol*/ softc->flags & DA_FLAG_CAN_ATA_DMA ? 3303 AP_PROTO_DMA : AP_PROTO_PIO_IN, 3304 /*data_ptr*/ (uint8_t *)log_dir, 3305 /*dxfer_len*/ sizeof(*log_dir), 3306 /*sense_len*/ SSD_FULL_SIZE, 3307 /*timeout*/ da_default_timeout * 1000); 3308 3309 if (retval != 0) { 3310 xpt_print(periph->path, "scsi_ata_read_log() failed!"); 3311 free(log_dir, M_SCSIDA); 3312 daprobedone(periph, start_ccb); 3313 break; 3314 } 3315 start_ccb->ccb_h.ccb_bp = NULL; 3316 start_ccb->ccb_h.ccb_state = DA_CCB_PROBE_ATA_LOGDIR; 3317 xpt_action(start_ccb); 3318 break; 3319 } 3320 case DA_STATE_PROBE_ATA_IDDIR: 3321 { 3322 struct ata_identify_log_pages *id_dir; 3323 int retval; 3324 3325 retval = 0; 3326 3327 /* 3328 * Check here to see whether the Identify Device log is 3329 * supported in the directory of logs. If so, continue 3330 * with requesting the log of identify device pages. 3331 */ 3332 if ((softc->flags & DA_FLAG_CAN_ATA_IDLOG) == 0) { 3333 daprobedone(periph, start_ccb); 3334 break; 3335 } 3336 3337 id_dir = malloc(sizeof(*id_dir), M_SCSIDA, M_NOWAIT | M_ZERO); 3338 if (id_dir == NULL) { 3339 xpt_print(periph->path, "Couldn't malloc id_dir " 3340 "data\n"); 3341 daprobedone(periph, start_ccb); 3342 break; 3343 } 3344 3345 retval = scsi_ata_read_log(&start_ccb->csio, 3346 /*retries*/ da_retry_count, 3347 /*cbfcnp*/ dadone, 3348 /*tag_action*/ MSG_SIMPLE_Q_TAG, 3349 /*log_address*/ ATA_IDENTIFY_DATA_LOG, 3350 /*page_number*/ ATA_IDL_PAGE_LIST, 3351 /*block_count*/ 1, 3352 /*protocol*/ softc->flags & DA_FLAG_CAN_ATA_DMA ? 3353 AP_PROTO_DMA : AP_PROTO_PIO_IN, 3354 /*data_ptr*/ (uint8_t *)id_dir, 3355 /*dxfer_len*/ sizeof(*id_dir), 3356 /*sense_len*/ SSD_FULL_SIZE, 3357 /*timeout*/ da_default_timeout * 1000); 3358 3359 if (retval != 0) { 3360 xpt_print(periph->path, "scsi_ata_read_log() failed!"); 3361 free(id_dir, M_SCSIDA); 3362 daprobedone(periph, start_ccb); 3363 break; 3364 } 3365 start_ccb->ccb_h.ccb_bp = NULL; 3366 start_ccb->ccb_h.ccb_state = DA_CCB_PROBE_ATA_IDDIR; 3367 xpt_action(start_ccb); 3368 break; 3369 } 3370 case DA_STATE_PROBE_ATA_SUP: 3371 { 3372 struct ata_identify_log_sup_cap *sup_cap; 3373 int retval; 3374 3375 retval = 0; 3376 3377 /* 3378 * Check here to see whether the Supported Capabilities log 3379 * is in the list of Identify Device logs. 3380 */ 3381 if ((softc->flags & DA_FLAG_CAN_ATA_SUPCAP) == 0) { 3382 daprobedone(periph, start_ccb); 3383 break; 3384 } 3385 3386 sup_cap = malloc(sizeof(*sup_cap), M_SCSIDA, M_NOWAIT|M_ZERO); 3387 if (sup_cap == NULL) { 3388 xpt_print(periph->path, "Couldn't malloc sup_cap " 3389 "data\n"); 3390 daprobedone(periph, start_ccb); 3391 break; 3392 } 3393 3394 retval = scsi_ata_read_log(&start_ccb->csio, 3395 /*retries*/ da_retry_count, 3396 /*cbfcnp*/ dadone, 3397 /*tag_action*/ MSG_SIMPLE_Q_TAG, 3398 /*log_address*/ ATA_IDENTIFY_DATA_LOG, 3399 /*page_number*/ ATA_IDL_SUP_CAP, 3400 /*block_count*/ 1, 3401 /*protocol*/ softc->flags & DA_FLAG_CAN_ATA_DMA ? 3402 AP_PROTO_DMA : AP_PROTO_PIO_IN, 3403 /*data_ptr*/ (uint8_t *)sup_cap, 3404 /*dxfer_len*/ sizeof(*sup_cap), 3405 /*sense_len*/ SSD_FULL_SIZE, 3406 /*timeout*/ da_default_timeout * 1000); 3407 3408 if (retval != 0) { 3409 xpt_print(periph->path, "scsi_ata_read_log() failed!"); 3410 free(sup_cap, M_SCSIDA); 3411 daprobedone(periph, start_ccb); 3412 break; 3413 3414 } 3415 3416 start_ccb->ccb_h.ccb_bp = NULL; 3417 start_ccb->ccb_h.ccb_state = DA_CCB_PROBE_ATA_SUP; 3418 xpt_action(start_ccb); 3419 break; 3420 } 3421 case DA_STATE_PROBE_ATA_ZONE: 3422 { 3423 struct ata_zoned_info_log *ata_zone; 3424 int retval; 3425 3426 retval = 0; 3427 3428 /* 3429 * Check here to see whether the zoned device information 3430 * page is supported. If so, continue on to request it. 3431 * If not, skip to DA_STATE_PROBE_LOG or done. 3432 */ 3433 if ((softc->flags & DA_FLAG_CAN_ATA_ZONE) == 0) { 3434 daprobedone(periph, start_ccb); 3435 break; 3436 } 3437 ata_zone = malloc(sizeof(*ata_zone), M_SCSIDA, 3438 M_NOWAIT|M_ZERO); 3439 if (ata_zone == NULL) { 3440 xpt_print(periph->path, "Couldn't malloc ata_zone " 3441 "data\n"); 3442 daprobedone(periph, start_ccb); 3443 break; 3444 } 3445 3446 retval = scsi_ata_read_log(&start_ccb->csio, 3447 /*retries*/ da_retry_count, 3448 /*cbfcnp*/ dadone, 3449 /*tag_action*/ MSG_SIMPLE_Q_TAG, 3450 /*log_address*/ ATA_IDENTIFY_DATA_LOG, 3451 /*page_number*/ ATA_IDL_ZDI, 3452 /*block_count*/ 1, 3453 /*protocol*/ softc->flags & DA_FLAG_CAN_ATA_DMA ? 3454 AP_PROTO_DMA : AP_PROTO_PIO_IN, 3455 /*data_ptr*/ (uint8_t *)ata_zone, 3456 /*dxfer_len*/ sizeof(*ata_zone), 3457 /*sense_len*/ SSD_FULL_SIZE, 3458 /*timeout*/ da_default_timeout * 1000); 3459 3460 if (retval != 0) { 3461 xpt_print(periph->path, "scsi_ata_read_log() failed!"); 3462 free(ata_zone, M_SCSIDA); 3463 daprobedone(periph, start_ccb); 3464 break; 3465 } 3466 start_ccb->ccb_h.ccb_bp = NULL; 3467 start_ccb->ccb_h.ccb_state = DA_CCB_PROBE_ATA_ZONE; 3468 xpt_action(start_ccb); 3469 3470 break; 3471 } 3472 case DA_STATE_PROBE_ZONE: 3473 { 3474 struct scsi_vpd_zoned_bdc *bdc; 3475 3476 /* 3477 * Note that this page will be supported for SCSI protocol 3478 * devices that support ZBC (SMR devices), as well as ATA 3479 * protocol devices that are behind a SAT (SCSI to ATA 3480 * Translation) layer that supports converting ZBC commands 3481 * to their ZAC equivalents. 3482 */ 3483 if (!scsi_vpd_supported_page(periph, SVPD_ZONED_BDC)) { 3484 daprobedone(periph, start_ccb); 3485 break; 3486 } 3487 bdc = (struct scsi_vpd_zoned_bdc *) 3488 malloc(sizeof(*bdc), M_SCSIDA, M_NOWAIT|M_ZERO); 3489 3490 if (bdc == NULL) { 3491 xpt_release_ccb(start_ccb); 3492 xpt_print(periph->path, "Couldn't malloc zone VPD " 3493 "data\n"); 3494 break; 3495 } 3496 scsi_inquiry(&start_ccb->csio, 3497 /*retries*/da_retry_count, 3498 /*cbfcnp*/dadone, 3499 /*tag_action*/MSG_SIMPLE_Q_TAG, 3500 /*inq_buf*/(u_int8_t *)bdc, 3501 /*inq_len*/sizeof(*bdc), 3502 /*evpd*/TRUE, 3503 /*page_code*/SVPD_ZONED_BDC, 3504 /*sense_len*/SSD_FULL_SIZE, 3505 /*timeout*/da_default_timeout * 1000); 3506 start_ccb->ccb_h.ccb_bp = NULL; 3507 start_ccb->ccb_h.ccb_state = DA_CCB_PROBE_ZONE; 3508 xpt_action(start_ccb); 3509 break; 3510 } 3511 } 3512 } 3513 3514 /* 3515 * In each of the methods below, while its the caller's 3516 * responsibility to ensure the request will fit into a 3517 * single device request, we might have changed the delete 3518 * method due to the device incorrectly advertising either 3519 * its supported methods or limits. 3520 * 3521 * To prevent this causing further issues we validate the 3522 * against the methods limits, and warn which would 3523 * otherwise be unnecessary. 3524 */ 3525 static void 3526 da_delete_unmap(struct cam_periph *periph, union ccb *ccb, struct bio *bp) 3527 { 3528 struct da_softc *softc = (struct da_softc *)periph->softc;; 3529 struct bio *bp1; 3530 uint8_t *buf = softc->unmap_buf; 3531 struct scsi_unmap_desc *d = (void *)&buf[UNMAP_HEAD_SIZE]; 3532 uint64_t lba, lastlba = (uint64_t)-1; 3533 uint64_t totalcount = 0; 3534 uint64_t count; 3535 uint32_t c, lastcount = 0, ranges = 0; 3536 3537 /* 3538 * Currently this doesn't take the UNMAP 3539 * Granularity and Granularity Alignment 3540 * fields into account. 3541 * 3542 * This could result in both unoptimal unmap 3543 * requests as as well as UNMAP calls unmapping 3544 * fewer LBA's than requested. 3545 */ 3546 3547 bzero(softc->unmap_buf, sizeof(softc->unmap_buf)); 3548 bp1 = bp; 3549 do { 3550 /* 3551 * Note: ada and da are different in how they store the 3552 * pending bp's in a trim. ada stores all of them in the 3553 * trim_req.bps. da stores all but the first one in the 3554 * delete_run_queue. ada then completes all the bps in 3555 * its adadone() loop. da completes all the bps in the 3556 * delete_run_queue in dadone, and relies on the biodone 3557 * after to complete. This should be reconciled since there's 3558 * no real reason to do it differently. XXX 3559 */ 3560 if (bp1 != bp) 3561 bioq_insert_tail(&softc->delete_run_queue, bp1); 3562 lba = bp1->bio_pblkno; 3563 count = bp1->bio_bcount / softc->params.secsize; 3564 3565 /* Try to extend the previous range. */ 3566 if (lba == lastlba) { 3567 c = omin(count, UNMAP_RANGE_MAX - lastcount); 3568 lastlba += c; 3569 lastcount += c; 3570 scsi_ulto4b(lastcount, d[ranges - 1].length); 3571 count -= c; 3572 lba += c; 3573 totalcount += c; 3574 } else if ((softc->quirks & DA_Q_STRICT_UNMAP) && 3575 softc->unmap_gran != 0) { 3576 /* Align length of the previous range. */ 3577 if ((c = lastcount % softc->unmap_gran) != 0) { 3578 if (lastcount <= c) { 3579 totalcount -= lastcount; 3580 lastlba = (uint64_t)-1; 3581 lastcount = 0; 3582 ranges--; 3583 } else { 3584 totalcount -= c; 3585 lastlba -= c; 3586 lastcount -= c; 3587 scsi_ulto4b(lastcount, d[ranges - 1].length); 3588 } 3589 } 3590 /* Align beginning of the new range. */ 3591 c = (lba - softc->unmap_gran_align) % softc->unmap_gran; 3592 if (c != 0) { 3593 c = softc->unmap_gran - c; 3594 if (count <= c) { 3595 count = 0; 3596 } else { 3597 lba += c; 3598 count -= c; 3599 } 3600 } 3601 } 3602 3603 while (count > 0) { 3604 c = omin(count, UNMAP_RANGE_MAX); 3605 if (totalcount + c > softc->unmap_max_lba || 3606 ranges >= softc->unmap_max_ranges) { 3607 xpt_print(periph->path, 3608 "%s issuing short delete %ld > %ld" 3609 "|| %d >= %d", 3610 da_delete_method_desc[softc->delete_method], 3611 totalcount + c, softc->unmap_max_lba, 3612 ranges, softc->unmap_max_ranges); 3613 break; 3614 } 3615 scsi_u64to8b(lba, d[ranges].lba); 3616 scsi_ulto4b(c, d[ranges].length); 3617 lba += c; 3618 totalcount += c; 3619 ranges++; 3620 count -= c; 3621 lastlba = lba; 3622 lastcount = c; 3623 } 3624 bp1 = cam_iosched_next_trim(softc->cam_iosched); 3625 if (bp1 == NULL) 3626 break; 3627 if (ranges >= softc->unmap_max_ranges || 3628 totalcount + bp1->bio_bcount / 3629 softc->params.secsize > softc->unmap_max_lba) { 3630 cam_iosched_put_back_trim(softc->cam_iosched, bp1); 3631 break; 3632 } 3633 } while (1); 3634 3635 /* Align length of the last range. */ 3636 if ((softc->quirks & DA_Q_STRICT_UNMAP) && softc->unmap_gran != 0 && 3637 (c = lastcount % softc->unmap_gran) != 0) { 3638 if (lastcount <= c) 3639 ranges--; 3640 else 3641 scsi_ulto4b(lastcount - c, d[ranges - 1].length); 3642 } 3643 3644 scsi_ulto2b(ranges * 16 + 6, &buf[0]); 3645 scsi_ulto2b(ranges * 16, &buf[2]); 3646 3647 scsi_unmap(&ccb->csio, 3648 /*retries*/da_retry_count, 3649 /*cbfcnp*/dadone, 3650 /*tag_action*/MSG_SIMPLE_Q_TAG, 3651 /*byte2*/0, 3652 /*data_ptr*/ buf, 3653 /*dxfer_len*/ ranges * 16 + 8, 3654 /*sense_len*/SSD_FULL_SIZE, 3655 da_default_timeout * 1000); 3656 ccb->ccb_h.ccb_state = DA_CCB_DELETE; 3657 ccb->ccb_h.flags |= CAM_UNLOCKED; 3658 cam_iosched_submit_trim(softc->cam_iosched); 3659 } 3660 3661 static void 3662 da_delete_trim(struct cam_periph *periph, union ccb *ccb, struct bio *bp) 3663 { 3664 struct da_softc *softc = (struct da_softc *)periph->softc; 3665 struct bio *bp1; 3666 uint8_t *buf = softc->unmap_buf; 3667 uint64_t lastlba = (uint64_t)-1; 3668 uint64_t count; 3669 uint64_t lba; 3670 uint32_t lastcount = 0, c, requestcount; 3671 int ranges = 0, off, block_count; 3672 3673 bzero(softc->unmap_buf, sizeof(softc->unmap_buf)); 3674 bp1 = bp; 3675 do { 3676 if (bp1 != bp)//XXX imp XXX 3677 bioq_insert_tail(&softc->delete_run_queue, bp1); 3678 lba = bp1->bio_pblkno; 3679 count = bp1->bio_bcount / softc->params.secsize; 3680 requestcount = count; 3681 3682 /* Try to extend the previous range. */ 3683 if (lba == lastlba) { 3684 c = omin(count, ATA_DSM_RANGE_MAX - lastcount); 3685 lastcount += c; 3686 off = (ranges - 1) * 8; 3687 buf[off + 6] = lastcount & 0xff; 3688 buf[off + 7] = (lastcount >> 8) & 0xff; 3689 count -= c; 3690 lba += c; 3691 } 3692 3693 while (count > 0) { 3694 c = omin(count, ATA_DSM_RANGE_MAX); 3695 off = ranges * 8; 3696 3697 buf[off + 0] = lba & 0xff; 3698 buf[off + 1] = (lba >> 8) & 0xff; 3699 buf[off + 2] = (lba >> 16) & 0xff; 3700 buf[off + 3] = (lba >> 24) & 0xff; 3701 buf[off + 4] = (lba >> 32) & 0xff; 3702 buf[off + 5] = (lba >> 40) & 0xff; 3703 buf[off + 6] = c & 0xff; 3704 buf[off + 7] = (c >> 8) & 0xff; 3705 lba += c; 3706 ranges++; 3707 count -= c; 3708 lastcount = c; 3709 if (count != 0 && ranges == softc->trim_max_ranges) { 3710 xpt_print(periph->path, 3711 "%s issuing short delete %ld > %ld\n", 3712 da_delete_method_desc[softc->delete_method], 3713 requestcount, 3714 (softc->trim_max_ranges - ranges) * 3715 ATA_DSM_RANGE_MAX); 3716 break; 3717 } 3718 } 3719 lastlba = lba; 3720 bp1 = cam_iosched_next_trim(softc->cam_iosched); 3721 if (bp1 == NULL) 3722 break; 3723 if (bp1->bio_bcount / softc->params.secsize > 3724 (softc->trim_max_ranges - ranges) * ATA_DSM_RANGE_MAX) { 3725 cam_iosched_put_back_trim(softc->cam_iosched, bp1); 3726 break; 3727 } 3728 } while (1); 3729 3730 block_count = howmany(ranges, ATA_DSM_BLK_RANGES); 3731 scsi_ata_trim(&ccb->csio, 3732 /*retries*/da_retry_count, 3733 /*cbfcnp*/dadone, 3734 /*tag_action*/MSG_SIMPLE_Q_TAG, 3735 block_count, 3736 /*data_ptr*/buf, 3737 /*dxfer_len*/block_count * ATA_DSM_BLK_SIZE, 3738 /*sense_len*/SSD_FULL_SIZE, 3739 da_default_timeout * 1000); 3740 ccb->ccb_h.ccb_state = DA_CCB_DELETE; 3741 ccb->ccb_h.flags |= CAM_UNLOCKED; 3742 cam_iosched_submit_trim(softc->cam_iosched); 3743 } 3744 3745 /* 3746 * We calculate ws_max_blks here based off d_delmaxsize instead 3747 * of using softc->ws_max_blks as it is absolute max for the 3748 * device not the protocol max which may well be lower. 3749 */ 3750 static void 3751 da_delete_ws(struct cam_periph *periph, union ccb *ccb, struct bio *bp) 3752 { 3753 struct da_softc *softc; 3754 struct bio *bp1; 3755 uint64_t ws_max_blks; 3756 uint64_t lba; 3757 uint64_t count; /* forward compat with WS32 */ 3758 3759 softc = (struct da_softc *)periph->softc; 3760 ws_max_blks = softc->disk->d_delmaxsize / softc->params.secsize; 3761 lba = bp->bio_pblkno; 3762 count = 0; 3763 bp1 = bp; 3764 do { 3765 if (bp1 != bp)//XXX imp XXX 3766 bioq_insert_tail(&softc->delete_run_queue, bp1); 3767 count += bp1->bio_bcount / softc->params.secsize; 3768 if (count > ws_max_blks) { 3769 xpt_print(periph->path, 3770 "%s issuing short delete %ld > %ld\n", 3771 da_delete_method_desc[softc->delete_method], 3772 count, ws_max_blks); 3773 count = omin(count, ws_max_blks); 3774 break; 3775 } 3776 bp1 = cam_iosched_next_trim(softc->cam_iosched); 3777 if (bp1 == NULL) 3778 break; 3779 if (lba + count != bp1->bio_pblkno || 3780 count + bp1->bio_bcount / 3781 softc->params.secsize > ws_max_blks) { 3782 cam_iosched_put_back_trim(softc->cam_iosched, bp1); 3783 break; 3784 } 3785 } while (1); 3786 3787 scsi_write_same(&ccb->csio, 3788 /*retries*/da_retry_count, 3789 /*cbfcnp*/dadone, 3790 /*tag_action*/MSG_SIMPLE_Q_TAG, 3791 /*byte2*/softc->delete_method == 3792 DA_DELETE_ZERO ? 0 : SWS_UNMAP, 3793 softc->delete_method == DA_DELETE_WS16 ? 16 : 10, 3794 /*lba*/lba, 3795 /*block_count*/count, 3796 /*data_ptr*/ __DECONST(void *, zero_region), 3797 /*dxfer_len*/ softc->params.secsize, 3798 /*sense_len*/SSD_FULL_SIZE, 3799 da_default_timeout * 1000); 3800 ccb->ccb_h.ccb_state = DA_CCB_DELETE; 3801 ccb->ccb_h.flags |= CAM_UNLOCKED; 3802 cam_iosched_submit_trim(softc->cam_iosched); 3803 } 3804 3805 static int 3806 cmd6workaround(union ccb *ccb) 3807 { 3808 struct scsi_rw_6 cmd6; 3809 struct scsi_rw_10 *cmd10; 3810 struct da_softc *softc; 3811 u_int8_t *cdb; 3812 struct bio *bp; 3813 int frozen; 3814 3815 cdb = ccb->csio.cdb_io.cdb_bytes; 3816 softc = (struct da_softc *)xpt_path_periph(ccb->ccb_h.path)->softc; 3817 3818 if (ccb->ccb_h.ccb_state == DA_CCB_DELETE) { 3819 da_delete_methods old_method = softc->delete_method; 3820 3821 /* 3822 * Typically there are two reasons for failure here 3823 * 1. Delete method was detected as supported but isn't 3824 * 2. Delete failed due to invalid params e.g. too big 3825 * 3826 * While we will attempt to choose an alternative delete method 3827 * this may result in short deletes if the existing delete 3828 * requests from geom are big for the new method chosen. 3829 * 3830 * This method assumes that the error which triggered this 3831 * will not retry the io otherwise a panic will occur 3832 */ 3833 dadeleteflag(softc, old_method, 0); 3834 dadeletemethodchoose(softc, DA_DELETE_DISABLE); 3835 if (softc->delete_method == DA_DELETE_DISABLE) 3836 xpt_print(ccb->ccb_h.path, 3837 "%s failed, disabling BIO_DELETE\n", 3838 da_delete_method_desc[old_method]); 3839 else 3840 xpt_print(ccb->ccb_h.path, 3841 "%s failed, switching to %s BIO_DELETE\n", 3842 da_delete_method_desc[old_method], 3843 da_delete_method_desc[softc->delete_method]); 3844 3845 while ((bp = bioq_takefirst(&softc->delete_run_queue)) != NULL) 3846 cam_iosched_queue_work(softc->cam_iosched, bp); 3847 cam_iosched_queue_work(softc->cam_iosched, 3848 (struct bio *)ccb->ccb_h.ccb_bp); 3849 ccb->ccb_h.ccb_bp = NULL; 3850 return (0); 3851 } 3852 3853 /* Detect unsupported PREVENT ALLOW MEDIUM REMOVAL. */ 3854 if ((ccb->ccb_h.flags & CAM_CDB_POINTER) == 0 && 3855 (*cdb == PREVENT_ALLOW) && 3856 (softc->quirks & DA_Q_NO_PREVENT) == 0) { 3857 if (bootverbose) 3858 xpt_print(ccb->ccb_h.path, 3859 "PREVENT ALLOW MEDIUM REMOVAL not supported.\n"); 3860 softc->quirks |= DA_Q_NO_PREVENT; 3861 return (0); 3862 } 3863 3864 /* Detect unsupported SYNCHRONIZE CACHE(10). */ 3865 if ((ccb->ccb_h.flags & CAM_CDB_POINTER) == 0 && 3866 (*cdb == SYNCHRONIZE_CACHE) && 3867 (softc->quirks & DA_Q_NO_SYNC_CACHE) == 0) { 3868 if (bootverbose) 3869 xpt_print(ccb->ccb_h.path, 3870 "SYNCHRONIZE CACHE(10) not supported.\n"); 3871 softc->quirks |= DA_Q_NO_SYNC_CACHE; 3872 softc->disk->d_flags &= ~DISKFLAG_CANFLUSHCACHE; 3873 return (0); 3874 } 3875 3876 /* Translation only possible if CDB is an array and cmd is R/W6 */ 3877 if ((ccb->ccb_h.flags & CAM_CDB_POINTER) != 0 || 3878 (*cdb != READ_6 && *cdb != WRITE_6)) 3879 return 0; 3880 3881 xpt_print(ccb->ccb_h.path, "READ(6)/WRITE(6) not supported, " 3882 "increasing minimum_cmd_size to 10.\n"); 3883 softc->minimum_cmd_size = 10; 3884 3885 bcopy(cdb, &cmd6, sizeof(struct scsi_rw_6)); 3886 cmd10 = (struct scsi_rw_10 *)cdb; 3887 cmd10->opcode = (cmd6.opcode == READ_6) ? READ_10 : WRITE_10; 3888 cmd10->byte2 = 0; 3889 scsi_ulto4b(scsi_3btoul(cmd6.addr), cmd10->addr); 3890 cmd10->reserved = 0; 3891 scsi_ulto2b(cmd6.length, cmd10->length); 3892 cmd10->control = cmd6.control; 3893 ccb->csio.cdb_len = sizeof(*cmd10); 3894 3895 /* Requeue request, unfreezing queue if necessary */ 3896 frozen = (ccb->ccb_h.status & CAM_DEV_QFRZN) != 0; 3897 ccb->ccb_h.status = CAM_REQUEUE_REQ; 3898 xpt_action(ccb); 3899 if (frozen) { 3900 cam_release_devq(ccb->ccb_h.path, 3901 /*relsim_flags*/0, 3902 /*reduction*/0, 3903 /*timeout*/0, 3904 /*getcount_only*/0); 3905 } 3906 return (ERESTART); 3907 } 3908 3909 static void 3910 dazonedone(struct cam_periph *periph, union ccb *ccb) 3911 { 3912 struct da_softc *softc; 3913 struct bio *bp; 3914 3915 softc = periph->softc; 3916 bp = (struct bio *)ccb->ccb_h.ccb_bp; 3917 3918 switch (bp->bio_zone.zone_cmd) { 3919 case DISK_ZONE_OPEN: 3920 case DISK_ZONE_CLOSE: 3921 case DISK_ZONE_FINISH: 3922 case DISK_ZONE_RWP: 3923 break; 3924 case DISK_ZONE_REPORT_ZONES: { 3925 uint32_t avail_len; 3926 struct disk_zone_report *rep; 3927 struct scsi_report_zones_hdr *hdr; 3928 struct scsi_report_zones_desc *desc; 3929 struct disk_zone_rep_entry *entry; 3930 uint32_t num_alloced, hdr_len, num_avail; 3931 uint32_t num_to_fill, i; 3932 int ata; 3933 3934 rep = &bp->bio_zone.zone_params.report; 3935 avail_len = ccb->csio.dxfer_len - ccb->csio.resid; 3936 /* 3937 * Note that bio_resid isn't normally used for zone 3938 * commands, but it is used by devstat_end_transaction_bio() 3939 * to determine how much data was transferred. Because 3940 * the size of the SCSI/ATA data structures is different 3941 * than the size of the BIO interface structures, the 3942 * amount of data actually transferred from the drive will 3943 * be different than the amount of data transferred to 3944 * the user. 3945 */ 3946 bp->bio_resid = ccb->csio.resid; 3947 num_alloced = rep->entries_allocated; 3948 hdr = (struct scsi_report_zones_hdr *)ccb->csio.data_ptr; 3949 if (avail_len < sizeof(*hdr)) { 3950 /* 3951 * Is there a better error than EIO here? We asked 3952 * for at least the header, and we got less than 3953 * that. 3954 */ 3955 bp->bio_error = EIO; 3956 bp->bio_flags |= BIO_ERROR; 3957 bp->bio_resid = bp->bio_bcount; 3958 break; 3959 } 3960 3961 if (softc->zone_interface == DA_ZONE_IF_ATA_PASS) 3962 ata = 1; 3963 else 3964 ata = 0; 3965 3966 hdr_len = ata ? le32dec(hdr->length) : 3967 scsi_4btoul(hdr->length); 3968 if (hdr_len > 0) 3969 rep->entries_available = hdr_len / sizeof(*desc); 3970 else 3971 rep->entries_available = 0; 3972 /* 3973 * NOTE: using the same values for the BIO version of the 3974 * same field as the SCSI/ATA values. This means we could 3975 * get some additional values that aren't defined in bio.h 3976 * if more values of the same field are defined later. 3977 */ 3978 rep->header.same = hdr->byte4 & SRZ_SAME_MASK; 3979 rep->header.maximum_lba = ata ? le64dec(hdr->maximum_lba) : 3980 scsi_8btou64(hdr->maximum_lba); 3981 /* 3982 * If the drive reports no entries that match the query, 3983 * we're done. 3984 */ 3985 if (hdr_len == 0) { 3986 rep->entries_filled = 0; 3987 break; 3988 } 3989 3990 num_avail = min((avail_len - sizeof(*hdr)) / sizeof(*desc), 3991 hdr_len / sizeof(*desc)); 3992 /* 3993 * If the drive didn't return any data, then we're done. 3994 */ 3995 if (num_avail == 0) { 3996 rep->entries_filled = 0; 3997 break; 3998 } 3999 4000 num_to_fill = min(num_avail, rep->entries_allocated); 4001 /* 4002 * If the user didn't allocate any entries for us to fill, 4003 * we're done. 4004 */ 4005 if (num_to_fill == 0) { 4006 rep->entries_filled = 0; 4007 break; 4008 } 4009 4010 for (i = 0, desc = &hdr->desc_list[0], entry=&rep->entries[0]; 4011 i < num_to_fill; i++, desc++, entry++) { 4012 /* 4013 * NOTE: we're mapping the values here directly 4014 * from the SCSI/ATA bit definitions to the bio.h 4015 * definitons. There is also a warning in 4016 * disk_zone.h, but the impact is that if 4017 * additional values are added in the SCSI/ATA 4018 * specs these will be visible to consumers of 4019 * this interface. 4020 */ 4021 entry->zone_type = desc->zone_type & SRZ_TYPE_MASK; 4022 entry->zone_condition = 4023 (desc->zone_flags & SRZ_ZONE_COND_MASK) >> 4024 SRZ_ZONE_COND_SHIFT; 4025 entry->zone_flags |= desc->zone_flags & 4026 (SRZ_ZONE_NON_SEQ|SRZ_ZONE_RESET); 4027 entry->zone_length = 4028 ata ? le64dec(desc->zone_length) : 4029 scsi_8btou64(desc->zone_length); 4030 entry->zone_start_lba = 4031 ata ? le64dec(desc->zone_start_lba) : 4032 scsi_8btou64(desc->zone_start_lba); 4033 entry->write_pointer_lba = 4034 ata ? le64dec(desc->write_pointer_lba) : 4035 scsi_8btou64(desc->write_pointer_lba); 4036 } 4037 rep->entries_filled = num_to_fill; 4038 break; 4039 } 4040 case DISK_ZONE_GET_PARAMS: 4041 default: 4042 /* 4043 * In theory we should not get a GET_PARAMS bio, since it 4044 * should be handled without queueing the command to the 4045 * drive. 4046 */ 4047 panic("%s: Invalid zone command %d", __func__, 4048 bp->bio_zone.zone_cmd); 4049 break; 4050 } 4051 4052 if (bp->bio_zone.zone_cmd == DISK_ZONE_REPORT_ZONES) 4053 free(ccb->csio.data_ptr, M_SCSIDA); 4054 } 4055 4056 static void 4057 dadone(struct cam_periph *periph, union ccb *done_ccb) 4058 { 4059 struct da_softc *softc; 4060 struct ccb_scsiio *csio; 4061 u_int32_t priority; 4062 da_ccb_state state; 4063 4064 softc = (struct da_softc *)periph->softc; 4065 priority = done_ccb->ccb_h.pinfo.priority; 4066 4067 CAM_DEBUG(periph->path, CAM_DEBUG_TRACE, ("dadone\n")); 4068 4069 csio = &done_ccb->csio; 4070 #if defined(BUF_TRACKING) || defined(FULL_BUF_TRACKING) 4071 if (csio->bio != NULL) 4072 biotrack(csio->bio, __func__); 4073 #endif 4074 state = csio->ccb_h.ccb_state & DA_CCB_TYPE_MASK; 4075 switch (state) { 4076 case DA_CCB_BUFFER_IO: 4077 case DA_CCB_DELETE: 4078 { 4079 struct bio *bp, *bp1; 4080 4081 cam_periph_lock(periph); 4082 bp = (struct bio *)done_ccb->ccb_h.ccb_bp; 4083 if ((done_ccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) { 4084 int error; 4085 int sf; 4086 4087 if ((csio->ccb_h.ccb_state & DA_CCB_RETRY_UA) != 0) 4088 sf = SF_RETRY_UA; 4089 else 4090 sf = 0; 4091 4092 error = daerror(done_ccb, CAM_RETRY_SELTO, sf); 4093 if (error == ERESTART) { 4094 /* 4095 * A retry was scheduled, so 4096 * just return. 4097 */ 4098 cam_periph_unlock(periph); 4099 return; 4100 } 4101 bp = (struct bio *)done_ccb->ccb_h.ccb_bp; 4102 if (error != 0) { 4103 int queued_error; 4104 4105 /* 4106 * return all queued I/O with EIO, so that 4107 * the client can retry these I/Os in the 4108 * proper order should it attempt to recover. 4109 */ 4110 queued_error = EIO; 4111 4112 if (error == ENXIO 4113 && (softc->flags & DA_FLAG_PACK_INVALID)== 0) { 4114 /* 4115 * Catastrophic error. Mark our pack as 4116 * invalid. 4117 */ 4118 /* 4119 * XXX See if this is really a media 4120 * XXX change first? 4121 */ 4122 xpt_print(periph->path, 4123 "Invalidating pack\n"); 4124 softc->flags |= DA_FLAG_PACK_INVALID; 4125 #ifdef CAM_IO_STATS 4126 softc->invalidations++; 4127 #endif 4128 queued_error = ENXIO; 4129 } 4130 cam_iosched_flush(softc->cam_iosched, NULL, 4131 queued_error); 4132 if (bp != NULL) { 4133 bp->bio_error = error; 4134 bp->bio_resid = bp->bio_bcount; 4135 bp->bio_flags |= BIO_ERROR; 4136 } 4137 } else if (bp != NULL) { 4138 if (state == DA_CCB_DELETE) 4139 bp->bio_resid = 0; 4140 else 4141 bp->bio_resid = csio->resid; 4142 bp->bio_error = 0; 4143 if (bp->bio_resid != 0) 4144 bp->bio_flags |= BIO_ERROR; 4145 } 4146 if ((done_ccb->ccb_h.status & CAM_DEV_QFRZN) != 0) 4147 cam_release_devq(done_ccb->ccb_h.path, 4148 /*relsim_flags*/0, 4149 /*reduction*/0, 4150 /*timeout*/0, 4151 /*getcount_only*/0); 4152 } else if (bp != NULL) { 4153 if ((done_ccb->ccb_h.status & CAM_DEV_QFRZN) != 0) 4154 panic("REQ_CMP with QFRZN"); 4155 if (bp->bio_cmd == BIO_ZONE) 4156 dazonedone(periph, done_ccb); 4157 else if (state == DA_CCB_DELETE) 4158 bp->bio_resid = 0; 4159 else 4160 bp->bio_resid = csio->resid; 4161 if ((csio->resid > 0) 4162 && (bp->bio_cmd != BIO_ZONE)) 4163 bp->bio_flags |= BIO_ERROR; 4164 if (softc->error_inject != 0) { 4165 bp->bio_error = softc->error_inject; 4166 bp->bio_resid = bp->bio_bcount; 4167 bp->bio_flags |= BIO_ERROR; 4168 softc->error_inject = 0; 4169 } 4170 } 4171 4172 biotrack(bp, __func__); 4173 LIST_REMOVE(&done_ccb->ccb_h, periph_links.le); 4174 if (LIST_EMPTY(&softc->pending_ccbs)) 4175 softc->flags |= DA_FLAG_WAS_OTAG; 4176 4177 cam_iosched_bio_complete(softc->cam_iosched, bp, done_ccb); 4178 xpt_release_ccb(done_ccb); 4179 if (state == DA_CCB_DELETE) { 4180 TAILQ_HEAD(, bio) queue; 4181 4182 TAILQ_INIT(&queue); 4183 TAILQ_CONCAT(&queue, &softc->delete_run_queue.queue, bio_queue); 4184 softc->delete_run_queue.insert_point = NULL; 4185 /* 4186 * Normally, the xpt_release_ccb() above would make sure 4187 * that when we have more work to do, that work would 4188 * get kicked off. However, we specifically keep 4189 * delete_running set to 0 before the call above to 4190 * allow other I/O to progress when many BIO_DELETE 4191 * requests are pushed down. We set delete_running to 0 4192 * and call daschedule again so that we don't stall if 4193 * there are no other I/Os pending apart from BIO_DELETEs. 4194 */ 4195 cam_iosched_trim_done(softc->cam_iosched); 4196 daschedule(periph); 4197 cam_periph_unlock(periph); 4198 while ((bp1 = TAILQ_FIRST(&queue)) != NULL) { 4199 TAILQ_REMOVE(&queue, bp1, bio_queue); 4200 bp1->bio_error = bp->bio_error; 4201 if (bp->bio_flags & BIO_ERROR) { 4202 bp1->bio_flags |= BIO_ERROR; 4203 bp1->bio_resid = bp1->bio_bcount; 4204 } else 4205 bp1->bio_resid = 0; 4206 biodone(bp1); 4207 } 4208 } else { 4209 daschedule(periph); 4210 cam_periph_unlock(periph); 4211 } 4212 if (bp != NULL) 4213 biodone(bp); 4214 return; 4215 } 4216 case DA_CCB_PROBE_RC: 4217 case DA_CCB_PROBE_RC16: 4218 { 4219 struct scsi_read_capacity_data *rdcap; 4220 struct scsi_read_capacity_data_long *rcaplong; 4221 char announce_buf[80]; 4222 int lbp; 4223 4224 lbp = 0; 4225 rdcap = NULL; 4226 rcaplong = NULL; 4227 if (state == DA_CCB_PROBE_RC) 4228 rdcap =(struct scsi_read_capacity_data *)csio->data_ptr; 4229 else 4230 rcaplong = (struct scsi_read_capacity_data_long *) 4231 csio->data_ptr; 4232 4233 if ((csio->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP) { 4234 struct disk_params *dp; 4235 uint32_t block_size; 4236 uint64_t maxsector; 4237 u_int lalba; /* Lowest aligned LBA. */ 4238 4239 if (state == DA_CCB_PROBE_RC) { 4240 block_size = scsi_4btoul(rdcap->length); 4241 maxsector = scsi_4btoul(rdcap->addr); 4242 lalba = 0; 4243 4244 /* 4245 * According to SBC-2, if the standard 10 4246 * byte READ CAPACITY command returns 2^32, 4247 * we should issue the 16 byte version of 4248 * the command, since the device in question 4249 * has more sectors than can be represented 4250 * with the short version of the command. 4251 */ 4252 if (maxsector == 0xffffffff) { 4253 free(rdcap, M_SCSIDA); 4254 xpt_release_ccb(done_ccb); 4255 softc->state = DA_STATE_PROBE_RC16; 4256 xpt_schedule(periph, priority); 4257 return; 4258 } 4259 } else { 4260 block_size = scsi_4btoul(rcaplong->length); 4261 maxsector = scsi_8btou64(rcaplong->addr); 4262 lalba = scsi_2btoul(rcaplong->lalba_lbp); 4263 } 4264 4265 /* 4266 * Because GEOM code just will panic us if we 4267 * give them an 'illegal' value we'll avoid that 4268 * here. 4269 */ 4270 if (block_size == 0) { 4271 block_size = 512; 4272 if (maxsector == 0) 4273 maxsector = -1; 4274 } 4275 if (block_size >= MAXPHYS) { 4276 xpt_print(periph->path, 4277 "unsupportable block size %ju\n", 4278 (uintmax_t) block_size); 4279 announce_buf[0] = '\0'; 4280 cam_periph_invalidate(periph); 4281 } else { 4282 /* 4283 * We pass rcaplong into dasetgeom(), 4284 * because it will only use it if it is 4285 * non-NULL. 4286 */ 4287 dasetgeom(periph, block_size, maxsector, 4288 rcaplong, sizeof(*rcaplong)); 4289 lbp = (lalba & SRC16_LBPME_A); 4290 dp = &softc->params; 4291 snprintf(announce_buf, sizeof(announce_buf), 4292 "%juMB (%ju %u byte sectors)", 4293 ((uintmax_t)dp->secsize * dp->sectors) / 4294 (1024 * 1024), 4295 (uintmax_t)dp->sectors, dp->secsize); 4296 } 4297 } else { 4298 int error; 4299 4300 announce_buf[0] = '\0'; 4301 4302 /* 4303 * Retry any UNIT ATTENTION type errors. They 4304 * are expected at boot. 4305 */ 4306 error = daerror(done_ccb, CAM_RETRY_SELTO, 4307 SF_RETRY_UA|SF_NO_PRINT); 4308 if (error == ERESTART) { 4309 /* 4310 * A retry was scheuled, so 4311 * just return. 4312 */ 4313 return; 4314 } else if (error != 0) { 4315 int asc, ascq; 4316 int sense_key, error_code; 4317 int have_sense; 4318 cam_status status; 4319 struct ccb_getdev cgd; 4320 4321 /* Don't wedge this device's queue */ 4322 status = done_ccb->ccb_h.status; 4323 if ((status & CAM_DEV_QFRZN) != 0) 4324 cam_release_devq(done_ccb->ccb_h.path, 4325 /*relsim_flags*/0, 4326 /*reduction*/0, 4327 /*timeout*/0, 4328 /*getcount_only*/0); 4329 4330 4331 xpt_setup_ccb(&cgd.ccb_h, 4332 done_ccb->ccb_h.path, 4333 CAM_PRIORITY_NORMAL); 4334 cgd.ccb_h.func_code = XPT_GDEV_TYPE; 4335 xpt_action((union ccb *)&cgd); 4336 4337 if (scsi_extract_sense_ccb(done_ccb, 4338 &error_code, &sense_key, &asc, &ascq)) 4339 have_sense = TRUE; 4340 else 4341 have_sense = FALSE; 4342 4343 /* 4344 * If we tried READ CAPACITY(16) and failed, 4345 * fallback to READ CAPACITY(10). 4346 */ 4347 if ((state == DA_CCB_PROBE_RC16) && 4348 (softc->flags & DA_FLAG_CAN_RC16) && 4349 (((csio->ccb_h.status & CAM_STATUS_MASK) == 4350 CAM_REQ_INVALID) || 4351 ((have_sense) && 4352 (error_code == SSD_CURRENT_ERROR) && 4353 (sense_key == SSD_KEY_ILLEGAL_REQUEST)))) { 4354 softc->flags &= ~DA_FLAG_CAN_RC16; 4355 free(rdcap, M_SCSIDA); 4356 xpt_release_ccb(done_ccb); 4357 softc->state = DA_STATE_PROBE_RC; 4358 xpt_schedule(periph, priority); 4359 return; 4360 } 4361 4362 /* 4363 * Attach to anything that claims to be a 4364 * direct access or optical disk device, 4365 * as long as it doesn't return a "Logical 4366 * unit not supported" (0x25) error. 4367 * "Internal Target Failure" (0x44) is also 4368 * special and typically means that the 4369 * device is a SATA drive behind a SATL 4370 * translation that's fallen into a 4371 * terminally fatal state. 4372 */ 4373 if ((have_sense) 4374 && (asc != 0x25) && (asc != 0x44) 4375 && (error_code == SSD_CURRENT_ERROR)) { 4376 const char *sense_key_desc; 4377 const char *asc_desc; 4378 4379 dasetgeom(periph, 512, -1, NULL, 0); 4380 scsi_sense_desc(sense_key, asc, ascq, 4381 &cgd.inq_data, 4382 &sense_key_desc, 4383 &asc_desc); 4384 snprintf(announce_buf, 4385 sizeof(announce_buf), 4386 "Attempt to query device " 4387 "size failed: %s, %s", 4388 sense_key_desc, 4389 asc_desc); 4390 } else { 4391 if (have_sense) 4392 scsi_sense_print( 4393 &done_ccb->csio); 4394 else { 4395 xpt_print(periph->path, 4396 "got CAM status %#x\n", 4397 done_ccb->ccb_h.status); 4398 } 4399 4400 xpt_print(periph->path, "fatal error, " 4401 "failed to attach to device\n"); 4402 4403 /* 4404 * Free up resources. 4405 */ 4406 cam_periph_invalidate(periph); 4407 } 4408 } 4409 } 4410 free(csio->data_ptr, M_SCSIDA); 4411 if (announce_buf[0] != '\0' && 4412 ((softc->flags & DA_FLAG_ANNOUNCED) == 0)) { 4413 /* 4414 * Create our sysctl variables, now that we know 4415 * we have successfully attached. 4416 */ 4417 /* increase the refcount */ 4418 if (cam_periph_acquire(periph) == CAM_REQ_CMP) { 4419 taskqueue_enqueue(taskqueue_thread, 4420 &softc->sysctl_task); 4421 xpt_announce_periph(periph, announce_buf); 4422 xpt_announce_quirks(periph, softc->quirks, 4423 DA_Q_BIT_STRING); 4424 } else { 4425 xpt_print(periph->path, "fatal error, " 4426 "could not acquire reference count\n"); 4427 } 4428 } 4429 4430 /* We already probed the device. */ 4431 if (softc->flags & DA_FLAG_PROBED) { 4432 daprobedone(periph, done_ccb); 4433 return; 4434 } 4435 4436 /* Ensure re-probe doesn't see old delete. */ 4437 softc->delete_available = 0; 4438 dadeleteflag(softc, DA_DELETE_ZERO, 1); 4439 if (lbp && (softc->quirks & DA_Q_NO_UNMAP) == 0) { 4440 /* 4441 * Based on older SBC-3 spec revisions 4442 * any of the UNMAP methods "may" be 4443 * available via LBP given this flag so 4444 * we flag all of them as available and 4445 * then remove those which further 4446 * probes confirm aren't available 4447 * later. 4448 * 4449 * We could also check readcap(16) p_type 4450 * flag to exclude one or more invalid 4451 * write same (X) types here 4452 */ 4453 dadeleteflag(softc, DA_DELETE_WS16, 1); 4454 dadeleteflag(softc, DA_DELETE_WS10, 1); 4455 dadeleteflag(softc, DA_DELETE_UNMAP, 1); 4456 4457 xpt_release_ccb(done_ccb); 4458 softc->state = DA_STATE_PROBE_LBP; 4459 xpt_schedule(periph, priority); 4460 return; 4461 } 4462 4463 xpt_release_ccb(done_ccb); 4464 softc->state = DA_STATE_PROBE_BDC; 4465 xpt_schedule(periph, priority); 4466 return; 4467 } 4468 case DA_CCB_PROBE_LBP: 4469 { 4470 struct scsi_vpd_logical_block_prov *lbp; 4471 4472 lbp = (struct scsi_vpd_logical_block_prov *)csio->data_ptr; 4473 4474 if ((csio->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP) { 4475 /* 4476 * T10/1799-D Revision 31 states at least one of these 4477 * must be supported but we don't currently enforce this. 4478 */ 4479 dadeleteflag(softc, DA_DELETE_WS16, 4480 (lbp->flags & SVPD_LBP_WS16)); 4481 dadeleteflag(softc, DA_DELETE_WS10, 4482 (lbp->flags & SVPD_LBP_WS10)); 4483 dadeleteflag(softc, DA_DELETE_UNMAP, 4484 (lbp->flags & SVPD_LBP_UNMAP)); 4485 } else { 4486 int error; 4487 error = daerror(done_ccb, CAM_RETRY_SELTO, 4488 SF_RETRY_UA|SF_NO_PRINT); 4489 if (error == ERESTART) 4490 return; 4491 else if (error != 0) { 4492 if ((done_ccb->ccb_h.status & CAM_DEV_QFRZN) != 0) { 4493 /* Don't wedge this device's queue */ 4494 cam_release_devq(done_ccb->ccb_h.path, 4495 /*relsim_flags*/0, 4496 /*reduction*/0, 4497 /*timeout*/0, 4498 /*getcount_only*/0); 4499 } 4500 4501 /* 4502 * Failure indicates we don't support any SBC-3 4503 * delete methods with UNMAP 4504 */ 4505 } 4506 } 4507 4508 free(lbp, M_SCSIDA); 4509 xpt_release_ccb(done_ccb); 4510 softc->state = DA_STATE_PROBE_BLK_LIMITS; 4511 xpt_schedule(periph, priority); 4512 return; 4513 } 4514 case DA_CCB_PROBE_BLK_LIMITS: 4515 { 4516 struct scsi_vpd_block_limits *block_limits; 4517 4518 block_limits = (struct scsi_vpd_block_limits *)csio->data_ptr; 4519 4520 if ((csio->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP) { 4521 uint32_t max_txfer_len = scsi_4btoul( 4522 block_limits->max_txfer_len); 4523 uint32_t max_unmap_lba_cnt = scsi_4btoul( 4524 block_limits->max_unmap_lba_cnt); 4525 uint32_t max_unmap_blk_cnt = scsi_4btoul( 4526 block_limits->max_unmap_blk_cnt); 4527 uint32_t unmap_gran = scsi_4btoul( 4528 block_limits->opt_unmap_grain); 4529 uint32_t unmap_gran_align = scsi_4btoul( 4530 block_limits->unmap_grain_align); 4531 uint64_t ws_max_blks = scsi_8btou64( 4532 block_limits->max_write_same_length); 4533 4534 if (max_txfer_len != 0) { 4535 softc->disk->d_maxsize = MIN(softc->maxio, 4536 (off_t)max_txfer_len * softc->params.secsize); 4537 } 4538 4539 /* 4540 * We should already support UNMAP but we check lba 4541 * and block count to be sure 4542 */ 4543 if (max_unmap_lba_cnt != 0x00L && 4544 max_unmap_blk_cnt != 0x00L) { 4545 softc->unmap_max_lba = max_unmap_lba_cnt; 4546 softc->unmap_max_ranges = min(max_unmap_blk_cnt, 4547 UNMAP_MAX_RANGES); 4548 if (unmap_gran > 1) { 4549 softc->unmap_gran = unmap_gran; 4550 if (unmap_gran_align & 0x80000000) { 4551 softc->unmap_gran_align = 4552 unmap_gran_align & 4553 0x7fffffff; 4554 } 4555 } 4556 } else { 4557 /* 4558 * Unexpected UNMAP limits which means the 4559 * device doesn't actually support UNMAP 4560 */ 4561 dadeleteflag(softc, DA_DELETE_UNMAP, 0); 4562 } 4563 4564 if (ws_max_blks != 0x00L) 4565 softc->ws_max_blks = ws_max_blks; 4566 } else { 4567 int error; 4568 error = daerror(done_ccb, CAM_RETRY_SELTO, 4569 SF_RETRY_UA|SF_NO_PRINT); 4570 if (error == ERESTART) 4571 return; 4572 else if (error != 0) { 4573 if ((done_ccb->ccb_h.status & CAM_DEV_QFRZN) != 0) { 4574 /* Don't wedge this device's queue */ 4575 cam_release_devq(done_ccb->ccb_h.path, 4576 /*relsim_flags*/0, 4577 /*reduction*/0, 4578 /*timeout*/0, 4579 /*getcount_only*/0); 4580 } 4581 4582 /* 4583 * Failure here doesn't mean UNMAP is not 4584 * supported as this is an optional page. 4585 */ 4586 softc->unmap_max_lba = 1; 4587 softc->unmap_max_ranges = 1; 4588 } 4589 } 4590 4591 free(block_limits, M_SCSIDA); 4592 xpt_release_ccb(done_ccb); 4593 softc->state = DA_STATE_PROBE_BDC; 4594 xpt_schedule(periph, priority); 4595 return; 4596 } 4597 case DA_CCB_PROBE_BDC: 4598 { 4599 struct scsi_vpd_block_device_characteristics *bdc; 4600 4601 bdc = (struct scsi_vpd_block_device_characteristics *) 4602 csio->data_ptr; 4603 4604 if ((csio->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP) { 4605 uint32_t valid_len; 4606 4607 /* 4608 * Disable queue sorting for non-rotational media 4609 * by default. 4610 */ 4611 u_int16_t old_rate = softc->disk->d_rotation_rate; 4612 4613 valid_len = csio->dxfer_len - csio->resid; 4614 if (SBDC_IS_PRESENT(bdc, valid_len, 4615 medium_rotation_rate)) { 4616 softc->disk->d_rotation_rate = 4617 scsi_2btoul(bdc->medium_rotation_rate); 4618 if (softc->disk->d_rotation_rate == 4619 SVPD_BDC_RATE_NON_ROTATING) { 4620 cam_iosched_set_sort_queue( 4621 softc->cam_iosched, 0); 4622 softc->rotating = 0; 4623 } 4624 if (softc->disk->d_rotation_rate != old_rate) { 4625 disk_attr_changed(softc->disk, 4626 "GEOM::rotation_rate", M_NOWAIT); 4627 } 4628 } 4629 if ((SBDC_IS_PRESENT(bdc, valid_len, flags)) 4630 && (softc->zone_mode == DA_ZONE_NONE)) { 4631 int ata_proto; 4632 4633 if (scsi_vpd_supported_page(periph, 4634 SVPD_ATA_INFORMATION)) 4635 ata_proto = 1; 4636 else 4637 ata_proto = 0; 4638 4639 /* 4640 * The Zoned field will only be set for 4641 * Drive Managed and Host Aware drives. If 4642 * they are Host Managed, the device type 4643 * in the standard INQUIRY data should be 4644 * set to T_ZBC_HM (0x14). 4645 */ 4646 if ((bdc->flags & SVPD_ZBC_MASK) == 4647 SVPD_HAW_ZBC) { 4648 softc->zone_mode = DA_ZONE_HOST_AWARE; 4649 softc->zone_interface = (ata_proto) ? 4650 DA_ZONE_IF_ATA_SAT : DA_ZONE_IF_SCSI; 4651 } else if ((bdc->flags & SVPD_ZBC_MASK) == 4652 SVPD_DM_ZBC) { 4653 softc->zone_mode =DA_ZONE_DRIVE_MANAGED; 4654 softc->zone_interface = (ata_proto) ? 4655 DA_ZONE_IF_ATA_SAT : DA_ZONE_IF_SCSI; 4656 } else if ((bdc->flags & SVPD_ZBC_MASK) != 4657 SVPD_ZBC_NR) { 4658 xpt_print(periph->path, "Unknown zoned " 4659 "type %#x", 4660 bdc->flags & SVPD_ZBC_MASK); 4661 } 4662 } 4663 } else { 4664 int error; 4665 error = daerror(done_ccb, CAM_RETRY_SELTO, 4666 SF_RETRY_UA|SF_NO_PRINT); 4667 if (error == ERESTART) 4668 return; 4669 else if (error != 0) { 4670 if ((done_ccb->ccb_h.status & CAM_DEV_QFRZN) != 0) { 4671 /* Don't wedge this device's queue */ 4672 cam_release_devq(done_ccb->ccb_h.path, 4673 /*relsim_flags*/0, 4674 /*reduction*/0, 4675 /*timeout*/0, 4676 /*getcount_only*/0); 4677 } 4678 } 4679 } 4680 4681 free(bdc, M_SCSIDA); 4682 xpt_release_ccb(done_ccb); 4683 softc->state = DA_STATE_PROBE_ATA; 4684 xpt_schedule(periph, priority); 4685 return; 4686 } 4687 case DA_CCB_PROBE_ATA: 4688 { 4689 int i; 4690 struct ata_params *ata_params; 4691 int continue_probe; 4692 int error; 4693 int16_t *ptr; 4694 4695 ata_params = (struct ata_params *)csio->data_ptr; 4696 ptr = (uint16_t *)ata_params; 4697 continue_probe = 0; 4698 error = 0; 4699 4700 if ((csio->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP) { 4701 uint16_t old_rate; 4702 4703 for (i = 0; i < sizeof(*ata_params) / 2; i++) 4704 ptr[i] = le16toh(ptr[i]); 4705 if (ata_params->support_dsm & ATA_SUPPORT_DSM_TRIM && 4706 (softc->quirks & DA_Q_NO_UNMAP) == 0) { 4707 dadeleteflag(softc, DA_DELETE_ATA_TRIM, 1); 4708 if (ata_params->max_dsm_blocks != 0) 4709 softc->trim_max_ranges = min( 4710 softc->trim_max_ranges, 4711 ata_params->max_dsm_blocks * 4712 ATA_DSM_BLK_RANGES); 4713 } 4714 /* 4715 * Disable queue sorting for non-rotational media 4716 * by default. 4717 */ 4718 old_rate = softc->disk->d_rotation_rate; 4719 softc->disk->d_rotation_rate = 4720 ata_params->media_rotation_rate; 4721 if (softc->disk->d_rotation_rate == 4722 ATA_RATE_NON_ROTATING) { 4723 cam_iosched_set_sort_queue(softc->cam_iosched, 0); 4724 softc->rotating = 0; 4725 } 4726 if (softc->disk->d_rotation_rate != old_rate) { 4727 disk_attr_changed(softc->disk, 4728 "GEOM::rotation_rate", M_NOWAIT); 4729 } 4730 4731 if (ata_params->capabilities1 & ATA_SUPPORT_DMA) 4732 softc->flags |= DA_FLAG_CAN_ATA_DMA; 4733 4734 if (ata_params->support.extension & 4735 ATA_SUPPORT_GENLOG) 4736 softc->flags |= DA_FLAG_CAN_ATA_LOG; 4737 4738 /* 4739 * At this point, if we have a SATA host aware drive, 4740 * we communicate via ATA passthrough unless the 4741 * SAT layer supports ZBC -> ZAC translation. In 4742 * that case, 4743 */ 4744 /* 4745 * XXX KDM figure out how to detect a host managed 4746 * SATA drive. 4747 */ 4748 if (softc->zone_mode == DA_ZONE_NONE) { 4749 /* 4750 * Note that we don't override the zone 4751 * mode or interface if it has already been 4752 * set. This is because it has either been 4753 * set as a quirk, or when we probed the 4754 * SCSI Block Device Characteristics page, 4755 * the zoned field was set. The latter 4756 * means that the SAT layer supports ZBC to 4757 * ZAC translation, and we would prefer to 4758 * use that if it is available. 4759 */ 4760 if ((ata_params->support3 & 4761 ATA_SUPPORT_ZONE_MASK) == 4762 ATA_SUPPORT_ZONE_HOST_AWARE) { 4763 softc->zone_mode = DA_ZONE_HOST_AWARE; 4764 softc->zone_interface = 4765 DA_ZONE_IF_ATA_PASS; 4766 } else if ((ata_params->support3 & 4767 ATA_SUPPORT_ZONE_MASK) == 4768 ATA_SUPPORT_ZONE_DEV_MANAGED) { 4769 softc->zone_mode =DA_ZONE_DRIVE_MANAGED; 4770 softc->zone_interface = 4771 DA_ZONE_IF_ATA_PASS; 4772 } 4773 } 4774 4775 } else { 4776 error = daerror(done_ccb, CAM_RETRY_SELTO, 4777 SF_RETRY_UA|SF_NO_PRINT); 4778 if (error == ERESTART) 4779 return; 4780 else if (error != 0) { 4781 if ((done_ccb->ccb_h.status & 4782 CAM_DEV_QFRZN) != 0) { 4783 /* Don't wedge this device's queue */ 4784 cam_release_devq(done_ccb->ccb_h.path, 4785 /*relsim_flags*/0, 4786 /*reduction*/0, 4787 /*timeout*/0, 4788 /*getcount_only*/0); 4789 } 4790 } 4791 } 4792 4793 free(ata_params, M_SCSIDA); 4794 if ((softc->zone_mode == DA_ZONE_HOST_AWARE) 4795 || (softc->zone_mode == DA_ZONE_HOST_MANAGED)) { 4796 /* 4797 * If the ATA IDENTIFY failed, we could be talking 4798 * to a SCSI drive, although that seems unlikely, 4799 * since the drive did report that it supported the 4800 * ATA Information VPD page. If the ATA IDENTIFY 4801 * succeeded, and the SAT layer doesn't support 4802 * ZBC -> ZAC translation, continue on to get the 4803 * directory of ATA logs, and complete the rest of 4804 * the ZAC probe. If the SAT layer does support 4805 * ZBC -> ZAC translation, we want to use that, 4806 * and we'll probe the SCSI Zoned Block Device 4807 * Characteristics VPD page next. 4808 */ 4809 if ((error == 0) 4810 && (softc->flags & DA_FLAG_CAN_ATA_LOG) 4811 && (softc->zone_interface == DA_ZONE_IF_ATA_PASS)) 4812 softc->state = DA_STATE_PROBE_ATA_LOGDIR; 4813 else 4814 softc->state = DA_STATE_PROBE_ZONE; 4815 continue_probe = 1; 4816 } 4817 if (continue_probe != 0) { 4818 xpt_release_ccb(done_ccb); 4819 xpt_schedule(periph, priority); 4820 return; 4821 } else 4822 daprobedone(periph, done_ccb); 4823 return; 4824 } 4825 case DA_CCB_PROBE_ATA_LOGDIR: 4826 { 4827 int error; 4828 4829 if ((csio->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP) { 4830 error = 0; 4831 softc->valid_logdir_len = 0; 4832 bzero(&softc->ata_logdir, sizeof(softc->ata_logdir)); 4833 softc->valid_logdir_len = 4834 csio->dxfer_len - csio->resid; 4835 if (softc->valid_logdir_len > 0) 4836 bcopy(csio->data_ptr, &softc->ata_logdir, 4837 min(softc->valid_logdir_len, 4838 sizeof(softc->ata_logdir))); 4839 /* 4840 * Figure out whether the Identify Device log is 4841 * supported. The General Purpose log directory 4842 * has a header, and lists the number of pages 4843 * available for each GP log identified by the 4844 * offset into the list. 4845 */ 4846 if ((softc->valid_logdir_len >= 4847 ((ATA_IDENTIFY_DATA_LOG + 1) * sizeof(uint16_t))) 4848 && (le16dec(softc->ata_logdir.header) == 4849 ATA_GP_LOG_DIR_VERSION) 4850 && (le16dec(&softc->ata_logdir.num_pages[ 4851 (ATA_IDENTIFY_DATA_LOG * 4852 sizeof(uint16_t)) - sizeof(uint16_t)]) > 0)){ 4853 softc->flags |= DA_FLAG_CAN_ATA_IDLOG; 4854 } else { 4855 softc->flags &= ~DA_FLAG_CAN_ATA_IDLOG; 4856 } 4857 } else { 4858 error = daerror(done_ccb, CAM_RETRY_SELTO, 4859 SF_RETRY_UA|SF_NO_PRINT); 4860 if (error == ERESTART) 4861 return; 4862 else if (error != 0) { 4863 /* 4864 * If we can't get the ATA log directory, 4865 * then ATA logs are effectively not 4866 * supported even if the bit is set in the 4867 * identify data. 4868 */ 4869 softc->flags &= ~(DA_FLAG_CAN_ATA_LOG | 4870 DA_FLAG_CAN_ATA_IDLOG); 4871 if ((done_ccb->ccb_h.status & 4872 CAM_DEV_QFRZN) != 0) { 4873 /* Don't wedge this device's queue */ 4874 cam_release_devq(done_ccb->ccb_h.path, 4875 /*relsim_flags*/0, 4876 /*reduction*/0, 4877 /*timeout*/0, 4878 /*getcount_only*/0); 4879 } 4880 } 4881 } 4882 4883 free(csio->data_ptr, M_SCSIDA); 4884 4885 if ((error == 0) 4886 && (softc->flags & DA_FLAG_CAN_ATA_IDLOG)) { 4887 softc->state = DA_STATE_PROBE_ATA_IDDIR; 4888 xpt_release_ccb(done_ccb); 4889 xpt_schedule(periph, priority); 4890 return; 4891 } 4892 daprobedone(periph, done_ccb); 4893 return; 4894 } 4895 case DA_CCB_PROBE_ATA_IDDIR: 4896 { 4897 int error; 4898 4899 if ((csio->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP) { 4900 off_t entries_offset, max_entries; 4901 error = 0; 4902 4903 softc->valid_iddir_len = 0; 4904 bzero(&softc->ata_iddir, sizeof(softc->ata_iddir)); 4905 softc->flags &= ~(DA_FLAG_CAN_ATA_SUPCAP | 4906 DA_FLAG_CAN_ATA_ZONE); 4907 softc->valid_iddir_len = 4908 csio->dxfer_len - csio->resid; 4909 if (softc->valid_iddir_len > 0) 4910 bcopy(csio->data_ptr, &softc->ata_iddir, 4911 min(softc->valid_iddir_len, 4912 sizeof(softc->ata_iddir))); 4913 4914 entries_offset = 4915 __offsetof(struct ata_identify_log_pages,entries); 4916 max_entries = softc->valid_iddir_len - entries_offset; 4917 if ((softc->valid_iddir_len > (entries_offset + 1)) 4918 && (le64dec(softc->ata_iddir.header) == 4919 ATA_IDLOG_REVISION) 4920 && (softc->ata_iddir.entry_count > 0)) { 4921 int num_entries, i; 4922 4923 num_entries = softc->ata_iddir.entry_count; 4924 num_entries = min(num_entries, 4925 softc->valid_iddir_len - entries_offset); 4926 for (i = 0; i < num_entries && 4927 i < max_entries; i++) { 4928 if (softc->ata_iddir.entries[i] == 4929 ATA_IDL_SUP_CAP) 4930 softc->flags |= 4931 DA_FLAG_CAN_ATA_SUPCAP; 4932 else if (softc->ata_iddir.entries[i]== 4933 ATA_IDL_ZDI) 4934 softc->flags |= 4935 DA_FLAG_CAN_ATA_ZONE; 4936 4937 if ((softc->flags & 4938 DA_FLAG_CAN_ATA_SUPCAP) 4939 && (softc->flags & 4940 DA_FLAG_CAN_ATA_ZONE)) 4941 break; 4942 } 4943 } 4944 } else { 4945 error = daerror(done_ccb, CAM_RETRY_SELTO, 4946 SF_RETRY_UA|SF_NO_PRINT); 4947 if (error == ERESTART) 4948 return; 4949 else if (error != 0) { 4950 /* 4951 * If we can't get the ATA Identify Data log 4952 * directory, then it effectively isn't 4953 * supported even if the ATA Log directory 4954 * a non-zero number of pages present for 4955 * this log. 4956 */ 4957 softc->flags &= ~DA_FLAG_CAN_ATA_IDLOG; 4958 if ((done_ccb->ccb_h.status & 4959 CAM_DEV_QFRZN) != 0) { 4960 /* Don't wedge this device's queue */ 4961 cam_release_devq(done_ccb->ccb_h.path, 4962 /*relsim_flags*/0, 4963 /*reduction*/0, 4964 /*timeout*/0, 4965 /*getcount_only*/0); 4966 } 4967 } 4968 } 4969 4970 free(csio->data_ptr, M_SCSIDA); 4971 4972 if ((error == 0) 4973 && (softc->flags & DA_FLAG_CAN_ATA_SUPCAP)) { 4974 softc->state = DA_STATE_PROBE_ATA_SUP; 4975 xpt_release_ccb(done_ccb); 4976 xpt_schedule(periph, priority); 4977 return; 4978 } 4979 daprobedone(periph, done_ccb); 4980 return; 4981 } 4982 case DA_CCB_PROBE_ATA_SUP: 4983 { 4984 int error; 4985 4986 if ((csio->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP) { 4987 uint32_t valid_len; 4988 size_t needed_size; 4989 struct ata_identify_log_sup_cap *sup_cap; 4990 error = 0; 4991 4992 sup_cap = (struct ata_identify_log_sup_cap *) 4993 csio->data_ptr; 4994 valid_len = csio->dxfer_len - csio->resid; 4995 needed_size = 4996 __offsetof(struct ata_identify_log_sup_cap, 4997 sup_zac_cap) + 1 + sizeof(sup_cap->sup_zac_cap); 4998 if (valid_len >= needed_size) { 4999 uint64_t zoned, zac_cap; 5000 5001 zoned = le64dec(sup_cap->zoned_cap); 5002 if (zoned & ATA_ZONED_VALID) { 5003 /* 5004 * This should have already been 5005 * set, because this is also in the 5006 * ATA identify data. 5007 */ 5008 if ((zoned & ATA_ZONED_MASK) == 5009 ATA_SUPPORT_ZONE_HOST_AWARE) 5010 softc->zone_mode = 5011 DA_ZONE_HOST_AWARE; 5012 else if ((zoned & ATA_ZONED_MASK) == 5013 ATA_SUPPORT_ZONE_DEV_MANAGED) 5014 softc->zone_mode = 5015 DA_ZONE_DRIVE_MANAGED; 5016 } 5017 5018 zac_cap = le64dec(sup_cap->sup_zac_cap); 5019 if (zac_cap & ATA_SUP_ZAC_CAP_VALID) { 5020 if (zac_cap & ATA_REPORT_ZONES_SUP) 5021 softc->zone_flags |= 5022 DA_ZONE_FLAG_RZ_SUP; 5023 if (zac_cap & ATA_ND_OPEN_ZONE_SUP) 5024 softc->zone_flags |= 5025 DA_ZONE_FLAG_OPEN_SUP; 5026 if (zac_cap & ATA_ND_CLOSE_ZONE_SUP) 5027 softc->zone_flags |= 5028 DA_ZONE_FLAG_CLOSE_SUP; 5029 if (zac_cap & ATA_ND_FINISH_ZONE_SUP) 5030 softc->zone_flags |= 5031 DA_ZONE_FLAG_FINISH_SUP; 5032 if (zac_cap & ATA_ND_RWP_SUP) 5033 softc->zone_flags |= 5034 DA_ZONE_FLAG_RWP_SUP; 5035 } else { 5036 /* 5037 * This field was introduced in 5038 * ACS-4, r08 on April 28th, 2015. 5039 * If the drive firmware was written 5040 * to an earlier spec, it won't have 5041 * the field. So, assume all 5042 * commands are supported. 5043 */ 5044 softc->zone_flags |= 5045 DA_ZONE_FLAG_SUP_MASK; 5046 } 5047 5048 } 5049 } else { 5050 error = daerror(done_ccb, CAM_RETRY_SELTO, 5051 SF_RETRY_UA|SF_NO_PRINT); 5052 if (error == ERESTART) 5053 return; 5054 else if (error != 0) { 5055 /* 5056 * If we can't get the ATA Identify Data 5057 * Supported Capabilities page, clear the 5058 * flag... 5059 */ 5060 softc->flags &= ~DA_FLAG_CAN_ATA_SUPCAP; 5061 /* 5062 * And clear zone capabilities. 5063 */ 5064 softc->zone_flags &= ~DA_ZONE_FLAG_SUP_MASK; 5065 if ((done_ccb->ccb_h.status & 5066 CAM_DEV_QFRZN) != 0) { 5067 /* Don't wedge this device's queue */ 5068 cam_release_devq(done_ccb->ccb_h.path, 5069 /*relsim_flags*/0, 5070 /*reduction*/0, 5071 /*timeout*/0, 5072 /*getcount_only*/0); 5073 } 5074 } 5075 } 5076 5077 free(csio->data_ptr, M_SCSIDA); 5078 5079 if ((error == 0) 5080 && (softc->flags & DA_FLAG_CAN_ATA_ZONE)) { 5081 softc->state = DA_STATE_PROBE_ATA_ZONE; 5082 xpt_release_ccb(done_ccb); 5083 xpt_schedule(periph, priority); 5084 return; 5085 } 5086 daprobedone(periph, done_ccb); 5087 return; 5088 } 5089 case DA_CCB_PROBE_ATA_ZONE: 5090 { 5091 int error; 5092 5093 if ((csio->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP) { 5094 struct ata_zoned_info_log *zi_log; 5095 uint32_t valid_len; 5096 size_t needed_size; 5097 5098 zi_log = (struct ata_zoned_info_log *)csio->data_ptr; 5099 5100 valid_len = csio->dxfer_len - csio->resid; 5101 needed_size = __offsetof(struct ata_zoned_info_log, 5102 version_info) + 1 + sizeof(zi_log->version_info); 5103 if (valid_len >= needed_size) { 5104 uint64_t tmpvar; 5105 5106 tmpvar = le64dec(zi_log->zoned_cap); 5107 if (tmpvar & ATA_ZDI_CAP_VALID) { 5108 if (tmpvar & ATA_ZDI_CAP_URSWRZ) 5109 softc->zone_flags |= 5110 DA_ZONE_FLAG_URSWRZ; 5111 else 5112 softc->zone_flags &= 5113 ~DA_ZONE_FLAG_URSWRZ; 5114 } 5115 tmpvar = le64dec(zi_log->optimal_seq_zones); 5116 if (tmpvar & ATA_ZDI_OPT_SEQ_VALID) { 5117 softc->zone_flags |= 5118 DA_ZONE_FLAG_OPT_SEQ_SET; 5119 softc->optimal_seq_zones = (tmpvar & 5120 ATA_ZDI_OPT_SEQ_MASK); 5121 } else { 5122 softc->zone_flags &= 5123 ~DA_ZONE_FLAG_OPT_SEQ_SET; 5124 softc->optimal_seq_zones = 0; 5125 } 5126 5127 tmpvar =le64dec(zi_log->optimal_nonseq_zones); 5128 if (tmpvar & ATA_ZDI_OPT_NS_VALID) { 5129 softc->zone_flags |= 5130 DA_ZONE_FLAG_OPT_NONSEQ_SET; 5131 softc->optimal_nonseq_zones = 5132 (tmpvar & ATA_ZDI_OPT_NS_MASK); 5133 } else { 5134 softc->zone_flags &= 5135 ~DA_ZONE_FLAG_OPT_NONSEQ_SET; 5136 softc->optimal_nonseq_zones = 0; 5137 } 5138 5139 tmpvar = le64dec(zi_log->max_seq_req_zones); 5140 if (tmpvar & ATA_ZDI_MAX_SEQ_VALID) { 5141 softc->zone_flags |= 5142 DA_ZONE_FLAG_MAX_SEQ_SET; 5143 softc->max_seq_zones = 5144 (tmpvar & ATA_ZDI_MAX_SEQ_MASK); 5145 } else { 5146 softc->zone_flags &= 5147 ~DA_ZONE_FLAG_MAX_SEQ_SET; 5148 softc->max_seq_zones = 0; 5149 } 5150 } 5151 } else { 5152 error = daerror(done_ccb, CAM_RETRY_SELTO, 5153 SF_RETRY_UA|SF_NO_PRINT); 5154 if (error == ERESTART) 5155 return; 5156 else if (error != 0) { 5157 softc->flags &= ~DA_FLAG_CAN_ATA_ZONE; 5158 softc->flags &= ~DA_ZONE_FLAG_SET_MASK; 5159 5160 if ((done_ccb->ccb_h.status & 5161 CAM_DEV_QFRZN) != 0) { 5162 /* Don't wedge this device's queue */ 5163 cam_release_devq(done_ccb->ccb_h.path, 5164 /*relsim_flags*/0, 5165 /*reduction*/0, 5166 /*timeout*/0, 5167 /*getcount_only*/0); 5168 } 5169 } 5170 5171 } 5172 free(csio->data_ptr, M_SCSIDA); 5173 5174 daprobedone(periph, done_ccb); 5175 return; 5176 } 5177 case DA_CCB_PROBE_ZONE: 5178 { 5179 int error; 5180 5181 if ((csio->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP) { 5182 uint32_t valid_len; 5183 size_t needed_len; 5184 struct scsi_vpd_zoned_bdc *zoned_bdc; 5185 5186 error = 0; 5187 zoned_bdc = (struct scsi_vpd_zoned_bdc *) 5188 csio->data_ptr; 5189 valid_len = csio->dxfer_len - csio->resid; 5190 needed_len = __offsetof(struct scsi_vpd_zoned_bdc, 5191 max_seq_req_zones) + 1 + 5192 sizeof(zoned_bdc->max_seq_req_zones); 5193 if ((valid_len >= needed_len) 5194 && (scsi_2btoul(zoned_bdc->page_length) >= 5195 SVPD_ZBDC_PL)) { 5196 if (zoned_bdc->flags & SVPD_ZBDC_URSWRZ) 5197 softc->zone_flags |= 5198 DA_ZONE_FLAG_URSWRZ; 5199 else 5200 softc->zone_flags &= 5201 ~DA_ZONE_FLAG_URSWRZ; 5202 softc->optimal_seq_zones = 5203 scsi_4btoul(zoned_bdc->optimal_seq_zones); 5204 softc->zone_flags |= DA_ZONE_FLAG_OPT_SEQ_SET; 5205 softc->optimal_nonseq_zones = scsi_4btoul( 5206 zoned_bdc->optimal_nonseq_zones); 5207 softc->zone_flags |= 5208 DA_ZONE_FLAG_OPT_NONSEQ_SET; 5209 softc->max_seq_zones = 5210 scsi_4btoul(zoned_bdc->max_seq_req_zones); 5211 softc->zone_flags |= DA_ZONE_FLAG_MAX_SEQ_SET; 5212 } 5213 /* 5214 * All of the zone commands are mandatory for SCSI 5215 * devices. 5216 * 5217 * XXX KDM this is valid as of September 2015. 5218 * Re-check this assumption once the SAT spec is 5219 * updated to support SCSI ZBC to ATA ZAC mapping. 5220 * Since ATA allows zone commands to be reported 5221 * as supported or not, this may not necessarily 5222 * be true for an ATA device behind a SAT (SCSI to 5223 * ATA Translation) layer. 5224 */ 5225 softc->zone_flags |= DA_ZONE_FLAG_SUP_MASK; 5226 } else { 5227 error = daerror(done_ccb, CAM_RETRY_SELTO, 5228 SF_RETRY_UA|SF_NO_PRINT); 5229 if (error == ERESTART) 5230 return; 5231 else if (error != 0) { 5232 if ((done_ccb->ccb_h.status & CAM_DEV_QFRZN) != 0) { 5233 /* Don't wedge this device's queue */ 5234 cam_release_devq(done_ccb->ccb_h.path, 5235 /*relsim_flags*/0, 5236 /*reduction*/0, 5237 /*timeout*/0, 5238 /*getcount_only*/0); 5239 } 5240 } 5241 } 5242 daprobedone(periph, done_ccb); 5243 return; 5244 } 5245 case DA_CCB_DUMP: 5246 /* No-op. We're polling */ 5247 return; 5248 case DA_CCB_TUR: 5249 { 5250 if ((done_ccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) { 5251 5252 if (daerror(done_ccb, CAM_RETRY_SELTO, 5253 SF_RETRY_UA | SF_NO_RECOVERY | SF_NO_PRINT) == 5254 ERESTART) 5255 return; 5256 if ((done_ccb->ccb_h.status & CAM_DEV_QFRZN) != 0) 5257 cam_release_devq(done_ccb->ccb_h.path, 5258 /*relsim_flags*/0, 5259 /*reduction*/0, 5260 /*timeout*/0, 5261 /*getcount_only*/0); 5262 } 5263 xpt_release_ccb(done_ccb); 5264 cam_periph_release_locked(periph); 5265 return; 5266 } 5267 default: 5268 break; 5269 } 5270 xpt_release_ccb(done_ccb); 5271 } 5272 5273 static void 5274 dareprobe(struct cam_periph *periph) 5275 { 5276 struct da_softc *softc; 5277 cam_status status; 5278 5279 softc = (struct da_softc *)periph->softc; 5280 5281 /* Probe in progress; don't interfere. */ 5282 if (softc->state != DA_STATE_NORMAL) 5283 return; 5284 5285 status = cam_periph_acquire(periph); 5286 KASSERT(status == CAM_REQ_CMP, 5287 ("dareprobe: cam_periph_acquire failed")); 5288 5289 if (softc->flags & DA_FLAG_CAN_RC16) 5290 softc->state = DA_STATE_PROBE_RC16; 5291 else 5292 softc->state = DA_STATE_PROBE_RC; 5293 5294 xpt_schedule(periph, CAM_PRIORITY_DEV); 5295 } 5296 5297 static int 5298 daerror(union ccb *ccb, u_int32_t cam_flags, u_int32_t sense_flags) 5299 { 5300 struct da_softc *softc; 5301 struct cam_periph *periph; 5302 int error, error_code, sense_key, asc, ascq; 5303 5304 #if defined(BUF_TRACKING) || defined(FULL_BUF_TRACKING) 5305 if (ccb->csio.bio != NULL) 5306 biotrack(ccb->csio.bio, __func__); 5307 #endif 5308 5309 periph = xpt_path_periph(ccb->ccb_h.path); 5310 softc = (struct da_softc *)periph->softc; 5311 5312 /* 5313 * Automatically detect devices that do not support 5314 * READ(6)/WRITE(6) and upgrade to using 10 byte cdbs. 5315 */ 5316 error = 0; 5317 if ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_INVALID) { 5318 error = cmd6workaround(ccb); 5319 } else if (scsi_extract_sense_ccb(ccb, 5320 &error_code, &sense_key, &asc, &ascq)) { 5321 if (sense_key == SSD_KEY_ILLEGAL_REQUEST) 5322 error = cmd6workaround(ccb); 5323 /* 5324 * If the target replied with CAPACITY DATA HAS CHANGED UA, 5325 * query the capacity and notify upper layers. 5326 */ 5327 else if (sense_key == SSD_KEY_UNIT_ATTENTION && 5328 asc == 0x2A && ascq == 0x09) { 5329 xpt_print(periph->path, "Capacity data has changed\n"); 5330 softc->flags &= ~DA_FLAG_PROBED; 5331 dareprobe(periph); 5332 sense_flags |= SF_NO_PRINT; 5333 } else if (sense_key == SSD_KEY_UNIT_ATTENTION && 5334 asc == 0x28 && ascq == 0x00) { 5335 softc->flags &= ~DA_FLAG_PROBED; 5336 disk_media_changed(softc->disk, M_NOWAIT); 5337 } else if (sense_key == SSD_KEY_UNIT_ATTENTION && 5338 asc == 0x3F && ascq == 0x03) { 5339 xpt_print(periph->path, "INQUIRY data has changed\n"); 5340 softc->flags &= ~DA_FLAG_PROBED; 5341 dareprobe(periph); 5342 sense_flags |= SF_NO_PRINT; 5343 } else if (sense_key == SSD_KEY_NOT_READY && 5344 asc == 0x3a && (softc->flags & DA_FLAG_PACK_INVALID) == 0) { 5345 softc->flags |= DA_FLAG_PACK_INVALID; 5346 disk_media_gone(softc->disk, M_NOWAIT); 5347 } 5348 } 5349 if (error == ERESTART) 5350 return (ERESTART); 5351 5352 #ifdef CAM_IO_STATS 5353 switch (ccb->ccb_h.status & CAM_STATUS_MASK) { 5354 case CAM_CMD_TIMEOUT: 5355 softc->timeouts++; 5356 break; 5357 case CAM_REQ_ABORTED: 5358 case CAM_REQ_CMP_ERR: 5359 case CAM_REQ_TERMIO: 5360 case CAM_UNREC_HBA_ERROR: 5361 case CAM_DATA_RUN_ERR: 5362 softc->errors++; 5363 break; 5364 default: 5365 break; 5366 } 5367 #endif 5368 5369 /* 5370 * XXX 5371 * Until we have a better way of doing pack validation, 5372 * don't treat UAs as errors. 5373 */ 5374 sense_flags |= SF_RETRY_UA; 5375 5376 if (softc->quirks & DA_Q_RETRY_BUSY) 5377 sense_flags |= SF_RETRY_BUSY; 5378 return(cam_periph_error(ccb, cam_flags, sense_flags, 5379 &softc->saved_ccb)); 5380 } 5381 5382 static void 5383 damediapoll(void *arg) 5384 { 5385 struct cam_periph *periph = arg; 5386 struct da_softc *softc = periph->softc; 5387 5388 if (!cam_iosched_has_work_flags(softc->cam_iosched, DA_WORK_TUR) && 5389 LIST_EMPTY(&softc->pending_ccbs)) { 5390 if (cam_periph_acquire(periph) == CAM_REQ_CMP) { 5391 cam_iosched_set_work_flags(softc->cam_iosched, DA_WORK_TUR); 5392 daschedule(periph); 5393 } 5394 } 5395 /* Queue us up again */ 5396 if (da_poll_period != 0) 5397 callout_schedule(&softc->mediapoll_c, da_poll_period * hz); 5398 } 5399 5400 static void 5401 daprevent(struct cam_periph *periph, int action) 5402 { 5403 struct da_softc *softc; 5404 union ccb *ccb; 5405 int error; 5406 5407 softc = (struct da_softc *)periph->softc; 5408 5409 if (((action == PR_ALLOW) 5410 && (softc->flags & DA_FLAG_PACK_LOCKED) == 0) 5411 || ((action == PR_PREVENT) 5412 && (softc->flags & DA_FLAG_PACK_LOCKED) != 0)) { 5413 return; 5414 } 5415 5416 ccb = cam_periph_getccb(periph, CAM_PRIORITY_NORMAL); 5417 5418 scsi_prevent(&ccb->csio, 5419 /*retries*/1, 5420 /*cbcfp*/dadone, 5421 MSG_SIMPLE_Q_TAG, 5422 action, 5423 SSD_FULL_SIZE, 5424 5000); 5425 5426 error = cam_periph_runccb(ccb, daerror, CAM_RETRY_SELTO, 5427 SF_RETRY_UA | SF_NO_PRINT, softc->disk->d_devstat); 5428 5429 if (error == 0) { 5430 if (action == PR_ALLOW) 5431 softc->flags &= ~DA_FLAG_PACK_LOCKED; 5432 else 5433 softc->flags |= DA_FLAG_PACK_LOCKED; 5434 } 5435 5436 xpt_release_ccb(ccb); 5437 } 5438 5439 static void 5440 dasetgeom(struct cam_periph *periph, uint32_t block_len, uint64_t maxsector, 5441 struct scsi_read_capacity_data_long *rcaplong, size_t rcap_len) 5442 { 5443 struct ccb_calc_geometry ccg; 5444 struct da_softc *softc; 5445 struct disk_params *dp; 5446 u_int lbppbe, lalba; 5447 int error; 5448 5449 softc = (struct da_softc *)periph->softc; 5450 5451 dp = &softc->params; 5452 dp->secsize = block_len; 5453 dp->sectors = maxsector + 1; 5454 if (rcaplong != NULL) { 5455 lbppbe = rcaplong->prot_lbppbe & SRC16_LBPPBE; 5456 lalba = scsi_2btoul(rcaplong->lalba_lbp); 5457 lalba &= SRC16_LALBA_A; 5458 } else { 5459 lbppbe = 0; 5460 lalba = 0; 5461 } 5462 5463 if (lbppbe > 0) { 5464 dp->stripesize = block_len << lbppbe; 5465 dp->stripeoffset = (dp->stripesize - block_len * lalba) % 5466 dp->stripesize; 5467 } else if (softc->quirks & DA_Q_4K) { 5468 dp->stripesize = 4096; 5469 dp->stripeoffset = 0; 5470 } else if (softc->unmap_gran != 0) { 5471 dp->stripesize = block_len * softc->unmap_gran; 5472 dp->stripeoffset = (dp->stripesize - block_len * 5473 softc->unmap_gran_align) % dp->stripesize; 5474 } else { 5475 dp->stripesize = 0; 5476 dp->stripeoffset = 0; 5477 } 5478 /* 5479 * Have the controller provide us with a geometry 5480 * for this disk. The only time the geometry 5481 * matters is when we boot and the controller 5482 * is the only one knowledgeable enough to come 5483 * up with something that will make this a bootable 5484 * device. 5485 */ 5486 xpt_setup_ccb(&ccg.ccb_h, periph->path, CAM_PRIORITY_NORMAL); 5487 ccg.ccb_h.func_code = XPT_CALC_GEOMETRY; 5488 ccg.block_size = dp->secsize; 5489 ccg.volume_size = dp->sectors; 5490 ccg.heads = 0; 5491 ccg.secs_per_track = 0; 5492 ccg.cylinders = 0; 5493 xpt_action((union ccb*)&ccg); 5494 if ((ccg.ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) { 5495 /* 5496 * We don't know what went wrong here- but just pick 5497 * a geometry so we don't have nasty things like divide 5498 * by zero. 5499 */ 5500 dp->heads = 255; 5501 dp->secs_per_track = 255; 5502 dp->cylinders = dp->sectors / (255 * 255); 5503 if (dp->cylinders == 0) { 5504 dp->cylinders = 1; 5505 } 5506 } else { 5507 dp->heads = ccg.heads; 5508 dp->secs_per_track = ccg.secs_per_track; 5509 dp->cylinders = ccg.cylinders; 5510 } 5511 5512 /* 5513 * If the user supplied a read capacity buffer, and if it is 5514 * different than the previous buffer, update the data in the EDT. 5515 * If it's the same, we don't bother. This avoids sending an 5516 * update every time someone opens this device. 5517 */ 5518 if ((rcaplong != NULL) 5519 && (bcmp(rcaplong, &softc->rcaplong, 5520 min(sizeof(softc->rcaplong), rcap_len)) != 0)) { 5521 struct ccb_dev_advinfo cdai; 5522 5523 xpt_setup_ccb(&cdai.ccb_h, periph->path, CAM_PRIORITY_NORMAL); 5524 cdai.ccb_h.func_code = XPT_DEV_ADVINFO; 5525 cdai.buftype = CDAI_TYPE_RCAPLONG; 5526 cdai.flags = CDAI_FLAG_STORE; 5527 cdai.bufsiz = rcap_len; 5528 cdai.buf = (uint8_t *)rcaplong; 5529 xpt_action((union ccb *)&cdai); 5530 if ((cdai.ccb_h.status & CAM_DEV_QFRZN) != 0) 5531 cam_release_devq(cdai.ccb_h.path, 0, 0, 0, FALSE); 5532 if (cdai.ccb_h.status != CAM_REQ_CMP) { 5533 xpt_print(periph->path, "%s: failed to set read " 5534 "capacity advinfo\n", __func__); 5535 /* Use cam_error_print() to decode the status */ 5536 cam_error_print((union ccb *)&cdai, CAM_ESF_CAM_STATUS, 5537 CAM_EPF_ALL); 5538 } else { 5539 bcopy(rcaplong, &softc->rcaplong, 5540 min(sizeof(softc->rcaplong), rcap_len)); 5541 } 5542 } 5543 5544 softc->disk->d_sectorsize = softc->params.secsize; 5545 softc->disk->d_mediasize = softc->params.secsize * (off_t)softc->params.sectors; 5546 softc->disk->d_stripesize = softc->params.stripesize; 5547 softc->disk->d_stripeoffset = softc->params.stripeoffset; 5548 /* XXX: these are not actually "firmware" values, so they may be wrong */ 5549 softc->disk->d_fwsectors = softc->params.secs_per_track; 5550 softc->disk->d_fwheads = softc->params.heads; 5551 softc->disk->d_devstat->block_size = softc->params.secsize; 5552 softc->disk->d_devstat->flags &= ~DEVSTAT_BS_UNAVAILABLE; 5553 5554 error = disk_resize(softc->disk, M_NOWAIT); 5555 if (error != 0) 5556 xpt_print(periph->path, "disk_resize(9) failed, error = %d\n", error); 5557 } 5558 5559 static void 5560 dasendorderedtag(void *arg) 5561 { 5562 struct da_softc *softc = arg; 5563 5564 if (da_send_ordered) { 5565 if (!LIST_EMPTY(&softc->pending_ccbs)) { 5566 if ((softc->flags & DA_FLAG_WAS_OTAG) == 0) 5567 softc->flags |= DA_FLAG_NEED_OTAG; 5568 softc->flags &= ~DA_FLAG_WAS_OTAG; 5569 } 5570 } 5571 /* Queue us up again */ 5572 callout_reset(&softc->sendordered_c, 5573 (da_default_timeout * hz) / DA_ORDEREDTAG_INTERVAL, 5574 dasendorderedtag, softc); 5575 } 5576 5577 /* 5578 * Step through all DA peripheral drivers, and if the device is still open, 5579 * sync the disk cache to physical media. 5580 */ 5581 static void 5582 dashutdown(void * arg, int howto) 5583 { 5584 struct cam_periph *periph; 5585 struct da_softc *softc; 5586 union ccb *ccb; 5587 int error; 5588 5589 CAM_PERIPH_FOREACH(periph, &dadriver) { 5590 softc = (struct da_softc *)periph->softc; 5591 if (SCHEDULER_STOPPED()) { 5592 /* If we paniced with the lock held, do not recurse. */ 5593 if (!cam_periph_owned(periph) && 5594 (softc->flags & DA_FLAG_OPEN)) { 5595 dadump(softc->disk, NULL, 0, 0, 0); 5596 } 5597 continue; 5598 } 5599 cam_periph_lock(periph); 5600 5601 /* 5602 * We only sync the cache if the drive is still open, and 5603 * if the drive is capable of it.. 5604 */ 5605 if (((softc->flags & DA_FLAG_OPEN) == 0) 5606 || (softc->quirks & DA_Q_NO_SYNC_CACHE)) { 5607 cam_periph_unlock(periph); 5608 continue; 5609 } 5610 5611 ccb = cam_periph_getccb(periph, CAM_PRIORITY_NORMAL); 5612 scsi_synchronize_cache(&ccb->csio, 5613 /*retries*/0, 5614 /*cbfcnp*/dadone, 5615 MSG_SIMPLE_Q_TAG, 5616 /*begin_lba*/0, /* whole disk */ 5617 /*lb_count*/0, 5618 SSD_FULL_SIZE, 5619 60 * 60 * 1000); 5620 5621 error = cam_periph_runccb(ccb, daerror, /*cam_flags*/0, 5622 /*sense_flags*/ SF_NO_RECOVERY | SF_NO_RETRY | SF_QUIET_IR, 5623 softc->disk->d_devstat); 5624 if (error != 0) 5625 xpt_print(periph->path, "Synchronize cache failed\n"); 5626 xpt_release_ccb(ccb); 5627 cam_periph_unlock(periph); 5628 } 5629 } 5630 5631 #else /* !_KERNEL */ 5632 5633 /* 5634 * XXX These are only left out of the kernel build to silence warnings. If, 5635 * for some reason these functions are used in the kernel, the ifdefs should 5636 * be moved so they are included both in the kernel and userland. 5637 */ 5638 void 5639 scsi_format_unit(struct ccb_scsiio *csio, u_int32_t retries, 5640 void (*cbfcnp)(struct cam_periph *, union ccb *), 5641 u_int8_t tag_action, u_int8_t byte2, u_int16_t ileave, 5642 u_int8_t *data_ptr, u_int32_t dxfer_len, u_int8_t sense_len, 5643 u_int32_t timeout) 5644 { 5645 struct scsi_format_unit *scsi_cmd; 5646 5647 scsi_cmd = (struct scsi_format_unit *)&csio->cdb_io.cdb_bytes; 5648 scsi_cmd->opcode = FORMAT_UNIT; 5649 scsi_cmd->byte2 = byte2; 5650 scsi_ulto2b(ileave, scsi_cmd->interleave); 5651 5652 cam_fill_csio(csio, 5653 retries, 5654 cbfcnp, 5655 /*flags*/ (dxfer_len > 0) ? CAM_DIR_OUT : CAM_DIR_NONE, 5656 tag_action, 5657 data_ptr, 5658 dxfer_len, 5659 sense_len, 5660 sizeof(*scsi_cmd), 5661 timeout); 5662 } 5663 5664 void 5665 scsi_read_defects(struct ccb_scsiio *csio, uint32_t retries, 5666 void (*cbfcnp)(struct cam_periph *, union ccb *), 5667 uint8_t tag_action, uint8_t list_format, 5668 uint32_t addr_desc_index, uint8_t *data_ptr, 5669 uint32_t dxfer_len, int minimum_cmd_size, 5670 uint8_t sense_len, uint32_t timeout) 5671 { 5672 uint8_t cdb_len; 5673 5674 /* 5675 * These conditions allow using the 10 byte command. Otherwise we 5676 * need to use the 12 byte command. 5677 */ 5678 if ((minimum_cmd_size <= 10) 5679 && (addr_desc_index == 0) 5680 && (dxfer_len <= SRDD10_MAX_LENGTH)) { 5681 struct scsi_read_defect_data_10 *cdb10; 5682 5683 cdb10 = (struct scsi_read_defect_data_10 *) 5684 &csio->cdb_io.cdb_bytes; 5685 5686 cdb_len = sizeof(*cdb10); 5687 bzero(cdb10, cdb_len); 5688 cdb10->opcode = READ_DEFECT_DATA_10; 5689 cdb10->format = list_format; 5690 scsi_ulto2b(dxfer_len, cdb10->alloc_length); 5691 } else { 5692 struct scsi_read_defect_data_12 *cdb12; 5693 5694 cdb12 = (struct scsi_read_defect_data_12 *) 5695 &csio->cdb_io.cdb_bytes; 5696 5697 cdb_len = sizeof(*cdb12); 5698 bzero(cdb12, cdb_len); 5699 cdb12->opcode = READ_DEFECT_DATA_12; 5700 cdb12->format = list_format; 5701 scsi_ulto4b(dxfer_len, cdb12->alloc_length); 5702 scsi_ulto4b(addr_desc_index, cdb12->address_descriptor_index); 5703 } 5704 5705 cam_fill_csio(csio, 5706 retries, 5707 cbfcnp, 5708 /*flags*/ CAM_DIR_IN, 5709 tag_action, 5710 data_ptr, 5711 dxfer_len, 5712 sense_len, 5713 cdb_len, 5714 timeout); 5715 } 5716 5717 void 5718 scsi_sanitize(struct ccb_scsiio *csio, u_int32_t retries, 5719 void (*cbfcnp)(struct cam_periph *, union ccb *), 5720 u_int8_t tag_action, u_int8_t byte2, u_int16_t control, 5721 u_int8_t *data_ptr, u_int32_t dxfer_len, u_int8_t sense_len, 5722 u_int32_t timeout) 5723 { 5724 struct scsi_sanitize *scsi_cmd; 5725 5726 scsi_cmd = (struct scsi_sanitize *)&csio->cdb_io.cdb_bytes; 5727 scsi_cmd->opcode = SANITIZE; 5728 scsi_cmd->byte2 = byte2; 5729 scsi_cmd->control = control; 5730 scsi_ulto2b(dxfer_len, scsi_cmd->length); 5731 5732 cam_fill_csio(csio, 5733 retries, 5734 cbfcnp, 5735 /*flags*/ (dxfer_len > 0) ? CAM_DIR_OUT : CAM_DIR_NONE, 5736 tag_action, 5737 data_ptr, 5738 dxfer_len, 5739 sense_len, 5740 sizeof(*scsi_cmd), 5741 timeout); 5742 } 5743 5744 #endif /* _KERNEL */ 5745 5746 void 5747 scsi_zbc_out(struct ccb_scsiio *csio, uint32_t retries, 5748 void (*cbfcnp)(struct cam_periph *, union ccb *), 5749 uint8_t tag_action, uint8_t service_action, uint64_t zone_id, 5750 uint8_t zone_flags, uint8_t *data_ptr, uint32_t dxfer_len, 5751 uint8_t sense_len, uint32_t timeout) 5752 { 5753 struct scsi_zbc_out *scsi_cmd; 5754 5755 scsi_cmd = (struct scsi_zbc_out *)&csio->cdb_io.cdb_bytes; 5756 scsi_cmd->opcode = ZBC_OUT; 5757 scsi_cmd->service_action = service_action; 5758 scsi_u64to8b(zone_id, scsi_cmd->zone_id); 5759 scsi_cmd->zone_flags = zone_flags; 5760 5761 cam_fill_csio(csio, 5762 retries, 5763 cbfcnp, 5764 /*flags*/ (dxfer_len > 0) ? CAM_DIR_OUT : CAM_DIR_NONE, 5765 tag_action, 5766 data_ptr, 5767 dxfer_len, 5768 sense_len, 5769 sizeof(*scsi_cmd), 5770 timeout); 5771 } 5772 5773 void 5774 scsi_zbc_in(struct ccb_scsiio *csio, uint32_t retries, 5775 void (*cbfcnp)(struct cam_periph *, union ccb *), 5776 uint8_t tag_action, uint8_t service_action, uint64_t zone_start_lba, 5777 uint8_t zone_options, uint8_t *data_ptr, uint32_t dxfer_len, 5778 uint8_t sense_len, uint32_t timeout) 5779 { 5780 struct scsi_zbc_in *scsi_cmd; 5781 5782 scsi_cmd = (struct scsi_zbc_in *)&csio->cdb_io.cdb_bytes; 5783 scsi_cmd->opcode = ZBC_IN; 5784 scsi_cmd->service_action = service_action; 5785 scsi_u64to8b(zone_start_lba, scsi_cmd->zone_start_lba); 5786 scsi_cmd->zone_options = zone_options; 5787 5788 cam_fill_csio(csio, 5789 retries, 5790 cbfcnp, 5791 /*flags*/ (dxfer_len > 0) ? CAM_DIR_IN : CAM_DIR_NONE, 5792 tag_action, 5793 data_ptr, 5794 dxfer_len, 5795 sense_len, 5796 sizeof(*scsi_cmd), 5797 timeout); 5798 5799 } 5800 5801 int 5802 scsi_ata_zac_mgmt_out(struct ccb_scsiio *csio, uint32_t retries, 5803 void (*cbfcnp)(struct cam_periph *, union ccb *), 5804 uint8_t tag_action, int use_ncq, 5805 uint8_t zm_action, uint64_t zone_id, uint8_t zone_flags, 5806 uint8_t *data_ptr, uint32_t dxfer_len, 5807 uint8_t *cdb_storage, size_t cdb_storage_len, 5808 uint8_t sense_len, uint32_t timeout) 5809 { 5810 uint8_t command_out, protocol, ata_flags; 5811 uint16_t features_out; 5812 uint32_t sectors_out, auxiliary; 5813 int retval; 5814 5815 retval = 0; 5816 5817 if (use_ncq == 0) { 5818 command_out = ATA_ZAC_MANAGEMENT_OUT; 5819 features_out = (zm_action & 0xf) | (zone_flags << 8); 5820 ata_flags = AP_FLAG_BYT_BLOK_BLOCKS; 5821 if (dxfer_len == 0) { 5822 protocol = AP_PROTO_NON_DATA; 5823 ata_flags |= AP_FLAG_TLEN_NO_DATA; 5824 sectors_out = 0; 5825 } else { 5826 protocol = AP_PROTO_DMA; 5827 ata_flags |= AP_FLAG_TLEN_SECT_CNT | 5828 AP_FLAG_TDIR_TO_DEV; 5829 sectors_out = ((dxfer_len >> 9) & 0xffff); 5830 } 5831 auxiliary = 0; 5832 } else { 5833 ata_flags = AP_FLAG_BYT_BLOK_BLOCKS; 5834 if (dxfer_len == 0) { 5835 command_out = ATA_NCQ_NON_DATA; 5836 features_out = ATA_NCQ_ZAC_MGMT_OUT; 5837 /* 5838 * We're assuming the SCSI to ATA translation layer 5839 * will set the NCQ tag number in the tag field. 5840 * That isn't clear from the SAT-4 spec (as of rev 05). 5841 */ 5842 sectors_out = 0; 5843 ata_flags |= AP_FLAG_TLEN_NO_DATA; 5844 } else { 5845 command_out = ATA_SEND_FPDMA_QUEUED; 5846 /* 5847 * Note that we're defaulting to normal priority, 5848 * and assuming that the SCSI to ATA translation 5849 * layer will insert the NCQ tag number in the tag 5850 * field. That isn't clear in the SAT-4 spec (as 5851 * of rev 05). 5852 */ 5853 sectors_out = ATA_SFPDMA_ZAC_MGMT_OUT << 8; 5854 5855 ata_flags |= AP_FLAG_TLEN_FEAT | 5856 AP_FLAG_TDIR_TO_DEV; 5857 5858 /* 5859 * For SEND FPDMA QUEUED, the transfer length is 5860 * encoded in the FEATURE register, and 0 means 5861 * that 65536 512 byte blocks are to be tranferred. 5862 * In practice, it seems unlikely that we'll see 5863 * a transfer that large, and it may confuse the 5864 * the SAT layer, because generally that means that 5865 * 0 bytes should be transferred. 5866 */ 5867 if (dxfer_len == (65536 * 512)) { 5868 features_out = 0; 5869 } else if (dxfer_len <= (65535 * 512)) { 5870 features_out = ((dxfer_len >> 9) & 0xffff); 5871 } else { 5872 /* The transfer is too big. */ 5873 retval = 1; 5874 goto bailout; 5875 } 5876 5877 } 5878 5879 auxiliary = (zm_action & 0xf) | (zone_flags << 8); 5880 protocol = AP_PROTO_FPDMA; 5881 } 5882 5883 protocol |= AP_EXTEND; 5884 5885 retval = scsi_ata_pass(csio, 5886 retries, 5887 cbfcnp, 5888 /*flags*/ (dxfer_len > 0) ? CAM_DIR_OUT : CAM_DIR_NONE, 5889 tag_action, 5890 /*protocol*/ protocol, 5891 /*ata_flags*/ ata_flags, 5892 /*features*/ features_out, 5893 /*sector_count*/ sectors_out, 5894 /*lba*/ zone_id, 5895 /*command*/ command_out, 5896 /*device*/ 0, 5897 /*icc*/ 0, 5898 /*auxiliary*/ auxiliary, 5899 /*control*/ 0, 5900 /*data_ptr*/ data_ptr, 5901 /*dxfer_len*/ dxfer_len, 5902 /*cdb_storage*/ cdb_storage, 5903 /*cdb_storage_len*/ cdb_storage_len, 5904 /*minimum_cmd_size*/ 0, 5905 /*sense_len*/ SSD_FULL_SIZE, 5906 /*timeout*/ timeout); 5907 5908 bailout: 5909 5910 return (retval); 5911 } 5912 5913 int 5914 scsi_ata_zac_mgmt_in(struct ccb_scsiio *csio, uint32_t retries, 5915 void (*cbfcnp)(struct cam_periph *, union ccb *), 5916 uint8_t tag_action, int use_ncq, 5917 uint8_t zm_action, uint64_t zone_id, uint8_t zone_flags, 5918 uint8_t *data_ptr, uint32_t dxfer_len, 5919 uint8_t *cdb_storage, size_t cdb_storage_len, 5920 uint8_t sense_len, uint32_t timeout) 5921 { 5922 uint8_t command_out, protocol; 5923 uint16_t features_out, sectors_out; 5924 uint32_t auxiliary; 5925 int ata_flags; 5926 int retval; 5927 5928 retval = 0; 5929 ata_flags = AP_FLAG_TDIR_FROM_DEV | AP_FLAG_BYT_BLOK_BLOCKS; 5930 5931 if (use_ncq == 0) { 5932 command_out = ATA_ZAC_MANAGEMENT_IN; 5933 /* XXX KDM put a macro here */ 5934 features_out = (zm_action & 0xf) | (zone_flags << 8); 5935 sectors_out = dxfer_len >> 9; /* XXX KDM macro */ 5936 protocol = AP_PROTO_DMA; 5937 ata_flags |= AP_FLAG_TLEN_SECT_CNT; 5938 auxiliary = 0; 5939 } else { 5940 ata_flags |= AP_FLAG_TLEN_FEAT; 5941 5942 command_out = ATA_RECV_FPDMA_QUEUED; 5943 sectors_out = ATA_RFPDMA_ZAC_MGMT_IN << 8; 5944 5945 /* 5946 * For RECEIVE FPDMA QUEUED, the transfer length is 5947 * encoded in the FEATURE register, and 0 means 5948 * that 65536 512 byte blocks are to be tranferred. 5949 * In practice, it seems unlikely that we'll see 5950 * a transfer that large, and it may confuse the 5951 * the SAT layer, because generally that means that 5952 * 0 bytes should be transferred. 5953 */ 5954 if (dxfer_len == (65536 * 512)) { 5955 features_out = 0; 5956 } else if (dxfer_len <= (65535 * 512)) { 5957 features_out = ((dxfer_len >> 9) & 0xffff); 5958 } else { 5959 /* The transfer is too big. */ 5960 retval = 1; 5961 goto bailout; 5962 } 5963 auxiliary = (zm_action & 0xf) | (zone_flags << 8), 5964 protocol = AP_PROTO_FPDMA; 5965 } 5966 5967 protocol |= AP_EXTEND; 5968 5969 retval = scsi_ata_pass(csio, 5970 retries, 5971 cbfcnp, 5972 /*flags*/ CAM_DIR_IN, 5973 tag_action, 5974 /*protocol*/ protocol, 5975 /*ata_flags*/ ata_flags, 5976 /*features*/ features_out, 5977 /*sector_count*/ sectors_out, 5978 /*lba*/ zone_id, 5979 /*command*/ command_out, 5980 /*device*/ 0, 5981 /*icc*/ 0, 5982 /*auxiliary*/ auxiliary, 5983 /*control*/ 0, 5984 /*data_ptr*/ data_ptr, 5985 /*dxfer_len*/ (dxfer_len >> 9) * 512, /* XXX KDM */ 5986 /*cdb_storage*/ cdb_storage, 5987 /*cdb_storage_len*/ cdb_storage_len, 5988 /*minimum_cmd_size*/ 0, 5989 /*sense_len*/ SSD_FULL_SIZE, 5990 /*timeout*/ timeout); 5991 5992 bailout: 5993 return (retval); 5994 } 5995