1zstd(1) -- zstd, zstdmt, unzstd, zstdcat - Compress or decompress .zst files 2============================================================================ 3 4SYNOPSIS 5-------- 6 7`zstd` [*OPTIONS*] [-|_INPUT-FILE_] [-o _OUTPUT-FILE_] 8 9`zstdmt` is equivalent to `zstd -T0` 10 11`unzstd` is equivalent to `zstd -d` 12 13`zstdcat` is equivalent to `zstd -dcf` 14 15 16DESCRIPTION 17----------- 18`zstd` is a fast lossless compression algorithm and data compression tool, 19with command line syntax similar to `gzip (1)` and `xz (1)`. 20It is based on the **LZ77** family, with further FSE & huff0 entropy stages. 21`zstd` offers highly configurable compression speed, 22with fast modes at > 200 MB/s per core, 23and strong modes nearing lzma compression ratios. 24It also features a very fast decoder, with speeds > 500 MB/s per core. 25 26`zstd` command line syntax is generally similar to gzip, 27but features the following differences : 28 29 - Source files are preserved by default. 30 It's possible to remove them automatically by using the `--rm` command. 31 - When compressing a single file, `zstd` displays progress notifications 32 and result summary by default. 33 Use `-q` to turn them off. 34 - `zstd` does not accept input from console, 35 but it properly accepts `stdin` when it's not the console. 36 - `zstd` displays a short help page when command line is an error. 37 Use `-q` to turn it off. 38 39`zstd` compresses or decompresses each _file_ according to the selected 40operation mode. 41If no _files_ are given or _file_ is `-`, `zstd` reads from standard input 42and writes the processed data to standard output. 43`zstd` will refuse to write compressed data to standard output 44if it is a terminal : it will display an error message and skip the _file_. 45Similarly, `zstd` will refuse to read compressed data from standard input 46if it is a terminal. 47 48Unless `--stdout` or `-o` is specified, _files_ are written to a new file 49whose name is derived from the source _file_ name: 50 51* When compressing, the suffix `.zst` is appended to the source filename to 52 get the target filename. 53* When decompressing, the `.zst` suffix is removed from the source filename to 54 get the target filename 55 56### Concatenation with .zst files 57It is possible to concatenate `.zst` files as is. 58`zstd` will decompress such files as if they were a single `.zst` file. 59 60OPTIONS 61------- 62 63### Integer suffixes and special values 64In most places where an integer argument is expected, 65an optional suffix is supported to easily indicate large integers. 66There must be no space between the integer and the suffix. 67 68* `KiB`: 69 Multiply the integer by 1,024 (2\^10). 70 `Ki`, `K`, and `KB` are accepted as synonyms for `KiB`. 71* `MiB`: 72 Multiply the integer by 1,048,576 (2\^20). 73 `Mi`, `M`, and `MB` are accepted as synonyms for `MiB`. 74 75### Operation mode 76If multiple operation mode options are given, 77the last one takes effect. 78 79* `-z`, `--compress`: 80 Compress. 81 This is the default operation mode when no operation mode option is specified 82 and no other operation mode is implied from the command name 83 (for example, `unzstd` implies `--decompress`). 84* `-d`, `--decompress`, `--uncompress`: 85 Decompress. 86* `-t`, `--test`: 87 Test the integrity of compressed _files_. 88 This option is equivalent to `--decompress --stdout` except that the 89 decompressed data is discarded instead of being written to standard output. 90 No files are created or removed. 91* `-b#`: 92 Benchmark file(s) using compression level # 93* `--train FILEs`: 94 Use FILEs as a training set to create a dictionary. 95 The training set should contain a lot of small files (> 100). 96* `-l`, `--list`: 97 Display information related to a zstd compressed file, such as size, ratio, and checksum. 98 Some of these fields may not be available. 99 This command can be augmented with the `-v` modifier. 100 101### Operation modifiers 102 103* `-#`: 104 `#` compression level \[1-19] (default: 3) 105* `--fast[=#]`: 106 switch to ultra-fast compression levels. 107 If `=#` is not present, it defaults to `1`. 108 The higher the value, the faster the compression speed, 109 at the cost of some compression ratio. 110 This setting overwrites compression level if one was set previously. 111 Similarly, if a compression level is set after `--fast`, it overrides it. 112* `--ultra`: 113 unlocks high compression levels 20+ (maximum 22), using a lot more memory. 114 Note that decompression will also require more memory when using these levels. 115* `--long[=#]`: 116 enables long distance matching with `#` `windowLog`, if not `#` is not 117 present it defaults to `27`. 118 This increases the window size (`windowLog`) and memory usage for both the 119 compressor and decompressor. 120 This setting is designed to improve the compression ratio for files with 121 long matches at a large distance. 122 123 Note: If `windowLog` is set to larger than 27, `--long=windowLog` or 124 `--memory=windowSize` needs to be passed to the decompressor. 125* `-T#`, `--threads=#`: 126 Compress using `#` working threads (default: 1). 127 If `#` is 0, attempt to detect and use the number of physical CPU cores. 128 In all cases, the nb of threads is capped to ZSTDMT_NBTHREADS_MAX==200. 129 This modifier does nothing if `zstd` is compiled without multithread support. 130* `--single-thread`: 131 Does not spawn a thread for compression, use a single thread for both I/O and compression. 132 In this mode, compression is serialized with I/O, which is slightly slower. 133 (This is different from `-T1`, which spawns 1 compression thread in parallel of I/O). 134 This mode is the only one available when multithread support is disabled. 135 Single-thread mode features lower memory usage. 136 Final compressed result is slightly different from `-T1`. 137* `--adapt[=min=#,max=#]` : 138 `zstd` will dynamically adapt compression level to perceived I/O conditions. 139 Compression level adaptation can be observed live by using command `-v`. 140 Adaptation can be constrained between supplied `min` and `max` levels. 141 The feature works when combined with multi-threading and `--long` mode. 142 It does not work with `--single-thread`. 143 It sets window size to 8 MB by default (can be changed manually, see `wlog`). 144 Due to the chaotic nature of dynamic adaptation, compressed result is not reproducible. 145 _note_ : at the time of this writing, `--adapt` can remain stuck at low speed 146 when combined with multiple worker threads (>=2). 147* `--stream-size=#` : 148 Sets the pledged source size of input coming from a stream. This value must be exact, as it 149 will be included in the produced frame header. Incorrect stream sizes will cause an error. 150 This information will be used to better optimize compression parameters, resulting in 151 better and potentially faster compression, especially for smaller source sizes. 152* `--size-hint=#`: 153 When handling input from a stream, `zstd` must guess how large the source size 154 will be when optimizing compression parameters. If the stream size is relatively 155 small, this guess may be a poor one, resulting in a higher compression ratio than 156 expected. This feature allows for controlling the guess when needed. 157 Exact guesses result in better compression ratios. Overestimates result in slightly 158 degraded compression ratios, while underestimates may result in significant degradation. 159* `--rsyncable` : 160 `zstd` will periodically synchronize the compression state to make the 161 compressed file more rsync-friendly. There is a negligible impact to 162 compression ratio, and the faster compression levels will see a small 163 compression speed hit. 164 This feature does not work with `--single-thread`. You probably don't want 165 to use it with long range mode, since it will decrease the effectiveness of 166 the synchronization points, but your milage may vary. 167* `-D file`: 168 use `file` as Dictionary to compress or decompress FILE(s) 169* `--no-dictID`: 170 do not store dictionary ID within frame header (dictionary compression). 171 The decoder will have to rely on implicit knowledge about which dictionary to use, 172 it won't be able to check if it's correct. 173* `-o file`: 174 save result into `file` (only possible with a single _INPUT-FILE_) 175* `-f`, `--force`: 176 overwrite output without prompting, and (de)compress symbolic links 177* `-c`, `--stdout`: 178 force write to standard output, even if it is the console 179* `--[no-]sparse`: 180 enable / disable sparse FS support, 181 to make files with many zeroes smaller on disk. 182 Creating sparse files may save disk space and speed up decompression by 183 reducing the amount of disk I/O. 184 default: enabled when output is into a file, 185 and disabled when output is stdout. 186 This setting overrides default and can force sparse mode over stdout. 187* `--rm`: 188 remove source file(s) after successful compression or decompression 189* `-k`, `--keep`: 190 keep source file(s) after successful compression or decompression. 191 This is the default behavior. 192* `-r`: 193 operate recursively on directories 194* `--output-dir-flat[=dir]`: 195 resulting files are stored into target `dir` directory, 196 instead of same directory as origin file. 197 Be aware that this command can introduce name collision issues, 198 if multiple files, from different directories, end up having the same name. 199 Collision resolution ensures first file with a given name will be present in `dir`, 200 while in combination with `-f`, the last file will be present instead. 201* `--format=FORMAT`: 202 compress and decompress in other formats. If compiled with 203 support, zstd can compress to or decompress from other compression algorithm 204 formats. Possibly available options are `zstd`, `gzip`, `xz`, `lzma`, and `lz4`. 205 If no such format is provided, `zstd` is the default. 206* `-h`/`-H`, `--help`: 207 display help/long help and exit 208* `-V`, `--version`: 209 display version number and exit. 210 Advanced : `-vV` also displays supported formats. 211 `-vvV` also displays POSIX support. 212* `-v`: 213 verbose mode 214* `-q`, `--quiet`: 215 suppress warnings, interactivity, and notifications. 216 specify twice to suppress errors too. 217* `--no-progress`: 218 do not display the progress bar, but keep all other messages. 219* `-C`, `--[no-]check`: 220 add integrity check computed from uncompressed data (default: enabled) 221* `--`: 222 All arguments after `--` are treated as files 223 224### Restricted usage of Environment Variables 225 226Using environment variables to set parameters has security implications. 227Therefore, this avenue is intentionally restricted. 228Only `ZSTD_CLEVEL` is supported currently, for setting compression level. 229`ZSTD_CLEVEL` can be used to set the level between 1 and 19 (the "normal" range). 230If the value of `ZSTD_CLEVEL` is not a valid integer, it will be ignored with a warning message. 231`ZSTD_CLEVEL` just replaces the default compression level (`3`). 232It can be overridden by corresponding command line arguments. 233 234 235DICTIONARY BUILDER 236------------------ 237`zstd` offers _dictionary_ compression, 238which greatly improves efficiency on small files and messages. 239It's possible to train `zstd` with a set of samples, 240the result of which is saved into a file called a `dictionary`. 241Then during compression and decompression, reference the same dictionary, 242using command `-D dictionaryFileName`. 243Compression of small files similar to the sample set will be greatly improved. 244 245* `--train FILEs`: 246 Use FILEs as training set to create a dictionary. 247 The training set should contain a lot of small files (> 100), 248 and weight typically 100x the target dictionary size 249 (for example, 10 MB for a 100 KB dictionary). 250 251 Supports multithreading if `zstd` is compiled with threading support. 252 Additional parameters can be specified with `--train-fastcover`. 253 The legacy dictionary builder can be accessed with `--train-legacy`. 254 The cover dictionary builder can be accessed with `--train-cover`. 255 Equivalent to `--train-fastcover=d=8,steps=4`. 256* `-o file`: 257 Dictionary saved into `file` (default name: dictionary). 258* `--maxdict=#`: 259 Limit dictionary to specified size (default: 112640). 260* `-#`: 261 Use `#` compression level during training (optional). 262 Will generate statistics more tuned for selected compression level, 263 resulting in a _small_ compression ratio improvement for this level. 264* `-B#`: 265 Split input files in blocks of size # (default: no split) 266* `--dictID=#`: 267 A dictionary ID is a locally unique ID that a decoder can use to verify it is 268 using the right dictionary. 269 By default, zstd will create a 4-bytes random number ID. 270 It's possible to give a precise number instead. 271 Short numbers have an advantage : an ID < 256 will only need 1 byte in the 272 compressed frame header, and an ID < 65536 will only need 2 bytes. 273 This compares favorably to 4 bytes default. 274 However, it's up to the dictionary manager to not assign twice the same ID to 275 2 different dictionaries. 276* `--train-cover[=k#,d=#,steps=#,split=#,shrink[=#]]`: 277 Select parameters for the default dictionary builder algorithm named cover. 278 If _d_ is not specified, then it tries _d_ = 6 and _d_ = 8. 279 If _k_ is not specified, then it tries _steps_ values in the range [50, 2000]. 280 If _steps_ is not specified, then the default value of 40 is used. 281 If _split_ is not specified or split <= 0, then the default value of 100 is used. 282 Requires that _d_ <= _k_. 283 If _shrink_ flag is not used, then the default value for _shrinkDict_ of 0 is used. 284 If _shrink_ is not specified, then the default value for _shrinkDictMaxRegression_ of 1 is used. 285 286 Selects segments of size _k_ with highest score to put in the dictionary. 287 The score of a segment is computed by the sum of the frequencies of all the 288 subsegments of size _d_. 289 Generally _d_ should be in the range [6, 8], occasionally up to 16, but the 290 algorithm will run faster with d <= _8_. 291 Good values for _k_ vary widely based on the input data, but a safe range is 292 [2 * _d_, 2000]. 293 If _split_ is 100, all input samples are used for both training and testing 294 to find optimal _d_ and _k_ to build dictionary. 295 Supports multithreading if `zstd` is compiled with threading support. 296 Having _shrink_ enabled takes a truncated dictionary of minimum size and doubles 297 in size until compression ratio of the truncated dictionary is at most 298 _shrinkDictMaxRegression%_ worse than the compression ratio of the largest dictionary. 299 300 Examples: 301 302 `zstd --train-cover FILEs` 303 304 `zstd --train-cover=k=50,d=8 FILEs` 305 306 `zstd --train-cover=d=8,steps=500 FILEs` 307 308 `zstd --train-cover=k=50 FILEs` 309 310 `zstd --train-cover=k=50,split=60 FILEs` 311 312 `zstd --train-cover=shrink FILEs` 313 314 `zstd --train-cover=shrink=2 FILEs` 315 316* `--train-fastcover[=k#,d=#,f=#,steps=#,split=#,accel=#]`: 317 Same as cover but with extra parameters _f_ and _accel_ and different default value of split 318 If _split_ is not specified, then it tries _split_ = 75. 319 If _f_ is not specified, then it tries _f_ = 20. 320 Requires that 0 < _f_ < 32. 321 If _accel_ is not specified, then it tries _accel_ = 1. 322 Requires that 0 < _accel_ <= 10. 323 Requires that _d_ = 6 or _d_ = 8. 324 325 _f_ is log of size of array that keeps track of frequency of subsegments of size _d_. 326 The subsegment is hashed to an index in the range [0,2^_f_ - 1]. 327 It is possible that 2 different subsegments are hashed to the same index, and they are considered as the same subsegment when computing frequency. 328 Using a higher _f_ reduces collision but takes longer. 329 330 Examples: 331 332 `zstd --train-fastcover FILEs` 333 334 `zstd --train-fastcover=d=8,f=15,accel=2 FILEs` 335 336* `--train-legacy[=selectivity=#]`: 337 Use legacy dictionary builder algorithm with the given dictionary 338 _selectivity_ (default: 9). 339 The smaller the _selectivity_ value, the denser the dictionary, 340 improving its efficiency but reducing its possible maximum size. 341 `--train-legacy=s=#` is also accepted. 342 343 Examples: 344 345 `zstd --train-legacy FILEs` 346 347 `zstd --train-legacy=selectivity=8 FILEs` 348 349 350BENCHMARK 351--------- 352 353* `-b#`: 354 benchmark file(s) using compression level # 355* `-e#`: 356 benchmark file(s) using multiple compression levels, from `-b#` to `-e#` (inclusive) 357* `-i#`: 358 minimum evaluation time, in seconds (default: 3s), benchmark mode only 359* `-B#`, `--block-size=#`: 360 cut file(s) into independent blocks of size # (default: no block) 361* `--priority=rt`: 362 set process priority to real-time 363 364**Output Format:** CompressionLevel#Filename : IntputSize -> OutputSize (CompressionRatio), CompressionSpeed, DecompressionSpeed 365 366**Methodology:** For both compression and decompression speed, the entire input is compressed/decompressed in-memory to measure speed. A run lasts at least 1 sec, so when files are small, they are compressed/decompressed several times per run, in order to improve measurement accuracy. 367 368ADVANCED COMPRESSION OPTIONS 369---------------------------- 370### --zstd[=options]: 371`zstd` provides 22 predefined compression levels. 372The selected or default predefined compression level can be changed with 373advanced compression options. 374The _options_ are provided as a comma-separated list. 375You may specify only the options you want to change and the rest will be 376taken from the selected or default compression level. 377The list of available _options_: 378 379- `strategy`=_strat_, `strat`=_strat_: 380 Specify a strategy used by a match finder. 381 382 There are 9 strategies numbered from 1 to 9, from faster to stronger: 383 1=ZSTD\_fast, 2=ZSTD\_dfast, 3=ZSTD\_greedy, 384 4=ZSTD\_lazy, 5=ZSTD\_lazy2, 6=ZSTD\_btlazy2, 385 7=ZSTD\_btopt, 8=ZSTD\_btultra, 9=ZSTD\_btultra2. 386 387- `windowLog`=_wlog_, `wlog`=_wlog_: 388 Specify the maximum number of bits for a match distance. 389 390 The higher number of increases the chance to find a match which usually 391 improves compression ratio. 392 It also increases memory requirements for the compressor and decompressor. 393 The minimum _wlog_ is 10 (1 KiB) and the maximum is 30 (1 GiB) on 32-bit 394 platforms and 31 (2 GiB) on 64-bit platforms. 395 396 Note: If `windowLog` is set to larger than 27, `--long=windowLog` or 397 `--memory=windowSize` needs to be passed to the decompressor. 398 399- `hashLog`=_hlog_, `hlog`=_hlog_: 400 Specify the maximum number of bits for a hash table. 401 402 Bigger hash tables cause less collisions which usually makes compression 403 faster, but requires more memory during compression. 404 405 The minimum _hlog_ is 6 (64 B) and the maximum is 26 (128 MiB). 406 407- `chainLog`=_clog_, `clog`=_clog_: 408 Specify the maximum number of bits for a hash chain or a binary tree. 409 410 Higher numbers of bits increases the chance to find a match which usually 411 improves compression ratio. 412 It also slows down compression speed and increases memory requirements for 413 compression. 414 This option is ignored for the ZSTD_fast strategy. 415 416 The minimum _clog_ is 6 (64 B) and the maximum is 28 (256 MiB). 417 418- `searchLog`=_slog_, `slog`=_slog_: 419 Specify the maximum number of searches in a hash chain or a binary tree 420 using logarithmic scale. 421 422 More searches increases the chance to find a match which usually increases 423 compression ratio but decreases compression speed. 424 425 The minimum _slog_ is 1 and the maximum is 26. 426 427- `minMatch`=_mml_, `mml`=_mml_: 428 Specify the minimum searched length of a match in a hash table. 429 430 Larger search lengths usually decrease compression ratio but improve 431 decompression speed. 432 433 The minimum _mml_ is 3 and the maximum is 7. 434 435- `targetLen`=_tlen_, `tlen`=_tlen_: 436 The impact of this field vary depending on selected strategy. 437 438 For ZSTD\_btopt, ZSTD\_btultra and ZSTD\_btultra2, it specifies 439 the minimum match length that causes match finder to stop searching. 440 A larger `targetLen` usually improves compression ratio 441 but decreases compression speed. 442 443 For ZSTD\_fast, it triggers ultra-fast mode when > 0. 444 The value represents the amount of data skipped between match sampling. 445 Impact is reversed : a larger `targetLen` increases compression speed 446 but decreases compression ratio. 447 448 For all other strategies, this field has no impact. 449 450 The minimum _tlen_ is 0 and the maximum is 999. 451 452- `overlapLog`=_ovlog_, `ovlog`=_ovlog_: 453 Determine `overlapSize`, amount of data reloaded from previous job. 454 This parameter is only available when multithreading is enabled. 455 Reloading more data improves compression ratio, but decreases speed. 456 457 The minimum _ovlog_ is 0, and the maximum is 9. 458 1 means "no overlap", hence completely independent jobs. 459 9 means "full overlap", meaning up to `windowSize` is reloaded from previous job. 460 Reducing _ovlog_ by 1 reduces the reloaded amount by a factor 2. 461 For example, 8 means "windowSize/2", and 6 means "windowSize/8". 462 Value 0 is special and means "default" : _ovlog_ is automatically determined by `zstd`. 463 In which case, _ovlog_ will range from 6 to 9, depending on selected _strat_. 464 465- `ldmHashLog`=_lhlog_, `lhlog`=_lhlog_: 466 Specify the maximum size for a hash table used for long distance matching. 467 468 This option is ignored unless long distance matching is enabled. 469 470 Bigger hash tables usually improve compression ratio at the expense of more 471 memory during compression and a decrease in compression speed. 472 473 The minimum _lhlog_ is 6 and the maximum is 26 (default: 20). 474 475- `ldmMinMatch`=_lmml_, `lmml`=_lmml_: 476 Specify the minimum searched length of a match for long distance matching. 477 478 This option is ignored unless long distance matching is enabled. 479 480 Larger/very small values usually decrease compression ratio. 481 482 The minimum _lmml_ is 4 and the maximum is 4096 (default: 64). 483 484- `ldmBucketSizeLog`=_lblog_, `lblog`=_lblog_: 485 Specify the size of each bucket for the hash table used for long distance 486 matching. 487 488 This option is ignored unless long distance matching is enabled. 489 490 Larger bucket sizes improve collision resolution but decrease compression 491 speed. 492 493 The minimum _lblog_ is 0 and the maximum is 8 (default: 3). 494 495- `ldmHashRateLog`=_lhrlog_, `lhrlog`=_lhrlog_: 496 Specify the frequency of inserting entries into the long distance matching 497 hash table. 498 499 This option is ignored unless long distance matching is enabled. 500 501 Larger values will improve compression speed. Deviating far from the 502 default value will likely result in a decrease in compression ratio. 503 504 The default value is `wlog - lhlog`. 505 506### Example 507The following parameters sets advanced compression options to something 508similar to predefined level 19 for files bigger than 256 KB: 509 510`--zstd`=wlog=23,clog=23,hlog=22,slog=6,mml=3,tlen=48,strat=6 511 512### -B#: 513Select the size of each compression job. 514This parameter is available only when multi-threading is enabled. 515Default value is `4 * windowSize`, which means it varies depending on compression level. 516`-B#` makes it possible to select a custom value. 517Note that job size must respect a minimum value which is enforced transparently. 518This minimum is either 1 MB, or `overlapSize`, whichever is largest. 519 520BUGS 521---- 522Report bugs at: https://github.com/facebook/zstd/issues 523 524AUTHOR 525------ 526Yann Collet 527