xref: /freebsd/sys/contrib/zstd/programs/zstd.1.md (revision 52f72944b8f5abb2386eae924357dee8aea17d5b)
1zstd(1) -- zstd, zstdmt, unzstd, zstdcat - Compress or decompress .zst files
2============================================================================
3
4SYNOPSIS
5--------
6
7`zstd` [*OPTIONS*] [-|_INPUT-FILE_] [-o _OUTPUT-FILE_]
8
9`zstdmt` is equivalent to `zstd -T0`
10
11`unzstd` is equivalent to `zstd -d`
12
13`zstdcat` is equivalent to `zstd -dcf`
14
15
16DESCRIPTION
17-----------
18`zstd` is a fast lossless compression algorithm and data compression tool,
19with command line syntax similar to `gzip (1)` and `xz (1)`.
20It is based on the **LZ77** family, with further FSE & huff0 entropy stages.
21`zstd` offers highly configurable compression speed,
22with fast modes at > 200 MB/s per code,
23and strong modes nearing lzma compression ratios.
24It also features a very fast decoder, with speeds > 500 MB/s per core.
25
26`zstd` command line syntax is generally similar to gzip,
27but features the following differences :
28
29  - Source files are preserved by default.
30    It's possible to remove them automatically by using the `--rm` command.
31  - When compressing a single file, `zstd` displays progress notifications
32    and result summary by default.
33    Use `-q` to turn them off.
34  - `zstd` does not accept input from console,
35    but it properly accepts `stdin` when it's not the console.
36  - `zstd` displays a short help page when command line is an error.
37    Use `-q` to turn it off.
38
39`zstd` compresses or decompresses each _file_ according to the selected
40operation mode.
41If no _files_ are given or _file_ is `-`, `zstd` reads from standard input
42and writes the processed data to standard output.
43`zstd` will refuse to write compressed data to standard output
44if it is a terminal : it will display an error message and skip the _file_.
45Similarly, `zstd` will refuse to read compressed data from standard input
46if it is a terminal.
47
48Unless `--stdout` or `-o` is specified, _files_ are written to a new file
49whose name is derived from the source _file_ name:
50
51* When compressing, the suffix `.zst` is appended to the source filename to
52  get the target filename.
53* When decompressing, the `.zst` suffix is removed from the source filename to
54  get the target filename
55
56### Concatenation with .zst files
57It is possible to concatenate `.zst` files as is.
58`zstd` will decompress such files as if they were a single `.zst` file.
59
60OPTIONS
61-------
62
63### Integer suffixes and special values
64In most places where an integer argument is expected,
65an optional suffix is supported to easily indicate large integers.
66There must be no space between the integer and the suffix.
67
68* `KiB`:
69    Multiply the integer by 1,024 (2\^10).
70    `Ki`, `K`, and `KB` are accepted as synonyms for `KiB`.
71* `MiB`:
72    Multiply the integer by 1,048,576 (2\^20).
73    `Mi`, `M`, and `MB` are accepted as synonyms for `MiB`.
74
75### Operation mode
76If multiple operation mode options are given,
77the last one takes effect.
78
79* `-z`, `--compress`:
80    Compress.
81    This is the default operation mode when no operation mode option is specified
82    and no other operation mode is implied from the command name
83    (for example, `unzstd` implies `--decompress`).
84* `-d`, `--decompress`, `--uncompress`:
85    Decompress.
86* `-t`, `--test`:
87    Test the integrity of compressed _files_.
88    This option is equivalent to `--decompress --stdout` except that the
89    decompressed data is discarded instead of being written to standard output.
90    No files are created or removed.
91* `-b#`:
92    Benchmark file(s) using compression level #
93* `--train FILEs`:
94    Use FILEs as a training set to create a dictionary.
95    The training set should contain a lot of small files (> 100).
96* `-l`, `--list`:
97    Display information related to a zstd compressed file, such as size, ratio, and checksum.
98    Some of these fields may not be available.
99    This command can be augmented with the `-v` modifier.
100
101### Operation modifiers
102
103* `-#`:
104    `#` compression level \[1-19] (default: 3)
105* `--ultra`:
106    unlocks high compression levels 20+ (maximum 22), using a lot more memory.
107    Note that decompression will also require more memory when using these levels.
108* `--long[=#]`:
109    enables long distance matching with `#` `windowLog`, if not `#` is not
110    present it defaults to `27`.
111    This increases the window size (`windowLog`) and memory usage for both the
112    compressor and decompressor.
113    This setting is designed to improve the compression ratio for files with
114    long matches at a large distance.
115
116    Note: If `windowLog` is set to larger than 27, `--long=windowLog` or
117    `--memory=windowSize` needs to be passed to the decompressor.
118* `-T#`, `--threads=#`:
119    Compress using `#` threads (default: 1).
120    If `#` is 0, attempt to detect and use the number of physical CPU cores.
121    In all cases, the nb of threads is capped to ZSTDMT_NBTHREADS_MAX==256.
122    This modifier does nothing if `zstd` is compiled without multithread support.
123* `-D file`:
124    use `file` as Dictionary to compress or decompress FILE(s)
125* `--nodictID`:
126    do not store dictionary ID within frame header (dictionary compression).
127    The decoder will have to rely on implicit knowledge about which dictionary to use,
128    it won't be able to check if it's correct.
129* `-o file`:
130    save result into `file` (only possible with a single _INPUT-FILE_)
131* `-f`, `--force`:
132    overwrite output without prompting, and (de)compress symbolic links
133* `-c`, `--stdout`:
134    force write to standard output, even if it is the console
135* `--[no-]sparse`:
136    enable / disable sparse FS support,
137    to make files with many zeroes smaller on disk.
138    Creating sparse files may save disk space and speed up decompression by
139    reducing the amount of disk I/O.
140    default : enabled when output is into a file,
141    and disabled when output is stdout.
142    This setting overrides default and can force sparse mode over stdout.
143* `--rm`:
144    remove source file(s) after successful compression or decompression
145* `-k`, `--keep`:
146    keep source file(s) after successful compression or decompression.
147    This is the default behavior.
148* `-r`:
149    operate recursively on dictionaries
150* `--format=FORMAT`:
151    compress and decompress in other formats. If compiled with
152    support, zstd can compress to or decompress from other compression algorithm
153    formats. Possibly available options are `gzip`, `xz`, `lzma`, and `lz4`.
154* `-h`/`-H`, `--help`:
155    display help/long help and exit
156* `-V`, `--version`:
157    display version number and exit.
158    Advanced : `-vV` also displays supported formats.
159    `-vvV` also displays POSIX support.
160* `-v`:
161    verbose mode
162* `-q`, `--quiet`:
163    suppress warnings, interactivity, and notifications.
164    specify twice to suppress errors too.
165* `-C`, `--[no-]check`:
166    add integrity check computed from uncompressed data (default : enabled)
167* `--`:
168    All arguments after `--` are treated as files
169
170
171DICTIONARY BUILDER
172------------------
173`zstd` offers _dictionary_ compression,
174useful for very small files and messages.
175It's possible to train `zstd` with some samples,
176the result of which is saved into a file called a `dictionary`.
177Then during compression and decompression, reference the same dictionary.
178It will improve compression ratio of small files.
179Typical gains range from 10% (at 64KB) to x5 better (at <1KB).
180
181* `--train FILEs`:
182    Use FILEs as training set to create a dictionary.
183    The training set should contain a lot of small files (> 100),
184    and weight typically 100x the target dictionary size
185    (for example, 10 MB for a 100 KB dictionary).
186
187    Supports multithreading if `zstd` is compiled with threading support.
188    Additional parameters can be specified with `--train-cover`.
189    The legacy dictionary builder can be accessed with `--train-legacy`.
190    Equivalent to `--train-cover=d=8,steps=4`.
191* `-o file`:
192    Dictionary saved into `file` (default name: dictionary).
193* `--maxdict=#`:
194    Limit dictionary to specified size (default: 112640).
195* `-B#`:
196    Split input files in blocks of size # (default: no split)
197* `--dictID=#`:
198    A dictionary ID is a locally unique ID that a decoder can use to verify it is
199    using the right dictionary.
200    By default, zstd will create a 4-bytes random number ID.
201    It's possible to give a precise number instead.
202    Short numbers have an advantage : an ID < 256 will only need 1 byte in the
203    compressed frame header, and an ID < 65536 will only need 2 bytes.
204    This compares favorably to 4 bytes default.
205    However, it's up to the dictionary manager to not assign twice the same ID to
206    2 different dictionaries.
207* `--train-cover[=k#,d=#,steps=#]`:
208    Select parameters for the default dictionary builder algorithm named cover.
209    If _d_ is not specified, then it tries _d_ = 6 and _d_ = 8.
210    If _k_ is not specified, then it tries _steps_ values in the range [50, 2000].
211    If _steps_ is not specified, then the default value of 40 is used.
212    Requires that _d_ <= _k_.
213
214    Selects segments of size _k_ with highest score to put in the dictionary.
215    The score of a segment is computed by the sum of the frequencies of all the
216    subsegments of size _d_.
217    Generally _d_ should be in the range [6, 8], occasionally up to 16, but the
218    algorithm will run faster with d <= _8_.
219    Good values for _k_ vary widely based on the input data, but a safe range is
220    [2 * _d_, 2000].
221    Supports multithreading if `zstd` is compiled with threading support.
222
223    Examples:
224
225    `zstd --train-cover FILEs`
226
227    `zstd --train-cover=k=50,d=8 FILEs`
228
229    `zstd --train-cover=d=8,steps=500 FILEs`
230
231    `zstd --train-cover=k=50 FILEs`
232
233* `--train-legacy[=selectivity=#]`:
234    Use legacy dictionary builder algorithm with the given dictionary
235    _selectivity_ (default: 9).
236    The smaller the _selectivity_ value, the denser the dictionary,
237    improving its efficiency but reducing its possible maximum size.
238    `--train-legacy=s=#` is also accepted.
239
240    Examples:
241
242    `zstd --train-legacy FILEs`
243
244    `zstd --train-legacy=selectivity=8 FILEs`
245
246
247BENCHMARK
248---------
249
250* `-b#`:
251    benchmark file(s) using compression level #
252* `-e#`:
253    benchmark file(s) using multiple compression levels, from `-b#` to `-e#` (inclusive)
254* `-i#`:
255    minimum evaluation time, in seconds (default : 3s), benchmark mode only
256* `-B#`, `--block-size=#`:
257    cut file(s) into independent blocks of size # (default: no block)
258* `--priority=rt`:
259    set process priority to real-time
260
261**Output Format:** CompressionLevel#Filename : IntputSize -> OutputSize (CompressionRatio), CompressionSpeed, DecompressionSpeed
262
263**Methodology:** For both compression and decompression speed, the entire input is compressed/decompressed in-memory to measure speed. A run lasts at least 1 sec, so when files are small, they are compressed/decompressed several times per run, in order to improve measurement accuracy.
264
265ADVANCED COMPRESSION OPTIONS
266----------------------------
267### --zstd[=options]:
268`zstd` provides 22 predefined compression levels.
269The selected or default predefined compression level can be changed with
270advanced compression options.
271The _options_ are provided as a comma-separated list.
272You may specify only the options you want to change and the rest will be
273taken from the selected or default compression level.
274The list of available _options_:
275
276- `strategy`=_strat_, `strat`=_strat_:
277    Specify a strategy used by a match finder.
278
279    There are 8 strategies numbered from 1 to 8, from faster to stronger:
280    1=ZSTD\_fast, 2=ZSTD\_dfast, 3=ZSTD\_greedy, 4=ZSTD\_lazy,
281    5=ZSTD\_lazy2, 6=ZSTD\_btlazy2, 7=ZSTD\_btopt, 8=ZSTD\_btultra.
282
283- `windowLog`=_wlog_, `wlog`=_wlog_:
284    Specify the maximum number of bits for a match distance.
285
286    The higher number of increases the chance to find a match which usually
287    improves compression ratio.
288    It also increases memory requirements for the compressor and decompressor.
289    The minimum _wlog_ is 10 (1 KiB) and the maximum is 30 (1 GiB) on 32-bit
290    platforms and 31 (2 GiB) on 64-bit platforms.
291
292    Note: If `windowLog` is set to larger than 27, `--long=windowLog` or
293    `--memory=windowSize` needs to be passed to the decompressor.
294
295- `hashLog`=_hlog_, `hlog`=_hlog_:
296    Specify the maximum number of bits for a hash table.
297
298    Bigger hash tables cause less collisions which usually makes compression
299    faster, but requires more memory during compression.
300
301    The minimum _hlog_ is 6 (64 B) and the maximum is 26 (128 MiB).
302
303- `chainLog`=_clog_, `clog`=_clog_:
304    Specify the maximum number of bits for a hash chain or a binary tree.
305
306    Higher numbers of bits increases the chance to find a match which usually
307    improves compression ratio.
308    It also slows down compression speed and increases memory requirements for
309    compression.
310    This option is ignored for the ZSTD_fast strategy.
311
312    The minimum _clog_ is 6 (64 B) and the maximum is 28 (256 MiB).
313
314- `searchLog`=_slog_, `slog`=_slog_:
315    Specify the maximum number of searches in a hash chain or a binary tree
316    using logarithmic scale.
317
318    More searches increases the chance to find a match which usually increases
319    compression ratio but decreases compression speed.
320
321    The minimum _slog_ is 1 and the maximum is 26.
322
323- `searchLength`=_slen_, `slen`=_slen_:
324    Specify the minimum searched length of a match in a hash table.
325
326    Larger search lengths usually decrease compression ratio but improve
327    decompression speed.
328
329    The minimum _slen_ is 3 and the maximum is 7.
330
331- `targetLen`=_tlen_, `tlen`=_tlen_:
332    Specify the minimum match length that causes a match finder to stop
333    searching for better matches.
334
335    A larger minimum match length usually improves compression ratio but
336    decreases compression speed.
337    This option is only used with strategies ZSTD_btopt and ZSTD_btultra.
338
339    The minimum _tlen_ is 4 and the maximum is 999.
340
341- `overlapLog`=_ovlog_,  `ovlog`=_ovlog_:
342    Determine `overlapSize`, amount of data reloaded from previous job.
343    This parameter is only available when multithreading is enabled.
344    Reloading more data improves compression ratio, but decreases speed.
345
346    The minimum _ovlog_ is 0, and the maximum is 9.
347    0 means "no overlap", hence completely independent jobs.
348    9 means "full overlap", meaning up to `windowSize` is reloaded from previous job.
349    Reducing _ovlog_ by 1 reduces the amount of reload by a factor 2.
350    Default _ovlog_ is 6, which means "reload `windowSize / 8`".
351    Exception : the maximum compression level (22) has a default _ovlog_ of 9.
352
353- `ldmHashLog`=_ldmhlog_, `ldmhlog`=_ldmhlog_:
354    Specify the maximum size for a hash table used for long distance matching.
355
356    This option is ignored unless long distance matching is enabled.
357
358    Bigger hash tables usually improve compression ratio at the expense of more
359    memory during compression and a decrease in compression speed.
360
361    The minimum _ldmhlog_ is 6 and the maximum is 26 (default: 20).
362
363- `ldmSearchLength`=_ldmslen_, `ldmslen`=_ldmslen_:
364    Specify the minimum searched length of a match for long distance matching.
365
366    This option is ignored unless long distance matching is enabled.
367
368    Larger/very small values usually decrease compression ratio.
369
370    The minumum _ldmslen_ is 4 and the maximum is 4096 (default: 64).
371
372- `ldmBucketSizeLog`=_ldmblog_, `ldmblog`=_ldmblog_:
373    Specify the size of each bucket for the hash table used for long distance
374    matching.
375
376    This option is ignored unless long distance matching is enabled.
377
378    Larger bucket sizes improve collision resolution but decrease compression
379    speed.
380
381    The minimum _ldmblog_ is 0 and the maximum is 8 (default: 3).
382
383- `ldmHashEveryLog`=_ldmhevery_, `ldmhevery`=_ldmhevery_:
384    Specify the frequency of inserting entries into the long distance matching
385    hash table.
386
387    This option is ignored unless long distance matching is enabled.
388
389    Larger values will improve compression speed. Deviating far from the
390    default value will likely result in a decrease in compression ratio.
391
392    The default value is `wlog - ldmhlog`.
393
394### -B#:
395Select the size of each compression job.
396This parameter is available only when multi-threading is enabled.
397Default value is `4 * windowSize`, which means it varies depending on compression level.
398`-B#` makes it possible to select a custom value.
399Note that job size must respect a minimum value which is enforced transparently.
400This minimum is either 1 MB, or `overlapSize`, whichever is largest.
401
402### Example
403The following parameters sets advanced compression options to those of
404predefined level 19 for files bigger than 256 KB:
405
406`--zstd`=windowLog=23,chainLog=23,hashLog=22,searchLog=6,searchLength=3,targetLength=48,strategy=6
407
408BUGS
409----
410Report bugs at: https://github.com/facebook/zstd/issues
411
412AUTHOR
413------
414Yann Collet
415