1 /*
2 * Copyright (c) Facebook, Inc.
3 * All rights reserved.
4 *
5 * This source code is licensed under both the BSD-style license (found in the
6 * LICENSE file in the root directory of this source tree) and the GPLv2 (found
7 * in the COPYING file in the root directory of this source tree).
8 * You may select, at your option, one of the above-listed licenses.
9 */
10
11 /*-*************************************
12 * Dependencies
13 ***************************************/
14 #include <stdio.h> /* fprintf */
15 #include <stdlib.h> /* malloc, free, qsort */
16 #include <string.h> /* memset */
17 #include <time.h> /* clock */
18
19 #ifndef ZDICT_STATIC_LINKING_ONLY
20 # define ZDICT_STATIC_LINKING_ONLY
21 #endif
22
23 #include "../common/mem.h" /* read */
24 #include "../common/pool.h"
25 #include "../common/threading.h"
26 #include "../common/zstd_internal.h" /* includes zstd.h */
27 #include "../compress/zstd_compress_internal.h" /* ZSTD_hash*() */
28 #include "../zdict.h"
29 #include "cover.h"
30
31
32 /*-*************************************
33 * Constants
34 ***************************************/
35 /**
36 * There are 32bit indexes used to ref samples, so limit samples size to 4GB
37 * on 64bit builds.
38 * For 32bit builds we choose 1 GB.
39 * Most 32bit platforms have 2GB user-mode addressable space and we allocate a large
40 * contiguous buffer, so 1GB is already a high limit.
41 */
42 #define FASTCOVER_MAX_SAMPLES_SIZE (sizeof(size_t) == 8 ? ((unsigned)-1) : ((unsigned)1 GB))
43 #define FASTCOVER_MAX_F 31
44 #define FASTCOVER_MAX_ACCEL 10
45 #define FASTCOVER_DEFAULT_SPLITPOINT 0.75
46 #define DEFAULT_F 20
47 #define DEFAULT_ACCEL 1
48
49
50 /*-*************************************
51 * Console display
52 ***************************************/
53 #ifndef LOCALDISPLAYLEVEL
54 static int g_displayLevel = 0;
55 #endif
56 #undef DISPLAY
57 #define DISPLAY(...) \
58 { \
59 fprintf(stderr, __VA_ARGS__); \
60 fflush(stderr); \
61 }
62 #undef LOCALDISPLAYLEVEL
63 #define LOCALDISPLAYLEVEL(displayLevel, l, ...) \
64 if (displayLevel >= l) { \
65 DISPLAY(__VA_ARGS__); \
66 } /* 0 : no display; 1: errors; 2: default; 3: details; 4: debug */
67 #undef DISPLAYLEVEL
68 #define DISPLAYLEVEL(l, ...) LOCALDISPLAYLEVEL(g_displayLevel, l, __VA_ARGS__)
69
70 #ifndef LOCALDISPLAYUPDATE
71 static const clock_t g_refreshRate = CLOCKS_PER_SEC * 15 / 100;
72 static clock_t g_time = 0;
73 #endif
74 #undef LOCALDISPLAYUPDATE
75 #define LOCALDISPLAYUPDATE(displayLevel, l, ...) \
76 if (displayLevel >= l) { \
77 if ((clock() - g_time > g_refreshRate) || (displayLevel >= 4)) { \
78 g_time = clock(); \
79 DISPLAY(__VA_ARGS__); \
80 } \
81 }
82 #undef DISPLAYUPDATE
83 #define DISPLAYUPDATE(l, ...) LOCALDISPLAYUPDATE(g_displayLevel, l, __VA_ARGS__)
84
85
86 /*-*************************************
87 * Hash Functions
88 ***************************************/
89 /**
90 * Hash the d-byte value pointed to by p and mod 2^f into the frequency vector
91 */
FASTCOVER_hashPtrToIndex(const void * p,U32 f,unsigned d)92 static size_t FASTCOVER_hashPtrToIndex(const void* p, U32 f, unsigned d) {
93 if (d == 6) {
94 return ZSTD_hash6Ptr(p, f);
95 }
96 return ZSTD_hash8Ptr(p, f);
97 }
98
99
100 /*-*************************************
101 * Acceleration
102 ***************************************/
103 typedef struct {
104 unsigned finalize; /* Percentage of training samples used for ZDICT_finalizeDictionary */
105 unsigned skip; /* Number of dmer skipped between each dmer counted in computeFrequency */
106 } FASTCOVER_accel_t;
107
108
109 static const FASTCOVER_accel_t FASTCOVER_defaultAccelParameters[FASTCOVER_MAX_ACCEL+1] = {
110 { 100, 0 }, /* accel = 0, should not happen because accel = 0 defaults to accel = 1 */
111 { 100, 0 }, /* accel = 1 */
112 { 50, 1 }, /* accel = 2 */
113 { 34, 2 }, /* accel = 3 */
114 { 25, 3 }, /* accel = 4 */
115 { 20, 4 }, /* accel = 5 */
116 { 17, 5 }, /* accel = 6 */
117 { 14, 6 }, /* accel = 7 */
118 { 13, 7 }, /* accel = 8 */
119 { 11, 8 }, /* accel = 9 */
120 { 10, 9 }, /* accel = 10 */
121 };
122
123
124 /*-*************************************
125 * Context
126 ***************************************/
127 typedef struct {
128 const BYTE *samples;
129 size_t *offsets;
130 const size_t *samplesSizes;
131 size_t nbSamples;
132 size_t nbTrainSamples;
133 size_t nbTestSamples;
134 size_t nbDmers;
135 U32 *freqs;
136 unsigned d;
137 unsigned f;
138 FASTCOVER_accel_t accelParams;
139 } FASTCOVER_ctx_t;
140
141
142 /*-*************************************
143 * Helper functions
144 ***************************************/
145 /**
146 * Selects the best segment in an epoch.
147 * Segments of are scored according to the function:
148 *
149 * Let F(d) be the frequency of all dmers with hash value d.
150 * Let S_i be hash value of the dmer at position i of segment S which has length k.
151 *
152 * Score(S) = F(S_1) + F(S_2) + ... + F(S_{k-d+1})
153 *
154 * Once the dmer with hash value d is in the dictionary we set F(d) = 0.
155 */
FASTCOVER_selectSegment(const FASTCOVER_ctx_t * ctx,U32 * freqs,U32 begin,U32 end,ZDICT_cover_params_t parameters,U16 * segmentFreqs)156 static COVER_segment_t FASTCOVER_selectSegment(const FASTCOVER_ctx_t *ctx,
157 U32 *freqs, U32 begin, U32 end,
158 ZDICT_cover_params_t parameters,
159 U16* segmentFreqs) {
160 /* Constants */
161 const U32 k = parameters.k;
162 const U32 d = parameters.d;
163 const U32 f = ctx->f;
164 const U32 dmersInK = k - d + 1;
165
166 /* Try each segment (activeSegment) and save the best (bestSegment) */
167 COVER_segment_t bestSegment = {0, 0, 0};
168 COVER_segment_t activeSegment;
169
170 /* Reset the activeDmers in the segment */
171 /* The activeSegment starts at the beginning of the epoch. */
172 activeSegment.begin = begin;
173 activeSegment.end = begin;
174 activeSegment.score = 0;
175
176 /* Slide the activeSegment through the whole epoch.
177 * Save the best segment in bestSegment.
178 */
179 while (activeSegment.end < end) {
180 /* Get hash value of current dmer */
181 const size_t idx = FASTCOVER_hashPtrToIndex(ctx->samples + activeSegment.end, f, d);
182
183 /* Add frequency of this index to score if this is the first occurrence of index in active segment */
184 if (segmentFreqs[idx] == 0) {
185 activeSegment.score += freqs[idx];
186 }
187 /* Increment end of segment and segmentFreqs*/
188 activeSegment.end += 1;
189 segmentFreqs[idx] += 1;
190 /* If the window is now too large, drop the first position */
191 if (activeSegment.end - activeSegment.begin == dmersInK + 1) {
192 /* Get hash value of the dmer to be eliminated from active segment */
193 const size_t delIndex = FASTCOVER_hashPtrToIndex(ctx->samples + activeSegment.begin, f, d);
194 segmentFreqs[delIndex] -= 1;
195 /* Subtract frequency of this index from score if this is the last occurrence of this index in active segment */
196 if (segmentFreqs[delIndex] == 0) {
197 activeSegment.score -= freqs[delIndex];
198 }
199 /* Increment start of segment */
200 activeSegment.begin += 1;
201 }
202
203 /* If this segment is the best so far save it */
204 if (activeSegment.score > bestSegment.score) {
205 bestSegment = activeSegment;
206 }
207 }
208
209 /* Zero out rest of segmentFreqs array */
210 while (activeSegment.begin < end) {
211 const size_t delIndex = FASTCOVER_hashPtrToIndex(ctx->samples + activeSegment.begin, f, d);
212 segmentFreqs[delIndex] -= 1;
213 activeSegment.begin += 1;
214 }
215
216 {
217 /* Zero the frequency of hash value of each dmer covered by the chosen segment. */
218 U32 pos;
219 for (pos = bestSegment.begin; pos != bestSegment.end; ++pos) {
220 const size_t i = FASTCOVER_hashPtrToIndex(ctx->samples + pos, f, d);
221 freqs[i] = 0;
222 }
223 }
224
225 return bestSegment;
226 }
227
228
FASTCOVER_checkParameters(ZDICT_cover_params_t parameters,size_t maxDictSize,unsigned f,unsigned accel)229 static int FASTCOVER_checkParameters(ZDICT_cover_params_t parameters,
230 size_t maxDictSize, unsigned f,
231 unsigned accel) {
232 /* k, d, and f are required parameters */
233 if (parameters.d == 0 || parameters.k == 0) {
234 return 0;
235 }
236 /* d has to be 6 or 8 */
237 if (parameters.d != 6 && parameters.d != 8) {
238 return 0;
239 }
240 /* k <= maxDictSize */
241 if (parameters.k > maxDictSize) {
242 return 0;
243 }
244 /* d <= k */
245 if (parameters.d > parameters.k) {
246 return 0;
247 }
248 /* 0 < f <= FASTCOVER_MAX_F*/
249 if (f > FASTCOVER_MAX_F || f == 0) {
250 return 0;
251 }
252 /* 0 < splitPoint <= 1 */
253 if (parameters.splitPoint <= 0 || parameters.splitPoint > 1) {
254 return 0;
255 }
256 /* 0 < accel <= 10 */
257 if (accel > 10 || accel == 0) {
258 return 0;
259 }
260 return 1;
261 }
262
263
264 /**
265 * Clean up a context initialized with `FASTCOVER_ctx_init()`.
266 */
267 static void
FASTCOVER_ctx_destroy(FASTCOVER_ctx_t * ctx)268 FASTCOVER_ctx_destroy(FASTCOVER_ctx_t* ctx)
269 {
270 if (!ctx) return;
271
272 free(ctx->freqs);
273 ctx->freqs = NULL;
274
275 free(ctx->offsets);
276 ctx->offsets = NULL;
277 }
278
279
280 /**
281 * Calculate for frequency of hash value of each dmer in ctx->samples
282 */
283 static void
FASTCOVER_computeFrequency(U32 * freqs,const FASTCOVER_ctx_t * ctx)284 FASTCOVER_computeFrequency(U32* freqs, const FASTCOVER_ctx_t* ctx)
285 {
286 const unsigned f = ctx->f;
287 const unsigned d = ctx->d;
288 const unsigned skip = ctx->accelParams.skip;
289 const unsigned readLength = MAX(d, 8);
290 size_t i;
291 assert(ctx->nbTrainSamples >= 5);
292 assert(ctx->nbTrainSamples <= ctx->nbSamples);
293 for (i = 0; i < ctx->nbTrainSamples; i++) {
294 size_t start = ctx->offsets[i]; /* start of current dmer */
295 size_t const currSampleEnd = ctx->offsets[i+1];
296 while (start + readLength <= currSampleEnd) {
297 const size_t dmerIndex = FASTCOVER_hashPtrToIndex(ctx->samples + start, f, d);
298 freqs[dmerIndex]++;
299 start = start + skip + 1;
300 }
301 }
302 }
303
304
305 /**
306 * Prepare a context for dictionary building.
307 * The context is only dependent on the parameter `d` and can used multiple
308 * times.
309 * Returns 0 on success or error code on error.
310 * The context must be destroyed with `FASTCOVER_ctx_destroy()`.
311 */
312 static size_t
FASTCOVER_ctx_init(FASTCOVER_ctx_t * ctx,const void * samplesBuffer,const size_t * samplesSizes,unsigned nbSamples,unsigned d,double splitPoint,unsigned f,FASTCOVER_accel_t accelParams)313 FASTCOVER_ctx_init(FASTCOVER_ctx_t* ctx,
314 const void* samplesBuffer,
315 const size_t* samplesSizes, unsigned nbSamples,
316 unsigned d, double splitPoint, unsigned f,
317 FASTCOVER_accel_t accelParams)
318 {
319 const BYTE* const samples = (const BYTE*)samplesBuffer;
320 const size_t totalSamplesSize = COVER_sum(samplesSizes, nbSamples);
321 /* Split samples into testing and training sets */
322 const unsigned nbTrainSamples = splitPoint < 1.0 ? (unsigned)((double)nbSamples * splitPoint) : nbSamples;
323 const unsigned nbTestSamples = splitPoint < 1.0 ? nbSamples - nbTrainSamples : nbSamples;
324 const size_t trainingSamplesSize = splitPoint < 1.0 ? COVER_sum(samplesSizes, nbTrainSamples) : totalSamplesSize;
325 const size_t testSamplesSize = splitPoint < 1.0 ? COVER_sum(samplesSizes + nbTrainSamples, nbTestSamples) : totalSamplesSize;
326
327 /* Checks */
328 if (totalSamplesSize < MAX(d, sizeof(U64)) ||
329 totalSamplesSize >= (size_t)FASTCOVER_MAX_SAMPLES_SIZE) {
330 DISPLAYLEVEL(1, "Total samples size is too large (%u MB), maximum size is %u MB\n",
331 (unsigned)(totalSamplesSize >> 20), (FASTCOVER_MAX_SAMPLES_SIZE >> 20));
332 return ERROR(srcSize_wrong);
333 }
334
335 /* Check if there are at least 5 training samples */
336 if (nbTrainSamples < 5) {
337 DISPLAYLEVEL(1, "Total number of training samples is %u and is invalid\n", nbTrainSamples);
338 return ERROR(srcSize_wrong);
339 }
340
341 /* Check if there's testing sample */
342 if (nbTestSamples < 1) {
343 DISPLAYLEVEL(1, "Total number of testing samples is %u and is invalid.\n", nbTestSamples);
344 return ERROR(srcSize_wrong);
345 }
346
347 /* Zero the context */
348 memset(ctx, 0, sizeof(*ctx));
349 DISPLAYLEVEL(2, "Training on %u samples of total size %u\n", nbTrainSamples,
350 (unsigned)trainingSamplesSize);
351 DISPLAYLEVEL(2, "Testing on %u samples of total size %u\n", nbTestSamples,
352 (unsigned)testSamplesSize);
353
354 ctx->samples = samples;
355 ctx->samplesSizes = samplesSizes;
356 ctx->nbSamples = nbSamples;
357 ctx->nbTrainSamples = nbTrainSamples;
358 ctx->nbTestSamples = nbTestSamples;
359 ctx->nbDmers = trainingSamplesSize - MAX(d, sizeof(U64)) + 1;
360 ctx->d = d;
361 ctx->f = f;
362 ctx->accelParams = accelParams;
363
364 /* The offsets of each file */
365 ctx->offsets = (size_t*)calloc((nbSamples + 1), sizeof(size_t));
366 if (ctx->offsets == NULL) {
367 DISPLAYLEVEL(1, "Failed to allocate scratch buffers \n");
368 FASTCOVER_ctx_destroy(ctx);
369 return ERROR(memory_allocation);
370 }
371
372 /* Fill offsets from the samplesSizes */
373 { U32 i;
374 ctx->offsets[0] = 0;
375 assert(nbSamples >= 5);
376 for (i = 1; i <= nbSamples; ++i) {
377 ctx->offsets[i] = ctx->offsets[i - 1] + samplesSizes[i - 1];
378 }
379 }
380
381 /* Initialize frequency array of size 2^f */
382 ctx->freqs = (U32*)calloc(((U64)1 << f), sizeof(U32));
383 if (ctx->freqs == NULL) {
384 DISPLAYLEVEL(1, "Failed to allocate frequency table \n");
385 FASTCOVER_ctx_destroy(ctx);
386 return ERROR(memory_allocation);
387 }
388
389 DISPLAYLEVEL(2, "Computing frequencies\n");
390 FASTCOVER_computeFrequency(ctx->freqs, ctx);
391
392 return 0;
393 }
394
395
396 /**
397 * Given the prepared context build the dictionary.
398 */
399 static size_t
FASTCOVER_buildDictionary(const FASTCOVER_ctx_t * ctx,U32 * freqs,void * dictBuffer,size_t dictBufferCapacity,ZDICT_cover_params_t parameters,U16 * segmentFreqs)400 FASTCOVER_buildDictionary(const FASTCOVER_ctx_t* ctx,
401 U32* freqs,
402 void* dictBuffer, size_t dictBufferCapacity,
403 ZDICT_cover_params_t parameters,
404 U16* segmentFreqs)
405 {
406 BYTE *const dict = (BYTE *)dictBuffer;
407 size_t tail = dictBufferCapacity;
408 /* Divide the data into epochs. We will select one segment from each epoch. */
409 const COVER_epoch_info_t epochs = COVER_computeEpochs(
410 (U32)dictBufferCapacity, (U32)ctx->nbDmers, parameters.k, 1);
411 const size_t maxZeroScoreRun = 10;
412 size_t zeroScoreRun = 0;
413 size_t epoch;
414 DISPLAYLEVEL(2, "Breaking content into %u epochs of size %u\n",
415 (U32)epochs.num, (U32)epochs.size);
416 /* Loop through the epochs until there are no more segments or the dictionary
417 * is full.
418 */
419 for (epoch = 0; tail > 0; epoch = (epoch + 1) % epochs.num) {
420 const U32 epochBegin = (U32)(epoch * epochs.size);
421 const U32 epochEnd = epochBegin + epochs.size;
422 size_t segmentSize;
423 /* Select a segment */
424 COVER_segment_t segment = FASTCOVER_selectSegment(
425 ctx, freqs, epochBegin, epochEnd, parameters, segmentFreqs);
426
427 /* If the segment covers no dmers, then we are out of content.
428 * There may be new content in other epochs, for continue for some time.
429 */
430 if (segment.score == 0) {
431 if (++zeroScoreRun >= maxZeroScoreRun) {
432 break;
433 }
434 continue;
435 }
436 zeroScoreRun = 0;
437
438 /* Trim the segment if necessary and if it is too small then we are done */
439 segmentSize = MIN(segment.end - segment.begin + parameters.d - 1, tail);
440 if (segmentSize < parameters.d) {
441 break;
442 }
443
444 /* We fill the dictionary from the back to allow the best segments to be
445 * referenced with the smallest offsets.
446 */
447 tail -= segmentSize;
448 memcpy(dict + tail, ctx->samples + segment.begin, segmentSize);
449 DISPLAYUPDATE(
450 2, "\r%u%% ",
451 (unsigned)(((dictBufferCapacity - tail) * 100) / dictBufferCapacity));
452 }
453 DISPLAYLEVEL(2, "\r%79s\r", "");
454 return tail;
455 }
456
457 /**
458 * Parameters for FASTCOVER_tryParameters().
459 */
460 typedef struct FASTCOVER_tryParameters_data_s {
461 const FASTCOVER_ctx_t* ctx;
462 COVER_best_t* best;
463 size_t dictBufferCapacity;
464 ZDICT_cover_params_t parameters;
465 } FASTCOVER_tryParameters_data_t;
466
467
468 /**
469 * Tries a set of parameters and updates the COVER_best_t with the results.
470 * This function is thread safe if zstd is compiled with multithreaded support.
471 * It takes its parameters as an *OWNING* opaque pointer to support threading.
472 */
FASTCOVER_tryParameters(void * opaque)473 static void FASTCOVER_tryParameters(void* opaque)
474 {
475 /* Save parameters as local variables */
476 FASTCOVER_tryParameters_data_t *const data = (FASTCOVER_tryParameters_data_t*)opaque;
477 const FASTCOVER_ctx_t *const ctx = data->ctx;
478 const ZDICT_cover_params_t parameters = data->parameters;
479 size_t dictBufferCapacity = data->dictBufferCapacity;
480 size_t totalCompressedSize = ERROR(GENERIC);
481 /* Initialize array to keep track of frequency of dmer within activeSegment */
482 U16* segmentFreqs = (U16*)calloc(((U64)1 << ctx->f), sizeof(U16));
483 /* Allocate space for hash table, dict, and freqs */
484 BYTE *const dict = (BYTE*)malloc(dictBufferCapacity);
485 COVER_dictSelection_t selection = COVER_dictSelectionError(ERROR(GENERIC));
486 U32* freqs = (U32*) malloc(((U64)1 << ctx->f) * sizeof(U32));
487 if (!segmentFreqs || !dict || !freqs) {
488 DISPLAYLEVEL(1, "Failed to allocate buffers: out of memory\n");
489 goto _cleanup;
490 }
491 /* Copy the frequencies because we need to modify them */
492 memcpy(freqs, ctx->freqs, ((U64)1 << ctx->f) * sizeof(U32));
493 /* Build the dictionary */
494 { const size_t tail = FASTCOVER_buildDictionary(ctx, freqs, dict, dictBufferCapacity,
495 parameters, segmentFreqs);
496
497 const unsigned nbFinalizeSamples = (unsigned)(ctx->nbTrainSamples * ctx->accelParams.finalize / 100);
498 selection = COVER_selectDict(dict + tail, dictBufferCapacity, dictBufferCapacity - tail,
499 ctx->samples, ctx->samplesSizes, nbFinalizeSamples, ctx->nbTrainSamples, ctx->nbSamples, parameters, ctx->offsets,
500 totalCompressedSize);
501
502 if (COVER_dictSelectionIsError(selection)) {
503 DISPLAYLEVEL(1, "Failed to select dictionary\n");
504 goto _cleanup;
505 }
506 }
507 _cleanup:
508 free(dict);
509 COVER_best_finish(data->best, parameters, selection);
510 free(data);
511 free(segmentFreqs);
512 COVER_dictSelectionFree(selection);
513 free(freqs);
514 }
515
516
517 static void
FASTCOVER_convertToCoverParams(ZDICT_fastCover_params_t fastCoverParams,ZDICT_cover_params_t * coverParams)518 FASTCOVER_convertToCoverParams(ZDICT_fastCover_params_t fastCoverParams,
519 ZDICT_cover_params_t* coverParams)
520 {
521 coverParams->k = fastCoverParams.k;
522 coverParams->d = fastCoverParams.d;
523 coverParams->steps = fastCoverParams.steps;
524 coverParams->nbThreads = fastCoverParams.nbThreads;
525 coverParams->splitPoint = fastCoverParams.splitPoint;
526 coverParams->zParams = fastCoverParams.zParams;
527 coverParams->shrinkDict = fastCoverParams.shrinkDict;
528 }
529
530
531 static void
FASTCOVER_convertToFastCoverParams(ZDICT_cover_params_t coverParams,ZDICT_fastCover_params_t * fastCoverParams,unsigned f,unsigned accel)532 FASTCOVER_convertToFastCoverParams(ZDICT_cover_params_t coverParams,
533 ZDICT_fastCover_params_t* fastCoverParams,
534 unsigned f, unsigned accel)
535 {
536 fastCoverParams->k = coverParams.k;
537 fastCoverParams->d = coverParams.d;
538 fastCoverParams->steps = coverParams.steps;
539 fastCoverParams->nbThreads = coverParams.nbThreads;
540 fastCoverParams->splitPoint = coverParams.splitPoint;
541 fastCoverParams->f = f;
542 fastCoverParams->accel = accel;
543 fastCoverParams->zParams = coverParams.zParams;
544 fastCoverParams->shrinkDict = coverParams.shrinkDict;
545 }
546
547
548 ZDICTLIB_API size_t
ZDICT_trainFromBuffer_fastCover(void * dictBuffer,size_t dictBufferCapacity,const void * samplesBuffer,const size_t * samplesSizes,unsigned nbSamples,ZDICT_fastCover_params_t parameters)549 ZDICT_trainFromBuffer_fastCover(void* dictBuffer, size_t dictBufferCapacity,
550 const void* samplesBuffer,
551 const size_t* samplesSizes, unsigned nbSamples,
552 ZDICT_fastCover_params_t parameters)
553 {
554 BYTE* const dict = (BYTE*)dictBuffer;
555 FASTCOVER_ctx_t ctx;
556 ZDICT_cover_params_t coverParams;
557 FASTCOVER_accel_t accelParams;
558 /* Initialize global data */
559 g_displayLevel = (int)parameters.zParams.notificationLevel;
560 /* Assign splitPoint and f if not provided */
561 parameters.splitPoint = 1.0;
562 parameters.f = parameters.f == 0 ? DEFAULT_F : parameters.f;
563 parameters.accel = parameters.accel == 0 ? DEFAULT_ACCEL : parameters.accel;
564 /* Convert to cover parameter */
565 memset(&coverParams, 0 , sizeof(coverParams));
566 FASTCOVER_convertToCoverParams(parameters, &coverParams);
567 /* Checks */
568 if (!FASTCOVER_checkParameters(coverParams, dictBufferCapacity, parameters.f,
569 parameters.accel)) {
570 DISPLAYLEVEL(1, "FASTCOVER parameters incorrect\n");
571 return ERROR(parameter_outOfBound);
572 }
573 if (nbSamples == 0) {
574 DISPLAYLEVEL(1, "FASTCOVER must have at least one input file\n");
575 return ERROR(srcSize_wrong);
576 }
577 if (dictBufferCapacity < ZDICT_DICTSIZE_MIN) {
578 DISPLAYLEVEL(1, "dictBufferCapacity must be at least %u\n",
579 ZDICT_DICTSIZE_MIN);
580 return ERROR(dstSize_tooSmall);
581 }
582 /* Assign corresponding FASTCOVER_accel_t to accelParams*/
583 accelParams = FASTCOVER_defaultAccelParameters[parameters.accel];
584 /* Initialize context */
585 {
586 size_t const initVal = FASTCOVER_ctx_init(&ctx, samplesBuffer, samplesSizes, nbSamples,
587 coverParams.d, parameters.splitPoint, parameters.f,
588 accelParams);
589 if (ZSTD_isError(initVal)) {
590 DISPLAYLEVEL(1, "Failed to initialize context\n");
591 return initVal;
592 }
593 }
594 COVER_warnOnSmallCorpus(dictBufferCapacity, ctx.nbDmers, g_displayLevel);
595 /* Build the dictionary */
596 DISPLAYLEVEL(2, "Building dictionary\n");
597 {
598 /* Initialize array to keep track of frequency of dmer within activeSegment */
599 U16* segmentFreqs = (U16 *)calloc(((U64)1 << parameters.f), sizeof(U16));
600 const size_t tail = FASTCOVER_buildDictionary(&ctx, ctx.freqs, dictBuffer,
601 dictBufferCapacity, coverParams, segmentFreqs);
602 const unsigned nbFinalizeSamples = (unsigned)(ctx.nbTrainSamples * ctx.accelParams.finalize / 100);
603 const size_t dictionarySize = ZDICT_finalizeDictionary(
604 dict, dictBufferCapacity, dict + tail, dictBufferCapacity - tail,
605 samplesBuffer, samplesSizes, nbFinalizeSamples, coverParams.zParams);
606 if (!ZSTD_isError(dictionarySize)) {
607 DISPLAYLEVEL(2, "Constructed dictionary of size %u\n",
608 (unsigned)dictionarySize);
609 }
610 FASTCOVER_ctx_destroy(&ctx);
611 free(segmentFreqs);
612 return dictionarySize;
613 }
614 }
615
616
617 ZDICTLIB_API size_t
ZDICT_optimizeTrainFromBuffer_fastCover(void * dictBuffer,size_t dictBufferCapacity,const void * samplesBuffer,const size_t * samplesSizes,unsigned nbSamples,ZDICT_fastCover_params_t * parameters)618 ZDICT_optimizeTrainFromBuffer_fastCover(
619 void* dictBuffer, size_t dictBufferCapacity,
620 const void* samplesBuffer,
621 const size_t* samplesSizes, unsigned nbSamples,
622 ZDICT_fastCover_params_t* parameters)
623 {
624 ZDICT_cover_params_t coverParams;
625 FASTCOVER_accel_t accelParams;
626 /* constants */
627 const unsigned nbThreads = parameters->nbThreads;
628 const double splitPoint =
629 parameters->splitPoint <= 0.0 ? FASTCOVER_DEFAULT_SPLITPOINT : parameters->splitPoint;
630 const unsigned kMinD = parameters->d == 0 ? 6 : parameters->d;
631 const unsigned kMaxD = parameters->d == 0 ? 8 : parameters->d;
632 const unsigned kMinK = parameters->k == 0 ? 50 : parameters->k;
633 const unsigned kMaxK = parameters->k == 0 ? 2000 : parameters->k;
634 const unsigned kSteps = parameters->steps == 0 ? 40 : parameters->steps;
635 const unsigned kStepSize = MAX((kMaxK - kMinK) / kSteps, 1);
636 const unsigned kIterations =
637 (1 + (kMaxD - kMinD) / 2) * (1 + (kMaxK - kMinK) / kStepSize);
638 const unsigned f = parameters->f == 0 ? DEFAULT_F : parameters->f;
639 const unsigned accel = parameters->accel == 0 ? DEFAULT_ACCEL : parameters->accel;
640 const unsigned shrinkDict = 0;
641 /* Local variables */
642 const int displayLevel = (int)parameters->zParams.notificationLevel;
643 unsigned iteration = 1;
644 unsigned d;
645 unsigned k;
646 COVER_best_t best;
647 POOL_ctx *pool = NULL;
648 int warned = 0;
649 /* Checks */
650 if (splitPoint <= 0 || splitPoint > 1) {
651 LOCALDISPLAYLEVEL(displayLevel, 1, "Incorrect splitPoint\n");
652 return ERROR(parameter_outOfBound);
653 }
654 if (accel == 0 || accel > FASTCOVER_MAX_ACCEL) {
655 LOCALDISPLAYLEVEL(displayLevel, 1, "Incorrect accel\n");
656 return ERROR(parameter_outOfBound);
657 }
658 if (kMinK < kMaxD || kMaxK < kMinK) {
659 LOCALDISPLAYLEVEL(displayLevel, 1, "Incorrect k\n");
660 return ERROR(parameter_outOfBound);
661 }
662 if (nbSamples == 0) {
663 LOCALDISPLAYLEVEL(displayLevel, 1, "FASTCOVER must have at least one input file\n");
664 return ERROR(srcSize_wrong);
665 }
666 if (dictBufferCapacity < ZDICT_DICTSIZE_MIN) {
667 LOCALDISPLAYLEVEL(displayLevel, 1, "dictBufferCapacity must be at least %u\n",
668 ZDICT_DICTSIZE_MIN);
669 return ERROR(dstSize_tooSmall);
670 }
671 if (nbThreads > 1) {
672 pool = POOL_create(nbThreads, 1);
673 if (!pool) {
674 return ERROR(memory_allocation);
675 }
676 }
677 /* Initialization */
678 COVER_best_init(&best);
679 memset(&coverParams, 0 , sizeof(coverParams));
680 FASTCOVER_convertToCoverParams(*parameters, &coverParams);
681 accelParams = FASTCOVER_defaultAccelParameters[accel];
682 /* Turn down global display level to clean up display at level 2 and below */
683 g_displayLevel = displayLevel == 0 ? 0 : displayLevel - 1;
684 /* Loop through d first because each new value needs a new context */
685 LOCALDISPLAYLEVEL(displayLevel, 2, "Trying %u different sets of parameters\n",
686 kIterations);
687 for (d = kMinD; d <= kMaxD; d += 2) {
688 /* Initialize the context for this value of d */
689 FASTCOVER_ctx_t ctx;
690 LOCALDISPLAYLEVEL(displayLevel, 3, "d=%u\n", d);
691 {
692 size_t const initVal = FASTCOVER_ctx_init(&ctx, samplesBuffer, samplesSizes, nbSamples, d, splitPoint, f, accelParams);
693 if (ZSTD_isError(initVal)) {
694 LOCALDISPLAYLEVEL(displayLevel, 1, "Failed to initialize context\n");
695 COVER_best_destroy(&best);
696 POOL_free(pool);
697 return initVal;
698 }
699 }
700 if (!warned) {
701 COVER_warnOnSmallCorpus(dictBufferCapacity, ctx.nbDmers, displayLevel);
702 warned = 1;
703 }
704 /* Loop through k reusing the same context */
705 for (k = kMinK; k <= kMaxK; k += kStepSize) {
706 /* Prepare the arguments */
707 FASTCOVER_tryParameters_data_t *data = (FASTCOVER_tryParameters_data_t *)malloc(
708 sizeof(FASTCOVER_tryParameters_data_t));
709 LOCALDISPLAYLEVEL(displayLevel, 3, "k=%u\n", k);
710 if (!data) {
711 LOCALDISPLAYLEVEL(displayLevel, 1, "Failed to allocate parameters\n");
712 COVER_best_destroy(&best);
713 FASTCOVER_ctx_destroy(&ctx);
714 POOL_free(pool);
715 return ERROR(memory_allocation);
716 }
717 data->ctx = &ctx;
718 data->best = &best;
719 data->dictBufferCapacity = dictBufferCapacity;
720 data->parameters = coverParams;
721 data->parameters.k = k;
722 data->parameters.d = d;
723 data->parameters.splitPoint = splitPoint;
724 data->parameters.steps = kSteps;
725 data->parameters.shrinkDict = shrinkDict;
726 data->parameters.zParams.notificationLevel = (unsigned)g_displayLevel;
727 /* Check the parameters */
728 if (!FASTCOVER_checkParameters(data->parameters, dictBufferCapacity,
729 data->ctx->f, accel)) {
730 DISPLAYLEVEL(1, "FASTCOVER parameters incorrect\n");
731 free(data);
732 continue;
733 }
734 /* Call the function and pass ownership of data to it */
735 COVER_best_start(&best);
736 if (pool) {
737 POOL_add(pool, &FASTCOVER_tryParameters, data);
738 } else {
739 FASTCOVER_tryParameters(data);
740 }
741 /* Print status */
742 LOCALDISPLAYUPDATE(displayLevel, 2, "\r%u%% ",
743 (unsigned)((iteration * 100) / kIterations));
744 ++iteration;
745 }
746 COVER_best_wait(&best);
747 FASTCOVER_ctx_destroy(&ctx);
748 }
749 LOCALDISPLAYLEVEL(displayLevel, 2, "\r%79s\r", "");
750 /* Fill the output buffer and parameters with output of the best parameters */
751 {
752 const size_t dictSize = best.dictSize;
753 if (ZSTD_isError(best.compressedSize)) {
754 const size_t compressedSize = best.compressedSize;
755 COVER_best_destroy(&best);
756 POOL_free(pool);
757 return compressedSize;
758 }
759 FASTCOVER_convertToFastCoverParams(best.parameters, parameters, f, accel);
760 memcpy(dictBuffer, best.dict, dictSize);
761 COVER_best_destroy(&best);
762 POOL_free(pool);
763 return dictSize;
764 }
765
766 }
767