Searched refs:matchIndex (Results 1 – 10 of 10) sorted by relevance
45 U32 const matchIndex = hashTable[h]; in ZSTD_updateDUBT() local52 *nextCandidatePtr = matchIndex; /* update BT like a chain */ in ZSTD_updateDUBT()84 …U32 matchIndex = *smallerPtr; /* this candidate is unsorted : next sorted candidate is reached t… in ZSTD_insertDUBT1() local96 while (nbCompares-- && (matchIndex > windowLow)) { in ZSTD_insertDUBT1()97 U32* const nextPtr = bt + 2*(matchIndex & btMask); in ZSTD_insertDUBT1()99 assert(matchIndex < current); in ZSTD_insertDUBT1()105 || (matchIndex+matchLength >= dictLimit) /* both in current segment*/ in ZSTD_insertDUBT1()108 || (matchIndex+matchLength >= dictLimit)) ? in ZSTD_insertDUBT1()110 …assert( (matchIndex+matchLength >= dictLimit) /* might be wrong if extDict is incorrectly set to… in ZSTD_insertDUBT1()112 match = mBase + matchIndex; in ZSTD_insertDUBT1()[all …]
381 U32 matchIndex = hashTable[h]; in ZSTD_insertBt1() local411 while (nbCompares-- && (matchIndex >= windowLow)) { in ZSTD_insertBt1()412 U32* const nextPtr = bt + 2*(matchIndex & btMask); in ZSTD_insertBt1()414 assert(matchIndex < current); in ZSTD_insertBt1()417 …const U32* predictPtr = bt + 2*((matchIndex-1) & btMask); /* written this way, as bt is a roll b… in ZSTD_insertBt1()418 if (matchIndex == predictedSmall) { in ZSTD_insertBt1()420 *smallerPtr = matchIndex; in ZSTD_insertBt1()421 … if (matchIndex <= btLow) { smallerPtr=&dummy32; break; } /* beyond tree size, stop the search */ in ZSTD_insertBt1()423 …matchIndex = nextPtr[1]; /* new matchIndex larger than previous (closer to current) */ in ZSTD_insertBt1()427 if (matchIndex == predictedLarge) { in ZSTD_insertBt1()[all …]
262 U32 const matchIndex = hashTable[h]; in ZSTD_compressBlock_fast_dictMatchState_generic() local263 const BYTE* match = base + matchIndex; in ZSTD_compressBlock_fast_dictMatchState_generic()276 } else if ( (matchIndex <= prefixStartIndex) ) { in ZSTD_compressBlock_fast_dictMatchState_generic()410 const U32 matchIndex = hashTable[h]; in ZSTD_compressBlock_fast_extDict_generic() local411 const BYTE* const matchBase = matchIndex < prefixStartIndex ? dictBase : base; in ZSTD_compressBlock_fast_extDict_generic()412 const BYTE* match = matchBase + matchIndex; in ZSTD_compressBlock_fast_extDict_generic()430 if ( (matchIndex < dictStartIndex) || in ZSTD_compressBlock_fast_extDict_generic()436 { const BYTE* const matchEnd = matchIndex < prefixStartIndex ? dictEnd : iend; in ZSTD_compressBlock_fast_extDict_generic()437 … const BYTE* const lowMatchPtr = matchIndex < prefixStartIndex ? dictStart : prefixStart; in ZSTD_compressBlock_fast_extDict_generic()438 U32 const offset = current - matchIndex; in ZSTD_compressBlock_fast_extDict_generic()
395 const U32 matchIndex = hashSmall[hSmall]; in ZSTD_compressBlock_doubleFast_extDict_generic() local396 const BYTE* const matchBase = matchIndex < prefixStartIndex ? dictBase : base; in ZSTD_compressBlock_doubleFast_extDict_generic()397 const BYTE* match = matchBase + matchIndex; in ZSTD_compressBlock_doubleFast_extDict_generic()430 } else if ((matchIndex > dictStartIndex) && (MEM_read32(match) == MEM_read32(ip))) { in ZSTD_compressBlock_doubleFast_extDict_generic()445 const BYTE* const matchEnd = matchIndex < prefixStartIndex ? dictEnd : iend; in ZSTD_compressBlock_doubleFast_extDict_generic()446 … const BYTE* const lowMatchPtr = matchIndex < prefixStartIndex ? dictStart : prefixStart; in ZSTD_compressBlock_doubleFast_extDict_generic()448 offset = current - matchIndex; in ZSTD_compressBlock_doubleFast_extDict_generic()
383 U32 const matchIndex = bestEntry->offset; in ZSTD_ldm_generateSequences_internal() local384 U32 const offset = current - matchIndex; in ZSTD_ldm_generateSequences_internal()
45 U32 const matchIndex = hashTable[h]; in ZSTD_updateDUBT() local52 *nextCandidatePtr = matchIndex; /* update BT like a chain */ in ZSTD_updateDUBT()84 …U32 matchIndex = *smallerPtr; /* this candidate is unsorted : next sorted candidate is reached t… in ZSTD_insertDUBT1() local96 for (; nbCompares && (matchIndex > windowLow); --nbCompares) { in ZSTD_insertDUBT1()97 U32* const nextPtr = bt + 2*(matchIndex & btMask); in ZSTD_insertDUBT1()99 assert(matchIndex < curr); in ZSTD_insertDUBT1()105 || (matchIndex+matchLength >= dictLimit) /* both in current segment*/ in ZSTD_insertDUBT1()108 || (matchIndex+matchLength >= dictLimit)) ? in ZSTD_insertDUBT1()110 …assert( (matchIndex+matchLength >= dictLimit) /* might be wrong if extDict is incorrectly set to… in ZSTD_insertDUBT1()112 match = mBase + matchIndex; in ZSTD_insertDUBT1()[all …]
424 U32 matchIndex = hashTable[h]; in ZSTD_insertBt1() local458 for (; nbCompares && (matchIndex >= windowLow); --nbCompares) { in ZSTD_insertBt1()459 U32* const nextPtr = bt + 2*(matchIndex & btMask); in ZSTD_insertBt1()461 assert(matchIndex < curr); in ZSTD_insertBt1()464 …const U32* predictPtr = bt + 2*((matchIndex-1) & btMask); /* written this way, as bt is a roll b… in ZSTD_insertBt1()465 if (matchIndex == predictedSmall) { in ZSTD_insertBt1()467 *smallerPtr = matchIndex; in ZSTD_insertBt1()468 … if (matchIndex <= btLow) { smallerPtr=&dummy32; break; } /* beyond tree size, stop the search */ in ZSTD_insertBt1()470 …matchIndex = nextPtr[1]; /* new matchIndex larger than previous (closer to current) */ in ZSTD_insertBt1()474 if (matchIndex == predictedLarge) { in ZSTD_insertBt1()[all …]
429 U32 const matchIndex = hashTable[h]; in ZSTD_compressBlock_fast_dictMatchState_generic() local430 const BYTE* match = base + matchIndex; in ZSTD_compressBlock_fast_dictMatchState_generic()443 } else if ( (matchIndex <= prefixStartIndex) ) { in ZSTD_compressBlock_fast_dictMatchState_generic()585 const U32 matchIndex = hashTable[h]; in ZSTD_compressBlock_fast_extDict_generic() local586 const BYTE* const matchBase = matchIndex < prefixStartIndex ? dictBase : base; in ZSTD_compressBlock_fast_extDict_generic()587 const BYTE* match = matchBase + matchIndex; in ZSTD_compressBlock_fast_extDict_generic()605 if ( (matchIndex < dictStartIndex) || in ZSTD_compressBlock_fast_extDict_generic()611 { const BYTE* const matchEnd = matchIndex < prefixStartIndex ? dictEnd : iend; in ZSTD_compressBlock_fast_extDict_generic()612 … const BYTE* const lowMatchPtr = matchIndex < prefixStartIndex ? dictStart : prefixStart; in ZSTD_compressBlock_fast_extDict_generic()613 U32 const offset = curr - matchIndex; in ZSTD_compressBlock_fast_extDict_generic()
566 const U32 matchIndex = hashSmall[hSmall]; in ZSTD_compressBlock_doubleFast_extDict_generic() local567 const BYTE* const matchBase = matchIndex < prefixStartIndex ? dictBase : base; in ZSTD_compressBlock_doubleFast_extDict_generic()568 const BYTE* match = matchBase + matchIndex; in ZSTD_compressBlock_doubleFast_extDict_generic()601 } else if ((matchIndex > dictStartIndex) && (MEM_read32(match) == MEM_read32(ip))) { in ZSTD_compressBlock_doubleFast_extDict_generic()616 const BYTE* const matchEnd = matchIndex < prefixStartIndex ? dictEnd : iend; in ZSTD_compressBlock_doubleFast_extDict_generic()617 … const BYTE* const lowMatchPtr = matchIndex < prefixStartIndex ? dictStart : prefixStart; in ZSTD_compressBlock_doubleFast_extDict_generic()619 offset = curr - matchIndex; in ZSTD_compressBlock_doubleFast_extDict_generic()
2107 int matchIndex = -1, lowIndex = -1; in ar5416GetTargetPowers() local2116 matchIndex = 0; in ar5416GetTargetPowers()2120 matchIndex = i; in ar5416GetTargetPowers()2129 if ((matchIndex == -1) && (lowIndex == -1)) { in ar5416GetTargetPowers()2131 matchIndex = i - 1; in ar5416GetTargetPowers()2135 if (matchIndex != -1) { in ar5416GetTargetPowers()2136 OS_MEMCPY(pNewPower, &powInfo[matchIndex], sizeof(*pNewPower)); in ar5416GetTargetPowers()2167 int matchIndex = -1, lowIndex = -1; in ar5416GetTargetPowersLeg() local2176 matchIndex = 0; in ar5416GetTargetPowersLeg()2180 matchIndex = i; in ar5416GetTargetPowersLeg()[all …]