Lines Matching refs:ip
367 #define INITBASE(base) const BYTE* const base = ip
527 const BYTE *ip = (BYTE *) source; in LZ4_compressCtx() local
529 const BYTE *anchor = ip; in LZ4_compressCtx()
530 const BYTE *const iend = ip + isize; in LZ4_compressCtx()
547 HashTable[LZ4_HASH_VALUE(ip)] = ip - base; in LZ4_compressCtx()
548 ip++; in LZ4_compressCtx()
549 forwardH = LZ4_HASH_VALUE(ip); in LZ4_compressCtx()
554 const BYTE *forwardIp = ip; in LZ4_compressCtx()
562 ip = forwardIp; in LZ4_compressCtx()
563 forwardIp = ip + step; in LZ4_compressCtx()
571 HashTable[h] = ip - base; in LZ4_compressCtx()
573 } while ((ref < ip - MAX_DISTANCE) || (A32(ref) != A32(ip))); in LZ4_compressCtx()
576 while ((ip > anchor) && (ref > (BYTE *) source) && in LZ4_compressCtx()
577 unlikely(ip[-1] == ref[-1])) { in LZ4_compressCtx()
578 ip--; in LZ4_compressCtx()
583 length = ip - anchor; in LZ4_compressCtx()
605 LZ4_WRITE_LITTLEENDIAN_16(op, ip - ref); in LZ4_compressCtx()
608 ip += MINMATCH; in LZ4_compressCtx()
610 anchor = ip; in LZ4_compressCtx()
611 while likely(ip < matchlimit - (STEPSIZE - 1)) { in LZ4_compressCtx()
612 UARCH diff = AARCH(ref) ^ AARCH(ip); in LZ4_compressCtx()
614 ip += STEPSIZE; in LZ4_compressCtx()
618 ip += LZ4_NbCommonBytes(diff); in LZ4_compressCtx()
622 if ((ip < (matchlimit - 3)) && (A32(ref) == A32(ip))) { in LZ4_compressCtx()
623 ip += 4; in LZ4_compressCtx()
627 if ((ip < (matchlimit - 1)) && (A16(ref) == A16(ip))) { in LZ4_compressCtx()
628 ip += 2; in LZ4_compressCtx()
631 if ((ip < matchlimit) && (*ref == *ip)) in LZ4_compressCtx()
632 ip++; in LZ4_compressCtx()
636 len = (ip - anchor); in LZ4_compressCtx()
656 if (ip > mflimit) { in LZ4_compressCtx()
657 anchor = ip; in LZ4_compressCtx()
661 HashTable[LZ4_HASH_VALUE(ip - 2)] = ip - 2 - base; in LZ4_compressCtx()
664 ref = base + HashTable[LZ4_HASH_VALUE(ip)]; in LZ4_compressCtx()
665 HashTable[LZ4_HASH_VALUE(ip)] = ip - base; in LZ4_compressCtx()
666 if ((ref > ip - (MAX_DISTANCE + 1)) && (A32(ref) == A32(ip))) { in LZ4_compressCtx()
672 anchor = ip++; in LZ4_compressCtx()
673 forwardH = LZ4_HASH_VALUE(ip); in LZ4_compressCtx()
722 const BYTE *ip = (BYTE *) source; in LZ4_compress64kCtx() local
723 const BYTE *anchor = ip; in LZ4_compress64kCtx()
724 const BYTE *const base = ip; in LZ4_compress64kCtx()
725 const BYTE *const iend = ip + isize; in LZ4_compress64kCtx()
741 ip++; in LZ4_compress64kCtx()
742 forwardH = LZ4_HASH64K_VALUE(ip); in LZ4_compress64kCtx()
747 const BYTE *forwardIp = ip; in LZ4_compress64kCtx()
755 ip = forwardIp; in LZ4_compress64kCtx()
756 forwardIp = ip + step; in LZ4_compress64kCtx()
764 HashTable[h] = ip - base; in LZ4_compress64kCtx()
766 } while (A32(ref) != A32(ip)); in LZ4_compress64kCtx()
769 while ((ip > anchor) && (ref > (BYTE *) source) && in LZ4_compress64kCtx()
770 (ip[-1] == ref[-1])) { in LZ4_compress64kCtx()
771 ip--; in LZ4_compress64kCtx()
776 length = ip - anchor; in LZ4_compress64kCtx()
798 LZ4_WRITE_LITTLEENDIAN_16(op, ip - ref); in LZ4_compress64kCtx()
801 ip += MINMATCH; in LZ4_compress64kCtx()
803 anchor = ip; in LZ4_compress64kCtx()
804 while (ip < matchlimit - (STEPSIZE - 1)) { in LZ4_compress64kCtx()
805 UARCH diff = AARCH(ref) ^ AARCH(ip); in LZ4_compress64kCtx()
807 ip += STEPSIZE; in LZ4_compress64kCtx()
811 ip += LZ4_NbCommonBytes(diff); in LZ4_compress64kCtx()
815 if ((ip < (matchlimit - 3)) && (A32(ref) == A32(ip))) { in LZ4_compress64kCtx()
816 ip += 4; in LZ4_compress64kCtx()
820 if ((ip < (matchlimit - 1)) && (A16(ref) == A16(ip))) { in LZ4_compress64kCtx()
821 ip += 2; in LZ4_compress64kCtx()
824 if ((ip < matchlimit) && (*ref == *ip)) in LZ4_compress64kCtx()
825 ip++; in LZ4_compress64kCtx()
829 len = (ip - anchor); in LZ4_compress64kCtx()
849 if (ip > mflimit) { in LZ4_compress64kCtx()
850 anchor = ip; in LZ4_compress64kCtx()
854 HashTable[LZ4_HASH64K_VALUE(ip - 2)] = ip - 2 - base; in LZ4_compress64kCtx()
857 ref = base + HashTable[LZ4_HASH64K_VALUE(ip)]; in LZ4_compress64kCtx()
858 HashTable[LZ4_HASH64K_VALUE(ip)] = ip - base; in LZ4_compress64kCtx()
859 if (A32(ref) == A32(ip)) { in LZ4_compress64kCtx()
865 anchor = ip++; in LZ4_compress64kCtx()
866 forwardH = LZ4_HASH64K_VALUE(ip); in LZ4_compress64kCtx()
936 const BYTE *restrict ip = (const BYTE *) source; in real_LZ4_uncompress() local
954 token = *ip++; in real_LZ4_uncompress()
957 for (; (len = *ip++) == 255; length += 255) { in real_LZ4_uncompress()
970 (void) memcpy(op, ip, length); in real_LZ4_uncompress()
971 ip += length; in real_LZ4_uncompress()
974 LZ4_WILDCOPY(ip, op, cpy); in real_LZ4_uncompress()
975 ip -= (op - cpy); in real_LZ4_uncompress()
979 LZ4_READ_LITTLEENDIAN_16(ref, cpy, ip); in real_LZ4_uncompress()
980 ip += 2; in real_LZ4_uncompress()
990 for (; *ip == 255; length += 255) { in real_LZ4_uncompress()
991 ip++; in real_LZ4_uncompress()
993 length += *ip++; in real_LZ4_uncompress()
1040 return (int)(((char *)ip) - source); in real_LZ4_uncompress()
1044 return (int)(-(((char *)ip) - source)); in real_LZ4_uncompress()
1052 const BYTE *restrict ip = (const BYTE *) source; in LZ4_uncompress_unknownOutputSize() local
1053 const BYTE *const iend = ip + isize; in LZ4_uncompress_unknownOutputSize()
1066 while (ip < iend) { in LZ4_uncompress_unknownOutputSize()
1071 token = *ip++; in LZ4_uncompress_unknownOutputSize()
1074 while ((ip < iend) && (s == 255)) { in LZ4_uncompress_unknownOutputSize()
1075 s = *ip++; in LZ4_uncompress_unknownOutputSize()
1085 (ip + length > iend - COPYLENGTH)) { in LZ4_uncompress_unknownOutputSize()
1089 if (ip + length != iend) in LZ4_uncompress_unknownOutputSize()
1095 (void) memcpy(op, ip, length); in LZ4_uncompress_unknownOutputSize()
1100 LZ4_WILDCOPY(ip, op, cpy); in LZ4_uncompress_unknownOutputSize()
1101 ip -= (op - cpy); in LZ4_uncompress_unknownOutputSize()
1105 LZ4_READ_LITTLEENDIAN_16(ref, cpy, ip); in LZ4_uncompress_unknownOutputSize()
1106 ip += 2; in LZ4_uncompress_unknownOutputSize()
1116 while (ip < iend) { in LZ4_uncompress_unknownOutputSize()
1117 int s = *ip++; in LZ4_uncompress_unknownOutputSize()
1173 return (int)(-(((char *)ip) - source)); in LZ4_uncompress_unknownOutputSize()