Lines Matching refs:ref

581 		const BYTE *ref;  in LZ4_compressCtx()  local
596 ref = base + HashTable[h]; in LZ4_compressCtx()
599 } while ((ref < ip - MAX_DISTANCE) || (A32(ref) != A32(ip))); in LZ4_compressCtx()
602 while ((ip > anchor) && (ref > (const BYTE *) source) && in LZ4_compressCtx()
603 unlikely(ip[-1] == ref[-1])) { in LZ4_compressCtx()
605 ref--; in LZ4_compressCtx()
631 LZ4_WRITE_LITTLEENDIAN_16(op, ip - ref); in LZ4_compressCtx()
635 ref += MINMATCH; /* MinMatch verified */ in LZ4_compressCtx()
638 UARCH diff = AARCH(ref) ^ AARCH(ip); in LZ4_compressCtx()
641 ref += STEPSIZE; in LZ4_compressCtx()
648 if ((ip < (matchlimit - 3)) && (A32(ref) == A32(ip))) { in LZ4_compressCtx()
650 ref += 4; in LZ4_compressCtx()
653 if ((ip < (matchlimit - 1)) && (A16(ref) == A16(ip))) { in LZ4_compressCtx()
655 ref += 2; in LZ4_compressCtx()
657 if ((ip < matchlimit) && (*ref == *ip)) in LZ4_compressCtx()
690 ref = base + HashTable[LZ4_HASH_VALUE(ip)]; in LZ4_compressCtx()
692 if ((ref > ip - (MAX_DISTANCE + 1)) && (A32(ref) == A32(ip))) { in LZ4_compressCtx()
774 const BYTE *ref; in LZ4_compress64kCtx() local
789 ref = base + HashTable[h]; in LZ4_compress64kCtx()
792 } while (A32(ref) != A32(ip)); in LZ4_compress64kCtx()
795 while ((ip > anchor) && (ref > (const BYTE *) source) && in LZ4_compress64kCtx()
796 (ip[-1] == ref[-1])) { in LZ4_compress64kCtx()
798 ref--; in LZ4_compress64kCtx()
824 LZ4_WRITE_LITTLEENDIAN_16(op, ip - ref); in LZ4_compress64kCtx()
828 ref += MINMATCH; /* MinMatch verified */ in LZ4_compress64kCtx()
831 UARCH diff = AARCH(ref) ^ AARCH(ip); in LZ4_compress64kCtx()
834 ref += STEPSIZE; in LZ4_compress64kCtx()
841 if ((ip < (matchlimit - 3)) && (A32(ref) == A32(ip))) { in LZ4_compress64kCtx()
843 ref += 4; in LZ4_compress64kCtx()
846 if ((ip < (matchlimit - 1)) && (A16(ref) == A16(ip))) { in LZ4_compress64kCtx()
848 ref += 2; in LZ4_compress64kCtx()
850 if ((ip < matchlimit) && (*ref == *ip)) in LZ4_compress64kCtx()
883 ref = base + HashTable[LZ4_HASH64K_VALUE(ip)]; in LZ4_compress64kCtx()
885 if (A32(ref) == A32(ip)) { in LZ4_compress64kCtx()
973 const BYTE *ref; in LZ4_uncompress_unknownOutputSize() local
1024 LZ4_READ_LITTLEENDIAN_16(ref, cpy, ip); in LZ4_uncompress_unknownOutputSize()
1026 if (ref < (BYTE * const) dest) in LZ4_uncompress_unknownOutputSize()
1044 if unlikely(op - ref < STEPSIZE) { in LZ4_uncompress_unknownOutputSize()
1046 size_t dec64 = dec64table[op-ref]; in LZ4_uncompress_unknownOutputSize()
1050 op[0] = ref[0]; in LZ4_uncompress_unknownOutputSize()
1051 op[1] = ref[1]; in LZ4_uncompress_unknownOutputSize()
1052 op[2] = ref[2]; in LZ4_uncompress_unknownOutputSize()
1053 op[3] = ref[3]; in LZ4_uncompress_unknownOutputSize()
1055 ref += 4; in LZ4_uncompress_unknownOutputSize()
1056 ref -= dec32table[op-ref]; in LZ4_uncompress_unknownOutputSize()
1057 A32(op) = A32(ref); in LZ4_uncompress_unknownOutputSize()
1059 ref -= dec64; in LZ4_uncompress_unknownOutputSize()
1061 LZ4_COPYSTEP(ref, op); in LZ4_uncompress_unknownOutputSize()
1071 LZ4_SECURECOPY(ref, op, (oend - COPYLENGTH)); in LZ4_uncompress_unknownOutputSize()
1073 *op++ = *ref++; in LZ4_uncompress_unknownOutputSize()
1083 LZ4_SECURECOPY(ref, op, cpy); in LZ4_uncompress_unknownOutputSize()