/lib/zstd/ |
D | entropy_common.c | 62 const BYTE *const iend = istart + hbSize; in FSE_readNCount() local 90 if (ip < iend - 5) { in FSE_readNCount() 109 if ((ip <= iend - 7) || (ip + (bitCount >> 3) <= iend - 4)) { in FSE_readNCount() 140 if ((ip <= iend - 7) || (ip + (bitCount >> 3) <= iend - 4)) { in FSE_readNCount() 144 bitCount -= (int)(8 * (iend - 4 - ip)); in FSE_readNCount() 145 ip = iend - 4; in FSE_readNCount()
|
D | compress.c | 998 const BYTE *const iend = ((const BYTE *)end) - HASH_READ_SIZE; in ZSTD_fillHashTable() local 1001 while (ip <= iend) { in ZSTD_fillHashTable() 1019 const BYTE *const iend = istart + srcSize; in ZSTD_compressBlock_fast_generic() local 1020 const BYTE *const ilimit = iend - HASH_READ_SIZE; in ZSTD_compressBlock_fast_generic() 1044 mLength = ZSTD_count(ip + 1 + 4, ip + 1 + 4 - offset_1, iend) + 4; in ZSTD_compressBlock_fast_generic() 1053 mLength = ZSTD_count(ip + 4, match + 4, iend) + 4; in ZSTD_compressBlock_fast_generic() 1077 size_t const rLength = ZSTD_count(ip + 4, ip + 4 - offset_2, iend) + 4; in ZSTD_compressBlock_fast_generic() 1098 size_t const lastLLSize = iend - anchor; in ZSTD_compressBlock_fast_generic() 1131 const BYTE *const iend = istart + srcSize; in ZSTD_compressBlock_fast_extDict_generic() local 1132 const BYTE *const ilimit = iend - 8; in ZSTD_compressBlock_fast_extDict_generic() [all …]
|
D | zstd_opt.h | 413 const BYTE *const iend = istart + srcSize; in ZSTD_compressBlock_opt_generic() local 414 const BYTE *const ilimit = iend - 8; in ZSTD_compressBlock_opt_generic() 453 mlen = (U32)ZSTD_count(ip + minMatch, ip + minMatch - repCur, iend) + minMatch; in ZSTD_compressBlock_opt_generic() 472 match_num = ZSTD_BtGetAllMatches_selectMLS(ctx, ip, iend, maxSearches, mls, matches, minMatch); in ZSTD_compressBlock_opt_generic() 557 mlen = (U32)ZSTD_count(inr + minMatch, inr + minMatch - repCur, iend) + minMatch; in ZSTD_compressBlock_opt_generic() 591 match_num = ZSTD_BtGetAllMatches_selectMLS(ctx, inr, iend, maxSearches, mls, matches, best_mlen); in ZSTD_compressBlock_opt_generic() 693 size_t const lastLLSize = iend - anchor; in ZSTD_compressBlock_opt_generic() 706 const BYTE *const iend = istart + srcSize; in ZSTD_compressBlock_opt_extDict_generic() local 707 const BYTE *const ilimit = iend - 8; in ZSTD_compressBlock_opt_extDict_generic() 757 const BYTE *const repEnd = repIndex < dictLimit ? dictEnd : iend; in ZSTD_compressBlock_opt_extDict_generic() [all …]
|
D | decompress.c | 797 const BYTE *const iend = istart + srcSize; in ZSTD_decodeSeqHeaders() local 813 if (ip + 2 > iend) in ZSTD_decodeSeqHeaders() 817 if (ip >= iend) in ZSTD_decodeSeqHeaders() 826 if (ip + 4 > iend) in ZSTD_decodeSeqHeaders() 836 … = ZSTD_buildSeqTable(dctx->entropy.LLTable, &dctx->LLTptr, LLtype, MaxLL, LLFSELog, ip, iend - ip, in ZSTD_decodeSeqHeaders() 843 … ZSTD_buildSeqTable(dctx->entropy.OFTable, &dctx->OFTptr, OFtype, MaxOff, OffFSELog, ip, iend - ip, in ZSTD_decodeSeqHeaders() 850 … = ZSTD_buildSeqTable(dctx->entropy.MLTable, &dctx->MLTptr, MLtype, MaxML, MLFSELog, ip, iend - ip, in ZSTD_decodeSeqHeaders() 1096 const BYTE *const iend = ip + seqSize; in ZSTD_decompressSequences() local 1124 CHECK_E(BIT_initDStream(&seqState.DStream, ip, iend - ip), corruption_detected); in ZSTD_decompressSequences() 1351 const BYTE *const iend = ip + seqSize; in ZSTD_decompressSequencesLong() local [all …]
|
D | fse_compress.c | 361 const BYTE *const iend = ip + sourceSize; in FSE_count_parallel_wksp() local 384 while (ip < iend - 15) { in FSE_count_parallel_wksp() 418 while (ip < iend) in FSE_count_parallel_wksp() 727 const BYTE *const iend = istart + srcSize; in FSE_compress_usingCTable_generic() local 728 const BYTE *ip = iend; in FSE_compress_usingCTable_generic()
|
D | huf_compress.c | 586 const BYTE *const iend = ip + srcSize; in HUF_compress4X_usingCTable() local 625 CHECK_V_F(cSize, HUF_compress1X_usingCTable(op, oend - op, ip, iend - ip, CTable)); in HUF_compress4X_usingCTable()
|
/lib/lz4/ |
D | lz4_decompress.c | 84 const BYTE * const iend = ip + srcSize; in LZ4_decompress_generic() local 98 const BYTE *const shortiend = iend - in LZ4_decompress_generic() 131 assert(!endOnInput || ip <= iend); in LZ4_decompress_generic() 195 if (unlikely(endOnInput ? ip >= iend - RUN_MASK : 0)) { in LZ4_decompress_generic() 203 ? ip < iend - RUN_MASK in LZ4_decompress_generic() 225 || (ip + length > iend - (2 + 1 + LASTLITERALS)))) in LZ4_decompress_generic() 237 && (ip + length > iend)) { in LZ4_decompress_generic() 256 && ((ip + length != iend) in LZ4_decompress_generic() 279 if (!partialDecoding || (cpy == oend) || (ip >= (iend - 2))) in LZ4_decompress_generic() 320 if ((endOnInput) && (ip > iend - LASTLITERALS)) in LZ4_decompress_generic()
|
D | lz4_compress.c | 196 const BYTE * const iend = ip + inputSize; in LZ4_compress_generic() local 197 const BYTE * const mflimit = iend - MFLIMIT; in LZ4_compress_generic() 198 const BYTE * const matchlimit = iend - LASTLITERALS; in LZ4_compress_generic() 431 size_t const lastRun = (size_t)(iend - anchor); in LZ4_compress_generic() 534 const BYTE * const iend = ip + *srcSizePtr; in LZ4_compress_destSize_generic() local 535 const BYTE * const mflimit = iend - MFLIMIT; in LZ4_compress_destSize_generic() 536 const BYTE * const matchlimit = iend - LASTLITERALS; in LZ4_compress_destSize_generic() 690 size_t lastRunSize = (size_t)(iend - anchor); in LZ4_compress_destSize_generic()
|
D | lz4hc_compress.c | 350 const BYTE * const iend = ip + inputSize; in LZ4HC_compress_generic() local 351 const BYTE * const mflimit = iend - MFLIMIT; in LZ4HC_compress_generic() 352 const BYTE * const matchlimit = (iend - LASTLITERALS); in LZ4HC_compress_generic() 556 int lastRun = (int)(iend - anchor); in LZ4HC_compress_generic() 573 LZ4_memcpy(op, anchor, iend - anchor); in LZ4HC_compress_generic() 574 op += iend - anchor; in LZ4HC_compress_generic()
|