/lib/lz4/ |
D | lz4_decompress.c | 123 const BYTE *match; in LZ4_decompress_generic() local 167 match = op - offset; in LZ4_decompress_generic() 168 assert(match <= op); /* check overflow */ in LZ4_decompress_generic() 173 (dict == withPrefix64k || match >= lowPrefix)) { in LZ4_decompress_generic() 175 LZ4_memcpy(op + 0, match + 0, 8); in LZ4_decompress_generic() 176 LZ4_memcpy(op + 8, match + 8, 8); in LZ4_decompress_generic() 177 LZ4_memcpy(op + 16, match + 16, 2); in LZ4_decompress_generic() 291 match = op - offset; in LZ4_decompress_generic() 297 if ((checkOffset) && (unlikely(match + dictSize < lowPrefix))) { in LZ4_decompress_generic() 337 if ((dict == usingExtDict) && (match < lowPrefix)) { in LZ4_decompress_generic() [all …]
|
D | lz4_compress.c | 246 const BYTE *match; in LZ4_compress_generic() local 265 match = LZ4_getPositionOnHash(h, in LZ4_compress_generic() 270 if (match < (const BYTE *)source) { in LZ4_compress_generic() 284 ? (match < lowRefLimit) in LZ4_compress_generic() 288 : (match + MAX_DISTANCE < ip)) in LZ4_compress_generic() 289 || (LZ4_read32(match + refDelta) in LZ4_compress_generic() 294 while (((ip > anchor) & (match + refDelta > lowLimit)) in LZ4_compress_generic() 295 && (unlikely(ip[-1] == match[refDelta - 1]))) { in LZ4_compress_generic() 297 match--; in LZ4_compress_generic() 331 LZ4_writeLE16(op, (U16)(ip - match)); in LZ4_compress_generic() [all …]
|
D | lz4hc_compress.c | 126 const BYTE * const match = base + matchIndex; in LZ4HC_InsertAndFindBestMatch() local 128 if (*(match + ml) == *(ip + ml) in LZ4HC_InsertAndFindBestMatch() 129 && (LZ4_read32(match) == LZ4_read32(ip))) { in LZ4HC_InsertAndFindBestMatch() 131 match + MINMATCH, iLimit) + MINMATCH; in LZ4HC_InsertAndFindBestMatch() 135 *matchpos = match; in LZ4HC_InsertAndFindBestMatch() 139 const BYTE * const match = dictBase + matchIndex; in LZ4HC_InsertAndFindBestMatch() local 141 if (LZ4_read32(match) == LZ4_read32(ip)) { in LZ4HC_InsertAndFindBestMatch() 149 match + MINMATCH, vLimit) + MINMATCH; in LZ4HC_InsertAndFindBestMatch() 267 const BYTE * const match, in LZ4HC_encodeSequence() argument 300 LZ4_writeLE16(*op, (U16)(*ip - match)); in LZ4HC_encodeSequence()
|
/lib/zlib_deflate/ |
D | deflate.c | 82 static void check_match (deflate_state *s, IPos start, IPos match, 565 register Byte *match; /* matched string */ in longest_match() local 608 match = s->window + cur_match; in longest_match() 617 if (*(ush*)(match+best_len-1) != scan_end || in longest_match() 618 *(ush*)match != scan_start) continue; in longest_match() 629 Assert(scan[2] == match[2], "scan[2]?"); in longest_match() 630 scan++, match++; in longest_match() 632 } while (*(ush*)(scan+=2) == *(ush*)(match+=2) && in longest_match() 633 *(ush*)(scan+=2) == *(ush*)(match+=2) && in longest_match() 634 *(ush*)(scan+=2) == *(ush*)(match+=2) && in longest_match() [all …]
|
/lib/zstd/compress/ |
D | zstd_lazy.c | 81 const BYTE* match; in ZSTD_insertDUBT1() local 112 match = mBase + matchIndex; in ZSTD_insertDUBT1() 113 matchLength += ZSTD_count(ip+matchLength, match+matchLength, iend); in ZSTD_insertDUBT1() 115 match = dictBase + matchIndex; in ZSTD_insertDUBT1() 116 …matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iend, dictEnd, prefixStart); in ZSTD_insertDUBT1() 118 match = base + matchIndex; /* preparation for next read of match[matchLength] */ in ZSTD_insertDUBT1() 128 if (match[matchLength] < ip[matchLength]) { /* necessarily within buffer */ in ZSTD_insertDUBT1() 191 const BYTE* match = dictBase + dictMatchIndex; in ZSTD_DUBT_findBetterDictMatch() local 192 …matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iend, dictEnd, prefixStart); in ZSTD_DUBT_findBetterDictMatch() 194 …match = base + dictMatchIndex + dictIndexDelta; /* to prepare for next usage of match[matchLengt… in ZSTD_DUBT_findBetterDictMatch() [all …]
|
D | zstd_double_fast.c | 136 const BYTE* match = base + matchIndexS; in ZSTD_compressBlock_doubleFast_generic() local 187 if (MEM_read32(match) == MEM_read32(ip)) { in ZSTD_compressBlock_doubleFast_generic() 193 match = dictBase + dictMatchIndexS; in ZSTD_compressBlock_doubleFast_generic() 196 if (match > dictStart && MEM_read32(match) == MEM_read32(ip)) { in ZSTD_compressBlock_doubleFast_generic() 238 mLength = ZSTD_count_2segments(ip+4, match+4, iend, dictEnd, prefixLowest) + 4; in ZSTD_compressBlock_doubleFast_generic() 240 …while (((ip>anchor) & (match>dictStart)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /… in ZSTD_compressBlock_doubleFast_generic() 242 mLength = ZSTD_count(ip+4, match+4, iend) + 4; in ZSTD_compressBlock_doubleFast_generic() 243 offset = (U32)(ip - match); in ZSTD_compressBlock_doubleFast_generic() 244 …while (((ip>anchor) & (match>prefixLowest)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; … in ZSTD_compressBlock_doubleFast_generic() 395 const BYTE* match = matchBase + matchIndex; in ZSTD_compressBlock_doubleFast_extDict_generic() local [all …]
|
D | zstd_fast.c | 263 const BYTE* match = base + matchIndex; in ZSTD_compressBlock_fast_dictMatchState_generic() local 297 } else if (MEM_read32(match) != MEM_read32(ip)) { in ZSTD_compressBlock_fast_dictMatchState_generic() 304 U32 const offset = (U32)(ip-match); in ZSTD_compressBlock_fast_dictMatchState_generic() 305 mLength = ZSTD_count(ip+4, match+4, iend) + 4; in ZSTD_compressBlock_fast_dictMatchState_generic() 306 while (((ip>anchor) & (match>prefixStart)) in ZSTD_compressBlock_fast_dictMatchState_generic() 307 && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */ in ZSTD_compressBlock_fast_dictMatchState_generic() 412 const BYTE* match = matchBase + matchIndex; in ZSTD_compressBlock_fast_extDict_generic() local 431 (MEM_read32(match) != MEM_read32(ip)) ) { in ZSTD_compressBlock_fast_extDict_generic() 439 … size_t mLength = ZSTD_count_2segments(ip+4, match+4, iend, matchEnd, prefixStart) + 4; in ZSTD_compressBlock_fast_extDict_generic() 440 …while (((ip>anchor) & (match>lowMatchPtr)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; }… in ZSTD_compressBlock_fast_extDict_generic()
|
D | zstd_opt.c | 400 const BYTE* match; in ZSTD_insertBt1() local 451 match = base + matchIndex; in ZSTD_insertBt1() 452 matchLength += ZSTD_count(ip+matchLength, match+matchLength, iend); in ZSTD_insertBt1() 454 match = dictBase + matchIndex; in ZSTD_insertBt1() 455 …matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iend, dictEnd, prefixStart); in ZSTD_insertBt1() 457 match = base + matchIndex; /* to prepare for next usage of match[matchLength] */ in ZSTD_insertBt1() 470 if (match[matchLength] < ip[matchLength]) { /* necessarily within buffer */ in ZSTD_insertBt1() 627 const BYTE* const match = base + matchIndex3; in ZSTD_insertBtAndGetAllMatches() local 628 mlen = ZSTD_count(ip, match, iLimit); in ZSTD_insertBtAndGetAllMatches() 630 const BYTE* const match = dictBase + matchIndex3; in ZSTD_insertBtAndGetAllMatches() local [all …]
|
D | zstd_compress_internal.h | 641 ZSTD_count_2segments(const BYTE* ip, const BYTE* match, in ZSTD_count_2segments() argument 644 const BYTE* const vEnd = MIN( ip + (mEnd - match), iEnd); in ZSTD_count_2segments() 645 size_t const matchLength = ZSTD_count(ip, match, vEnd); in ZSTD_count_2segments() 646 if (match + matchLength != mEnd) return matchLength; in ZSTD_count_2segments() 648 DEBUGLOG(7, "distance from match beginning to end dictionary = %zi", mEnd - match); in ZSTD_count_2segments()
|
/lib/ |
D | glob.c | 71 bool match = false, inverted = (*pat == '!'); in glob_match() local 95 match |= (a <= c && c <= b); in glob_match() 98 if (match == inverted) in glob_match()
|
D | globtest.c | 21 bool match = glob_match(pat, str); in test() local 22 bool success = match == expected; in test() 39 printk(message, pat, str, mismatch + 3*match); in test()
|
D | test_bpf.c | 2308 bool match = __bpf_match_jmp_cond(dst, (s32)imm, op); in __bpf_emit_jmp_imm() local 2311 insns[i++] = BPF_ALU32_IMM(BPF_MOV, R0, match); in __bpf_emit_jmp_imm() 2315 if (!match) in __bpf_emit_jmp_imm() 2331 bool match = __bpf_match_jmp_cond((s32)dst, (s32)imm, op); in __bpf_emit_jmp32_imm() local 2336 if (!match) in __bpf_emit_jmp32_imm() 2352 bool match = __bpf_match_jmp_cond(dst, src, op); in __bpf_emit_jmp_reg() local 2358 if (!match) in __bpf_emit_jmp_reg() 2374 bool match = __bpf_match_jmp_cond((s32)dst, (s32)src, op); in __bpf_emit_jmp32_reg() local 2380 if (!match) in __bpf_emit_jmp32_reg()
|
D | Kconfig | 37 Drivers may use these helpers to match the bit indices as described
|
D | Kconfig.debug | 386 For modules whose split BTF does not match vmlinux, load without
|
/lib/zstd/decompress/ |
D | zstd_decompress_block.c | 661 const BYTE* match; member 773 const BYTE* match = oLitEnd - sequence.offset; in ZSTD_execSequenceEnd() local 791 match = dictEnd - (prefixStart-match); in ZSTD_execSequenceEnd() 792 if (match + sequence.matchLength <= dictEnd) { in ZSTD_execSequenceEnd() 793 ZSTD_memmove(oLitEnd, match, sequence.matchLength); in ZSTD_execSequenceEnd() 797 { size_t const length1 = dictEnd - match; in ZSTD_execSequenceEnd() 798 ZSTD_memmove(oLitEnd, match, length1); in ZSTD_execSequenceEnd() 801 match = prefixStart; in ZSTD_execSequenceEnd() 803 ZSTD_safecopy(op, oend_w, match, sequence.matchLength, ZSTD_overlap_src_before_dst); in ZSTD_execSequenceEnd() 818 const BYTE* match = oLitEnd - sequence.offset; in ZSTD_execSequence() local [all …]
|
/lib/kunit/ |
D | resource.c | 63 int kunit_destroy_resource(struct kunit *test, kunit_resource_match_t match, in kunit_destroy_resource() argument 66 struct kunit_resource *res = kunit_find_resource(test, match, in kunit_destroy_resource()
|
/lib/842/ |
D | 842_compress.c | 383 int i, match, b = 0; in check_template() local 391 match = check_index(p, 2, b >> 1); in check_template() 393 match = check_index(p, 4, b >> 2); in check_template() 395 match = check_index(p, 8, 0); in check_template() 398 if (!match) in check_template()
|