| /lib/lz4/ |
| D | lz4_decompress.c | 123 const BYTE *match; in LZ4_decompress_generic() local 167 match = op - offset; in LZ4_decompress_generic() 168 assert(match <= op); /* check overflow */ in LZ4_decompress_generic() 173 (dict == withPrefix64k || match >= lowPrefix)) { in LZ4_decompress_generic() 175 LZ4_memcpy(op + 0, match + 0, 8); in LZ4_decompress_generic() 176 LZ4_memcpy(op + 8, match + 8, 8); in LZ4_decompress_generic() 177 LZ4_memcpy(op + 16, match + 16, 2); in LZ4_decompress_generic() 291 match = op - offset; in LZ4_decompress_generic() 297 if ((checkOffset) && (unlikely(match + dictSize < lowPrefix))) { in LZ4_decompress_generic() 337 if ((dict == usingExtDict) && (match < lowPrefix)) { in LZ4_decompress_generic() [all …]
|
| D | lz4_compress.c | 246 const BYTE *match; in LZ4_compress_generic() local 265 match = LZ4_getPositionOnHash(h, in LZ4_compress_generic() 270 if (match < (const BYTE *)source) { in LZ4_compress_generic() 284 ? (match < lowRefLimit) in LZ4_compress_generic() 288 : (match + MAX_DISTANCE < ip)) in LZ4_compress_generic() 289 || (LZ4_read32(match + refDelta) in LZ4_compress_generic() 294 while (((ip > anchor) & (match + refDelta > lowLimit)) in LZ4_compress_generic() 295 && (unlikely(ip[-1] == match[refDelta - 1]))) { in LZ4_compress_generic() 297 match--; in LZ4_compress_generic() 331 LZ4_writeLE16(op, (U16)(ip - match)); in LZ4_compress_generic() [all …]
|
| D | lz4hc_compress.c | 126 const BYTE * const match = base + matchIndex; in LZ4HC_InsertAndFindBestMatch() local 128 if (*(match + ml) == *(ip + ml) in LZ4HC_InsertAndFindBestMatch() 129 && (LZ4_read32(match) == LZ4_read32(ip))) { in LZ4HC_InsertAndFindBestMatch() 131 match + MINMATCH, iLimit) + MINMATCH; in LZ4HC_InsertAndFindBestMatch() 135 *matchpos = match; in LZ4HC_InsertAndFindBestMatch() 139 const BYTE * const match = dictBase + matchIndex; in LZ4HC_InsertAndFindBestMatch() local 141 if (LZ4_read32(match) == LZ4_read32(ip)) { in LZ4HC_InsertAndFindBestMatch() 149 match + MINMATCH, vLimit) + MINMATCH; in LZ4HC_InsertAndFindBestMatch() 267 const BYTE * const match, in LZ4HC_encodeSequence() argument 300 LZ4_writeLE16(*op, (U16)(*ip - match)); in LZ4HC_encodeSequence()
|
| /lib/zlib_deflate/ |
| D | deflate.c | 82 static void check_match (deflate_state *s, IPos start, IPos match, 572 register Byte *match; /* matched string */ in longest_match() local 615 match = s->window + cur_match; in longest_match() 624 if (*(ush*)(match+best_len-1) != scan_end || in longest_match() 625 *(ush*)match != scan_start) continue; in longest_match() 636 Assert(scan[2] == match[2], "scan[2]?"); in longest_match() 637 scan++, match++; in longest_match() 639 } while (*(ush*)(scan+=2) == *(ush*)(match+=2) && in longest_match() 640 *(ush*)(scan+=2) == *(ush*)(match+=2) && in longest_match() 641 *(ush*)(scan+=2) == *(ush*)(match+=2) && in longest_match() [all …]
|
| /lib/ |
| D | glob.c | 73 bool match = false, inverted = (*pat == '!'); in glob_match() local 97 match |= (a <= c && c <= b); in glob_match() 100 if (match == inverted) in glob_match()
|
| D | globtest.c | 21 bool match = glob_match(pat, str); in test() local 22 bool success = match == expected; in test() 39 printk(message, pat, str, mismatch + 3*match); in test()
|
| D | test_bpf.c | 2308 bool match = __bpf_match_jmp_cond(dst, (s32)imm, op); in __bpf_emit_jmp_imm() local 2311 insns[i++] = BPF_ALU32_IMM(BPF_MOV, R0, match); in __bpf_emit_jmp_imm() 2315 if (!match) in __bpf_emit_jmp_imm() 2331 bool match = __bpf_match_jmp_cond((s32)dst, (s32)imm, op); in __bpf_emit_jmp32_imm() local 2336 if (!match) in __bpf_emit_jmp32_imm() 2352 bool match = __bpf_match_jmp_cond(dst, src, op); in __bpf_emit_jmp_reg() local 2358 if (!match) in __bpf_emit_jmp_reg() 2374 bool match = __bpf_match_jmp_cond((s32)dst, (s32)src, op); in __bpf_emit_jmp32_reg() local 2380 if (!match) in __bpf_emit_jmp32_reg()
|
| D | Kconfig | 38 Drivers may use these helpers to match the bit indices as described
|
| D | Kconfig.debug | 423 For modules whose split BTF does not match vmlinux, load without
|
| /lib/zstd/compress/ |
| D | zstd_lazy.c | 81 const BYTE* match; in ZSTD_insertDUBT1() local 112 match = mBase + matchIndex; in ZSTD_insertDUBT1() 113 matchLength += ZSTD_count(ip+matchLength, match+matchLength, iend); in ZSTD_insertDUBT1() 115 match = dictBase + matchIndex; in ZSTD_insertDUBT1() 116 …matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iend, dictEnd, prefixStart); in ZSTD_insertDUBT1() 118 match = base + matchIndex; /* preparation for next read of match[matchLength] */ in ZSTD_insertDUBT1() 128 if (match[matchLength] < ip[matchLength]) { /* necessarily within buffer */ in ZSTD_insertDUBT1() 191 const BYTE* match = dictBase + dictMatchIndex; in ZSTD_DUBT_findBetterDictMatch() local 192 …matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iend, dictEnd, prefixStart); in ZSTD_DUBT_findBetterDictMatch() 194 …match = base + dictMatchIndex + dictIndexDelta; /* to prepare for next usage of match[matchLengt… in ZSTD_DUBT_findBetterDictMatch() [all …]
|
| D | zstd_double_fast.c | 318 const BYTE* match = base + matchIndexS; in ZSTD_compressBlock_doubleFast_dictMatchState_generic() local 358 if (MEM_read32(match) == MEM_read32(ip)) { in ZSTD_compressBlock_doubleFast_dictMatchState_generic() 364 match = dictBase + dictMatchIndexS; in ZSTD_compressBlock_doubleFast_dictMatchState_generic() 367 if (match > dictStart && MEM_read32(match) == MEM_read32(ip)) { in ZSTD_compressBlock_doubleFast_dictMatchState_generic() 409 mLength = ZSTD_count_2segments(ip+4, match+4, iend, dictEnd, prefixLowest) + 4; in ZSTD_compressBlock_doubleFast_dictMatchState_generic() 411 …while (((ip>anchor) & (match>dictStart)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /… in ZSTD_compressBlock_doubleFast_dictMatchState_generic() 413 mLength = ZSTD_count(ip+4, match+4, iend) + 4; in ZSTD_compressBlock_doubleFast_dictMatchState_generic() 414 offset = (U32)(ip - match); in ZSTD_compressBlock_doubleFast_dictMatchState_generic() 415 …while (((ip>anchor) & (match>prefixLowest)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; … in ZSTD_compressBlock_doubleFast_dictMatchState_generic() 568 const BYTE* match = matchBase + matchIndex; in ZSTD_compressBlock_doubleFast_extDict_generic() local [all …]
|
| D | zstd_fast.c | 430 const BYTE* match = base + matchIndex; in ZSTD_compressBlock_fast_dictMatchState_generic() local 464 } else if (MEM_read32(match) != MEM_read32(ip)) { in ZSTD_compressBlock_fast_dictMatchState_generic() 471 U32 const offset = (U32)(ip-match); in ZSTD_compressBlock_fast_dictMatchState_generic() 472 mLength = ZSTD_count(ip+4, match+4, iend) + 4; in ZSTD_compressBlock_fast_dictMatchState_generic() 473 while (((ip>anchor) & (match>prefixStart)) in ZSTD_compressBlock_fast_dictMatchState_generic() 474 && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */ in ZSTD_compressBlock_fast_dictMatchState_generic() 587 const BYTE* match = matchBase + matchIndex; in ZSTD_compressBlock_fast_extDict_generic() local 606 (MEM_read32(match) != MEM_read32(ip)) ) { in ZSTD_compressBlock_fast_extDict_generic() 614 … size_t mLength = ZSTD_count_2segments(ip+4, match+4, iend, matchEnd, prefixStart) + 4; in ZSTD_compressBlock_fast_extDict_generic() 615 …while (((ip>anchor) & (match>lowMatchPtr)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; }… in ZSTD_compressBlock_fast_extDict_generic()
|
| D | zstd_opt.c | 431 const BYTE* match; in ZSTD_insertBt1() local 486 match = base + matchIndex; in ZSTD_insertBt1() 487 matchLength += ZSTD_count(ip+matchLength, match+matchLength, iend); in ZSTD_insertBt1() 489 match = dictBase + matchIndex; in ZSTD_insertBt1() 490 …matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iend, dictEnd, prefixStart); in ZSTD_insertBt1() 492 match = base + matchIndex; /* to prepare for next usage of match[matchLength] */ in ZSTD_insertBt1() 505 if (match[matchLength] < ip[matchLength]) { /* necessarily within buffer */ in ZSTD_insertBt1() 662 const BYTE* const match = base + matchIndex3; in ZSTD_insertBtAndGetAllMatches() local 663 mlen = ZSTD_count(ip, match, iLimit); in ZSTD_insertBtAndGetAllMatches() 665 const BYTE* const match = dictBase + matchIndex3; in ZSTD_insertBtAndGetAllMatches() local [all …]
|
| D | zstd_compress_internal.h | 767 ZSTD_count_2segments(const BYTE* ip, const BYTE* match, in ZSTD_count_2segments() argument 770 const BYTE* const vEnd = MIN( ip + (mEnd - match), iEnd); in ZSTD_count_2segments() 771 size_t const matchLength = ZSTD_count(ip, match, vEnd); in ZSTD_count_2segments() 772 if (match + matchLength != mEnd) return matchLength; in ZSTD_count_2segments() 774 DEBUGLOG(7, "distance from match beginning to end dictionary = %zi", mEnd - match); in ZSTD_count_2segments()
|
| /lib/zstd/decompress/ |
| D | zstd_decompress_block.c | 869 const BYTE* match = oLitEnd - sequence.offset; in ZSTD_execSequenceEnd() local 887 match = dictEnd - (prefixStart - match); in ZSTD_execSequenceEnd() 888 if (match + sequence.matchLength <= dictEnd) { in ZSTD_execSequenceEnd() 889 ZSTD_memmove(oLitEnd, match, sequence.matchLength); in ZSTD_execSequenceEnd() 893 { size_t const length1 = dictEnd - match; in ZSTD_execSequenceEnd() 894 ZSTD_memmove(oLitEnd, match, length1); in ZSTD_execSequenceEnd() 897 match = prefixStart; in ZSTD_execSequenceEnd() 900 ZSTD_safecopy(op, oend_w, match, sequence.matchLength, ZSTD_overlap_src_before_dst); in ZSTD_execSequenceEnd() 916 const BYTE* match = oLitEnd - sequence.offset; in ZSTD_execSequenceEndSplitLitBuffer() local 935 match = dictEnd - (prefixStart - match); in ZSTD_execSequenceEndSplitLitBuffer() [all …]
|
| /lib/kunit/ |
| D | resource.c | 63 int kunit_destroy_resource(struct kunit *test, kunit_resource_match_t match, in kunit_destroy_resource() argument 66 struct kunit_resource *res = kunit_find_resource(test, match, in kunit_destroy_resource()
|
| /lib/842/ |
| D | 842_compress.c | 383 int i, match, b = 0; in check_template() local 391 match = check_index(p, 2, b >> 1); in check_template() 393 match = check_index(p, 4, b >> 2); in check_template() 395 match = check_index(p, 8, 0); in check_template() 398 if (!match) in check_template()
|