/lib/lzo/ |
D | lzo1x_decompress_safe.c | 24 #define HAVE_OP(x) ((size_t)(op_end - op) >= (size_t)(x)) 42 unsigned char *op; in lzo1x_decompress_safe() local 52 op = out; in lzo1x_decompress_safe() 98 unsigned char *oe = op + t; in lzo1x_decompress_safe() 100 COPY8(op, ip); in lzo1x_decompress_safe() 101 op += 8; in lzo1x_decompress_safe() 103 COPY8(op, ip); in lzo1x_decompress_safe() 104 op += 8; in lzo1x_decompress_safe() 108 op = oe; in lzo1x_decompress_safe() 115 *op++ = *ip++; in lzo1x_decompress_safe() [all …]
|
D | lzo1x_compress.c | 28 unsigned char *op; in lzo1x_1_do_compress() local 34 op = out; in lzo1x_1_do_compress() 120 op[*state_offset] |= t; in lzo1x_1_do_compress() 121 COPY4(op, ii); in lzo1x_1_do_compress() 122 op += t; in lzo1x_1_do_compress() 124 *op++ = (t - 3); in lzo1x_1_do_compress() 125 COPY8(op, ii); in lzo1x_1_do_compress() 126 COPY8(op + 8, ii + 8); in lzo1x_1_do_compress() 127 op += t; in lzo1x_1_do_compress() 130 *op++ = (t - 3); in lzo1x_1_do_compress() [all …]
|
/lib/zlib_inflate/ |
D | inffast.c | 110 unsigned op; /* code bits, operation, extra bits, or */ in inflate_fast() local 148 op = (unsigned)(this.bits); in inflate_fast() 149 hold >>= op; in inflate_fast() 150 bits -= op; in inflate_fast() 151 op = (unsigned)(this.op); in inflate_fast() 152 if (op == 0) { /* literal */ in inflate_fast() 155 else if (op & 16) { /* length base */ in inflate_fast() 157 op &= 15; /* number of extra bits */ in inflate_fast() 158 if (op) { in inflate_fast() 159 if (bits < op) { in inflate_fast() [all …]
|
D | inftrees.c | 104 this.op = (unsigned char)64; /* invalid code marker */ in zlib_inflate_table() 206 this.op = (unsigned char)0; in zlib_inflate_table() 210 this.op = (unsigned char)(extra[work[sym]]); in zlib_inflate_table() 214 this.op = (unsigned char)(32 + 64); /* end of block */ in zlib_inflate_table() 271 (*table)[low].op = (unsigned char)curr; in zlib_inflate_table() 284 this.op = (unsigned char)64; /* invalid code marker */ in zlib_inflate_table()
|
D | inflate.c | 576 if (this.op && (this.op & 0xf0) == 0) { in zlib_inflate() 580 (BITS(last.bits + last.op) >> last.bits)]; in zlib_inflate() 588 if ((int)(this.op) == 0) { in zlib_inflate() 592 if (this.op & 32) { in zlib_inflate() 596 if (this.op & 64) { in zlib_inflate() 601 state->extra = (unsigned)(this.op) & 15; in zlib_inflate() 618 if ((this.op & 0xf0) == 0) { in zlib_inflate() 622 (BITS(last.bits + last.op) >> last.bits)]; in zlib_inflate() 629 if (this.op & 64) { in zlib_inflate() 635 state->extra = (unsigned)(this.op) & 15; in zlib_inflate()
|
D | inftrees.h | 28 unsigned char op; /* operation, extra bits, table bits */ member
|
/lib/lz4/ |
D | lz4_decompress.c | 86 BYTE *op = (BYTE *) dst; in LZ4_decompress_generic() local 87 BYTE * const oend = op + outputSize; in LZ4_decompress_generic() 107 assert(lowPrefix <= op); in LZ4_decompress_generic() 151 (op <= shortoend))) { in LZ4_decompress_generic() 153 memcpy(op, ip, endOnInput ? 16 : 8); in LZ4_decompress_generic() 154 op += length; ip += length; in LZ4_decompress_generic() 164 match = op - offset; in LZ4_decompress_generic() 165 assert(match <= op); /* check overflow */ in LZ4_decompress_generic() 172 memcpy(op + 0, match + 0, 8); in LZ4_decompress_generic() 173 memcpy(op + 8, match + 8, 8); in LZ4_decompress_generic() [all …]
|
D | lz4_compress.c | 200 BYTE *op = (BYTE *) dest; in LZ4_compress_generic() local 201 BYTE * const olimit = op + maxOutputSize; in LZ4_compress_generic() 304 token = op++; in LZ4_compress_generic() 308 (unlikely(op + litLength + in LZ4_compress_generic() 319 *op++ = 255; in LZ4_compress_generic() 320 *op++ = (BYTE)len; in LZ4_compress_generic() 325 LZ4_wildCopy(op, anchor, op + litLength); in LZ4_compress_generic() 326 op += litLength; in LZ4_compress_generic() 331 LZ4_writeLE16(op, (U16)(ip - match)); in LZ4_compress_generic() 332 op += 2; in LZ4_compress_generic() [all …]
|
D | lz4hc_compress.c | 264 BYTE **op, in LZ4HC_encodeSequence() argument 276 token = (*op)++; in LZ4HC_encodeSequence() 279 && ((*op + (length>>8) in LZ4HC_encodeSequence() 290 *(*op)++ = 255; in LZ4HC_encodeSequence() 291 *(*op)++ = (BYTE)len; in LZ4HC_encodeSequence() 296 LZ4_wildCopy(*op, *anchor, (*op) + length); in LZ4HC_encodeSequence() 297 *op += length; in LZ4HC_encodeSequence() 300 LZ4_writeLE16(*op, (U16)(*ip - match)); in LZ4HC_encodeSequence() 301 *op += 2; in LZ4HC_encodeSequence() 307 && (*op + (length>>8) in LZ4HC_encodeSequence() [all …]
|
/lib/ |
D | atomic64.c | 69 #define ATOMIC64_OP(op, c_op) \ argument 70 void atomic64_##op(s64 a, atomic64_t *v) \ 79 EXPORT_SYMBOL(atomic64_##op); 81 #define ATOMIC64_OP_RETURN(op, c_op) \ argument 82 s64 atomic64_##op##_return(s64 a, atomic64_t *v) \ 93 EXPORT_SYMBOL(atomic64_##op##_return); 95 #define ATOMIC64_FETCH_OP(op, c_op) \ argument 96 s64 atomic64_fetch_##op(s64 a, atomic64_t *v) \ 108 EXPORT_SYMBOL(atomic64_fetch_##op); 110 #define ATOMIC64_OPS(op, c_op) \ argument [all …]
|
D | asn1_decoder.c | 175 enum asn1_opcode op; in asn1_ber_decoder() local 205 op = machine[pc]; in asn1_ber_decoder() 206 if (unlikely(pc + asn1_op_lengths[op] > machlen)) in asn1_ber_decoder() 212 if (op <= ASN1_OP__MATCHES_TAG) { in asn1_ber_decoder() 216 if ((op & ASN1_OP_MATCH__COND && flags & FLAG_MATCHED) || in asn1_ber_decoder() 217 (op & ASN1_OP_MATCH__SKIP && dp == datalen)) { in asn1_ber_decoder() 219 pc += asn1_op_lengths[op]; in asn1_ber_decoder() 233 if (op & ASN1_OP_MATCH__ANY) { in asn1_ber_decoder() 250 if (op & ASN1_OP_MATCH__SKIP) { in asn1_ber_decoder() 251 pc += asn1_op_lengths[op]; in asn1_ber_decoder() [all …]
|
D | atomic64_test.c | 20 #define TEST(bit, op, c_op, val) \ argument 24 atomic##bit##_##op(val, &v); \ 36 #define FAMILY_TEST(test, bit, op, args...) \ argument 38 test(bit, op, ##args); \ 39 test(bit, op##_acquire, ##args); \ 40 test(bit, op##_release, ##args); \ 41 test(bit, op##_relaxed, ##args); \ 44 #define TEST_RETURN(bit, op, c_op, val) \ argument 49 BUG_ON(atomic##bit##_##op(val, &v) != r); \ 53 #define TEST_FETCH(bit, op, c_op, val) \ argument [all …]
|
D | packing.c | 89 enum packing_op op, u8 quirks) in packing() argument 113 if (op == PACK && value_width < 64 && (*uval >= (1ull << value_width))) in packing() 121 if (op == UNPACK) in packing() 176 if (op == UNPACK) { in packing()
|
D | smp_processor_id.c | 62 notrace void __this_cpu_preempt_check(const char *op) in __this_cpu_preempt_check() argument 64 check_preemption_disabled("__this_cpu_", op); in __this_cpu_preempt_check()
|
D | ubsan.c | 162 void *rhs, char op) in handle_overflow() argument 180 op, in handle_overflow()
|
D | dynamic_debug.c | 399 int op = '=', i; in ddebug_parse_flags() local 405 op = *str++; in ddebug_parse_flags() 411 vpr_info("op='%c'\n", op); in ddebug_parse_flags() 428 switch (op) { in ddebug_parse_flags()
|
D | test_overflow.c | 218 #define check_one_op(t, fmt, op, sym, a, b, r, of) do { \ argument 222 _of = check_ ## op ## _overflow(a, b, &_r); \
|
/lib/zstd/ |
D | huf_compress.c | 85 BYTE *op = ostart; in HUF_compressWeights_wksp() local 128 CHECK_V_F(hSize, FSE_writeNCount(op, oend - op, norm, maxSymbolValue, tableLog)); in HUF_compressWeights_wksp() 129 op += hSize; in HUF_compressWeights_wksp() 135 CHECK_V_F(cSize, FSE_compress_usingCTable(op, oend - op, weightTable, wtSize, CTable)); in HUF_compressWeights_wksp() 138 op += cSize; in HUF_compressWeights_wksp() 141 return op - ostart; in HUF_compressWeights_wksp() 154 BYTE *op = (BYTE *)dst; in HUF_writeCTable_wksp() local 184 …CHECK_V_F(hSize, HUF_compressWeights_wksp(op + 1, maxDstSize - 1, huffWeight, maxSymbolValue, work… in HUF_writeCTable_wksp() 186 op[0] = (BYTE)hSize; in HUF_writeCTable_wksp() 196 op[0] = (BYTE)(128 /*special case*/ + (maxSymbolValue - 1)); in HUF_writeCTable_wksp() [all …]
|
D | fse_decompress.c | 218 BYTE *op = ostart; in FSE_decompress_usingDTable_generic() local 219 BYTE *const omax = op + maxDstSize; in FSE_decompress_usingDTable_generic() 235 for (; (BIT_reloadDStream(&bitD) == BIT_DStream_unfinished) & (op < olimit); op += 4) { in FSE_decompress_usingDTable_generic() 236 op[0] = FSE_GETSYMBOL(&state1); in FSE_decompress_usingDTable_generic() 241 op[1] = FSE_GETSYMBOL(&state2); in FSE_decompress_usingDTable_generic() 246 op += 2; in FSE_decompress_usingDTable_generic() 251 op[2] = FSE_GETSYMBOL(&state1); in FSE_decompress_usingDTable_generic() 256 op[3] = FSE_GETSYMBOL(&state2); in FSE_decompress_usingDTable_generic() 262 if (op > (omax - 2)) in FSE_decompress_usingDTable_generic() 264 *op++ = FSE_GETSYMBOL(&state1); in FSE_decompress_usingDTable_generic() [all …]
|
D | decompress.c | 880 size_t ZSTD_execSequenceLast7(BYTE *op, BYTE *const oend, seq_t sequence, const BYTE **litPtr, cons… in ZSTD_execSequenceLast7() argument 883 BYTE *const oLitEnd = op + sequence.litLength; in ZSTD_execSequenceLast7() 885 BYTE *const oMatchEnd = op + sequenceLength; /* risk : address space overflow (32-bits) */ in ZSTD_execSequenceLast7() 899 if (op < oend_w) { in ZSTD_execSequenceLast7() 900 ZSTD_wildcopy(op, *litPtr, oend_w - op); in ZSTD_execSequenceLast7() 901 *litPtr += oend_w - op; in ZSTD_execSequenceLast7() 902 op = oend_w; in ZSTD_execSequenceLast7() 904 while (op < oLitEnd) in ZSTD_execSequenceLast7() 905 *op++ = *(*litPtr)++; in ZSTD_execSequenceLast7() 921 op = oLitEnd + length1; in ZSTD_execSequenceLast7() [all …]
|
D | zstd_internal.h | 138 BYTE* op = (BYTE*)dst; in ZSTD_wildcopy() local 139 BYTE* const oend = op + length; in ZSTD_wildcopy() 148 ZSTD_copy8(op, ip); in ZSTD_wildcopy() 149 op += 8; in ZSTD_wildcopy() 151 } while (op < oend); in ZSTD_wildcopy()
|
D | compress.c | 600 BYTE *op = ostart; in ZSTD_compressSequences_internal() local 623 size_t const cSize = ZSTD_compressLiterals(zc, op, dstCapacity, literals, litSize); in ZSTD_compressSequences_internal() 626 op += cSize; in ZSTD_compressSequences_internal() 630 if ((oend - op) < 3 /*max nbSeq Size*/ + 1 /*seqHead */) in ZSTD_compressSequences_internal() 633 *op++ = (BYTE)nbSeq; in ZSTD_compressSequences_internal() 635 op[0] = (BYTE)((nbSeq >> 8) + 0x80), op[1] = (BYTE)nbSeq, op += 2; in ZSTD_compressSequences_internal() 637 op[0] = 0xFF, ZSTD_writeLE16(op + 1, (U16)(nbSeq - LONGNBSEQ)), op += 3; in ZSTD_compressSequences_internal() 639 return op - ostart; in ZSTD_compressSequences_internal() 642 seqHead = op++; in ZSTD_compressSequences_internal() 655 *op++ = llCodeTable[0]; in ZSTD_compressSequences_internal() [all …]
|
D | huf_decompress.c | 201 BYTE *op = (BYTE *)dst; in HUF_decompress1X2_usingDTable_internal() local 202 BYTE *const oend = op + dstSize; in HUF_decompress1X2_usingDTable_internal() 215 HUF_decodeStreamX2(op, &bitD, oend, dt, dtLog); in HUF_decompress1X2_usingDTable_internal() 595 static U32 HUF_decodeSymbolX4(void *op, BIT_DStream_t *DStream, const HUF_DEltX4 *dt, const U32 dtL… in HUF_decodeSymbolX4() argument 598 memcpy(op, dt + val, 2); in HUF_decodeSymbolX4() 603 static U32 HUF_decodeLastSymbolX4(void *op, BIT_DStream_t *DStream, const HUF_DEltX4 *dt, const U32… in HUF_decodeLastSymbolX4() argument 606 memcpy(op, dt + val, 1); in HUF_decodeLastSymbolX4()
|
/lib/842/ |
D | 842_decompress.c | 234 u8 op = decomp_ops[o][i]; in do_op() local 236 pr_debug("op is %x\n", op); in do_op() 238 switch (op & OP_ACTION) { in do_op() 240 ret = do_data(p, op & OP_AMOUNT); in do_op() 243 ret = do_index(p, op & OP_AMOUNT); in do_op() 248 pr_err("Internal error, invalid op %x\n", op); in do_op() 282 u64 op, rep, tmp, bytes, total; in sw842_decompress() local 297 ret = next_bits(&p, &op, OP_BITS); in sw842_decompress() 301 pr_debug("template is %lx\n", (unsigned long)op); in sw842_decompress() 303 switch (op) { in sw842_decompress() [all …]
|