/lib/lz4/ |
D | lz4_decompress.c | 59 BYTE *op = (BYTE *) dest; in lz4_uncompress() local 60 BYTE * const oend = op + osize; in lz4_uncompress() 82 cpy = op + length; in lz4_uncompress() 91 memcpy(op, ip, length); in lz4_uncompress() 95 LZ4_WILDCOPY(ip, op, cpy); in lz4_uncompress() 96 ip -= (op - cpy); in lz4_uncompress() 97 op = cpy; in lz4_uncompress() 118 if (unlikely((op - ref) < STEPSIZE)) { in lz4_uncompress() 120 int dec64 = dec64table[op - ref]; in lz4_uncompress() 124 op[0] = ref[0]; in lz4_uncompress() [all …]
|
D | lz4_compress.c | 70 u8 *op = (u8 *) dest; in lz4_compressctx() local 71 u8 *const oend = op + maxoutputsize; in lz4_compressctx() 119 token = op++; in lz4_compressctx() 121 if (unlikely(op + length + (2 + 1 + LASTLITERALS) + in lz4_compressctx() 130 *op++ = 255; in lz4_compressctx() 131 *op++ = (u8)len; in lz4_compressctx() 136 LZ4_BLINDCOPY(anchor, op, length); in lz4_compressctx() 139 LZ4_WRITE_LITTLEENDIAN_16(op, (u16)(ip - ref)); in lz4_compressctx() 176 if (unlikely(op + (1 + LASTLITERALS) + (length >> 8) > oend)) in lz4_compressctx() 182 *op++ = 255; in lz4_compressctx() [all …]
|
D | lz4hc_compress.c | 269 static inline int lz4_encodesequence(const u8 **ip, u8 **op, const u8 **anchor, in lz4_encodesequence() argument 277 token = (*op)++; in lz4_encodesequence() 282 *(*op)++ = 255; in lz4_encodesequence() 283 *(*op)++ = (u8)len; in lz4_encodesequence() 288 LZ4_BLINDCOPY(*anchor, *op, length); in lz4_encodesequence() 291 LZ4_WRITE_LITTLEENDIAN_16(*op, (u16)(*ip - ref)); in lz4_encodesequence() 299 *(*op)++ = 255; in lz4_encodesequence() 300 *(*op)++ = 255; in lz4_encodesequence() 304 *(*op)++ = 255; in lz4_encodesequence() 306 *(*op)++ = (u8)len; in lz4_encodesequence() [all …]
|
/lib/lzo/ |
D | lzo1x_decompress_safe.c | 23 #define HAVE_OP(x) ((size_t)(op_end - op) >= (size_t)(x)) 41 unsigned char *op; in lzo1x_decompress_safe() local 49 op = out; in lzo1x_decompress_safe() 87 unsigned char *oe = op + t; in lzo1x_decompress_safe() 89 COPY8(op, ip); in lzo1x_decompress_safe() 90 op += 8; in lzo1x_decompress_safe() 92 COPY8(op, ip); in lzo1x_decompress_safe() 93 op += 8; in lzo1x_decompress_safe() 97 op = oe; in lzo1x_decompress_safe() 104 *op++ = *ip++; in lzo1x_decompress_safe() [all …]
|
D | lzo1x_compress.c | 26 unsigned char *op; in lzo1x_1_do_compress() local 32 op = out; in lzo1x_1_do_compress() 58 op[-2] |= t; in lzo1x_1_do_compress() 59 COPY4(op, ii); in lzo1x_1_do_compress() 60 op += t; in lzo1x_1_do_compress() 62 *op++ = (t - 3); in lzo1x_1_do_compress() 63 COPY8(op, ii); in lzo1x_1_do_compress() 64 COPY8(op + 8, ii + 8); in lzo1x_1_do_compress() 65 op += t; in lzo1x_1_do_compress() 68 *op++ = (t - 3); in lzo1x_1_do_compress() [all …]
|
/lib/zlib_inflate/ |
D | inffast.c | 110 unsigned op; /* code bits, operation, extra bits, or */ in inflate_fast() local 148 op = (unsigned)(this.bits); in inflate_fast() 149 hold >>= op; in inflate_fast() 150 bits -= op; in inflate_fast() 151 op = (unsigned)(this.op); in inflate_fast() 152 if (op == 0) { /* literal */ in inflate_fast() 155 else if (op & 16) { /* length base */ in inflate_fast() 157 op &= 15; /* number of extra bits */ in inflate_fast() 158 if (op) { in inflate_fast() 159 if (bits < op) { in inflate_fast() [all …]
|
D | inftrees.c | 104 this.op = (unsigned char)64; /* invalid code marker */ in zlib_inflate_table() 206 this.op = (unsigned char)0; in zlib_inflate_table() 210 this.op = (unsigned char)(extra[work[sym]]); in zlib_inflate_table() 214 this.op = (unsigned char)(32 + 64); /* end of block */ in zlib_inflate_table() 271 (*table)[low].op = (unsigned char)curr; in zlib_inflate_table() 284 this.op = (unsigned char)64; /* invalid code marker */ in zlib_inflate_table()
|
D | inflate.c | 569 if (this.op && (this.op & 0xf0) == 0) { in zlib_inflate() 573 (BITS(last.bits + last.op) >> last.bits)]; in zlib_inflate() 581 if ((int)(this.op) == 0) { in zlib_inflate() 585 if (this.op & 32) { in zlib_inflate() 589 if (this.op & 64) { in zlib_inflate() 594 state->extra = (unsigned)(this.op) & 15; in zlib_inflate() 609 if ((this.op & 0xf0) == 0) { in zlib_inflate() 613 (BITS(last.bits + last.op) >> last.bits)]; in zlib_inflate() 620 if (this.op & 64) { in zlib_inflate() 626 state->extra = (unsigned)(this.op) & 15; in zlib_inflate()
|
D | inftrees.h | 28 unsigned char op; /* operation, extra bits, table bits */ member
|
/lib/ |
D | asn1_decoder.c | 173 enum asn1_opcode op; in asn1_ber_decoder() local 202 op = machine[pc]; in asn1_ber_decoder() 203 if (unlikely(pc + asn1_op_lengths[op] > machlen)) in asn1_ber_decoder() 209 if (op <= ASN1_OP__MATCHES_TAG) { in asn1_ber_decoder() 213 if ((op & ASN1_OP_MATCH__COND && flags & FLAG_MATCHED) || in asn1_ber_decoder() 214 (op & ASN1_OP_MATCH__SKIP && dp == datalen)) { in asn1_ber_decoder() 215 pc += asn1_op_lengths[op]; in asn1_ber_decoder() 229 if (op & ASN1_OP_MATCH__ANY) { in asn1_ber_decoder() 246 if (op & ASN1_OP_MATCH__SKIP) { in asn1_ber_decoder() 247 pc += asn1_op_lengths[op]; in asn1_ber_decoder() [all …]
|
D | atomic64.c | 73 #define ATOMIC64_OP(op, c_op) \ argument 74 void atomic64_##op(long long a, atomic64_t *v) \ 83 EXPORT_SYMBOL(atomic64_##op); 85 #define ATOMIC64_OP_RETURN(op, c_op) \ argument 86 long long atomic64_##op##_return(long long a, atomic64_t *v) \ 97 EXPORT_SYMBOL(atomic64_##op##_return); 99 #define ATOMIC64_OPS(op, c_op) \ argument 100 ATOMIC64_OP(op, c_op) \ 101 ATOMIC64_OP_RETURN(op, c_op)
|
D | smp_processor_id.c | 60 notrace void __this_cpu_preempt_check(const char *op) in __this_cpu_preempt_check() argument 62 check_preemption_disabled("__this_cpu_", op); in __this_cpu_preempt_check()
|
D | dynamic_debug.c | 390 int op = '=', i; in ddebug_parse_flags() local 396 op = *str++; in ddebug_parse_flags() 402 vpr_info("op='%c'\n", op); in ddebug_parse_flags() 419 switch (op) { in ddebug_parse_flags()
|