/lib/ |
D | test-string_helpers.c | 15 char *in, size_t p, in test_string_check_buf() argument 25 in, p, true); in test_string_check_buf() 35 const char *in; member 42 .in = "\\f\\ \\n\\r\\t\\v", 47 .in = "\\40\\1\\387\\0064\\05\\040\\8a\\110\\777", 52 .in = "\\xv\\xa\\x2c\\xD\\x6f2", 57 .in = "\\h\\\\\\\"\\a\\e\\", 67 char *in = kmalloc(q_real, GFP_KERNEL); in test_string_unescape() local 72 if (!in || !out_test || !out_real) in test_string_unescape() 76 const char *s = strings[i].in; in test_string_unescape() [all …]
|
D | kfifo.c | 21 return (fifo->mask + 1) - (fifo->in - fifo->out); in kfifo_unused() 33 fifo->in = 0; in __kfifo_alloc() 58 fifo->in = 0; in __kfifo_free() 74 fifo->in = 0; in __kfifo_init() 122 kfifo_copy_in(fifo, buf, len, fifo->in); in __kfifo_in() 123 fifo->in += len; in __kfifo_in() 157 l = fifo->in - fifo->out; in __kfifo_out_peek() 225 ret = kfifo_copy_from_user(fifo, from, len, fifo->in, copied); in __kfifo_from_user() 231 fifo->in += len; in __kfifo_from_user() 281 l = fifo->in - fifo->out; in __kfifo_to_user() [all …]
|
D | decompress_unxz.c | 251 STATIC int INIT unxz(unsigned char *in, long in_size, in unxz() argument 287 if (in == NULL) { in unxz() 289 in = malloc(XZ_IOBUF_SIZE); in unxz() 290 if (in == NULL) in unxz() 294 b.in = in; in unxz() 309 in_size = fill(in, XZ_IOBUF_SIZE); in unxz() 340 free(in); in unxz()
|
D | digsig.c | 78 MPI in = NULL, res = NULL, pkey[2]; in digsig_verify_rsa() local 135 in = mpi_read_from_buffer(sig, &nret); in digsig_verify_rsa() 136 if (IS_ERR(in)) { in digsig_verify_rsa() 137 err = PTR_ERR(in); in digsig_verify_rsa() 141 res = mpi_alloc(mpi_get_nlimbs(in) * 2); in digsig_verify_rsa() 145 err = mpi_powm(res, in, pkey[1], pkey[0]); in digsig_verify_rsa() 173 mpi_free(in); in digsig_verify_rsa()
|
D | test_siphash.c | 115 u8 in[64] __aligned(SIPHASH_ALIGNMENT); in siphash_test_init() local 121 in[i] = i; in siphash_test_init() 123 if (siphash(in, i, &test_key_siphash) != in siphash_test_init() 133 if (hsiphash(in, i, &test_key_hsiphash) != in siphash_test_init()
|
D | sg_split.c | 20 static int sg_calculate_split(struct scatterlist *in, int nents, int nb_splits, in sg_calculate_split() argument 35 for_each_sg(in, sg, nents, i) { in sg_calculate_split() 148 int sg_split(struct scatterlist *in, const int in_mapped_nents, in sg_split() argument 161 ret = sg_calculate_split(in, sg_nents(in), nb_splits, skip, split_sizes, in sg_split() 180 ret = sg_calculate_split(in, in_mapped_nents, nb_splits, skip, in sg_split()
|
D | Kconfig.kgdb | 14 CONFIG_FRAME_POINTER to aid in producing more reliable stack 15 backtraces in the external debugger. Documentation of 17 as well as in Documentation/dev-tools/kgdb.rst. If 30 to break in initially. 50 to pass in a kernel parameter 63 bool "KGDB: Allow debugging with traps in notifiers" 121 kernel is still usable in this situation. 123 No guarantees that the kernel is still usable in this situation.
|
D | test_bitmap.c | 213 const char *in; member 298 size_t len = strlen(ptest.in); in __test_bitmap_parselist() 302 err = bitmap_parselist_user(ptest.in, len, in __test_bitmap_parselist() 308 err = bitmap_parselist(ptest.in, bmap, ptest.nbits); in __test_bitmap_parselist() 314 mode, i, ptest.in, err, ptest.errno); in __test_bitmap_parselist() 321 mode, i, ptest.in, bmap[0], in __test_bitmap_parselist() 328 mode, i, ptest.in, time); in __test_bitmap_parselist()
|
D | Kconfig.debug | 26 in task context) or a caller "processor id" (if not in task context) 45 Setting a default here is equivalent to passing in loglevel=<x> in 50 usage in the kernel. That is controlled by the MESSAGE_LOGLEVEL_DEFAULT 76 by default. To change that, use loglevel=<x> in the kernel bootargs, 85 specified in milliseconds on the kernel command line, 108 implicitly compiles in all pr_debug() and dev_dbg() calls, which 112 pr_debug() calls in it are enabled by default, but can be 119 which is contained in the 'debugfs' filesystem. Thus, the debugfs 148 // enable all the messages in file svcsock.c 152 // enable all the messages in the NFS server module [all …]
|
D | Kconfig | 35 in the data sheets of the peripherals they are in control of. 37 When in doubt, say N. 84 bool "Access I/O in non-MMIO mode" 88 hosts which can not be accessed in MMIO mode. Using the logical PIO 97 When in doubt, say N. 102 This option is provided for the case where no in-kernel-tree 110 This option is provided for the case where no in-kernel-tree 120 This option is only needed if a module that's not in the 127 This option is provided for the case where no in-kernel-tree 137 This option is provided for the case where no in-kernel-tree [all …]
|
D | Kconfig.ubsan | 10 behaviours in runtime. Various types of checks may be enabled 20 # use -Wmaybe-uninitilized in allmodconfig builds. 21 # So dependsy bellow used to disable this option in allmodconfig
|
D | chacha.c | 102 void hchacha_block(const u32 *in, u32 *out, int nrounds) in hchacha_block() argument 106 memcpy(x, in, 64); in hchacha_block()
|
D | Kconfig.kasan | 46 Supported in both GCC and Clang. With GCC it requires version 4.9.2 100 memory accesses. This is faster than outline (in some workloads 112 causes excessive stack usage in a lot of functions, see 118 to avoid cluttering the output in stack overflow warnings,
|
/lib/crypto/ |
D | aes.c | 147 static __always_inline u32 subshift(u32 in[], int pos) in subshift() argument 149 return (aes_sbox[in[pos] & 0xff]) ^ in subshift() 150 (aes_sbox[(in[(pos + 1) % 4] >> 8) & 0xff] << 8) ^ in subshift() 151 (aes_sbox[(in[(pos + 2) % 4] >> 16) & 0xff] << 16) ^ in subshift() 152 (aes_sbox[(in[(pos + 3) % 4] >> 24) & 0xff] << 24); in subshift() 155 static __always_inline u32 inv_subshift(u32 in[], int pos) in inv_subshift() argument 157 return (aes_inv_sbox[in[pos] & 0xff]) ^ in inv_subshift() 158 (aes_inv_sbox[(in[(pos + 3) % 4] >> 8) & 0xff] << 8) ^ in inv_subshift() 159 (aes_inv_sbox[(in[(pos + 2) % 4] >> 16) & 0xff] << 16) ^ in inv_subshift() 160 (aes_inv_sbox[(in[(pos + 1) % 4] >> 24) & 0xff] << 24); in inv_subshift() [all …]
|
D | arc4.c | 37 void arc4_crypt(struct arc4_ctx *ctx, u8 *out, const u8 *in, unsigned int len) in arc4_crypt() argument 61 *out++ = *in++ ^ S[a]; in arc4_crypt()
|
/lib/zlib_inflate/ |
D | inffast.c | 91 const unsigned char *in; /* local strm->next_in */ in inflate_fast() local 118 in = strm->next_in - OFF; in inflate_fast() 119 last = in + (strm->avail_in - 5); in inflate_fast() 141 hold += (unsigned long)(PUP(in)) << bits; in inflate_fast() 143 hold += (unsigned long)(PUP(in)) << bits; in inflate_fast() 160 hold += (unsigned long)(PUP(in)) << bits; in inflate_fast() 168 hold += (unsigned long)(PUP(in)) << bits; in inflate_fast() 170 hold += (unsigned long)(PUP(in)) << bits; in inflate_fast() 183 hold += (unsigned long)(PUP(in)) << bits; in inflate_fast() 186 hold += (unsigned long)(PUP(in)) << bits; in inflate_fast() [all …]
|
/lib/lzo/ |
D | lzo1x_compress.c | 22 lzo1x_1_do_compress(const unsigned char *in, size_t in_len, in lzo1x_1_do_compress() argument 29 const unsigned char * const in_end = in + in_len; in lzo1x_1_do_compress() 30 const unsigned char * const ip_end = in + in_len - 20; in lzo1x_1_do_compress() 35 ip = in; in lzo1x_1_do_compress() 109 m_pos = in + dict[t]; in lzo1x_1_do_compress() 110 dict[t] = (lzo_dict_t) (ip - in); in lzo1x_1_do_compress() 291 int lzogeneric1x_1_compress(const unsigned char *in, size_t in_len, in lzogeneric1x_1_compress() argument 295 const unsigned char *ip = in; in lzogeneric1x_1_compress() 331 const unsigned char *ii = in + in_len - t; in lzogeneric1x_1_compress() 368 int lzo1x_1_compress(const unsigned char *in, size_t in_len, in lzo1x_1_compress() argument [all …]
|
D | lzo1x_decompress_safe.c | 39 int lzo1x_decompress_safe(const unsigned char *in, size_t in_len, in lzo1x_decompress_safe() argument 47 const unsigned char * const ip_end = in + in_len; in lzo1x_decompress_safe() 53 ip = in; in lzo1x_decompress_safe()
|
/lib/842/ |
D | 842_decompress.c | 51 u8 *in; member 89 u8 *in = p->in, b = p->bit, bits = b + n; in next_bits() local 110 *d = *in >> (8 - bits); in next_bits() 112 *d = be16_to_cpu(get_unaligned((__be16 *)in)) >> (16 - bits); in next_bits() 114 *d = be32_to_cpu(get_unaligned((__be32 *)in)) >> (32 - bits); in next_bits() 116 *d = be64_to_cpu(get_unaligned((__be64 *)in)) >> (64 - bits); in next_bits() 123 p->in += p->bit / 8; in next_bits() 277 int sw842_decompress(const u8 *in, unsigned int ilen, in sw842_decompress() argument 285 p.in = (u8 *)in; in sw842_decompress()
|
D | 842_compress.c | 89 u8 *in; member 110 be##b##_to_cpu(get_unaligned((__be##b *)((p)->in + (o)))) 145 (unsigned int)((p)->in - (p)->instart), \ 343 ret = add_bits(p, p->in[i], 8); in add_short_data_template() 425 u64 pos = p->in - p->instart; in update_hashtables() 477 int sw842_compress(const u8 *in, unsigned int ilen, in sw842_compress() argument 492 p->in = (u8 *)in; in sw842_compress() 493 p->instart = p->in; in sw842_compress() 514 last = ~get_unaligned((u64 *)p->in); in sw842_compress() 517 next = get_unaligned((u64 *)p->in); in sw842_compress() [all …]
|
/lib/xz/ |
D | xz_dec_lzma2.c | 109 const uint8_t *in; member 390 memcpy(dict->buf + dict->pos, b->in + b->in_pos, copy_size); in dict_uncompressed() 400 memcpy(b->out + b->out_pos, b->in + b->in_pos, in dict_uncompressed() 455 rc->code = (rc->code << 8) + b->in[b->in_pos++]; in rc_read_init() 482 rc->code = (rc->code << RC_SHIFT_BITS) + rc->in[rc->in_pos++]; in rc_normalize() 859 memcpy(s->temp.buf + s->temp.size, b->in + b->in_pos, tmp); in lzma2_lzma() 874 s->rc.in = s->temp.buf; in lzma2_lzma() 895 s->rc.in = b->in; in lzma2_lzma() 919 memcpy(s->temp.buf, b->in + b->in_pos, in_avail); in lzma2_lzma() 970 tmp = b->in[b->in_pos++]; in xz_dec_lzma2_run() [all …]
|
D | xz_dec_stream.c | 162 memcpy(s->temp.buf + s->temp.pos, b->in + b->in_pos, copy_size); in fill_temp() 175 static enum xz_ret dec_vli(struct xz_dec *s, const uint8_t *in, in dec_vli() argument 184 byte = in[*in_pos]; in dec_vli() 285 s->crc32 = xz_crc32(b->in + s->in_start, in_used, s->crc32); in index_update() 301 ret = dec_vli(s, b->in, &b->in_pos, b->in_size); in dec_index() 352 if (((s->crc32 >> s->pos) & 0xFF) != b->in[b->in_pos++]) in crc32_validate() 594 if (b->in[b->in_pos] == 0) { in dec_main() 605 = ((uint32_t)b->in[b->in_pos] + 1) * 4; in dec_main() 646 if (b->in[b->in_pos++] != 0) in dec_main() 688 if (b->in[b->in_pos++] != 0) in dec_main()
|
/lib/raid6/ |
D | s390vx.uc | 28 * The SHLBYTE() operation shifts each of the 16 bytes in 29 * vector register y left by 1 bit and stores the result in 38 * For each of the 16 bytes in the vector register y the MASK() 40 * or 0x00 if the high bit is 0. The result is stored in vector
|
D | altivec.uc | 20 * <benh> hpa: in process, 22 * bracked this with preempt_disable/enable or in a lock) 56 * The MASK() operation returns 0xFF in any byte for which the high
|
/lib/math/ |
D | Kconfig | 6 calculations are in fixed point. Module will be called cordic.
|