/crypto/ |
D | aegis128-core.c | 77 union aegis_block tmp; in crypto_aegis128_update() local 80 tmp = state->blocks[AEGIS128_STATE_BLOCKS - 1]; in crypto_aegis128_update() 84 crypto_aegis_aesenc(&state->blocks[0], &tmp, &state->blocks[0]); in crypto_aegis128_update() 161 union aegis_block tmp; in crypto_aegis128_encrypt_chunk() local 170 tmp = state->blocks[2]; in crypto_aegis128_encrypt_chunk() 171 crypto_aegis_block_and(&tmp, &state->blocks[3]); in crypto_aegis128_encrypt_chunk() 172 crypto_aegis_block_xor(&tmp, &state->blocks[4]); in crypto_aegis128_encrypt_chunk() 173 crypto_aegis_block_xor(&tmp, &state->blocks[1]); in crypto_aegis128_encrypt_chunk() 174 crypto_aegis_block_xor(&tmp, src_blk); in crypto_aegis128_encrypt_chunk() 178 *dst_blk = tmp; in crypto_aegis128_encrypt_chunk() [all …]
|
D | ecc.c | 603 u64 tmp[2 * ECC_MAX_DIGITS]; in vli_mmod_slow() local 604 u64 *v[2] = { tmp, product }; in vli_mmod_slow() 676 const u64 *curve_prime, u64 *tmp) in vli_mmod_fast_192() argument 683 vli_set(tmp, &product[3], ndigits); in vli_mmod_fast_192() 684 carry = vli_add(result, result, tmp, ndigits); in vli_mmod_fast_192() 686 tmp[0] = 0; in vli_mmod_fast_192() 687 tmp[1] = product[3]; in vli_mmod_fast_192() 688 tmp[2] = product[4]; in vli_mmod_fast_192() 689 carry += vli_add(result, result, tmp, ndigits); in vli_mmod_fast_192() 691 tmp[0] = tmp[1] = product[5]; in vli_mmod_fast_192() [all …]
|
D | jitterentropy.c | 204 __u64 tmp = time << (DATA_SIZE_BITS - i); in jent_lfsr_time() local 206 tmp = tmp >> (DATA_SIZE_BITS - 1); in jent_lfsr_time() 218 tmp ^= ((new >> 63) & 1); in jent_lfsr_time() 219 tmp ^= ((new >> 60) & 1); in jent_lfsr_time() 220 tmp ^= ((new >> 55) & 1); in jent_lfsr_time() 221 tmp ^= ((new >> 30) & 1); in jent_lfsr_time() 222 tmp ^= ((new >> 27) & 1); in jent_lfsr_time() 223 tmp ^= ((new >> 22) & 1); in jent_lfsr_time() 225 new ^= tmp; in jent_lfsr_time()
|
D | authencesn.c | 109 u32 tmp[2]; in crypto_authenc_esn_genicv_tail() local 112 scatterwalk_map_and_copy(tmp, dst, 4, 4, 0); in crypto_authenc_esn_genicv_tail() 113 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0); in crypto_authenc_esn_genicv_tail() 114 scatterwalk_map_and_copy(tmp, dst, 0, 8, 1); in crypto_authenc_esn_genicv_tail() 143 u32 tmp[2]; in crypto_authenc_esn_genicv() local 149 scatterwalk_map_and_copy(tmp, dst, 0, 8, 0); in crypto_authenc_esn_genicv() 150 scatterwalk_map_and_copy(tmp, dst, 4, 4, 1); in crypto_authenc_esn_genicv() 151 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 1); in crypto_authenc_esn_genicv() 245 u32 tmp[2]; in crypto_authenc_esn_decrypt_tail() local 251 scatterwalk_map_and_copy(tmp, dst, 4, 4, 0); in crypto_authenc_esn_decrypt_tail() [all …]
|
D | jitterentropy-kcapi.c | 95 __u64 tmp = 0; in jent_get_nstime() local 97 tmp = random_get_entropy(); in jent_get_nstime() 104 if (tmp == 0) in jent_get_nstime() 105 tmp = ktime_get_ns(); in jent_get_nstime() 107 *out = tmp; in jent_get_nstime()
|
D | cfb.c | 46 u8 tmp[MAX_CIPHER_BLOCKSIZE + MAX_CIPHER_ALIGNMASK]; in crypto_cfb_final() local 47 u8 *stream = PTR_ALIGN(tmp + 0, alignmask + 1); in crypto_cfb_final() 87 u8 tmp[MAX_CIPHER_BLOCKSIZE]; in crypto_cfb_encrypt_inplace() local 90 crypto_cfb_encrypt_one(tfm, iv, tmp); in crypto_cfb_encrypt_inplace() 91 crypto_xor(src, tmp, bsize); in crypto_cfb_encrypt_inplace() 157 u8 tmp[MAX_CIPHER_BLOCKSIZE]; in crypto_cfb_decrypt_inplace() local 160 crypto_cfb_encrypt_one(tfm, iv, tmp); in crypto_cfb_decrypt_inplace() 162 crypto_xor(src, tmp, bsize); in crypto_cfb_decrypt_inplace()
|
D | cipher.c | 67 u8 *tmp = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); in cipher_crypt_unaligned() local 69 memcpy(tmp, src, size); in cipher_crypt_unaligned() 70 fn(tfm, tmp, tmp); in cipher_crypt_unaligned() 71 memcpy(dst, tmp, size); in cipher_crypt_unaligned()
|
D | ansi_cprng.c | 83 unsigned char tmp[DEFAULT_BLK_SZ]; in _get_more_prng_bytes() local 105 memcpy(tmp, ctx->DT, DEFAULT_BLK_SZ); in _get_more_prng_bytes() 107 hexdump("tmp stage 0: ", tmp, DEFAULT_BLK_SZ); in _get_more_prng_bytes() 116 xor_vectors(ctx->I, ctx->V, tmp, DEFAULT_BLK_SZ); in _get_more_prng_bytes() 117 hexdump("tmp stage 1: ", tmp, DEFAULT_BLK_SZ); in _get_more_prng_bytes() 146 xor_vectors(ctx->rand_data, ctx->I, tmp, in _get_more_prng_bytes() 149 hexdump("tmp stage 2: ", tmp, DEFAULT_BLK_SZ); in _get_more_prng_bytes() 155 crypto_cipher_encrypt_one(ctx->tfm, output, tmp); in _get_more_prng_bytes()
|
D | camellia_generic.c | 1004 u32 tmp[4]; in camellia_encrypt() local 1006 tmp[0] = be32_to_cpu(src[0]); in camellia_encrypt() 1007 tmp[1] = be32_to_cpu(src[1]); in camellia_encrypt() 1008 tmp[2] = be32_to_cpu(src[2]); in camellia_encrypt() 1009 tmp[3] = be32_to_cpu(src[3]); in camellia_encrypt() 1016 camellia_do_encrypt(cctx->key_table, tmp, max); in camellia_encrypt() 1019 dst[0] = cpu_to_be32(tmp[2]); in camellia_encrypt() 1020 dst[1] = cpu_to_be32(tmp[3]); in camellia_encrypt() 1021 dst[2] = cpu_to_be32(tmp[0]); in camellia_encrypt() 1022 dst[3] = cpu_to_be32(tmp[1]); in camellia_encrypt() [all …]
|
D | twofish_common.c | 470 tmp = poly_to_exp[key[i] - 1]; \ 471 (a) ^= exp_to_poly[tmp + (w)]; \ 472 (b) ^= exp_to_poly[tmp + (x)]; \ 473 (c) ^= exp_to_poly[tmp + (y)]; \ 474 (d) ^= exp_to_poly[tmp + (z)]; \ 583 u8 tmp; in __twofish_setkey() local
|
D | scatterwalk.c | 59 struct scatterlist tmp[2]; in scatterwalk_map_and_copy() local 64 sg = scatterwalk_ffwd(tmp, sg, start); in scatterwalk_map_and_copy()
|
D | sm3_generic.c | 57 unsigned int tmp; in sm3_expand() local 64 tmp = w[i - 16] ^ w[i - 9] ^ rol32(w[i - 3], 15); in sm3_expand() 65 w[i] = p1(tmp) ^ (rol32(w[i - 13], 7)) ^ w[i - 6]; in sm3_expand()
|
D | ghash-generic.c | 122 u8 *tmp = dst + (GHASH_BLOCK_SIZE - dctx->bytes); in ghash_flush() local 125 *tmp++ ^= 0; in ghash_flush()
|
D | ctr.c | 33 u8 tmp[MAX_CIPHER_BLOCKSIZE + MAX_CIPHER_ALIGNMASK]; in crypto_ctr_crypt_final() local 34 u8 *keystream = PTR_ALIGN(tmp + 0, alignmask + 1); in crypto_ctr_crypt_final() 81 u8 tmp[MAX_CIPHER_BLOCKSIZE + MAX_CIPHER_ALIGNMASK]; in crypto_ctr_crypt_inplace() local 82 u8 *keystream = PTR_ALIGN(tmp + 0, alignmask + 1); in crypto_ctr_crypt_inplace()
|
D | lrw.c | 75 be128 tmp = { 0 }; in setkey() local 97 setbit128_bbe(&tmp, i); in setkey() 98 ctx->mulinc[i] = tmp; in setkey()
|
D | blkcipher.c | 177 u8 *tmp = walk->page; in blkcipher_next_copy() local 180 memcpy(tmp, walk->src.virt.addr, walk->nbytes); in blkcipher_next_copy() 183 walk->src.virt.addr = tmp; in blkcipher_next_copy() 184 walk->dst.virt.addr = tmp; in blkcipher_next_copy()
|
D | skcipher.c | 183 struct skcipher_walk_buffer *p, *tmp; in skcipher_walk_complete() local 185 list_for_each_entry_safe(p, tmp, &walk->buffers, entry) { in skcipher_walk_complete() 288 u8 *tmp = walk->page; in skcipher_next_copy() local 291 memcpy(tmp, walk->src.virt.addr, walk->nbytes); in skcipher_next_copy() 294 walk->src.virt.addr = tmp; in skcipher_next_copy() 295 walk->dst.virt.addr = tmp; in skcipher_next_copy()
|
D | ablkcipher.c | 42 struct ablkcipher_buffer *p, *tmp; in __ablkcipher_walk_complete() local 44 list_for_each_entry_safe(p, tmp, &walk->buffers, entry) { in __ablkcipher_walk_complete()
|
D | algif_aead.c | 101 struct af_alg_tsgl *tsgl, *tmp; in _aead_recvmsg() local 182 list_for_each_entry_safe(tsgl, tmp, &ctx->tsgl_list, list) { in _aead_recvmsg()
|
D | drbg.c | 827 unsigned char *tmp = drbg->scratchpad + drbg_statelen(drbg); in drbg_hash_df() local 842 ret = drbg_kcapi_hash(drbg, tmp, entropylist); in drbg_hash_df() 849 memcpy(outval + len, tmp, blocklen); in drbg_hash_df() 854 memset(tmp, 0, drbg_blocklen(drbg)); in drbg_hash_df()
|
D | af_alg.c | 655 struct af_alg_rsgl *rsgl, *tmp; in af_alg_free_areq_sgls() local 660 list_for_each_entry_safe(rsgl, tmp, &areq->rsgl_list, list) { in af_alg_free_areq_sgls()
|
/crypto/asymmetric_keys/ |
D | verify_pefile.c | 245 unsigned *canon, tmp, loop, i, hashed_bytes; in pefile_digest_pe_contents() local 255 tmp = ctx->image_checksum_offset + sizeof(uint32_t); in pefile_digest_pe_contents() 256 ret = crypto_shash_update(desc, pebuf + tmp, in pefile_digest_pe_contents() 257 ctx->cert_dirent_offset - tmp); in pefile_digest_pe_contents() 261 tmp = ctx->cert_dirent_offset + sizeof(struct data_dirent); in pefile_digest_pe_contents() 262 ret = crypto_shash_update(desc, pebuf + tmp, ctx->header_size - tmp); in pefile_digest_pe_contents() 303 tmp = hashed_bytes + ctx->certs_size; in pefile_digest_pe_contents() 306 pelen - tmp); in pefile_digest_pe_contents()
|
/crypto/async_tx/ |
D | async_xor.c | 38 dma_addr_t tmp; in do_async_xor() local 62 tmp = src_list[0]; in do_async_xor() 80 src_list[0] = tmp; in do_async_xor()
|