Home
last modified time | relevance | path

Searched refs:dst (Results 1 – 25 of 79) sorted by relevance

1234

/crypto/
Dauthencesn.c39 struct scatterlist dst[2]; member
108 struct scatterlist *dst = req->dst; in crypto_authenc_esn_genicv_tail() local
112 scatterwalk_map_and_copy(tmp, dst, 4, 4, 0); in crypto_authenc_esn_genicv_tail()
113 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0); in crypto_authenc_esn_genicv_tail()
114 scatterwalk_map_and_copy(tmp, dst, 0, 8, 1); in crypto_authenc_esn_genicv_tail()
116 scatterwalk_map_and_copy(hash, dst, assoclen + cryptlen, authsize, 1); in crypto_authenc_esn_genicv_tail()
142 struct scatterlist *dst = req->dst; in crypto_authenc_esn_genicv() local
149 scatterwalk_map_and_copy(tmp, dst, 0, 8, 0); in crypto_authenc_esn_genicv()
150 scatterwalk_map_and_copy(tmp, dst, 4, 4, 1); in crypto_authenc_esn_genicv()
151 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 1); in crypto_authenc_esn_genicv()
[all …]
Daegis.h27 static __always_inline void crypto_aegis_block_xor(union aegis_block *dst, in crypto_aegis_block_xor() argument
30 dst->words64[0] ^= src->words64[0]; in crypto_aegis_block_xor()
31 dst->words64[1] ^= src->words64[1]; in crypto_aegis_block_xor()
34 static __always_inline void crypto_aegis_block_and(union aegis_block *dst, in crypto_aegis_block_and() argument
37 dst->words64[0] &= src->words64[0]; in crypto_aegis_block_and()
38 dst->words64[1] &= src->words64[1]; in crypto_aegis_block_and()
41 static __always_inline void crypto_aegis_aesenc(union aegis_block *dst, in crypto_aegis_aesenc() argument
54 dst->words32[0] = cpu_to_le32(d0) ^ key->words32[0]; in crypto_aegis_aesenc()
55 dst->words32[1] = cpu_to_le32(d1) ^ key->words32[1]; in crypto_aegis_aesenc()
56 dst->words32[2] = cpu_to_le32(d2) ^ key->words32[2]; in crypto_aegis_aesenc()
[all …]
Dcfb.c36 const u8 *src, u8 *dst) in crypto_cfb_encrypt_one() argument
38 crypto_cipher_encrypt_one(skcipher_cipher_simple(tfm), dst, src); in crypto_cfb_encrypt_one()
49 u8 *dst = walk->dst.virt.addr; in crypto_cfb_final() local
54 crypto_xor_cpy(dst, stream, src, nbytes); in crypto_cfb_final()
63 u8 *dst = walk->dst.virt.addr; in crypto_cfb_encrypt_segment() local
67 crypto_cfb_encrypt_one(tfm, iv, dst); in crypto_cfb_encrypt_segment()
68 crypto_xor(dst, src, bsize); in crypto_cfb_encrypt_segment()
69 iv = dst; in crypto_cfb_encrypt_segment()
72 dst += bsize; in crypto_cfb_encrypt_segment()
112 if (walk.src.virt.addr == walk.dst.virt.addr) in crypto_cfb_encrypt()
[all …]
Dlz4.c54 u8 *dst, unsigned int *dlen, void *ctx) in __lz4_compress_crypto() argument
56 int out_len = LZ4_compress_default(src, dst, in __lz4_compress_crypto()
67 unsigned int slen, u8 *dst, unsigned int *dlen, in lz4_scompress() argument
70 return __lz4_compress_crypto(src, slen, dst, dlen, ctx); in lz4_scompress()
74 unsigned int slen, u8 *dst, unsigned int *dlen) in lz4_compress_crypto() argument
78 return __lz4_compress_crypto(src, slen, dst, dlen, ctx->lz4_comp_mem); in lz4_compress_crypto()
82 u8 *dst, unsigned int *dlen, void *ctx) in __lz4_decompress_crypto() argument
84 int out_len = LZ4_decompress_safe(src, dst, slen, *dlen); in __lz4_decompress_crypto()
94 unsigned int slen, u8 *dst, unsigned int *dlen, in lz4_sdecompress() argument
97 return __lz4_decompress_crypto(src, slen, dst, dlen, NULL); in lz4_sdecompress()
[all …]
Dlz4hc.c53 u8 *dst, unsigned int *dlen, void *ctx) in __lz4hc_compress_crypto() argument
55 int out_len = LZ4_compress_HC(src, dst, slen, in __lz4hc_compress_crypto()
66 unsigned int slen, u8 *dst, unsigned int *dlen, in lz4hc_scompress() argument
69 return __lz4hc_compress_crypto(src, slen, dst, dlen, ctx); in lz4hc_scompress()
73 unsigned int slen, u8 *dst, in lz4hc_compress_crypto() argument
78 return __lz4hc_compress_crypto(src, slen, dst, dlen, in lz4hc_compress_crypto()
83 u8 *dst, unsigned int *dlen, void *ctx) in __lz4hc_decompress_crypto() argument
85 int out_len = LZ4_decompress_safe(src, dst, slen, *dlen); in __lz4hc_decompress_crypto()
95 unsigned int slen, u8 *dst, unsigned int *dlen, in lz4hc_sdecompress() argument
98 return __lz4hc_decompress_crypto(src, slen, dst, dlen, NULL); in lz4hc_sdecompress()
[all …]
Dkeywrap.c128 struct scatterlist *src, *dst; in crypto_kw_decrypt() local
149 dst = req->dst; in crypto_kw_decrypt()
170 crypto_kw_scatterlist_ff(&dst_walk, dst, nbytes); in crypto_kw_decrypt()
179 src = req->dst; in crypto_kw_decrypt()
180 dst = req->dst; in crypto_kw_decrypt()
197 struct scatterlist *src, *dst; in crypto_kw_encrypt() local
222 dst = req->dst; in crypto_kw_encrypt()
229 scatterwalk_start(&dst_walk, dst); in crypto_kw_encrypt()
251 src = req->dst; in crypto_kw_encrypt()
252 dst = req->dst; in crypto_kw_encrypt()
Dlzo-rle.c53 u8 *dst, unsigned int *dlen, void *ctx) in __lzorle_compress() argument
58 err = lzorle1x_1_compress(src, slen, dst, &tmp_len, ctx); in __lzorle_compress()
68 unsigned int slen, u8 *dst, unsigned int *dlen) in lzorle_compress() argument
72 return __lzorle_compress(src, slen, dst, dlen, ctx->lzorle_comp_mem); in lzorle_compress()
76 unsigned int slen, u8 *dst, unsigned int *dlen, in lzorle_scompress() argument
79 return __lzorle_compress(src, slen, dst, dlen, ctx); in lzorle_scompress()
83 u8 *dst, unsigned int *dlen) in __lzorle_decompress() argument
88 err = lzo1x_decompress_safe(src, slen, dst, &tmp_len); in __lzorle_decompress()
98 unsigned int slen, u8 *dst, unsigned int *dlen) in lzorle_decompress() argument
100 return __lzorle_decompress(src, slen, dst, dlen); in lzorle_decompress()
[all …]
Dlzo.c53 u8 *dst, unsigned int *dlen, void *ctx) in __lzo_compress() argument
58 err = lzo1x_1_compress(src, slen, dst, &tmp_len, ctx); in __lzo_compress()
68 unsigned int slen, u8 *dst, unsigned int *dlen) in lzo_compress() argument
72 return __lzo_compress(src, slen, dst, dlen, ctx->lzo_comp_mem); in lzo_compress()
76 unsigned int slen, u8 *dst, unsigned int *dlen, in lzo_scompress() argument
79 return __lzo_compress(src, slen, dst, dlen, ctx); in lzo_scompress()
83 u8 *dst, unsigned int *dlen) in __lzo_decompress() argument
88 err = lzo1x_decompress_safe(src, slen, dst, &tmp_len); in __lzo_decompress()
98 unsigned int slen, u8 *dst, unsigned int *dlen) in lzo_decompress() argument
100 return __lzo_decompress(src, slen, dst, dlen); in lzo_decompress()
[all …]
Dghash-generic.c85 u8 *dst = dctx->buffer; in ghash_update() local
89 u8 *pos = dst + (GHASH_BLOCK_SIZE - dctx->bytes); in ghash_update()
98 gf128mul_4k_lle((be128 *)dst, ctx->gf128); in ghash_update()
102 crypto_xor(dst, src, GHASH_BLOCK_SIZE); in ghash_update()
103 gf128mul_4k_lle((be128 *)dst, ctx->gf128); in ghash_update()
111 *dst++ ^= *src++; in ghash_update()
119 u8 *dst = dctx->buffer; in ghash_flush() local
122 u8 *tmp = dst + (GHASH_BLOCK_SIZE - dctx->bytes); in ghash_flush()
127 gf128mul_4k_lle((be128 *)dst, ctx->gf128); in ghash_flush()
133 static int ghash_final(struct shash_desc *desc, u8 *dst) in ghash_final() argument
[all …]
Dcipher.c62 u8 *dst, const u8 *src) in cipher_crypt_unaligned() argument
71 memcpy(dst, tmp, size); in cipher_crypt_unaligned()
75 u8 *dst, const u8 *src) in cipher_encrypt_unaligned() argument
80 if (unlikely(((unsigned long)dst | (unsigned long)src) & alignmask)) { in cipher_encrypt_unaligned()
81 cipher_crypt_unaligned(cipher->cia_encrypt, tfm, dst, src); in cipher_encrypt_unaligned()
85 cipher->cia_encrypt(tfm, dst, src); in cipher_encrypt_unaligned()
89 u8 *dst, const u8 *src) in cipher_decrypt_unaligned() argument
94 if (unlikely(((unsigned long)dst | (unsigned long)src) & alignmask)) { in cipher_decrypt_unaligned()
95 cipher_crypt_unaligned(cipher->cia_decrypt, tfm, dst, src); in cipher_decrypt_unaligned()
99 cipher->cia_decrypt(tfm, dst, src); in cipher_decrypt_unaligned()
Dpcbc.c26 u8 *dst = walk->dst.virt.addr; in crypto_pcbc_encrypt_segment() local
31 crypto_cipher_encrypt_one(tfm, dst, iv); in crypto_pcbc_encrypt_segment()
32 crypto_xor_cpy(iv, dst, src, bsize); in crypto_pcbc_encrypt_segment()
35 dst += bsize; in crypto_pcbc_encrypt_segment()
74 if (walk.src.virt.addr == walk.dst.virt.addr) in crypto_pcbc_encrypt()
93 u8 *dst = walk->dst.virt.addr; in crypto_pcbc_decrypt_segment() local
97 crypto_cipher_decrypt_one(tfm, dst, src); in crypto_pcbc_decrypt_segment()
98 crypto_xor(dst, iv, bsize); in crypto_pcbc_decrypt_segment()
99 crypto_xor_cpy(iv, dst, src, bsize); in crypto_pcbc_decrypt_segment()
102 dst += bsize; in crypto_pcbc_decrypt_segment()
[all …]
Dzstd.c151 u8 *dst, unsigned int *dlen, void *ctx) in __zstd_compress() argument
157 out_len = ZSTD_compressCCtx(zctx->cctx, dst, *dlen, src, slen, params); in __zstd_compress()
165 unsigned int slen, u8 *dst, unsigned int *dlen) in zstd_compress() argument
169 return __zstd_compress(src, slen, dst, dlen, ctx); in zstd_compress()
173 unsigned int slen, u8 *dst, unsigned int *dlen, in zstd_scompress() argument
176 return __zstd_compress(src, slen, dst, dlen, ctx); in zstd_scompress()
180 u8 *dst, unsigned int *dlen, void *ctx) in __zstd_decompress() argument
185 out_len = ZSTD_decompressDCtx(zctx->dctx, dst, *dlen, src, slen); in __zstd_decompress()
193 unsigned int slen, u8 *dst, unsigned int *dlen) in zstd_decompress() argument
197 return __zstd_decompress(src, slen, dst, dlen, ctx); in zstd_decompress()
[all …]
Daegis128-neon.c12 void crypto_aegis128_encrypt_chunk_neon(void *state, void *dst, const void *src,
14 void crypto_aegis128_decrypt_chunk_neon(void *state, void *dst, const void *src,
35 void crypto_aegis128_encrypt_chunk_simd(union aegis_block *state, u8 *dst, in crypto_aegis128_encrypt_chunk_simd() argument
39 crypto_aegis128_encrypt_chunk_neon(state, dst, src, size); in crypto_aegis128_encrypt_chunk_simd()
43 void crypto_aegis128_decrypt_chunk_simd(union aegis_block *state, u8 *dst, in crypto_aegis128_decrypt_chunk_simd() argument
47 crypto_aegis128_decrypt_chunk_neon(state, dst, src, size); in crypto_aegis128_decrypt_chunk_simd()
D842.c67 u8 *dst, unsigned int *dlen) in crypto842_compress() argument
71 return sw842_compress(src, slen, dst, dlen, ctx->wmem); in crypto842_compress()
76 u8 *dst, unsigned int *dlen, void *ctx) in crypto842_scompress() argument
78 return sw842_compress(src, slen, dst, dlen, ctx); in crypto842_scompress()
83 u8 *dst, unsigned int *dlen) in crypto842_decompress() argument
85 return sw842_decompress(src, slen, dst, dlen); in crypto842_decompress()
90 u8 *dst, unsigned int *dlen, void *ctx) in crypto842_sdecompress() argument
92 return sw842_decompress(src, slen, dst, dlen); in crypto842_sdecompress()
Dscatterwalk.c21 void *dst = out ? sgdata : buf; in memcpy_dir() local
23 memcpy(dst, src, nbytes); in memcpy_dir()
72 struct scatterlist *scatterwalk_ffwd(struct scatterlist dst[2], in scatterwalk_ffwd()
87 sg_init_table(dst, 2); in scatterwalk_ffwd()
88 sg_set_page(dst, sg_page(src), src->length - len, src->offset + len); in scatterwalk_ffwd()
89 scatterwalk_crypto_chain(dst, sg_next(src), 2); in scatterwalk_ffwd()
91 return dst; in scatterwalk_ffwd()
Ddes_generic.c40 static void crypto_des_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) in crypto_des_encrypt() argument
44 des_encrypt(dctx, dst, src); in crypto_des_encrypt()
47 static void crypto_des_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) in crypto_des_decrypt() argument
51 des_decrypt(dctx, dst, src); in crypto_des_decrypt()
75 static void crypto_des3_ede_encrypt(struct crypto_tfm *tfm, u8 *dst, in crypto_des3_ede_encrypt() argument
80 des3_ede_encrypt(dctx, dst, src); in crypto_des3_ede_encrypt()
83 static void crypto_des3_ede_decrypt(struct crypto_tfm *tfm, u8 *dst, in crypto_des3_ede_decrypt() argument
88 des3_ede_decrypt(dctx, dst, src); in crypto_des3_ede_decrypt()
Dablkcipher.c26 struct scatter_walk dst; member
37 scatterwalk_copychunks(p->data, &p->dst, p->len, 1); in ablkcipher_buffer_write()
55 p->dst = walk->out; in ablkcipher_queue_write()
145 void *src, *dst, *base; in ablkcipher_next_slow() local
158 dst = (u8 *)ALIGN((unsigned long)base, alignmask + 1); in ablkcipher_next_slow()
159 src = dst = ablkcipher_get_spot(dst, bsize); in ablkcipher_next_slow()
162 p->data = dst; in ablkcipher_next_slow()
172 *dst_p = dst; in ablkcipher_next_slow()
207 walk->dst.page = scatterwalk_page(&walk->out); in ablkcipher_next_fast()
208 walk->dst.offset = offset_in_page(walk->out.offset); in ablkcipher_next_fast()
[all …]
Daegis128-core.c42 void (*crypt_chunk)(struct aegis_state *state, u8 *dst,
70 void crypto_aegis128_encrypt_chunk_simd(struct aegis_state *state, u8 *dst,
72 void crypto_aegis128_decrypt_chunk_simd(struct aegis_state *state, u8 *dst,
158 static void crypto_aegis128_encrypt_chunk(struct aegis_state *state, u8 *dst, in crypto_aegis128_encrypt_chunk() argument
163 if (AEGIS_ALIGNED(src) && AEGIS_ALIGNED(dst)) { in crypto_aegis128_encrypt_chunk()
166 (union aegis_block *)dst; in crypto_aegis128_encrypt_chunk()
182 dst += AEGIS_BLOCK_SIZE; in crypto_aegis128_encrypt_chunk()
194 memcpy(dst, tmp.bytes, AEGIS_BLOCK_SIZE); in crypto_aegis128_encrypt_chunk()
198 dst += AEGIS_BLOCK_SIZE; in crypto_aegis128_encrypt_chunk()
215 memcpy(dst, msg.bytes, size); in crypto_aegis128_encrypt_chunk()
[all …]
Dccm.c37 struct scatterlist dst[3]; member
47 struct scatterlist dst[3]; member
243 scatterwalk_map_and_copy(odata, req->dst, in crypto_ccm_encrypt_done()
282 if (req->src != req->dst) { in crypto_ccm_init_crypt()
283 sg_init_table(pctx->dst, 3); in crypto_ccm_init_crypt()
284 sg_set_buf(pctx->dst, tag, 16); in crypto_ccm_init_crypt()
285 sg = scatterwalk_ffwd(pctx->dst + 1, req->dst, req->assoclen); in crypto_ccm_init_crypt()
286 if (sg != pctx->dst + 1) in crypto_ccm_init_crypt()
287 sg_chain(pctx->dst, 2, sg); in crypto_ccm_init_crypt()
299 struct scatterlist *dst; in crypto_ccm_encrypt() local
[all …]
Dscompress.c30 void *dst; member
77 vfree(scratch->dst); in crypto_scomp_free_scratches()
79 scratch->dst = NULL; in crypto_scomp_free_scratches()
100 scratch->dst = mem; in crypto_scomp_alloc_scratches()
132 if (req->dst && !req->dlen) in scomp_acomp_comp_decomp()
144 scratch->dst, &req->dlen, *ctx); in scomp_acomp_comp_decomp()
147 scratch->dst, &req->dlen, *ctx); in scomp_acomp_comp_decomp()
149 if (!req->dst) { in scomp_acomp_comp_decomp()
150 req->dst = sgl_alloc(req->dlen, GFP_ATOMIC, NULL); in scomp_acomp_comp_decomp()
151 if (!req->dst) { in scomp_acomp_comp_decomp()
[all …]
Ddeflate.c177 u8 *dst, unsigned int *dlen, void *ctx) in __deflate_compress() argument
191 stream->next_out = (u8 *)dst; in __deflate_compress()
206 unsigned int slen, u8 *dst, unsigned int *dlen) in deflate_compress() argument
210 return __deflate_compress(src, slen, dst, dlen, dctx); in deflate_compress()
214 unsigned int slen, u8 *dst, unsigned int *dlen, in deflate_scompress() argument
217 return __deflate_compress(src, slen, dst, dlen, ctx); in deflate_scompress()
221 u8 *dst, unsigned int *dlen, void *ctx) in __deflate_decompress() argument
236 stream->next_out = (u8 *)dst; in __deflate_decompress()
262 unsigned int slen, u8 *dst, unsigned int *dlen) in deflate_decompress() argument
266 return __deflate_decompress(src, slen, dst, dlen, dctx); in deflate_decompress()
[all …]
Dtea.c57 static void tea_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) in tea_encrypt() argument
63 __le32 *out = (__le32 *)dst; in tea_encrypt()
85 static void tea_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) in tea_decrypt() argument
91 __le32 *out = (__le32 *)dst; in tea_decrypt()
130 static void xtea_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) in xtea_encrypt() argument
136 __le32 *out = (__le32 *)dst; in xtea_encrypt()
151 static void xtea_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) in xtea_decrypt() argument
156 __le32 *out = (__le32 *)dst; in xtea_decrypt()
174 static void xeta_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) in xeta_encrypt() argument
180 __le32 *out = (__le32 *)dst; in xeta_encrypt()
[all …]
Dauthenc.c36 struct scatterlist dst[2]; member
134 scatterwalk_map_and_copy(ahreq->result, req->dst, in authenc_geniv_ahash_done()
158 ahash_request_set_crypt(ahreq, req->dst, hash, in crypto_authenc_genicv()
167 scatterwalk_map_and_copy(hash, req->dst, req->assoclen + req->cryptlen, in crypto_authenc_genicv()
196 skcipher_request_set_crypt(skreq, req->src, req->dst, req->assoclen, in crypto_authenc_copy_assoc()
213 struct scatterlist *src, *dst; in crypto_authenc_encrypt() local
217 dst = src; in crypto_authenc_encrypt()
219 if (req->src != req->dst) { in crypto_authenc_encrypt()
224 dst = scatterwalk_ffwd(areq_ctx->dst, req->dst, req->assoclen); in crypto_authenc_encrypt()
230 skcipher_request_set_crypt(skreq, src, dst, cryptlen, req->iv); in crypto_authenc_encrypt()
[all …]
Ddh_helper.c15 static inline u8 *dh_pack_data(u8 *dst, u8 *end, const void *src, size_t size) in dh_pack_data() argument
17 if (!dst || size > end - dst) in dh_pack_data()
19 memcpy(dst, src, size); in dh_pack_data()
20 return dst + size; in dh_pack_data()
23 static inline const u8 *dh_unpack_data(void *dst, const void *src, size_t size) in dh_unpack_data() argument
25 memcpy(dst, src, size); in dh_unpack_data()
Dcompress.c17 u8 *dst, unsigned int *dlen) in crypto_compress() argument
19 return tfm->__crt_alg->cra_compress.coa_compress(tfm, src, slen, dst, in crypto_compress()
25 u8 *dst, unsigned int *dlen) in crypto_decompress() argument
27 return tfm->__crt_alg->cra_compress.coa_decompress(tfm, src, slen, dst, in crypto_decompress()

1234