/net/ipv6/ |
D | esp6.c | 53 static void *esp_alloc_tmp(struct crypto_aead *aead, int nfrags, int seqihlen) in esp_alloc_tmp() argument 59 len += crypto_aead_ivsize(aead); in esp_alloc_tmp() 62 len += crypto_aead_alignmask(aead) & in esp_alloc_tmp() 67 len += sizeof(struct aead_request) + crypto_aead_reqsize(aead); in esp_alloc_tmp() 80 static inline u8 *esp_tmp_iv(struct crypto_aead *aead, void *tmp, int seqhilen) in esp_tmp_iv() argument 82 return crypto_aead_ivsize(aead) ? in esp_tmp_iv() 84 crypto_aead_alignmask(aead) + 1) : tmp + seqhilen; in esp_tmp_iv() 87 static inline struct aead_request *esp_tmp_req(struct crypto_aead *aead, u8 *iv) in esp_tmp_req() argument 91 req = (void *)PTR_ALIGN(iv + crypto_aead_ivsize(aead), in esp_tmp_req() 93 aead_request_set_tfm(req, aead); in esp_tmp_req() [all …]
|
D | esp6_offload.c | 186 struct crypto_aead *aead; in esp6_gso_segment() local 199 aead = x->data; in esp6_gso_segment() 205 if (!pskb_may_pull(skb, sizeof(*esph) + crypto_aead_ivsize(aead))) in esp6_gso_segment() 208 __skb_pull(skb, sizeof(*esph) + crypto_aead_ivsize(aead)); in esp6_gso_segment() 226 struct crypto_aead *aead = x->data; in esp6_input_tail() local 229 if (!pskb_may_pull(skb, sizeof(struct ip_esp_hdr) + crypto_aead_ivsize(aead))) in esp6_input_tail() 246 struct crypto_aead *aead; in esp6_xmit() local 267 aead = x->data; in esp6_xmit() 268 alen = crypto_aead_authsize(aead); in esp6_xmit() 273 blksize = ALIGN(crypto_aead_blocksize(aead), 4); in esp6_xmit()
|
/net/ipv4/ |
D | esp4.c | 44 static void *esp_alloc_tmp(struct crypto_aead *aead, int nfrags, int extralen) in esp_alloc_tmp() argument 50 len += crypto_aead_ivsize(aead); in esp_alloc_tmp() 53 len += crypto_aead_alignmask(aead) & in esp_alloc_tmp() 58 len += sizeof(struct aead_request) + crypto_aead_reqsize(aead); in esp_alloc_tmp() 71 static inline u8 *esp_tmp_iv(struct crypto_aead *aead, void *tmp, int extralen) in esp_tmp_iv() argument 73 return crypto_aead_ivsize(aead) ? in esp_tmp_iv() 75 crypto_aead_alignmask(aead) + 1) : tmp + extralen; in esp_tmp_iv() 78 static inline struct aead_request *esp_tmp_req(struct crypto_aead *aead, u8 *iv) in esp_tmp_req() argument 82 req = (void *)PTR_ALIGN(iv + crypto_aead_ivsize(aead), in esp_tmp_req() 84 aead_request_set_tfm(req, aead); in esp_tmp_req() [all …]
|
D | esp4_offload.c | 152 struct crypto_aead *aead; in esp4_gso_segment() local 165 aead = x->data; in esp4_gso_segment() 171 if (!pskb_may_pull(skb, sizeof(*esph) + crypto_aead_ivsize(aead))) in esp4_gso_segment() 174 __skb_pull(skb, sizeof(*esph) + crypto_aead_ivsize(aead)); in esp4_gso_segment() 194 struct crypto_aead *aead = x->data; in esp_input_tail() local 197 if (!pskb_may_pull(skb, sizeof(struct ip_esp_hdr) + crypto_aead_ivsize(aead))) in esp_input_tail() 213 struct crypto_aead *aead; in esp_xmit() local 236 aead = x->data; in esp_xmit() 237 alen = crypto_aead_authsize(aead); in esp_xmit() 242 blksize = ALIGN(crypto_aead_blocksize(aead), 4); in esp_xmit()
|
/net/xfrm/ |
D | xfrm_algo.c | 31 .aead = { 50 .aead = { 69 .aead = { 88 .aead = { 107 .aead = { 126 .aead = { 145 .aead = { 164 .aead = { 734 const struct xfrm_aead_name *aead = data; in xfrm_aead_name_match() local 735 const char *name = aead->name; in xfrm_aead_name_match() [all …]
|
D | xfrm_state.c | 484 kfree(x->aead); in ___xfrm_state_destroy() 1492 if (orig->aead) { in xfrm_state_clone() 1493 x->aead = xfrm_algo_aead_clone(orig->aead); in xfrm_state_clone() 1495 if (!x->aead) in xfrm_state_clone() 2530 struct crypto_aead *aead; in xfrm_state_mtu() local 2537 aead = x->data; in xfrm_state_mtu() 2538 blksize = ALIGN(crypto_aead_blocksize(aead), 4); in xfrm_state_mtu() 2555 return ((mtu - x->props.header_len - crypto_aead_authsize(aead) - in xfrm_state_mtu()
|
D | xfrm_user.c | 420 x->aead = p; in attach_aead() 421 x->geniv = algo->uinfo.aead.geniv; in attach_aead() 908 if (x->aead) { in copy_to_user_state_extra() 909 ret = nla_put(skb, XFRMA_ALG_AEAD, aead_len(x->aead), x->aead); in copy_to_user_state_extra() 2893 if (x->aead) in xfrm_sa_len() 2894 l += nla_total_size(aead_len(x->aead)); in xfrm_sa_len()
|
/net/tls/ |
D | tls_device_fallback.c | 49 struct crypto_aead *aead, char *aad, in tls_enc_record() argument 121 struct crypto_aead *aead) in tls_init_aead_request() argument 123 aead_request_set_tfm(aead_req, aead); in tls_init_aead_request() 127 static struct aead_request *tls_alloc_aead_request(struct crypto_aead *aead, in tls_alloc_aead_request() argument 131 crypto_aead_reqsize(aead); in tls_alloc_aead_request() 136 tls_init_aead_request(aead_req, aead); in tls_alloc_aead_request() 141 struct crypto_aead *aead, struct scatterlist *sg_in, in tls_enc_records() argument 152 rc = tls_enc_record(aead_req, aead, aad, iv, in tls_enc_records()
|
D | tls_sw.c | 2320 struct crypto_aead **aead; in tls_set_sw_offload() local 2364 aead = &sw_ctx_tx->aead_send; in tls_set_sw_offload() 2374 aead = &sw_ctx_rx->aead_recv; in tls_set_sw_offload() 2473 if (!*aead) { in tls_set_sw_offload() 2474 *aead = crypto_alloc_aead(cipher_name, 0, 0); in tls_set_sw_offload() 2475 if (IS_ERR(*aead)) { in tls_set_sw_offload() 2476 rc = PTR_ERR(*aead); in tls_set_sw_offload() 2477 *aead = NULL; in tls_set_sw_offload() 2484 rc = crypto_aead_setkey(*aead, key, keysize); in tls_set_sw_offload() 2489 rc = crypto_aead_setauthsize(*aead, prot->tag_size); in tls_set_sw_offload() [all …]
|