Lines Matching +full:xo +full:- +full:1
1 // SPDX-License-Identifier: GPL-2.0-only
36 #define ESP_SKB_CB(__skb) ((struct esp_skb_cb *)&((__skb)->cb[0]))
56 ~(crypto_tfm_ctx_alignment() - 1); in esp_alloc_tmp()
77 crypto_aead_alignmask(aead) + 1) : tmp + extralen; in esp_tmp_iv()
93 return (void *)ALIGN((unsigned long)(req + 1) + in esp_req_sg()
101 struct crypto_aead *aead = x->data; in esp_ssg_unref()
107 if (x->props.flags & XFRM_STATE_ESN) in esp_ssg_unref()
115 * Skip the first sg which comes from skb->data. in esp_ssg_unref()
117 if (req->src != req->dst) in esp_ssg_unref()
118 for (sg = sg_next(req->src); sg; sg = sg_next(sg)) in esp_ssg_unref()
132 sock_put(esk->sk); in esp_free_tcp_sk()
138 struct xfrm_encap_tmpl *encap = x->encap; in esp_find_tcp_sk()
144 sk = rcu_dereference(x->encap_sk); in esp_find_tcp_sk()
145 if (sk && sk->sk_state == TCP_ESTABLISHED) in esp_find_tcp_sk()
148 spin_lock_bh(&x->lock); in esp_find_tcp_sk()
149 sport = encap->encap_sport; in esp_find_tcp_sk()
150 dport = encap->encap_dport; in esp_find_tcp_sk()
151 nsk = rcu_dereference_protected(x->encap_sk, in esp_find_tcp_sk()
152 lockdep_is_held(&x->lock)); in esp_find_tcp_sk()
156 spin_unlock_bh(&x->lock); in esp_find_tcp_sk()
157 return ERR_PTR(-ENOMEM); in esp_find_tcp_sk()
159 RCU_INIT_POINTER(x->encap_sk, NULL); in esp_find_tcp_sk()
160 esk->sk = sk; in esp_find_tcp_sk()
161 call_rcu(&esk->rcu, esp_free_tcp_sk); in esp_find_tcp_sk()
163 spin_unlock_bh(&x->lock); in esp_find_tcp_sk()
165 sk = inet_lookup_established(xs_net(x), &tcp_hashinfo, x->id.daddr.a4, in esp_find_tcp_sk()
166 dport, x->props.saddr.a4, sport, 0); in esp_find_tcp_sk()
168 return ERR_PTR(-ENOENT); in esp_find_tcp_sk()
172 return ERR_PTR(-EINVAL); in esp_find_tcp_sk()
175 spin_lock_bh(&x->lock); in esp_find_tcp_sk()
176 nsk = rcu_dereference_protected(x->encap_sk, in esp_find_tcp_sk()
177 lockdep_is_held(&x->lock)); in esp_find_tcp_sk()
178 if (encap->encap_sport != sport || in esp_find_tcp_sk()
179 encap->encap_dport != dport) { in esp_find_tcp_sk()
181 sk = nsk ?: ERR_PTR(-EREMCHG); in esp_find_tcp_sk()
185 rcu_assign_pointer(x->encap_sk, sk); in esp_find_tcp_sk()
187 spin_unlock_bh(&x->lock); in esp_find_tcp_sk()
220 struct xfrm_state *x = dst->xfrm; in esp_output_tcp_encap_cb()
237 return err ?: -EINPROGRESS; in esp_output_tail_tcp()
244 return -EOPNOTSUPP; in esp_output_tail_tcp()
250 struct sk_buff *skb = base->data; in esp_output_done()
251 struct xfrm_offload *xo = xfrm_offload(skb); in esp_output_done() local
255 if (xo && (xo->flags & XFRM_DEV_RESUME)) { in esp_output_done()
258 x = sp->xvec[sp->len - 1]; in esp_output_done()
260 x = skb_dst(skb)->xfrm; in esp_output_done()
263 tmp = ESP_SKB_CB(skb)->tmp; in esp_output_done()
267 if (xo && (xo->flags & XFRM_DEV_RESUME)) { in esp_output_done()
274 skb_push(skb, skb->data - skb_mac_header(skb)); in esp_output_done()
279 x->encap && x->encap->encap_type == TCP_ENCAP_ESPINTCP) in esp_output_done()
282 xfrm_output_resume(skb->sk, skb, err); in esp_output_done()
289 struct ip_esp_hdr *esph = (void *)(skb->data + offset); in esp_restore_header()
290 void *tmp = ESP_SKB_CB(skb)->tmp; in esp_restore_header()
293 esph->seq_no = esph->spi; in esp_restore_header()
294 esph->spi = *seqhi; in esp_restore_header()
299 void *tmp = ESP_SKB_CB(skb)->tmp; in esp_output_restore_header()
302 esp_restore_header(skb, skb_transport_offset(skb) + extra->esphoff - in esp_output_restore_header()
315 if ((x->props.flags & XFRM_STATE_ESN)) { in esp_output_set_extra()
317 struct xfrm_offload *xo = xfrm_offload(skb); in esp_output_set_extra() local
319 if (xo) in esp_output_set_extra()
320 seqhi = xo->seq.hi; in esp_output_set_extra()
322 seqhi = XFRM_SKB_CB(skb)->seq.output.hi; in esp_output_set_extra()
324 extra->esphoff = (unsigned char *)esph - in esp_output_set_extra()
326 esph = (struct ip_esp_hdr *)((unsigned char *)esph - 4); in esp_output_set_extra()
327 extra->seqhi = esph->spi; in esp_output_set_extra()
328 esph->seq_no = htonl(seqhi); in esp_output_set_extra()
331 esph->spi = x->id.spi; in esp_output_set_extra()
338 struct sk_buff *skb = base->data; in esp_output_done_esn()
354 len = skb->len + esp->tailen - skb_transport_offset(skb); in esp_output_udp_encap()
356 return ERR_PTR(-EMSGSIZE); in esp_output_udp_encap()
358 uh = (struct udphdr *)esp->esph; in esp_output_udp_encap()
359 uh->source = sport; in esp_output_udp_encap()
360 uh->dest = dport; in esp_output_udp_encap()
361 uh->len = htons(len); in esp_output_udp_encap()
362 uh->check = 0; in esp_output_udp_encap()
367 udpdata32 = (__be32 *)(uh + 1); in esp_output_udp_encap()
368 udpdata32[0] = udpdata32[1] = 0; in esp_output_udp_encap()
372 return (struct ip_esp_hdr *)(uh + 1); in esp_output_udp_encap()
380 __be16 *lenp = (void *)esp->esph; in esp_output_tcp_encap()
385 len = skb->len + esp->tailen - skb_transport_offset(skb); in esp_output_tcp_encap()
387 return ERR_PTR(-EMSGSIZE); in esp_output_tcp_encap()
397 esph = (struct ip_esp_hdr *)(lenp + 1); in esp_output_tcp_encap()
406 return ERR_PTR(-EOPNOTSUPP); in esp_output_tcp_encap()
413 struct xfrm_encap_tmpl *encap = x->encap; in esp_output_encap()
418 spin_lock_bh(&x->lock); in esp_output_encap()
419 sport = encap->encap_sport; in esp_output_encap()
420 dport = encap->encap_dport; in esp_output_encap()
421 encap_type = encap->encap_type; in esp_output_encap()
422 spin_unlock_bh(&x->lock); in esp_output_encap()
438 esp->esph = esph; in esp_output_encap()
450 int tailen = esp->tailen; in esp_output_head()
453 /* this is non-NULL only with TCP/UDP Encapsulation */ in esp_output_head()
454 if (x->encap) { in esp_output_head()
461 allocsz = ALIGN(skb->data_len + tailen, L1_CACHE_BYTES); in esp_output_head()
467 nfrags = 1; in esp_output_head()
472 } else if ((skb_shinfo(skb)->nr_frags < MAX_SKB_FRAGS) in esp_output_head()
475 struct sock *sk = skb->sk; in esp_output_head()
476 struct page_frag *pfrag = &x->xfrag; in esp_output_head()
478 esp->inplace = false; in esp_output_head()
482 spin_lock_bh(&x->lock); in esp_output_head()
485 spin_unlock_bh(&x->lock); in esp_output_head()
489 page = pfrag->page; in esp_output_head()
492 tail = page_address(page) + pfrag->offset; in esp_output_head()
494 esp_output_fill_trailer(tail, esp->tfclen, esp->plen, esp->proto); in esp_output_head()
496 nfrags = skb_shinfo(skb)->nr_frags; in esp_output_head()
498 __skb_fill_page_desc(skb, nfrags, page, pfrag->offset, in esp_output_head()
500 skb_shinfo(skb)->nr_frags = ++nfrags; in esp_output_head()
502 pfrag->offset = pfrag->offset + allocsize; in esp_output_head()
504 spin_unlock_bh(&x->lock); in esp_output_head()
508 skb->len += tailen; in esp_output_head()
509 skb->data_len += tailen; in esp_output_head()
510 skb->truesize += tailen; in esp_output_head()
512 refcount_add(tailen, &sk->sk_wmem_alloc); in esp_output_head()
519 esph_offset = (unsigned char *)esp->esph - skb_transport_header(skb); in esp_output_head()
525 esp->esph = (struct ip_esp_hdr *)(skb_transport_header(skb) + esph_offset); in esp_output_head()
528 esp_output_fill_trailer(tail, esp->tfclen, esp->plen, esp->proto); in esp_output_head()
550 int err = -ENOMEM; in esp_output_tail()
555 if (x->props.flags & XFRM_STATE_ESN) { in esp_output_tail()
560 aead = x->data; in esp_output_tail()
564 tmp = esp_alloc_tmp(aead, esp->nfrags + 2, extralen); in esp_output_tail()
573 if (esp->inplace) in esp_output_tail()
576 dsg = &sg[esp->nfrags]; in esp_output_tail()
578 esph = esp_output_set_extra(skb, x, esp->esph, extra); in esp_output_tail()
579 esp->esph = esph; in esp_output_tail()
581 sg_init_table(sg, esp->nfrags); in esp_output_tail()
583 (unsigned char *)esph - skb->data, in esp_output_tail()
584 assoclen + ivlen + esp->clen + alen); in esp_output_tail()
588 if (!esp->inplace) { in esp_output_tail()
590 struct page_frag *pfrag = &x->xfrag; in esp_output_tail()
592 allocsize = ALIGN(skb->data_len, L1_CACHE_BYTES); in esp_output_tail()
594 spin_lock_bh(&x->lock); in esp_output_tail()
596 spin_unlock_bh(&x->lock); in esp_output_tail()
600 skb_shinfo(skb)->nr_frags = 1; in esp_output_tail()
602 page = pfrag->page; in esp_output_tail()
605 __skb_fill_page_desc(skb, 0, page, pfrag->offset, skb->data_len); in esp_output_tail()
606 pfrag->offset = pfrag->offset + allocsize; in esp_output_tail()
607 spin_unlock_bh(&x->lock); in esp_output_tail()
609 sg_init_table(dsg, skb_shinfo(skb)->nr_frags + 1); in esp_output_tail()
611 (unsigned char *)esph - skb->data, in esp_output_tail()
612 assoclen + ivlen + esp->clen + alen); in esp_output_tail()
617 if ((x->props.flags & XFRM_STATE_ESN)) in esp_output_tail()
622 aead_request_set_crypt(req, sg, dsg, ivlen + esp->clen, iv); in esp_output_tail()
626 memcpy(iv + ivlen - min(ivlen, 8), (u8 *)&esp->seqno + 8 - min(ivlen, 8), in esp_output_tail()
629 ESP_SKB_CB(skb)->tmp = tmp; in esp_output_tail()
633 case -EINPROGRESS: in esp_output_tail()
636 case -ENOSPC: in esp_output_tail()
641 if ((x->props.flags & XFRM_STATE_ESN)) in esp_output_tail()
648 if (!err && x->encap && x->encap->encap_type == TCP_ENCAP_ESPINTCP) in esp_output_tail()
673 aead = x->data; in esp_output()
677 if (x->tfcpad) { in esp_output()
681 padto = min(x->tfcpad, __xfrm_state_mtu(x, dst->child_mtu_cached)); in esp_output()
682 if (skb->len < padto) in esp_output()
683 esp.tfclen = padto - skb->len; in esp_output()
686 esp.clen = ALIGN(skb->len + 2 + esp.tfclen, blksize); in esp_output()
687 esp.plen = esp.clen - skb->len - esp.tfclen; in esp_output()
697 esph->spi = x->id.spi; in esp_output()
699 esph->seq_no = htonl(XFRM_SKB_CB(skb)->seq.output.low); in esp_output()
700 esp.seqno = cpu_to_be64(XFRM_SKB_CB(skb)->seq.output.low + in esp_output()
701 ((u64)XFRM_SKB_CB(skb)->seq.output.hi << 32)); in esp_output()
703 skb_push(skb, -skb_network_offset(skb)); in esp_output()
711 struct xfrm_offload *xo = xfrm_offload(skb); in esp_remove_trailer() local
712 struct crypto_aead *aead = x->data; in esp_remove_trailer()
721 elen = skb->len - hlen; in esp_remove_trailer()
723 if (xo && (xo->flags & XFRM_ESP_NO_TRAILER)) { in esp_remove_trailer()
724 ret = xo->proto; in esp_remove_trailer()
728 if (skb_copy_bits(skb, skb->len - alen - 2, nexthdr, 2)) in esp_remove_trailer()
731 ret = -EINVAL; in esp_remove_trailer()
735 padlen + 2, elen - alen); in esp_remove_trailer()
740 if (skb->ip_summed == CHECKSUM_COMPLETE) { in esp_remove_trailer()
741 csumdiff = skb_checksum(skb, skb->len - trimlen, trimlen, 0); in esp_remove_trailer()
742 skb->csum = csum_block_sub(skb->csum, csumdiff, in esp_remove_trailer()
743 skb->len - trimlen); in esp_remove_trailer()
745 pskb_trim(skb, skb->len - trimlen); in esp_remove_trailer()
747 ret = nexthdr[1]; in esp_remove_trailer()
757 struct xfrm_offload *xo = xfrm_offload(skb); in esp_input_done2() local
758 struct crypto_aead *aead = x->data; in esp_input_done2()
762 if (!xo || (xo && !(xo->flags & CRYPTO_DONE))) in esp_input_done2()
763 kfree(ESP_SKB_CB(skb)->tmp); in esp_input_done2()
773 ihl = iph->ihl * 4; in esp_input_done2()
775 if (x->encap) { in esp_input_done2()
776 struct xfrm_encap_tmpl *encap = x->encap; in esp_input_done2()
781 switch (x->encap->encap_type) { in esp_input_done2()
783 source = th->source; in esp_input_done2()
787 source = uh->source; in esp_input_done2()
790 WARN_ON_ONCE(1); in esp_input_done2()
791 err = -EINVAL; in esp_input_done2()
796 * 1) if the NAT-T peer's IP or port changed then in esp_input_done2()
801 if (iph->saddr != x->props.saddr.a4 || in esp_input_done2()
802 source != encap->encap_sport) { in esp_input_done2()
805 ipaddr.a4 = iph->saddr; in esp_input_done2()
819 * of NAT-T in Transport Mode, or in esp_input_done2()
820 * perform other post-processing fixes in esp_input_done2()
821 * as per draft-ietf-ipsec-udp-encaps-06, in esp_input_done2()
824 if (x->props.mode == XFRM_MODE_TRANSPORT) in esp_input_done2()
825 skb->ip_summed = CHECKSUM_UNNECESSARY; in esp_input_done2()
829 if (x->props.mode == XFRM_MODE_TUNNEL) in esp_input_done2()
832 skb_set_transport_header(skb, -ihl); in esp_input_done2()
836 err = -EINVAL; in esp_input_done2()
845 struct sk_buff *skb = base->data; in esp_input_done()
865 if ((x->props.flags & XFRM_STATE_ESN)) { in esp_input_set_header()
867 *seqhi = esph->spi; in esp_input_set_header()
868 esph->spi = esph->seq_no; in esp_input_set_header()
869 esph->seq_no = XFRM_SKB_CB(skb)->seq.input.hi; in esp_input_set_header()
875 struct sk_buff *skb = base->data; in esp_input_done_esn()
882 * Note: detecting truncated vs. non-truncated authentication data is very
888 struct crypto_aead *aead = x->data; in esp_input()
892 int elen = skb->len - sizeof(struct ip_esp_hdr) - ivlen; in esp_input()
900 int err = -EINVAL; in esp_input()
911 if (x->props.flags & XFRM_STATE_ESN) { in esp_input()
918 nfrags = 1; in esp_input()
922 nfrags = skb_shinfo(skb)->nr_frags; in esp_input()
936 err = -ENOMEM; in esp_input()
941 ESP_SKB_CB(skb)->tmp = tmp; in esp_input()
950 err = skb_to_sgvec(skb, sg, 0, skb->len); in esp_input()
956 skb->ip_summed = CHECKSUM_NONE; in esp_input()
958 if ((x->props.flags & XFRM_STATE_ESN)) in esp_input()
967 if (err == -EINPROGRESS) in esp_input()
970 if ((x->props.flags & XFRM_STATE_ESN)) in esp_input()
981 struct net *net = dev_net(skb->dev); in esp4_err()
982 const struct iphdr *iph = (const struct iphdr *)skb->data; in esp4_err()
983 struct ip_esp_hdr *esph = (struct ip_esp_hdr *)(skb->data+(iph->ihl<<2)); in esp4_err()
986 switch (icmp_hdr(skb)->type) { in esp4_err()
988 if (icmp_hdr(skb)->code != ICMP_FRAG_NEEDED) in esp4_err()
996 x = xfrm_state_lookup(net, skb->mark, (const xfrm_address_t *)&iph->daddr, in esp4_err()
997 esph->spi, IPPROTO_ESP, AF_INET); in esp4_err()
1001 if (icmp_hdr(skb)->type == ICMP_DEST_UNREACH) in esp4_err()
1012 struct crypto_aead *aead = x->data; in esp_destroy()
1026 err = -ENAMETOOLONG; in esp_init_aead()
1028 x->geniv, x->aead->alg_name) >= CRYPTO_MAX_ALG_NAME) in esp_init_aead()
1036 x->data = aead; in esp_init_aead()
1038 err = crypto_aead_setkey(aead, x->aead->alg_key, in esp_init_aead()
1039 (x->aead->alg_key_len + 7) / 8); in esp_init_aead()
1043 err = crypto_aead_setauthsize(aead, x->aead->alg_icv_len / 8); in esp_init_aead()
1062 err = -EINVAL; in esp_init_authenc()
1063 if (!x->ealg) in esp_init_authenc()
1066 err = -ENAMETOOLONG; in esp_init_authenc()
1068 if ((x->props.flags & XFRM_STATE_ESN)) { in esp_init_authenc()
1071 x->geniv ?: "", x->geniv ? "(" : "", in esp_init_authenc()
1072 x->aalg ? x->aalg->alg_name : "digest_null", in esp_init_authenc()
1073 x->ealg->alg_name, in esp_init_authenc()
1074 x->geniv ? ")" : "") >= CRYPTO_MAX_ALG_NAME) in esp_init_authenc()
1079 x->geniv ?: "", x->geniv ? "(" : "", in esp_init_authenc()
1080 x->aalg ? x->aalg->alg_name : "digest_null", in esp_init_authenc()
1081 x->ealg->alg_name, in esp_init_authenc()
1082 x->geniv ? ")" : "") >= CRYPTO_MAX_ALG_NAME) in esp_init_authenc()
1091 x->data = aead; in esp_init_authenc()
1093 keylen = (x->aalg ? (x->aalg->alg_key_len + 7) / 8 : 0) + in esp_init_authenc()
1094 (x->ealg->alg_key_len + 7) / 8 + RTA_SPACE(sizeof(*param)); in esp_init_authenc()
1095 err = -ENOMEM; in esp_init_authenc()
1102 rta->rta_type = CRYPTO_AUTHENC_KEYA_PARAM; in esp_init_authenc()
1103 rta->rta_len = RTA_LENGTH(sizeof(*param)); in esp_init_authenc()
1107 if (x->aalg) { in esp_init_authenc()
1110 memcpy(p, x->aalg->alg_key, (x->aalg->alg_key_len + 7) / 8); in esp_init_authenc()
1111 p += (x->aalg->alg_key_len + 7) / 8; in esp_init_authenc()
1113 aalg_desc = xfrm_aalg_get_byname(x->aalg->alg_name, 0); in esp_init_authenc()
1116 err = -EINVAL; in esp_init_authenc()
1117 if (aalg_desc->uinfo.auth.icv_fullbits / 8 != in esp_init_authenc()
1120 x->aalg->alg_name, in esp_init_authenc()
1122 aalg_desc->uinfo.auth.icv_fullbits / 8); in esp_init_authenc()
1127 aead, x->aalg->alg_trunc_len / 8); in esp_init_authenc()
1132 param->enckeylen = cpu_to_be32((x->ealg->alg_key_len + 7) / 8); in esp_init_authenc()
1133 memcpy(p, x->ealg->alg_key, (x->ealg->alg_key_len + 7) / 8); in esp_init_authenc()
1150 x->data = NULL; in esp_init_state()
1152 if (x->aead) in esp_init_state()
1160 aead = x->data; in esp_init_state()
1162 x->props.header_len = sizeof(struct ip_esp_hdr) + in esp_init_state()
1164 if (x->props.mode == XFRM_MODE_TUNNEL) in esp_init_state()
1165 x->props.header_len += sizeof(struct iphdr); in esp_init_state()
1166 else if (x->props.mode == XFRM_MODE_BEET && x->sel.family != AF_INET6) in esp_init_state()
1167 x->props.header_len += IPV4_BEET_PHMAXLEN; in esp_init_state()
1168 if (x->encap) { in esp_init_state()
1169 struct xfrm_encap_tmpl *encap = x->encap; in esp_init_state()
1171 switch (encap->encap_type) { in esp_init_state()
1173 err = -EINVAL; in esp_init_state()
1176 x->props.header_len += sizeof(struct udphdr); in esp_init_state()
1179 x->props.header_len += sizeof(struct udphdr) + 2 * sizeof(u32); in esp_init_state()
1186 x->props.header_len += 2; in esp_init_state()
1193 x->props.trailer_len = align + 1 + crypto_aead_authsize(aead); in esp_init_state()
1228 return -EAGAIN; in esp4_init()
1233 return -EAGAIN; in esp4_init()