Lines Matching +full:xo +full:- +full:1
1 // SPDX-License-Identifier: GPL-2.0-only
37 #define ESP_SKB_CB(__skb) ((struct esp_skb_cb *)&((__skb)->cb[0]))
57 ~(crypto_tfm_ctx_alignment() - 1);
78 crypto_aead_alignmask(aead) + 1) : tmp + extralen;
94 return (void *)ALIGN((unsigned long)(req + 1) +
101 struct crypto_aead *aead = x->data;
107 if (x->props.flags & XFRM_STATE_ESN)
114 * Skip the first sg which comes from skb->data.
116 if (req->src != req->dst)
117 for (sg = sg_next(req->src); sg; sg = sg_next(sg))
119 skb->pp_recycle);
125 struct xfrm_encap_tmpl *encap = x->encap;
130 spin_lock_bh(&x->lock);
131 sport = encap->encap_sport;
132 dport = encap->encap_dport;
133 spin_unlock_bh(&x->lock);
135 sk = inet_lookup_established(net, x->id.daddr.a4, dport,
136 x->props.saddr.a4, sport, 0);
138 return ERR_PTR(-ENOENT);
142 return ERR_PTR(-EINVAL);
180 struct xfrm_state *x = dst->xfrm;
197 return err ?: -EINPROGRESS;
202 WARN_ON(1);
203 return -EOPNOTSUPP;
210 struct xfrm_offload *xo = xfrm_offload(skb);
214 if (xo && (xo->flags & XFRM_DEV_RESUME)) {
217 x = sp->xvec[sp->len - 1];
219 x = skb_dst(skb)->xfrm;
222 tmp = ESP_SKB_CB(skb)->tmp;
226 if (xo && (xo->flags & XFRM_DEV_RESUME)) {
233 skb_push(skb, skb->data - skb_mac_header(skb));
238 x->encap && x->encap->encap_type == TCP_ENCAP_ESPINTCP)
248 struct ip_esp_hdr *esph = (void *)(skb->data + offset);
249 void *tmp = ESP_SKB_CB(skb)->tmp;
252 esph->seq_no = esph->spi;
253 esph->spi = *seqhi;
258 void *tmp = ESP_SKB_CB(skb)->tmp;
261 esp_restore_header(skb, skb_transport_offset(skb) + extra->esphoff -
274 if ((x->props.flags & XFRM_STATE_ESN)) {
276 struct xfrm_offload *xo = xfrm_offload(skb);
278 if (xo)
279 seqhi = xo->seq.hi;
281 seqhi = XFRM_SKB_CB(skb)->seq.output.hi;
283 extra->esphoff = (unsigned char *)esph -
285 esph = (struct ip_esp_hdr *)((unsigned char *)esph - 4);
286 extra->seqhi = esph->spi;
287 esph->seq_no = htonl(seqhi);
290 esph->spi = x->id.spi;
311 struct xfrm_offload *xo = xfrm_offload(skb);
313 len = skb->len + esp->tailen - skb_transport_offset(skb);
315 return ERR_PTR(-EMSGSIZE);
317 uh = (struct udphdr *)esp->esph;
318 uh->source = sport;
319 uh->dest = dport;
320 uh->len = htons(len);
321 uh->check = 0;
323 /* For IPv4 ESP with UDP encapsulation, if xo is not null, the skb is in the crypto offload
327 if (!xo || encap_type != UDP_ENCAP_ESPINUDP)
330 return (struct ip_esp_hdr *)(uh + 1);
338 __be16 *lenp = (void *)esp->esph;
343 len = skb->len + esp->tailen - skb_transport_offset(skb);
345 return ERR_PTR(-EMSGSIZE);
357 esph = (struct ip_esp_hdr *)(lenp + 1);
366 return ERR_PTR(-EOPNOTSUPP);
373 struct xfrm_encap_tmpl *encap = x->encap;
378 spin_lock_bh(&x->lock);
379 sport = encap->encap_sport;
380 dport = encap->encap_dport;
381 encap_type = encap->encap_type;
382 spin_unlock_bh(&x->lock);
397 esp->esph = esph;
409 int tailen = esp->tailen;
411 /* this is non-NULL only with TCP/UDP Encapsulation */
412 if (x->encap) {
420 ALIGN(skb->data_len, L1_CACHE_BYTES) > PAGE_SIZE)
425 nfrags = 1;
430 } else if ((skb_shinfo(skb)->nr_frags < MAX_SKB_FRAGS)
433 struct sock *sk = skb->sk;
434 struct page_frag *pfrag = &x->xfrag;
436 esp->inplace = false;
440 spin_lock_bh(&x->lock);
443 spin_unlock_bh(&x->lock);
447 page = pfrag->page;
450 tail = page_address(page) + pfrag->offset;
452 esp_output_fill_trailer(tail, esp->tfclen, esp->plen, esp->proto);
454 nfrags = skb_shinfo(skb)->nr_frags;
456 __skb_fill_page_desc(skb, nfrags, page, pfrag->offset,
458 skb_shinfo(skb)->nr_frags = ++nfrags;
460 pfrag->offset = pfrag->offset + allocsize;
462 spin_unlock_bh(&x->lock);
468 refcount_add(tailen, &sk->sk_wmem_alloc);
475 esph_offset = (unsigned char *)esp->esph - skb_transport_header(skb);
481 esp->esph = (struct ip_esp_hdr *)(skb_transport_header(skb) + esph_offset);
484 esp_output_fill_trailer(tail, esp->tfclen, esp->plen, esp->proto);
506 int err = -ENOMEM;
511 if (x->props.flags & XFRM_STATE_ESN) {
516 aead = x->data;
520 tmp = esp_alloc_tmp(aead, esp->nfrags + 2, extralen);
529 if (esp->inplace)
532 dsg = &sg[esp->nfrags];
534 esph = esp_output_set_extra(skb, x, esp->esph, extra);
535 esp->esph = esph;
537 sg_init_table(sg, esp->nfrags);
539 (unsigned char *)esph - skb->data,
540 assoclen + ivlen + esp->clen + alen);
544 if (!esp->inplace) {
546 struct page_frag *pfrag = &x->xfrag;
548 allocsize = ALIGN(skb->data_len, L1_CACHE_BYTES);
550 spin_lock_bh(&x->lock);
552 spin_unlock_bh(&x->lock);
556 skb_shinfo(skb)->nr_frags = 1;
558 page = pfrag->page;
561 __skb_fill_page_desc(skb, 0, page, pfrag->offset, skb->data_len);
562 pfrag->offset = pfrag->offset + allocsize;
563 spin_unlock_bh(&x->lock);
565 sg_init_table(dsg, skb_shinfo(skb)->nr_frags + 1);
567 (unsigned char *)esph - skb->data,
568 assoclen + ivlen + esp->clen + alen);
573 if ((x->props.flags & XFRM_STATE_ESN))
578 aead_request_set_crypt(req, sg, dsg, ivlen + esp->clen, iv);
582 memcpy(iv + ivlen - min(ivlen, 8), (u8 *)&esp->seqno + 8 - min(ivlen, 8),
585 ESP_SKB_CB(skb)->tmp = tmp;
589 case -EINPROGRESS:
592 case -ENOSPC:
597 if ((x->props.flags & XFRM_STATE_ESN))
604 if (!err && x->encap && x->encap->encap_type == TCP_ENCAP_ESPINTCP)
629 aead = x->data;
633 if (x->tfcpad) {
637 padto = min(x->tfcpad, xfrm_state_mtu(x, dst->child_mtu_cached));
638 if (skb->len < padto)
639 esp.tfclen = padto - skb->len;
642 esp.clen = ALIGN(skb->len + 2 + esp.tfclen, blksize);
643 esp.plen = esp.clen - skb->len - esp.tfclen;
653 esph->spi = x->id.spi;
655 esph->seq_no = htonl(XFRM_SKB_CB(skb)->seq.output.low);
656 esp.seqno = cpu_to_be64(XFRM_SKB_CB(skb)->seq.output.low +
657 ((u64)XFRM_SKB_CB(skb)->seq.output.hi << 32));
659 skb_push(skb, -skb_network_offset(skb));
667 struct crypto_aead *aead = x->data;
676 elen = skb->len - hlen;
678 if (skb_copy_bits(skb, skb->len - alen - 2, nexthdr, 2))
681 ret = -EINVAL;
685 padlen + 2, elen - alen);
690 if (skb->ip_summed == CHECKSUM_COMPLETE) {
691 csumdiff = skb_checksum(skb, skb->len - trimlen, trimlen, 0);
692 skb->csum = csum_block_sub(skb->csum, csumdiff,
693 skb->len - trimlen);
695 ret = pskb_trim(skb, skb->len - trimlen);
699 ret = nexthdr[1];
709 struct xfrm_offload *xo = xfrm_offload(skb);
710 struct crypto_aead *aead = x->data;
714 if (!xo || !(xo->flags & CRYPTO_DONE))
715 kfree(ESP_SKB_CB(skb)->tmp);
725 ihl = iph->ihl * 4;
727 if (x->encap) {
728 struct xfrm_encap_tmpl *encap = x->encap;
733 switch (x->encap->encap_type) {
735 source = th->source;
738 source = uh->source;
741 WARN_ON_ONCE(1);
742 err = -EINVAL;
747 * 1) if the NAT-T peer's IP or port changed then
752 if (iph->saddr != x->props.saddr.a4 ||
753 source != encap->encap_sport) {
756 ipaddr.a4 = iph->saddr;
770 * of NAT-T in Transport Mode, or
771 * perform other post-processing fixes
772 * as per draft-ietf-ipsec-udp-encaps-06,
775 if (x->props.mode == XFRM_MODE_TRANSPORT)
776 skb->ip_summed = CHECKSUM_UNNECESSARY;
780 if (x->props.mode == XFRM_MODE_TUNNEL ||
781 x->props.mode == XFRM_MODE_IPTFS)
784 skb_set_transport_header(skb, -ihl);
788 err = -EINVAL;
817 if ((x->props.flags & XFRM_STATE_ESN)) {
819 *seqhi = esph->spi;
820 esph->spi = esph->seq_no;
821 esph->seq_no = XFRM_SKB_CB(skb)->seq.input.hi;
834 * Note: detecting truncated vs. non-truncated authentication data is very
840 struct crypto_aead *aead = x->data;
844 int elen = skb->len - sizeof(struct ip_esp_hdr) - ivlen;
852 int err = -EINVAL;
863 if (x->props.flags & XFRM_STATE_ESN) {
870 nfrags = 1;
874 nfrags = skb_shinfo(skb)->nr_frags;
888 err = -ENOMEM;
893 ESP_SKB_CB(skb)->tmp = tmp;
902 err = skb_to_sgvec(skb, sg, 0, skb->len);
908 skb->ip_summed = CHECKSUM_NONE;
910 if ((x->props.flags & XFRM_STATE_ESN))
919 if (err == -EINPROGRESS)
922 if ((x->props.flags & XFRM_STATE_ESN))
933 struct net *net = dev_net(skb->dev);
934 const struct iphdr *iph = (const struct iphdr *)skb->data;
935 struct ip_esp_hdr *esph = (struct ip_esp_hdr *)(skb->data+(iph->ihl<<2));
938 switch (icmp_hdr(skb)->type) {
940 if (icmp_hdr(skb)->code != ICMP_FRAG_NEEDED)
949 x = xfrm_state_lookup(net, skb->mark, (const xfrm_address_t *)&iph->daddr,
950 esph->spi, IPPROTO_ESP, AF_INET);
954 if (icmp_hdr(skb)->type == ICMP_DEST_UNREACH)
965 struct crypto_aead *aead = x->data;
980 x->geniv, x->aead->alg_name) >= CRYPTO_MAX_ALG_NAME) {
982 return -ENAMETOOLONG;
990 x->data = aead;
992 err = crypto_aead_setkey(aead, x->aead->alg_key,
993 (x->aead->alg_key_len + 7) / 8);
997 err = crypto_aead_setauthsize(aead, x->aead->alg_icv_len / 8);
1020 err = -ENAMETOOLONG;
1022 if ((x->props.flags & XFRM_STATE_ESN)) {
1025 x->geniv ?: "", x->geniv ? "(" : "",
1026 x->aalg ? x->aalg->alg_name : "digest_null",
1027 x->ealg->alg_name,
1028 x->geniv ? ")" : "") >= CRYPTO_MAX_ALG_NAME) {
1035 x->geniv ?: "", x->geniv ? "(" : "",
1036 x->aalg ? x->aalg->alg_name : "digest_null",
1037 x->ealg->alg_name,
1038 x->geniv ? ")" : "") >= CRYPTO_MAX_ALG_NAME) {
1051 x->data = aead;
1053 keylen = (x->aalg ? (x->aalg->alg_key_len + 7) / 8 : 0) +
1054 (x->ealg->alg_key_len + 7) / 8 + RTA_SPACE(sizeof(*param));
1055 err = -ENOMEM;
1062 rta->rta_type = CRYPTO_AUTHENC_KEYA_PARAM;
1063 rta->rta_len = RTA_LENGTH(sizeof(*param));
1067 if (x->aalg) {
1070 memcpy(p, x->aalg->alg_key, (x->aalg->alg_key_len + 7) / 8);
1071 p += (x->aalg->alg_key_len + 7) / 8;
1073 aalg_desc = xfrm_aalg_get_byname(x->aalg->alg_name, 0);
1076 err = -EINVAL;
1077 if (aalg_desc->uinfo.auth.icv_fullbits / 8 !=
1084 aead, x->aalg->alg_trunc_len / 8);
1091 param->enckeylen = cpu_to_be32((x->ealg->alg_key_len + 7) / 8);
1092 memcpy(p, x->ealg->alg_key, (x->ealg->alg_key_len + 7) / 8);
1109 x->data = NULL;
1111 if (x->aead) {
1113 } else if (x->ealg) {
1117 err = -EINVAL;
1123 aead = x->data;
1125 x->props.header_len = sizeof(struct ip_esp_hdr) +
1127 if (x->props.mode == XFRM_MODE_TUNNEL)
1128 x->props.header_len += sizeof(struct iphdr);
1129 else if (x->props.mode == XFRM_MODE_BEET && x->sel.family != AF_INET6)
1130 x->props.header_len += IPV4_BEET_PHMAXLEN;
1131 if (x->encap) {
1132 struct xfrm_encap_tmpl *encap = x->encap;
1134 switch (encap->encap_type) {
1137 err = -EINVAL;
1140 x->props.header_len += sizeof(struct udphdr);
1147 x->props.header_len += 2;
1154 x->props.trailer_len = align + 1 + crypto_aead_authsize(aead);
1188 return -EAGAIN;
1193 return -EAGAIN;