1 /* 2 * DCCP over IPv6 3 * Linux INET6 implementation 4 * 5 * Based on net/dccp6/ipv6.c 6 * 7 * Arnaldo Carvalho de Melo <acme@ghostprotocols.net> 8 * 9 * This program is free software; you can redistribute it and/or 10 * modify it under the terms of the GNU General Public License 11 * as published by the Free Software Foundation; either version 12 * 2 of the License, or (at your option) any later version. 13 */ 14 15 #include <linux/module.h> 16 #include <linux/random.h> 17 #include <linux/slab.h> 18 #include <linux/xfrm.h> 19 20 #include <net/addrconf.h> 21 #include <net/inet_common.h> 22 #include <net/inet_hashtables.h> 23 #include <net/inet_sock.h> 24 #include <net/inet6_connection_sock.h> 25 #include <net/inet6_hashtables.h> 26 #include <net/ip6_route.h> 27 #include <net/ipv6.h> 28 #include <net/protocol.h> 29 #include <net/transp_v6.h> 30 #include <net/ip6_checksum.h> 31 #include <net/xfrm.h> 32 33 #include "dccp.h" 34 #include "ipv6.h" 35 #include "feat.h" 36 37 /* The per-net dccp.v6_ctl_sk is used for sending RSTs and ACKs */ 38 39 static const struct inet_connection_sock_af_ops dccp_ipv6_mapped; 40 static const struct inet_connection_sock_af_ops dccp_ipv6_af_ops; 41 42 static void dccp_v6_hash(struct sock *sk) 43 { 44 if (sk->sk_state != DCCP_CLOSED) { 45 if (inet_csk(sk)->icsk_af_ops == &dccp_ipv6_mapped) { 46 inet_hash(sk); 47 return; 48 } 49 local_bh_disable(); 50 __inet6_hash(sk, NULL); 51 local_bh_enable(); 52 } 53 } 54 55 /* add pseudo-header to DCCP checksum stored in skb->csum */ 56 static inline __sum16 dccp_v6_csum_finish(struct sk_buff *skb, 57 struct in6_addr *saddr, 58 struct in6_addr *daddr) 59 { 60 return csum_ipv6_magic(saddr, daddr, skb->len, IPPROTO_DCCP, skb->csum); 61 } 62 63 static inline void dccp_v6_send_check(struct sock *sk, struct sk_buff *skb) 64 { 65 struct ipv6_pinfo *np = inet6_sk(sk); 66 struct dccp_hdr *dh = dccp_hdr(skb); 67 68 dccp_csum_outgoing(skb); 69 dh->dccph_checksum = dccp_v6_csum_finish(skb, &np->saddr, &np->daddr); 70 } 71 72 static inline __u32 secure_dccpv6_sequence_number(__be32 *saddr, __be32 *daddr, 73 __be16 sport, __be16 dport ) 74 { 75 return secure_tcpv6_sequence_number(saddr, daddr, sport, dport); 76 } 77 78 static inline __u32 dccp_v6_init_sequence(struct sk_buff *skb) 79 { 80 return secure_dccpv6_sequence_number(ipv6_hdr(skb)->daddr.s6_addr32, 81 ipv6_hdr(skb)->saddr.s6_addr32, 82 dccp_hdr(skb)->dccph_dport, 83 dccp_hdr(skb)->dccph_sport ); 84 85 } 86 87 static void dccp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt, 88 u8 type, u8 code, int offset, __be32 info) 89 { 90 struct ipv6hdr *hdr = (struct ipv6hdr *)skb->data; 91 const struct dccp_hdr *dh = (struct dccp_hdr *)(skb->data + offset); 92 struct dccp_sock *dp; 93 struct ipv6_pinfo *np; 94 struct sock *sk; 95 int err; 96 __u64 seq; 97 struct net *net = dev_net(skb->dev); 98 99 if (skb->len < offset + sizeof(*dh) || 100 skb->len < offset + __dccp_basic_hdr_len(dh)) { 101 ICMP6_INC_STATS_BH(net, __in6_dev_get(skb->dev), 102 ICMP6_MIB_INERRORS); 103 return; 104 } 105 106 sk = inet6_lookup(net, &dccp_hashinfo, 107 &hdr->daddr, dh->dccph_dport, 108 &hdr->saddr, dh->dccph_sport, inet6_iif(skb)); 109 110 if (sk == NULL) { 111 ICMP6_INC_STATS_BH(net, __in6_dev_get(skb->dev), 112 ICMP6_MIB_INERRORS); 113 return; 114 } 115 116 if (sk->sk_state == DCCP_TIME_WAIT) { 117 inet_twsk_put(inet_twsk(sk)); 118 return; 119 } 120 121 bh_lock_sock(sk); 122 if (sock_owned_by_user(sk)) 123 NET_INC_STATS_BH(net, LINUX_MIB_LOCKDROPPEDICMPS); 124 125 if (sk->sk_state == DCCP_CLOSED) 126 goto out; 127 128 dp = dccp_sk(sk); 129 seq = dccp_hdr_seq(dh); 130 if ((1 << sk->sk_state) & ~(DCCPF_REQUESTING | DCCPF_LISTEN) && 131 !between48(seq, dp->dccps_awl, dp->dccps_awh)) { 132 NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS); 133 goto out; 134 } 135 136 np = inet6_sk(sk); 137 138 if (type == ICMPV6_PKT_TOOBIG) { 139 struct dst_entry *dst = NULL; 140 141 if (sock_owned_by_user(sk)) 142 goto out; 143 if ((1 << sk->sk_state) & (DCCPF_LISTEN | DCCPF_CLOSED)) 144 goto out; 145 146 /* icmp should have updated the destination cache entry */ 147 dst = __sk_dst_check(sk, np->dst_cookie); 148 if (dst == NULL) { 149 struct inet_sock *inet = inet_sk(sk); 150 struct flowi fl; 151 152 /* BUGGG_FUTURE: Again, it is not clear how 153 to handle rthdr case. Ignore this complexity 154 for now. 155 */ 156 memset(&fl, 0, sizeof(fl)); 157 fl.proto = IPPROTO_DCCP; 158 ipv6_addr_copy(&fl.fl6_dst, &np->daddr); 159 ipv6_addr_copy(&fl.fl6_src, &np->saddr); 160 fl.oif = sk->sk_bound_dev_if; 161 fl.fl_ip_dport = inet->inet_dport; 162 fl.fl_ip_sport = inet->inet_sport; 163 security_sk_classify_flow(sk, &fl); 164 165 err = ip6_dst_lookup(sk, &dst, &fl); 166 if (err) { 167 sk->sk_err_soft = -err; 168 goto out; 169 } 170 171 err = xfrm_lookup(net, &dst, &fl, sk, 0); 172 if (err < 0) { 173 sk->sk_err_soft = -err; 174 goto out; 175 } 176 } else 177 dst_hold(dst); 178 179 if (inet_csk(sk)->icsk_pmtu_cookie > dst_mtu(dst)) { 180 dccp_sync_mss(sk, dst_mtu(dst)); 181 } /* else let the usual retransmit timer handle it */ 182 dst_release(dst); 183 goto out; 184 } 185 186 icmpv6_err_convert(type, code, &err); 187 188 /* Might be for an request_sock */ 189 switch (sk->sk_state) { 190 struct request_sock *req, **prev; 191 case DCCP_LISTEN: 192 if (sock_owned_by_user(sk)) 193 goto out; 194 195 req = inet6_csk_search_req(sk, &prev, dh->dccph_dport, 196 &hdr->daddr, &hdr->saddr, 197 inet6_iif(skb)); 198 if (req == NULL) 199 goto out; 200 201 /* 202 * ICMPs are not backlogged, hence we cannot get an established 203 * socket here. 204 */ 205 WARN_ON(req->sk != NULL); 206 207 if (seq != dccp_rsk(req)->dreq_iss) { 208 NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS); 209 goto out; 210 } 211 212 inet_csk_reqsk_queue_drop(sk, req, prev); 213 goto out; 214 215 case DCCP_REQUESTING: 216 case DCCP_RESPOND: /* Cannot happen. 217 It can, it SYNs are crossed. --ANK */ 218 if (!sock_owned_by_user(sk)) { 219 DCCP_INC_STATS_BH(DCCP_MIB_ATTEMPTFAILS); 220 sk->sk_err = err; 221 /* 222 * Wake people up to see the error 223 * (see connect in sock.c) 224 */ 225 sk->sk_error_report(sk); 226 dccp_done(sk); 227 } else 228 sk->sk_err_soft = err; 229 goto out; 230 } 231 232 if (!sock_owned_by_user(sk) && np->recverr) { 233 sk->sk_err = err; 234 sk->sk_error_report(sk); 235 } else 236 sk->sk_err_soft = err; 237 238 out: 239 bh_unlock_sock(sk); 240 sock_put(sk); 241 } 242 243 244 static int dccp_v6_send_response(struct sock *sk, struct request_sock *req, 245 struct request_values *rv_unused) 246 { 247 struct inet6_request_sock *ireq6 = inet6_rsk(req); 248 struct ipv6_pinfo *np = inet6_sk(sk); 249 struct sk_buff *skb; 250 struct ipv6_txoptions *opt = NULL; 251 struct in6_addr *final_p, final; 252 struct flowi fl; 253 int err = -1; 254 struct dst_entry *dst; 255 256 memset(&fl, 0, sizeof(fl)); 257 fl.proto = IPPROTO_DCCP; 258 ipv6_addr_copy(&fl.fl6_dst, &ireq6->rmt_addr); 259 ipv6_addr_copy(&fl.fl6_src, &ireq6->loc_addr); 260 fl.fl6_flowlabel = 0; 261 fl.oif = ireq6->iif; 262 fl.fl_ip_dport = inet_rsk(req)->rmt_port; 263 fl.fl_ip_sport = inet_rsk(req)->loc_port; 264 security_req_classify_flow(req, &fl); 265 266 opt = np->opt; 267 268 final_p = fl6_update_dst(&fl, opt, &final); 269 270 err = ip6_dst_lookup(sk, &dst, &fl); 271 if (err) 272 goto done; 273 274 if (final_p) 275 ipv6_addr_copy(&fl.fl6_dst, final_p); 276 277 err = xfrm_lookup(sock_net(sk), &dst, &fl, sk, 0); 278 if (err < 0) 279 goto done; 280 281 skb = dccp_make_response(sk, dst, req); 282 if (skb != NULL) { 283 struct dccp_hdr *dh = dccp_hdr(skb); 284 285 dh->dccph_checksum = dccp_v6_csum_finish(skb, 286 &ireq6->loc_addr, 287 &ireq6->rmt_addr); 288 ipv6_addr_copy(&fl.fl6_dst, &ireq6->rmt_addr); 289 err = ip6_xmit(sk, skb, &fl, opt); 290 err = net_xmit_eval(err); 291 } 292 293 done: 294 if (opt != NULL && opt != np->opt) 295 sock_kfree_s(sk, opt, opt->tot_len); 296 dst_release(dst); 297 return err; 298 } 299 300 static void dccp_v6_reqsk_destructor(struct request_sock *req) 301 { 302 dccp_feat_list_purge(&dccp_rsk(req)->dreq_featneg); 303 if (inet6_rsk(req)->pktopts != NULL) 304 kfree_skb(inet6_rsk(req)->pktopts); 305 } 306 307 static void dccp_v6_ctl_send_reset(struct sock *sk, struct sk_buff *rxskb) 308 { 309 struct ipv6hdr *rxip6h; 310 struct sk_buff *skb; 311 struct flowi fl; 312 struct net *net = dev_net(skb_dst(rxskb)->dev); 313 struct sock *ctl_sk = net->dccp.v6_ctl_sk; 314 struct dst_entry *dst; 315 316 if (dccp_hdr(rxskb)->dccph_type == DCCP_PKT_RESET) 317 return; 318 319 if (!ipv6_unicast_destination(rxskb)) 320 return; 321 322 skb = dccp_ctl_make_reset(ctl_sk, rxskb); 323 if (skb == NULL) 324 return; 325 326 rxip6h = ipv6_hdr(rxskb); 327 dccp_hdr(skb)->dccph_checksum = dccp_v6_csum_finish(skb, &rxip6h->saddr, 328 &rxip6h->daddr); 329 330 memset(&fl, 0, sizeof(fl)); 331 ipv6_addr_copy(&fl.fl6_dst, &rxip6h->saddr); 332 ipv6_addr_copy(&fl.fl6_src, &rxip6h->daddr); 333 334 fl.proto = IPPROTO_DCCP; 335 fl.oif = inet6_iif(rxskb); 336 fl.fl_ip_dport = dccp_hdr(skb)->dccph_dport; 337 fl.fl_ip_sport = dccp_hdr(skb)->dccph_sport; 338 security_skb_classify_flow(rxskb, &fl); 339 340 /* sk = NULL, but it is safe for now. RST socket required. */ 341 if (!ip6_dst_lookup(ctl_sk, &dst, &fl)) { 342 if (xfrm_lookup(net, &dst, &fl, NULL, 0) >= 0) { 343 skb_dst_set(skb, dst); 344 ip6_xmit(ctl_sk, skb, &fl, NULL); 345 DCCP_INC_STATS_BH(DCCP_MIB_OUTSEGS); 346 DCCP_INC_STATS_BH(DCCP_MIB_OUTRSTS); 347 return; 348 } 349 } 350 351 kfree_skb(skb); 352 } 353 354 static struct request_sock_ops dccp6_request_sock_ops = { 355 .family = AF_INET6, 356 .obj_size = sizeof(struct dccp6_request_sock), 357 .rtx_syn_ack = dccp_v6_send_response, 358 .send_ack = dccp_reqsk_send_ack, 359 .destructor = dccp_v6_reqsk_destructor, 360 .send_reset = dccp_v6_ctl_send_reset, 361 }; 362 363 static struct sock *dccp_v6_hnd_req(struct sock *sk,struct sk_buff *skb) 364 { 365 const struct dccp_hdr *dh = dccp_hdr(skb); 366 const struct ipv6hdr *iph = ipv6_hdr(skb); 367 struct sock *nsk; 368 struct request_sock **prev; 369 /* Find possible connection requests. */ 370 struct request_sock *req = inet6_csk_search_req(sk, &prev, 371 dh->dccph_sport, 372 &iph->saddr, 373 &iph->daddr, 374 inet6_iif(skb)); 375 if (req != NULL) 376 return dccp_check_req(sk, skb, req, prev); 377 378 nsk = __inet6_lookup_established(sock_net(sk), &dccp_hashinfo, 379 &iph->saddr, dh->dccph_sport, 380 &iph->daddr, ntohs(dh->dccph_dport), 381 inet6_iif(skb)); 382 if (nsk != NULL) { 383 if (nsk->sk_state != DCCP_TIME_WAIT) { 384 bh_lock_sock(nsk); 385 return nsk; 386 } 387 inet_twsk_put(inet_twsk(nsk)); 388 return NULL; 389 } 390 391 return sk; 392 } 393 394 static int dccp_v6_conn_request(struct sock *sk, struct sk_buff *skb) 395 { 396 struct request_sock *req; 397 struct dccp_request_sock *dreq; 398 struct inet6_request_sock *ireq6; 399 struct ipv6_pinfo *np = inet6_sk(sk); 400 const __be32 service = dccp_hdr_request(skb)->dccph_req_service; 401 struct dccp_skb_cb *dcb = DCCP_SKB_CB(skb); 402 403 if (skb->protocol == htons(ETH_P_IP)) 404 return dccp_v4_conn_request(sk, skb); 405 406 if (!ipv6_unicast_destination(skb)) 407 return 0; /* discard, don't send a reset here */ 408 409 if (dccp_bad_service_code(sk, service)) { 410 dcb->dccpd_reset_code = DCCP_RESET_CODE_BAD_SERVICE_CODE; 411 goto drop; 412 } 413 /* 414 * There are no SYN attacks on IPv6, yet... 415 */ 416 dcb->dccpd_reset_code = DCCP_RESET_CODE_TOO_BUSY; 417 if (inet_csk_reqsk_queue_is_full(sk)) 418 goto drop; 419 420 if (sk_acceptq_is_full(sk) && inet_csk_reqsk_queue_young(sk) > 1) 421 goto drop; 422 423 req = inet6_reqsk_alloc(&dccp6_request_sock_ops); 424 if (req == NULL) 425 goto drop; 426 427 if (dccp_reqsk_init(req, dccp_sk(sk), skb)) 428 goto drop_and_free; 429 430 dreq = dccp_rsk(req); 431 if (dccp_parse_options(sk, dreq, skb)) 432 goto drop_and_free; 433 434 if (security_inet_conn_request(sk, skb, req)) 435 goto drop_and_free; 436 437 ireq6 = inet6_rsk(req); 438 ipv6_addr_copy(&ireq6->rmt_addr, &ipv6_hdr(skb)->saddr); 439 ipv6_addr_copy(&ireq6->loc_addr, &ipv6_hdr(skb)->daddr); 440 441 if (ipv6_opt_accepted(sk, skb) || 442 np->rxopt.bits.rxinfo || np->rxopt.bits.rxoinfo || 443 np->rxopt.bits.rxhlim || np->rxopt.bits.rxohlim) { 444 atomic_inc(&skb->users); 445 ireq6->pktopts = skb; 446 } 447 ireq6->iif = sk->sk_bound_dev_if; 448 449 /* So that link locals have meaning */ 450 if (!sk->sk_bound_dev_if && 451 ipv6_addr_type(&ireq6->rmt_addr) & IPV6_ADDR_LINKLOCAL) 452 ireq6->iif = inet6_iif(skb); 453 454 /* 455 * Step 3: Process LISTEN state 456 * 457 * Set S.ISR, S.GSR, S.SWL, S.SWH from packet or Init Cookie 458 * 459 * In fact we defer setting S.GSR, S.SWL, S.SWH to 460 * dccp_create_openreq_child. 461 */ 462 dreq->dreq_isr = dcb->dccpd_seq; 463 dreq->dreq_iss = dccp_v6_init_sequence(skb); 464 dreq->dreq_service = service; 465 466 if (dccp_v6_send_response(sk, req, NULL)) 467 goto drop_and_free; 468 469 inet6_csk_reqsk_queue_hash_add(sk, req, DCCP_TIMEOUT_INIT); 470 return 0; 471 472 drop_and_free: 473 reqsk_free(req); 474 drop: 475 DCCP_INC_STATS_BH(DCCP_MIB_ATTEMPTFAILS); 476 return -1; 477 } 478 479 static struct sock *dccp_v6_request_recv_sock(struct sock *sk, 480 struct sk_buff *skb, 481 struct request_sock *req, 482 struct dst_entry *dst) 483 { 484 struct inet6_request_sock *ireq6 = inet6_rsk(req); 485 struct ipv6_pinfo *newnp, *np = inet6_sk(sk); 486 struct inet_sock *newinet; 487 struct dccp_sock *newdp; 488 struct dccp6_sock *newdp6; 489 struct sock *newsk; 490 struct ipv6_txoptions *opt; 491 492 if (skb->protocol == htons(ETH_P_IP)) { 493 /* 494 * v6 mapped 495 */ 496 newsk = dccp_v4_request_recv_sock(sk, skb, req, dst); 497 if (newsk == NULL) 498 return NULL; 499 500 newdp6 = (struct dccp6_sock *)newsk; 501 newdp = dccp_sk(newsk); 502 newinet = inet_sk(newsk); 503 newinet->pinet6 = &newdp6->inet6; 504 newnp = inet6_sk(newsk); 505 506 memcpy(newnp, np, sizeof(struct ipv6_pinfo)); 507 508 ipv6_addr_set_v4mapped(newinet->inet_daddr, &newnp->daddr); 509 510 ipv6_addr_set_v4mapped(newinet->inet_saddr, &newnp->saddr); 511 512 ipv6_addr_copy(&newnp->rcv_saddr, &newnp->saddr); 513 514 inet_csk(newsk)->icsk_af_ops = &dccp_ipv6_mapped; 515 newsk->sk_backlog_rcv = dccp_v4_do_rcv; 516 newnp->pktoptions = NULL; 517 newnp->opt = NULL; 518 newnp->mcast_oif = inet6_iif(skb); 519 newnp->mcast_hops = ipv6_hdr(skb)->hop_limit; 520 521 /* 522 * No need to charge this sock to the relevant IPv6 refcnt debug socks count 523 * here, dccp_create_openreq_child now does this for us, see the comment in 524 * that function for the gory details. -acme 525 */ 526 527 /* It is tricky place. Until this moment IPv4 tcp 528 worked with IPv6 icsk.icsk_af_ops. 529 Sync it now. 530 */ 531 dccp_sync_mss(newsk, inet_csk(newsk)->icsk_pmtu_cookie); 532 533 return newsk; 534 } 535 536 opt = np->opt; 537 538 if (sk_acceptq_is_full(sk)) 539 goto out_overflow; 540 541 if (dst == NULL) { 542 struct in6_addr *final_p, final; 543 struct flowi fl; 544 545 memset(&fl, 0, sizeof(fl)); 546 fl.proto = IPPROTO_DCCP; 547 ipv6_addr_copy(&fl.fl6_dst, &ireq6->rmt_addr); 548 final_p = fl6_update_dst(&fl, opt, &final); 549 ipv6_addr_copy(&fl.fl6_src, &ireq6->loc_addr); 550 fl.oif = sk->sk_bound_dev_if; 551 fl.fl_ip_dport = inet_rsk(req)->rmt_port; 552 fl.fl_ip_sport = inet_rsk(req)->loc_port; 553 security_sk_classify_flow(sk, &fl); 554 555 if (ip6_dst_lookup(sk, &dst, &fl)) 556 goto out; 557 558 if (final_p) 559 ipv6_addr_copy(&fl.fl6_dst, final_p); 560 561 if ((xfrm_lookup(sock_net(sk), &dst, &fl, sk, 0)) < 0) 562 goto out; 563 } 564 565 newsk = dccp_create_openreq_child(sk, req, skb); 566 if (newsk == NULL) 567 goto out; 568 569 /* 570 * No need to charge this sock to the relevant IPv6 refcnt debug socks 571 * count here, dccp_create_openreq_child now does this for us, see the 572 * comment in that function for the gory details. -acme 573 */ 574 575 __ip6_dst_store(newsk, dst, NULL, NULL); 576 newsk->sk_route_caps = dst->dev->features & ~(NETIF_F_IP_CSUM | 577 NETIF_F_TSO); 578 newdp6 = (struct dccp6_sock *)newsk; 579 newinet = inet_sk(newsk); 580 newinet->pinet6 = &newdp6->inet6; 581 newdp = dccp_sk(newsk); 582 newnp = inet6_sk(newsk); 583 584 memcpy(newnp, np, sizeof(struct ipv6_pinfo)); 585 586 ipv6_addr_copy(&newnp->daddr, &ireq6->rmt_addr); 587 ipv6_addr_copy(&newnp->saddr, &ireq6->loc_addr); 588 ipv6_addr_copy(&newnp->rcv_saddr, &ireq6->loc_addr); 589 newsk->sk_bound_dev_if = ireq6->iif; 590 591 /* Now IPv6 options... 592 593 First: no IPv4 options. 594 */ 595 newinet->opt = NULL; 596 597 /* Clone RX bits */ 598 newnp->rxopt.all = np->rxopt.all; 599 600 /* Clone pktoptions received with SYN */ 601 newnp->pktoptions = NULL; 602 if (ireq6->pktopts != NULL) { 603 newnp->pktoptions = skb_clone(ireq6->pktopts, GFP_ATOMIC); 604 kfree_skb(ireq6->pktopts); 605 ireq6->pktopts = NULL; 606 if (newnp->pktoptions) 607 skb_set_owner_r(newnp->pktoptions, newsk); 608 } 609 newnp->opt = NULL; 610 newnp->mcast_oif = inet6_iif(skb); 611 newnp->mcast_hops = ipv6_hdr(skb)->hop_limit; 612 613 /* 614 * Clone native IPv6 options from listening socket (if any) 615 * 616 * Yes, keeping reference count would be much more clever, but we make 617 * one more one thing there: reattach optmem to newsk. 618 */ 619 if (opt != NULL) { 620 newnp->opt = ipv6_dup_options(newsk, opt); 621 if (opt != np->opt) 622 sock_kfree_s(sk, opt, opt->tot_len); 623 } 624 625 inet_csk(newsk)->icsk_ext_hdr_len = 0; 626 if (newnp->opt != NULL) 627 inet_csk(newsk)->icsk_ext_hdr_len = (newnp->opt->opt_nflen + 628 newnp->opt->opt_flen); 629 630 dccp_sync_mss(newsk, dst_mtu(dst)); 631 632 newinet->inet_daddr = newinet->inet_saddr = LOOPBACK4_IPV6; 633 newinet->inet_rcv_saddr = LOOPBACK4_IPV6; 634 635 __inet6_hash(newsk, NULL); 636 __inet_inherit_port(sk, newsk); 637 638 return newsk; 639 640 out_overflow: 641 NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_LISTENOVERFLOWS); 642 out: 643 NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_LISTENDROPS); 644 if (opt != NULL && opt != np->opt) 645 sock_kfree_s(sk, opt, opt->tot_len); 646 dst_release(dst); 647 return NULL; 648 } 649 650 /* The socket must have it's spinlock held when we get 651 * here. 652 * 653 * We have a potential double-lock case here, so even when 654 * doing backlog processing we use the BH locking scheme. 655 * This is because we cannot sleep with the original spinlock 656 * held. 657 */ 658 static int dccp_v6_do_rcv(struct sock *sk, struct sk_buff *skb) 659 { 660 struct ipv6_pinfo *np = inet6_sk(sk); 661 struct sk_buff *opt_skb = NULL; 662 663 /* Imagine: socket is IPv6. IPv4 packet arrives, 664 goes to IPv4 receive handler and backlogged. 665 From backlog it always goes here. Kerboom... 666 Fortunately, dccp_rcv_established and rcv_established 667 handle them correctly, but it is not case with 668 dccp_v6_hnd_req and dccp_v6_ctl_send_reset(). --ANK 669 */ 670 671 if (skb->protocol == htons(ETH_P_IP)) 672 return dccp_v4_do_rcv(sk, skb); 673 674 if (sk_filter(sk, skb)) 675 goto discard; 676 677 /* 678 * socket locking is here for SMP purposes as backlog rcv is currently 679 * called with bh processing disabled. 680 */ 681 682 /* Do Stevens' IPV6_PKTOPTIONS. 683 684 Yes, guys, it is the only place in our code, where we 685 may make it not affecting IPv4. 686 The rest of code is protocol independent, 687 and I do not like idea to uglify IPv4. 688 689 Actually, all the idea behind IPV6_PKTOPTIONS 690 looks not very well thought. For now we latch 691 options, received in the last packet, enqueued 692 by tcp. Feel free to propose better solution. 693 --ANK (980728) 694 */ 695 if (np->rxopt.all) 696 /* 697 * FIXME: Add handling of IPV6_PKTOPTIONS skb. See the comments below 698 * (wrt ipv6_pktopions) and net/ipv6/tcp_ipv6.c for an example. 699 */ 700 opt_skb = skb_clone(skb, GFP_ATOMIC); 701 702 if (sk->sk_state == DCCP_OPEN) { /* Fast path */ 703 if (dccp_rcv_established(sk, skb, dccp_hdr(skb), skb->len)) 704 goto reset; 705 if (opt_skb) { 706 /* XXX This is where we would goto ipv6_pktoptions. */ 707 __kfree_skb(opt_skb); 708 } 709 return 0; 710 } 711 712 /* 713 * Step 3: Process LISTEN state 714 * If S.state == LISTEN, 715 * If P.type == Request or P contains a valid Init Cookie option, 716 * (* Must scan the packet's options to check for Init 717 * Cookies. Only Init Cookies are processed here, 718 * however; other options are processed in Step 8. This 719 * scan need only be performed if the endpoint uses Init 720 * Cookies *) 721 * (* Generate a new socket and switch to that socket *) 722 * Set S := new socket for this port pair 723 * S.state = RESPOND 724 * Choose S.ISS (initial seqno) or set from Init Cookies 725 * Initialize S.GAR := S.ISS 726 * Set S.ISR, S.GSR, S.SWL, S.SWH from packet or Init Cookies 727 * Continue with S.state == RESPOND 728 * (* A Response packet will be generated in Step 11 *) 729 * Otherwise, 730 * Generate Reset(No Connection) unless P.type == Reset 731 * Drop packet and return 732 * 733 * NOTE: the check for the packet types is done in 734 * dccp_rcv_state_process 735 */ 736 if (sk->sk_state == DCCP_LISTEN) { 737 struct sock *nsk = dccp_v6_hnd_req(sk, skb); 738 739 if (nsk == NULL) 740 goto discard; 741 /* 742 * Queue it on the new socket if the new socket is active, 743 * otherwise we just shortcircuit this and continue with 744 * the new socket.. 745 */ 746 if (nsk != sk) { 747 if (dccp_child_process(sk, nsk, skb)) 748 goto reset; 749 if (opt_skb != NULL) 750 __kfree_skb(opt_skb); 751 return 0; 752 } 753 } 754 755 if (dccp_rcv_state_process(sk, skb, dccp_hdr(skb), skb->len)) 756 goto reset; 757 if (opt_skb) { 758 /* XXX This is where we would goto ipv6_pktoptions. */ 759 __kfree_skb(opt_skb); 760 } 761 return 0; 762 763 reset: 764 dccp_v6_ctl_send_reset(sk, skb); 765 discard: 766 if (opt_skb != NULL) 767 __kfree_skb(opt_skb); 768 kfree_skb(skb); 769 return 0; 770 } 771 772 static int dccp_v6_rcv(struct sk_buff *skb) 773 { 774 const struct dccp_hdr *dh; 775 struct sock *sk; 776 int min_cov; 777 778 /* Step 1: Check header basics */ 779 780 if (dccp_invalid_packet(skb)) 781 goto discard_it; 782 783 /* Step 1: If header checksum is incorrect, drop packet and return. */ 784 if (dccp_v6_csum_finish(skb, &ipv6_hdr(skb)->saddr, 785 &ipv6_hdr(skb)->daddr)) { 786 DCCP_WARN("dropped packet with invalid checksum\n"); 787 goto discard_it; 788 } 789 790 dh = dccp_hdr(skb); 791 792 DCCP_SKB_CB(skb)->dccpd_seq = dccp_hdr_seq(dh); 793 DCCP_SKB_CB(skb)->dccpd_type = dh->dccph_type; 794 795 if (dccp_packet_without_ack(skb)) 796 DCCP_SKB_CB(skb)->dccpd_ack_seq = DCCP_PKT_WITHOUT_ACK_SEQ; 797 else 798 DCCP_SKB_CB(skb)->dccpd_ack_seq = dccp_hdr_ack_seq(skb); 799 800 /* Step 2: 801 * Look up flow ID in table and get corresponding socket */ 802 sk = __inet6_lookup_skb(&dccp_hashinfo, skb, 803 dh->dccph_sport, dh->dccph_dport); 804 /* 805 * Step 2: 806 * If no socket ... 807 */ 808 if (sk == NULL) { 809 dccp_pr_debug("failed to look up flow ID in table and " 810 "get corresponding socket\n"); 811 goto no_dccp_socket; 812 } 813 814 /* 815 * Step 2: 816 * ... or S.state == TIMEWAIT, 817 * Generate Reset(No Connection) unless P.type == Reset 818 * Drop packet and return 819 */ 820 if (sk->sk_state == DCCP_TIME_WAIT) { 821 dccp_pr_debug("sk->sk_state == DCCP_TIME_WAIT: do_time_wait\n"); 822 inet_twsk_put(inet_twsk(sk)); 823 goto no_dccp_socket; 824 } 825 826 /* 827 * RFC 4340, sec. 9.2.1: Minimum Checksum Coverage 828 * o if MinCsCov = 0, only packets with CsCov = 0 are accepted 829 * o if MinCsCov > 0, also accept packets with CsCov >= MinCsCov 830 */ 831 min_cov = dccp_sk(sk)->dccps_pcrlen; 832 if (dh->dccph_cscov && (min_cov == 0 || dh->dccph_cscov < min_cov)) { 833 dccp_pr_debug("Packet CsCov %d does not satisfy MinCsCov %d\n", 834 dh->dccph_cscov, min_cov); 835 /* FIXME: send Data Dropped option (see also dccp_v4_rcv) */ 836 goto discard_and_relse; 837 } 838 839 if (!xfrm6_policy_check(sk, XFRM_POLICY_IN, skb)) 840 goto discard_and_relse; 841 842 return sk_receive_skb(sk, skb, 1) ? -1 : 0; 843 844 no_dccp_socket: 845 if (!xfrm6_policy_check(NULL, XFRM_POLICY_IN, skb)) 846 goto discard_it; 847 /* 848 * Step 2: 849 * If no socket ... 850 * Generate Reset(No Connection) unless P.type == Reset 851 * Drop packet and return 852 */ 853 if (dh->dccph_type != DCCP_PKT_RESET) { 854 DCCP_SKB_CB(skb)->dccpd_reset_code = 855 DCCP_RESET_CODE_NO_CONNECTION; 856 dccp_v6_ctl_send_reset(sk, skb); 857 } 858 859 discard_it: 860 kfree_skb(skb); 861 return 0; 862 863 discard_and_relse: 864 sock_put(sk); 865 goto discard_it; 866 } 867 868 static int dccp_v6_connect(struct sock *sk, struct sockaddr *uaddr, 869 int addr_len) 870 { 871 struct sockaddr_in6 *usin = (struct sockaddr_in6 *)uaddr; 872 struct inet_connection_sock *icsk = inet_csk(sk); 873 struct inet_sock *inet = inet_sk(sk); 874 struct ipv6_pinfo *np = inet6_sk(sk); 875 struct dccp_sock *dp = dccp_sk(sk); 876 struct in6_addr *saddr = NULL, *final_p, final; 877 struct flowi fl; 878 struct dst_entry *dst; 879 int addr_type; 880 int err; 881 882 dp->dccps_role = DCCP_ROLE_CLIENT; 883 884 if (addr_len < SIN6_LEN_RFC2133) 885 return -EINVAL; 886 887 if (usin->sin6_family != AF_INET6) 888 return -EAFNOSUPPORT; 889 890 memset(&fl, 0, sizeof(fl)); 891 892 if (np->sndflow) { 893 fl.fl6_flowlabel = usin->sin6_flowinfo & IPV6_FLOWINFO_MASK; 894 IP6_ECN_flow_init(fl.fl6_flowlabel); 895 if (fl.fl6_flowlabel & IPV6_FLOWLABEL_MASK) { 896 struct ip6_flowlabel *flowlabel; 897 flowlabel = fl6_sock_lookup(sk, fl.fl6_flowlabel); 898 if (flowlabel == NULL) 899 return -EINVAL; 900 ipv6_addr_copy(&usin->sin6_addr, &flowlabel->dst); 901 fl6_sock_release(flowlabel); 902 } 903 } 904 /* 905 * connect() to INADDR_ANY means loopback (BSD'ism). 906 */ 907 if (ipv6_addr_any(&usin->sin6_addr)) 908 usin->sin6_addr.s6_addr[15] = 1; 909 910 addr_type = ipv6_addr_type(&usin->sin6_addr); 911 912 if (addr_type & IPV6_ADDR_MULTICAST) 913 return -ENETUNREACH; 914 915 if (addr_type & IPV6_ADDR_LINKLOCAL) { 916 if (addr_len >= sizeof(struct sockaddr_in6) && 917 usin->sin6_scope_id) { 918 /* If interface is set while binding, indices 919 * must coincide. 920 */ 921 if (sk->sk_bound_dev_if && 922 sk->sk_bound_dev_if != usin->sin6_scope_id) 923 return -EINVAL; 924 925 sk->sk_bound_dev_if = usin->sin6_scope_id; 926 } 927 928 /* Connect to link-local address requires an interface */ 929 if (!sk->sk_bound_dev_if) 930 return -EINVAL; 931 } 932 933 ipv6_addr_copy(&np->daddr, &usin->sin6_addr); 934 np->flow_label = fl.fl6_flowlabel; 935 936 /* 937 * DCCP over IPv4 938 */ 939 if (addr_type == IPV6_ADDR_MAPPED) { 940 u32 exthdrlen = icsk->icsk_ext_hdr_len; 941 struct sockaddr_in sin; 942 943 SOCK_DEBUG(sk, "connect: ipv4 mapped\n"); 944 945 if (__ipv6_only_sock(sk)) 946 return -ENETUNREACH; 947 948 sin.sin_family = AF_INET; 949 sin.sin_port = usin->sin6_port; 950 sin.sin_addr.s_addr = usin->sin6_addr.s6_addr32[3]; 951 952 icsk->icsk_af_ops = &dccp_ipv6_mapped; 953 sk->sk_backlog_rcv = dccp_v4_do_rcv; 954 955 err = dccp_v4_connect(sk, (struct sockaddr *)&sin, sizeof(sin)); 956 if (err) { 957 icsk->icsk_ext_hdr_len = exthdrlen; 958 icsk->icsk_af_ops = &dccp_ipv6_af_ops; 959 sk->sk_backlog_rcv = dccp_v6_do_rcv; 960 goto failure; 961 } 962 ipv6_addr_set_v4mapped(inet->inet_saddr, &np->saddr); 963 ipv6_addr_set_v4mapped(inet->inet_rcv_saddr, &np->rcv_saddr); 964 965 return err; 966 } 967 968 if (!ipv6_addr_any(&np->rcv_saddr)) 969 saddr = &np->rcv_saddr; 970 971 fl.proto = IPPROTO_DCCP; 972 ipv6_addr_copy(&fl.fl6_dst, &np->daddr); 973 ipv6_addr_copy(&fl.fl6_src, saddr ? saddr : &np->saddr); 974 fl.oif = sk->sk_bound_dev_if; 975 fl.fl_ip_dport = usin->sin6_port; 976 fl.fl_ip_sport = inet->inet_sport; 977 security_sk_classify_flow(sk, &fl); 978 979 final_p = fl6_update_dst(&fl, np->opt, &final); 980 981 err = ip6_dst_lookup(sk, &dst, &fl); 982 if (err) 983 goto failure; 984 985 if (final_p) 986 ipv6_addr_copy(&fl.fl6_dst, final_p); 987 988 err = __xfrm_lookup(sock_net(sk), &dst, &fl, sk, XFRM_LOOKUP_WAIT); 989 if (err < 0) { 990 if (err == -EREMOTE) 991 err = ip6_dst_blackhole(sk, &dst, &fl); 992 if (err < 0) 993 goto failure; 994 } 995 996 if (saddr == NULL) { 997 saddr = &fl.fl6_src; 998 ipv6_addr_copy(&np->rcv_saddr, saddr); 999 } 1000 1001 /* set the source address */ 1002 ipv6_addr_copy(&np->saddr, saddr); 1003 inet->inet_rcv_saddr = LOOPBACK4_IPV6; 1004 1005 __ip6_dst_store(sk, dst, NULL, NULL); 1006 1007 icsk->icsk_ext_hdr_len = 0; 1008 if (np->opt != NULL) 1009 icsk->icsk_ext_hdr_len = (np->opt->opt_flen + 1010 np->opt->opt_nflen); 1011 1012 inet->inet_dport = usin->sin6_port; 1013 1014 dccp_set_state(sk, DCCP_REQUESTING); 1015 err = inet6_hash_connect(&dccp_death_row, sk); 1016 if (err) 1017 goto late_failure; 1018 1019 dp->dccps_iss = secure_dccpv6_sequence_number(np->saddr.s6_addr32, 1020 np->daddr.s6_addr32, 1021 inet->inet_sport, 1022 inet->inet_dport); 1023 err = dccp_connect(sk); 1024 if (err) 1025 goto late_failure; 1026 1027 return 0; 1028 1029 late_failure: 1030 dccp_set_state(sk, DCCP_CLOSED); 1031 __sk_dst_reset(sk); 1032 failure: 1033 inet->inet_dport = 0; 1034 sk->sk_route_caps = 0; 1035 return err; 1036 } 1037 1038 static const struct inet_connection_sock_af_ops dccp_ipv6_af_ops = { 1039 .queue_xmit = inet6_csk_xmit, 1040 .send_check = dccp_v6_send_check, 1041 .rebuild_header = inet6_sk_rebuild_header, 1042 .conn_request = dccp_v6_conn_request, 1043 .syn_recv_sock = dccp_v6_request_recv_sock, 1044 .net_header_len = sizeof(struct ipv6hdr), 1045 .setsockopt = ipv6_setsockopt, 1046 .getsockopt = ipv6_getsockopt, 1047 .addr2sockaddr = inet6_csk_addr2sockaddr, 1048 .sockaddr_len = sizeof(struct sockaddr_in6), 1049 .bind_conflict = inet6_csk_bind_conflict, 1050 #ifdef CONFIG_COMPAT 1051 .compat_setsockopt = compat_ipv6_setsockopt, 1052 .compat_getsockopt = compat_ipv6_getsockopt, 1053 #endif 1054 }; 1055 1056 /* 1057 * DCCP over IPv4 via INET6 API 1058 */ 1059 static const struct inet_connection_sock_af_ops dccp_ipv6_mapped = { 1060 .queue_xmit = ip_queue_xmit, 1061 .send_check = dccp_v4_send_check, 1062 .rebuild_header = inet_sk_rebuild_header, 1063 .conn_request = dccp_v6_conn_request, 1064 .syn_recv_sock = dccp_v6_request_recv_sock, 1065 .net_header_len = sizeof(struct iphdr), 1066 .setsockopt = ipv6_setsockopt, 1067 .getsockopt = ipv6_getsockopt, 1068 .addr2sockaddr = inet6_csk_addr2sockaddr, 1069 .sockaddr_len = sizeof(struct sockaddr_in6), 1070 #ifdef CONFIG_COMPAT 1071 .compat_setsockopt = compat_ipv6_setsockopt, 1072 .compat_getsockopt = compat_ipv6_getsockopt, 1073 #endif 1074 }; 1075 1076 /* NOTE: A lot of things set to zero explicitly by call to 1077 * sk_alloc() so need not be done here. 1078 */ 1079 static int dccp_v6_init_sock(struct sock *sk) 1080 { 1081 static __u8 dccp_v6_ctl_sock_initialized; 1082 int err = dccp_init_sock(sk, dccp_v6_ctl_sock_initialized); 1083 1084 if (err == 0) { 1085 if (unlikely(!dccp_v6_ctl_sock_initialized)) 1086 dccp_v6_ctl_sock_initialized = 1; 1087 inet_csk(sk)->icsk_af_ops = &dccp_ipv6_af_ops; 1088 } 1089 1090 return err; 1091 } 1092 1093 static void dccp_v6_destroy_sock(struct sock *sk) 1094 { 1095 dccp_destroy_sock(sk); 1096 inet6_destroy_sock(sk); 1097 } 1098 1099 static struct timewait_sock_ops dccp6_timewait_sock_ops = { 1100 .twsk_obj_size = sizeof(struct dccp6_timewait_sock), 1101 }; 1102 1103 static struct proto dccp_v6_prot = { 1104 .name = "DCCPv6", 1105 .owner = THIS_MODULE, 1106 .close = dccp_close, 1107 .connect = dccp_v6_connect, 1108 .disconnect = dccp_disconnect, 1109 .ioctl = dccp_ioctl, 1110 .init = dccp_v6_init_sock, 1111 .setsockopt = dccp_setsockopt, 1112 .getsockopt = dccp_getsockopt, 1113 .sendmsg = dccp_sendmsg, 1114 .recvmsg = dccp_recvmsg, 1115 .backlog_rcv = dccp_v6_do_rcv, 1116 .hash = dccp_v6_hash, 1117 .unhash = inet_unhash, 1118 .accept = inet_csk_accept, 1119 .get_port = inet_csk_get_port, 1120 .shutdown = dccp_shutdown, 1121 .destroy = dccp_v6_destroy_sock, 1122 .orphan_count = &dccp_orphan_count, 1123 .max_header = MAX_DCCP_HEADER, 1124 .obj_size = sizeof(struct dccp6_sock), 1125 .slab_flags = SLAB_DESTROY_BY_RCU, 1126 .rsk_prot = &dccp6_request_sock_ops, 1127 .twsk_prot = &dccp6_timewait_sock_ops, 1128 .h.hashinfo = &dccp_hashinfo, 1129 #ifdef CONFIG_COMPAT 1130 .compat_setsockopt = compat_dccp_setsockopt, 1131 .compat_getsockopt = compat_dccp_getsockopt, 1132 #endif 1133 }; 1134 1135 static const struct inet6_protocol dccp_v6_protocol = { 1136 .handler = dccp_v6_rcv, 1137 .err_handler = dccp_v6_err, 1138 .flags = INET6_PROTO_NOPOLICY | INET6_PROTO_FINAL, 1139 }; 1140 1141 static const struct proto_ops inet6_dccp_ops = { 1142 .family = PF_INET6, 1143 .owner = THIS_MODULE, 1144 .release = inet6_release, 1145 .bind = inet6_bind, 1146 .connect = inet_stream_connect, 1147 .socketpair = sock_no_socketpair, 1148 .accept = inet_accept, 1149 .getname = inet6_getname, 1150 .poll = dccp_poll, 1151 .ioctl = inet6_ioctl, 1152 .listen = inet_dccp_listen, 1153 .shutdown = inet_shutdown, 1154 .setsockopt = sock_common_setsockopt, 1155 .getsockopt = sock_common_getsockopt, 1156 .sendmsg = inet_sendmsg, 1157 .recvmsg = sock_common_recvmsg, 1158 .mmap = sock_no_mmap, 1159 .sendpage = sock_no_sendpage, 1160 #ifdef CONFIG_COMPAT 1161 .compat_setsockopt = compat_sock_common_setsockopt, 1162 .compat_getsockopt = compat_sock_common_getsockopt, 1163 #endif 1164 }; 1165 1166 static struct inet_protosw dccp_v6_protosw = { 1167 .type = SOCK_DCCP, 1168 .protocol = IPPROTO_DCCP, 1169 .prot = &dccp_v6_prot, 1170 .ops = &inet6_dccp_ops, 1171 .flags = INET_PROTOSW_ICSK, 1172 }; 1173 1174 static int __net_init dccp_v6_init_net(struct net *net) 1175 { 1176 if (dccp_hashinfo.bhash == NULL) 1177 return -ESOCKTNOSUPPORT; 1178 1179 return inet_ctl_sock_create(&net->dccp.v6_ctl_sk, PF_INET6, 1180 SOCK_DCCP, IPPROTO_DCCP, net); 1181 } 1182 1183 static void __net_exit dccp_v6_exit_net(struct net *net) 1184 { 1185 inet_ctl_sock_destroy(net->dccp.v6_ctl_sk); 1186 } 1187 1188 static struct pernet_operations dccp_v6_ops = { 1189 .init = dccp_v6_init_net, 1190 .exit = dccp_v6_exit_net, 1191 }; 1192 1193 static int __init dccp_v6_init(void) 1194 { 1195 int err = proto_register(&dccp_v6_prot, 1); 1196 1197 if (err != 0) 1198 goto out; 1199 1200 err = inet6_add_protocol(&dccp_v6_protocol, IPPROTO_DCCP); 1201 if (err != 0) 1202 goto out_unregister_proto; 1203 1204 inet6_register_protosw(&dccp_v6_protosw); 1205 1206 err = register_pernet_subsys(&dccp_v6_ops); 1207 if (err != 0) 1208 goto out_destroy_ctl_sock; 1209 out: 1210 return err; 1211 1212 out_destroy_ctl_sock: 1213 inet6_del_protocol(&dccp_v6_protocol, IPPROTO_DCCP); 1214 inet6_unregister_protosw(&dccp_v6_protosw); 1215 out_unregister_proto: 1216 proto_unregister(&dccp_v6_prot); 1217 goto out; 1218 } 1219 1220 static void __exit dccp_v6_exit(void) 1221 { 1222 unregister_pernet_subsys(&dccp_v6_ops); 1223 inet6_del_protocol(&dccp_v6_protocol, IPPROTO_DCCP); 1224 inet6_unregister_protosw(&dccp_v6_protosw); 1225 proto_unregister(&dccp_v6_prot); 1226 } 1227 1228 module_init(dccp_v6_init); 1229 module_exit(dccp_v6_exit); 1230 1231 /* 1232 * __stringify doesn't likes enums, so use SOCK_DCCP (6) and IPPROTO_DCCP (33) 1233 * values directly, Also cover the case where the protocol is not specified, 1234 * i.e. net-pf-PF_INET6-proto-0-type-SOCK_DCCP 1235 */ 1236 MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_INET6, 33, 6); 1237 MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_INET6, 0, 6); 1238 MODULE_LICENSE("GPL"); 1239 MODULE_AUTHOR("Arnaldo Carvalho de Melo <acme@mandriva.com>"); 1240 MODULE_DESCRIPTION("DCCPv6 - Datagram Congestion Controlled Protocol"); 1241