1 // SPDX-License-Identifier: GPL-2.0 2 /* Multipath TCP 3 * 4 * Copyright (c) 2021, Red Hat. 5 */ 6 7 #define pr_fmt(fmt) "MPTCP: " fmt 8 9 #include <linux/kernel.h> 10 #include <linux/module.h> 11 #include <net/sock.h> 12 #include <net/protocol.h> 13 #include <net/tcp.h> 14 #include <net/mptcp.h> 15 #include "protocol.h" 16 17 #define MIN_INFO_OPTLEN_SIZE 16 18 #define MIN_FULL_INFO_OPTLEN_SIZE 40 19 20 static struct sock *__mptcp_tcp_fallback(struct mptcp_sock *msk) 21 { 22 msk_owned_by_me(msk); 23 24 if (likely(!__mptcp_check_fallback(msk))) 25 return NULL; 26 27 return msk->first; 28 } 29 30 static u32 sockopt_seq_reset(const struct sock *sk) 31 { 32 sock_owned_by_me(sk); 33 34 /* Highbits contain state. Allows to distinguish sockopt_seq 35 * of listener and established: 36 * s0 = new_listener() 37 * sockopt(s0) - seq is 1 38 * s1 = accept(s0) - s1 inherits seq 1 if listener sk (s0) 39 * sockopt(s0) - seq increments to 2 on s0 40 * sockopt(s1) // seq increments to 2 on s1 (different option) 41 * new ssk completes join, inherits options from s0 // seq 2 42 * Needs sync from mptcp join logic, but ssk->seq == msk->seq 43 * 44 * Set High order bits to sk_state so ssk->seq == msk->seq test 45 * will fail. 46 */ 47 48 return (u32)sk->sk_state << 24u; 49 } 50 51 static void sockopt_seq_inc(struct mptcp_sock *msk) 52 { 53 u32 seq = (msk->setsockopt_seq + 1) & 0x00ffffff; 54 55 msk->setsockopt_seq = sockopt_seq_reset((struct sock *)msk) + seq; 56 } 57 58 static int mptcp_get_int_option(struct mptcp_sock *msk, sockptr_t optval, 59 unsigned int optlen, int *val) 60 { 61 if (optlen < sizeof(int)) 62 return -EINVAL; 63 64 if (copy_from_sockptr(val, optval, sizeof(*val))) 65 return -EFAULT; 66 67 return 0; 68 } 69 70 static void mptcp_sol_socket_sync_intval(struct mptcp_sock *msk, int optname, int val) 71 { 72 struct mptcp_subflow_context *subflow; 73 struct sock *sk = (struct sock *)msk; 74 75 lock_sock(sk); 76 sockopt_seq_inc(msk); 77 78 mptcp_for_each_subflow(msk, subflow) { 79 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 80 bool slow = lock_sock_fast(ssk); 81 82 switch (optname) { 83 case SO_DEBUG: 84 sock_valbool_flag(ssk, SOCK_DBG, !!val); 85 break; 86 case SO_KEEPALIVE: 87 if (ssk->sk_prot->keepalive) 88 ssk->sk_prot->keepalive(ssk, !!val); 89 sock_valbool_flag(ssk, SOCK_KEEPOPEN, !!val); 90 break; 91 case SO_PRIORITY: 92 WRITE_ONCE(ssk->sk_priority, val); 93 break; 94 case SO_SNDBUF: 95 case SO_SNDBUFFORCE: 96 ssk->sk_userlocks |= SOCK_SNDBUF_LOCK; 97 WRITE_ONCE(ssk->sk_sndbuf, sk->sk_sndbuf); 98 mptcp_subflow_ctx(ssk)->cached_sndbuf = sk->sk_sndbuf; 99 break; 100 case SO_RCVBUF: 101 case SO_RCVBUFFORCE: 102 ssk->sk_userlocks |= SOCK_RCVBUF_LOCK; 103 WRITE_ONCE(ssk->sk_rcvbuf, sk->sk_rcvbuf); 104 break; 105 case SO_MARK: 106 if (READ_ONCE(ssk->sk_mark) != sk->sk_mark) { 107 WRITE_ONCE(ssk->sk_mark, sk->sk_mark); 108 sk_dst_reset(ssk); 109 } 110 break; 111 case SO_INCOMING_CPU: 112 WRITE_ONCE(ssk->sk_incoming_cpu, val); 113 break; 114 } 115 116 subflow->setsockopt_seq = msk->setsockopt_seq; 117 unlock_sock_fast(ssk, slow); 118 } 119 120 release_sock(sk); 121 } 122 123 static int mptcp_sol_socket_intval(struct mptcp_sock *msk, int optname, int val) 124 { 125 sockptr_t optval = KERNEL_SOCKPTR(&val); 126 struct sock *sk = (struct sock *)msk; 127 int ret; 128 129 ret = sock_setsockopt(sk->sk_socket, SOL_SOCKET, optname, 130 optval, sizeof(val)); 131 if (ret) 132 return ret; 133 134 mptcp_sol_socket_sync_intval(msk, optname, val); 135 return 0; 136 } 137 138 static void mptcp_so_incoming_cpu(struct mptcp_sock *msk, int val) 139 { 140 struct sock *sk = (struct sock *)msk; 141 142 WRITE_ONCE(sk->sk_incoming_cpu, val); 143 144 mptcp_sol_socket_sync_intval(msk, SO_INCOMING_CPU, val); 145 } 146 147 static int mptcp_setsockopt_sol_socket_tstamp(struct mptcp_sock *msk, int optname, int val) 148 { 149 sockptr_t optval = KERNEL_SOCKPTR(&val); 150 struct mptcp_subflow_context *subflow; 151 struct sock *sk = (struct sock *)msk; 152 int ret; 153 154 ret = sock_setsockopt(sk->sk_socket, SOL_SOCKET, optname, 155 optval, sizeof(val)); 156 if (ret) 157 return ret; 158 159 lock_sock(sk); 160 mptcp_for_each_subflow(msk, subflow) { 161 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 162 bool slow = lock_sock_fast(ssk); 163 164 sock_set_timestamp(sk, optname, !!val); 165 unlock_sock_fast(ssk, slow); 166 } 167 168 release_sock(sk); 169 return 0; 170 } 171 172 static int mptcp_setsockopt_sol_socket_int(struct mptcp_sock *msk, int optname, 173 sockptr_t optval, 174 unsigned int optlen) 175 { 176 int val, ret; 177 178 ret = mptcp_get_int_option(msk, optval, optlen, &val); 179 if (ret) 180 return ret; 181 182 switch (optname) { 183 case SO_KEEPALIVE: 184 case SO_DEBUG: 185 case SO_MARK: 186 case SO_PRIORITY: 187 case SO_SNDBUF: 188 case SO_SNDBUFFORCE: 189 case SO_RCVBUF: 190 case SO_RCVBUFFORCE: 191 return mptcp_sol_socket_intval(msk, optname, val); 192 case SO_INCOMING_CPU: 193 mptcp_so_incoming_cpu(msk, val); 194 return 0; 195 case SO_TIMESTAMP_OLD: 196 case SO_TIMESTAMP_NEW: 197 case SO_TIMESTAMPNS_OLD: 198 case SO_TIMESTAMPNS_NEW: 199 return mptcp_setsockopt_sol_socket_tstamp(msk, optname, val); 200 } 201 202 return -ENOPROTOOPT; 203 } 204 205 static int mptcp_setsockopt_sol_socket_timestamping(struct mptcp_sock *msk, 206 int optname, 207 sockptr_t optval, 208 unsigned int optlen) 209 { 210 struct mptcp_subflow_context *subflow; 211 struct sock *sk = (struct sock *)msk; 212 struct so_timestamping timestamping; 213 int ret; 214 215 if (optlen == sizeof(timestamping)) { 216 if (copy_from_sockptr(×tamping, optval, 217 sizeof(timestamping))) 218 return -EFAULT; 219 } else if (optlen == sizeof(int)) { 220 memset(×tamping, 0, sizeof(timestamping)); 221 222 if (copy_from_sockptr(×tamping.flags, optval, sizeof(int))) 223 return -EFAULT; 224 } else { 225 return -EINVAL; 226 } 227 228 ret = sock_setsockopt(sk->sk_socket, SOL_SOCKET, optname, 229 KERNEL_SOCKPTR(×tamping), 230 sizeof(timestamping)); 231 if (ret) 232 return ret; 233 234 lock_sock(sk); 235 236 mptcp_for_each_subflow(msk, subflow) { 237 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 238 bool slow = lock_sock_fast(ssk); 239 240 sock_set_timestamping(sk, optname, timestamping); 241 unlock_sock_fast(ssk, slow); 242 } 243 244 release_sock(sk); 245 246 return 0; 247 } 248 249 static int mptcp_setsockopt_sol_socket_linger(struct mptcp_sock *msk, sockptr_t optval, 250 unsigned int optlen) 251 { 252 struct mptcp_subflow_context *subflow; 253 struct sock *sk = (struct sock *)msk; 254 struct linger ling; 255 sockptr_t kopt; 256 int ret; 257 258 if (optlen < sizeof(ling)) 259 return -EINVAL; 260 261 if (copy_from_sockptr(&ling, optval, sizeof(ling))) 262 return -EFAULT; 263 264 kopt = KERNEL_SOCKPTR(&ling); 265 ret = sock_setsockopt(sk->sk_socket, SOL_SOCKET, SO_LINGER, kopt, sizeof(ling)); 266 if (ret) 267 return ret; 268 269 lock_sock(sk); 270 sockopt_seq_inc(msk); 271 mptcp_for_each_subflow(msk, subflow) { 272 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 273 bool slow = lock_sock_fast(ssk); 274 275 if (!ling.l_onoff) { 276 sock_reset_flag(ssk, SOCK_LINGER); 277 } else { 278 ssk->sk_lingertime = sk->sk_lingertime; 279 sock_set_flag(ssk, SOCK_LINGER); 280 } 281 282 subflow->setsockopt_seq = msk->setsockopt_seq; 283 unlock_sock_fast(ssk, slow); 284 } 285 286 release_sock(sk); 287 return 0; 288 } 289 290 static int mptcp_setsockopt_sol_socket(struct mptcp_sock *msk, int optname, 291 sockptr_t optval, unsigned int optlen) 292 { 293 struct sock *sk = (struct sock *)msk; 294 struct sock *ssk; 295 int ret; 296 297 switch (optname) { 298 case SO_REUSEPORT: 299 case SO_REUSEADDR: 300 case SO_BINDTODEVICE: 301 case SO_BINDTOIFINDEX: 302 lock_sock(sk); 303 ssk = __mptcp_nmpc_sk(msk); 304 if (IS_ERR(ssk)) { 305 release_sock(sk); 306 return PTR_ERR(ssk); 307 } 308 309 ret = sk_setsockopt(ssk, SOL_SOCKET, optname, optval, optlen); 310 if (ret == 0) { 311 if (optname == SO_REUSEPORT) 312 sk->sk_reuseport = ssk->sk_reuseport; 313 else if (optname == SO_REUSEADDR) 314 sk->sk_reuse = ssk->sk_reuse; 315 else if (optname == SO_BINDTODEVICE) 316 sk->sk_bound_dev_if = ssk->sk_bound_dev_if; 317 else if (optname == SO_BINDTOIFINDEX) 318 sk->sk_bound_dev_if = ssk->sk_bound_dev_if; 319 } 320 release_sock(sk); 321 return ret; 322 case SO_KEEPALIVE: 323 case SO_PRIORITY: 324 case SO_SNDBUF: 325 case SO_SNDBUFFORCE: 326 case SO_RCVBUF: 327 case SO_RCVBUFFORCE: 328 case SO_MARK: 329 case SO_INCOMING_CPU: 330 case SO_DEBUG: 331 case SO_TIMESTAMP_OLD: 332 case SO_TIMESTAMP_NEW: 333 case SO_TIMESTAMPNS_OLD: 334 case SO_TIMESTAMPNS_NEW: 335 return mptcp_setsockopt_sol_socket_int(msk, optname, optval, 336 optlen); 337 case SO_TIMESTAMPING_OLD: 338 case SO_TIMESTAMPING_NEW: 339 return mptcp_setsockopt_sol_socket_timestamping(msk, optname, 340 optval, optlen); 341 case SO_LINGER: 342 return mptcp_setsockopt_sol_socket_linger(msk, optval, optlen); 343 case SO_RCVLOWAT: 344 case SO_RCVTIMEO_OLD: 345 case SO_RCVTIMEO_NEW: 346 case SO_SNDTIMEO_OLD: 347 case SO_SNDTIMEO_NEW: 348 case SO_BUSY_POLL: 349 case SO_PREFER_BUSY_POLL: 350 case SO_BUSY_POLL_BUDGET: 351 /* No need to copy: only relevant for msk */ 352 return sock_setsockopt(sk->sk_socket, SOL_SOCKET, optname, optval, optlen); 353 case SO_NO_CHECK: 354 case SO_DONTROUTE: 355 case SO_BROADCAST: 356 case SO_BSDCOMPAT: 357 case SO_PASSCRED: 358 case SO_PASSPIDFD: 359 case SO_PASSSEC: 360 case SO_RXQ_OVFL: 361 case SO_WIFI_STATUS: 362 case SO_NOFCS: 363 case SO_SELECT_ERR_QUEUE: 364 return 0; 365 } 366 367 /* SO_OOBINLINE is not supported, let's avoid the related mess 368 * SO_ATTACH_FILTER, SO_ATTACH_BPF, SO_ATTACH_REUSEPORT_CBPF, 369 * SO_DETACH_REUSEPORT_BPF, SO_DETACH_FILTER, SO_LOCK_FILTER, 370 * we must be careful with subflows 371 * 372 * SO_ATTACH_REUSEPORT_EBPF is not supported, at it checks 373 * explicitly the sk_protocol field 374 * 375 * SO_PEEK_OFF is unsupported, as it is for plain TCP 376 * SO_MAX_PACING_RATE is unsupported, we must be careful with subflows 377 * SO_CNX_ADVICE is currently unsupported, could possibly be relevant, 378 * but likely needs careful design 379 * 380 * SO_ZEROCOPY is currently unsupported, TODO in sndmsg 381 * SO_TXTIME is currently unsupported 382 */ 383 384 return -EOPNOTSUPP; 385 } 386 387 static int mptcp_setsockopt_v6(struct mptcp_sock *msk, int optname, 388 sockptr_t optval, unsigned int optlen) 389 { 390 struct sock *sk = (struct sock *)msk; 391 int ret = -EOPNOTSUPP; 392 struct sock *ssk; 393 394 switch (optname) { 395 case IPV6_V6ONLY: 396 case IPV6_TRANSPARENT: 397 case IPV6_FREEBIND: 398 lock_sock(sk); 399 ssk = __mptcp_nmpc_sk(msk); 400 if (IS_ERR(ssk)) { 401 release_sock(sk); 402 return PTR_ERR(ssk); 403 } 404 405 ret = tcp_setsockopt(ssk, SOL_IPV6, optname, optval, optlen); 406 if (ret != 0) { 407 release_sock(sk); 408 return ret; 409 } 410 411 sockopt_seq_inc(msk); 412 413 switch (optname) { 414 case IPV6_V6ONLY: 415 sk->sk_ipv6only = ssk->sk_ipv6only; 416 break; 417 case IPV6_TRANSPARENT: 418 inet_assign_bit(TRANSPARENT, sk, 419 inet_test_bit(TRANSPARENT, ssk)); 420 break; 421 case IPV6_FREEBIND: 422 inet_assign_bit(FREEBIND, sk, 423 inet_test_bit(FREEBIND, ssk)); 424 break; 425 } 426 427 release_sock(sk); 428 break; 429 } 430 431 return ret; 432 } 433 434 static bool mptcp_supported_sockopt(int level, int optname) 435 { 436 if (level == SOL_IP) { 437 switch (optname) { 438 /* should work fine */ 439 case IP_FREEBIND: 440 case IP_TRANSPARENT: 441 case IP_BIND_ADDRESS_NO_PORT: 442 case IP_LOCAL_PORT_RANGE: 443 444 /* the following are control cmsg related */ 445 case IP_PKTINFO: 446 case IP_RECVTTL: 447 case IP_RECVTOS: 448 case IP_RECVOPTS: 449 case IP_RETOPTS: 450 case IP_PASSSEC: 451 case IP_RECVORIGDSTADDR: 452 case IP_CHECKSUM: 453 case IP_RECVFRAGSIZE: 454 455 /* common stuff that need some love */ 456 case IP_TOS: 457 case IP_TTL: 458 case IP_MTU_DISCOVER: 459 case IP_RECVERR: 460 461 /* possibly less common may deserve some love */ 462 case IP_MINTTL: 463 464 /* the following is apparently a no-op for plain TCP */ 465 case IP_RECVERR_RFC4884: 466 return true; 467 } 468 469 /* IP_OPTIONS is not supported, needs subflow care */ 470 /* IP_HDRINCL, IP_NODEFRAG are not supported, RAW specific */ 471 /* IP_MULTICAST_TTL, IP_MULTICAST_LOOP, IP_UNICAST_IF, 472 * IP_ADD_MEMBERSHIP, IP_ADD_SOURCE_MEMBERSHIP, IP_DROP_MEMBERSHIP, 473 * IP_DROP_SOURCE_MEMBERSHIP, IP_BLOCK_SOURCE, IP_UNBLOCK_SOURCE, 474 * MCAST_JOIN_GROUP, MCAST_LEAVE_GROUP MCAST_JOIN_SOURCE_GROUP, 475 * MCAST_LEAVE_SOURCE_GROUP, MCAST_BLOCK_SOURCE, MCAST_UNBLOCK_SOURCE, 476 * MCAST_MSFILTER, IP_MULTICAST_ALL are not supported, better not deal 477 * with mcast stuff 478 */ 479 /* IP_IPSEC_POLICY, IP_XFRM_POLICY are nut supported, unrelated here */ 480 return false; 481 } 482 if (level == SOL_IPV6) { 483 switch (optname) { 484 case IPV6_V6ONLY: 485 486 /* the following are control cmsg related */ 487 case IPV6_RECVPKTINFO: 488 case IPV6_2292PKTINFO: 489 case IPV6_RECVHOPLIMIT: 490 case IPV6_2292HOPLIMIT: 491 case IPV6_RECVRTHDR: 492 case IPV6_2292RTHDR: 493 case IPV6_RECVHOPOPTS: 494 case IPV6_2292HOPOPTS: 495 case IPV6_RECVDSTOPTS: 496 case IPV6_2292DSTOPTS: 497 case IPV6_RECVTCLASS: 498 case IPV6_FLOWINFO: 499 case IPV6_RECVPATHMTU: 500 case IPV6_RECVORIGDSTADDR: 501 case IPV6_RECVFRAGSIZE: 502 503 /* the following ones need some love but are quite common */ 504 case IPV6_TCLASS: 505 case IPV6_TRANSPARENT: 506 case IPV6_FREEBIND: 507 case IPV6_PKTINFO: 508 case IPV6_2292PKTOPTIONS: 509 case IPV6_UNICAST_HOPS: 510 case IPV6_MTU_DISCOVER: 511 case IPV6_MTU: 512 case IPV6_RECVERR: 513 case IPV6_FLOWINFO_SEND: 514 case IPV6_FLOWLABEL_MGR: 515 case IPV6_MINHOPCOUNT: 516 case IPV6_DONTFRAG: 517 case IPV6_AUTOFLOWLABEL: 518 519 /* the following one is a no-op for plain TCP */ 520 case IPV6_RECVERR_RFC4884: 521 return true; 522 } 523 524 /* IPV6_HOPOPTS, IPV6_RTHDRDSTOPTS, IPV6_RTHDR, IPV6_DSTOPTS are 525 * not supported 526 */ 527 /* IPV6_MULTICAST_HOPS, IPV6_MULTICAST_LOOP, IPV6_UNICAST_IF, 528 * IPV6_MULTICAST_IF, IPV6_ADDRFORM, 529 * IPV6_ADD_MEMBERSHIP, IPV6_DROP_MEMBERSHIP, IPV6_JOIN_ANYCAST, 530 * IPV6_LEAVE_ANYCAST, IPV6_MULTICAST_ALL, MCAST_JOIN_GROUP, MCAST_LEAVE_GROUP, 531 * MCAST_JOIN_SOURCE_GROUP, MCAST_LEAVE_SOURCE_GROUP, 532 * MCAST_BLOCK_SOURCE, MCAST_UNBLOCK_SOURCE, MCAST_MSFILTER 533 * are not supported better not deal with mcast 534 */ 535 /* IPV6_ROUTER_ALERT, IPV6_ROUTER_ALERT_ISOLATE are not supported, since are evil */ 536 537 /* IPV6_IPSEC_POLICY, IPV6_XFRM_POLICY are not supported */ 538 /* IPV6_ADDR_PREFERENCES is not supported, we must be careful with subflows */ 539 return false; 540 } 541 if (level == SOL_TCP) { 542 switch (optname) { 543 /* the following are no-op or should work just fine */ 544 case TCP_THIN_DUPACK: 545 case TCP_DEFER_ACCEPT: 546 547 /* the following need some love */ 548 case TCP_MAXSEG: 549 case TCP_NODELAY: 550 case TCP_THIN_LINEAR_TIMEOUTS: 551 case TCP_CONGESTION: 552 case TCP_CORK: 553 case TCP_KEEPIDLE: 554 case TCP_KEEPINTVL: 555 case TCP_KEEPCNT: 556 case TCP_SYNCNT: 557 case TCP_SAVE_SYN: 558 case TCP_LINGER2: 559 case TCP_WINDOW_CLAMP: 560 case TCP_QUICKACK: 561 case TCP_USER_TIMEOUT: 562 case TCP_TIMESTAMP: 563 case TCP_NOTSENT_LOWAT: 564 case TCP_TX_DELAY: 565 case TCP_INQ: 566 case TCP_FASTOPEN: 567 case TCP_FASTOPEN_CONNECT: 568 case TCP_FASTOPEN_KEY: 569 case TCP_FASTOPEN_NO_COOKIE: 570 return true; 571 } 572 573 /* TCP_MD5SIG, TCP_MD5SIG_EXT are not supported, MD5 is not compatible with MPTCP */ 574 575 /* TCP_REPAIR, TCP_REPAIR_QUEUE, TCP_QUEUE_SEQ, TCP_REPAIR_OPTIONS, 576 * TCP_REPAIR_WINDOW are not supported, better avoid this mess 577 */ 578 } 579 return false; 580 } 581 582 static int mptcp_setsockopt_sol_tcp_congestion(struct mptcp_sock *msk, sockptr_t optval, 583 unsigned int optlen) 584 { 585 struct mptcp_subflow_context *subflow; 586 struct sock *sk = (struct sock *)msk; 587 char name[TCP_CA_NAME_MAX]; 588 bool cap_net_admin; 589 int ret; 590 591 if (optlen < 1) 592 return -EINVAL; 593 594 ret = strncpy_from_sockptr(name, optval, 595 min_t(long, TCP_CA_NAME_MAX - 1, optlen)); 596 if (ret < 0) 597 return -EFAULT; 598 599 name[ret] = 0; 600 601 cap_net_admin = ns_capable(sock_net(sk)->user_ns, CAP_NET_ADMIN); 602 603 ret = 0; 604 lock_sock(sk); 605 sockopt_seq_inc(msk); 606 mptcp_for_each_subflow(msk, subflow) { 607 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 608 int err; 609 610 lock_sock(ssk); 611 err = tcp_set_congestion_control(ssk, name, true, cap_net_admin); 612 if (err < 0 && ret == 0) 613 ret = err; 614 subflow->setsockopt_seq = msk->setsockopt_seq; 615 release_sock(ssk); 616 } 617 618 if (ret == 0) 619 strscpy(msk->ca_name, name, sizeof(msk->ca_name)); 620 621 release_sock(sk); 622 return ret; 623 } 624 625 static int __mptcp_setsockopt_set_val(struct mptcp_sock *msk, int max, 626 int (*set_val)(struct sock *, int), 627 int *msk_val, int val) 628 { 629 struct mptcp_subflow_context *subflow; 630 int err = 0; 631 632 mptcp_for_each_subflow(msk, subflow) { 633 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 634 int ret; 635 636 lock_sock(ssk); 637 ret = set_val(ssk, val); 638 err = err ? : ret; 639 release_sock(ssk); 640 } 641 642 if (!err) { 643 *msk_val = val; 644 sockopt_seq_inc(msk); 645 } 646 647 return err; 648 } 649 650 static int __mptcp_setsockopt_sol_tcp_cork(struct mptcp_sock *msk, int val) 651 { 652 struct mptcp_subflow_context *subflow; 653 struct sock *sk = (struct sock *)msk; 654 655 sockopt_seq_inc(msk); 656 msk->cork = !!val; 657 mptcp_for_each_subflow(msk, subflow) { 658 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 659 660 lock_sock(ssk); 661 __tcp_sock_set_cork(ssk, !!val); 662 release_sock(ssk); 663 } 664 if (!val) 665 mptcp_check_and_set_pending(sk); 666 667 return 0; 668 } 669 670 static int __mptcp_setsockopt_sol_tcp_nodelay(struct mptcp_sock *msk, int val) 671 { 672 struct mptcp_subflow_context *subflow; 673 struct sock *sk = (struct sock *)msk; 674 675 sockopt_seq_inc(msk); 676 msk->nodelay = !!val; 677 mptcp_for_each_subflow(msk, subflow) { 678 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 679 680 lock_sock(ssk); 681 __tcp_sock_set_nodelay(ssk, !!val); 682 release_sock(ssk); 683 } 684 if (val) 685 mptcp_check_and_set_pending(sk); 686 return 0; 687 } 688 689 static int mptcp_setsockopt_sol_ip_set(struct mptcp_sock *msk, int optname, 690 sockptr_t optval, unsigned int optlen) 691 { 692 struct sock *sk = (struct sock *)msk; 693 struct sock *ssk; 694 int err; 695 696 err = ip_setsockopt(sk, SOL_IP, optname, optval, optlen); 697 if (err != 0) 698 return err; 699 700 lock_sock(sk); 701 702 ssk = __mptcp_nmpc_sk(msk); 703 if (IS_ERR(ssk)) { 704 release_sock(sk); 705 return PTR_ERR(ssk); 706 } 707 708 switch (optname) { 709 case IP_FREEBIND: 710 inet_assign_bit(FREEBIND, ssk, inet_test_bit(FREEBIND, sk)); 711 break; 712 case IP_TRANSPARENT: 713 inet_assign_bit(TRANSPARENT, ssk, 714 inet_test_bit(TRANSPARENT, sk)); 715 break; 716 case IP_BIND_ADDRESS_NO_PORT: 717 inet_assign_bit(BIND_ADDRESS_NO_PORT, ssk, 718 inet_test_bit(BIND_ADDRESS_NO_PORT, sk)); 719 break; 720 case IP_LOCAL_PORT_RANGE: 721 WRITE_ONCE(inet_sk(ssk)->local_port_range, 722 READ_ONCE(inet_sk(sk)->local_port_range)); 723 break; 724 default: 725 release_sock(sk); 726 WARN_ON_ONCE(1); 727 return -EOPNOTSUPP; 728 } 729 730 sockopt_seq_inc(msk); 731 release_sock(sk); 732 return 0; 733 } 734 735 static int mptcp_setsockopt_v4_set_tos(struct mptcp_sock *msk, int optname, 736 sockptr_t optval, unsigned int optlen) 737 { 738 struct mptcp_subflow_context *subflow; 739 struct sock *sk = (struct sock *)msk; 740 int err, val; 741 742 err = ip_setsockopt(sk, SOL_IP, optname, optval, optlen); 743 744 if (err != 0) 745 return err; 746 747 lock_sock(sk); 748 sockopt_seq_inc(msk); 749 val = READ_ONCE(inet_sk(sk)->tos); 750 mptcp_for_each_subflow(msk, subflow) { 751 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 752 bool slow; 753 754 slow = lock_sock_fast(ssk); 755 __ip_sock_set_tos(ssk, val); 756 unlock_sock_fast(ssk, slow); 757 } 758 release_sock(sk); 759 760 return 0; 761 } 762 763 static int mptcp_setsockopt_v4(struct mptcp_sock *msk, int optname, 764 sockptr_t optval, unsigned int optlen) 765 { 766 switch (optname) { 767 case IP_FREEBIND: 768 case IP_TRANSPARENT: 769 case IP_BIND_ADDRESS_NO_PORT: 770 case IP_LOCAL_PORT_RANGE: 771 return mptcp_setsockopt_sol_ip_set(msk, optname, optval, optlen); 772 case IP_TOS: 773 return mptcp_setsockopt_v4_set_tos(msk, optname, optval, optlen); 774 } 775 776 return -EOPNOTSUPP; 777 } 778 779 static int mptcp_setsockopt_first_sf_only(struct mptcp_sock *msk, int level, int optname, 780 sockptr_t optval, unsigned int optlen) 781 { 782 struct sock *sk = (struct sock *)msk; 783 struct sock *ssk; 784 int ret; 785 786 /* Limit to first subflow, before the connection establishment */ 787 lock_sock(sk); 788 ssk = __mptcp_nmpc_sk(msk); 789 if (IS_ERR(ssk)) { 790 ret = PTR_ERR(ssk); 791 goto unlock; 792 } 793 794 ret = tcp_setsockopt(ssk, level, optname, optval, optlen); 795 796 unlock: 797 release_sock(sk); 798 return ret; 799 } 800 801 static int mptcp_setsockopt_all_sf(struct mptcp_sock *msk, int level, 802 int optname, sockptr_t optval, 803 unsigned int optlen) 804 { 805 struct mptcp_subflow_context *subflow; 806 int ret = 0; 807 808 mptcp_for_each_subflow(msk, subflow) { 809 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 810 811 ret = tcp_setsockopt(ssk, level, optname, optval, optlen); 812 if (ret) 813 break; 814 } 815 return ret; 816 } 817 818 static int mptcp_setsockopt_sol_tcp(struct mptcp_sock *msk, int optname, 819 sockptr_t optval, unsigned int optlen) 820 { 821 struct sock *sk = (void *)msk; 822 int ret, val; 823 824 switch (optname) { 825 case TCP_ULP: 826 return -EOPNOTSUPP; 827 case TCP_CONGESTION: 828 return mptcp_setsockopt_sol_tcp_congestion(msk, optval, optlen); 829 case TCP_DEFER_ACCEPT: 830 /* See tcp.c: TCP_DEFER_ACCEPT does not fail */ 831 mptcp_setsockopt_first_sf_only(msk, SOL_TCP, optname, optval, optlen); 832 return 0; 833 case TCP_FASTOPEN: 834 case TCP_FASTOPEN_CONNECT: 835 case TCP_FASTOPEN_KEY: 836 case TCP_FASTOPEN_NO_COOKIE: 837 return mptcp_setsockopt_first_sf_only(msk, SOL_TCP, optname, 838 optval, optlen); 839 } 840 841 ret = mptcp_get_int_option(msk, optval, optlen, &val); 842 if (ret) 843 return ret; 844 845 lock_sock(sk); 846 switch (optname) { 847 case TCP_INQ: 848 if (val < 0 || val > 1) 849 ret = -EINVAL; 850 else 851 msk->recvmsg_inq = !!val; 852 break; 853 case TCP_NOTSENT_LOWAT: 854 WRITE_ONCE(msk->notsent_lowat, val); 855 mptcp_write_space(sk); 856 break; 857 case TCP_CORK: 858 ret = __mptcp_setsockopt_sol_tcp_cork(msk, val); 859 break; 860 case TCP_NODELAY: 861 ret = __mptcp_setsockopt_sol_tcp_nodelay(msk, val); 862 break; 863 case TCP_KEEPIDLE: 864 ret = __mptcp_setsockopt_set_val(msk, MAX_TCP_KEEPIDLE, 865 &tcp_sock_set_keepidle_locked, 866 &msk->keepalive_idle, val); 867 break; 868 case TCP_KEEPINTVL: 869 ret = __mptcp_setsockopt_set_val(msk, MAX_TCP_KEEPINTVL, 870 &tcp_sock_set_keepintvl, 871 &msk->keepalive_intvl, val); 872 break; 873 case TCP_KEEPCNT: 874 ret = __mptcp_setsockopt_set_val(msk, MAX_TCP_KEEPCNT, 875 &tcp_sock_set_keepcnt, 876 &msk->keepalive_cnt, 877 val); 878 break; 879 case TCP_MAXSEG: 880 msk->maxseg = val; 881 ret = mptcp_setsockopt_all_sf(msk, SOL_TCP, optname, optval, 882 optlen); 883 break; 884 default: 885 ret = -ENOPROTOOPT; 886 } 887 888 release_sock(sk); 889 return ret; 890 } 891 892 int mptcp_setsockopt(struct sock *sk, int level, int optname, 893 sockptr_t optval, unsigned int optlen) 894 { 895 struct mptcp_sock *msk = mptcp_sk(sk); 896 struct sock *ssk; 897 898 pr_debug("msk=%p\n", msk); 899 900 if (level == SOL_SOCKET) 901 return mptcp_setsockopt_sol_socket(msk, optname, optval, optlen); 902 903 if (!mptcp_supported_sockopt(level, optname)) 904 return -ENOPROTOOPT; 905 906 /* @@ the meaning of setsockopt() when the socket is connected and 907 * there are multiple subflows is not yet defined. It is up to the 908 * MPTCP-level socket to configure the subflows until the subflow 909 * is in TCP fallback, when TCP socket options are passed through 910 * to the one remaining subflow. 911 */ 912 lock_sock(sk); 913 ssk = __mptcp_tcp_fallback(msk); 914 release_sock(sk); 915 if (ssk) 916 return tcp_setsockopt(ssk, level, optname, optval, optlen); 917 918 if (level == SOL_IP) 919 return mptcp_setsockopt_v4(msk, optname, optval, optlen); 920 921 if (level == SOL_IPV6) 922 return mptcp_setsockopt_v6(msk, optname, optval, optlen); 923 924 if (level == SOL_TCP) 925 return mptcp_setsockopt_sol_tcp(msk, optname, optval, optlen); 926 927 return -EOPNOTSUPP; 928 } 929 930 static int mptcp_getsockopt_first_sf_only(struct mptcp_sock *msk, int level, int optname, 931 char __user *optval, int __user *optlen) 932 { 933 struct sock *sk = (struct sock *)msk; 934 struct sock *ssk; 935 int ret; 936 937 lock_sock(sk); 938 ssk = msk->first; 939 if (ssk) 940 goto get; 941 942 ssk = __mptcp_nmpc_sk(msk); 943 if (IS_ERR(ssk)) { 944 ret = PTR_ERR(ssk); 945 goto out; 946 } 947 948 get: 949 ret = tcp_getsockopt(ssk, level, optname, optval, optlen); 950 951 out: 952 release_sock(sk); 953 return ret; 954 } 955 956 void mptcp_diag_fill_info(struct mptcp_sock *msk, struct mptcp_info *info) 957 { 958 struct sock *sk = (struct sock *)msk; 959 u32 flags = 0; 960 bool slow; 961 u32 now; 962 963 memset(info, 0, sizeof(*info)); 964 965 info->mptcpi_extra_subflows = READ_ONCE(msk->pm.extra_subflows); 966 info->mptcpi_add_addr_signal = READ_ONCE(msk->pm.add_addr_signaled); 967 info->mptcpi_add_addr_accepted = READ_ONCE(msk->pm.add_addr_accepted); 968 info->mptcpi_local_addr_used = READ_ONCE(msk->pm.local_addr_used); 969 970 if (inet_sk_state_load(sk) == TCP_LISTEN) 971 return; 972 973 /* The following limits only make sense for the in-kernel PM */ 974 if (mptcp_pm_is_kernel(msk)) { 975 info->mptcpi_limit_extra_subflows = 976 mptcp_pm_get_limit_extra_subflows(msk); 977 info->mptcpi_endp_signal_max = 978 mptcp_pm_get_endp_signal_max(msk); 979 info->mptcpi_limit_add_addr_accepted = 980 mptcp_pm_get_limit_add_addr_accepted(msk); 981 info->mptcpi_endp_subflow_max = 982 mptcp_pm_get_endp_subflow_max(msk); 983 info->mptcpi_endp_laminar_max = 984 mptcp_pm_get_endp_laminar_max(msk); 985 } 986 987 if (__mptcp_check_fallback(msk)) 988 flags |= MPTCP_INFO_FLAG_FALLBACK; 989 if (READ_ONCE(msk->can_ack)) 990 flags |= MPTCP_INFO_FLAG_REMOTE_KEY_RECEIVED; 991 info->mptcpi_flags = flags; 992 993 slow = lock_sock_fast(sk); 994 info->mptcpi_csum_enabled = READ_ONCE(msk->csum_enabled); 995 info->mptcpi_token = msk->token; 996 info->mptcpi_write_seq = msk->write_seq; 997 info->mptcpi_retransmits = inet_csk(sk)->icsk_retransmits; 998 info->mptcpi_bytes_sent = msk->bytes_sent; 999 info->mptcpi_bytes_received = msk->bytes_received; 1000 info->mptcpi_bytes_retrans = msk->bytes_retrans; 1001 info->mptcpi_subflows_total = info->mptcpi_extra_subflows + 1002 __mptcp_has_initial_subflow(msk); 1003 now = tcp_jiffies32; 1004 info->mptcpi_last_data_sent = jiffies_to_msecs(now - msk->last_data_sent); 1005 info->mptcpi_last_data_recv = jiffies_to_msecs(now - msk->last_data_recv); 1006 unlock_sock_fast(sk, slow); 1007 1008 mptcp_data_lock(sk); 1009 info->mptcpi_last_ack_recv = jiffies_to_msecs(now - msk->last_ack_recv); 1010 info->mptcpi_snd_una = msk->snd_una; 1011 info->mptcpi_rcv_nxt = msk->ack_seq; 1012 info->mptcpi_bytes_acked = msk->bytes_acked; 1013 mptcp_data_unlock(sk); 1014 } 1015 EXPORT_SYMBOL_GPL(mptcp_diag_fill_info); 1016 1017 static int mptcp_getsockopt_info(struct mptcp_sock *msk, char __user *optval, int __user *optlen) 1018 { 1019 struct mptcp_info m_info; 1020 int len; 1021 1022 if (get_user(len, optlen)) 1023 return -EFAULT; 1024 1025 /* When used only to check if a fallback to TCP happened. */ 1026 if (len == 0) 1027 return 0; 1028 1029 len = min_t(unsigned int, len, sizeof(struct mptcp_info)); 1030 1031 mptcp_diag_fill_info(msk, &m_info); 1032 1033 if (put_user(len, optlen)) 1034 return -EFAULT; 1035 1036 if (copy_to_user(optval, &m_info, len)) 1037 return -EFAULT; 1038 1039 return 0; 1040 } 1041 1042 static int mptcp_put_subflow_data(struct mptcp_subflow_data *sfd, 1043 char __user *optval, 1044 u32 copied, 1045 int __user *optlen) 1046 { 1047 u32 copylen = min_t(u32, sfd->size_subflow_data, sizeof(*sfd)); 1048 1049 if (copied) 1050 copied += sfd->size_subflow_data; 1051 else 1052 copied = copylen; 1053 1054 if (put_user(copied, optlen)) 1055 return -EFAULT; 1056 1057 if (copy_to_user(optval, sfd, copylen)) 1058 return -EFAULT; 1059 1060 return 0; 1061 } 1062 1063 static int mptcp_get_subflow_data(struct mptcp_subflow_data *sfd, 1064 char __user *optval, 1065 int __user *optlen) 1066 { 1067 int len, copylen; 1068 1069 if (get_user(len, optlen)) 1070 return -EFAULT; 1071 1072 /* if mptcp_subflow_data size is changed, need to adjust 1073 * this function to deal with programs using old version. 1074 */ 1075 BUILD_BUG_ON(sizeof(*sfd) != MIN_INFO_OPTLEN_SIZE); 1076 1077 if (len < MIN_INFO_OPTLEN_SIZE) 1078 return -EINVAL; 1079 1080 memset(sfd, 0, sizeof(*sfd)); 1081 1082 copylen = min_t(unsigned int, len, sizeof(*sfd)); 1083 if (copy_from_user(sfd, optval, copylen)) 1084 return -EFAULT; 1085 1086 /* size_subflow_data is u32, but len is signed */ 1087 if (sfd->size_subflow_data > INT_MAX || 1088 sfd->size_user > INT_MAX) 1089 return -EINVAL; 1090 1091 if (sfd->size_subflow_data < MIN_INFO_OPTLEN_SIZE || 1092 sfd->size_subflow_data > len) 1093 return -EINVAL; 1094 1095 if (sfd->num_subflows || sfd->size_kernel) 1096 return -EINVAL; 1097 1098 return len - sfd->size_subflow_data; 1099 } 1100 1101 static int mptcp_getsockopt_tcpinfo(struct mptcp_sock *msk, char __user *optval, 1102 int __user *optlen) 1103 { 1104 struct mptcp_subflow_context *subflow; 1105 struct sock *sk = (struct sock *)msk; 1106 unsigned int sfcount = 0, copied = 0; 1107 struct mptcp_subflow_data sfd; 1108 char __user *infoptr; 1109 int len; 1110 1111 len = mptcp_get_subflow_data(&sfd, optval, optlen); 1112 if (len < 0) 1113 return len; 1114 1115 sfd.size_kernel = sizeof(struct tcp_info); 1116 sfd.size_user = min_t(unsigned int, sfd.size_user, 1117 sizeof(struct tcp_info)); 1118 1119 infoptr = optval + sfd.size_subflow_data; 1120 1121 lock_sock(sk); 1122 1123 mptcp_for_each_subflow(msk, subflow) { 1124 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 1125 1126 ++sfcount; 1127 1128 if (len && len >= sfd.size_user) { 1129 struct tcp_info info; 1130 1131 tcp_get_info(ssk, &info); 1132 1133 if (copy_to_user(infoptr, &info, sfd.size_user)) { 1134 release_sock(sk); 1135 return -EFAULT; 1136 } 1137 1138 infoptr += sfd.size_user; 1139 copied += sfd.size_user; 1140 len -= sfd.size_user; 1141 } 1142 } 1143 1144 release_sock(sk); 1145 1146 sfd.num_subflows = sfcount; 1147 1148 if (mptcp_put_subflow_data(&sfd, optval, copied, optlen)) 1149 return -EFAULT; 1150 1151 return 0; 1152 } 1153 1154 static void mptcp_get_sub_addrs(const struct sock *sk, struct mptcp_subflow_addrs *a) 1155 { 1156 const struct inet_sock *inet = inet_sk(sk); 1157 1158 memset(a, 0, sizeof(*a)); 1159 1160 if (sk->sk_family == AF_INET) { 1161 a->sin_local.sin_family = AF_INET; 1162 a->sin_local.sin_port = inet->inet_sport; 1163 a->sin_local.sin_addr.s_addr = inet->inet_rcv_saddr; 1164 1165 if (!a->sin_local.sin_addr.s_addr) 1166 a->sin_local.sin_addr.s_addr = inet->inet_saddr; 1167 1168 a->sin_remote.sin_family = AF_INET; 1169 a->sin_remote.sin_port = inet->inet_dport; 1170 a->sin_remote.sin_addr.s_addr = inet->inet_daddr; 1171 #if IS_ENABLED(CONFIG_IPV6) 1172 } else if (sk->sk_family == AF_INET6) { 1173 const struct ipv6_pinfo *np = inet6_sk(sk); 1174 1175 if (WARN_ON_ONCE(!np)) 1176 return; 1177 1178 a->sin6_local.sin6_family = AF_INET6; 1179 a->sin6_local.sin6_port = inet->inet_sport; 1180 1181 if (ipv6_addr_any(&sk->sk_v6_rcv_saddr)) 1182 a->sin6_local.sin6_addr = np->saddr; 1183 else 1184 a->sin6_local.sin6_addr = sk->sk_v6_rcv_saddr; 1185 1186 a->sin6_remote.sin6_family = AF_INET6; 1187 a->sin6_remote.sin6_port = inet->inet_dport; 1188 a->sin6_remote.sin6_addr = sk->sk_v6_daddr; 1189 #endif 1190 } 1191 } 1192 1193 static int mptcp_getsockopt_subflow_addrs(struct mptcp_sock *msk, char __user *optval, 1194 int __user *optlen) 1195 { 1196 struct mptcp_subflow_context *subflow; 1197 struct sock *sk = (struct sock *)msk; 1198 unsigned int sfcount = 0, copied = 0; 1199 struct mptcp_subflow_data sfd; 1200 char __user *addrptr; 1201 int len; 1202 1203 len = mptcp_get_subflow_data(&sfd, optval, optlen); 1204 if (len < 0) 1205 return len; 1206 1207 sfd.size_kernel = sizeof(struct mptcp_subflow_addrs); 1208 sfd.size_user = min_t(unsigned int, sfd.size_user, 1209 sizeof(struct mptcp_subflow_addrs)); 1210 1211 addrptr = optval + sfd.size_subflow_data; 1212 1213 lock_sock(sk); 1214 1215 mptcp_for_each_subflow(msk, subflow) { 1216 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 1217 1218 ++sfcount; 1219 1220 if (len && len >= sfd.size_user) { 1221 struct mptcp_subflow_addrs a; 1222 1223 mptcp_get_sub_addrs(ssk, &a); 1224 1225 if (copy_to_user(addrptr, &a, sfd.size_user)) { 1226 release_sock(sk); 1227 return -EFAULT; 1228 } 1229 1230 addrptr += sfd.size_user; 1231 copied += sfd.size_user; 1232 len -= sfd.size_user; 1233 } 1234 } 1235 1236 release_sock(sk); 1237 1238 sfd.num_subflows = sfcount; 1239 1240 if (mptcp_put_subflow_data(&sfd, optval, copied, optlen)) 1241 return -EFAULT; 1242 1243 return 0; 1244 } 1245 1246 static int mptcp_get_full_info(struct mptcp_full_info *mfi, 1247 char __user *optval, 1248 int __user *optlen) 1249 { 1250 int len; 1251 1252 BUILD_BUG_ON(offsetof(struct mptcp_full_info, mptcp_info) != 1253 MIN_FULL_INFO_OPTLEN_SIZE); 1254 1255 if (get_user(len, optlen)) 1256 return -EFAULT; 1257 1258 if (len < MIN_FULL_INFO_OPTLEN_SIZE) 1259 return -EINVAL; 1260 1261 memset(mfi, 0, sizeof(*mfi)); 1262 if (copy_from_user(mfi, optval, MIN_FULL_INFO_OPTLEN_SIZE)) 1263 return -EFAULT; 1264 1265 if (mfi->size_tcpinfo_kernel || 1266 mfi->size_sfinfo_kernel || 1267 mfi->num_subflows) 1268 return -EINVAL; 1269 1270 if (mfi->size_sfinfo_user > INT_MAX || 1271 mfi->size_tcpinfo_user > INT_MAX) 1272 return -EINVAL; 1273 1274 return len - MIN_FULL_INFO_OPTLEN_SIZE; 1275 } 1276 1277 static int mptcp_put_full_info(struct mptcp_full_info *mfi, 1278 char __user *optval, 1279 u32 copylen, 1280 int __user *optlen) 1281 { 1282 copylen += MIN_FULL_INFO_OPTLEN_SIZE; 1283 if (put_user(copylen, optlen)) 1284 return -EFAULT; 1285 1286 if (copy_to_user(optval, mfi, copylen)) 1287 return -EFAULT; 1288 return 0; 1289 } 1290 1291 static int mptcp_getsockopt_full_info(struct mptcp_sock *msk, char __user *optval, 1292 int __user *optlen) 1293 { 1294 unsigned int sfcount = 0, copylen = 0; 1295 struct mptcp_subflow_context *subflow; 1296 struct sock *sk = (struct sock *)msk; 1297 void __user *tcpinfoptr, *sfinfoptr; 1298 struct mptcp_full_info mfi; 1299 int len; 1300 1301 len = mptcp_get_full_info(&mfi, optval, optlen); 1302 if (len < 0) 1303 return len; 1304 1305 /* don't bother filling the mptcp info if there is not enough 1306 * user-space-provided storage 1307 */ 1308 if (len > 0) { 1309 mptcp_diag_fill_info(msk, &mfi.mptcp_info); 1310 copylen += min_t(unsigned int, len, sizeof(struct mptcp_info)); 1311 } 1312 1313 mfi.size_tcpinfo_kernel = sizeof(struct tcp_info); 1314 mfi.size_tcpinfo_user = min_t(unsigned int, mfi.size_tcpinfo_user, 1315 sizeof(struct tcp_info)); 1316 sfinfoptr = u64_to_user_ptr(mfi.subflow_info); 1317 mfi.size_sfinfo_kernel = sizeof(struct mptcp_subflow_info); 1318 mfi.size_sfinfo_user = min_t(unsigned int, mfi.size_sfinfo_user, 1319 sizeof(struct mptcp_subflow_info)); 1320 tcpinfoptr = u64_to_user_ptr(mfi.tcp_info); 1321 1322 lock_sock(sk); 1323 mptcp_for_each_subflow(msk, subflow) { 1324 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 1325 struct mptcp_subflow_info sfinfo; 1326 struct tcp_info tcp_info; 1327 1328 if (sfcount++ >= mfi.size_arrays_user) 1329 continue; 1330 1331 /* fetch addr/tcp_info only if the user space buffers 1332 * are wide enough 1333 */ 1334 memset(&sfinfo, 0, sizeof(sfinfo)); 1335 sfinfo.id = subflow->subflow_id; 1336 if (mfi.size_sfinfo_user > 1337 offsetof(struct mptcp_subflow_info, addrs)) 1338 mptcp_get_sub_addrs(ssk, &sfinfo.addrs); 1339 if (copy_to_user(sfinfoptr, &sfinfo, mfi.size_sfinfo_user)) 1340 goto fail_release; 1341 1342 if (mfi.size_tcpinfo_user) { 1343 tcp_get_info(ssk, &tcp_info); 1344 if (copy_to_user(tcpinfoptr, &tcp_info, 1345 mfi.size_tcpinfo_user)) 1346 goto fail_release; 1347 } 1348 1349 tcpinfoptr += mfi.size_tcpinfo_user; 1350 sfinfoptr += mfi.size_sfinfo_user; 1351 } 1352 release_sock(sk); 1353 1354 mfi.num_subflows = sfcount; 1355 if (mptcp_put_full_info(&mfi, optval, copylen, optlen)) 1356 return -EFAULT; 1357 1358 return 0; 1359 1360 fail_release: 1361 release_sock(sk); 1362 return -EFAULT; 1363 } 1364 1365 static int mptcp_put_int_option(struct mptcp_sock *msk, char __user *optval, 1366 int __user *optlen, int val) 1367 { 1368 int len; 1369 1370 if (get_user(len, optlen)) 1371 return -EFAULT; 1372 if (len < 0) 1373 return -EINVAL; 1374 1375 if (len < sizeof(int) && len > 0 && val >= 0 && val <= 255) { 1376 unsigned char ucval = (unsigned char)val; 1377 1378 len = 1; 1379 if (put_user(len, optlen)) 1380 return -EFAULT; 1381 if (copy_to_user(optval, &ucval, 1)) 1382 return -EFAULT; 1383 } else { 1384 len = min_t(unsigned int, len, sizeof(int)); 1385 if (put_user(len, optlen)) 1386 return -EFAULT; 1387 if (copy_to_user(optval, &val, len)) 1388 return -EFAULT; 1389 } 1390 1391 return 0; 1392 } 1393 1394 static int mptcp_getsockopt_sol_tcp(struct mptcp_sock *msk, int optname, 1395 char __user *optval, int __user *optlen) 1396 { 1397 struct sock *sk = (void *)msk; 1398 1399 switch (optname) { 1400 case TCP_ULP: 1401 case TCP_CONGESTION: 1402 case TCP_INFO: 1403 case TCP_CC_INFO: 1404 case TCP_DEFER_ACCEPT: 1405 case TCP_FASTOPEN: 1406 case TCP_FASTOPEN_CONNECT: 1407 case TCP_FASTOPEN_KEY: 1408 case TCP_FASTOPEN_NO_COOKIE: 1409 return mptcp_getsockopt_first_sf_only(msk, SOL_TCP, optname, 1410 optval, optlen); 1411 case TCP_INQ: 1412 return mptcp_put_int_option(msk, optval, optlen, msk->recvmsg_inq); 1413 case TCP_CORK: 1414 return mptcp_put_int_option(msk, optval, optlen, msk->cork); 1415 case TCP_NODELAY: 1416 return mptcp_put_int_option(msk, optval, optlen, msk->nodelay); 1417 case TCP_KEEPIDLE: 1418 return mptcp_put_int_option(msk, optval, optlen, 1419 msk->keepalive_idle ? : 1420 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_keepalive_time) / HZ); 1421 case TCP_KEEPINTVL: 1422 return mptcp_put_int_option(msk, optval, optlen, 1423 msk->keepalive_intvl ? : 1424 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_keepalive_intvl) / HZ); 1425 case TCP_KEEPCNT: 1426 return mptcp_put_int_option(msk, optval, optlen, 1427 msk->keepalive_cnt ? : 1428 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_keepalive_probes)); 1429 case TCP_NOTSENT_LOWAT: 1430 return mptcp_put_int_option(msk, optval, optlen, msk->notsent_lowat); 1431 case TCP_IS_MPTCP: 1432 return mptcp_put_int_option(msk, optval, optlen, 1); 1433 case TCP_MAXSEG: 1434 return mptcp_getsockopt_first_sf_only(msk, SOL_TCP, optname, 1435 optval, optlen); 1436 } 1437 return -EOPNOTSUPP; 1438 } 1439 1440 static int mptcp_getsockopt_v4(struct mptcp_sock *msk, int optname, 1441 char __user *optval, int __user *optlen) 1442 { 1443 struct sock *sk = (void *)msk; 1444 1445 switch (optname) { 1446 case IP_TOS: 1447 return mptcp_put_int_option(msk, optval, optlen, READ_ONCE(inet_sk(sk)->tos)); 1448 case IP_FREEBIND: 1449 return mptcp_put_int_option(msk, optval, optlen, 1450 inet_test_bit(FREEBIND, sk)); 1451 case IP_TRANSPARENT: 1452 return mptcp_put_int_option(msk, optval, optlen, 1453 inet_test_bit(TRANSPARENT, sk)); 1454 case IP_BIND_ADDRESS_NO_PORT: 1455 return mptcp_put_int_option(msk, optval, optlen, 1456 inet_test_bit(BIND_ADDRESS_NO_PORT, sk)); 1457 case IP_LOCAL_PORT_RANGE: 1458 return mptcp_put_int_option(msk, optval, optlen, 1459 READ_ONCE(inet_sk(sk)->local_port_range)); 1460 } 1461 1462 return -EOPNOTSUPP; 1463 } 1464 1465 static int mptcp_getsockopt_v6(struct mptcp_sock *msk, int optname, 1466 char __user *optval, int __user *optlen) 1467 { 1468 struct sock *sk = (void *)msk; 1469 1470 switch (optname) { 1471 case IPV6_V6ONLY: 1472 return mptcp_put_int_option(msk, optval, optlen, 1473 sk->sk_ipv6only); 1474 case IPV6_TRANSPARENT: 1475 return mptcp_put_int_option(msk, optval, optlen, 1476 inet_test_bit(TRANSPARENT, sk)); 1477 case IPV6_FREEBIND: 1478 return mptcp_put_int_option(msk, optval, optlen, 1479 inet_test_bit(FREEBIND, sk)); 1480 } 1481 1482 return -EOPNOTSUPP; 1483 } 1484 1485 static int mptcp_getsockopt_sol_mptcp(struct mptcp_sock *msk, int optname, 1486 char __user *optval, int __user *optlen) 1487 { 1488 switch (optname) { 1489 case MPTCP_INFO: 1490 return mptcp_getsockopt_info(msk, optval, optlen); 1491 case MPTCP_FULL_INFO: 1492 return mptcp_getsockopt_full_info(msk, optval, optlen); 1493 case MPTCP_TCPINFO: 1494 return mptcp_getsockopt_tcpinfo(msk, optval, optlen); 1495 case MPTCP_SUBFLOW_ADDRS: 1496 return mptcp_getsockopt_subflow_addrs(msk, optval, optlen); 1497 } 1498 1499 return -EOPNOTSUPP; 1500 } 1501 1502 int mptcp_getsockopt(struct sock *sk, int level, int optname, 1503 char __user *optval, int __user *option) 1504 { 1505 struct mptcp_sock *msk = mptcp_sk(sk); 1506 struct sock *ssk; 1507 1508 pr_debug("msk=%p\n", msk); 1509 1510 /* @@ the meaning of setsockopt() when the socket is connected and 1511 * there are multiple subflows is not yet defined. It is up to the 1512 * MPTCP-level socket to configure the subflows until the subflow 1513 * is in TCP fallback, when socket options are passed through 1514 * to the one remaining subflow. 1515 */ 1516 lock_sock(sk); 1517 ssk = __mptcp_tcp_fallback(msk); 1518 release_sock(sk); 1519 if (ssk) 1520 return tcp_getsockopt(ssk, level, optname, optval, option); 1521 1522 if (level == SOL_IP) 1523 return mptcp_getsockopt_v4(msk, optname, optval, option); 1524 if (level == SOL_IPV6) 1525 return mptcp_getsockopt_v6(msk, optname, optval, option); 1526 if (level == SOL_TCP) 1527 return mptcp_getsockopt_sol_tcp(msk, optname, optval, option); 1528 if (level == SOL_MPTCP) 1529 return mptcp_getsockopt_sol_mptcp(msk, optname, optval, option); 1530 return -EOPNOTSUPP; 1531 } 1532 1533 static void sync_socket_options(struct mptcp_sock *msk, struct sock *ssk) 1534 { 1535 static const unsigned int tx_rx_locks = SOCK_RCVBUF_LOCK | SOCK_SNDBUF_LOCK; 1536 struct sock *sk = (struct sock *)msk; 1537 bool keep_open; 1538 1539 keep_open = sock_flag(sk, SOCK_KEEPOPEN); 1540 if (ssk->sk_prot->keepalive) 1541 ssk->sk_prot->keepalive(ssk, keep_open); 1542 sock_valbool_flag(ssk, SOCK_KEEPOPEN, keep_open); 1543 1544 ssk->sk_priority = sk->sk_priority; 1545 ssk->sk_bound_dev_if = sk->sk_bound_dev_if; 1546 ssk->sk_incoming_cpu = sk->sk_incoming_cpu; 1547 ssk->sk_ipv6only = sk->sk_ipv6only; 1548 __ip_sock_set_tos(ssk, inet_sk(sk)->tos); 1549 1550 if (sk->sk_userlocks & tx_rx_locks) { 1551 ssk->sk_userlocks |= sk->sk_userlocks & tx_rx_locks; 1552 if (sk->sk_userlocks & SOCK_SNDBUF_LOCK) { 1553 WRITE_ONCE(ssk->sk_sndbuf, sk->sk_sndbuf); 1554 mptcp_subflow_ctx(ssk)->cached_sndbuf = sk->sk_sndbuf; 1555 } 1556 if (sk->sk_userlocks & SOCK_RCVBUF_LOCK) 1557 WRITE_ONCE(ssk->sk_rcvbuf, sk->sk_rcvbuf); 1558 } 1559 1560 if (sock_flag(sk, SOCK_LINGER)) { 1561 ssk->sk_lingertime = sk->sk_lingertime; 1562 sock_set_flag(ssk, SOCK_LINGER); 1563 } else { 1564 sock_reset_flag(ssk, SOCK_LINGER); 1565 } 1566 1567 if (sk->sk_mark != ssk->sk_mark) { 1568 ssk->sk_mark = sk->sk_mark; 1569 sk_dst_reset(ssk); 1570 } 1571 1572 sock_valbool_flag(ssk, SOCK_DBG, sock_flag(sk, SOCK_DBG)); 1573 1574 if (inet_csk(sk)->icsk_ca_ops != inet_csk(ssk)->icsk_ca_ops) 1575 tcp_set_congestion_control(ssk, msk->ca_name, false, true); 1576 __tcp_sock_set_cork(ssk, !!msk->cork); 1577 __tcp_sock_set_nodelay(ssk, !!msk->nodelay); 1578 tcp_sock_set_keepidle_locked(ssk, msk->keepalive_idle); 1579 tcp_sock_set_keepintvl(ssk, msk->keepalive_intvl); 1580 tcp_sock_set_keepcnt(ssk, msk->keepalive_cnt); 1581 tcp_sock_set_maxseg(ssk, msk->maxseg); 1582 1583 inet_assign_bit(TRANSPARENT, ssk, inet_test_bit(TRANSPARENT, sk)); 1584 inet_assign_bit(FREEBIND, ssk, inet_test_bit(FREEBIND, sk)); 1585 inet_assign_bit(BIND_ADDRESS_NO_PORT, ssk, inet_test_bit(BIND_ADDRESS_NO_PORT, sk)); 1586 WRITE_ONCE(inet_sk(ssk)->local_port_range, READ_ONCE(inet_sk(sk)->local_port_range)); 1587 } 1588 1589 void mptcp_sockopt_sync_locked(struct mptcp_sock *msk, struct sock *ssk) 1590 { 1591 struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk); 1592 1593 msk_owned_by_me(msk); 1594 1595 ssk->sk_rcvlowat = 0; 1596 1597 /* subflows must ignore any latency-related settings: will not affect 1598 * the user-space - only the msk is relevant - but will foul the 1599 * mptcp scheduler 1600 */ 1601 tcp_sk(ssk)->notsent_lowat = UINT_MAX; 1602 1603 if (READ_ONCE(subflow->setsockopt_seq) != msk->setsockopt_seq) { 1604 sync_socket_options(msk, ssk); 1605 1606 subflow->setsockopt_seq = msk->setsockopt_seq; 1607 } 1608 } 1609 1610 /* unfortunately this is different enough from the tcp version so 1611 * that we can't factor it out 1612 */ 1613 int mptcp_set_rcvlowat(struct sock *sk, int val) 1614 { 1615 struct mptcp_subflow_context *subflow; 1616 int space, cap; 1617 1618 /* bpf can land here with a wrong sk type */ 1619 if (sk->sk_protocol == IPPROTO_TCP) 1620 return -EINVAL; 1621 1622 if (sk->sk_userlocks & SOCK_RCVBUF_LOCK) 1623 cap = sk->sk_rcvbuf >> 1; 1624 else 1625 cap = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2]) >> 1; 1626 val = min(val, cap); 1627 WRITE_ONCE(sk->sk_rcvlowat, val ? : 1); 1628 1629 /* Check if we need to signal EPOLLIN right now */ 1630 if (mptcp_epollin_ready(sk)) 1631 sk->sk_data_ready(sk); 1632 1633 if (sk->sk_userlocks & SOCK_RCVBUF_LOCK) 1634 return 0; 1635 1636 space = mptcp_space_from_win(sk, val); 1637 if (space <= sk->sk_rcvbuf) 1638 return 0; 1639 1640 /* propagate the rcvbuf changes to all the subflows */ 1641 WRITE_ONCE(sk->sk_rcvbuf, space); 1642 mptcp_for_each_subflow(mptcp_sk(sk), subflow) { 1643 struct sock *ssk = mptcp_subflow_tcp_sock(subflow); 1644 bool slow; 1645 1646 slow = lock_sock_fast(ssk); 1647 WRITE_ONCE(ssk->sk_rcvbuf, space); 1648 WRITE_ONCE(tcp_sk(ssk)->window_clamp, val); 1649 unlock_sock_fast(ssk, slow); 1650 } 1651 return 0; 1652 } 1653