xref: /linux/net/ipv4/ip_sockglue.c (revision 1553a1c48281243359a9529a10ddb551f3b967ab)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * INET		An implementation of the TCP/IP protocol suite for the LINUX
4  *		operating system.  INET is implemented using the  BSD Socket
5  *		interface as the means of communication with the user level.
6  *
7  *		The IP to API glue.
8  *
9  * Authors:	see ip.c
10  *
11  * Fixes:
12  *		Many		:	Split from ip.c , see ip.c for history.
13  *		Martin Mares	:	TOS setting fixed.
14  *		Alan Cox	:	Fixed a couple of oopses in Martin's
15  *					TOS tweaks.
16  *		Mike McLagan	:	Routing by source
17  */
18 
19 #include <linux/module.h>
20 #include <linux/types.h>
21 #include <linux/mm.h>
22 #include <linux/skbuff.h>
23 #include <linux/ip.h>
24 #include <linux/icmp.h>
25 #include <linux/inetdevice.h>
26 #include <linux/netdevice.h>
27 #include <linux/slab.h>
28 #include <net/sock.h>
29 #include <net/ip.h>
30 #include <net/icmp.h>
31 #include <net/tcp_states.h>
32 #include <linux/udp.h>
33 #include <linux/igmp.h>
34 #include <linux/netfilter.h>
35 #include <linux/route.h>
36 #include <linux/mroute.h>
37 #include <net/inet_ecn.h>
38 #include <net/route.h>
39 #include <net/xfrm.h>
40 #include <net/compat.h>
41 #include <net/checksum.h>
42 #if IS_ENABLED(CONFIG_IPV6)
43 #include <net/transp_v6.h>
44 #endif
45 #include <net/ip_fib.h>
46 
47 #include <linux/errqueue.h>
48 #include <linux/uaccess.h>
49 
50 /*
51  *	SOL_IP control messages.
52  */
53 
54 static void ip_cmsg_recv_pktinfo(struct msghdr *msg, struct sk_buff *skb)
55 {
56 	struct in_pktinfo info = *PKTINFO_SKB_CB(skb);
57 
58 	info.ipi_addr.s_addr = ip_hdr(skb)->daddr;
59 
60 	put_cmsg(msg, SOL_IP, IP_PKTINFO, sizeof(info), &info);
61 }
62 
63 static void ip_cmsg_recv_ttl(struct msghdr *msg, struct sk_buff *skb)
64 {
65 	int ttl = ip_hdr(skb)->ttl;
66 	put_cmsg(msg, SOL_IP, IP_TTL, sizeof(int), &ttl);
67 }
68 
69 static void ip_cmsg_recv_tos(struct msghdr *msg, struct sk_buff *skb)
70 {
71 	put_cmsg(msg, SOL_IP, IP_TOS, 1, &ip_hdr(skb)->tos);
72 }
73 
74 static void ip_cmsg_recv_opts(struct msghdr *msg, struct sk_buff *skb)
75 {
76 	if (IPCB(skb)->opt.optlen == 0)
77 		return;
78 
79 	put_cmsg(msg, SOL_IP, IP_RECVOPTS, IPCB(skb)->opt.optlen,
80 		 ip_hdr(skb) + 1);
81 }
82 
83 
84 static void ip_cmsg_recv_retopts(struct net *net, struct msghdr *msg,
85 				 struct sk_buff *skb)
86 {
87 	unsigned char optbuf[sizeof(struct ip_options) + 40];
88 	struct ip_options *opt = (struct ip_options *)optbuf;
89 
90 	if (IPCB(skb)->opt.optlen == 0)
91 		return;
92 
93 	if (ip_options_echo(net, opt, skb)) {
94 		msg->msg_flags |= MSG_CTRUNC;
95 		return;
96 	}
97 	ip_options_undo(opt);
98 
99 	put_cmsg(msg, SOL_IP, IP_RETOPTS, opt->optlen, opt->__data);
100 }
101 
102 static void ip_cmsg_recv_fragsize(struct msghdr *msg, struct sk_buff *skb)
103 {
104 	int val;
105 
106 	if (IPCB(skb)->frag_max_size == 0)
107 		return;
108 
109 	val = IPCB(skb)->frag_max_size;
110 	put_cmsg(msg, SOL_IP, IP_RECVFRAGSIZE, sizeof(val), &val);
111 }
112 
113 static void ip_cmsg_recv_checksum(struct msghdr *msg, struct sk_buff *skb,
114 				  int tlen, int offset)
115 {
116 	__wsum csum = skb->csum;
117 
118 	if (skb->ip_summed != CHECKSUM_COMPLETE)
119 		return;
120 
121 	if (offset != 0) {
122 		int tend_off = skb_transport_offset(skb) + tlen;
123 		csum = csum_sub(csum, skb_checksum(skb, tend_off, offset, 0));
124 	}
125 
126 	put_cmsg(msg, SOL_IP, IP_CHECKSUM, sizeof(__wsum), &csum);
127 }
128 
129 static void ip_cmsg_recv_security(struct msghdr *msg, struct sk_buff *skb)
130 {
131 	char *secdata;
132 	u32 seclen, secid;
133 	int err;
134 
135 	err = security_socket_getpeersec_dgram(NULL, skb, &secid);
136 	if (err)
137 		return;
138 
139 	err = security_secid_to_secctx(secid, &secdata, &seclen);
140 	if (err)
141 		return;
142 
143 	put_cmsg(msg, SOL_IP, SCM_SECURITY, seclen, secdata);
144 	security_release_secctx(secdata, seclen);
145 }
146 
147 static void ip_cmsg_recv_dstaddr(struct msghdr *msg, struct sk_buff *skb)
148 {
149 	__be16 _ports[2], *ports;
150 	struct sockaddr_in sin;
151 
152 	/* All current transport protocols have the port numbers in the
153 	 * first four bytes of the transport header and this function is
154 	 * written with this assumption in mind.
155 	 */
156 	ports = skb_header_pointer(skb, skb_transport_offset(skb),
157 				   sizeof(_ports), &_ports);
158 	if (!ports)
159 		return;
160 
161 	sin.sin_family = AF_INET;
162 	sin.sin_addr.s_addr = ip_hdr(skb)->daddr;
163 	sin.sin_port = ports[1];
164 	memset(sin.sin_zero, 0, sizeof(sin.sin_zero));
165 
166 	put_cmsg(msg, SOL_IP, IP_ORIGDSTADDR, sizeof(sin), &sin);
167 }
168 
169 void ip_cmsg_recv_offset(struct msghdr *msg, struct sock *sk,
170 			 struct sk_buff *skb, int tlen, int offset)
171 {
172 	unsigned long flags = inet_cmsg_flags(inet_sk(sk));
173 
174 	if (!flags)
175 		return;
176 
177 	/* Ordered by supposed usage frequency */
178 	if (flags & IP_CMSG_PKTINFO) {
179 		ip_cmsg_recv_pktinfo(msg, skb);
180 
181 		flags &= ~IP_CMSG_PKTINFO;
182 		if (!flags)
183 			return;
184 	}
185 
186 	if (flags & IP_CMSG_TTL) {
187 		ip_cmsg_recv_ttl(msg, skb);
188 
189 		flags &= ~IP_CMSG_TTL;
190 		if (!flags)
191 			return;
192 	}
193 
194 	if (flags & IP_CMSG_TOS) {
195 		ip_cmsg_recv_tos(msg, skb);
196 
197 		flags &= ~IP_CMSG_TOS;
198 		if (!flags)
199 			return;
200 	}
201 
202 	if (flags & IP_CMSG_RECVOPTS) {
203 		ip_cmsg_recv_opts(msg, skb);
204 
205 		flags &= ~IP_CMSG_RECVOPTS;
206 		if (!flags)
207 			return;
208 	}
209 
210 	if (flags & IP_CMSG_RETOPTS) {
211 		ip_cmsg_recv_retopts(sock_net(sk), msg, skb);
212 
213 		flags &= ~IP_CMSG_RETOPTS;
214 		if (!flags)
215 			return;
216 	}
217 
218 	if (flags & IP_CMSG_PASSSEC) {
219 		ip_cmsg_recv_security(msg, skb);
220 
221 		flags &= ~IP_CMSG_PASSSEC;
222 		if (!flags)
223 			return;
224 	}
225 
226 	if (flags & IP_CMSG_ORIGDSTADDR) {
227 		ip_cmsg_recv_dstaddr(msg, skb);
228 
229 		flags &= ~IP_CMSG_ORIGDSTADDR;
230 		if (!flags)
231 			return;
232 	}
233 
234 	if (flags & IP_CMSG_CHECKSUM)
235 		ip_cmsg_recv_checksum(msg, skb, tlen, offset);
236 
237 	if (flags & IP_CMSG_RECVFRAGSIZE)
238 		ip_cmsg_recv_fragsize(msg, skb);
239 }
240 EXPORT_SYMBOL(ip_cmsg_recv_offset);
241 
242 int ip_cmsg_send(struct sock *sk, struct msghdr *msg, struct ipcm_cookie *ipc,
243 		 bool allow_ipv6)
244 {
245 	int err, val;
246 	struct cmsghdr *cmsg;
247 	struct net *net = sock_net(sk);
248 
249 	for_each_cmsghdr(cmsg, msg) {
250 		if (!CMSG_OK(msg, cmsg))
251 			return -EINVAL;
252 #if IS_ENABLED(CONFIG_IPV6)
253 		if (allow_ipv6 &&
254 		    cmsg->cmsg_level == SOL_IPV6 &&
255 		    cmsg->cmsg_type == IPV6_PKTINFO) {
256 			struct in6_pktinfo *src_info;
257 
258 			if (cmsg->cmsg_len < CMSG_LEN(sizeof(*src_info)))
259 				return -EINVAL;
260 			src_info = (struct in6_pktinfo *)CMSG_DATA(cmsg);
261 			if (!ipv6_addr_v4mapped(&src_info->ipi6_addr))
262 				return -EINVAL;
263 			if (src_info->ipi6_ifindex)
264 				ipc->oif = src_info->ipi6_ifindex;
265 			ipc->addr = src_info->ipi6_addr.s6_addr32[3];
266 			continue;
267 		}
268 #endif
269 		if (cmsg->cmsg_level == SOL_SOCKET) {
270 			err = __sock_cmsg_send(sk, cmsg, &ipc->sockc);
271 			if (err)
272 				return err;
273 			continue;
274 		}
275 
276 		if (cmsg->cmsg_level != SOL_IP)
277 			continue;
278 		switch (cmsg->cmsg_type) {
279 		case IP_RETOPTS:
280 			err = cmsg->cmsg_len - sizeof(struct cmsghdr);
281 
282 			/* Our caller is responsible for freeing ipc->opt */
283 			err = ip_options_get(net, &ipc->opt,
284 					     KERNEL_SOCKPTR(CMSG_DATA(cmsg)),
285 					     err < 40 ? err : 40);
286 			if (err)
287 				return err;
288 			break;
289 		case IP_PKTINFO:
290 		{
291 			struct in_pktinfo *info;
292 			if (cmsg->cmsg_len != CMSG_LEN(sizeof(struct in_pktinfo)))
293 				return -EINVAL;
294 			info = (struct in_pktinfo *)CMSG_DATA(cmsg);
295 			if (info->ipi_ifindex)
296 				ipc->oif = info->ipi_ifindex;
297 			ipc->addr = info->ipi_spec_dst.s_addr;
298 			break;
299 		}
300 		case IP_TTL:
301 			if (cmsg->cmsg_len != CMSG_LEN(sizeof(int)))
302 				return -EINVAL;
303 			val = *(int *)CMSG_DATA(cmsg);
304 			if (val < 1 || val > 255)
305 				return -EINVAL;
306 			ipc->ttl = val;
307 			break;
308 		case IP_TOS:
309 			if (cmsg->cmsg_len == CMSG_LEN(sizeof(int)))
310 				val = *(int *)CMSG_DATA(cmsg);
311 			else if (cmsg->cmsg_len == CMSG_LEN(sizeof(u8)))
312 				val = *(u8 *)CMSG_DATA(cmsg);
313 			else
314 				return -EINVAL;
315 			if (val < 0 || val > 255)
316 				return -EINVAL;
317 			ipc->tos = val;
318 			ipc->priority = rt_tos2priority(ipc->tos);
319 			break;
320 		case IP_PROTOCOL:
321 			if (cmsg->cmsg_len != CMSG_LEN(sizeof(int)))
322 				return -EINVAL;
323 			val = *(int *)CMSG_DATA(cmsg);
324 			if (val < 1 || val > 255)
325 				return -EINVAL;
326 			ipc->protocol = val;
327 			break;
328 		default:
329 			return -EINVAL;
330 		}
331 	}
332 	return 0;
333 }
334 
335 static void ip_ra_destroy_rcu(struct rcu_head *head)
336 {
337 	struct ip_ra_chain *ra = container_of(head, struct ip_ra_chain, rcu);
338 
339 	sock_put(ra->saved_sk);
340 	kfree(ra);
341 }
342 
343 int ip_ra_control(struct sock *sk, unsigned char on,
344 		  void (*destructor)(struct sock *))
345 {
346 	struct ip_ra_chain *ra, *new_ra;
347 	struct ip_ra_chain __rcu **rap;
348 	struct net *net = sock_net(sk);
349 
350 	if (sk->sk_type != SOCK_RAW || inet_sk(sk)->inet_num == IPPROTO_RAW)
351 		return -EINVAL;
352 
353 	new_ra = on ? kmalloc(sizeof(*new_ra), GFP_KERNEL) : NULL;
354 	if (on && !new_ra)
355 		return -ENOMEM;
356 
357 	mutex_lock(&net->ipv4.ra_mutex);
358 	for (rap = &net->ipv4.ra_chain;
359 	     (ra = rcu_dereference_protected(*rap,
360 			lockdep_is_held(&net->ipv4.ra_mutex))) != NULL;
361 	     rap = &ra->next) {
362 		if (ra->sk == sk) {
363 			if (on) {
364 				mutex_unlock(&net->ipv4.ra_mutex);
365 				kfree(new_ra);
366 				return -EADDRINUSE;
367 			}
368 			/* dont let ip_call_ra_chain() use sk again */
369 			ra->sk = NULL;
370 			RCU_INIT_POINTER(*rap, ra->next);
371 			mutex_unlock(&net->ipv4.ra_mutex);
372 
373 			if (ra->destructor)
374 				ra->destructor(sk);
375 			/*
376 			 * Delay sock_put(sk) and kfree(ra) after one rcu grace
377 			 * period. This guarantee ip_call_ra_chain() dont need
378 			 * to mess with socket refcounts.
379 			 */
380 			ra->saved_sk = sk;
381 			call_rcu(&ra->rcu, ip_ra_destroy_rcu);
382 			return 0;
383 		}
384 	}
385 	if (!new_ra) {
386 		mutex_unlock(&net->ipv4.ra_mutex);
387 		return -ENOBUFS;
388 	}
389 	new_ra->sk = sk;
390 	new_ra->destructor = destructor;
391 
392 	RCU_INIT_POINTER(new_ra->next, ra);
393 	rcu_assign_pointer(*rap, new_ra);
394 	sock_hold(sk);
395 	mutex_unlock(&net->ipv4.ra_mutex);
396 
397 	return 0;
398 }
399 
400 static void ipv4_icmp_error_rfc4884(const struct sk_buff *skb,
401 				    struct sock_ee_data_rfc4884 *out)
402 {
403 	switch (icmp_hdr(skb)->type) {
404 	case ICMP_DEST_UNREACH:
405 	case ICMP_TIME_EXCEEDED:
406 	case ICMP_PARAMETERPROB:
407 		ip_icmp_error_rfc4884(skb, out, sizeof(struct icmphdr),
408 				      icmp_hdr(skb)->un.reserved[1] * 4);
409 	}
410 }
411 
412 void ip_icmp_error(struct sock *sk, struct sk_buff *skb, int err,
413 		   __be16 port, u32 info, u8 *payload)
414 {
415 	struct sock_exterr_skb *serr;
416 
417 	skb = skb_clone(skb, GFP_ATOMIC);
418 	if (!skb)
419 		return;
420 
421 	serr = SKB_EXT_ERR(skb);
422 	serr->ee.ee_errno = err;
423 	serr->ee.ee_origin = SO_EE_ORIGIN_ICMP;
424 	serr->ee.ee_type = icmp_hdr(skb)->type;
425 	serr->ee.ee_code = icmp_hdr(skb)->code;
426 	serr->ee.ee_pad = 0;
427 	serr->ee.ee_info = info;
428 	serr->ee.ee_data = 0;
429 	serr->addr_offset = (u8 *)&(((struct iphdr *)(icmp_hdr(skb) + 1))->daddr) -
430 				   skb_network_header(skb);
431 	serr->port = port;
432 
433 	if (skb_pull(skb, payload - skb->data)) {
434 		if (inet_test_bit(RECVERR_RFC4884, sk))
435 			ipv4_icmp_error_rfc4884(skb, &serr->ee.ee_rfc4884);
436 
437 		skb_reset_transport_header(skb);
438 		if (sock_queue_err_skb(sk, skb) == 0)
439 			return;
440 	}
441 	kfree_skb(skb);
442 }
443 EXPORT_SYMBOL_GPL(ip_icmp_error);
444 
445 void ip_local_error(struct sock *sk, int err, __be32 daddr, __be16 port, u32 info)
446 {
447 	struct sock_exterr_skb *serr;
448 	struct iphdr *iph;
449 	struct sk_buff *skb;
450 
451 	if (!inet_test_bit(RECVERR, sk))
452 		return;
453 
454 	skb = alloc_skb(sizeof(struct iphdr), GFP_ATOMIC);
455 	if (!skb)
456 		return;
457 
458 	skb_put(skb, sizeof(struct iphdr));
459 	skb_reset_network_header(skb);
460 	iph = ip_hdr(skb);
461 	iph->daddr = daddr;
462 
463 	serr = SKB_EXT_ERR(skb);
464 	serr->ee.ee_errno = err;
465 	serr->ee.ee_origin = SO_EE_ORIGIN_LOCAL;
466 	serr->ee.ee_type = 0;
467 	serr->ee.ee_code = 0;
468 	serr->ee.ee_pad = 0;
469 	serr->ee.ee_info = info;
470 	serr->ee.ee_data = 0;
471 	serr->addr_offset = (u8 *)&iph->daddr - skb_network_header(skb);
472 	serr->port = port;
473 
474 	__skb_pull(skb, skb_tail_pointer(skb) - skb->data);
475 	skb_reset_transport_header(skb);
476 
477 	if (sock_queue_err_skb(sk, skb))
478 		kfree_skb(skb);
479 }
480 
481 /* For some errors we have valid addr_offset even with zero payload and
482  * zero port. Also, addr_offset should be supported if port is set.
483  */
484 static inline bool ipv4_datagram_support_addr(struct sock_exterr_skb *serr)
485 {
486 	return serr->ee.ee_origin == SO_EE_ORIGIN_ICMP ||
487 	       serr->ee.ee_origin == SO_EE_ORIGIN_LOCAL || serr->port;
488 }
489 
490 /* IPv4 supports cmsg on all imcp errors and some timestamps
491  *
492  * Timestamp code paths do not initialize the fields expected by cmsg:
493  * the PKTINFO fields in skb->cb[]. Fill those in here.
494  */
495 static bool ipv4_datagram_support_cmsg(const struct sock *sk,
496 				       struct sk_buff *skb,
497 				       int ee_origin)
498 {
499 	struct in_pktinfo *info;
500 
501 	if (ee_origin == SO_EE_ORIGIN_ICMP)
502 		return true;
503 
504 	if (ee_origin == SO_EE_ORIGIN_LOCAL)
505 		return false;
506 
507 	/* Support IP_PKTINFO on tstamp packets if requested, to correlate
508 	 * timestamp with egress dev. Not possible for packets without iif
509 	 * or without payload (SOF_TIMESTAMPING_OPT_TSONLY).
510 	 */
511 	info = PKTINFO_SKB_CB(skb);
512 	if (!(READ_ONCE(sk->sk_tsflags) & SOF_TIMESTAMPING_OPT_CMSG) ||
513 	    !info->ipi_ifindex)
514 		return false;
515 
516 	info->ipi_spec_dst.s_addr = ip_hdr(skb)->saddr;
517 	return true;
518 }
519 
520 /*
521  *	Handle MSG_ERRQUEUE
522  */
523 int ip_recv_error(struct sock *sk, struct msghdr *msg, int len, int *addr_len)
524 {
525 	struct sock_exterr_skb *serr;
526 	struct sk_buff *skb;
527 	DECLARE_SOCKADDR(struct sockaddr_in *, sin, msg->msg_name);
528 	struct {
529 		struct sock_extended_err ee;
530 		struct sockaddr_in	 offender;
531 	} errhdr;
532 	int err;
533 	int copied;
534 
535 	err = -EAGAIN;
536 	skb = sock_dequeue_err_skb(sk);
537 	if (!skb)
538 		goto out;
539 
540 	copied = skb->len;
541 	if (copied > len) {
542 		msg->msg_flags |= MSG_TRUNC;
543 		copied = len;
544 	}
545 	err = skb_copy_datagram_msg(skb, 0, msg, copied);
546 	if (unlikely(err)) {
547 		kfree_skb(skb);
548 		return err;
549 	}
550 	sock_recv_timestamp(msg, sk, skb);
551 
552 	serr = SKB_EXT_ERR(skb);
553 
554 	if (sin && ipv4_datagram_support_addr(serr)) {
555 		sin->sin_family = AF_INET;
556 		sin->sin_addr.s_addr = *(__be32 *)(skb_network_header(skb) +
557 						   serr->addr_offset);
558 		sin->sin_port = serr->port;
559 		memset(&sin->sin_zero, 0, sizeof(sin->sin_zero));
560 		*addr_len = sizeof(*sin);
561 	}
562 
563 	memcpy(&errhdr.ee, &serr->ee, sizeof(struct sock_extended_err));
564 	sin = &errhdr.offender;
565 	memset(sin, 0, sizeof(*sin));
566 
567 	if (ipv4_datagram_support_cmsg(sk, skb, serr->ee.ee_origin)) {
568 		sin->sin_family = AF_INET;
569 		sin->sin_addr.s_addr = ip_hdr(skb)->saddr;
570 		if (inet_cmsg_flags(inet_sk(sk)))
571 			ip_cmsg_recv(msg, skb);
572 	}
573 
574 	put_cmsg(msg, SOL_IP, IP_RECVERR, sizeof(errhdr), &errhdr);
575 
576 	/* Now we could try to dump offended packet options */
577 
578 	msg->msg_flags |= MSG_ERRQUEUE;
579 	err = copied;
580 
581 	consume_skb(skb);
582 out:
583 	return err;
584 }
585 
586 void __ip_sock_set_tos(struct sock *sk, int val)
587 {
588 	u8 old_tos = inet_sk(sk)->tos;
589 
590 	if (sk->sk_type == SOCK_STREAM) {
591 		val &= ~INET_ECN_MASK;
592 		val |= old_tos & INET_ECN_MASK;
593 	}
594 	if (old_tos != val) {
595 		WRITE_ONCE(inet_sk(sk)->tos, val);
596 		WRITE_ONCE(sk->sk_priority, rt_tos2priority(val));
597 		sk_dst_reset(sk);
598 	}
599 }
600 
601 void ip_sock_set_tos(struct sock *sk, int val)
602 {
603 	sockopt_lock_sock(sk);
604 	__ip_sock_set_tos(sk, val);
605 	sockopt_release_sock(sk);
606 }
607 EXPORT_SYMBOL(ip_sock_set_tos);
608 
609 void ip_sock_set_freebind(struct sock *sk)
610 {
611 	inet_set_bit(FREEBIND, sk);
612 }
613 EXPORT_SYMBOL(ip_sock_set_freebind);
614 
615 void ip_sock_set_recverr(struct sock *sk)
616 {
617 	inet_set_bit(RECVERR, sk);
618 }
619 EXPORT_SYMBOL(ip_sock_set_recverr);
620 
621 int ip_sock_set_mtu_discover(struct sock *sk, int val)
622 {
623 	if (val < IP_PMTUDISC_DONT || val > IP_PMTUDISC_OMIT)
624 		return -EINVAL;
625 	WRITE_ONCE(inet_sk(sk)->pmtudisc, val);
626 	return 0;
627 }
628 EXPORT_SYMBOL(ip_sock_set_mtu_discover);
629 
630 void ip_sock_set_pktinfo(struct sock *sk)
631 {
632 	inet_set_bit(PKTINFO, sk);
633 }
634 EXPORT_SYMBOL(ip_sock_set_pktinfo);
635 
636 /*
637  *	Socket option code for IP. This is the end of the line after any
638  *	TCP,UDP etc options on an IP socket.
639  */
640 static bool setsockopt_needs_rtnl(int optname)
641 {
642 	switch (optname) {
643 	case IP_ADD_MEMBERSHIP:
644 	case IP_ADD_SOURCE_MEMBERSHIP:
645 	case IP_BLOCK_SOURCE:
646 	case IP_DROP_MEMBERSHIP:
647 	case IP_DROP_SOURCE_MEMBERSHIP:
648 	case IP_MSFILTER:
649 	case IP_UNBLOCK_SOURCE:
650 	case MCAST_BLOCK_SOURCE:
651 	case MCAST_MSFILTER:
652 	case MCAST_JOIN_GROUP:
653 	case MCAST_JOIN_SOURCE_GROUP:
654 	case MCAST_LEAVE_GROUP:
655 	case MCAST_LEAVE_SOURCE_GROUP:
656 	case MCAST_UNBLOCK_SOURCE:
657 		return true;
658 	}
659 	return false;
660 }
661 
662 static int set_mcast_msfilter(struct sock *sk, int ifindex,
663 			      int numsrc, int fmode,
664 			      struct sockaddr_storage *group,
665 			      struct sockaddr_storage *list)
666 {
667 	struct ip_msfilter *msf;
668 	struct sockaddr_in *psin;
669 	int err, i;
670 
671 	msf = kmalloc(IP_MSFILTER_SIZE(numsrc), GFP_KERNEL);
672 	if (!msf)
673 		return -ENOBUFS;
674 
675 	psin = (struct sockaddr_in *)group;
676 	if (psin->sin_family != AF_INET)
677 		goto Eaddrnotavail;
678 	msf->imsf_multiaddr = psin->sin_addr.s_addr;
679 	msf->imsf_interface = 0;
680 	msf->imsf_fmode = fmode;
681 	msf->imsf_numsrc = numsrc;
682 	for (i = 0; i < numsrc; ++i) {
683 		psin = (struct sockaddr_in *)&list[i];
684 
685 		if (psin->sin_family != AF_INET)
686 			goto Eaddrnotavail;
687 		msf->imsf_slist_flex[i] = psin->sin_addr.s_addr;
688 	}
689 	err = ip_mc_msfilter(sk, msf, ifindex);
690 	kfree(msf);
691 	return err;
692 
693 Eaddrnotavail:
694 	kfree(msf);
695 	return -EADDRNOTAVAIL;
696 }
697 
698 static int copy_group_source_from_sockptr(struct group_source_req *greqs,
699 		sockptr_t optval, int optlen)
700 {
701 	if (in_compat_syscall()) {
702 		struct compat_group_source_req gr32;
703 
704 		if (optlen != sizeof(gr32))
705 			return -EINVAL;
706 		if (copy_from_sockptr(&gr32, optval, sizeof(gr32)))
707 			return -EFAULT;
708 		greqs->gsr_interface = gr32.gsr_interface;
709 		greqs->gsr_group = gr32.gsr_group;
710 		greqs->gsr_source = gr32.gsr_source;
711 	} else {
712 		if (optlen != sizeof(*greqs))
713 			return -EINVAL;
714 		if (copy_from_sockptr(greqs, optval, sizeof(*greqs)))
715 			return -EFAULT;
716 	}
717 
718 	return 0;
719 }
720 
721 static int do_mcast_group_source(struct sock *sk, int optname,
722 		sockptr_t optval, int optlen)
723 {
724 	struct group_source_req greqs;
725 	struct ip_mreq_source mreqs;
726 	struct sockaddr_in *psin;
727 	int omode, add, err;
728 
729 	err = copy_group_source_from_sockptr(&greqs, optval, optlen);
730 	if (err)
731 		return err;
732 
733 	if (greqs.gsr_group.ss_family != AF_INET ||
734 	    greqs.gsr_source.ss_family != AF_INET)
735 		return -EADDRNOTAVAIL;
736 
737 	psin = (struct sockaddr_in *)&greqs.gsr_group;
738 	mreqs.imr_multiaddr = psin->sin_addr.s_addr;
739 	psin = (struct sockaddr_in *)&greqs.gsr_source;
740 	mreqs.imr_sourceaddr = psin->sin_addr.s_addr;
741 	mreqs.imr_interface = 0; /* use index for mc_source */
742 
743 	if (optname == MCAST_BLOCK_SOURCE) {
744 		omode = MCAST_EXCLUDE;
745 		add = 1;
746 	} else if (optname == MCAST_UNBLOCK_SOURCE) {
747 		omode = MCAST_EXCLUDE;
748 		add = 0;
749 	} else if (optname == MCAST_JOIN_SOURCE_GROUP) {
750 		struct ip_mreqn mreq;
751 
752 		psin = (struct sockaddr_in *)&greqs.gsr_group;
753 		mreq.imr_multiaddr = psin->sin_addr;
754 		mreq.imr_address.s_addr = 0;
755 		mreq.imr_ifindex = greqs.gsr_interface;
756 		err = ip_mc_join_group_ssm(sk, &mreq, MCAST_INCLUDE);
757 		if (err && err != -EADDRINUSE)
758 			return err;
759 		greqs.gsr_interface = mreq.imr_ifindex;
760 		omode = MCAST_INCLUDE;
761 		add = 1;
762 	} else /* MCAST_LEAVE_SOURCE_GROUP */ {
763 		omode = MCAST_INCLUDE;
764 		add = 0;
765 	}
766 	return ip_mc_source(add, omode, sk, &mreqs, greqs.gsr_interface);
767 }
768 
769 static int ip_set_mcast_msfilter(struct sock *sk, sockptr_t optval, int optlen)
770 {
771 	struct group_filter *gsf = NULL;
772 	int err;
773 
774 	if (optlen < GROUP_FILTER_SIZE(0))
775 		return -EINVAL;
776 	if (optlen > READ_ONCE(sock_net(sk)->core.sysctl_optmem_max))
777 		return -ENOBUFS;
778 
779 	gsf = memdup_sockptr(optval, optlen);
780 	if (IS_ERR(gsf))
781 		return PTR_ERR(gsf);
782 
783 	/* numsrc >= (4G-140)/128 overflow in 32 bits */
784 	err = -ENOBUFS;
785 	if (gsf->gf_numsrc >= 0x1ffffff ||
786 	    gsf->gf_numsrc > READ_ONCE(sock_net(sk)->ipv4.sysctl_igmp_max_msf))
787 		goto out_free_gsf;
788 
789 	err = -EINVAL;
790 	if (GROUP_FILTER_SIZE(gsf->gf_numsrc) > optlen)
791 		goto out_free_gsf;
792 
793 	err = set_mcast_msfilter(sk, gsf->gf_interface, gsf->gf_numsrc,
794 				 gsf->gf_fmode, &gsf->gf_group,
795 				 gsf->gf_slist_flex);
796 out_free_gsf:
797 	kfree(gsf);
798 	return err;
799 }
800 
801 static int compat_ip_set_mcast_msfilter(struct sock *sk, sockptr_t optval,
802 		int optlen)
803 {
804 	const int size0 = offsetof(struct compat_group_filter, gf_slist_flex);
805 	struct compat_group_filter *gf32;
806 	unsigned int n;
807 	void *p;
808 	int err;
809 
810 	if (optlen < size0)
811 		return -EINVAL;
812 	if (optlen > READ_ONCE(sock_net(sk)->core.sysctl_optmem_max) - 4)
813 		return -ENOBUFS;
814 
815 	p = kmalloc(optlen + 4, GFP_KERNEL);
816 	if (!p)
817 		return -ENOMEM;
818 	gf32 = p + 4; /* we want ->gf_group and ->gf_slist_flex aligned */
819 
820 	err = -EFAULT;
821 	if (copy_from_sockptr(gf32, optval, optlen))
822 		goto out_free_gsf;
823 
824 	/* numsrc >= (4G-140)/128 overflow in 32 bits */
825 	n = gf32->gf_numsrc;
826 	err = -ENOBUFS;
827 	if (n >= 0x1ffffff)
828 		goto out_free_gsf;
829 
830 	err = -EINVAL;
831 	if (offsetof(struct compat_group_filter, gf_slist_flex[n]) > optlen)
832 		goto out_free_gsf;
833 
834 	/* numsrc >= (4G-140)/128 overflow in 32 bits */
835 	err = -ENOBUFS;
836 	if (n > READ_ONCE(sock_net(sk)->ipv4.sysctl_igmp_max_msf))
837 		goto out_free_gsf;
838 	err = set_mcast_msfilter(sk, gf32->gf_interface, n, gf32->gf_fmode,
839 				 &gf32->gf_group, gf32->gf_slist_flex);
840 out_free_gsf:
841 	kfree(p);
842 	return err;
843 }
844 
845 static int ip_mcast_join_leave(struct sock *sk, int optname,
846 		sockptr_t optval, int optlen)
847 {
848 	struct ip_mreqn mreq = { };
849 	struct sockaddr_in *psin;
850 	struct group_req greq;
851 
852 	if (optlen < sizeof(struct group_req))
853 		return -EINVAL;
854 	if (copy_from_sockptr(&greq, optval, sizeof(greq)))
855 		return -EFAULT;
856 
857 	psin = (struct sockaddr_in *)&greq.gr_group;
858 	if (psin->sin_family != AF_INET)
859 		return -EINVAL;
860 	mreq.imr_multiaddr = psin->sin_addr;
861 	mreq.imr_ifindex = greq.gr_interface;
862 	if (optname == MCAST_JOIN_GROUP)
863 		return ip_mc_join_group(sk, &mreq);
864 	return ip_mc_leave_group(sk, &mreq);
865 }
866 
867 static int compat_ip_mcast_join_leave(struct sock *sk, int optname,
868 		sockptr_t optval, int optlen)
869 {
870 	struct compat_group_req greq;
871 	struct ip_mreqn mreq = { };
872 	struct sockaddr_in *psin;
873 
874 	if (optlen < sizeof(struct compat_group_req))
875 		return -EINVAL;
876 	if (copy_from_sockptr(&greq, optval, sizeof(greq)))
877 		return -EFAULT;
878 
879 	psin = (struct sockaddr_in *)&greq.gr_group;
880 	if (psin->sin_family != AF_INET)
881 		return -EINVAL;
882 	mreq.imr_multiaddr = psin->sin_addr;
883 	mreq.imr_ifindex = greq.gr_interface;
884 
885 	if (optname == MCAST_JOIN_GROUP)
886 		return ip_mc_join_group(sk, &mreq);
887 	return ip_mc_leave_group(sk, &mreq);
888 }
889 
890 DEFINE_STATIC_KEY_FALSE(ip4_min_ttl);
891 
892 int do_ip_setsockopt(struct sock *sk, int level, int optname,
893 		     sockptr_t optval, unsigned int optlen)
894 {
895 	struct inet_sock *inet = inet_sk(sk);
896 	struct net *net = sock_net(sk);
897 	int val = 0, err, retv;
898 	bool needs_rtnl = setsockopt_needs_rtnl(optname);
899 
900 	switch (optname) {
901 	case IP_PKTINFO:
902 	case IP_RECVTTL:
903 	case IP_RECVOPTS:
904 	case IP_RECVTOS:
905 	case IP_RETOPTS:
906 	case IP_TOS:
907 	case IP_TTL:
908 	case IP_HDRINCL:
909 	case IP_MTU_DISCOVER:
910 	case IP_RECVERR:
911 	case IP_ROUTER_ALERT:
912 	case IP_FREEBIND:
913 	case IP_PASSSEC:
914 	case IP_TRANSPARENT:
915 	case IP_MINTTL:
916 	case IP_NODEFRAG:
917 	case IP_BIND_ADDRESS_NO_PORT:
918 	case IP_UNICAST_IF:
919 	case IP_MULTICAST_TTL:
920 	case IP_MULTICAST_ALL:
921 	case IP_MULTICAST_LOOP:
922 	case IP_RECVORIGDSTADDR:
923 	case IP_CHECKSUM:
924 	case IP_RECVFRAGSIZE:
925 	case IP_RECVERR_RFC4884:
926 	case IP_LOCAL_PORT_RANGE:
927 		if (optlen >= sizeof(int)) {
928 			if (copy_from_sockptr(&val, optval, sizeof(val)))
929 				return -EFAULT;
930 		} else if (optlen >= sizeof(char)) {
931 			unsigned char ucval;
932 
933 			if (copy_from_sockptr(&ucval, optval, sizeof(ucval)))
934 				return -EFAULT;
935 			val = (int) ucval;
936 		}
937 	}
938 
939 	/* If optlen==0, it is equivalent to val == 0 */
940 
941 	if (optname == IP_ROUTER_ALERT) {
942 		retv = ip_ra_control(sk, val ? 1 : 0, NULL);
943 		if (retv == 0)
944 			inet_assign_bit(RTALERT, sk, val);
945 		return retv;
946 	}
947 	if (ip_mroute_opt(optname))
948 		return ip_mroute_setsockopt(sk, optname, optval, optlen);
949 
950 	/* Handle options that can be set without locking the socket. */
951 	switch (optname) {
952 	case IP_PKTINFO:
953 		inet_assign_bit(PKTINFO, sk, val);
954 		return 0;
955 	case IP_RECVTTL:
956 		inet_assign_bit(TTL, sk, val);
957 		return 0;
958 	case IP_RECVTOS:
959 		inet_assign_bit(TOS, sk, val);
960 		return 0;
961 	case IP_RECVOPTS:
962 		inet_assign_bit(RECVOPTS, sk, val);
963 		return 0;
964 	case IP_RETOPTS:
965 		inet_assign_bit(RETOPTS, sk, val);
966 		return 0;
967 	case IP_PASSSEC:
968 		inet_assign_bit(PASSSEC, sk, val);
969 		return 0;
970 	case IP_RECVORIGDSTADDR:
971 		inet_assign_bit(ORIGDSTADDR, sk, val);
972 		return 0;
973 	case IP_RECVFRAGSIZE:
974 		if (sk->sk_type != SOCK_RAW && sk->sk_type != SOCK_DGRAM)
975 			return -EINVAL;
976 		inet_assign_bit(RECVFRAGSIZE, sk, val);
977 		return 0;
978 	case IP_RECVERR:
979 		inet_assign_bit(RECVERR, sk, val);
980 		if (!val)
981 			skb_errqueue_purge(&sk->sk_error_queue);
982 		return 0;
983 	case IP_RECVERR_RFC4884:
984 		if (val < 0 || val > 1)
985 			return -EINVAL;
986 		inet_assign_bit(RECVERR_RFC4884, sk, val);
987 		return 0;
988 	case IP_FREEBIND:
989 		if (optlen < 1)
990 			return -EINVAL;
991 		inet_assign_bit(FREEBIND, sk, val);
992 		return 0;
993 	case IP_HDRINCL:
994 		if (sk->sk_type != SOCK_RAW)
995 			return -ENOPROTOOPT;
996 		inet_assign_bit(HDRINCL, sk, val);
997 		return 0;
998 	case IP_MULTICAST_LOOP:
999 		if (optlen < 1)
1000 			return -EINVAL;
1001 		inet_assign_bit(MC_LOOP, sk, val);
1002 		return 0;
1003 	case IP_MULTICAST_ALL:
1004 		if (optlen < 1)
1005 			return -EINVAL;
1006 		if (val != 0 && val != 1)
1007 			return -EINVAL;
1008 		inet_assign_bit(MC_ALL, sk, val);
1009 		return 0;
1010 	case IP_TRANSPARENT:
1011 		if (!!val && !sockopt_ns_capable(sock_net(sk)->user_ns, CAP_NET_RAW) &&
1012 		    !sockopt_ns_capable(sock_net(sk)->user_ns, CAP_NET_ADMIN))
1013 			return -EPERM;
1014 		if (optlen < 1)
1015 			return -EINVAL;
1016 		inet_assign_bit(TRANSPARENT, sk, val);
1017 		return 0;
1018 	case IP_NODEFRAG:
1019 		if (sk->sk_type != SOCK_RAW)
1020 			return -ENOPROTOOPT;
1021 		inet_assign_bit(NODEFRAG, sk, val);
1022 		return 0;
1023 	case IP_BIND_ADDRESS_NO_PORT:
1024 		inet_assign_bit(BIND_ADDRESS_NO_PORT, sk, val);
1025 		return 0;
1026 	case IP_TTL:
1027 		if (optlen < 1)
1028 			return -EINVAL;
1029 		if (val != -1 && (val < 1 || val > 255))
1030 			return -EINVAL;
1031 		WRITE_ONCE(inet->uc_ttl, val);
1032 		return 0;
1033 	case IP_MINTTL:
1034 		if (optlen < 1)
1035 			return -EINVAL;
1036 		if (val < 0 || val > 255)
1037 			return -EINVAL;
1038 
1039 		if (val)
1040 			static_branch_enable(&ip4_min_ttl);
1041 
1042 		WRITE_ONCE(inet->min_ttl, val);
1043 		return 0;
1044 	case IP_MULTICAST_TTL:
1045 		if (sk->sk_type == SOCK_STREAM)
1046 			return -EINVAL;
1047 		if (optlen < 1)
1048 			return -EINVAL;
1049 		if (val == -1)
1050 			val = 1;
1051 		if (val < 0 || val > 255)
1052 			return -EINVAL;
1053 		WRITE_ONCE(inet->mc_ttl, val);
1054 		return 0;
1055 	case IP_MTU_DISCOVER:
1056 		return ip_sock_set_mtu_discover(sk, val);
1057 	case IP_TOS:	/* This sets both TOS and Precedence */
1058 		ip_sock_set_tos(sk, val);
1059 		return 0;
1060 	case IP_LOCAL_PORT_RANGE:
1061 	{
1062 		u16 lo = val;
1063 		u16 hi = val >> 16;
1064 
1065 		if (optlen != sizeof(u32))
1066 			return -EINVAL;
1067 		if (lo != 0 && hi != 0 && lo > hi)
1068 			return -EINVAL;
1069 
1070 		WRITE_ONCE(inet->local_port_range, val);
1071 		return 0;
1072 	}
1073 	}
1074 
1075 	err = 0;
1076 	if (needs_rtnl)
1077 		rtnl_lock();
1078 	sockopt_lock_sock(sk);
1079 
1080 	switch (optname) {
1081 	case IP_OPTIONS:
1082 	{
1083 		struct ip_options_rcu *old, *opt = NULL;
1084 
1085 		if (optlen > 40)
1086 			goto e_inval;
1087 		err = ip_options_get(sock_net(sk), &opt, optval, optlen);
1088 		if (err)
1089 			break;
1090 		old = rcu_dereference_protected(inet->inet_opt,
1091 						lockdep_sock_is_held(sk));
1092 		if (inet_test_bit(IS_ICSK, sk)) {
1093 			struct inet_connection_sock *icsk = inet_csk(sk);
1094 #if IS_ENABLED(CONFIG_IPV6)
1095 			if (sk->sk_family == PF_INET ||
1096 			    (!((1 << sk->sk_state) &
1097 			       (TCPF_LISTEN | TCPF_CLOSE)) &&
1098 			     inet->inet_daddr != LOOPBACK4_IPV6)) {
1099 #endif
1100 				if (old)
1101 					icsk->icsk_ext_hdr_len -= old->opt.optlen;
1102 				if (opt)
1103 					icsk->icsk_ext_hdr_len += opt->opt.optlen;
1104 				icsk->icsk_sync_mss(sk, icsk->icsk_pmtu_cookie);
1105 #if IS_ENABLED(CONFIG_IPV6)
1106 			}
1107 #endif
1108 		}
1109 		rcu_assign_pointer(inet->inet_opt, opt);
1110 		if (old)
1111 			kfree_rcu(old, rcu);
1112 		break;
1113 	}
1114 	case IP_CHECKSUM:
1115 		if (val) {
1116 			if (!(inet_test_bit(CHECKSUM, sk))) {
1117 				inet_inc_convert_csum(sk);
1118 				inet_set_bit(CHECKSUM, sk);
1119 			}
1120 		} else {
1121 			if (inet_test_bit(CHECKSUM, sk)) {
1122 				inet_dec_convert_csum(sk);
1123 				inet_clear_bit(CHECKSUM, sk);
1124 			}
1125 		}
1126 		break;
1127 	case IP_UNICAST_IF:
1128 	{
1129 		struct net_device *dev = NULL;
1130 		int ifindex;
1131 		int midx;
1132 
1133 		if (optlen != sizeof(int))
1134 			goto e_inval;
1135 
1136 		ifindex = (__force int)ntohl((__force __be32)val);
1137 		if (ifindex == 0) {
1138 			WRITE_ONCE(inet->uc_index, 0);
1139 			err = 0;
1140 			break;
1141 		}
1142 
1143 		dev = dev_get_by_index(sock_net(sk), ifindex);
1144 		err = -EADDRNOTAVAIL;
1145 		if (!dev)
1146 			break;
1147 
1148 		midx = l3mdev_master_ifindex(dev);
1149 		dev_put(dev);
1150 
1151 		err = -EINVAL;
1152 		if (sk->sk_bound_dev_if && midx != sk->sk_bound_dev_if)
1153 			break;
1154 
1155 		WRITE_ONCE(inet->uc_index, ifindex);
1156 		err = 0;
1157 		break;
1158 	}
1159 	case IP_MULTICAST_IF:
1160 	{
1161 		struct ip_mreqn mreq;
1162 		struct net_device *dev = NULL;
1163 		int midx;
1164 
1165 		if (sk->sk_type == SOCK_STREAM)
1166 			goto e_inval;
1167 		/*
1168 		 *	Check the arguments are allowable
1169 		 */
1170 
1171 		if (optlen < sizeof(struct in_addr))
1172 			goto e_inval;
1173 
1174 		err = -EFAULT;
1175 		if (optlen >= sizeof(struct ip_mreqn)) {
1176 			if (copy_from_sockptr(&mreq, optval, sizeof(mreq)))
1177 				break;
1178 		} else {
1179 			memset(&mreq, 0, sizeof(mreq));
1180 			if (optlen >= sizeof(struct ip_mreq)) {
1181 				if (copy_from_sockptr(&mreq, optval,
1182 						      sizeof(struct ip_mreq)))
1183 					break;
1184 			} else if (optlen >= sizeof(struct in_addr)) {
1185 				if (copy_from_sockptr(&mreq.imr_address, optval,
1186 						      sizeof(struct in_addr)))
1187 					break;
1188 			}
1189 		}
1190 
1191 		if (!mreq.imr_ifindex) {
1192 			if (mreq.imr_address.s_addr == htonl(INADDR_ANY)) {
1193 				WRITE_ONCE(inet->mc_index, 0);
1194 				WRITE_ONCE(inet->mc_addr, 0);
1195 				err = 0;
1196 				break;
1197 			}
1198 			dev = ip_dev_find(sock_net(sk), mreq.imr_address.s_addr);
1199 			if (dev)
1200 				mreq.imr_ifindex = dev->ifindex;
1201 		} else
1202 			dev = dev_get_by_index(sock_net(sk), mreq.imr_ifindex);
1203 
1204 
1205 		err = -EADDRNOTAVAIL;
1206 		if (!dev)
1207 			break;
1208 
1209 		midx = l3mdev_master_ifindex(dev);
1210 
1211 		dev_put(dev);
1212 
1213 		err = -EINVAL;
1214 		if (sk->sk_bound_dev_if &&
1215 		    mreq.imr_ifindex != sk->sk_bound_dev_if &&
1216 		    midx != sk->sk_bound_dev_if)
1217 			break;
1218 
1219 		WRITE_ONCE(inet->mc_index, mreq.imr_ifindex);
1220 		WRITE_ONCE(inet->mc_addr, mreq.imr_address.s_addr);
1221 		err = 0;
1222 		break;
1223 	}
1224 
1225 	case IP_ADD_MEMBERSHIP:
1226 	case IP_DROP_MEMBERSHIP:
1227 	{
1228 		struct ip_mreqn mreq;
1229 
1230 		err = -EPROTO;
1231 		if (inet_test_bit(IS_ICSK, sk))
1232 			break;
1233 
1234 		if (optlen < sizeof(struct ip_mreq))
1235 			goto e_inval;
1236 		err = -EFAULT;
1237 		if (optlen >= sizeof(struct ip_mreqn)) {
1238 			if (copy_from_sockptr(&mreq, optval, sizeof(mreq)))
1239 				break;
1240 		} else {
1241 			memset(&mreq, 0, sizeof(mreq));
1242 			if (copy_from_sockptr(&mreq, optval,
1243 					      sizeof(struct ip_mreq)))
1244 				break;
1245 		}
1246 
1247 		if (optname == IP_ADD_MEMBERSHIP)
1248 			err = ip_mc_join_group(sk, &mreq);
1249 		else
1250 			err = ip_mc_leave_group(sk, &mreq);
1251 		break;
1252 	}
1253 	case IP_MSFILTER:
1254 	{
1255 		struct ip_msfilter *msf;
1256 
1257 		if (optlen < IP_MSFILTER_SIZE(0))
1258 			goto e_inval;
1259 		if (optlen > READ_ONCE(net->core.sysctl_optmem_max)) {
1260 			err = -ENOBUFS;
1261 			break;
1262 		}
1263 		msf = memdup_sockptr(optval, optlen);
1264 		if (IS_ERR(msf)) {
1265 			err = PTR_ERR(msf);
1266 			break;
1267 		}
1268 		/* numsrc >= (1G-4) overflow in 32 bits */
1269 		if (msf->imsf_numsrc >= 0x3ffffffcU ||
1270 		    msf->imsf_numsrc > READ_ONCE(net->ipv4.sysctl_igmp_max_msf)) {
1271 			kfree(msf);
1272 			err = -ENOBUFS;
1273 			break;
1274 		}
1275 		if (IP_MSFILTER_SIZE(msf->imsf_numsrc) > optlen) {
1276 			kfree(msf);
1277 			err = -EINVAL;
1278 			break;
1279 		}
1280 		err = ip_mc_msfilter(sk, msf, 0);
1281 		kfree(msf);
1282 		break;
1283 	}
1284 	case IP_BLOCK_SOURCE:
1285 	case IP_UNBLOCK_SOURCE:
1286 	case IP_ADD_SOURCE_MEMBERSHIP:
1287 	case IP_DROP_SOURCE_MEMBERSHIP:
1288 	{
1289 		struct ip_mreq_source mreqs;
1290 		int omode, add;
1291 
1292 		if (optlen != sizeof(struct ip_mreq_source))
1293 			goto e_inval;
1294 		if (copy_from_sockptr(&mreqs, optval, sizeof(mreqs))) {
1295 			err = -EFAULT;
1296 			break;
1297 		}
1298 		if (optname == IP_BLOCK_SOURCE) {
1299 			omode = MCAST_EXCLUDE;
1300 			add = 1;
1301 		} else if (optname == IP_UNBLOCK_SOURCE) {
1302 			omode = MCAST_EXCLUDE;
1303 			add = 0;
1304 		} else if (optname == IP_ADD_SOURCE_MEMBERSHIP) {
1305 			struct ip_mreqn mreq;
1306 
1307 			mreq.imr_multiaddr.s_addr = mreqs.imr_multiaddr;
1308 			mreq.imr_address.s_addr = mreqs.imr_interface;
1309 			mreq.imr_ifindex = 0;
1310 			err = ip_mc_join_group_ssm(sk, &mreq, MCAST_INCLUDE);
1311 			if (err && err != -EADDRINUSE)
1312 				break;
1313 			omode = MCAST_INCLUDE;
1314 			add = 1;
1315 		} else /* IP_DROP_SOURCE_MEMBERSHIP */ {
1316 			omode = MCAST_INCLUDE;
1317 			add = 0;
1318 		}
1319 		err = ip_mc_source(add, omode, sk, &mreqs, 0);
1320 		break;
1321 	}
1322 	case MCAST_JOIN_GROUP:
1323 	case MCAST_LEAVE_GROUP:
1324 		if (in_compat_syscall())
1325 			err = compat_ip_mcast_join_leave(sk, optname, optval,
1326 							 optlen);
1327 		else
1328 			err = ip_mcast_join_leave(sk, optname, optval, optlen);
1329 		break;
1330 	case MCAST_JOIN_SOURCE_GROUP:
1331 	case MCAST_LEAVE_SOURCE_GROUP:
1332 	case MCAST_BLOCK_SOURCE:
1333 	case MCAST_UNBLOCK_SOURCE:
1334 		err = do_mcast_group_source(sk, optname, optval, optlen);
1335 		break;
1336 	case MCAST_MSFILTER:
1337 		if (in_compat_syscall())
1338 			err = compat_ip_set_mcast_msfilter(sk, optval, optlen);
1339 		else
1340 			err = ip_set_mcast_msfilter(sk, optval, optlen);
1341 		break;
1342 	case IP_IPSEC_POLICY:
1343 	case IP_XFRM_POLICY:
1344 		err = -EPERM;
1345 		if (!sockopt_ns_capable(sock_net(sk)->user_ns, CAP_NET_ADMIN))
1346 			break;
1347 		err = xfrm_user_policy(sk, optname, optval, optlen);
1348 		break;
1349 
1350 	default:
1351 		err = -ENOPROTOOPT;
1352 		break;
1353 	}
1354 	sockopt_release_sock(sk);
1355 	if (needs_rtnl)
1356 		rtnl_unlock();
1357 	return err;
1358 
1359 e_inval:
1360 	sockopt_release_sock(sk);
1361 	if (needs_rtnl)
1362 		rtnl_unlock();
1363 	return -EINVAL;
1364 }
1365 
1366 /**
1367  * ipv4_pktinfo_prepare - transfer some info from rtable to skb
1368  * @sk: socket
1369  * @skb: buffer
1370  * @drop_dst: if true, drops skb dst
1371  *
1372  * To support IP_CMSG_PKTINFO option, we store rt_iif and specific
1373  * destination in skb->cb[] before dst drop.
1374  * This way, receiver doesn't make cache line misses to read rtable.
1375  */
1376 void ipv4_pktinfo_prepare(const struct sock *sk, struct sk_buff *skb, bool drop_dst)
1377 {
1378 	struct in_pktinfo *pktinfo = PKTINFO_SKB_CB(skb);
1379 	bool prepare = inet_test_bit(PKTINFO, sk) ||
1380 		       ipv6_sk_rxinfo(sk);
1381 
1382 	if (prepare && skb_rtable(skb)) {
1383 		/* skb->cb is overloaded: prior to this point it is IP{6}CB
1384 		 * which has interface index (iif) as the first member of the
1385 		 * underlying inet{6}_skb_parm struct. This code then overlays
1386 		 * PKTINFO_SKB_CB and in_pktinfo also has iif as the first
1387 		 * element so the iif is picked up from the prior IPCB. If iif
1388 		 * is the loopback interface, then return the sending interface
1389 		 * (e.g., process binds socket to eth0 for Tx which is
1390 		 * redirected to loopback in the rtable/dst).
1391 		 */
1392 		struct rtable *rt = skb_rtable(skb);
1393 		bool l3slave = ipv4_l3mdev_skb(IPCB(skb)->flags);
1394 
1395 		if (pktinfo->ipi_ifindex == LOOPBACK_IFINDEX)
1396 			pktinfo->ipi_ifindex = inet_iif(skb);
1397 		else if (l3slave && rt && rt->rt_iif)
1398 			pktinfo->ipi_ifindex = rt->rt_iif;
1399 
1400 		pktinfo->ipi_spec_dst.s_addr = fib_compute_spec_dst(skb);
1401 	} else {
1402 		pktinfo->ipi_ifindex = 0;
1403 		pktinfo->ipi_spec_dst.s_addr = 0;
1404 	}
1405 	if (drop_dst)
1406 		skb_dst_drop(skb);
1407 }
1408 
1409 int ip_setsockopt(struct sock *sk, int level, int optname, sockptr_t optval,
1410 		unsigned int optlen)
1411 {
1412 	int err;
1413 
1414 	if (level != SOL_IP)
1415 		return -ENOPROTOOPT;
1416 
1417 	err = do_ip_setsockopt(sk, level, optname, optval, optlen);
1418 #ifdef CONFIG_NETFILTER
1419 	/* we need to exclude all possible ENOPROTOOPTs except default case */
1420 	if (err == -ENOPROTOOPT && optname != IP_HDRINCL &&
1421 			optname != IP_IPSEC_POLICY &&
1422 			optname != IP_XFRM_POLICY &&
1423 			!ip_mroute_opt(optname))
1424 		err = nf_setsockopt(sk, PF_INET, optname, optval, optlen);
1425 #endif
1426 	return err;
1427 }
1428 EXPORT_SYMBOL(ip_setsockopt);
1429 
1430 /*
1431  *	Get the options. Note for future reference. The GET of IP options gets
1432  *	the _received_ ones. The set sets the _sent_ ones.
1433  */
1434 
1435 static bool getsockopt_needs_rtnl(int optname)
1436 {
1437 	switch (optname) {
1438 	case IP_MSFILTER:
1439 	case MCAST_MSFILTER:
1440 		return true;
1441 	}
1442 	return false;
1443 }
1444 
1445 static int ip_get_mcast_msfilter(struct sock *sk, sockptr_t optval,
1446 				 sockptr_t optlen, int len)
1447 {
1448 	const int size0 = offsetof(struct group_filter, gf_slist_flex);
1449 	struct group_filter gsf;
1450 	int num, gsf_size;
1451 	int err;
1452 
1453 	if (len < size0)
1454 		return -EINVAL;
1455 	if (copy_from_sockptr(&gsf, optval, size0))
1456 		return -EFAULT;
1457 
1458 	num = gsf.gf_numsrc;
1459 	err = ip_mc_gsfget(sk, &gsf, optval,
1460 			   offsetof(struct group_filter, gf_slist_flex));
1461 	if (err)
1462 		return err;
1463 	if (gsf.gf_numsrc < num)
1464 		num = gsf.gf_numsrc;
1465 	gsf_size = GROUP_FILTER_SIZE(num);
1466 	if (copy_to_sockptr(optlen, &gsf_size, sizeof(int)) ||
1467 	    copy_to_sockptr(optval, &gsf, size0))
1468 		return -EFAULT;
1469 	return 0;
1470 }
1471 
1472 static int compat_ip_get_mcast_msfilter(struct sock *sk, sockptr_t optval,
1473 					sockptr_t optlen, int len)
1474 {
1475 	const int size0 = offsetof(struct compat_group_filter, gf_slist_flex);
1476 	struct compat_group_filter gf32;
1477 	struct group_filter gf;
1478 	int num;
1479 	int err;
1480 
1481 	if (len < size0)
1482 		return -EINVAL;
1483 	if (copy_from_sockptr(&gf32, optval, size0))
1484 		return -EFAULT;
1485 
1486 	gf.gf_interface = gf32.gf_interface;
1487 	gf.gf_fmode = gf32.gf_fmode;
1488 	num = gf.gf_numsrc = gf32.gf_numsrc;
1489 	gf.gf_group = gf32.gf_group;
1490 
1491 	err = ip_mc_gsfget(sk, &gf, optval,
1492 			   offsetof(struct compat_group_filter, gf_slist_flex));
1493 	if (err)
1494 		return err;
1495 	if (gf.gf_numsrc < num)
1496 		num = gf.gf_numsrc;
1497 	len = GROUP_FILTER_SIZE(num) - (sizeof(gf) - sizeof(gf32));
1498 	if (copy_to_sockptr(optlen, &len, sizeof(int)) ||
1499 	    copy_to_sockptr_offset(optval, offsetof(struct compat_group_filter, gf_fmode),
1500 				   &gf.gf_fmode, sizeof(gf.gf_fmode)) ||
1501 	    copy_to_sockptr_offset(optval, offsetof(struct compat_group_filter, gf_numsrc),
1502 				   &gf.gf_numsrc, sizeof(gf.gf_numsrc)))
1503 		return -EFAULT;
1504 	return 0;
1505 }
1506 
1507 int do_ip_getsockopt(struct sock *sk, int level, int optname,
1508 		     sockptr_t optval, sockptr_t optlen)
1509 {
1510 	struct inet_sock *inet = inet_sk(sk);
1511 	bool needs_rtnl = getsockopt_needs_rtnl(optname);
1512 	int val, err = 0;
1513 	int len;
1514 
1515 	if (level != SOL_IP)
1516 		return -EOPNOTSUPP;
1517 
1518 	if (ip_mroute_opt(optname))
1519 		return ip_mroute_getsockopt(sk, optname, optval, optlen);
1520 
1521 	if (copy_from_sockptr(&len, optlen, sizeof(int)))
1522 		return -EFAULT;
1523 	if (len < 0)
1524 		return -EINVAL;
1525 
1526 	/* Handle options that can be read without locking the socket. */
1527 	switch (optname) {
1528 	case IP_PKTINFO:
1529 		val = inet_test_bit(PKTINFO, sk);
1530 		goto copyval;
1531 	case IP_RECVTTL:
1532 		val = inet_test_bit(TTL, sk);
1533 		goto copyval;
1534 	case IP_RECVTOS:
1535 		val = inet_test_bit(TOS, sk);
1536 		goto copyval;
1537 	case IP_RECVOPTS:
1538 		val = inet_test_bit(RECVOPTS, sk);
1539 		goto copyval;
1540 	case IP_RETOPTS:
1541 		val = inet_test_bit(RETOPTS, sk);
1542 		goto copyval;
1543 	case IP_PASSSEC:
1544 		val = inet_test_bit(PASSSEC, sk);
1545 		goto copyval;
1546 	case IP_RECVORIGDSTADDR:
1547 		val = inet_test_bit(ORIGDSTADDR, sk);
1548 		goto copyval;
1549 	case IP_CHECKSUM:
1550 		val = inet_test_bit(CHECKSUM, sk);
1551 		goto copyval;
1552 	case IP_RECVFRAGSIZE:
1553 		val = inet_test_bit(RECVFRAGSIZE, sk);
1554 		goto copyval;
1555 	case IP_RECVERR:
1556 		val = inet_test_bit(RECVERR, sk);
1557 		goto copyval;
1558 	case IP_RECVERR_RFC4884:
1559 		val = inet_test_bit(RECVERR_RFC4884, sk);
1560 		goto copyval;
1561 	case IP_FREEBIND:
1562 		val = inet_test_bit(FREEBIND, sk);
1563 		goto copyval;
1564 	case IP_HDRINCL:
1565 		val = inet_test_bit(HDRINCL, sk);
1566 		goto copyval;
1567 	case IP_MULTICAST_LOOP:
1568 		val = inet_test_bit(MC_LOOP, sk);
1569 		goto copyval;
1570 	case IP_MULTICAST_ALL:
1571 		val = inet_test_bit(MC_ALL, sk);
1572 		goto copyval;
1573 	case IP_TRANSPARENT:
1574 		val = inet_test_bit(TRANSPARENT, sk);
1575 		goto copyval;
1576 	case IP_NODEFRAG:
1577 		val = inet_test_bit(NODEFRAG, sk);
1578 		goto copyval;
1579 	case IP_BIND_ADDRESS_NO_PORT:
1580 		val = inet_test_bit(BIND_ADDRESS_NO_PORT, sk);
1581 		goto copyval;
1582 	case IP_ROUTER_ALERT:
1583 		val = inet_test_bit(RTALERT, sk);
1584 		goto copyval;
1585 	case IP_TTL:
1586 		val = READ_ONCE(inet->uc_ttl);
1587 		if (val < 0)
1588 			val = READ_ONCE(sock_net(sk)->ipv4.sysctl_ip_default_ttl);
1589 		goto copyval;
1590 	case IP_MINTTL:
1591 		val = READ_ONCE(inet->min_ttl);
1592 		goto copyval;
1593 	case IP_MULTICAST_TTL:
1594 		val = READ_ONCE(inet->mc_ttl);
1595 		goto copyval;
1596 	case IP_MTU_DISCOVER:
1597 		val = READ_ONCE(inet->pmtudisc);
1598 		goto copyval;
1599 	case IP_TOS:
1600 		val = READ_ONCE(inet->tos);
1601 		goto copyval;
1602 	case IP_OPTIONS:
1603 	{
1604 		unsigned char optbuf[sizeof(struct ip_options)+40];
1605 		struct ip_options *opt = (struct ip_options *)optbuf;
1606 		struct ip_options_rcu *inet_opt;
1607 
1608 		rcu_read_lock();
1609 		inet_opt = rcu_dereference(inet->inet_opt);
1610 		opt->optlen = 0;
1611 		if (inet_opt)
1612 			memcpy(optbuf, &inet_opt->opt,
1613 			       sizeof(struct ip_options) +
1614 			       inet_opt->opt.optlen);
1615 		rcu_read_unlock();
1616 
1617 		if (opt->optlen == 0) {
1618 			len = 0;
1619 			return copy_to_sockptr(optlen, &len, sizeof(int));
1620 		}
1621 
1622 		ip_options_undo(opt);
1623 
1624 		len = min_t(unsigned int, len, opt->optlen);
1625 		if (copy_to_sockptr(optlen, &len, sizeof(int)))
1626 			return -EFAULT;
1627 		if (copy_to_sockptr(optval, opt->__data, len))
1628 			return -EFAULT;
1629 		return 0;
1630 	}
1631 	case IP_MTU:
1632 	{
1633 		struct dst_entry *dst;
1634 		val = 0;
1635 		dst = sk_dst_get(sk);
1636 		if (dst) {
1637 			val = dst_mtu(dst);
1638 			dst_release(dst);
1639 		}
1640 		if (!val)
1641 			return -ENOTCONN;
1642 		goto copyval;
1643 	}
1644 	case IP_PKTOPTIONS:
1645 	{
1646 		struct msghdr msg;
1647 
1648 		if (sk->sk_type != SOCK_STREAM)
1649 			return -ENOPROTOOPT;
1650 
1651 		if (optval.is_kernel) {
1652 			msg.msg_control_is_user = false;
1653 			msg.msg_control = optval.kernel;
1654 		} else {
1655 			msg.msg_control_is_user = true;
1656 			msg.msg_control_user = optval.user;
1657 		}
1658 		msg.msg_controllen = len;
1659 		msg.msg_flags = in_compat_syscall() ? MSG_CMSG_COMPAT : 0;
1660 
1661 		if (inet_test_bit(PKTINFO, sk)) {
1662 			struct in_pktinfo info;
1663 
1664 			info.ipi_addr.s_addr = READ_ONCE(inet->inet_rcv_saddr);
1665 			info.ipi_spec_dst.s_addr = READ_ONCE(inet->inet_rcv_saddr);
1666 			info.ipi_ifindex = READ_ONCE(inet->mc_index);
1667 			put_cmsg(&msg, SOL_IP, IP_PKTINFO, sizeof(info), &info);
1668 		}
1669 		if (inet_test_bit(TTL, sk)) {
1670 			int hlim = READ_ONCE(inet->mc_ttl);
1671 
1672 			put_cmsg(&msg, SOL_IP, IP_TTL, sizeof(hlim), &hlim);
1673 		}
1674 		if (inet_test_bit(TOS, sk)) {
1675 			int tos = READ_ONCE(inet->rcv_tos);
1676 			put_cmsg(&msg, SOL_IP, IP_TOS, sizeof(tos), &tos);
1677 		}
1678 		len -= msg.msg_controllen;
1679 		return copy_to_sockptr(optlen, &len, sizeof(int));
1680 	}
1681 	case IP_UNICAST_IF:
1682 		val = (__force int)htonl((__u32) READ_ONCE(inet->uc_index));
1683 		goto copyval;
1684 	case IP_MULTICAST_IF:
1685 	{
1686 		struct in_addr addr;
1687 		len = min_t(unsigned int, len, sizeof(struct in_addr));
1688 		addr.s_addr = READ_ONCE(inet->mc_addr);
1689 
1690 		if (copy_to_sockptr(optlen, &len, sizeof(int)))
1691 			return -EFAULT;
1692 		if (copy_to_sockptr(optval, &addr, len))
1693 			return -EFAULT;
1694 		return 0;
1695 	}
1696 	case IP_LOCAL_PORT_RANGE:
1697 		val = READ_ONCE(inet->local_port_range);
1698 		goto copyval;
1699 	}
1700 
1701 	if (needs_rtnl)
1702 		rtnl_lock();
1703 	sockopt_lock_sock(sk);
1704 
1705 	switch (optname) {
1706 	case IP_MSFILTER:
1707 	{
1708 		struct ip_msfilter msf;
1709 
1710 		if (len < IP_MSFILTER_SIZE(0)) {
1711 			err = -EINVAL;
1712 			goto out;
1713 		}
1714 		if (copy_from_sockptr(&msf, optval, IP_MSFILTER_SIZE(0))) {
1715 			err = -EFAULT;
1716 			goto out;
1717 		}
1718 		err = ip_mc_msfget(sk, &msf, optval, optlen);
1719 		goto out;
1720 	}
1721 	case MCAST_MSFILTER:
1722 		if (in_compat_syscall())
1723 			err = compat_ip_get_mcast_msfilter(sk, optval, optlen,
1724 							   len);
1725 		else
1726 			err = ip_get_mcast_msfilter(sk, optval, optlen, len);
1727 		goto out;
1728 	case IP_PROTOCOL:
1729 		val = inet_sk(sk)->inet_num;
1730 		break;
1731 	default:
1732 		sockopt_release_sock(sk);
1733 		return -ENOPROTOOPT;
1734 	}
1735 	sockopt_release_sock(sk);
1736 copyval:
1737 	if (len < sizeof(int) && len > 0 && val >= 0 && val <= 255) {
1738 		unsigned char ucval = (unsigned char)val;
1739 		len = 1;
1740 		if (copy_to_sockptr(optlen, &len, sizeof(int)))
1741 			return -EFAULT;
1742 		if (copy_to_sockptr(optval, &ucval, 1))
1743 			return -EFAULT;
1744 	} else {
1745 		len = min_t(unsigned int, sizeof(int), len);
1746 		if (copy_to_sockptr(optlen, &len, sizeof(int)))
1747 			return -EFAULT;
1748 		if (copy_to_sockptr(optval, &val, len))
1749 			return -EFAULT;
1750 	}
1751 	return 0;
1752 
1753 out:
1754 	sockopt_release_sock(sk);
1755 	if (needs_rtnl)
1756 		rtnl_unlock();
1757 	return err;
1758 }
1759 
1760 int ip_getsockopt(struct sock *sk, int level,
1761 		  int optname, char __user *optval, int __user *optlen)
1762 {
1763 	int err;
1764 
1765 	err = do_ip_getsockopt(sk, level, optname,
1766 			       USER_SOCKPTR(optval), USER_SOCKPTR(optlen));
1767 
1768 #ifdef CONFIG_NETFILTER
1769 	/* we need to exclude all possible ENOPROTOOPTs except default case */
1770 	if (err == -ENOPROTOOPT && optname != IP_PKTOPTIONS &&
1771 			!ip_mroute_opt(optname)) {
1772 		int len;
1773 
1774 		if (get_user(len, optlen))
1775 			return -EFAULT;
1776 
1777 		err = nf_getsockopt(sk, PF_INET, optname, optval, &len);
1778 		if (err >= 0)
1779 			err = put_user(len, optlen);
1780 		return err;
1781 	}
1782 #endif
1783 	return err;
1784 }
1785 EXPORT_SYMBOL(ip_getsockopt);
1786