xref: /linux/net/bridge/br_netfilter_hooks.c (revision 3fd6c59042dbba50391e30862beac979491145fe)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  *	Handle firewalling
4  *	Linux ethernet bridge
5  *
6  *	Authors:
7  *	Lennert Buytenhek		<buytenh@gnu.org>
8  *	Bart De Schuymer		<bdschuym@pandora.be>
9  *
10  *	Lennert dedicates this file to Kerstin Wurdinger.
11  */
12 
13 #include <linux/module.h>
14 #include <linux/kernel.h>
15 #include <linux/slab.h>
16 #include <linux/ip.h>
17 #include <linux/netdevice.h>
18 #include <linux/skbuff.h>
19 #include <linux/if_arp.h>
20 #include <linux/if_ether.h>
21 #include <linux/if_vlan.h>
22 #include <linux/if_pppox.h>
23 #include <linux/ppp_defs.h>
24 #include <linux/netfilter_bridge.h>
25 #include <uapi/linux/netfilter_bridge.h>
26 #include <linux/netfilter_ipv4.h>
27 #include <linux/netfilter_ipv6.h>
28 #include <linux/netfilter_arp.h>
29 #include <linux/in_route.h>
30 #include <linux/rculist.h>
31 #include <linux/inetdevice.h>
32 
33 #include <net/ip.h>
34 #include <net/ipv6.h>
35 #include <net/addrconf.h>
36 #include <net/dst_metadata.h>
37 #include <net/route.h>
38 #include <net/netfilter/br_netfilter.h>
39 #include <net/netns/generic.h>
40 #include <net/inet_dscp.h>
41 
42 #include <linux/uaccess.h>
43 #include "br_private.h"
44 #ifdef CONFIG_SYSCTL
45 #include <linux/sysctl.h>
46 #endif
47 
48 #if IS_ENABLED(CONFIG_NF_CONNTRACK)
49 #include <net/netfilter/nf_conntrack_core.h>
50 #endif
51 
52 static unsigned int brnf_net_id __read_mostly;
53 
54 struct brnf_net {
55 	bool enabled;
56 
57 #ifdef CONFIG_SYSCTL
58 	struct ctl_table_header *ctl_hdr;
59 #endif
60 
61 	/* default value is 1 */
62 	int call_iptables;
63 	int call_ip6tables;
64 	int call_arptables;
65 
66 	/* default value is 0 */
67 	int filter_vlan_tagged;
68 	int filter_pppoe_tagged;
69 	int pass_vlan_indev;
70 };
71 
72 #define IS_IP(skb) \
73 	(!skb_vlan_tag_present(skb) && skb->protocol == htons(ETH_P_IP))
74 
75 #define IS_IPV6(skb) \
76 	(!skb_vlan_tag_present(skb) && skb->protocol == htons(ETH_P_IPV6))
77 
78 #define IS_ARP(skb) \
79 	(!skb_vlan_tag_present(skb) && skb->protocol == htons(ETH_P_ARP))
80 
vlan_proto(const struct sk_buff * skb)81 static inline __be16 vlan_proto(const struct sk_buff *skb)
82 {
83 	if (skb_vlan_tag_present(skb))
84 		return skb->protocol;
85 	else if (skb->protocol == htons(ETH_P_8021Q))
86 		return vlan_eth_hdr(skb)->h_vlan_encapsulated_proto;
87 	else
88 		return 0;
89 }
90 
is_vlan_ip(const struct sk_buff * skb,const struct net * net)91 static inline bool is_vlan_ip(const struct sk_buff *skb, const struct net *net)
92 {
93 	struct brnf_net *brnet = net_generic(net, brnf_net_id);
94 
95 	return vlan_proto(skb) == htons(ETH_P_IP) && brnet->filter_vlan_tagged;
96 }
97 
is_vlan_ipv6(const struct sk_buff * skb,const struct net * net)98 static inline bool is_vlan_ipv6(const struct sk_buff *skb,
99 				const struct net *net)
100 {
101 	struct brnf_net *brnet = net_generic(net, brnf_net_id);
102 
103 	return vlan_proto(skb) == htons(ETH_P_IPV6) &&
104 	       brnet->filter_vlan_tagged;
105 }
106 
is_vlan_arp(const struct sk_buff * skb,const struct net * net)107 static inline bool is_vlan_arp(const struct sk_buff *skb, const struct net *net)
108 {
109 	struct brnf_net *brnet = net_generic(net, brnf_net_id);
110 
111 	return vlan_proto(skb) == htons(ETH_P_ARP) && brnet->filter_vlan_tagged;
112 }
113 
pppoe_proto(const struct sk_buff * skb)114 static inline __be16 pppoe_proto(const struct sk_buff *skb)
115 {
116 	return *((__be16 *)(skb_mac_header(skb) + ETH_HLEN +
117 			    sizeof(struct pppoe_hdr)));
118 }
119 
is_pppoe_ip(const struct sk_buff * skb,const struct net * net)120 static inline bool is_pppoe_ip(const struct sk_buff *skb, const struct net *net)
121 {
122 	struct brnf_net *brnet = net_generic(net, brnf_net_id);
123 
124 	return skb->protocol == htons(ETH_P_PPP_SES) &&
125 	       pppoe_proto(skb) == htons(PPP_IP) && brnet->filter_pppoe_tagged;
126 }
127 
is_pppoe_ipv6(const struct sk_buff * skb,const struct net * net)128 static inline bool is_pppoe_ipv6(const struct sk_buff *skb,
129 				 const struct net *net)
130 {
131 	struct brnf_net *brnet = net_generic(net, brnf_net_id);
132 
133 	return skb->protocol == htons(ETH_P_PPP_SES) &&
134 	       pppoe_proto(skb) == htons(PPP_IPV6) &&
135 	       brnet->filter_pppoe_tagged;
136 }
137 
138 /* largest possible L2 header, see br_nf_dev_queue_xmit() */
139 #define NF_BRIDGE_MAX_MAC_HEADER_LENGTH (PPPOE_SES_HLEN + ETH_HLEN)
140 
141 struct brnf_frag_data {
142 	local_lock_t bh_lock;
143 	char mac[NF_BRIDGE_MAX_MAC_HEADER_LENGTH];
144 	u8 encap_size;
145 	u8 size;
146 	u16 vlan_tci;
147 	__be16 vlan_proto;
148 };
149 
150 static DEFINE_PER_CPU(struct brnf_frag_data, brnf_frag_data_storage) = {
151 	.bh_lock = INIT_LOCAL_LOCK(bh_lock),
152 };
153 
nf_bridge_info_free(struct sk_buff * skb)154 static void nf_bridge_info_free(struct sk_buff *skb)
155 {
156 	skb_ext_del(skb, SKB_EXT_BRIDGE_NF);
157 }
158 
bridge_parent(const struct net_device * dev)159 static inline struct net_device *bridge_parent(const struct net_device *dev)
160 {
161 	struct net_bridge_port *port;
162 
163 	port = br_port_get_rcu(dev);
164 	return port ? port->br->dev : NULL;
165 }
166 
nf_bridge_unshare(struct sk_buff * skb)167 static inline struct nf_bridge_info *nf_bridge_unshare(struct sk_buff *skb)
168 {
169 	return skb_ext_add(skb, SKB_EXT_BRIDGE_NF);
170 }
171 
nf_bridge_encap_header_len(const struct sk_buff * skb)172 unsigned int nf_bridge_encap_header_len(const struct sk_buff *skb)
173 {
174 	switch (skb->protocol) {
175 	case __cpu_to_be16(ETH_P_8021Q):
176 		return VLAN_HLEN;
177 	case __cpu_to_be16(ETH_P_PPP_SES):
178 		return PPPOE_SES_HLEN;
179 	default:
180 		return 0;
181 	}
182 }
183 
nf_bridge_pull_encap_header(struct sk_buff * skb)184 static inline void nf_bridge_pull_encap_header(struct sk_buff *skb)
185 {
186 	unsigned int len = nf_bridge_encap_header_len(skb);
187 
188 	skb_pull(skb, len);
189 	skb->network_header += len;
190 }
191 
nf_bridge_pull_encap_header_rcsum(struct sk_buff * skb)192 static inline void nf_bridge_pull_encap_header_rcsum(struct sk_buff *skb)
193 {
194 	unsigned int len = nf_bridge_encap_header_len(skb);
195 
196 	skb_pull_rcsum(skb, len);
197 	skb->network_header += len;
198 }
199 
200 /* When handing a packet over to the IP layer
201  * check whether we have a skb that is in the
202  * expected format
203  */
204 
br_validate_ipv4(struct net * net,struct sk_buff * skb)205 static int br_validate_ipv4(struct net *net, struct sk_buff *skb)
206 {
207 	const struct iphdr *iph;
208 	u32 len;
209 
210 	if (!pskb_may_pull(skb, sizeof(struct iphdr)))
211 		goto inhdr_error;
212 
213 	iph = ip_hdr(skb);
214 
215 	/* Basic sanity checks */
216 	if (iph->ihl < 5 || iph->version != 4)
217 		goto inhdr_error;
218 
219 	if (!pskb_may_pull(skb, iph->ihl*4))
220 		goto inhdr_error;
221 
222 	iph = ip_hdr(skb);
223 	if (unlikely(ip_fast_csum((u8 *)iph, iph->ihl)))
224 		goto csum_error;
225 
226 	len = skb_ip_totlen(skb);
227 	if (skb->len < len) {
228 		__IP_INC_STATS(net, IPSTATS_MIB_INTRUNCATEDPKTS);
229 		goto drop;
230 	} else if (len < (iph->ihl*4))
231 		goto inhdr_error;
232 
233 	if (pskb_trim_rcsum(skb, len)) {
234 		__IP_INC_STATS(net, IPSTATS_MIB_INDISCARDS);
235 		goto drop;
236 	}
237 
238 	memset(IPCB(skb), 0, sizeof(struct inet_skb_parm));
239 	/* We should really parse IP options here but until
240 	 * somebody who actually uses IP options complains to
241 	 * us we'll just silently ignore the options because
242 	 * we're lazy!
243 	 */
244 	return 0;
245 
246 csum_error:
247 	__IP_INC_STATS(net, IPSTATS_MIB_CSUMERRORS);
248 inhdr_error:
249 	__IP_INC_STATS(net, IPSTATS_MIB_INHDRERRORS);
250 drop:
251 	return -1;
252 }
253 
nf_bridge_update_protocol(struct sk_buff * skb)254 void nf_bridge_update_protocol(struct sk_buff *skb)
255 {
256 	const struct nf_bridge_info *nf_bridge = nf_bridge_info_get(skb);
257 
258 	switch (nf_bridge->orig_proto) {
259 	case BRNF_PROTO_8021Q:
260 		skb->protocol = htons(ETH_P_8021Q);
261 		break;
262 	case BRNF_PROTO_PPPOE:
263 		skb->protocol = htons(ETH_P_PPP_SES);
264 		break;
265 	case BRNF_PROTO_UNCHANGED:
266 		break;
267 	}
268 }
269 
270 /* Obtain the correct destination MAC address, while preserving the original
271  * source MAC address. If we already know this address, we just copy it. If we
272  * don't, we use the neighbour framework to find out. In both cases, we make
273  * sure that br_handle_frame_finish() is called afterwards.
274  */
br_nf_pre_routing_finish_bridge(struct net * net,struct sock * sk,struct sk_buff * skb)275 int br_nf_pre_routing_finish_bridge(struct net *net, struct sock *sk, struct sk_buff *skb)
276 {
277 	struct neighbour *neigh;
278 	struct dst_entry *dst;
279 
280 	skb->dev = bridge_parent(skb->dev);
281 	if (!skb->dev)
282 		goto free_skb;
283 	dst = skb_dst(skb);
284 	neigh = dst_neigh_lookup_skb(dst, skb);
285 	if (neigh) {
286 		struct nf_bridge_info *nf_bridge = nf_bridge_info_get(skb);
287 		int ret;
288 
289 		if ((READ_ONCE(neigh->nud_state) & NUD_CONNECTED) &&
290 		    READ_ONCE(neigh->hh.hh_len)) {
291 			struct net_device *br_indev;
292 
293 			br_indev = nf_bridge_get_physindev(skb, net);
294 			if (!br_indev) {
295 				neigh_release(neigh);
296 				goto free_skb;
297 			}
298 
299 			neigh_hh_bridge(&neigh->hh, skb);
300 			skb->dev = br_indev;
301 
302 			ret = br_handle_frame_finish(net, sk, skb);
303 		} else {
304 			/* the neighbour function below overwrites the complete
305 			 * MAC header, so we save the Ethernet source address and
306 			 * protocol number.
307 			 */
308 			skb_copy_from_linear_data_offset(skb,
309 							 -(ETH_HLEN-ETH_ALEN),
310 							 nf_bridge->neigh_header,
311 							 ETH_HLEN-ETH_ALEN);
312 			/* tell br_dev_xmit to continue with forwarding */
313 			nf_bridge->bridged_dnat = 1;
314 			/* FIXME Need to refragment */
315 			ret = READ_ONCE(neigh->output)(neigh, skb);
316 		}
317 		neigh_release(neigh);
318 		return ret;
319 	}
320 free_skb:
321 	kfree_skb(skb);
322 	return 0;
323 }
324 
325 static inline bool
br_nf_ipv4_daddr_was_changed(const struct sk_buff * skb,const struct nf_bridge_info * nf_bridge)326 br_nf_ipv4_daddr_was_changed(const struct sk_buff *skb,
327 			     const struct nf_bridge_info *nf_bridge)
328 {
329 	return ip_hdr(skb)->daddr != nf_bridge->ipv4_daddr;
330 }
331 
332 /* This requires some explaining. If DNAT has taken place,
333  * we will need to fix up the destination Ethernet address.
334  * This is also true when SNAT takes place (for the reply direction).
335  *
336  * There are two cases to consider:
337  * 1. The packet was DNAT'ed to a device in the same bridge
338  *    port group as it was received on. We can still bridge
339  *    the packet.
340  * 2. The packet was DNAT'ed to a different device, either
341  *    a non-bridged device or another bridge port group.
342  *    The packet will need to be routed.
343  *
344  * The correct way of distinguishing between these two cases is to
345  * call ip_route_input() and to look at skb->dst->dev, which is
346  * changed to the destination device if ip_route_input() succeeds.
347  *
348  * Let's first consider the case that ip_route_input() succeeds:
349  *
350  * If the output device equals the logical bridge device the packet
351  * came in on, we can consider this bridging. The corresponding MAC
352  * address will be obtained in br_nf_pre_routing_finish_bridge.
353  * Otherwise, the packet is considered to be routed and we just
354  * change the destination MAC address so that the packet will
355  * later be passed up to the IP stack to be routed. For a redirected
356  * packet, ip_route_input() will give back the localhost as output device,
357  * which differs from the bridge device.
358  *
359  * Let's now consider the case that ip_route_input() fails:
360  *
361  * This can be because the destination address is martian, in which case
362  * the packet will be dropped.
363  * If IP forwarding is disabled, ip_route_input() will fail, while
364  * ip_route_output_key() can return success. The source
365  * address for ip_route_output_key() is set to zero, so ip_route_output_key()
366  * thinks we're handling a locally generated packet and won't care
367  * if IP forwarding is enabled. If the output device equals the logical bridge
368  * device, we proceed as if ip_route_input() succeeded. If it differs from the
369  * logical bridge port or if ip_route_output_key() fails we drop the packet.
370  */
br_nf_pre_routing_finish(struct net * net,struct sock * sk,struct sk_buff * skb)371 static int br_nf_pre_routing_finish(struct net *net, struct sock *sk, struct sk_buff *skb)
372 {
373 	struct nf_bridge_info *nf_bridge = nf_bridge_info_get(skb);
374 	struct net_device *dev = skb->dev, *br_indev;
375 	const struct iphdr *iph = ip_hdr(skb);
376 	enum skb_drop_reason reason;
377 	struct rtable *rt;
378 
379 	br_indev = nf_bridge_get_physindev(skb, net);
380 	if (!br_indev) {
381 		kfree_skb(skb);
382 		return 0;
383 	}
384 
385 	nf_bridge->frag_max_size = IPCB(skb)->frag_max_size;
386 
387 	if (nf_bridge->pkt_otherhost) {
388 		skb->pkt_type = PACKET_OTHERHOST;
389 		nf_bridge->pkt_otherhost = false;
390 	}
391 	nf_bridge->in_prerouting = 0;
392 	if (br_nf_ipv4_daddr_was_changed(skb, nf_bridge)) {
393 		reason = ip_route_input(skb, iph->daddr, iph->saddr,
394 					ip4h_dscp(iph), dev);
395 		if (reason) {
396 			struct in_device *in_dev = __in_dev_get_rcu(dev);
397 
398 			/* If err equals -EHOSTUNREACH the error is due to a
399 			 * martian destination or due to the fact that
400 			 * forwarding is disabled. For most martian packets,
401 			 * ip_route_output_key() will fail. It won't fail for 2 types of
402 			 * martian destinations: loopback destinations and destination
403 			 * 0.0.0.0. In both cases the packet will be dropped because the
404 			 * destination is the loopback device and not the bridge. */
405 			if (reason != SKB_DROP_REASON_IP_INADDRERRORS || !in_dev ||
406 			    IN_DEV_FORWARD(in_dev))
407 				goto free_skb;
408 
409 			rt = ip_route_output(net, iph->daddr, 0,
410 					     ip4h_dscp(iph), 0,
411 					     RT_SCOPE_UNIVERSE);
412 			if (!IS_ERR(rt)) {
413 				/* - Bridged-and-DNAT'ed traffic doesn't
414 				 *   require ip_forwarding. */
415 				if (rt->dst.dev == dev) {
416 					skb_dst_drop(skb);
417 					skb_dst_set(skb, &rt->dst);
418 					goto bridged_dnat;
419 				}
420 				ip_rt_put(rt);
421 			}
422 free_skb:
423 			kfree_skb(skb);
424 			return 0;
425 		} else {
426 			if (skb_dst(skb)->dev == dev) {
427 bridged_dnat:
428 				skb->dev = br_indev;
429 				nf_bridge_update_protocol(skb);
430 				nf_bridge_push_encap_header(skb);
431 				br_nf_hook_thresh(NF_BR_PRE_ROUTING,
432 						  net, sk, skb, skb->dev,
433 						  NULL,
434 						  br_nf_pre_routing_finish_bridge);
435 				return 0;
436 			}
437 			ether_addr_copy(eth_hdr(skb)->h_dest, dev->dev_addr);
438 			skb->pkt_type = PACKET_HOST;
439 		}
440 	} else {
441 		rt = bridge_parent_rtable(br_indev);
442 		if (!rt) {
443 			kfree_skb(skb);
444 			return 0;
445 		}
446 		skb_dst_drop(skb);
447 		skb_dst_set_noref(skb, &rt->dst);
448 	}
449 
450 	skb->dev = br_indev;
451 	nf_bridge_update_protocol(skb);
452 	nf_bridge_push_encap_header(skb);
453 	br_nf_hook_thresh(NF_BR_PRE_ROUTING, net, sk, skb, skb->dev, NULL,
454 			  br_handle_frame_finish);
455 	return 0;
456 }
457 
brnf_get_logical_dev(struct sk_buff * skb,const struct net_device * dev,const struct net * net)458 static struct net_device *brnf_get_logical_dev(struct sk_buff *skb,
459 					       const struct net_device *dev,
460 					       const struct net *net)
461 {
462 	struct net_device *vlan, *br;
463 	struct brnf_net *brnet = net_generic(net, brnf_net_id);
464 
465 	br = bridge_parent(dev);
466 
467 	if (brnet->pass_vlan_indev == 0 || !skb_vlan_tag_present(skb))
468 		return br;
469 
470 	vlan = __vlan_find_dev_deep_rcu(br, skb->vlan_proto,
471 				    skb_vlan_tag_get(skb) & VLAN_VID_MASK);
472 
473 	return vlan ? vlan : br;
474 }
475 
476 /* Some common code for IPv4/IPv6 */
setup_pre_routing(struct sk_buff * skb,const struct net * net)477 struct net_device *setup_pre_routing(struct sk_buff *skb, const struct net *net)
478 {
479 	struct nf_bridge_info *nf_bridge = nf_bridge_info_get(skb);
480 
481 	if (skb->pkt_type == PACKET_OTHERHOST) {
482 		skb->pkt_type = PACKET_HOST;
483 		nf_bridge->pkt_otherhost = true;
484 	}
485 
486 	nf_bridge->in_prerouting = 1;
487 	nf_bridge->physinif = skb->dev->ifindex;
488 	skb->dev = brnf_get_logical_dev(skb, skb->dev, net);
489 
490 	if (skb->protocol == htons(ETH_P_8021Q))
491 		nf_bridge->orig_proto = BRNF_PROTO_8021Q;
492 	else if (skb->protocol == htons(ETH_P_PPP_SES))
493 		nf_bridge->orig_proto = BRNF_PROTO_PPPOE;
494 
495 	/* Must drop socket now because of tproxy. */
496 	skb_orphan(skb);
497 	return skb->dev;
498 }
499 
500 /* Direct IPv6 traffic to br_nf_pre_routing_ipv6.
501  * Replicate the checks that IPv4 does on packet reception.
502  * Set skb->dev to the bridge device (i.e. parent of the
503  * receiving device) to make netfilter happy, the REDIRECT
504  * target in particular.  Save the original destination IP
505  * address to be able to detect DNAT afterwards. */
br_nf_pre_routing(void * priv,struct sk_buff * skb,const struct nf_hook_state * state)506 static unsigned int br_nf_pre_routing(void *priv,
507 				      struct sk_buff *skb,
508 				      const struct nf_hook_state *state)
509 {
510 	struct nf_bridge_info *nf_bridge;
511 	struct net_bridge_port *p;
512 	struct net_bridge *br;
513 	__u32 len = nf_bridge_encap_header_len(skb);
514 	struct brnf_net *brnet;
515 
516 	if (unlikely(!pskb_may_pull(skb, len)))
517 		return NF_DROP_REASON(skb, SKB_DROP_REASON_PKT_TOO_SMALL, 0);
518 
519 	p = br_port_get_rcu(state->in);
520 	if (p == NULL)
521 		return NF_DROP_REASON(skb, SKB_DROP_REASON_DEV_READY, 0);
522 	br = p->br;
523 
524 	brnet = net_generic(state->net, brnf_net_id);
525 	if (IS_IPV6(skb) || is_vlan_ipv6(skb, state->net) ||
526 	    is_pppoe_ipv6(skb, state->net)) {
527 		if (!brnet->call_ip6tables &&
528 		    !br_opt_get(br, BROPT_NF_CALL_IP6TABLES))
529 			return NF_ACCEPT;
530 		if (!ipv6_mod_enabled()) {
531 			pr_warn_once("Module ipv6 is disabled, so call_ip6tables is not supported.");
532 			return NF_DROP_REASON(skb, SKB_DROP_REASON_IPV6DISABLED, 0);
533 		}
534 
535 		nf_bridge_pull_encap_header_rcsum(skb);
536 		return br_nf_pre_routing_ipv6(priv, skb, state);
537 	}
538 
539 	if (!brnet->call_iptables && !br_opt_get(br, BROPT_NF_CALL_IPTABLES))
540 		return NF_ACCEPT;
541 
542 	if (!IS_IP(skb) && !is_vlan_ip(skb, state->net) &&
543 	    !is_pppoe_ip(skb, state->net))
544 		return NF_ACCEPT;
545 
546 	nf_bridge_pull_encap_header_rcsum(skb);
547 
548 	if (br_validate_ipv4(state->net, skb))
549 		return NF_DROP_REASON(skb, SKB_DROP_REASON_IP_INHDR, 0);
550 
551 	if (!nf_bridge_alloc(skb))
552 		return NF_DROP_REASON(skb, SKB_DROP_REASON_NOMEM, 0);
553 	if (!setup_pre_routing(skb, state->net))
554 		return NF_DROP_REASON(skb, SKB_DROP_REASON_DEV_READY, 0);
555 
556 	nf_bridge = nf_bridge_info_get(skb);
557 	nf_bridge->ipv4_daddr = ip_hdr(skb)->daddr;
558 
559 	skb->protocol = htons(ETH_P_IP);
560 	skb->transport_header = skb->network_header + ip_hdr(skb)->ihl * 4;
561 
562 	NF_HOOK(NFPROTO_IPV4, NF_INET_PRE_ROUTING, state->net, state->sk, skb,
563 		skb->dev, NULL,
564 		br_nf_pre_routing_finish);
565 
566 	return NF_STOLEN;
567 }
568 
569 #if IS_ENABLED(CONFIG_NF_CONNTRACK)
570 /* conntracks' nf_confirm logic cannot handle cloned skbs referencing
571  * the same nf_conn entry, which will happen for multicast (broadcast)
572  * Frames on bridges.
573  *
574  * Example:
575  *      macvlan0
576  *      br0
577  *  ethX  ethY
578  *
579  * ethX (or Y) receives multicast or broadcast packet containing
580  * an IP packet, not yet in conntrack table.
581  *
582  * 1. skb passes through bridge and fake-ip (br_netfilter)Prerouting.
583  *    -> skb->_nfct now references a unconfirmed entry
584  * 2. skb is broad/mcast packet. bridge now passes clones out on each bridge
585  *    interface.
586  * 3. skb gets passed up the stack.
587  * 4. In macvlan case, macvlan driver retains clone(s) of the mcast skb
588  *    and schedules a work queue to send them out on the lower devices.
589  *
590  *    The clone skb->_nfct is not a copy, it is the same entry as the
591  *    original skb.  The macvlan rx handler then returns RX_HANDLER_PASS.
592  * 5. Normal conntrack hooks (in NF_INET_LOCAL_IN) confirm the orig skb.
593  *
594  * The Macvlan broadcast worker and normal confirm path will race.
595  *
596  * This race will not happen if step 2 already confirmed a clone. In that
597  * case later steps perform skb_clone() with skb->_nfct already confirmed (in
598  * hash table).  This works fine.
599  *
600  * But such confirmation won't happen when eb/ip/nftables rules dropped the
601  * packets before they reached the nf_confirm step in postrouting.
602  *
603  * Work around this problem by explicit confirmation of the entry at
604  * LOCAL_IN time, before upper layer has a chance to clone the unconfirmed
605  * entry.
606  *
607  */
br_nf_local_in(void * priv,struct sk_buff * skb,const struct nf_hook_state * state)608 static unsigned int br_nf_local_in(void *priv,
609 				   struct sk_buff *skb,
610 				   const struct nf_hook_state *state)
611 {
612 	bool promisc = BR_INPUT_SKB_CB(skb)->promisc;
613 	struct nf_conntrack *nfct = skb_nfct(skb);
614 	const struct nf_ct_hook *ct_hook;
615 	struct nf_conn *ct;
616 	int ret;
617 
618 	if (promisc) {
619 		nf_reset_ct(skb);
620 		return NF_ACCEPT;
621 	}
622 
623 	if (!nfct || skb->pkt_type == PACKET_HOST)
624 		return NF_ACCEPT;
625 
626 	ct = container_of(nfct, struct nf_conn, ct_general);
627 	if (likely(nf_ct_is_confirmed(ct)))
628 		return NF_ACCEPT;
629 
630 	if (WARN_ON_ONCE(refcount_read(&nfct->use) != 1)) {
631 		nf_reset_ct(skb);
632 		return NF_ACCEPT;
633 	}
634 
635 	WARN_ON_ONCE(skb_shared(skb));
636 
637 	/* We can't call nf_confirm here, it would create a dependency
638 	 * on nf_conntrack module.
639 	 */
640 	ct_hook = rcu_dereference(nf_ct_hook);
641 	if (!ct_hook) {
642 		skb->_nfct = 0ul;
643 		nf_conntrack_put(nfct);
644 		return NF_ACCEPT;
645 	}
646 
647 	nf_bridge_pull_encap_header(skb);
648 	ret = ct_hook->confirm(skb);
649 	switch (ret & NF_VERDICT_MASK) {
650 	case NF_STOLEN:
651 		return NF_STOLEN;
652 	default:
653 		nf_bridge_push_encap_header(skb);
654 		break;
655 	}
656 
657 	ct = container_of(nfct, struct nf_conn, ct_general);
658 	WARN_ON_ONCE(!nf_ct_is_confirmed(ct));
659 
660 	return ret;
661 }
662 #endif
663 
664 /* PF_BRIDGE/FORWARD *************************************************/
br_nf_forward_finish(struct net * net,struct sock * sk,struct sk_buff * skb)665 static int br_nf_forward_finish(struct net *net, struct sock *sk, struct sk_buff *skb)
666 {
667 	struct nf_bridge_info *nf_bridge = nf_bridge_info_get(skb);
668 	struct net_device *in;
669 
670 	if (!IS_ARP(skb) && !is_vlan_arp(skb, net)) {
671 
672 		if (skb->protocol == htons(ETH_P_IP))
673 			nf_bridge->frag_max_size = IPCB(skb)->frag_max_size;
674 
675 		if (skb->protocol == htons(ETH_P_IPV6))
676 			nf_bridge->frag_max_size = IP6CB(skb)->frag_max_size;
677 
678 		in = nf_bridge_get_physindev(skb, net);
679 		if (!in) {
680 			kfree_skb(skb);
681 			return 0;
682 		}
683 		if (nf_bridge->pkt_otherhost) {
684 			skb->pkt_type = PACKET_OTHERHOST;
685 			nf_bridge->pkt_otherhost = false;
686 		}
687 		nf_bridge_update_protocol(skb);
688 	} else {
689 		in = *((struct net_device **)(skb->cb));
690 	}
691 	nf_bridge_push_encap_header(skb);
692 
693 	br_nf_hook_thresh(NF_BR_FORWARD, net, sk, skb, in, skb->dev,
694 			  br_forward_finish);
695 	return 0;
696 }
697 
698 
br_nf_forward_ip(struct sk_buff * skb,const struct nf_hook_state * state,u8 pf)699 static unsigned int br_nf_forward_ip(struct sk_buff *skb,
700 				     const struct nf_hook_state *state,
701 				     u8 pf)
702 {
703 	struct nf_bridge_info *nf_bridge;
704 	struct net_device *parent;
705 
706 	nf_bridge = nf_bridge_info_get(skb);
707 	if (!nf_bridge)
708 		return NF_ACCEPT;
709 
710 	/* Need exclusive nf_bridge_info since we might have multiple
711 	 * different physoutdevs. */
712 	if (!nf_bridge_unshare(skb))
713 		return NF_DROP_REASON(skb, SKB_DROP_REASON_NOMEM, 0);
714 
715 	nf_bridge = nf_bridge_info_get(skb);
716 	if (!nf_bridge)
717 		return NF_DROP_REASON(skb, SKB_DROP_REASON_NOMEM, 0);
718 
719 	parent = bridge_parent(state->out);
720 	if (!parent)
721 		return NF_DROP_REASON(skb, SKB_DROP_REASON_DEV_READY, 0);
722 
723 	nf_bridge_pull_encap_header(skb);
724 
725 	if (skb->pkt_type == PACKET_OTHERHOST) {
726 		skb->pkt_type = PACKET_HOST;
727 		nf_bridge->pkt_otherhost = true;
728 	}
729 
730 	if (pf == NFPROTO_IPV4) {
731 		if (br_validate_ipv4(state->net, skb))
732 			return NF_DROP_REASON(skb, SKB_DROP_REASON_IP_INHDR, 0);
733 		IPCB(skb)->frag_max_size = nf_bridge->frag_max_size;
734 		skb->protocol = htons(ETH_P_IP);
735 	} else if (pf == NFPROTO_IPV6) {
736 		if (br_validate_ipv6(state->net, skb))
737 			return NF_DROP_REASON(skb, SKB_DROP_REASON_IP_INHDR, 0);
738 		IP6CB(skb)->frag_max_size = nf_bridge->frag_max_size;
739 		skb->protocol = htons(ETH_P_IPV6);
740 	} else {
741 		WARN_ON_ONCE(1);
742 		return NF_DROP;
743 	}
744 
745 	nf_bridge->physoutdev = skb->dev;
746 
747 	NF_HOOK(pf, NF_INET_FORWARD, state->net, NULL, skb,
748 		brnf_get_logical_dev(skb, state->in, state->net),
749 		parent,	br_nf_forward_finish);
750 
751 	return NF_STOLEN;
752 }
753 
br_nf_forward_arp(struct sk_buff * skb,const struct nf_hook_state * state)754 static unsigned int br_nf_forward_arp(struct sk_buff *skb,
755 				      const struct nf_hook_state *state)
756 {
757 	struct net_bridge_port *p;
758 	struct net_bridge *br;
759 	struct net_device **d = (struct net_device **)(skb->cb);
760 	struct brnf_net *brnet;
761 
762 	p = br_port_get_rcu(state->out);
763 	if (p == NULL)
764 		return NF_ACCEPT;
765 	br = p->br;
766 
767 	brnet = net_generic(state->net, brnf_net_id);
768 	if (!brnet->call_arptables && !br_opt_get(br, BROPT_NF_CALL_ARPTABLES))
769 		return NF_ACCEPT;
770 
771 	if (is_vlan_arp(skb, state->net))
772 		nf_bridge_pull_encap_header(skb);
773 
774 	if (unlikely(!pskb_may_pull(skb, sizeof(struct arphdr))))
775 		return NF_DROP_REASON(skb, SKB_DROP_REASON_PKT_TOO_SMALL, 0);
776 
777 	if (arp_hdr(skb)->ar_pln != 4) {
778 		if (is_vlan_arp(skb, state->net))
779 			nf_bridge_push_encap_header(skb);
780 		return NF_ACCEPT;
781 	}
782 	*d = state->in;
783 	NF_HOOK(NFPROTO_ARP, NF_ARP_FORWARD, state->net, state->sk, skb,
784 		state->in, state->out, br_nf_forward_finish);
785 
786 	return NF_STOLEN;
787 }
788 
789 /* This is the 'purely bridged' case.  For IP, we pass the packet to
790  * netfilter with indev and outdev set to the bridge device,
791  * but we are still able to filter on the 'real' indev/outdev
792  * because of the physdev module. For ARP, indev and outdev are the
793  * bridge ports.
794  */
br_nf_forward(void * priv,struct sk_buff * skb,const struct nf_hook_state * state)795 static unsigned int br_nf_forward(void *priv,
796 				  struct sk_buff *skb,
797 				  const struct nf_hook_state *state)
798 {
799 	if (IS_IP(skb) || is_vlan_ip(skb, state->net) ||
800 	    is_pppoe_ip(skb, state->net))
801 		return br_nf_forward_ip(skb, state, NFPROTO_IPV4);
802 	if (IS_IPV6(skb) || is_vlan_ipv6(skb, state->net) ||
803 	    is_pppoe_ipv6(skb, state->net))
804 		return br_nf_forward_ip(skb, state, NFPROTO_IPV6);
805 	if (IS_ARP(skb) || is_vlan_arp(skb, state->net))
806 		return br_nf_forward_arp(skb, state);
807 
808 	return NF_ACCEPT;
809 }
810 
br_nf_push_frag_xmit(struct net * net,struct sock * sk,struct sk_buff * skb)811 static int br_nf_push_frag_xmit(struct net *net, struct sock *sk, struct sk_buff *skb)
812 {
813 	struct brnf_frag_data *data;
814 	int err;
815 
816 	data = this_cpu_ptr(&brnf_frag_data_storage);
817 	err = skb_cow_head(skb, data->size);
818 
819 	if (err) {
820 		kfree_skb(skb);
821 		return 0;
822 	}
823 
824 	if (data->vlan_proto)
825 		__vlan_hwaccel_put_tag(skb, data->vlan_proto, data->vlan_tci);
826 
827 	skb_copy_to_linear_data_offset(skb, -data->size, data->mac, data->size);
828 	__skb_push(skb, data->encap_size);
829 
830 	nf_bridge_info_free(skb);
831 	return br_dev_queue_push_xmit(net, sk, skb);
832 }
833 
834 static int
br_nf_ip_fragment(struct net * net,struct sock * sk,struct sk_buff * skb,int (* output)(struct net *,struct sock *,struct sk_buff *))835 br_nf_ip_fragment(struct net *net, struct sock *sk, struct sk_buff *skb,
836 		  int (*output)(struct net *, struct sock *, struct sk_buff *))
837 {
838 	unsigned int mtu = ip_skb_dst_mtu(sk, skb);
839 	struct iphdr *iph = ip_hdr(skb);
840 
841 	if (unlikely(((iph->frag_off & htons(IP_DF)) && !skb->ignore_df) ||
842 		     (IPCB(skb)->frag_max_size &&
843 		      IPCB(skb)->frag_max_size > mtu))) {
844 		IP_INC_STATS(net, IPSTATS_MIB_FRAGFAILS);
845 		kfree_skb(skb);
846 		return -EMSGSIZE;
847 	}
848 
849 	return ip_do_fragment(net, sk, skb, output);
850 }
851 
nf_bridge_mtu_reduction(const struct sk_buff * skb)852 static unsigned int nf_bridge_mtu_reduction(const struct sk_buff *skb)
853 {
854 	const struct nf_bridge_info *nf_bridge = nf_bridge_info_get(skb);
855 
856 	if (nf_bridge->orig_proto == BRNF_PROTO_PPPOE)
857 		return PPPOE_SES_HLEN;
858 	return 0;
859 }
860 
br_nf_dev_queue_xmit(struct net * net,struct sock * sk,struct sk_buff * skb)861 static int br_nf_dev_queue_xmit(struct net *net, struct sock *sk, struct sk_buff *skb)
862 {
863 	struct nf_bridge_info *nf_bridge = nf_bridge_info_get(skb);
864 	unsigned int mtu, mtu_reserved;
865 	int ret;
866 
867 	mtu_reserved = nf_bridge_mtu_reduction(skb);
868 	mtu = skb->dev->mtu;
869 
870 	if (nf_bridge->pkt_otherhost) {
871 		skb->pkt_type = PACKET_OTHERHOST;
872 		nf_bridge->pkt_otherhost = false;
873 	}
874 
875 	if (nf_bridge->frag_max_size && nf_bridge->frag_max_size < mtu)
876 		mtu = nf_bridge->frag_max_size;
877 
878 	nf_bridge_update_protocol(skb);
879 	nf_bridge_push_encap_header(skb);
880 
881 	if (skb_is_gso(skb) || skb->len + mtu_reserved <= mtu) {
882 		nf_bridge_info_free(skb);
883 		return br_dev_queue_push_xmit(net, sk, skb);
884 	}
885 
886 	/* Fragmentation on metadata/template dst is not supported */
887 	if (unlikely(!skb_valid_dst(skb)))
888 		goto drop;
889 
890 	/* This is wrong! We should preserve the original fragment
891 	 * boundaries by preserving frag_list rather than refragmenting.
892 	 */
893 	if (IS_ENABLED(CONFIG_NF_DEFRAG_IPV4) &&
894 	    skb->protocol == htons(ETH_P_IP)) {
895 		struct brnf_frag_data *data;
896 
897 		if (br_validate_ipv4(net, skb))
898 			goto drop;
899 
900 		IPCB(skb)->frag_max_size = nf_bridge->frag_max_size;
901 
902 		local_lock_nested_bh(&brnf_frag_data_storage.bh_lock);
903 		data = this_cpu_ptr(&brnf_frag_data_storage);
904 
905 		if (skb_vlan_tag_present(skb)) {
906 			data->vlan_tci = skb->vlan_tci;
907 			data->vlan_proto = skb->vlan_proto;
908 		} else {
909 			data->vlan_proto = 0;
910 		}
911 
912 		data->encap_size = nf_bridge_encap_header_len(skb);
913 		data->size = ETH_HLEN + data->encap_size;
914 
915 		skb_copy_from_linear_data_offset(skb, -data->size, data->mac,
916 						 data->size);
917 
918 		ret = br_nf_ip_fragment(net, sk, skb, br_nf_push_frag_xmit);
919 		local_unlock_nested_bh(&brnf_frag_data_storage.bh_lock);
920 		return ret;
921 	}
922 	if (IS_ENABLED(CONFIG_NF_DEFRAG_IPV6) &&
923 	    skb->protocol == htons(ETH_P_IPV6)) {
924 		const struct nf_ipv6_ops *v6ops = nf_get_ipv6_ops();
925 		struct brnf_frag_data *data;
926 
927 		if (br_validate_ipv6(net, skb))
928 			goto drop;
929 
930 		IP6CB(skb)->frag_max_size = nf_bridge->frag_max_size;
931 
932 		local_lock_nested_bh(&brnf_frag_data_storage.bh_lock);
933 		data = this_cpu_ptr(&brnf_frag_data_storage);
934 		data->encap_size = nf_bridge_encap_header_len(skb);
935 		data->size = ETH_HLEN + data->encap_size;
936 
937 		skb_copy_from_linear_data_offset(skb, -data->size, data->mac,
938 						 data->size);
939 
940 		if (v6ops) {
941 			ret = v6ops->fragment(net, sk, skb, br_nf_push_frag_xmit);
942 			local_unlock_nested_bh(&brnf_frag_data_storage.bh_lock);
943 			return ret;
944 		}
945 		local_unlock_nested_bh(&brnf_frag_data_storage.bh_lock);
946 
947 		kfree_skb(skb);
948 		return -EMSGSIZE;
949 	}
950 	nf_bridge_info_free(skb);
951 	return br_dev_queue_push_xmit(net, sk, skb);
952  drop:
953 	kfree_skb(skb);
954 	return 0;
955 }
956 
957 /* PF_BRIDGE/POST_ROUTING ********************************************/
br_nf_post_routing(void * priv,struct sk_buff * skb,const struct nf_hook_state * state)958 static unsigned int br_nf_post_routing(void *priv,
959 				       struct sk_buff *skb,
960 				       const struct nf_hook_state *state)
961 {
962 	struct nf_bridge_info *nf_bridge = nf_bridge_info_get(skb);
963 	struct net_device *realoutdev = bridge_parent(skb->dev);
964 	u_int8_t pf;
965 
966 	/* if nf_bridge is set, but ->physoutdev is NULL, this packet came in
967 	 * on a bridge, but was delivered locally and is now being routed:
968 	 *
969 	 * POST_ROUTING was already invoked from the ip stack.
970 	 */
971 	if (!nf_bridge || !nf_bridge->physoutdev)
972 		return NF_ACCEPT;
973 
974 	if (!realoutdev)
975 		return NF_DROP_REASON(skb, SKB_DROP_REASON_DEV_READY, 0);
976 
977 	if (IS_IP(skb) || is_vlan_ip(skb, state->net) ||
978 	    is_pppoe_ip(skb, state->net))
979 		pf = NFPROTO_IPV4;
980 	else if (IS_IPV6(skb) || is_vlan_ipv6(skb, state->net) ||
981 		 is_pppoe_ipv6(skb, state->net))
982 		pf = NFPROTO_IPV6;
983 	else
984 		return NF_ACCEPT;
985 
986 	if (skb->pkt_type == PACKET_OTHERHOST) {
987 		skb->pkt_type = PACKET_HOST;
988 		nf_bridge->pkt_otherhost = true;
989 	}
990 
991 	nf_bridge_pull_encap_header(skb);
992 	if (pf == NFPROTO_IPV4)
993 		skb->protocol = htons(ETH_P_IP);
994 	else
995 		skb->protocol = htons(ETH_P_IPV6);
996 
997 	NF_HOOK(pf, NF_INET_POST_ROUTING, state->net, state->sk, skb,
998 		NULL, realoutdev,
999 		br_nf_dev_queue_xmit);
1000 
1001 	return NF_STOLEN;
1002 }
1003 
1004 /* IP/SABOTAGE *****************************************************/
1005 /* Don't hand locally destined packets to PF_INET(6)/PRE_ROUTING
1006  * for the second time. */
ip_sabotage_in(void * priv,struct sk_buff * skb,const struct nf_hook_state * state)1007 static unsigned int ip_sabotage_in(void *priv,
1008 				   struct sk_buff *skb,
1009 				   const struct nf_hook_state *state)
1010 {
1011 	struct nf_bridge_info *nf_bridge = nf_bridge_info_get(skb);
1012 
1013 	if (nf_bridge) {
1014 		if (nf_bridge->sabotage_in_done)
1015 			return NF_ACCEPT;
1016 
1017 		if (!nf_bridge->in_prerouting &&
1018 		    !netif_is_l3_master(skb->dev) &&
1019 		    !netif_is_l3_slave(skb->dev)) {
1020 			nf_bridge->sabotage_in_done = 1;
1021 			state->okfn(state->net, state->sk, skb);
1022 			return NF_STOLEN;
1023 		}
1024 	}
1025 
1026 	return NF_ACCEPT;
1027 }
1028 
1029 /* This is called when br_netfilter has called into iptables/netfilter,
1030  * and DNAT has taken place on a bridge-forwarded packet.
1031  *
1032  * neigh->output has created a new MAC header, with local br0 MAC
1033  * as saddr.
1034  *
1035  * This restores the original MAC saddr of the bridged packet
1036  * before invoking bridge forward logic to transmit the packet.
1037  */
br_nf_pre_routing_finish_bridge_slow(struct sk_buff * skb)1038 static void br_nf_pre_routing_finish_bridge_slow(struct sk_buff *skb)
1039 {
1040 	struct nf_bridge_info *nf_bridge = nf_bridge_info_get(skb);
1041 	struct net_device *br_indev;
1042 
1043 	br_indev = nf_bridge_get_physindev(skb, dev_net(skb->dev));
1044 	if (!br_indev) {
1045 		kfree_skb(skb);
1046 		return;
1047 	}
1048 
1049 	skb_pull(skb, ETH_HLEN);
1050 	nf_bridge->bridged_dnat = 0;
1051 
1052 	BUILD_BUG_ON(sizeof(nf_bridge->neigh_header) != (ETH_HLEN - ETH_ALEN));
1053 
1054 	skb_copy_to_linear_data_offset(skb, -(ETH_HLEN - ETH_ALEN),
1055 				       nf_bridge->neigh_header,
1056 				       ETH_HLEN - ETH_ALEN);
1057 	skb->dev = br_indev;
1058 
1059 	nf_bridge->physoutdev = NULL;
1060 	br_handle_frame_finish(dev_net(skb->dev), NULL, skb);
1061 }
1062 
br_nf_dev_xmit(struct sk_buff * skb)1063 static int br_nf_dev_xmit(struct sk_buff *skb)
1064 {
1065 	const struct nf_bridge_info *nf_bridge = nf_bridge_info_get(skb);
1066 
1067 	if (nf_bridge && nf_bridge->bridged_dnat) {
1068 		br_nf_pre_routing_finish_bridge_slow(skb);
1069 		return 1;
1070 	}
1071 	return 0;
1072 }
1073 
1074 static const struct nf_br_ops br_ops = {
1075 	.br_dev_xmit_hook =	br_nf_dev_xmit,
1076 };
1077 
1078 /* For br_nf_post_routing, we need (prio = NF_BR_PRI_LAST), because
1079  * br_dev_queue_push_xmit is called afterwards */
1080 static const struct nf_hook_ops br_nf_ops[] = {
1081 	{
1082 		.hook = br_nf_pre_routing,
1083 		.pf = NFPROTO_BRIDGE,
1084 		.hooknum = NF_BR_PRE_ROUTING,
1085 		.priority = NF_BR_PRI_BRNF,
1086 	},
1087 #if IS_ENABLED(CONFIG_NF_CONNTRACK)
1088 	{
1089 		.hook = br_nf_local_in,
1090 		.pf = NFPROTO_BRIDGE,
1091 		.hooknum = NF_BR_LOCAL_IN,
1092 		.priority = NF_BR_PRI_LAST,
1093 	},
1094 #endif
1095 	{
1096 		.hook = br_nf_forward,
1097 		.pf = NFPROTO_BRIDGE,
1098 		.hooknum = NF_BR_FORWARD,
1099 		.priority = NF_BR_PRI_BRNF,
1100 	},
1101 	{
1102 		.hook = br_nf_post_routing,
1103 		.pf = NFPROTO_BRIDGE,
1104 		.hooknum = NF_BR_POST_ROUTING,
1105 		.priority = NF_BR_PRI_LAST,
1106 	},
1107 	{
1108 		.hook = ip_sabotage_in,
1109 		.pf = NFPROTO_IPV4,
1110 		.hooknum = NF_INET_PRE_ROUTING,
1111 		.priority = NF_IP_PRI_FIRST,
1112 	},
1113 	{
1114 		.hook = ip_sabotage_in,
1115 		.pf = NFPROTO_IPV6,
1116 		.hooknum = NF_INET_PRE_ROUTING,
1117 		.priority = NF_IP6_PRI_FIRST,
1118 	},
1119 };
1120 
brnf_device_event(struct notifier_block * unused,unsigned long event,void * ptr)1121 static int brnf_device_event(struct notifier_block *unused, unsigned long event,
1122 			     void *ptr)
1123 {
1124 	struct net_device *dev = netdev_notifier_info_to_dev(ptr);
1125 	struct brnf_net *brnet;
1126 	struct net *net;
1127 	int ret;
1128 
1129 	if (event != NETDEV_REGISTER || !netif_is_bridge_master(dev))
1130 		return NOTIFY_DONE;
1131 
1132 	ASSERT_RTNL();
1133 
1134 	net = dev_net(dev);
1135 	brnet = net_generic(net, brnf_net_id);
1136 	if (brnet->enabled)
1137 		return NOTIFY_OK;
1138 
1139 	ret = nf_register_net_hooks(net, br_nf_ops, ARRAY_SIZE(br_nf_ops));
1140 	if (ret)
1141 		return NOTIFY_BAD;
1142 
1143 	brnet->enabled = true;
1144 	return NOTIFY_OK;
1145 }
1146 
1147 static struct notifier_block brnf_notifier __read_mostly = {
1148 	.notifier_call = brnf_device_event,
1149 };
1150 
1151 /* recursively invokes nf_hook_slow (again), skipping already-called
1152  * hooks (< NF_BR_PRI_BRNF).
1153  *
1154  * Called with rcu read lock held.
1155  */
br_nf_hook_thresh(unsigned int hook,struct net * net,struct sock * sk,struct sk_buff * skb,struct net_device * indev,struct net_device * outdev,int (* okfn)(struct net *,struct sock *,struct sk_buff *))1156 int br_nf_hook_thresh(unsigned int hook, struct net *net,
1157 		      struct sock *sk, struct sk_buff *skb,
1158 		      struct net_device *indev,
1159 		      struct net_device *outdev,
1160 		      int (*okfn)(struct net *, struct sock *,
1161 				  struct sk_buff *))
1162 {
1163 	const struct nf_hook_entries *e;
1164 	struct nf_hook_state state;
1165 	struct nf_hook_ops **ops;
1166 	unsigned int i;
1167 	int ret;
1168 
1169 	e = rcu_dereference(net->nf.hooks_bridge[hook]);
1170 	if (!e)
1171 		return okfn(net, sk, skb);
1172 
1173 	ops = nf_hook_entries_get_hook_ops(e);
1174 	for (i = 0; i < e->num_hook_entries; i++) {
1175 		/* These hooks have already been called */
1176 		if (ops[i]->priority < NF_BR_PRI_BRNF)
1177 			continue;
1178 
1179 		/* These hooks have not been called yet, run them. */
1180 		if (ops[i]->priority > NF_BR_PRI_BRNF)
1181 			break;
1182 
1183 		/* take a closer look at NF_BR_PRI_BRNF. */
1184 		if (ops[i]->hook == br_nf_pre_routing) {
1185 			/* This hook diverted the skb to this function,
1186 			 * hooks after this have not been run yet.
1187 			 */
1188 			i++;
1189 			break;
1190 		}
1191 	}
1192 
1193 	nf_hook_state_init(&state, hook, NFPROTO_BRIDGE, indev, outdev,
1194 			   sk, net, okfn);
1195 
1196 	ret = nf_hook_slow(skb, &state, e, i);
1197 	if (ret == 1)
1198 		ret = okfn(net, sk, skb);
1199 
1200 	return ret;
1201 }
1202 
1203 #ifdef CONFIG_SYSCTL
1204 static
brnf_sysctl_call_tables(const struct ctl_table * ctl,int write,void * buffer,size_t * lenp,loff_t * ppos)1205 int brnf_sysctl_call_tables(const struct ctl_table *ctl, int write,
1206 			    void *buffer, size_t *lenp, loff_t *ppos)
1207 {
1208 	int ret;
1209 
1210 	ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
1211 
1212 	if (write && *(int *)(ctl->data))
1213 		*(int *)(ctl->data) = 1;
1214 	return ret;
1215 }
1216 
1217 static struct ctl_table brnf_table[] = {
1218 	{
1219 		.procname	= "bridge-nf-call-arptables",
1220 		.maxlen		= sizeof(int),
1221 		.mode		= 0644,
1222 		.proc_handler	= brnf_sysctl_call_tables,
1223 	},
1224 	{
1225 		.procname	= "bridge-nf-call-iptables",
1226 		.maxlen		= sizeof(int),
1227 		.mode		= 0644,
1228 		.proc_handler	= brnf_sysctl_call_tables,
1229 	},
1230 	{
1231 		.procname	= "bridge-nf-call-ip6tables",
1232 		.maxlen		= sizeof(int),
1233 		.mode		= 0644,
1234 		.proc_handler	= brnf_sysctl_call_tables,
1235 	},
1236 	{
1237 		.procname	= "bridge-nf-filter-vlan-tagged",
1238 		.maxlen		= sizeof(int),
1239 		.mode		= 0644,
1240 		.proc_handler	= brnf_sysctl_call_tables,
1241 	},
1242 	{
1243 		.procname	= "bridge-nf-filter-pppoe-tagged",
1244 		.maxlen		= sizeof(int),
1245 		.mode		= 0644,
1246 		.proc_handler	= brnf_sysctl_call_tables,
1247 	},
1248 	{
1249 		.procname	= "bridge-nf-pass-vlan-input-dev",
1250 		.maxlen		= sizeof(int),
1251 		.mode		= 0644,
1252 		.proc_handler	= brnf_sysctl_call_tables,
1253 	},
1254 };
1255 
br_netfilter_sysctl_default(struct brnf_net * brnf)1256 static inline void br_netfilter_sysctl_default(struct brnf_net *brnf)
1257 {
1258 	brnf->call_iptables = 1;
1259 	brnf->call_ip6tables = 1;
1260 	brnf->call_arptables = 1;
1261 	brnf->filter_vlan_tagged = 0;
1262 	brnf->filter_pppoe_tagged = 0;
1263 	brnf->pass_vlan_indev = 0;
1264 }
1265 
br_netfilter_sysctl_init_net(struct net * net)1266 static int br_netfilter_sysctl_init_net(struct net *net)
1267 {
1268 	struct ctl_table *table = brnf_table;
1269 	struct brnf_net *brnet;
1270 
1271 	if (!net_eq(net, &init_net)) {
1272 		table = kmemdup(table, sizeof(brnf_table), GFP_KERNEL);
1273 		if (!table)
1274 			return -ENOMEM;
1275 	}
1276 
1277 	brnet = net_generic(net, brnf_net_id);
1278 	table[0].data = &brnet->call_arptables;
1279 	table[1].data = &brnet->call_iptables;
1280 	table[2].data = &brnet->call_ip6tables;
1281 	table[3].data = &brnet->filter_vlan_tagged;
1282 	table[4].data = &brnet->filter_pppoe_tagged;
1283 	table[5].data = &brnet->pass_vlan_indev;
1284 
1285 	br_netfilter_sysctl_default(brnet);
1286 
1287 	brnet->ctl_hdr = register_net_sysctl_sz(net, "net/bridge", table,
1288 						ARRAY_SIZE(brnf_table));
1289 	if (!brnet->ctl_hdr) {
1290 		if (!net_eq(net, &init_net))
1291 			kfree(table);
1292 
1293 		return -ENOMEM;
1294 	}
1295 
1296 	return 0;
1297 }
1298 
br_netfilter_sysctl_exit_net(struct net * net,struct brnf_net * brnet)1299 static void br_netfilter_sysctl_exit_net(struct net *net,
1300 					 struct brnf_net *brnet)
1301 {
1302 	const struct ctl_table *table = brnet->ctl_hdr->ctl_table_arg;
1303 
1304 	unregister_net_sysctl_table(brnet->ctl_hdr);
1305 	if (!net_eq(net, &init_net))
1306 		kfree(table);
1307 }
1308 
brnf_init_net(struct net * net)1309 static int __net_init brnf_init_net(struct net *net)
1310 {
1311 	return br_netfilter_sysctl_init_net(net);
1312 }
1313 #endif
1314 
brnf_exit_net(struct net * net)1315 static void __net_exit brnf_exit_net(struct net *net)
1316 {
1317 	struct brnf_net *brnet;
1318 
1319 	brnet = net_generic(net, brnf_net_id);
1320 	if (brnet->enabled) {
1321 		nf_unregister_net_hooks(net, br_nf_ops, ARRAY_SIZE(br_nf_ops));
1322 		brnet->enabled = false;
1323 	}
1324 
1325 #ifdef CONFIG_SYSCTL
1326 	br_netfilter_sysctl_exit_net(net, brnet);
1327 #endif
1328 }
1329 
1330 static struct pernet_operations brnf_net_ops __read_mostly = {
1331 #ifdef CONFIG_SYSCTL
1332 	.init = brnf_init_net,
1333 #endif
1334 	.exit = brnf_exit_net,
1335 	.id   = &brnf_net_id,
1336 	.size = sizeof(struct brnf_net),
1337 };
1338 
br_netfilter_init(void)1339 static int __init br_netfilter_init(void)
1340 {
1341 	int ret;
1342 
1343 	ret = register_pernet_subsys(&brnf_net_ops);
1344 	if (ret < 0)
1345 		return ret;
1346 
1347 	ret = register_netdevice_notifier(&brnf_notifier);
1348 	if (ret < 0) {
1349 		unregister_pernet_subsys(&brnf_net_ops);
1350 		return ret;
1351 	}
1352 
1353 	RCU_INIT_POINTER(nf_br_ops, &br_ops);
1354 	printk(KERN_NOTICE "Bridge firewalling registered\n");
1355 	return 0;
1356 }
1357 
br_netfilter_fini(void)1358 static void __exit br_netfilter_fini(void)
1359 {
1360 	RCU_INIT_POINTER(nf_br_ops, NULL);
1361 	unregister_netdevice_notifier(&brnf_notifier);
1362 	unregister_pernet_subsys(&brnf_net_ops);
1363 }
1364 
1365 module_init(br_netfilter_init);
1366 module_exit(br_netfilter_fini);
1367 
1368 MODULE_LICENSE("GPL");
1369 MODULE_AUTHOR("Lennert Buytenhek <buytenh@gnu.org>");
1370 MODULE_AUTHOR("Bart De Schuymer <bdschuym@pandora.be>");
1371 MODULE_DESCRIPTION("Linux ethernet netfilter firewall bridge");
1372