xref: /linux/net/openvswitch/vport-netdev.c (revision 4e95bc268b915c3a19ec8b9110f61e4ea41a1ed0)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright (c) 2007-2012 Nicira, Inc.
4  */
5 
6 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
7 
8 #include <linux/if_arp.h>
9 #include <linux/if_bridge.h>
10 #include <linux/if_vlan.h>
11 #include <linux/kernel.h>
12 #include <linux/llc.h>
13 #include <linux/rtnetlink.h>
14 #include <linux/skbuff.h>
15 #include <linux/openvswitch.h>
16 #include <linux/export.h>
17 
18 #include <net/ip_tunnels.h>
19 #include <net/rtnetlink.h>
20 
21 #include "datapath.h"
22 #include "vport.h"
23 #include "vport-internal_dev.h"
24 #include "vport-netdev.h"
25 
26 static struct vport_ops ovs_netdev_vport_ops;
27 
28 /* Must be called with rcu_read_lock. */
29 static void netdev_port_receive(struct sk_buff *skb)
30 {
31 	struct vport *vport;
32 
33 	vport = ovs_netdev_get_vport(skb->dev);
34 	if (unlikely(!vport))
35 		goto error;
36 
37 	if (unlikely(skb_warn_if_lro(skb)))
38 		goto error;
39 
40 	/* Make our own copy of the packet.  Otherwise we will mangle the
41 	 * packet for anyone who came before us (e.g. tcpdump via AF_PACKET).
42 	 */
43 	skb = skb_share_check(skb, GFP_ATOMIC);
44 	if (unlikely(!skb))
45 		return;
46 
47 	if (skb->dev->type == ARPHRD_ETHER) {
48 		skb_push(skb, ETH_HLEN);
49 		skb_postpush_rcsum(skb, skb->data, ETH_HLEN);
50 	}
51 	ovs_vport_receive(vport, skb, skb_tunnel_info(skb));
52 	return;
53 error:
54 	kfree_skb(skb);
55 }
56 
57 /* Called with rcu_read_lock and bottom-halves disabled. */
58 static rx_handler_result_t netdev_frame_hook(struct sk_buff **pskb)
59 {
60 	struct sk_buff *skb = *pskb;
61 
62 	if (unlikely(skb->pkt_type == PACKET_LOOPBACK))
63 		return RX_HANDLER_PASS;
64 
65 	netdev_port_receive(skb);
66 	return RX_HANDLER_CONSUMED;
67 }
68 
69 static struct net_device *get_dpdev(const struct datapath *dp)
70 {
71 	struct vport *local;
72 
73 	local = ovs_vport_ovsl(dp, OVSP_LOCAL);
74 	return local->dev;
75 }
76 
77 struct vport *ovs_netdev_link(struct vport *vport, const char *name)
78 {
79 	int err;
80 
81 	vport->dev = dev_get_by_name(ovs_dp_get_net(vport->dp), name);
82 	if (!vport->dev) {
83 		err = -ENODEV;
84 		goto error_free_vport;
85 	}
86 
87 	if (vport->dev->flags & IFF_LOOPBACK ||
88 	    (vport->dev->type != ARPHRD_ETHER &&
89 	     vport->dev->type != ARPHRD_NONE) ||
90 	    ovs_is_internal_dev(vport->dev)) {
91 		err = -EINVAL;
92 		goto error_put;
93 	}
94 
95 	rtnl_lock();
96 	err = netdev_master_upper_dev_link(vport->dev,
97 					   get_dpdev(vport->dp),
98 					   NULL, NULL, NULL);
99 	if (err)
100 		goto error_unlock;
101 
102 	err = netdev_rx_handler_register(vport->dev, netdev_frame_hook,
103 					 vport);
104 	if (err)
105 		goto error_master_upper_dev_unlink;
106 
107 	dev_disable_lro(vport->dev);
108 	dev_set_promiscuity(vport->dev, 1);
109 	vport->dev->priv_flags |= IFF_OVS_DATAPATH;
110 	rtnl_unlock();
111 
112 	return vport;
113 
114 error_master_upper_dev_unlink:
115 	netdev_upper_dev_unlink(vport->dev, get_dpdev(vport->dp));
116 error_unlock:
117 	rtnl_unlock();
118 error_put:
119 	dev_put(vport->dev);
120 error_free_vport:
121 	ovs_vport_free(vport);
122 	return ERR_PTR(err);
123 }
124 EXPORT_SYMBOL_GPL(ovs_netdev_link);
125 
126 static struct vport *netdev_create(const struct vport_parms *parms)
127 {
128 	struct vport *vport;
129 
130 	vport = ovs_vport_alloc(0, &ovs_netdev_vport_ops, parms);
131 	if (IS_ERR(vport))
132 		return vport;
133 
134 	return ovs_netdev_link(vport, parms->name);
135 }
136 
137 static void vport_netdev_free(struct rcu_head *rcu)
138 {
139 	struct vport *vport = container_of(rcu, struct vport, rcu);
140 
141 	if (vport->dev)
142 		dev_put(vport->dev);
143 	ovs_vport_free(vport);
144 }
145 
146 void ovs_netdev_detach_dev(struct vport *vport)
147 {
148 	ASSERT_RTNL();
149 	vport->dev->priv_flags &= ~IFF_OVS_DATAPATH;
150 	netdev_rx_handler_unregister(vport->dev);
151 	netdev_upper_dev_unlink(vport->dev,
152 				netdev_master_upper_dev_get(vport->dev));
153 	dev_set_promiscuity(vport->dev, -1);
154 }
155 
156 static void netdev_destroy(struct vport *vport)
157 {
158 	rtnl_lock();
159 	if (netif_is_ovs_port(vport->dev))
160 		ovs_netdev_detach_dev(vport);
161 	rtnl_unlock();
162 
163 	call_rcu(&vport->rcu, vport_netdev_free);
164 }
165 
166 void ovs_netdev_tunnel_destroy(struct vport *vport)
167 {
168 	rtnl_lock();
169 	if (netif_is_ovs_port(vport->dev))
170 		ovs_netdev_detach_dev(vport);
171 
172 	/* We can be invoked by both explicit vport deletion and
173 	 * underlying netdev deregistration; delete the link only
174 	 * if it's not already shutting down.
175 	 */
176 	if (vport->dev->reg_state == NETREG_REGISTERED)
177 		rtnl_delete_link(vport->dev);
178 	dev_put(vport->dev);
179 	vport->dev = NULL;
180 	rtnl_unlock();
181 
182 	call_rcu(&vport->rcu, vport_netdev_free);
183 }
184 EXPORT_SYMBOL_GPL(ovs_netdev_tunnel_destroy);
185 
186 /* Returns null if this device is not attached to a datapath. */
187 struct vport *ovs_netdev_get_vport(struct net_device *dev)
188 {
189 	if (likely(netif_is_ovs_port(dev)))
190 		return (struct vport *)
191 			rcu_dereference_rtnl(dev->rx_handler_data);
192 	else
193 		return NULL;
194 }
195 
196 static struct vport_ops ovs_netdev_vport_ops = {
197 	.type		= OVS_VPORT_TYPE_NETDEV,
198 	.create		= netdev_create,
199 	.destroy	= netdev_destroy,
200 	.send		= dev_queue_xmit,
201 };
202 
203 int __init ovs_netdev_init(void)
204 {
205 	return ovs_vport_ops_register(&ovs_netdev_vport_ops);
206 }
207 
208 void ovs_netdev_exit(void)
209 {
210 	ovs_vport_ops_unregister(&ovs_netdev_vport_ops);
211 }
212