1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3 * Copyright (c) 2007-2012 Nicira, Inc.
4 */
5
6 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
7
8 #include <linux/if_arp.h>
9 #include <linux/if_bridge.h>
10 #include <linux/if_vlan.h>
11 #include <linux/kernel.h>
12 #include <linux/llc.h>
13 #include <linux/rtnetlink.h>
14 #include <linux/skbuff.h>
15 #include <linux/openvswitch.h>
16 #include <linux/export.h>
17
18 #include <net/ip_tunnels.h>
19 #include <net/rtnetlink.h>
20
21 #include "datapath.h"
22 #include "vport.h"
23 #include "vport-internal_dev.h"
24 #include "vport-netdev.h"
25
26 static struct vport_ops ovs_netdev_vport_ops;
27
28 /* Must be called with rcu_read_lock. */
netdev_port_receive(struct sk_buff * skb)29 static void netdev_port_receive(struct sk_buff *skb)
30 {
31 struct vport *vport;
32
33 vport = ovs_netdev_get_vport(skb->dev);
34 if (unlikely(!vport))
35 goto error;
36
37 if (unlikely(skb_warn_if_lro(skb)))
38 goto error;
39
40 /* Make our own copy of the packet. Otherwise we will mangle the
41 * packet for anyone who came before us (e.g. tcpdump via AF_PACKET).
42 */
43 skb = skb_share_check(skb, GFP_ATOMIC);
44 if (unlikely(!skb))
45 return;
46
47 if (skb->dev->type == ARPHRD_ETHER)
48 skb_push_rcsum(skb, ETH_HLEN);
49
50 ovs_vport_receive(vport, skb, skb_tunnel_info(skb));
51 return;
52 error:
53 kfree_skb(skb);
54 }
55
56 /* Called with rcu_read_lock and bottom-halves disabled. */
netdev_frame_hook(struct sk_buff ** pskb)57 static rx_handler_result_t netdev_frame_hook(struct sk_buff **pskb)
58 {
59 struct sk_buff *skb = *pskb;
60
61 if (unlikely(skb->pkt_type == PACKET_LOOPBACK))
62 return RX_HANDLER_PASS;
63
64 netdev_port_receive(skb);
65 return RX_HANDLER_CONSUMED;
66 }
67
get_dpdev(const struct datapath * dp)68 static struct net_device *get_dpdev(const struct datapath *dp)
69 {
70 struct vport *local;
71
72 local = ovs_vport_ovsl(dp, OVSP_LOCAL);
73 return local->dev;
74 }
75
ovs_netdev_link(struct vport * vport,bool tunnel)76 struct vport *ovs_netdev_link(struct vport *vport, bool tunnel)
77 {
78 int err;
79
80 if (WARN_ON_ONCE(!vport->dev)) {
81 err = -ENODEV;
82 goto error_free_vport;
83 }
84
85 rtnl_lock();
86 err = netdev_master_upper_dev_link(vport->dev,
87 get_dpdev(vport->dp),
88 NULL, NULL, NULL);
89 if (err)
90 goto error_put_unlock;
91
92 err = netdev_rx_handler_register(vport->dev, netdev_frame_hook,
93 vport);
94 if (err)
95 goto error_master_upper_dev_unlink;
96
97 dev_disable_lro(vport->dev);
98 dev_set_promiscuity(vport->dev, 1);
99 vport->dev->priv_flags |= IFF_OVS_DATAPATH;
100 rtnl_unlock();
101
102 return vport;
103
104 error_master_upper_dev_unlink:
105 netdev_upper_dev_unlink(vport->dev, get_dpdev(vport->dp));
106 error_put_unlock:
107 if (tunnel && vport->dev->reg_state == NETREG_REGISTERED)
108 rtnl_delete_link(vport->dev, 0, NULL);
109 netdev_put(vport->dev, &vport->dev_tracker);
110 rtnl_unlock();
111 error_free_vport:
112 ovs_vport_free(vport);
113 return ERR_PTR(err);
114 }
115 EXPORT_SYMBOL_GPL(ovs_netdev_link);
116
netdev_create(const struct vport_parms * parms)117 static struct vport *netdev_create(const struct vport_parms *parms)
118 {
119 struct vport *vport;
120 int err;
121
122 vport = ovs_vport_alloc(0, &ovs_netdev_vport_ops, parms);
123 if (IS_ERR(vport))
124 return vport;
125
126 vport->dev = dev_get_by_name(ovs_dp_get_net(vport->dp), parms->name);
127 if (!vport->dev) {
128 err = -ENODEV;
129 goto error_free_vport;
130 }
131 netdev_tracker_alloc(vport->dev, &vport->dev_tracker, GFP_KERNEL);
132
133 /* Ensure that the provided name is not an alias. */
134 if (strcmp(parms->name, ovs_vport_name(vport))) {
135 err = -ENODEV;
136 goto error_put;
137 }
138
139 if (vport->dev->flags & IFF_LOOPBACK ||
140 (vport->dev->type != ARPHRD_ETHER &&
141 vport->dev->type != ARPHRD_NONE) ||
142 ovs_is_internal_dev(vport->dev)) {
143 err = -EINVAL;
144 goto error_put;
145 }
146
147 return ovs_netdev_link(vport, false);
148 error_put:
149 netdev_put(vport->dev, &vport->dev_tracker);
150 error_free_vport:
151 ovs_vport_free(vport);
152 return ERR_PTR(err);
153 }
154
vport_netdev_free(struct rcu_head * rcu)155 static void vport_netdev_free(struct rcu_head *rcu)
156 {
157 struct vport *vport = container_of(rcu, struct vport, rcu);
158
159 netdev_put(vport->dev, &vport->dev_tracker);
160 ovs_vport_free(vport);
161 }
162
ovs_netdev_detach_dev(struct vport * vport)163 void ovs_netdev_detach_dev(struct vport *vport)
164 {
165 ASSERT_RTNL();
166 netdev_rx_handler_unregister(vport->dev);
167 netdev_upper_dev_unlink(vport->dev,
168 netdev_master_upper_dev_get(vport->dev));
169 dev_set_promiscuity(vport->dev, -1);
170
171 /* paired with smp_mb() in netdev_destroy() */
172 smp_wmb();
173
174 vport->dev->priv_flags &= ~IFF_OVS_DATAPATH;
175 }
176
netdev_destroy(struct vport * vport)177 static void netdev_destroy(struct vport *vport)
178 {
179 /* When called from ovs_db_notify_wq() after a dp_device_event(), the
180 * port has already been detached, so we can avoid taking the RTNL by
181 * checking this first.
182 */
183 if (netif_is_ovs_port(vport->dev)) {
184 rtnl_lock();
185 /* Check again while holding the lock to ensure we don't race
186 * with the netdev notifier and detach twice.
187 */
188 if (netif_is_ovs_port(vport->dev))
189 ovs_netdev_detach_dev(vport);
190 rtnl_unlock();
191 }
192
193 /* paired with smp_wmb() in ovs_netdev_detach_dev() */
194 smp_mb();
195
196 call_rcu(&vport->rcu, vport_netdev_free);
197 }
198
ovs_netdev_tunnel_destroy(struct vport * vport)199 void ovs_netdev_tunnel_destroy(struct vport *vport)
200 {
201 rtnl_lock();
202 if (netif_is_ovs_port(vport->dev))
203 ovs_netdev_detach_dev(vport);
204
205 /* We can be invoked by both explicit vport deletion and
206 * underlying netdev deregistration; delete the link only
207 * if it's not already shutting down.
208 */
209 if (vport->dev->reg_state == NETREG_REGISTERED)
210 rtnl_delete_link(vport->dev, 0, NULL);
211
212 /* We can't put the device reference yet, since it can still be in
213 * use, but rtnl_unlock()->netdev_run_todo() will block until all
214 * the references are released, so the RCU call must be before it.
215 */
216 call_rcu(&vport->rcu, vport_netdev_free);
217 rtnl_unlock();
218 }
219 EXPORT_SYMBOL_GPL(ovs_netdev_tunnel_destroy);
220
221 /* Returns null if this device is not attached to a datapath. */
ovs_netdev_get_vport(struct net_device * dev)222 struct vport *ovs_netdev_get_vport(struct net_device *dev)
223 {
224 if (likely(netif_is_ovs_port(dev)))
225 return (struct vport *)
226 rcu_dereference_rtnl(dev->rx_handler_data);
227 else
228 return NULL;
229 }
230
231 static struct vport_ops ovs_netdev_vport_ops = {
232 .type = OVS_VPORT_TYPE_NETDEV,
233 .create = netdev_create,
234 .destroy = netdev_destroy,
235 .send = dev_queue_xmit,
236 };
237
ovs_netdev_init(void)238 int __init ovs_netdev_init(void)
239 {
240 return ovs_vport_ops_register(&ovs_netdev_vport_ops);
241 }
242
ovs_netdev_exit(void)243 void ovs_netdev_exit(void)
244 {
245 ovs_vport_ops_unregister(&ovs_netdev_vport_ops);
246 }
247