vport-internal_dev.c 7.8 KB
Newer Older
1
/*
2
 * Copyright (c) 2007-2012 Nicira, Inc.
3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26
 *
 * This program is free software; you can redistribute it and/or
 * modify it under the terms of version 2 of the GNU General Public
 * License as published by the Free Software Foundation.
 *
 * This program is distributed in the hope that it will be useful, but
 * WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
 * General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program; if not, write to the Free Software
 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
 * 02110-1301, USA
 */

#include <linux/hardirq.h>
#include <linux/if_vlan.h>
#include <linux/kernel.h>
#include <linux/netdevice.h>
#include <linux/etherdevice.h>
#include <linux/ethtool.h>
#include <linux/skbuff.h>

27 28
#include <net/dst.h>
#include <net/xfrm.h>
29
#include <net/rtnetlink.h>
30

31 32 33 34 35 36 37 38
#include "datapath.h"
#include "vport-internal_dev.h"
#include "vport-netdev.h"

struct internal_dev {
	struct vport *vport;
};

39 40
static struct vport_ops ovs_internal_vport_ops;

41 42 43 44 45 46 47 48
static struct internal_dev *internal_dev_priv(struct net_device *netdev)
{
	return netdev_priv(netdev);
}

/* Called with rcu_read_lock_bh. */
static int internal_dev_xmit(struct sk_buff *skb, struct net_device *netdev)
{
49 50 51
	int len, err;

	len = skb->len;
52
	rcu_read_lock();
53
	err = ovs_vport_receive(internal_dev_priv(netdev)->vport, skb, NULL);
54
	rcu_read_unlock();
55 56 57 58 59 60 61 62 63 64 65

	if (likely(!err)) {
		struct pcpu_sw_netstats *tstats = this_cpu_ptr(netdev->tstats);

		u64_stats_update_begin(&tstats->syncp);
		tstats->tx_bytes += len;
		tstats->tx_packets++;
		u64_stats_update_end(&tstats->syncp);
	} else {
		netdev->stats.tx_errors++;
	}
66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83
	return 0;
}

static int internal_dev_open(struct net_device *netdev)
{
	netif_start_queue(netdev);
	return 0;
}

static int internal_dev_stop(struct net_device *netdev)
{
	netif_stop_queue(netdev);
	return 0;
}

static void internal_dev_getinfo(struct net_device *netdev,
				 struct ethtool_drvinfo *info)
{
84
	strlcpy(info->driver, "openvswitch", sizeof(info->driver));
85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108
}

static const struct ethtool_ops internal_dev_ethtool_ops = {
	.get_drvinfo	= internal_dev_getinfo,
	.get_link	= ethtool_op_get_link,
};

static int internal_dev_change_mtu(struct net_device *netdev, int new_mtu)
{
	if (new_mtu < 68)
		return -EINVAL;

	netdev->mtu = new_mtu;
	return 0;
}

static void internal_dev_destructor(struct net_device *dev)
{
	struct vport *vport = ovs_internal_dev_get_vport(dev);

	ovs_vport_free(vport);
	free_netdev(dev);
}

109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140
static struct rtnl_link_stats64 *
internal_get_stats(struct net_device *dev, struct rtnl_link_stats64 *stats)
{
	int i;

	memset(stats, 0, sizeof(*stats));
	stats->rx_errors  = dev->stats.rx_errors;
	stats->tx_errors  = dev->stats.tx_errors;
	stats->tx_dropped = dev->stats.tx_dropped;
	stats->rx_dropped = dev->stats.rx_dropped;

	for_each_possible_cpu(i) {
		const struct pcpu_sw_netstats *percpu_stats;
		struct pcpu_sw_netstats local_stats;
		unsigned int start;

		percpu_stats = per_cpu_ptr(dev->tstats, i);

		do {
			start = u64_stats_fetch_begin_irq(&percpu_stats->syncp);
			local_stats = *percpu_stats;
		} while (u64_stats_fetch_retry_irq(&percpu_stats->syncp, start));

		stats->rx_bytes         += local_stats.rx_bytes;
		stats->rx_packets       += local_stats.rx_packets;
		stats->tx_bytes         += local_stats.tx_bytes;
		stats->tx_packets       += local_stats.tx_packets;
	}

	return stats;
}

141
static void internal_set_rx_headroom(struct net_device *dev, int new_hr)
142 143 144 145
{
	dev->needed_headroom = new_hr;
}

146 147 148 149
static const struct net_device_ops internal_dev_netdev_ops = {
	.ndo_open = internal_dev_open,
	.ndo_stop = internal_dev_stop,
	.ndo_start_xmit = internal_dev_xmit,
150
	.ndo_set_mac_address = eth_mac_addr,
151
	.ndo_change_mtu = internal_dev_change_mtu,
152
	.ndo_get_stats64 = internal_get_stats,
153
	.ndo_set_rx_headroom = internal_set_rx_headroom,
154 155
};

156 157 158 159
static struct rtnl_link_ops internal_dev_link_ops __read_mostly = {
	.kind = "openvswitch",
};

160 161 162 163 164 165 166
static void do_setup(struct net_device *netdev)
{
	ether_setup(netdev);

	netdev->netdev_ops = &internal_dev_netdev_ops;

	netdev->priv_flags &= ~IFF_TX_SKB_SHARING;
167
	netdev->priv_flags |= IFF_LIVE_ADDR_CHANGE | IFF_OPENVSWITCH |
168
			      IFF_PHONY_HEADROOM | IFF_NO_QUEUE;
169
	netdev->destructor = internal_dev_destructor;
170
	netdev->ethtool_ops = &internal_dev_ethtool_ops;
171
	netdev->rtnl_link_ops = &internal_dev_link_ops;
172 173

	netdev->features = NETIF_F_LLTX | NETIF_F_SG | NETIF_F_FRAGLIST |
174 175
			   NETIF_F_HIGHDMA | NETIF_F_HW_CSUM |
			   NETIF_F_GSO_SOFTWARE | NETIF_F_GSO_ENCAP_ALL;
176 177

	netdev->vlan_features = netdev->features;
178
	netdev->hw_enc_features = netdev->features;
179
	netdev->features |= NETIF_F_HW_VLAN_CTAG_TX;
180
	netdev->hw_features = netdev->features & ~NETIF_F_LLTX;
181

182
	eth_hw_addr_random(netdev);
183 184 185 186 187 188 189 190
}

static struct vport *internal_dev_create(const struct vport_parms *parms)
{
	struct vport *vport;
	struct internal_dev *internal_dev;
	int err;

191
	vport = ovs_vport_alloc(0, &ovs_internal_vport_ops, parms);
192 193 194 195 196
	if (IS_ERR(vport)) {
		err = PTR_ERR(vport);
		goto error;
	}

197
	vport->dev = alloc_netdev(sizeof(struct internal_dev),
198
				  parms->name, NET_NAME_USER, do_setup);
199
	if (!vport->dev) {
200 201 202
		err = -ENOMEM;
		goto error_free_vport;
	}
203 204 205 206 207
	vport->dev->tstats = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats);
	if (!vport->dev->tstats) {
		err = -ENOMEM;
		goto error_free_netdev;
	}
208
	vport->dev->needed_headroom = vport->dp->max_headroom;
209

210 211
	dev_net_set(vport->dev, ovs_dp_get_net(vport->dp));
	internal_dev = internal_dev_priv(vport->dev);
212 213
	internal_dev->vport = vport;

214 215
	/* Restrict bridge port to current netns. */
	if (vport->port_no == OVSP_LOCAL)
216
		vport->dev->features |= NETIF_F_NETNS_LOCAL;
217

218
	rtnl_lock();
219
	err = register_netdevice(vport->dev);
220
	if (err)
221
		goto error_unlock;
222

223
	dev_set_promiscuity(vport->dev, 1);
224
	rtnl_unlock();
225
	netif_start_queue(vport->dev);
226 227 228

	return vport;

229
error_unlock:
230
	rtnl_unlock();
231 232
	free_percpu(vport->dev->tstats);
error_free_netdev:
233
	free_netdev(vport->dev);
234 235 236 237 238 239 240 241
error_free_vport:
	ovs_vport_free(vport);
error:
	return ERR_PTR(err);
}

static void internal_dev_destroy(struct vport *vport)
{
242
	netif_stop_queue(vport->dev);
243
	rtnl_lock();
244
	dev_set_promiscuity(vport->dev, -1);
245 246

	/* unregister_netdevice() waits for an RCU grace period. */
247
	unregister_netdevice(vport->dev);
248
	free_percpu(vport->dev->tstats);
249
	rtnl_unlock();
250 251
}

252
static netdev_tx_t internal_dev_recv(struct sk_buff *skb)
253
{
254
	struct net_device *netdev = skb->dev;
255
	struct pcpu_sw_netstats *stats;
256

257 258
	if (unlikely(!(netdev->flags & IFF_UP))) {
		kfree_skb(skb);
259
		netdev->stats.rx_dropped++;
260
		return NETDEV_TX_OK;
261 262
	}

263 264 265 266
	skb_dst_drop(skb);
	nf_reset(skb);
	secpath_reset(skb);

267 268
	skb->pkt_type = PACKET_HOST;
	skb->protocol = eth_type_trans(skb, netdev);
269
	skb_postpull_rcsum(skb, eth_hdr(skb), ETH_HLEN);
270

271 272 273 274 275
	stats = this_cpu_ptr(netdev->tstats);
	u64_stats_update_begin(&stats->syncp);
	stats->rx_packets++;
	stats->rx_bytes += skb->len;
	u64_stats_update_end(&stats->syncp);
276

277
	netif_rx(skb);
278
	return NETDEV_TX_OK;
279 280
}

281
static struct vport_ops ovs_internal_vport_ops = {
282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299
	.type		= OVS_VPORT_TYPE_INTERNAL,
	.create		= internal_dev_create,
	.destroy	= internal_dev_destroy,
	.send		= internal_dev_recv,
};

int ovs_is_internal_dev(const struct net_device *netdev)
{
	return netdev->netdev_ops == &internal_dev_netdev_ops;
}

struct vport *ovs_internal_dev_get_vport(struct net_device *netdev)
{
	if (!ovs_is_internal_dev(netdev))
		return NULL;

	return internal_dev_priv(netdev)->vport;
}
300 301 302

int ovs_internal_dev_rtnl_link_register(void)
{
303 304 305 306 307 308 309 310 311 312 313
	int err;

	err = rtnl_link_register(&internal_dev_link_ops);
	if (err < 0)
		return err;

	err = ovs_vport_ops_register(&ovs_internal_vport_ops);
	if (err < 0)
		rtnl_link_unregister(&internal_dev_link_ops);

	return err;
314 315 316 317
}

void ovs_internal_dev_rtnl_link_unregister(void)
{
318
	ovs_vport_ops_unregister(&ovs_internal_vport_ops);
319 320
	rtnl_link_unregister(&internal_dev_link_ops);
}