vport-internal_dev.c 7.8 KB
Newer Older
1
/*
2
 * Copyright (c) 2007-2012 Nicira, Inc.
3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26
 *
 * This program is free software; you can redistribute it and/or
 * modify it under the terms of version 2 of the GNU General Public
 * License as published by the Free Software Foundation.
 *
 * This program is distributed in the hope that it will be useful, but
 * WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
 * General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program; if not, write to the Free Software
 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
 * 02110-1301, USA
 */

#include <linux/hardirq.h>
#include <linux/if_vlan.h>
#include <linux/kernel.h>
#include <linux/netdevice.h>
#include <linux/etherdevice.h>
#include <linux/ethtool.h>
#include <linux/skbuff.h>

27 28
#include <net/dst.h>
#include <net/xfrm.h>
29
#include <net/rtnetlink.h>
30

31 32 33 34 35 36 37 38
#include "datapath.h"
#include "vport-internal_dev.h"
#include "vport-netdev.h"

struct internal_dev {
	struct vport *vport;
};

39 40
static struct vport_ops ovs_internal_vport_ops;

41 42 43 44 45 46 47 48
static struct internal_dev *internal_dev_priv(struct net_device *netdev)
{
	return netdev_priv(netdev);
}

/* Called with rcu_read_lock_bh. */
static int internal_dev_xmit(struct sk_buff *skb, struct net_device *netdev)
{
49 50 51
	int len, err;

	len = skb->len;
52
	rcu_read_lock();
53
	err = ovs_vport_receive(internal_dev_priv(netdev)->vport, skb, NULL);
54
	rcu_read_unlock();
55 56 57 58 59 60 61 62 63 64 65

	if (likely(!err)) {
		struct pcpu_sw_netstats *tstats = this_cpu_ptr(netdev->tstats);

		u64_stats_update_begin(&tstats->syncp);
		tstats->tx_bytes += len;
		tstats->tx_packets++;
		u64_stats_update_end(&tstats->syncp);
	} else {
		netdev->stats.tx_errors++;
	}
66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83
	return 0;
}

static int internal_dev_open(struct net_device *netdev)
{
	netif_start_queue(netdev);
	return 0;
}

static int internal_dev_stop(struct net_device *netdev)
{
	netif_stop_queue(netdev);
	return 0;
}

static void internal_dev_getinfo(struct net_device *netdev,
				 struct ethtool_drvinfo *info)
{
84
	strlcpy(info->driver, "openvswitch", sizeof(info->driver));
85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108
}

static const struct ethtool_ops internal_dev_ethtool_ops = {
	.get_drvinfo	= internal_dev_getinfo,
	.get_link	= ethtool_op_get_link,
};

static int internal_dev_change_mtu(struct net_device *netdev, int new_mtu)
{
	if (new_mtu < 68)
		return -EINVAL;

	netdev->mtu = new_mtu;
	return 0;
}

static void internal_dev_destructor(struct net_device *dev)
{
	struct vport *vport = ovs_internal_dev_get_vport(dev);

	ovs_vport_free(vport);
	free_netdev(dev);
}

109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140
static struct rtnl_link_stats64 *
internal_get_stats(struct net_device *dev, struct rtnl_link_stats64 *stats)
{
	int i;

	memset(stats, 0, sizeof(*stats));
	stats->rx_errors  = dev->stats.rx_errors;
	stats->tx_errors  = dev->stats.tx_errors;
	stats->tx_dropped = dev->stats.tx_dropped;
	stats->rx_dropped = dev->stats.rx_dropped;

	for_each_possible_cpu(i) {
		const struct pcpu_sw_netstats *percpu_stats;
		struct pcpu_sw_netstats local_stats;
		unsigned int start;

		percpu_stats = per_cpu_ptr(dev->tstats, i);

		do {
			start = u64_stats_fetch_begin_irq(&percpu_stats->syncp);
			local_stats = *percpu_stats;
		} while (u64_stats_fetch_retry_irq(&percpu_stats->syncp, start));

		stats->rx_bytes         += local_stats.rx_bytes;
		stats->rx_packets       += local_stats.rx_packets;
		stats->tx_bytes         += local_stats.tx_bytes;
		stats->tx_packets       += local_stats.tx_packets;
	}

	return stats;
}

141
static void internal_set_rx_headroom(struct net_device *dev, int new_hr)
142 143 144 145
{
	dev->needed_headroom = new_hr;
}

146 147 148 149
static const struct net_device_ops internal_dev_netdev_ops = {
	.ndo_open = internal_dev_open,
	.ndo_stop = internal_dev_stop,
	.ndo_start_xmit = internal_dev_xmit,
150
	.ndo_set_mac_address = eth_mac_addr,
151
	.ndo_change_mtu = internal_dev_change_mtu,
152
	.ndo_get_stats64 = internal_get_stats,
153
	.ndo_set_rx_headroom = internal_set_rx_headroom,
154 155
};

156 157 158 159
static struct rtnl_link_ops internal_dev_link_ops __read_mostly = {
	.kind = "openvswitch",
};

160 161 162 163 164 165 166
static void do_setup(struct net_device *netdev)
{
	ether_setup(netdev);

	netdev->netdev_ops = &internal_dev_netdev_ops;

	netdev->priv_flags &= ~IFF_TX_SKB_SHARING;
167 168
	netdev->priv_flags |= IFF_LIVE_ADDR_CHANGE | IFF_OPENVSWITCH |
			      IFF_PHONY_HEADROOM;
169
	netdev->destructor = internal_dev_destructor;
170
	netdev->ethtool_ops = &internal_dev_ethtool_ops;
171
	netdev->rtnl_link_ops = &internal_dev_link_ops;
172 173 174
	netdev->tx_queue_len = 0;

	netdev->features = NETIF_F_LLTX | NETIF_F_SG | NETIF_F_FRAGLIST |
175 176
			   NETIF_F_HIGHDMA | NETIF_F_HW_CSUM |
			   NETIF_F_GSO_SOFTWARE | NETIF_F_GSO_ENCAP_ALL;
177 178

	netdev->vlan_features = netdev->features;
179
	netdev->hw_enc_features = netdev->features;
180
	netdev->features |= NETIF_F_HW_VLAN_CTAG_TX;
181
	netdev->hw_features = netdev->features & ~NETIF_F_LLTX;
182

183
	eth_hw_addr_random(netdev);
184 185 186 187 188 189 190 191
}

static struct vport *internal_dev_create(const struct vport_parms *parms)
{
	struct vport *vport;
	struct internal_dev *internal_dev;
	int err;

192
	vport = ovs_vport_alloc(0, &ovs_internal_vport_ops, parms);
193 194 195 196 197
	if (IS_ERR(vport)) {
		err = PTR_ERR(vport);
		goto error;
	}

198 199 200
	vport->dev = alloc_netdev(sizeof(struct internal_dev),
				  parms->name, NET_NAME_UNKNOWN, do_setup);
	if (!vport->dev) {
201 202 203
		err = -ENOMEM;
		goto error_free_vport;
	}
204 205 206 207 208
	vport->dev->tstats = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats);
	if (!vport->dev->tstats) {
		err = -ENOMEM;
		goto error_free_netdev;
	}
209
	vport->dev->needed_headroom = vport->dp->max_headroom;
210

211 212
	dev_net_set(vport->dev, ovs_dp_get_net(vport->dp));
	internal_dev = internal_dev_priv(vport->dev);
213 214
	internal_dev->vport = vport;

215 216
	/* Restrict bridge port to current netns. */
	if (vport->port_no == OVSP_LOCAL)
217
		vport->dev->features |= NETIF_F_NETNS_LOCAL;
218

219
	rtnl_lock();
220
	err = register_netdevice(vport->dev);
221
	if (err)
222
		goto error_unlock;
223

224
	dev_set_promiscuity(vport->dev, 1);
225
	rtnl_unlock();
226
	netif_start_queue(vport->dev);
227 228 229

	return vport;

230
error_unlock:
231
	rtnl_unlock();
232 233
	free_percpu(vport->dev->tstats);
error_free_netdev:
234
	free_netdev(vport->dev);
235 236 237 238 239 240 241 242
error_free_vport:
	ovs_vport_free(vport);
error:
	return ERR_PTR(err);
}

static void internal_dev_destroy(struct vport *vport)
{
243
	netif_stop_queue(vport->dev);
244
	rtnl_lock();
245
	dev_set_promiscuity(vport->dev, -1);
246 247

	/* unregister_netdevice() waits for an RCU grace period. */
248
	unregister_netdevice(vport->dev);
249
	free_percpu(vport->dev->tstats);
250
	rtnl_unlock();
251 252
}

253
static netdev_tx_t internal_dev_recv(struct sk_buff *skb)
254
{
255
	struct net_device *netdev = skb->dev;
256
	struct pcpu_sw_netstats *stats;
257

258 259
	if (unlikely(!(netdev->flags & IFF_UP))) {
		kfree_skb(skb);
260
		netdev->stats.rx_dropped++;
261
		return NETDEV_TX_OK;
262 263
	}

264 265 266 267
	skb_dst_drop(skb);
	nf_reset(skb);
	secpath_reset(skb);

268 269
	skb->pkt_type = PACKET_HOST;
	skb->protocol = eth_type_trans(skb, netdev);
270
	skb_postpull_rcsum(skb, eth_hdr(skb), ETH_HLEN);
271

272 273 274 275 276
	stats = this_cpu_ptr(netdev->tstats);
	u64_stats_update_begin(&stats->syncp);
	stats->rx_packets++;
	stats->rx_bytes += skb->len;
	u64_stats_update_end(&stats->syncp);
277

278
	netif_rx(skb);
279
	return NETDEV_TX_OK;
280 281
}

282
static struct vport_ops ovs_internal_vport_ops = {
283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300
	.type		= OVS_VPORT_TYPE_INTERNAL,
	.create		= internal_dev_create,
	.destroy	= internal_dev_destroy,
	.send		= internal_dev_recv,
};

int ovs_is_internal_dev(const struct net_device *netdev)
{
	return netdev->netdev_ops == &internal_dev_netdev_ops;
}

struct vport *ovs_internal_dev_get_vport(struct net_device *netdev)
{
	if (!ovs_is_internal_dev(netdev))
		return NULL;

	return internal_dev_priv(netdev)->vport;
}
301 302 303

int ovs_internal_dev_rtnl_link_register(void)
{
304 305 306 307 308 309 310 311 312 313 314
	int err;

	err = rtnl_link_register(&internal_dev_link_ops);
	if (err < 0)
		return err;

	err = ovs_vport_ops_register(&ovs_internal_vport_ops);
	if (err < 0)
		rtnl_link_unregister(&internal_dev_link_ops);

	return err;
315 316 317 318
}

void ovs_internal_dev_rtnl_link_unregister(void)
{
319
	ovs_vport_ops_unregister(&ovs_internal_vport_ops);
320 321
	rtnl_link_unregister(&internal_dev_link_ops);
}