提交 c9d0dc4b 编写于 作者: D David S. Miller

Merge branch 'qualcomm-rmnet-Add-64-bit-stats-and-GRO'

Subash Abhinov Kasiviswanathan says:

====================
net: qualcomm: rmnet: Add 64 bit stats and GRO

This series adds support for 64 bit per cpu stats and GRO

Patches 1-2 are cleanups of return code and a redundant condition
Patch 3 adds support for 64 bit per cpu stats
Patch 4 adds support for GRO using GRO cells

v1->v2: Since gro_cells_init() could potentially fail, move it from device
setup to ndo_init() as mentioned by Eric.
====================
Signed-off-by: NDavid S. Miller <davem@davemloft.net>
...@@ -5,6 +5,7 @@ ...@@ -5,6 +5,7 @@
menuconfig RMNET menuconfig RMNET
tristate "RmNet MAP driver" tristate "RmNet MAP driver"
default n default n
select GRO_CELLS
---help--- ---help---
If you select this, you will enable the RMNET module which is used If you select this, you will enable the RMNET module which is used
for handling data in the multiplexing and aggregation protocol (MAP) for handling data in the multiplexing and aggregation protocol (MAP)
......
...@@ -14,6 +14,7 @@ ...@@ -14,6 +14,7 @@
*/ */
#include <linux/skbuff.h> #include <linux/skbuff.h>
#include <net/gro_cells.h>
#ifndef _RMNET_CONFIG_H_ #ifndef _RMNET_CONFIG_H_
#define _RMNET_CONFIG_H_ #define _RMNET_CONFIG_H_
...@@ -41,9 +42,24 @@ struct rmnet_port { ...@@ -41,9 +42,24 @@ struct rmnet_port {
extern struct rtnl_link_ops rmnet_link_ops; extern struct rtnl_link_ops rmnet_link_ops;
struct rmnet_vnd_stats {
u64 rx_pkts;
u64 rx_bytes;
u64 tx_pkts;
u64 tx_bytes;
u32 tx_drops;
};
struct rmnet_pcpu_stats {
struct rmnet_vnd_stats stats;
struct u64_stats_sync syncp;
};
struct rmnet_priv { struct rmnet_priv {
u8 mux_id; u8 mux_id;
struct net_device *real_dev; struct net_device *real_dev;
struct rmnet_pcpu_stats __percpu *pcpu_stats;
struct gro_cells gro_cells;
}; };
struct rmnet_port *rmnet_get_port(struct net_device *real_dev); struct rmnet_port *rmnet_get_port(struct net_device *real_dev);
......
...@@ -43,22 +43,23 @@ static void rmnet_set_skb_proto(struct sk_buff *skb) ...@@ -43,22 +43,23 @@ static void rmnet_set_skb_proto(struct sk_buff *skb)
/* Generic handler */ /* Generic handler */
static rx_handler_result_t static void
rmnet_deliver_skb(struct sk_buff *skb) rmnet_deliver_skb(struct sk_buff *skb)
{ {
struct rmnet_priv *priv = netdev_priv(skb->dev);
skb_reset_transport_header(skb); skb_reset_transport_header(skb);
skb_reset_network_header(skb); skb_reset_network_header(skb);
rmnet_vnd_rx_fixup(skb, skb->dev); rmnet_vnd_rx_fixup(skb, skb->dev);
skb->pkt_type = PACKET_HOST; skb->pkt_type = PACKET_HOST;
skb_set_mac_header(skb, 0); skb_set_mac_header(skb, 0);
netif_receive_skb(skb); gro_cells_receive(&priv->gro_cells, skb);
return RX_HANDLER_CONSUMED;
} }
/* MAP handler */ /* MAP handler */
static rx_handler_result_t static void
__rmnet_map_ingress_handler(struct sk_buff *skb, __rmnet_map_ingress_handler(struct sk_buff *skb,
struct rmnet_port *port) struct rmnet_port *port)
{ {
...@@ -84,38 +85,33 @@ __rmnet_map_ingress_handler(struct sk_buff *skb, ...@@ -84,38 +85,33 @@ __rmnet_map_ingress_handler(struct sk_buff *skb,
if (!ep) if (!ep)
goto free_skb; goto free_skb;
if (port->ingress_data_format & RMNET_INGRESS_FORMAT_DEMUXING) skb->dev = ep->egress_dev;
skb->dev = ep->egress_dev;
/* Subtract MAP header */ /* Subtract MAP header */
skb_pull(skb, sizeof(struct rmnet_map_header)); skb_pull(skb, sizeof(struct rmnet_map_header));
skb_trim(skb, len); skb_trim(skb, len);
rmnet_set_skb_proto(skb); rmnet_set_skb_proto(skb);
return rmnet_deliver_skb(skb); rmnet_deliver_skb(skb);
return;
free_skb: free_skb:
kfree_skb(skb); kfree_skb(skb);
return RX_HANDLER_CONSUMED;
} }
static rx_handler_result_t static void
rmnet_map_ingress_handler(struct sk_buff *skb, rmnet_map_ingress_handler(struct sk_buff *skb,
struct rmnet_port *port) struct rmnet_port *port)
{ {
struct sk_buff *skbn; struct sk_buff *skbn;
int rc;
if (port->ingress_data_format & RMNET_INGRESS_FORMAT_DEAGGREGATION) { if (port->ingress_data_format & RMNET_INGRESS_FORMAT_DEAGGREGATION) {
while ((skbn = rmnet_map_deaggregate(skb)) != NULL) while ((skbn = rmnet_map_deaggregate(skb)) != NULL)
__rmnet_map_ingress_handler(skbn, port); __rmnet_map_ingress_handler(skbn, port);
consume_skb(skb); consume_skb(skb);
rc = RX_HANDLER_CONSUMED;
} else { } else {
rc = __rmnet_map_ingress_handler(skb, port); __rmnet_map_ingress_handler(skb, port);
} }
return rc;
} }
static int rmnet_map_egress_handler(struct sk_buff *skb, static int rmnet_map_egress_handler(struct sk_buff *skb,
...@@ -149,15 +145,13 @@ static int rmnet_map_egress_handler(struct sk_buff *skb, ...@@ -149,15 +145,13 @@ static int rmnet_map_egress_handler(struct sk_buff *skb,
return RMNET_MAP_SUCCESS; return RMNET_MAP_SUCCESS;
} }
static rx_handler_result_t static void
rmnet_bridge_handler(struct sk_buff *skb, struct net_device *bridge_dev) rmnet_bridge_handler(struct sk_buff *skb, struct net_device *bridge_dev)
{ {
if (bridge_dev) { if (bridge_dev) {
skb->dev = bridge_dev; skb->dev = bridge_dev;
dev_queue_xmit(skb); dev_queue_xmit(skb);
} }
return RX_HANDLER_CONSUMED;
} }
/* Ingress / Egress Entry Points */ /* Ingress / Egress Entry Points */
...@@ -168,13 +162,12 @@ rmnet_bridge_handler(struct sk_buff *skb, struct net_device *bridge_dev) ...@@ -168,13 +162,12 @@ rmnet_bridge_handler(struct sk_buff *skb, struct net_device *bridge_dev)
*/ */
rx_handler_result_t rmnet_rx_handler(struct sk_buff **pskb) rx_handler_result_t rmnet_rx_handler(struct sk_buff **pskb)
{ {
int rc = RX_HANDLER_CONSUMED;
struct sk_buff *skb = *pskb; struct sk_buff *skb = *pskb;
struct rmnet_port *port; struct rmnet_port *port;
struct net_device *dev; struct net_device *dev;
if (!skb) if (!skb)
return RX_HANDLER_CONSUMED; goto done;
dev = skb->dev; dev = skb->dev;
port = rmnet_get_port(dev); port = rmnet_get_port(dev);
...@@ -182,14 +175,15 @@ rx_handler_result_t rmnet_rx_handler(struct sk_buff **pskb) ...@@ -182,14 +175,15 @@ rx_handler_result_t rmnet_rx_handler(struct sk_buff **pskb)
switch (port->rmnet_mode) { switch (port->rmnet_mode) {
case RMNET_EPMODE_VND: case RMNET_EPMODE_VND:
if (port->ingress_data_format & RMNET_INGRESS_FORMAT_MAP) if (port->ingress_data_format & RMNET_INGRESS_FORMAT_MAP)
rc = rmnet_map_ingress_handler(skb, port); rmnet_map_ingress_handler(skb, port);
break; break;
case RMNET_EPMODE_BRIDGE: case RMNET_EPMODE_BRIDGE:
rc = rmnet_bridge_handler(skb, port->bridge_ep); rmnet_bridge_handler(skb, port->bridge_ep);
break; break;
} }
return rc; done:
return RX_HANDLER_CONSUMED;
} }
/* Modifies packet as per logical endpoint configuration and egress data format /* Modifies packet as per logical endpoint configuration and egress data format
......
...@@ -80,7 +80,6 @@ u8 rmnet_map_demultiplex(struct sk_buff *skb); ...@@ -80,7 +80,6 @@ u8 rmnet_map_demultiplex(struct sk_buff *skb);
struct sk_buff *rmnet_map_deaggregate(struct sk_buff *skb); struct sk_buff *rmnet_map_deaggregate(struct sk_buff *skb);
struct rmnet_map_header *rmnet_map_add_map_header(struct sk_buff *skb, struct rmnet_map_header *rmnet_map_add_map_header(struct sk_buff *skb,
int hdrlen, int pad); int hdrlen, int pad);
rx_handler_result_t rmnet_map_command(struct sk_buff *skb, void rmnet_map_command(struct sk_buff *skb, struct rmnet_port *port);
struct rmnet_port *port);
#endif /* _RMNET_MAP_H_ */ #endif /* _RMNET_MAP_H_ */
...@@ -76,8 +76,7 @@ static void rmnet_map_send_ack(struct sk_buff *skb, ...@@ -76,8 +76,7 @@ static void rmnet_map_send_ack(struct sk_buff *skb,
/* Process MAP command frame and send N/ACK message as appropriate. Message cmd /* Process MAP command frame and send N/ACK message as appropriate. Message cmd
* name is decoded here and appropriate handler is called. * name is decoded here and appropriate handler is called.
*/ */
rx_handler_result_t rmnet_map_command(struct sk_buff *skb, void rmnet_map_command(struct sk_buff *skb, struct rmnet_port *port)
struct rmnet_port *port)
{ {
struct rmnet_map_control_command *cmd; struct rmnet_map_control_command *cmd;
unsigned char command_name; unsigned char command_name;
...@@ -102,5 +101,4 @@ rx_handler_result_t rmnet_map_command(struct sk_buff *skb, ...@@ -102,5 +101,4 @@ rx_handler_result_t rmnet_map_command(struct sk_buff *skb,
} }
if (rc == RMNET_MAP_COMMAND_ACK) if (rc == RMNET_MAP_COMMAND_ACK)
rmnet_map_send_ack(skb, rc); rmnet_map_send_ack(skb, rc);
return RX_HANDLER_CONSUMED;
} }
...@@ -27,14 +27,28 @@ ...@@ -27,14 +27,28 @@
void rmnet_vnd_rx_fixup(struct sk_buff *skb, struct net_device *dev) void rmnet_vnd_rx_fixup(struct sk_buff *skb, struct net_device *dev)
{ {
dev->stats.rx_packets++; struct rmnet_priv *priv = netdev_priv(dev);
dev->stats.rx_bytes += skb->len; struct rmnet_pcpu_stats *pcpu_ptr;
pcpu_ptr = this_cpu_ptr(priv->pcpu_stats);
u64_stats_update_begin(&pcpu_ptr->syncp);
pcpu_ptr->stats.rx_pkts++;
pcpu_ptr->stats.rx_bytes += skb->len;
u64_stats_update_end(&pcpu_ptr->syncp);
} }
void rmnet_vnd_tx_fixup(struct sk_buff *skb, struct net_device *dev) void rmnet_vnd_tx_fixup(struct sk_buff *skb, struct net_device *dev)
{ {
dev->stats.tx_packets++; struct rmnet_priv *priv = netdev_priv(dev);
dev->stats.tx_bytes += skb->len; struct rmnet_pcpu_stats *pcpu_ptr;
pcpu_ptr = this_cpu_ptr(priv->pcpu_stats);
u64_stats_update_begin(&pcpu_ptr->syncp);
pcpu_ptr->stats.tx_pkts++;
pcpu_ptr->stats.tx_bytes += skb->len;
u64_stats_update_end(&pcpu_ptr->syncp);
} }
/* Network Device Operations */ /* Network Device Operations */
...@@ -48,7 +62,7 @@ static netdev_tx_t rmnet_vnd_start_xmit(struct sk_buff *skb, ...@@ -48,7 +62,7 @@ static netdev_tx_t rmnet_vnd_start_xmit(struct sk_buff *skb,
if (priv->real_dev) { if (priv->real_dev) {
rmnet_egress_handler(skb); rmnet_egress_handler(skb);
} else { } else {
dev->stats.tx_dropped++; this_cpu_inc(priv->pcpu_stats->stats.tx_drops);
kfree_skb(skb); kfree_skb(skb);
} }
return NETDEV_TX_OK; return NETDEV_TX_OK;
...@@ -70,12 +84,72 @@ static int rmnet_vnd_get_iflink(const struct net_device *dev) ...@@ -70,12 +84,72 @@ static int rmnet_vnd_get_iflink(const struct net_device *dev)
return priv->real_dev->ifindex; return priv->real_dev->ifindex;
} }
static int rmnet_vnd_init(struct net_device *dev)
{
struct rmnet_priv *priv = netdev_priv(dev);
int err;
priv->pcpu_stats = alloc_percpu(struct rmnet_pcpu_stats);
if (!priv->pcpu_stats)
return -ENOMEM;
err = gro_cells_init(&priv->gro_cells, dev);
if (err) {
free_percpu(priv->pcpu_stats);
return err;
}
return 0;
}
static void rmnet_vnd_uninit(struct net_device *dev)
{
struct rmnet_priv *priv = netdev_priv(dev);
gro_cells_destroy(&priv->gro_cells);
free_percpu(priv->pcpu_stats);
}
static void rmnet_get_stats64(struct net_device *dev,
struct rtnl_link_stats64 *s)
{
struct rmnet_priv *priv = netdev_priv(dev);
struct rmnet_vnd_stats total_stats;
struct rmnet_pcpu_stats *pcpu_ptr;
unsigned int cpu, start;
memset(&total_stats, 0, sizeof(struct rmnet_vnd_stats));
for_each_possible_cpu(cpu) {
pcpu_ptr = this_cpu_ptr(priv->pcpu_stats);
do {
start = u64_stats_fetch_begin_irq(&pcpu_ptr->syncp);
total_stats.rx_pkts += pcpu_ptr->stats.rx_pkts;
total_stats.rx_bytes += pcpu_ptr->stats.rx_bytes;
total_stats.tx_pkts += pcpu_ptr->stats.tx_pkts;
total_stats.tx_bytes += pcpu_ptr->stats.tx_bytes;
} while (u64_stats_fetch_retry_irq(&pcpu_ptr->syncp, start));
total_stats.tx_drops += pcpu_ptr->stats.tx_drops;
}
s->rx_packets = total_stats.rx_pkts;
s->rx_bytes = total_stats.rx_bytes;
s->tx_packets = total_stats.tx_pkts;
s->tx_bytes = total_stats.tx_bytes;
s->tx_dropped = total_stats.tx_drops;
}
static const struct net_device_ops rmnet_vnd_ops = { static const struct net_device_ops rmnet_vnd_ops = {
.ndo_start_xmit = rmnet_vnd_start_xmit, .ndo_start_xmit = rmnet_vnd_start_xmit,
.ndo_change_mtu = rmnet_vnd_change_mtu, .ndo_change_mtu = rmnet_vnd_change_mtu,
.ndo_get_iflink = rmnet_vnd_get_iflink, .ndo_get_iflink = rmnet_vnd_get_iflink,
.ndo_add_slave = rmnet_add_bridge, .ndo_add_slave = rmnet_add_bridge,
.ndo_del_slave = rmnet_del_bridge, .ndo_del_slave = rmnet_del_bridge,
.ndo_init = rmnet_vnd_init,
.ndo_uninit = rmnet_vnd_uninit,
.ndo_get_stats64 = rmnet_get_stats64,
}; };
/* Called by kernel whenever a new rmnet<n> device is created. Sets MTU, /* Called by kernel whenever a new rmnet<n> device is created. Sets MTU,
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册