提交 eca1d5c2 编写于 作者: J Julian Wiedmann 提交者: David S. Miller

s390/qeth: move cast type selection into fill_header()

The cast type currently gets selected in .ndo_start_xmit, and is then
piped through several layers until it's stored into the HW header.
Push the selection down into qeth_l?_fill_header() to (1) reduce the
number of xmit-wide parameters, and (2) merge the two route validation
checks into just one.
Signed-off-by: NJulian Wiedmann <jwi@linux.ibm.com>
Signed-off-by: NDavid S. Miller <davem@davemloft.net>
上级 980f4568
...@@ -1040,11 +1040,10 @@ int qeth_stop(struct net_device *dev); ...@@ -1040,11 +1040,10 @@ int qeth_stop(struct net_device *dev);
int qeth_vm_request_mac(struct qeth_card *card); int qeth_vm_request_mac(struct qeth_card *card);
int qeth_xmit(struct qeth_card *card, struct sk_buff *skb, int qeth_xmit(struct qeth_card *card, struct sk_buff *skb,
struct qeth_qdio_out_q *queue, int ipv, int cast_type, struct qeth_qdio_out_q *queue, int ipv,
void (*fill_header)(struct qeth_qdio_out_q *queue, void (*fill_header)(struct qeth_qdio_out_q *queue,
struct qeth_hdr *hdr, struct sk_buff *skb, struct qeth_hdr *hdr, struct sk_buff *skb,
int ipv, int cast_type, int ipv, unsigned int data_len));
unsigned int data_len));
/* exports for OSN */ /* exports for OSN */
int qeth_osn_assist(struct net_device *, void *, int); int qeth_osn_assist(struct net_device *, void *, int);
......
...@@ -3908,11 +3908,10 @@ static void qeth_fill_tso_ext(struct qeth_hdr_tso *hdr, ...@@ -3908,11 +3908,10 @@ static void qeth_fill_tso_ext(struct qeth_hdr_tso *hdr,
} }
int qeth_xmit(struct qeth_card *card, struct sk_buff *skb, int qeth_xmit(struct qeth_card *card, struct sk_buff *skb,
struct qeth_qdio_out_q *queue, int ipv, int cast_type, struct qeth_qdio_out_q *queue, int ipv,
void (*fill_header)(struct qeth_qdio_out_q *queue, void (*fill_header)(struct qeth_qdio_out_q *queue,
struct qeth_hdr *hdr, struct sk_buff *skb, struct qeth_hdr *hdr, struct sk_buff *skb,
int ipv, int cast_type, int ipv, unsigned int data_len))
unsigned int data_len))
{ {
unsigned int proto_len, hw_hdr_len; unsigned int proto_len, hw_hdr_len;
unsigned int frame_len = skb->len; unsigned int frame_len = skb->len;
...@@ -3946,7 +3945,7 @@ int qeth_xmit(struct qeth_card *card, struct sk_buff *skb, ...@@ -3946,7 +3945,7 @@ int qeth_xmit(struct qeth_card *card, struct sk_buff *skb,
data_offset = push_len + proto_len; data_offset = push_len + proto_len;
} }
memset(hdr, 0, hw_hdr_len); memset(hdr, 0, hw_hdr_len);
fill_header(queue, hdr, skb, ipv, cast_type, frame_len); fill_header(queue, hdr, skb, ipv, frame_len);
if (is_tso) if (is_tso)
qeth_fill_tso_ext((struct qeth_hdr_tso *) hdr, qeth_fill_tso_ext((struct qeth_hdr_tso *) hdr,
frame_len - proto_len, skb, proto_len); frame_len - proto_len, skb, proto_len);
......
...@@ -164,8 +164,9 @@ static void qeth_l2_drain_rx_mode_cache(struct qeth_card *card) ...@@ -164,8 +164,9 @@ static void qeth_l2_drain_rx_mode_cache(struct qeth_card *card)
static void qeth_l2_fill_header(struct qeth_qdio_out_q *queue, static void qeth_l2_fill_header(struct qeth_qdio_out_q *queue,
struct qeth_hdr *hdr, struct sk_buff *skb, struct qeth_hdr *hdr, struct sk_buff *skb,
int ipv, int cast_type, unsigned int data_len) int ipv, unsigned int data_len)
{ {
int cast_type = qeth_get_ether_cast_type(skb);
struct vlan_ethhdr *veth = vlan_eth_hdr(skb); struct vlan_ethhdr *veth = vlan_eth_hdr(skb);
hdr->hdr.l2.pkt_length = data_len; hdr->hdr.l2.pkt_length = data_len;
...@@ -598,7 +599,6 @@ static netdev_tx_t qeth_l2_hard_start_xmit(struct sk_buff *skb, ...@@ -598,7 +599,6 @@ static netdev_tx_t qeth_l2_hard_start_xmit(struct sk_buff *skb,
rc = qeth_l2_xmit_osn(card, skb, queue); rc = qeth_l2_xmit_osn(card, skb, queue);
else else
rc = qeth_xmit(card, skb, queue, qeth_get_ip_version(skb), rc = qeth_xmit(card, skb, queue, qeth_get_ip_version(skb),
qeth_get_ether_cast_type(skb),
qeth_l2_fill_header); qeth_l2_fill_header);
if (!rc) { if (!rc) {
......
...@@ -1939,12 +1939,13 @@ static u8 qeth_l3_cast_type_to_flag(int cast_type) ...@@ -1939,12 +1939,13 @@ static u8 qeth_l3_cast_type_to_flag(int cast_type)
static void qeth_l3_fill_header(struct qeth_qdio_out_q *queue, static void qeth_l3_fill_header(struct qeth_qdio_out_q *queue,
struct qeth_hdr *hdr, struct sk_buff *skb, struct qeth_hdr *hdr, struct sk_buff *skb,
int ipv, int cast_type, unsigned int data_len) int ipv, unsigned int data_len)
{ {
struct qeth_hdr_layer3 *l3_hdr = &hdr->hdr.l3; struct qeth_hdr_layer3 *l3_hdr = &hdr->hdr.l3;
struct vlan_ethhdr *veth = vlan_eth_hdr(skb); struct vlan_ethhdr *veth = vlan_eth_hdr(skb);
struct qeth_card *card = queue->card; struct qeth_card *card = queue->card;
struct dst_entry *dst; struct dst_entry *dst;
int cast_type;
hdr->hdr.l3.length = data_len; hdr->hdr.l3.length = data_len;
...@@ -1981,25 +1982,22 @@ static void qeth_l3_fill_header(struct qeth_qdio_out_q *queue, ...@@ -1981,25 +1982,22 @@ static void qeth_l3_fill_header(struct qeth_qdio_out_q *queue,
hdr->hdr.l3.vlan_id = ntohs(veth->h_vlan_TCI); hdr->hdr.l3.vlan_id = ntohs(veth->h_vlan_TCI);
} }
l3_hdr->flags = qeth_l3_cast_type_to_flag(cast_type);
/* OSA only: */
if (!ipv) {
l3_hdr->flags |= QETH_HDR_PASSTHRU;
return;
}
rcu_read_lock(); rcu_read_lock();
dst = qeth_dst_check_rcu(skb, ipv); dst = qeth_dst_check_rcu(skb, ipv);
if (IS_IQD(card) && skb_get_queue_mapping(skb) != QETH_IQD_MCAST_TXQ)
cast_type = RTN_UNICAST;
else
cast_type = qeth_l3_get_cast_type_rcu(skb, dst, ipv);
l3_hdr->flags |= qeth_l3_cast_type_to_flag(cast_type);
if (ipv == 4) { if (ipv == 4) {
struct rtable *rt = (struct rtable *) dst; struct rtable *rt = (struct rtable *) dst;
*((__be32 *) &hdr->hdr.l3.next_hop.ipv4.addr) = (rt) ? *((__be32 *) &hdr->hdr.l3.next_hop.ipv4.addr) = (rt) ?
rt_nexthop(rt, ip_hdr(skb)->daddr) : rt_nexthop(rt, ip_hdr(skb)->daddr) :
ip_hdr(skb)->daddr; ip_hdr(skb)->daddr;
} else { } else if (ipv == 6) {
/* IPv6 */
struct rt6_info *rt = (struct rt6_info *) dst; struct rt6_info *rt = (struct rt6_info *) dst;
if (rt && !ipv6_addr_any(&rt->rt6i_gateway)) if (rt && !ipv6_addr_any(&rt->rt6i_gateway))
...@@ -2010,6 +2008,9 @@ static void qeth_l3_fill_header(struct qeth_qdio_out_q *queue, ...@@ -2010,6 +2008,9 @@ static void qeth_l3_fill_header(struct qeth_qdio_out_q *queue,
hdr->hdr.l3.flags |= QETH_HDR_IPV6; hdr->hdr.l3.flags |= QETH_HDR_IPV6;
if (!IS_IQD(card)) if (!IS_IQD(card))
hdr->hdr.l3.flags |= QETH_HDR_PASSTHRU; hdr->hdr.l3.flags |= QETH_HDR_PASSTHRU;
} else {
/* OSA only: */
l3_hdr->flags |= QETH_HDR_PASSTHRU;
} }
rcu_read_unlock(); rcu_read_unlock();
} }
...@@ -2029,7 +2030,7 @@ static void qeth_l3_fixup_headers(struct sk_buff *skb) ...@@ -2029,7 +2030,7 @@ static void qeth_l3_fixup_headers(struct sk_buff *skb)
} }
static int qeth_l3_xmit(struct qeth_card *card, struct sk_buff *skb, static int qeth_l3_xmit(struct qeth_card *card, struct sk_buff *skb,
struct qeth_qdio_out_q *queue, int ipv, int cast_type) struct qeth_qdio_out_q *queue, int ipv)
{ {
unsigned int hw_hdr_len; unsigned int hw_hdr_len;
int rc; int rc;
...@@ -2043,7 +2044,7 @@ static int qeth_l3_xmit(struct qeth_card *card, struct sk_buff *skb, ...@@ -2043,7 +2044,7 @@ static int qeth_l3_xmit(struct qeth_card *card, struct sk_buff *skb,
skb_pull(skb, ETH_HLEN); skb_pull(skb, ETH_HLEN);
qeth_l3_fixup_headers(skb); qeth_l3_fixup_headers(skb);
return qeth_xmit(card, skb, queue, ipv, cast_type, qeth_l3_fill_header); return qeth_xmit(card, skb, queue, ipv, qeth_l3_fill_header);
} }
static netdev_tx_t qeth_l3_hard_start_xmit(struct sk_buff *skb, static netdev_tx_t qeth_l3_hard_start_xmit(struct sk_buff *skb,
...@@ -2054,7 +2055,7 @@ static netdev_tx_t qeth_l3_hard_start_xmit(struct sk_buff *skb, ...@@ -2054,7 +2055,7 @@ static netdev_tx_t qeth_l3_hard_start_xmit(struct sk_buff *skb,
int ipv = qeth_get_ip_version(skb); int ipv = qeth_get_ip_version(skb);
struct qeth_qdio_out_q *queue; struct qeth_qdio_out_q *queue;
int tx_bytes = skb->len; int tx_bytes = skb->len;
int cast_type, rc; int rc;
if (IS_IQD(card)) { if (IS_IQD(card)) {
queue = card->qdio.out_qs[qeth_iqd_translate_txq(dev, txq)]; queue = card->qdio.out_qs[qeth_iqd_translate_txq(dev, txq)];
...@@ -2065,24 +2066,18 @@ static netdev_tx_t qeth_l3_hard_start_xmit(struct sk_buff *skb, ...@@ -2065,24 +2066,18 @@ static netdev_tx_t qeth_l3_hard_start_xmit(struct sk_buff *skb,
(card->options.cq == QETH_CQ_ENABLED && (card->options.cq == QETH_CQ_ENABLED &&
skb->protocol != htons(ETH_P_AF_IUCV))) skb->protocol != htons(ETH_P_AF_IUCV)))
goto tx_drop; goto tx_drop;
if (txq == QETH_IQD_MCAST_TXQ)
cast_type = qeth_l3_get_cast_type(skb);
else
cast_type = RTN_UNICAST;
} else { } else {
queue = card->qdio.out_qs[txq]; queue = card->qdio.out_qs[txq];
cast_type = qeth_l3_get_cast_type(skb);
} }
if (cast_type == RTN_BROADCAST && !card->info.broadcast_capable) if (!(dev->flags & IFF_BROADCAST) &&
qeth_l3_get_cast_type(skb) == RTN_BROADCAST)
goto tx_drop; goto tx_drop;
if (ipv == 4 || IS_IQD(card)) if (ipv == 4 || IS_IQD(card))
rc = qeth_l3_xmit(card, skb, queue, ipv, cast_type); rc = qeth_l3_xmit(card, skb, queue, ipv);
else else
rc = qeth_xmit(card, skb, queue, ipv, cast_type, rc = qeth_xmit(card, skb, queue, ipv, qeth_l3_fill_header);
qeth_l3_fill_header);
if (!rc) { if (!rc) {
QETH_TXQ_STAT_INC(queue, tx_packets); QETH_TXQ_STAT_INC(queue, tx_packets);
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册