提交 18604c54 编写于 作者: J Jan-Bernd Themann 提交者: Jeff Garzik

ehea: NAPI multi queue TX/RX path for SMP

This patch provides a functionality that allows parallel
RX processing on multiple RX queues by using dummy netdevices.
Signed-off-by: NJan-Bernd Themann <themann@de.ibm.com>
Signed-off-by: NJeff Garzik <jeff@garzik.org>
上级 1acf2318
...@@ -39,7 +39,7 @@ ...@@ -39,7 +39,7 @@
#include <asm/io.h> #include <asm/io.h>
#define DRV_NAME "ehea" #define DRV_NAME "ehea"
#define DRV_VERSION "EHEA_0048" #define DRV_VERSION "EHEA_0052"
#define EHEA_MSG_DEFAULT (NETIF_MSG_LINK | NETIF_MSG_TIMER \ #define EHEA_MSG_DEFAULT (NETIF_MSG_LINK | NETIF_MSG_TIMER \
| NETIF_MSG_RX_ERR | NETIF_MSG_TX_ERR) | NETIF_MSG_RX_ERR | NETIF_MSG_TX_ERR)
...@@ -78,8 +78,6 @@ ...@@ -78,8 +78,6 @@
#define EHEA_RQ2_PKT_SIZE 1522 #define EHEA_RQ2_PKT_SIZE 1522
#define EHEA_L_PKT_SIZE 256 /* low latency */ #define EHEA_L_PKT_SIZE 256 /* low latency */
#define EHEA_POLL_MAX_RWQE 1000
/* Send completion signaling */ /* Send completion signaling */
#define EHEA_SIG_IV_LONG 1 #define EHEA_SIG_IV_LONG 1
...@@ -357,8 +355,8 @@ struct ehea_port_res { ...@@ -357,8 +355,8 @@ struct ehea_port_res {
struct ehea_qp *qp; struct ehea_qp *qp;
struct ehea_cq *send_cq; struct ehea_cq *send_cq;
struct ehea_cq *recv_cq; struct ehea_cq *recv_cq;
struct ehea_eq *send_eq; struct ehea_eq *eq;
struct ehea_eq *recv_eq; struct net_device *d_netdev;
spinlock_t send_lock; spinlock_t send_lock;
struct ehea_q_skb_arr rq1_skba; struct ehea_q_skb_arr rq1_skba;
struct ehea_q_skb_arr rq2_skba; struct ehea_q_skb_arr rq2_skba;
...@@ -372,7 +370,6 @@ struct ehea_port_res { ...@@ -372,7 +370,6 @@ struct ehea_port_res {
int swqe_count; int swqe_count;
u32 swqe_id_counter; u32 swqe_id_counter;
u64 tx_packets; u64 tx_packets;
struct tasklet_struct send_comp_task;
spinlock_t recv_lock; spinlock_t recv_lock;
struct port_state p_state; struct port_state p_state;
u64 rx_packets; u64 rx_packets;
...@@ -416,7 +413,9 @@ struct ehea_port { ...@@ -416,7 +413,9 @@ struct ehea_port {
char int_aff_name[EHEA_IRQ_NAME_SIZE]; char int_aff_name[EHEA_IRQ_NAME_SIZE];
int allmulti; /* Indicates IFF_ALLMULTI state */ int allmulti; /* Indicates IFF_ALLMULTI state */
int promisc; /* Indicates IFF_PROMISC state */ int promisc; /* Indicates IFF_PROMISC state */
int num_tx_qps;
int num_add_tx_qps; int num_add_tx_qps;
int num_mcs;
int resets; int resets;
u64 mac_addr; u64 mac_addr;
u32 logical_port_id; u32 logical_port_id;
......
...@@ -51,13 +51,18 @@ static int rq1_entries = EHEA_DEF_ENTRIES_RQ1; ...@@ -51,13 +51,18 @@ static int rq1_entries = EHEA_DEF_ENTRIES_RQ1;
static int rq2_entries = EHEA_DEF_ENTRIES_RQ2; static int rq2_entries = EHEA_DEF_ENTRIES_RQ2;
static int rq3_entries = EHEA_DEF_ENTRIES_RQ3; static int rq3_entries = EHEA_DEF_ENTRIES_RQ3;
static int sq_entries = EHEA_DEF_ENTRIES_SQ; static int sq_entries = EHEA_DEF_ENTRIES_SQ;
static int use_mcs = 0;
static int num_tx_qps = EHEA_NUM_TX_QP;
module_param(msg_level, int, 0); module_param(msg_level, int, 0);
module_param(rq1_entries, int, 0); module_param(rq1_entries, int, 0);
module_param(rq2_entries, int, 0); module_param(rq2_entries, int, 0);
module_param(rq3_entries, int, 0); module_param(rq3_entries, int, 0);
module_param(sq_entries, int, 0); module_param(sq_entries, int, 0);
module_param(use_mcs, int, 0);
module_param(num_tx_qps, int, 0);
MODULE_PARM_DESC(num_tx_qps, "Number of TX-QPS");
MODULE_PARM_DESC(msg_level, "msg_level"); MODULE_PARM_DESC(msg_level, "msg_level");
MODULE_PARM_DESC(rq3_entries, "Number of entries for Receive Queue 3 " MODULE_PARM_DESC(rq3_entries, "Number of entries for Receive Queue 3 "
"[2^x - 1], x = [6..14]. Default = " "[2^x - 1], x = [6..14]. Default = "
...@@ -71,6 +76,7 @@ MODULE_PARM_DESC(rq1_entries, "Number of entries for Receive Queue 1 " ...@@ -71,6 +76,7 @@ MODULE_PARM_DESC(rq1_entries, "Number of entries for Receive Queue 1 "
MODULE_PARM_DESC(sq_entries, " Number of entries for the Send Queue " MODULE_PARM_DESC(sq_entries, " Number of entries for the Send Queue "
"[2^x - 1], x = [6..14]. Default = " "[2^x - 1], x = [6..14]. Default = "
__MODULE_STRING(EHEA_DEF_ENTRIES_SQ) ")"); __MODULE_STRING(EHEA_DEF_ENTRIES_SQ) ")");
MODULE_PARM_DESC(use_mcs, " 0:NAPI, 1:Multiple receive queues, Default = 1 ");
void ehea_dump(void *adr, int len, char *msg) { void ehea_dump(void *adr, int len, char *msg) {
int x; int x;
...@@ -197,7 +203,7 @@ static int ehea_refill_rq_def(struct ehea_port_res *pr, ...@@ -197,7 +203,7 @@ static int ehea_refill_rq_def(struct ehea_port_res *pr,
struct sk_buff *skb = netdev_alloc_skb(dev, packet_size); struct sk_buff *skb = netdev_alloc_skb(dev, packet_size);
if (!skb) { if (!skb) {
ehea_error("%s: no mem for skb/%d wqes filled", ehea_error("%s: no mem for skb/%d wqes filled",
dev->name, i); pr->port->netdev->name, i);
q_skba->os_skbs = fill_wqes - i; q_skba->os_skbs = fill_wqes - i;
ret = -ENOMEM; ret = -ENOMEM;
break; break;
...@@ -345,10 +351,11 @@ static int ehea_treat_poll_error(struct ehea_port_res *pr, int rq, ...@@ -345,10 +351,11 @@ static int ehea_treat_poll_error(struct ehea_port_res *pr, int rq,
return 0; return 0;
} }
static int ehea_poll(struct net_device *dev, int *budget) static struct ehea_cqe *ehea_proc_rwqes(struct net_device *dev,
struct ehea_port_res *pr,
int *budget)
{ {
struct ehea_port *port = netdev_priv(dev); struct ehea_port *port = pr->port;
struct ehea_port_res *pr = &port->port_res[0];
struct ehea_qp *qp = pr->qp; struct ehea_qp *qp = pr->qp;
struct ehea_cqe *cqe; struct ehea_cqe *cqe;
struct sk_buff *skb; struct sk_buff *skb;
...@@ -359,14 +366,12 @@ static int ehea_poll(struct net_device *dev, int *budget) ...@@ -359,14 +366,12 @@ static int ehea_poll(struct net_device *dev, int *budget)
int skb_arr_rq2_len = pr->rq2_skba.len; int skb_arr_rq2_len = pr->rq2_skba.len;
int skb_arr_rq3_len = pr->rq3_skba.len; int skb_arr_rq3_len = pr->rq3_skba.len;
int processed, processed_rq1, processed_rq2, processed_rq3; int processed, processed_rq1, processed_rq2, processed_rq3;
int wqe_index, last_wqe_index, rq, intreq, my_quota, port_reset; int wqe_index, last_wqe_index, rq, my_quota, port_reset;
processed = processed_rq1 = processed_rq2 = processed_rq3 = 0; processed = processed_rq1 = processed_rq2 = processed_rq3 = 0;
last_wqe_index = 0; last_wqe_index = 0;
my_quota = min(*budget, dev->quota); my_quota = min(*budget, dev->quota);
my_quota = min(my_quota, EHEA_POLL_MAX_RWQE);
/* rq0 is low latency RQ */
cqe = ehea_poll_rq1(qp, &wqe_index); cqe = ehea_poll_rq1(qp, &wqe_index);
while ((my_quota > 0) && cqe) { while ((my_quota > 0) && cqe) {
ehea_inc_rq1(qp); ehea_inc_rq1(qp);
...@@ -386,7 +391,8 @@ static int ehea_poll(struct net_device *dev, int *budget) ...@@ -386,7 +391,8 @@ static int ehea_poll(struct net_device *dev, int *budget)
if (unlikely(!skb)) { if (unlikely(!skb)) {
if (netif_msg_rx_err(port)) if (netif_msg_rx_err(port))
ehea_error("LL rq1: skb=NULL"); ehea_error("LL rq1: skb=NULL");
skb = netdev_alloc_skb(dev,
skb = netdev_alloc_skb(port->netdev,
EHEA_L_PKT_SIZE); EHEA_L_PKT_SIZE);
if (!skb) if (!skb)
break; break;
...@@ -402,7 +408,7 @@ static int ehea_poll(struct net_device *dev, int *budget) ...@@ -402,7 +408,7 @@ static int ehea_poll(struct net_device *dev, int *budget)
ehea_error("rq2: skb=NULL"); ehea_error("rq2: skb=NULL");
break; break;
} }
ehea_fill_skb(dev, skb, cqe); ehea_fill_skb(port->netdev, skb, cqe);
processed_rq2++; processed_rq2++;
} else { /* RQ3 */ } else { /* RQ3 */
skb = get_skb_by_index(skb_arr_rq3, skb = get_skb_by_index(skb_arr_rq3,
...@@ -412,7 +418,7 @@ static int ehea_poll(struct net_device *dev, int *budget) ...@@ -412,7 +418,7 @@ static int ehea_poll(struct net_device *dev, int *budget)
ehea_error("rq3: skb=NULL"); ehea_error("rq3: skb=NULL");
break; break;
} }
ehea_fill_skb(dev, skb, cqe); ehea_fill_skb(port->netdev, skb, cqe);
processed_rq3++; processed_rq3++;
} }
...@@ -421,8 +427,7 @@ static int ehea_poll(struct net_device *dev, int *budget) ...@@ -421,8 +427,7 @@ static int ehea_poll(struct net_device *dev, int *budget)
cqe->vlan_tag); cqe->vlan_tag);
else else
netif_receive_skb(skb); netif_receive_skb(skb);
} else {
} else { /* Error occured */
pr->p_state.poll_receive_errors++; pr->p_state.poll_receive_errors++;
port_reset = ehea_treat_poll_error(pr, rq, cqe, port_reset = ehea_treat_poll_error(pr, rq, cqe,
&processed_rq2, &processed_rq2,
...@@ -433,32 +438,18 @@ static int ehea_poll(struct net_device *dev, int *budget) ...@@ -433,32 +438,18 @@ static int ehea_poll(struct net_device *dev, int *budget)
cqe = ehea_poll_rq1(qp, &wqe_index); cqe = ehea_poll_rq1(qp, &wqe_index);
} }
dev->quota -= processed;
*budget -= processed;
pr->p_state.ehea_poll += 1;
pr->rx_packets += processed; pr->rx_packets += processed;
*budget -= processed;
ehea_refill_rq1(pr, last_wqe_index, processed_rq1); ehea_refill_rq1(pr, last_wqe_index, processed_rq1);
ehea_refill_rq2(pr, processed_rq2); ehea_refill_rq2(pr, processed_rq2);
ehea_refill_rq3(pr, processed_rq3); ehea_refill_rq3(pr, processed_rq3);
intreq = ((pr->p_state.ehea_poll & 0xF) == 0xF); cqe = ehea_poll_rq1(qp, &wqe_index);
return cqe;
if (!cqe || intreq) {
netif_rx_complete(dev);
ehea_reset_cq_ep(pr->recv_cq);
ehea_reset_cq_n1(pr->recv_cq);
cqe = hw_qeit_get_valid(&qp->hw_rqueue1);
if (!cqe || intreq)
return 0;
if (!netif_rx_reschedule(dev, my_quota))
return 0;
}
return 1;
} }
void free_sent_skbs(struct ehea_cqe *cqe, struct ehea_port_res *pr) static void ehea_free_sent_skbs(struct ehea_cqe *cqe, struct ehea_port_res *pr)
{ {
struct sk_buff *skb; struct sk_buff *skb;
int index, max_index_mask, i; int index, max_index_mask, i;
...@@ -479,26 +470,19 @@ void free_sent_skbs(struct ehea_cqe *cqe, struct ehea_port_res *pr) ...@@ -479,26 +470,19 @@ void free_sent_skbs(struct ehea_cqe *cqe, struct ehea_port_res *pr)
} }
} }
#define MAX_SENDCOMP_QUOTA 400 static struct ehea_cqe *ehea_proc_cqes(struct ehea_port_res *pr, int my_quota)
void ehea_send_irq_tasklet(unsigned long data)
{ {
struct ehea_port_res *pr = (struct ehea_port_res*)data;
struct ehea_cq *send_cq = pr->send_cq; struct ehea_cq *send_cq = pr->send_cq;
struct ehea_cqe *cqe; struct ehea_cqe *cqe;
int quota = MAX_SENDCOMP_QUOTA; int quota = my_quota;
int cqe_counter = 0; int cqe_counter = 0;
int swqe_av = 0; int swqe_av = 0;
unsigned long flags; unsigned long flags;
do { cqe = ehea_poll_cq(send_cq);
cqe = ehea_poll_cq(send_cq); while(cqe && (quota > 0)) {
if (!cqe) { ehea_inc_cq(send_cq);
ehea_reset_cq_ep(send_cq);
ehea_reset_cq_n1(send_cq);
cqe = ehea_poll_cq(send_cq);
if (!cqe)
break;
}
cqe_counter++; cqe_counter++;
rmb(); rmb();
if (cqe->status & EHEA_CQE_STAT_ERR_MASK) { if (cqe->status & EHEA_CQE_STAT_ERR_MASK) {
...@@ -515,16 +499,19 @@ void ehea_send_irq_tasklet(unsigned long data) ...@@ -515,16 +499,19 @@ void ehea_send_irq_tasklet(unsigned long data)
if (likely(EHEA_BMASK_GET(EHEA_WR_ID_TYPE, cqe->wr_id) if (likely(EHEA_BMASK_GET(EHEA_WR_ID_TYPE, cqe->wr_id)
== EHEA_SWQE2_TYPE)) == EHEA_SWQE2_TYPE))
free_sent_skbs(cqe, pr); ehea_free_sent_skbs(cqe, pr);
swqe_av += EHEA_BMASK_GET(EHEA_WR_ID_REFILL, cqe->wr_id); swqe_av += EHEA_BMASK_GET(EHEA_WR_ID_REFILL, cqe->wr_id);
quota--; quota--;
} while (quota > 0);
cqe = ehea_poll_cq(send_cq);
};
ehea_update_feca(send_cq, cqe_counter); ehea_update_feca(send_cq, cqe_counter);
atomic_add(swqe_av, &pr->swqe_avail); atomic_add(swqe_av, &pr->swqe_avail);
spin_lock_irqsave(&pr->netif_queue, flags); spin_lock_irqsave(&pr->netif_queue, flags);
if (pr->queue_stopped && (atomic_read(&pr->swqe_avail) if (pr->queue_stopped && (atomic_read(&pr->swqe_avail)
>= pr->swqe_refill_th)) { >= pr->swqe_refill_th)) {
netif_wake_queue(pr->port->netdev); netif_wake_queue(pr->port->netdev);
...@@ -532,22 +519,55 @@ void ehea_send_irq_tasklet(unsigned long data) ...@@ -532,22 +519,55 @@ void ehea_send_irq_tasklet(unsigned long data)
} }
spin_unlock_irqrestore(&pr->netif_queue, flags); spin_unlock_irqrestore(&pr->netif_queue, flags);
if (unlikely(cqe)) return cqe;
tasklet_hi_schedule(&pr->send_comp_task);
} }
static irqreturn_t ehea_send_irq_handler(int irq, void *param) #define EHEA_NAPI_POLL_NUM_BEFORE_IRQ 16
static int ehea_poll(struct net_device *dev, int *budget)
{ {
struct ehea_port_res *pr = param; struct ehea_port_res *pr = dev->priv;
tasklet_hi_schedule(&pr->send_comp_task); struct ehea_cqe *cqe;
return IRQ_HANDLED; struct ehea_cqe *cqe_skb = NULL;
int force_irq, wqe_index;
cqe = ehea_poll_rq1(pr->qp, &wqe_index);
cqe_skb = ehea_poll_cq(pr->send_cq);
force_irq = (pr->poll_counter > EHEA_NAPI_POLL_NUM_BEFORE_IRQ);
if ((!cqe && !cqe_skb) || force_irq) {
pr->poll_counter = 0;
netif_rx_complete(dev);
ehea_reset_cq_ep(pr->recv_cq);
ehea_reset_cq_ep(pr->send_cq);
ehea_reset_cq_n1(pr->recv_cq);
ehea_reset_cq_n1(pr->send_cq);
cqe = ehea_poll_rq1(pr->qp, &wqe_index);
cqe_skb = ehea_poll_cq(pr->send_cq);
if (!cqe && !cqe_skb)
return 0;
if (!netif_rx_reschedule(dev, dev->quota))
return 0;
}
cqe = ehea_proc_rwqes(dev, pr, budget);
cqe_skb = ehea_proc_cqes(pr, 300);
if (cqe || cqe_skb)
pr->poll_counter++;
return 1;
} }
static irqreturn_t ehea_recv_irq_handler(int irq, void *param) static irqreturn_t ehea_recv_irq_handler(int irq, void *param)
{ {
struct ehea_port_res *pr = param; struct ehea_port_res *pr = param;
struct ehea_port *port = pr->port;
netif_rx_schedule(port->netdev); netif_rx_schedule(pr->d_netdev);
return IRQ_HANDLED; return IRQ_HANDLED;
} }
...@@ -650,19 +670,25 @@ int ehea_sense_port_attr(struct ehea_port *port) ...@@ -650,19 +670,25 @@ int ehea_sense_port_attr(struct ehea_port *port)
} }
port->autoneg = 1; port->autoneg = 1;
port->num_mcs = cb0->num_default_qps;
/* Number of default QPs */ /* Number of default QPs */
port->num_def_qps = cb0->num_default_qps; if (use_mcs)
port->num_def_qps = cb0->num_default_qps;
else
port->num_def_qps = 1;
if (!port->num_def_qps) { if (!port->num_def_qps) {
ret = -EINVAL; ret = -EINVAL;
goto out_free; goto out_free;
} }
if (port->num_def_qps >= EHEA_NUM_TX_QP) port->num_tx_qps = num_tx_qps;
if (port->num_def_qps >= port->num_tx_qps)
port->num_add_tx_qps = 0; port->num_add_tx_qps = 0;
else else
port->num_add_tx_qps = EHEA_NUM_TX_QP - port->num_def_qps; port->num_add_tx_qps = port->num_tx_qps - port->num_def_qps;
ret = 0; ret = 0;
out_free: out_free:
...@@ -882,23 +908,6 @@ static int ehea_reg_interrupts(struct net_device *dev) ...@@ -882,23 +908,6 @@ static int ehea_reg_interrupts(struct net_device *dev)
struct ehea_port_res *pr; struct ehea_port_res *pr;
int i, ret; int i, ret;
for (i = 0; i < port->num_def_qps; i++) {
pr = &port->port_res[i];
snprintf(pr->int_recv_name, EHEA_IRQ_NAME_SIZE - 1
, "%s-recv%d", dev->name, i);
ret = ibmebus_request_irq(NULL, pr->recv_eq->attr.ist1,
ehea_recv_irq_handler,
IRQF_DISABLED, pr->int_recv_name, pr);
if (ret) {
ehea_error("failed registering irq for ehea_recv_int:"
"port_res_nr:%d, ist=%X", i,
pr->recv_eq->attr.ist1);
goto out_free_seq;
}
if (netif_msg_ifup(port))
ehea_info("irq_handle 0x%X for funct ehea_recv_int %d "
"registered", pr->recv_eq->attr.ist1, i);
}
snprintf(port->int_aff_name, EHEA_IRQ_NAME_SIZE - 1, "%s-aff", snprintf(port->int_aff_name, EHEA_IRQ_NAME_SIZE - 1, "%s-aff",
dev->name); dev->name);
...@@ -916,41 +925,41 @@ static int ehea_reg_interrupts(struct net_device *dev) ...@@ -916,41 +925,41 @@ static int ehea_reg_interrupts(struct net_device *dev)
ehea_info("irq_handle 0x%X for function qp_aff_irq_handler " ehea_info("irq_handle 0x%X for function qp_aff_irq_handler "
"registered", port->qp_eq->attr.ist1); "registered", port->qp_eq->attr.ist1);
for (i = 0; i < port->num_def_qps + port->num_add_tx_qps; i++) { for (i = 0; i < port->num_def_qps + port->num_add_tx_qps; i++) {
pr = &port->port_res[i]; pr = &port->port_res[i];
snprintf(pr->int_send_name, EHEA_IRQ_NAME_SIZE - 1, snprintf(pr->int_send_name, EHEA_IRQ_NAME_SIZE - 1,
"%s-send%d", dev->name, i); "%s-queue%d", dev->name, i);
ret = ibmebus_request_irq(NULL, pr->send_eq->attr.ist1, ret = ibmebus_request_irq(NULL, pr->eq->attr.ist1,
ehea_send_irq_handler, ehea_recv_irq_handler,
IRQF_DISABLED, pr->int_send_name, IRQF_DISABLED, pr->int_send_name,
pr); pr);
if (ret) { if (ret) {
ehea_error("failed registering irq for ehea_send " ehea_error("failed registering irq for ehea_queue "
"port_res_nr:%d, ist=%X", i, "port_res_nr:%d, ist=%X", i,
pr->send_eq->attr.ist1); pr->eq->attr.ist1);
goto out_free_req; goto out_free_req;
} }
if (netif_msg_ifup(port)) if (netif_msg_ifup(port))
ehea_info("irq_handle 0x%X for function ehea_send_int " ehea_info("irq_handle 0x%X for function ehea_queue_int "
"%d registered", pr->send_eq->attr.ist1, i); "%d registered", pr->eq->attr.ist1, i);
} }
out: out:
return ret; return ret;
out_free_req: out_free_req:
while (--i >= 0) { while (--i >= 0) {
u32 ist = port->port_res[i].send_eq->attr.ist1; u32 ist = port->port_res[i].eq->attr.ist1;
ibmebus_free_irq(NULL, ist, &port->port_res[i]); ibmebus_free_irq(NULL, ist, &port->port_res[i]);
} }
out_free_qpeq: out_free_qpeq:
ibmebus_free_irq(NULL, port->qp_eq->attr.ist1, port); ibmebus_free_irq(NULL, port->qp_eq->attr.ist1, port);
i = port->num_def_qps; i = port->num_def_qps;
out_free_seq:
while (--i >= 0) {
u32 ist = port->port_res[i].recv_eq->attr.ist1;
ibmebus_free_irq(NULL, ist, &port->port_res[i]);
}
goto out; goto out;
} }
static void ehea_free_interrupts(struct net_device *dev) static void ehea_free_interrupts(struct net_device *dev)
...@@ -960,21 +969,13 @@ static void ehea_free_interrupts(struct net_device *dev) ...@@ -960,21 +969,13 @@ static void ehea_free_interrupts(struct net_device *dev)
int i; int i;
/* send */ /* send */
for (i = 0; i < port->num_def_qps + port->num_add_tx_qps; i++) { for (i = 0; i < port->num_def_qps + port->num_add_tx_qps; i++) {
pr = &port->port_res[i]; pr = &port->port_res[i];
ibmebus_free_irq(NULL, pr->send_eq->attr.ist1, pr); ibmebus_free_irq(NULL, pr->eq->attr.ist1, pr);
if (netif_msg_intr(port)) if (netif_msg_intr(port))
ehea_info("free send irq for res %d with handle 0x%X", ehea_info("free send irq for res %d with handle 0x%X",
i, pr->send_eq->attr.ist1); i, pr->eq->attr.ist1);
}
/* receive */
for (i = 0; i < port->num_def_qps; i++) {
pr = &port->port_res[i];
ibmebus_free_irq(NULL, pr->recv_eq->attr.ist1, pr);
if (netif_msg_intr(port))
ehea_info("free recv irq for res %d with handle 0x%X",
i, pr->recv_eq->attr.ist1);
} }
/* associated events */ /* associated events */
...@@ -1003,9 +1004,14 @@ static int ehea_configure_port(struct ehea_port *port) ...@@ -1003,9 +1004,14 @@ static int ehea_configure_port(struct ehea_port *port)
PXLY_RC_VLAN_FILTER) PXLY_RC_VLAN_FILTER)
| EHEA_BMASK_SET(PXLY_RC_JUMBO_FRAME, 1); | EHEA_BMASK_SET(PXLY_RC_JUMBO_FRAME, 1);
for (i = 0; i < port->num_def_qps; i++) for (i = 0; i < port->num_mcs; i++)
cb0->default_qpn_arr[i] = port->port_res[0].qp->init_attr.qp_nr; if (use_mcs)
cb0->default_qpn_arr[i] =
port->port_res[i].qp->init_attr.qp_nr;
else
cb0->default_qpn_arr[i] =
port->port_res[0].qp->init_attr.qp_nr;
if (netif_msg_ifup(port)) if (netif_msg_ifup(port))
ehea_dump(cb0, sizeof(*cb0), "ehea_configure_port"); ehea_dump(cb0, sizeof(*cb0), "ehea_configure_port");
...@@ -1108,20 +1114,14 @@ static int ehea_init_port_res(struct ehea_port *port, struct ehea_port_res *pr, ...@@ -1108,20 +1114,14 @@ static int ehea_init_port_res(struct ehea_port *port, struct ehea_port_res *pr,
spin_lock_init(&pr->xmit_lock); spin_lock_init(&pr->xmit_lock);
spin_lock_init(&pr->netif_queue); spin_lock_init(&pr->netif_queue);
pr->recv_eq = ehea_create_eq(adapter, eq_type, EHEA_MAX_ENTRIES_EQ, 0); pr->eq = ehea_create_eq(adapter, eq_type, EHEA_MAX_ENTRIES_EQ, 0);
if (!pr->recv_eq) { if (!pr->eq) {
ehea_error("create_eq failed (recv_eq)"); ehea_error("create_eq failed (eq)");
goto out_free;
}
pr->send_eq = ehea_create_eq(adapter, eq_type, EHEA_MAX_ENTRIES_EQ, 0);
if (!pr->send_eq) {
ehea_error("create_eq failed (send_eq)");
goto out_free; goto out_free;
} }
pr->recv_cq = ehea_create_cq(adapter, pr_cfg->max_entries_rcq, pr->recv_cq = ehea_create_cq(adapter, pr_cfg->max_entries_rcq,
pr->recv_eq->fw_handle, pr->eq->fw_handle,
port->logical_port_id); port->logical_port_id);
if (!pr->recv_cq) { if (!pr->recv_cq) {
ehea_error("create_cq failed (cq_recv)"); ehea_error("create_cq failed (cq_recv)");
...@@ -1129,7 +1129,7 @@ static int ehea_init_port_res(struct ehea_port *port, struct ehea_port_res *pr, ...@@ -1129,7 +1129,7 @@ static int ehea_init_port_res(struct ehea_port *port, struct ehea_port_res *pr,
} }
pr->send_cq = ehea_create_cq(adapter, pr_cfg->max_entries_scq, pr->send_cq = ehea_create_cq(adapter, pr_cfg->max_entries_scq,
pr->send_eq->fw_handle, pr->eq->fw_handle,
port->logical_port_id); port->logical_port_id);
if (!pr->send_cq) { if (!pr->send_cq) {
ehea_error("create_cq failed (cq_send)"); ehea_error("create_cq failed (cq_send)");
...@@ -1194,11 +1194,20 @@ static int ehea_init_port_res(struct ehea_port *port, struct ehea_port_res *pr, ...@@ -1194,11 +1194,20 @@ static int ehea_init_port_res(struct ehea_port *port, struct ehea_port_res *pr,
ret = -EIO; ret = -EIO;
goto out_free; goto out_free;
} }
tasklet_init(&pr->send_comp_task, ehea_send_irq_tasklet,
(unsigned long)pr);
atomic_set(&pr->swqe_avail, init_attr->act_nr_send_wqes - 1); atomic_set(&pr->swqe_avail, init_attr->act_nr_send_wqes - 1);
kfree(init_attr); kfree(init_attr);
pr->d_netdev = alloc_netdev(0, "", ether_setup);
if (!pr->d_netdev)
goto out_free;
pr->d_netdev->priv = pr;
pr->d_netdev->weight = 64;
pr->d_netdev->poll = ehea_poll;
set_bit(__LINK_STATE_START, &pr->d_netdev->state);
strcpy(pr->d_netdev->name, port->netdev->name);
ret = 0; ret = 0;
goto out; goto out;
...@@ -1211,8 +1220,7 @@ static int ehea_init_port_res(struct ehea_port *port, struct ehea_port_res *pr, ...@@ -1211,8 +1220,7 @@ static int ehea_init_port_res(struct ehea_port *port, struct ehea_port_res *pr,
ehea_destroy_qp(pr->qp); ehea_destroy_qp(pr->qp);
ehea_destroy_cq(pr->send_cq); ehea_destroy_cq(pr->send_cq);
ehea_destroy_cq(pr->recv_cq); ehea_destroy_cq(pr->recv_cq);
ehea_destroy_eq(pr->send_eq); ehea_destroy_eq(pr->eq);
ehea_destroy_eq(pr->recv_eq);
out: out:
return ret; return ret;
} }
...@@ -1221,13 +1229,14 @@ static int ehea_clean_portres(struct ehea_port *port, struct ehea_port_res *pr) ...@@ -1221,13 +1229,14 @@ static int ehea_clean_portres(struct ehea_port *port, struct ehea_port_res *pr)
{ {
int ret, i; int ret, i;
free_netdev(pr->d_netdev);
ret = ehea_destroy_qp(pr->qp); ret = ehea_destroy_qp(pr->qp);
if (!ret) { if (!ret) {
ehea_destroy_cq(pr->send_cq); ehea_destroy_cq(pr->send_cq);
ehea_destroy_cq(pr->recv_cq); ehea_destroy_cq(pr->recv_cq);
ehea_destroy_eq(pr->send_eq); ehea_destroy_eq(pr->eq);
ehea_destroy_eq(pr->recv_eq);
for (i = 0; i < pr->rq1_skba.len; i++) for (i = 0; i < pr->rq1_skba.len; i++)
if (pr->rq1_skba.arr[i]) if (pr->rq1_skba.arr[i])
...@@ -1792,6 +1801,22 @@ static void ehea_xmit3(struct sk_buff *skb, struct net_device *dev, ...@@ -1792,6 +1801,22 @@ static void ehea_xmit3(struct sk_buff *skb, struct net_device *dev,
dev_kfree_skb(skb); dev_kfree_skb(skb);
} }
static inline int ehea_hash_skb(struct sk_buff *skb, int num_qps)
{
struct tcphdr *tcp;
u32 tmp;
if ((skb->protocol == htons(ETH_P_IP)) &&
(skb->nh.iph->protocol == IPPROTO_TCP)) {
tcp = (struct tcphdr*)(skb->nh.raw + (skb->nh.iph->ihl * 4));
tmp = (tcp->source + (tcp->dest << 16)) % 31;
tmp += skb->nh.iph->daddr % 31;
return tmp % num_qps;
}
else
return 0;
}
static int ehea_start_xmit(struct sk_buff *skb, struct net_device *dev) static int ehea_start_xmit(struct sk_buff *skb, struct net_device *dev)
{ {
struct ehea_port *port = netdev_priv(dev); struct ehea_port *port = netdev_priv(dev);
...@@ -1799,9 +1824,18 @@ static int ehea_start_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -1799,9 +1824,18 @@ static int ehea_start_xmit(struct sk_buff *skb, struct net_device *dev)
unsigned long flags; unsigned long flags;
u32 lkey; u32 lkey;
int swqe_index; int swqe_index;
struct ehea_port_res *pr = &port->port_res[0]; struct ehea_port_res *pr;
pr = &port->port_res[ehea_hash_skb(skb, port->num_tx_qps)];
spin_lock(&pr->xmit_lock); if (!spin_trylock(&pr->xmit_lock))
return NETDEV_TX_BUSY;
if (pr->queue_stopped) {
spin_unlock(&pr->xmit_lock);
return NETDEV_TX_BUSY;
}
swqe = ehea_get_swqe(pr->qp, &swqe_index); swqe = ehea_get_swqe(pr->qp, &swqe_index);
memset(swqe, 0, SWQE_HEADER_SIZE); memset(swqe, 0, SWQE_HEADER_SIZE);
...@@ -2060,7 +2094,7 @@ static int ehea_port_res_setup(struct ehea_port *port, int def_qps, ...@@ -2060,7 +2094,7 @@ static int ehea_port_res_setup(struct ehea_port *port, int def_qps,
} }
pr_cfg.max_entries_rcq = rq1_entries + rq2_entries + rq3_entries; pr_cfg.max_entries_rcq = rq1_entries + rq2_entries + rq3_entries;
pr_cfg.max_entries_scq = sq_entries; pr_cfg.max_entries_scq = sq_entries * 2;
pr_cfg.max_entries_sq = sq_entries; pr_cfg.max_entries_sq = sq_entries;
pr_cfg.max_entries_rq1 = rq1_entries; pr_cfg.max_entries_rq1 = rq1_entries;
pr_cfg.max_entries_rq2 = rq2_entries; pr_cfg.max_entries_rq2 = rq2_entries;
...@@ -2208,8 +2242,10 @@ static int ehea_down(struct net_device *dev) ...@@ -2208,8 +2242,10 @@ static int ehea_down(struct net_device *dev)
ehea_drop_multicast_list(dev); ehea_drop_multicast_list(dev);
ehea_free_interrupts(dev); ehea_free_interrupts(dev);
for (i = 0; i < port->num_def_qps + port->num_add_tx_qps; i++) for (i = 0; i < port->num_def_qps; i++)
tasklet_kill(&port->port_res[i].send_comp_task); while (test_bit(__LINK_STATE_RX_SCHED,
&port->port_res[i].d_netdev->state))
msleep(1);
ehea_broadcast_reg_helper(port, H_DEREG_BCMC); ehea_broadcast_reg_helper(port, H_DEREG_BCMC);
ret = ehea_clean_all_portres(port); ret = ehea_clean_all_portres(port);
......
...@@ -320,6 +320,11 @@ static inline struct ehea_cqe *ehea_poll_rq1(struct ehea_qp *qp, int *wqe_index) ...@@ -320,6 +320,11 @@ static inline struct ehea_cqe *ehea_poll_rq1(struct ehea_qp *qp, int *wqe_index)
return hw_qeit_get_valid(queue); return hw_qeit_get_valid(queue);
} }
static inline void ehea_inc_cq(struct ehea_cq *cq)
{
hw_qeit_inc(&cq->hw_queue);
}
static inline void ehea_inc_rq1(struct ehea_qp *qp) static inline void ehea_inc_rq1(struct ehea_qp *qp)
{ {
hw_qeit_inc(&qp->hw_rqueue1); hw_qeit_inc(&qp->hw_rqueue1);
...@@ -327,7 +332,7 @@ static inline void ehea_inc_rq1(struct ehea_qp *qp) ...@@ -327,7 +332,7 @@ static inline void ehea_inc_rq1(struct ehea_qp *qp)
static inline struct ehea_cqe *ehea_poll_cq(struct ehea_cq *my_cq) static inline struct ehea_cqe *ehea_poll_cq(struct ehea_cq *my_cq)
{ {
return hw_qeit_get_inc_valid(&my_cq->hw_queue); return hw_qeit_get_valid(&my_cq->hw_queue);
} }
#define EHEA_CQ_REGISTER_ORIG 0 #define EHEA_CQ_REGISTER_ORIG 0
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册