提交 2528ea60 编写于 作者: M Mike Marciniszyn 提交者: Roland Dreier

IB/qib: Change receive queue/QPN selection

The basic idea is that on SusieQ, the difficult part of mapping QPN to
context is handled by the mapping registers so the generic QPN
allocation doesn't need to worry about chip specifics.  For Monty and
Linda, there is no mapping table so the qpt->mask (same as
dd->qpn_mask), is used to see if the QPN to context falls within
[zero..dd->n_krcv_queues).
Signed-off-by: NMike Marciniszyn <mike.marciniszyn@qlogic.com>
Signed-off-by: NRoland Dreier <rolandd@cisco.com>
上级 19ede2e4
...@@ -2297,7 +2297,7 @@ static void qib_7220_config_ctxts(struct qib_devdata *dd) ...@@ -2297,7 +2297,7 @@ static void qib_7220_config_ctxts(struct qib_devdata *dd)
nchipctxts = qib_read_kreg32(dd, kr_portcnt); nchipctxts = qib_read_kreg32(dd, kr_portcnt);
dd->cspec->numctxts = nchipctxts; dd->cspec->numctxts = nchipctxts;
if (qib_n_krcv_queues > 1) { if (qib_n_krcv_queues > 1) {
dd->qpn_mask = 0x3f; dd->qpn_mask = 0x3e;
dd->first_user_ctxt = qib_n_krcv_queues * dd->num_pports; dd->first_user_ctxt = qib_n_krcv_queues * dd->num_pports;
if (dd->first_user_ctxt > nchipctxts) if (dd->first_user_ctxt > nchipctxts)
dd->first_user_ctxt = nchipctxts; dd->first_user_ctxt = nchipctxts;
......
...@@ -3515,11 +3515,6 @@ static void qib_7322_config_ctxts(struct qib_devdata *dd) ...@@ -3515,11 +3515,6 @@ static void qib_7322_config_ctxts(struct qib_devdata *dd)
nchipctxts = qib_read_kreg32(dd, kr_contextcnt); nchipctxts = qib_read_kreg32(dd, kr_contextcnt);
dd->cspec->numctxts = nchipctxts; dd->cspec->numctxts = nchipctxts;
if (qib_n_krcv_queues > 1 && dd->num_pports) { if (qib_n_krcv_queues > 1 && dd->num_pports) {
/*
* Set the mask for which bits from the QPN are used
* to select a context number.
*/
dd->qpn_mask = 0x3f;
dd->first_user_ctxt = NUM_IB_PORTS + dd->first_user_ctxt = NUM_IB_PORTS +
(qib_n_krcv_queues - 1) * dd->num_pports; (qib_n_krcv_queues - 1) * dd->num_pports;
if (dd->first_user_ctxt > nchipctxts) if (dd->first_user_ctxt > nchipctxts)
...@@ -5865,7 +5860,8 @@ static void write_7322_initregs(struct qib_devdata *dd) ...@@ -5865,7 +5860,8 @@ static void write_7322_initregs(struct qib_devdata *dd)
unsigned n, regno; unsigned n, regno;
unsigned long flags; unsigned long flags;
if (!dd->qpn_mask || !dd->pport[pidx].link_speed_supported) if (dd->n_krcv_queues < 2 ||
!dd->pport[pidx].link_speed_supported)
continue; continue;
ppd = &dd->pport[pidx]; ppd = &dd->pport[pidx];
......
...@@ -48,13 +48,12 @@ static inline unsigned mk_qpn(struct qib_qpn_table *qpt, ...@@ -48,13 +48,12 @@ static inline unsigned mk_qpn(struct qib_qpn_table *qpt,
static inline unsigned find_next_offset(struct qib_qpn_table *qpt, static inline unsigned find_next_offset(struct qib_qpn_table *qpt,
struct qpn_map *map, unsigned off, struct qpn_map *map, unsigned off,
unsigned r) unsigned n)
{ {
if (qpt->mask) { if (qpt->mask) {
off++; off++;
if ((off & qpt->mask) >> 1 != r) if (((off & qpt->mask) >> 1) >= n)
off = ((off & qpt->mask) ? off = (off | qpt->mask) + 2;
(off | qpt->mask) + 1 : off) | (r << 1);
} else } else
off = find_next_zero_bit(map->page, BITS_PER_PAGE, off); off = find_next_zero_bit(map->page, BITS_PER_PAGE, off);
return off; return off;
...@@ -123,7 +122,6 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt, ...@@ -123,7 +122,6 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt,
u32 i, offset, max_scan, qpn; u32 i, offset, max_scan, qpn;
struct qpn_map *map; struct qpn_map *map;
u32 ret; u32 ret;
int r;
if (type == IB_QPT_SMI || type == IB_QPT_GSI) { if (type == IB_QPT_SMI || type == IB_QPT_GSI) {
unsigned n; unsigned n;
...@@ -139,15 +137,11 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt, ...@@ -139,15 +137,11 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt,
goto bail; goto bail;
} }
r = smp_processor_id();
if (r >= dd->n_krcv_queues)
r %= dd->n_krcv_queues;
qpn = qpt->last + 1; qpn = qpt->last + 1;
if (qpn >= QPN_MAX) if (qpn >= QPN_MAX)
qpn = 2; qpn = 2;
if (qpt->mask && ((qpn & qpt->mask) >> 1) != r) if (qpt->mask && ((qpn & qpt->mask) >> 1) >= dd->n_krcv_queues)
qpn = ((qpn & qpt->mask) ? (qpn | qpt->mask) + 1 : qpn) | qpn = (qpn | qpt->mask) + 2;
(r << 1);
offset = qpn & BITS_PER_PAGE_MASK; offset = qpn & BITS_PER_PAGE_MASK;
map = &qpt->map[qpn / BITS_PER_PAGE]; map = &qpt->map[qpn / BITS_PER_PAGE];
max_scan = qpt->nmaps - !offset; max_scan = qpt->nmaps - !offset;
...@@ -163,7 +157,8 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt, ...@@ -163,7 +157,8 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt,
ret = qpn; ret = qpn;
goto bail; goto bail;
} }
offset = find_next_offset(qpt, map, offset, r); offset = find_next_offset(qpt, map, offset,
dd->n_krcv_queues);
qpn = mk_qpn(qpt, map, offset); qpn = mk_qpn(qpt, map, offset);
/* /*
* This test differs from alloc_pidmap(). * This test differs from alloc_pidmap().
...@@ -183,13 +178,13 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt, ...@@ -183,13 +178,13 @@ static int alloc_qpn(struct qib_devdata *dd, struct qib_qpn_table *qpt,
if (qpt->nmaps == QPNMAP_ENTRIES) if (qpt->nmaps == QPNMAP_ENTRIES)
break; break;
map = &qpt->map[qpt->nmaps++]; map = &qpt->map[qpt->nmaps++];
offset = qpt->mask ? (r << 1) : 0; offset = 0;
} else if (map < &qpt->map[qpt->nmaps]) { } else if (map < &qpt->map[qpt->nmaps]) {
++map; ++map;
offset = qpt->mask ? (r << 1) : 0; offset = 0;
} else { } else {
map = &qpt->map[0]; map = &qpt->map[0];
offset = qpt->mask ? (r << 1) : 2; offset = 2;
} }
qpn = mk_qpn(qpt, map, offset); qpn = mk_qpn(qpt, map, offset);
} }
...@@ -1065,7 +1060,6 @@ struct ib_qp *qib_create_qp(struct ib_pd *ibpd, ...@@ -1065,7 +1060,6 @@ struct ib_qp *qib_create_qp(struct ib_pd *ibpd,
} }
qp->ibqp.qp_num = err; qp->ibqp.qp_num = err;
qp->port_num = init_attr->port_num; qp->port_num = init_attr->port_num;
qp->processor_id = smp_processor_id();
qib_reset_qp(qp, init_attr->qp_type); qib_reset_qp(qp, init_attr->qp_type);
break; break;
......
...@@ -435,7 +435,6 @@ struct qib_qp { ...@@ -435,7 +435,6 @@ struct qib_qp {
spinlock_t r_lock; /* used for APM */ spinlock_t r_lock; /* used for APM */
spinlock_t s_lock; spinlock_t s_lock;
atomic_t s_dma_busy; atomic_t s_dma_busy;
unsigned processor_id; /* Processor ID QP is bound to */
u32 s_flags; u32 s_flags;
u32 s_cur_size; /* size of send packet in bytes */ u32 s_cur_size; /* size of send packet in bytes */
u32 s_len; /* total length of s_sge */ u32 s_len; /* total length of s_sge */
...@@ -813,13 +812,8 @@ extern struct workqueue_struct *qib_cq_wq; ...@@ -813,13 +812,8 @@ extern struct workqueue_struct *qib_cq_wq;
*/ */
static inline void qib_schedule_send(struct qib_qp *qp) static inline void qib_schedule_send(struct qib_qp *qp)
{ {
if (qib_send_ok(qp)) { if (qib_send_ok(qp))
if (qp->processor_id == smp_processor_id()) queue_work(qib_wq, &qp->s_work);
queue_work(qib_wq, &qp->s_work);
else
queue_work_on(qp->processor_id,
qib_wq, &qp->s_work);
}
} }
static inline int qib_pkey_ok(u16 pkey1, u16 pkey2) static inline int qib_pkey_ok(u16 pkey1, u16 pkey2)
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册