提交 4e396db8 编写于 作者: L Linus Torvalds

Merge branch 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/roland/infiniband

* 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/roland/infiniband:
  mlx4_core: Fix thinko in QP destroy (incorrect bitmap_free)
  RDMA/cxgb3: Set the max_qp_init_rd_atom attribute in query_device
  IB/ehca: Fix static rate calculation
  IB/ehca: Return physical link information in query_port()
  IB/ipath: Fix race with ACK retry timeout list management
  IB/ipath: Fix memory leak in ipath_resize_cq() if copy_to_user() fails
  mlx4_core: Fix possible bad free in mlx4_buf_free()
...@@ -1000,6 +1000,7 @@ static int iwch_query_device(struct ib_device *ibdev, ...@@ -1000,6 +1000,7 @@ static int iwch_query_device(struct ib_device *ibdev,
props->max_sge = dev->attr.max_sge_per_wr; props->max_sge = dev->attr.max_sge_per_wr;
props->max_sge_rd = 1; props->max_sge_rd = 1;
props->max_qp_rd_atom = dev->attr.max_rdma_reads_per_qp; props->max_qp_rd_atom = dev->attr.max_rdma_reads_per_qp;
props->max_qp_init_rd_atom = dev->attr.max_rdma_reads_per_qp;
props->max_cq = dev->attr.max_cqs; props->max_cq = dev->attr.max_cqs;
props->max_cqe = dev->attr.max_cqes_per_cq; props->max_cqe = dev->attr.max_cqes_per_cq;
props->max_mr = dev->attr.max_mem_regs; props->max_mr = dev->attr.max_mem_regs;
......
...@@ -50,6 +50,38 @@ ...@@ -50,6 +50,38 @@
static struct kmem_cache *av_cache; static struct kmem_cache *av_cache;
int ehca_calc_ipd(struct ehca_shca *shca, int port,
enum ib_rate path_rate, u32 *ipd)
{
int path = ib_rate_to_mult(path_rate);
int link, ret;
struct ib_port_attr pa;
if (path_rate == IB_RATE_PORT_CURRENT) {
*ipd = 0;
return 0;
}
if (unlikely(path < 0)) {
ehca_err(&shca->ib_device, "Invalid static rate! path_rate=%x",
path_rate);
return -EINVAL;
}
ret = ehca_query_port(&shca->ib_device, port, &pa);
if (unlikely(ret < 0)) {
ehca_err(&shca->ib_device, "Failed to query port ret=%i", ret);
return ret;
}
link = ib_width_enum_to_int(pa.active_width) * pa.active_speed;
/* IPD = round((link / path) - 1) */
*ipd = ((link + (path >> 1)) / path) - 1;
return 0;
}
struct ib_ah *ehca_create_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr) struct ib_ah *ehca_create_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr)
{ {
int ret; int ret;
...@@ -69,15 +101,13 @@ struct ib_ah *ehca_create_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr) ...@@ -69,15 +101,13 @@ struct ib_ah *ehca_create_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr)
av->av.slid_path_bits = ah_attr->src_path_bits; av->av.slid_path_bits = ah_attr->src_path_bits;
if (ehca_static_rate < 0) { if (ehca_static_rate < 0) {
int ah_mult = ib_rate_to_mult(ah_attr->static_rate); u32 ipd;
int ehca_mult = if (ehca_calc_ipd(shca, ah_attr->port_num,
ib_rate_to_mult(shca->sport[ah_attr->port_num].rate ); ah_attr->static_rate, &ipd)) {
ret = -EINVAL;
if (ah_mult >= ehca_mult) goto create_ah_exit1;
av->av.ipd = 0; }
else av->av.ipd = ipd;
av->av.ipd = (ah_mult > 0) ?
((ehca_mult - 1) / ah_mult) : 0;
} else } else
av->av.ipd = ehca_static_rate; av->av.ipd = ehca_static_rate;
......
...@@ -95,7 +95,6 @@ struct ehca_sma_attr { ...@@ -95,7 +95,6 @@ struct ehca_sma_attr {
struct ehca_sport { struct ehca_sport {
struct ib_cq *ibcq_aqp1; struct ib_cq *ibcq_aqp1;
struct ib_qp *ibqp_aqp1; struct ib_qp *ibqp_aqp1;
enum ib_rate rate;
enum ib_port_state port_state; enum ib_port_state port_state;
struct ehca_sma_attr saved_attr; struct ehca_sma_attr saved_attr;
}; };
......
...@@ -151,7 +151,6 @@ int ehca_query_port(struct ib_device *ibdev, ...@@ -151,7 +151,6 @@ int ehca_query_port(struct ib_device *ibdev,
} }
memset(props, 0, sizeof(struct ib_port_attr)); memset(props, 0, sizeof(struct ib_port_attr));
props->state = rblock->state;
switch (rblock->max_mtu) { switch (rblock->max_mtu) {
case 0x1: case 0x1:
...@@ -188,11 +187,20 @@ int ehca_query_port(struct ib_device *ibdev, ...@@ -188,11 +187,20 @@ int ehca_query_port(struct ib_device *ibdev,
props->subnet_timeout = rblock->subnet_timeout; props->subnet_timeout = rblock->subnet_timeout;
props->init_type_reply = rblock->init_type_reply; props->init_type_reply = rblock->init_type_reply;
props->active_width = IB_WIDTH_12X; if (rblock->state && rblock->phys_width) {
props->active_speed = 0x1; props->phys_state = rblock->phys_pstate;
props->state = rblock->phys_state;
/* at the moment (logical) link state is always LINK_UP */ props->active_width = rblock->phys_width;
props->phys_state = 0x5; props->active_speed = rblock->phys_speed;
} else {
/* old firmware releases don't report physical
* port info, so use default values
*/
props->phys_state = 5;
props->state = rblock->state;
props->active_width = IB_WIDTH_12X;
props->active_speed = 0x1;
}
query_port1: query_port1:
ehca_free_fw_ctrlblock(rblock); ehca_free_fw_ctrlblock(rblock);
......
...@@ -189,6 +189,9 @@ int ehca_mmap(struct ib_ucontext *context, struct vm_area_struct *vma); ...@@ -189,6 +189,9 @@ int ehca_mmap(struct ib_ucontext *context, struct vm_area_struct *vma);
void ehca_poll_eqs(unsigned long data); void ehca_poll_eqs(unsigned long data);
int ehca_calc_ipd(struct ehca_shca *shca, int port,
enum ib_rate path_rate, u32 *ipd);
#ifdef CONFIG_PPC_64K_PAGES #ifdef CONFIG_PPC_64K_PAGES
void *ehca_alloc_fw_ctrlblock(gfp_t flags); void *ehca_alloc_fw_ctrlblock(gfp_t flags);
void ehca_free_fw_ctrlblock(void *ptr); void ehca_free_fw_ctrlblock(void *ptr);
......
...@@ -327,9 +327,6 @@ static int ehca_sense_attributes(struct ehca_shca *shca) ...@@ -327,9 +327,6 @@ static int ehca_sense_attributes(struct ehca_shca *shca)
shca->hw_level = ehca_hw_level; shca->hw_level = ehca_hw_level;
ehca_gen_dbg(" ... hardware level=%x", shca->hw_level); ehca_gen_dbg(" ... hardware level=%x", shca->hw_level);
shca->sport[0].rate = IB_RATE_30_GBPS;
shca->sport[1].rate = IB_RATE_30_GBPS;
shca->hca_cap = rblock->hca_cap_indicators; shca->hca_cap = rblock->hca_cap_indicators;
ehca_gen_dbg(" ... HCA capabilities:"); ehca_gen_dbg(" ... HCA capabilities:");
for (i = 0; i < ARRAY_SIZE(hca_cap_descr); i++) for (i = 0; i < ARRAY_SIZE(hca_cap_descr); i++)
......
...@@ -1196,10 +1196,6 @@ static int internal_modify_qp(struct ib_qp *ibqp, ...@@ -1196,10 +1196,6 @@ static int internal_modify_qp(struct ib_qp *ibqp,
update_mask |= EHCA_BMASK_SET(MQPCB_MASK_QKEY, 1); update_mask |= EHCA_BMASK_SET(MQPCB_MASK_QKEY, 1);
} }
if (attr_mask & IB_QP_AV) { if (attr_mask & IB_QP_AV) {
int ah_mult = ib_rate_to_mult(attr->ah_attr.static_rate);
int ehca_mult = ib_rate_to_mult(shca->sport[my_qp->
init_attr.port_num].rate);
mqpcb->dlid = attr->ah_attr.dlid; mqpcb->dlid = attr->ah_attr.dlid;
update_mask |= EHCA_BMASK_SET(MQPCB_MASK_DLID, 1); update_mask |= EHCA_BMASK_SET(MQPCB_MASK_DLID, 1);
mqpcb->source_path_bits = attr->ah_attr.src_path_bits; mqpcb->source_path_bits = attr->ah_attr.src_path_bits;
...@@ -1207,11 +1203,12 @@ static int internal_modify_qp(struct ib_qp *ibqp, ...@@ -1207,11 +1203,12 @@ static int internal_modify_qp(struct ib_qp *ibqp,
mqpcb->service_level = attr->ah_attr.sl; mqpcb->service_level = attr->ah_attr.sl;
update_mask |= EHCA_BMASK_SET(MQPCB_MASK_SERVICE_LEVEL, 1); update_mask |= EHCA_BMASK_SET(MQPCB_MASK_SERVICE_LEVEL, 1);
if (ah_mult < ehca_mult) if (ehca_calc_ipd(shca, my_qp->init_attr.port_num,
mqpcb->max_static_rate = (ah_mult > 0) ? attr->ah_attr.static_rate,
((ehca_mult - 1) / ah_mult) : 0; &mqpcb->max_static_rate)) {
else ret = -EINVAL;
mqpcb->max_static_rate = 0; goto modify_qp_exit2;
}
update_mask |= EHCA_BMASK_SET(MQPCB_MASK_MAX_STATIC_RATE, 1); update_mask |= EHCA_BMASK_SET(MQPCB_MASK_MAX_STATIC_RATE, 1);
/* /*
...@@ -1280,10 +1277,6 @@ static int internal_modify_qp(struct ib_qp *ibqp, ...@@ -1280,10 +1277,6 @@ static int internal_modify_qp(struct ib_qp *ibqp,
(MQPCB_MASK_RDMA_ATOMIC_OUTST_DEST_QP, 1); (MQPCB_MASK_RDMA_ATOMIC_OUTST_DEST_QP, 1);
} }
if (attr_mask & IB_QP_ALT_PATH) { if (attr_mask & IB_QP_ALT_PATH) {
int ah_mult = ib_rate_to_mult(attr->alt_ah_attr.static_rate);
int ehca_mult = ib_rate_to_mult(
shca->sport[my_qp->init_attr.port_num].rate);
if (attr->alt_port_num < 1 if (attr->alt_port_num < 1
|| attr->alt_port_num > shca->num_ports) { || attr->alt_port_num > shca->num_ports) {
ret = -EINVAL; ret = -EINVAL;
...@@ -1309,10 +1302,12 @@ static int internal_modify_qp(struct ib_qp *ibqp, ...@@ -1309,10 +1302,12 @@ static int internal_modify_qp(struct ib_qp *ibqp,
mqpcb->source_path_bits_al = attr->alt_ah_attr.src_path_bits; mqpcb->source_path_bits_al = attr->alt_ah_attr.src_path_bits;
mqpcb->service_level_al = attr->alt_ah_attr.sl; mqpcb->service_level_al = attr->alt_ah_attr.sl;
if (ah_mult > 0 && ah_mult < ehca_mult) if (ehca_calc_ipd(shca, my_qp->init_attr.port_num,
mqpcb->max_static_rate_al = (ehca_mult - 1) / ah_mult; attr->alt_ah_attr.static_rate,
else &mqpcb->max_static_rate_al)) {
mqpcb->max_static_rate_al = 0; ret = -EINVAL;
goto modify_qp_exit2;
}
/* OpenIB doesn't support alternate retry counts - copy them */ /* OpenIB doesn't support alternate retry counts - copy them */
mqpcb->retry_count_al = mqpcb->retry_count; mqpcb->retry_count_al = mqpcb->retry_count;
......
...@@ -402,7 +402,11 @@ struct hipz_query_port { ...@@ -402,7 +402,11 @@ struct hipz_query_port {
u64 max_msg_sz; u64 max_msg_sz;
u32 max_mtu; u32 max_mtu;
u32 vl_cap; u32 vl_cap;
u8 reserved2[1900]; u32 phys_pstate;
u32 phys_state;
u32 phys_speed;
u32 phys_width;
u8 reserved2[1884];
u64 guid_entries[255]; u64 guid_entries[255];
} __attribute__ ((packed)); } __attribute__ ((packed));
......
...@@ -404,7 +404,7 @@ int ipath_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) ...@@ -404,7 +404,7 @@ int ipath_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata)
ret = ib_copy_to_udata(udata, &offset, sizeof(offset)); ret = ib_copy_to_udata(udata, &offset, sizeof(offset));
if (ret) if (ret)
goto bail; goto bail_free;
} }
spin_lock_irq(&cq->lock); spin_lock_irq(&cq->lock);
...@@ -424,10 +424,8 @@ int ipath_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) ...@@ -424,10 +424,8 @@ int ipath_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata)
else else
n = head - tail; n = head - tail;
if (unlikely((u32)cqe < n)) { if (unlikely((u32)cqe < n)) {
spin_unlock_irq(&cq->lock);
vfree(wc);
ret = -EOVERFLOW; ret = -EOVERFLOW;
goto bail; goto bail_unlock;
} }
for (n = 0; tail != head; n++) { for (n = 0; tail != head; n++) {
if (cq->ip) if (cq->ip)
...@@ -459,7 +457,12 @@ int ipath_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) ...@@ -459,7 +457,12 @@ int ipath_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata)
} }
ret = 0; ret = 0;
goto bail;
bail_unlock:
spin_unlock_irq(&cq->lock);
bail_free:
vfree(wc);
bail: bail:
return ret; return ret;
} }
...@@ -959,8 +959,9 @@ static int do_rc_ack(struct ipath_qp *qp, u32 aeth, u32 psn, int opcode, ...@@ -959,8 +959,9 @@ static int do_rc_ack(struct ipath_qp *qp, u32 aeth, u32 psn, int opcode,
/* If this is a partial ACK, reset the retransmit timer. */ /* If this is a partial ACK, reset the retransmit timer. */
if (qp->s_last != qp->s_tail) { if (qp->s_last != qp->s_tail) {
spin_lock(&dev->pending_lock); spin_lock(&dev->pending_lock);
list_add_tail(&qp->timerwait, if (list_empty(&qp->timerwait))
&dev->pending[dev->pending_index]); list_add_tail(&qp->timerwait,
&dev->pending[dev->pending_index]);
spin_unlock(&dev->pending_lock); spin_unlock(&dev->pending_lock);
/* /*
* If we get a partial ACK for a resent operation, * If we get a partial ACK for a resent operation,
......
...@@ -171,9 +171,10 @@ void mlx4_buf_free(struct mlx4_dev *dev, int size, struct mlx4_buf *buf) ...@@ -171,9 +171,10 @@ void mlx4_buf_free(struct mlx4_dev *dev, int size, struct mlx4_buf *buf)
buf->u.direct.map); buf->u.direct.map);
else { else {
for (i = 0; i < buf->nbufs; ++i) for (i = 0; i < buf->nbufs; ++i)
dma_free_coherent(&dev->pdev->dev, PAGE_SIZE, if (buf->u.page_list[i].buf)
buf->u.page_list[i].buf, dma_free_coherent(&dev->pdev->dev, PAGE_SIZE,
buf->u.page_list[i].map); buf->u.page_list[i].buf,
buf->u.page_list[i].map);
kfree(buf->u.page_list); kfree(buf->u.page_list);
} }
} }
......
...@@ -240,7 +240,7 @@ void mlx4_qp_free(struct mlx4_dev *dev, struct mlx4_qp *qp) ...@@ -240,7 +240,7 @@ void mlx4_qp_free(struct mlx4_dev *dev, struct mlx4_qp *qp)
mlx4_table_put(dev, &qp_table->auxc_table, qp->qpn); mlx4_table_put(dev, &qp_table->auxc_table, qp->qpn);
mlx4_table_put(dev, &qp_table->qp_table, qp->qpn); mlx4_table_put(dev, &qp_table->qp_table, qp->qpn);
if (qp->qpn < dev->caps.sqp_start + 8) if (qp->qpn >= dev->caps.sqp_start + 8)
mlx4_bitmap_free(&qp_table->bitmap, qp->qpn); mlx4_bitmap_free(&qp_table->bitmap, qp->qpn);
} }
EXPORT_SYMBOL_GPL(mlx4_qp_free); EXPORT_SYMBOL_GPL(mlx4_qp_free);
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册