提交 a26b5fce 编写于 作者: L Linus Torvalds

Merge branch 'for-linus' of master.kernel.org:/pub/scm/linux/kernel/git/roland/infiniband

* 'for-linus' of master.kernel.org:/pub/scm/linux/kernel/git/roland/infiniband:
  IB/iser: Handle aborting a command after it is sent
  IB/mthca: Fix thinko in init_mr_table()
  RDMA/cxgb3: Fix resource leak in cxio_hal_init_ctrl_qp()
...@@ -498,9 +498,9 @@ static int cxio_hal_init_ctrl_qp(struct cxio_rdev *rdev_p) ...@@ -498,9 +498,9 @@ static int cxio_hal_init_ctrl_qp(struct cxio_rdev *rdev_p)
u64 sge_cmd, ctx0, ctx1; u64 sge_cmd, ctx0, ctx1;
u64 base_addr; u64 base_addr;
struct t3_modify_qp_wr *wqe; struct t3_modify_qp_wr *wqe;
struct sk_buff *skb = alloc_skb(sizeof(*wqe), GFP_KERNEL); struct sk_buff *skb;
skb = alloc_skb(sizeof(*wqe), GFP_KERNEL);
if (!skb) { if (!skb) {
PDBG("%s alloc_skb failed\n", __FUNCTION__); PDBG("%s alloc_skb failed\n", __FUNCTION__);
return -ENOMEM; return -ENOMEM;
...@@ -508,7 +508,7 @@ static int cxio_hal_init_ctrl_qp(struct cxio_rdev *rdev_p) ...@@ -508,7 +508,7 @@ static int cxio_hal_init_ctrl_qp(struct cxio_rdev *rdev_p)
err = cxio_hal_init_ctrl_cq(rdev_p); err = cxio_hal_init_ctrl_cq(rdev_p);
if (err) { if (err) {
PDBG("%s err %d initializing ctrl_cq\n", __FUNCTION__, err); PDBG("%s err %d initializing ctrl_cq\n", __FUNCTION__, err);
return err; goto err;
} }
rdev_p->ctrl_qp.workq = dma_alloc_coherent( rdev_p->ctrl_qp.workq = dma_alloc_coherent(
&(rdev_p->rnic_info.pdev->dev), &(rdev_p->rnic_info.pdev->dev),
...@@ -518,7 +518,8 @@ static int cxio_hal_init_ctrl_qp(struct cxio_rdev *rdev_p) ...@@ -518,7 +518,8 @@ static int cxio_hal_init_ctrl_qp(struct cxio_rdev *rdev_p)
GFP_KERNEL); GFP_KERNEL);
if (!rdev_p->ctrl_qp.workq) { if (!rdev_p->ctrl_qp.workq) {
PDBG("%s dma_alloc_coherent failed\n", __FUNCTION__); PDBG("%s dma_alloc_coherent failed\n", __FUNCTION__);
return -ENOMEM; err = -ENOMEM;
goto err;
} }
pci_unmap_addr_set(&rdev_p->ctrl_qp, mapping, pci_unmap_addr_set(&rdev_p->ctrl_qp, mapping,
rdev_p->ctrl_qp.dma_addr); rdev_p->ctrl_qp.dma_addr);
...@@ -556,6 +557,9 @@ static int cxio_hal_init_ctrl_qp(struct cxio_rdev *rdev_p) ...@@ -556,6 +557,9 @@ static int cxio_hal_init_ctrl_qp(struct cxio_rdev *rdev_p)
rdev_p->ctrl_qp.workq, 1 << T3_CTRL_QP_SIZE_LOG2); rdev_p->ctrl_qp.workq, 1 << T3_CTRL_QP_SIZE_LOG2);
skb->priority = CPL_PRIORITY_CONTROL; skb->priority = CPL_PRIORITY_CONTROL;
return (cxgb3_ofld_send(rdev_p->t3cdev_p, skb)); return (cxgb3_ofld_send(rdev_p->t3cdev_p, skb));
err:
kfree_skb(skb);
return err;
} }
static int cxio_hal_destroy_ctrl_qp(struct cxio_rdev *rdev_p) static int cxio_hal_destroy_ctrl_qp(struct cxio_rdev *rdev_p)
......
...@@ -881,8 +881,8 @@ int mthca_init_mr_table(struct mthca_dev *dev) ...@@ -881,8 +881,8 @@ int mthca_init_mr_table(struct mthca_dev *dev)
} }
mpts = mtts = 1 << i; mpts = mtts = 1 << i;
} else { } else {
mpts = dev->limits.num_mtt_segs; mtts = dev->limits.num_mtt_segs;
mtts = dev->limits.num_mpts; mpts = dev->limits.num_mpts;
} }
if (!mthca_is_memfree(dev) && if (!mthca_is_memfree(dev) &&
......
...@@ -658,6 +658,7 @@ void iser_ctask_rdma_finalize(struct iscsi_iser_cmd_task *iser_ctask) ...@@ -658,6 +658,7 @@ void iser_ctask_rdma_finalize(struct iscsi_iser_cmd_task *iser_ctask)
{ {
int deferred; int deferred;
int is_rdma_aligned = 1; int is_rdma_aligned = 1;
struct iser_regd_buf *regd;
/* if we were reading, copy back to unaligned sglist, /* if we were reading, copy back to unaligned sglist,
* anyway dma_unmap and free the copy * anyway dma_unmap and free the copy
...@@ -672,20 +673,20 @@ void iser_ctask_rdma_finalize(struct iscsi_iser_cmd_task *iser_ctask) ...@@ -672,20 +673,20 @@ void iser_ctask_rdma_finalize(struct iscsi_iser_cmd_task *iser_ctask)
} }
if (iser_ctask->dir[ISER_DIR_IN]) { if (iser_ctask->dir[ISER_DIR_IN]) {
deferred = iser_regd_buff_release regd = &iser_ctask->rdma_regd[ISER_DIR_IN];
(&iser_ctask->rdma_regd[ISER_DIR_IN]); deferred = iser_regd_buff_release(regd);
if (deferred) { if (deferred) {
iser_err("References remain for BUF-IN rdma reg\n"); iser_err("%d references remain for BUF-IN rdma reg\n",
BUG(); atomic_read(&regd->ref_count));
} }
} }
if (iser_ctask->dir[ISER_DIR_OUT]) { if (iser_ctask->dir[ISER_DIR_OUT]) {
deferred = iser_regd_buff_release regd = &iser_ctask->rdma_regd[ISER_DIR_OUT];
(&iser_ctask->rdma_regd[ISER_DIR_OUT]); deferred = iser_regd_buff_release(regd);
if (deferred) { if (deferred) {
iser_err("References remain for BUF-OUT rdma reg\n"); iser_err("%d references remain for BUF-OUT rdma reg\n",
BUG(); atomic_read(&regd->ref_count));
} }
} }
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册