Merge branch 'for-linus' of master.kernel.org:/pub/scm/linux/kernel/git/roland/infiniband
* 'for-linus' of master.kernel.org:/pub/scm/linux/kernel/git/roland/infiniband: IB/iser: Handle aborting a command after it is sent IB/mthca: Fix thinko in init_mr_table() RDMA/cxgb3: Fix resource leak in cxio_hal_init_ctrl_qp()
This commit is contained in:
@@ -498,9 +498,9 @@ static int cxio_hal_init_ctrl_qp(struct cxio_rdev *rdev_p)
|
|||||||
u64 sge_cmd, ctx0, ctx1;
|
u64 sge_cmd, ctx0, ctx1;
|
||||||
u64 base_addr;
|
u64 base_addr;
|
||||||
struct t3_modify_qp_wr *wqe;
|
struct t3_modify_qp_wr *wqe;
|
||||||
struct sk_buff *skb = alloc_skb(sizeof(*wqe), GFP_KERNEL);
|
struct sk_buff *skb;
|
||||||
|
|
||||||
|
|
||||||
|
skb = alloc_skb(sizeof(*wqe), GFP_KERNEL);
|
||||||
if (!skb) {
|
if (!skb) {
|
||||||
PDBG("%s alloc_skb failed\n", __FUNCTION__);
|
PDBG("%s alloc_skb failed\n", __FUNCTION__);
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
@@ -508,7 +508,7 @@ static int cxio_hal_init_ctrl_qp(struct cxio_rdev *rdev_p)
|
|||||||
err = cxio_hal_init_ctrl_cq(rdev_p);
|
err = cxio_hal_init_ctrl_cq(rdev_p);
|
||||||
if (err) {
|
if (err) {
|
||||||
PDBG("%s err %d initializing ctrl_cq\n", __FUNCTION__, err);
|
PDBG("%s err %d initializing ctrl_cq\n", __FUNCTION__, err);
|
||||||
return err;
|
goto err;
|
||||||
}
|
}
|
||||||
rdev_p->ctrl_qp.workq = dma_alloc_coherent(
|
rdev_p->ctrl_qp.workq = dma_alloc_coherent(
|
||||||
&(rdev_p->rnic_info.pdev->dev),
|
&(rdev_p->rnic_info.pdev->dev),
|
||||||
@@ -518,7 +518,8 @@ static int cxio_hal_init_ctrl_qp(struct cxio_rdev *rdev_p)
|
|||||||
GFP_KERNEL);
|
GFP_KERNEL);
|
||||||
if (!rdev_p->ctrl_qp.workq) {
|
if (!rdev_p->ctrl_qp.workq) {
|
||||||
PDBG("%s dma_alloc_coherent failed\n", __FUNCTION__);
|
PDBG("%s dma_alloc_coherent failed\n", __FUNCTION__);
|
||||||
return -ENOMEM;
|
err = -ENOMEM;
|
||||||
|
goto err;
|
||||||
}
|
}
|
||||||
pci_unmap_addr_set(&rdev_p->ctrl_qp, mapping,
|
pci_unmap_addr_set(&rdev_p->ctrl_qp, mapping,
|
||||||
rdev_p->ctrl_qp.dma_addr);
|
rdev_p->ctrl_qp.dma_addr);
|
||||||
@@ -556,6 +557,9 @@ static int cxio_hal_init_ctrl_qp(struct cxio_rdev *rdev_p)
|
|||||||
rdev_p->ctrl_qp.workq, 1 << T3_CTRL_QP_SIZE_LOG2);
|
rdev_p->ctrl_qp.workq, 1 << T3_CTRL_QP_SIZE_LOG2);
|
||||||
skb->priority = CPL_PRIORITY_CONTROL;
|
skb->priority = CPL_PRIORITY_CONTROL;
|
||||||
return (cxgb3_ofld_send(rdev_p->t3cdev_p, skb));
|
return (cxgb3_ofld_send(rdev_p->t3cdev_p, skb));
|
||||||
|
err:
|
||||||
|
kfree_skb(skb);
|
||||||
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int cxio_hal_destroy_ctrl_qp(struct cxio_rdev *rdev_p)
|
static int cxio_hal_destroy_ctrl_qp(struct cxio_rdev *rdev_p)
|
||||||
|
@@ -881,8 +881,8 @@ int mthca_init_mr_table(struct mthca_dev *dev)
|
|||||||
}
|
}
|
||||||
mpts = mtts = 1 << i;
|
mpts = mtts = 1 << i;
|
||||||
} else {
|
} else {
|
||||||
mpts = dev->limits.num_mtt_segs;
|
mtts = dev->limits.num_mtt_segs;
|
||||||
mtts = dev->limits.num_mpts;
|
mpts = dev->limits.num_mpts;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!mthca_is_memfree(dev) &&
|
if (!mthca_is_memfree(dev) &&
|
||||||
|
@@ -658,6 +658,7 @@ void iser_ctask_rdma_finalize(struct iscsi_iser_cmd_task *iser_ctask)
|
|||||||
{
|
{
|
||||||
int deferred;
|
int deferred;
|
||||||
int is_rdma_aligned = 1;
|
int is_rdma_aligned = 1;
|
||||||
|
struct iser_regd_buf *regd;
|
||||||
|
|
||||||
/* if we were reading, copy back to unaligned sglist,
|
/* if we were reading, copy back to unaligned sglist,
|
||||||
* anyway dma_unmap and free the copy
|
* anyway dma_unmap and free the copy
|
||||||
@@ -672,20 +673,20 @@ void iser_ctask_rdma_finalize(struct iscsi_iser_cmd_task *iser_ctask)
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (iser_ctask->dir[ISER_DIR_IN]) {
|
if (iser_ctask->dir[ISER_DIR_IN]) {
|
||||||
deferred = iser_regd_buff_release
|
regd = &iser_ctask->rdma_regd[ISER_DIR_IN];
|
||||||
(&iser_ctask->rdma_regd[ISER_DIR_IN]);
|
deferred = iser_regd_buff_release(regd);
|
||||||
if (deferred) {
|
if (deferred) {
|
||||||
iser_err("References remain for BUF-IN rdma reg\n");
|
iser_err("%d references remain for BUF-IN rdma reg\n",
|
||||||
BUG();
|
atomic_read(®d->ref_count));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (iser_ctask->dir[ISER_DIR_OUT]) {
|
if (iser_ctask->dir[ISER_DIR_OUT]) {
|
||||||
deferred = iser_regd_buff_release
|
regd = &iser_ctask->rdma_regd[ISER_DIR_OUT];
|
||||||
(&iser_ctask->rdma_regd[ISER_DIR_OUT]);
|
deferred = iser_regd_buff_release(regd);
|
||||||
if (deferred) {
|
if (deferred) {
|
||||||
iser_err("References remain for BUF-OUT rdma reg\n");
|
iser_err("%d references remain for BUF-OUT rdma reg\n",
|
||||||
BUG();
|
atomic_read(®d->ref_count));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Reference in New Issue
Block a user