IB/ehca: Fix warnings issued by checkpatch.pl
Run the existing ehca code through checkpatch.pl and clean up the worst of the coding style violations. Signed-off-by: Joachim Fenkes <fenkes@de.ibm.com> Signed-off-by: Roland Dreier <rolandd@cisco.com>
This commit is contained in:
committed by
Roland Dreier
parent
187c72e31f
commit
2b94397adc
@@ -96,7 +96,8 @@ int ehca_create_eq(struct ehca_shca *shca,
|
||||
for (i = 0; i < nr_pages; i++) {
|
||||
u64 rpage;
|
||||
|
||||
if (!(vpage = ipz_qpageit_get_inc(&eq->ipz_queue))) {
|
||||
vpage = ipz_qpageit_get_inc(&eq->ipz_queue);
|
||||
if (!vpage) {
|
||||
ret = H_RESOURCE;
|
||||
goto create_eq_exit2;
|
||||
}
|
||||
|
@@ -127,6 +127,7 @@ int ehca_query_port(struct ib_device *ibdev,
|
||||
u8 port, struct ib_port_attr *props)
|
||||
{
|
||||
int ret = 0;
|
||||
u64 h_ret;
|
||||
struct ehca_shca *shca = container_of(ibdev, struct ehca_shca,
|
||||
ib_device);
|
||||
struct hipz_query_port *rblock;
|
||||
@@ -137,7 +138,8 @@ int ehca_query_port(struct ib_device *ibdev,
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
if (hipz_h_query_port(shca->ipz_hca_handle, port, rblock) != H_SUCCESS) {
|
||||
h_ret = hipz_h_query_port(shca->ipz_hca_handle, port, rblock);
|
||||
if (h_ret != H_SUCCESS) {
|
||||
ehca_err(&shca->ib_device, "Can't query port properties");
|
||||
ret = -EINVAL;
|
||||
goto query_port1;
|
||||
@@ -197,6 +199,7 @@ int ehca_query_sma_attr(struct ehca_shca *shca,
|
||||
u8 port, struct ehca_sma_attr *attr)
|
||||
{
|
||||
int ret = 0;
|
||||
u64 h_ret;
|
||||
struct hipz_query_port *rblock;
|
||||
|
||||
rblock = ehca_alloc_fw_ctrlblock(GFP_ATOMIC);
|
||||
@@ -205,7 +208,8 @@ int ehca_query_sma_attr(struct ehca_shca *shca,
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
if (hipz_h_query_port(shca->ipz_hca_handle, port, rblock) != H_SUCCESS) {
|
||||
h_ret = hipz_h_query_port(shca->ipz_hca_handle, port, rblock);
|
||||
if (h_ret != H_SUCCESS) {
|
||||
ehca_err(&shca->ib_device, "Can't query port properties");
|
||||
ret = -EINVAL;
|
||||
goto query_sma_attr1;
|
||||
@@ -230,9 +234,11 @@ query_sma_attr1:
|
||||
int ehca_query_pkey(struct ib_device *ibdev, u8 port, u16 index, u16 *pkey)
|
||||
{
|
||||
int ret = 0;
|
||||
struct ehca_shca *shca = container_of(ibdev, struct ehca_shca, ib_device);
|
||||
u64 h_ret;
|
||||
struct ehca_shca *shca;
|
||||
struct hipz_query_port *rblock;
|
||||
|
||||
shca = container_of(ibdev, struct ehca_shca, ib_device);
|
||||
if (index > 16) {
|
||||
ehca_err(&shca->ib_device, "Invalid index: %x.", index);
|
||||
return -EINVAL;
|
||||
@@ -244,7 +250,8 @@ int ehca_query_pkey(struct ib_device *ibdev, u8 port, u16 index, u16 *pkey)
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
if (hipz_h_query_port(shca->ipz_hca_handle, port, rblock) != H_SUCCESS) {
|
||||
h_ret = hipz_h_query_port(shca->ipz_hca_handle, port, rblock);
|
||||
if (h_ret != H_SUCCESS) {
|
||||
ehca_err(&shca->ib_device, "Can't query port properties");
|
||||
ret = -EINVAL;
|
||||
goto query_pkey1;
|
||||
@@ -262,6 +269,7 @@ int ehca_query_gid(struct ib_device *ibdev, u8 port,
|
||||
int index, union ib_gid *gid)
|
||||
{
|
||||
int ret = 0;
|
||||
u64 h_ret;
|
||||
struct ehca_shca *shca = container_of(ibdev, struct ehca_shca,
|
||||
ib_device);
|
||||
struct hipz_query_port *rblock;
|
||||
@@ -277,7 +285,8 @@ int ehca_query_gid(struct ib_device *ibdev, u8 port,
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
if (hipz_h_query_port(shca->ipz_hca_handle, port, rblock) != H_SUCCESS) {
|
||||
h_ret = hipz_h_query_port(shca->ipz_hca_handle, port, rblock);
|
||||
if (h_ret != H_SUCCESS) {
|
||||
ehca_err(&shca->ib_device, "Can't query port properties");
|
||||
ret = -EINVAL;
|
||||
goto query_gid1;
|
||||
@@ -302,11 +311,12 @@ int ehca_modify_port(struct ib_device *ibdev,
|
||||
struct ib_port_modify *props)
|
||||
{
|
||||
int ret = 0;
|
||||
struct ehca_shca *shca = container_of(ibdev, struct ehca_shca, ib_device);
|
||||
struct ehca_shca *shca;
|
||||
struct hipz_query_port *rblock;
|
||||
u32 cap;
|
||||
u64 hret;
|
||||
|
||||
shca = container_of(ibdev, struct ehca_shca, ib_device);
|
||||
if ((props->set_port_cap_mask | props->clr_port_cap_mask)
|
||||
& ~allowed_port_caps) {
|
||||
ehca_err(&shca->ib_device, "Non-changeable bits set in masks "
|
||||
@@ -325,7 +335,8 @@ int ehca_modify_port(struct ib_device *ibdev,
|
||||
goto modify_port1;
|
||||
}
|
||||
|
||||
if (hipz_h_query_port(shca->ipz_hca_handle, port, rblock) != H_SUCCESS) {
|
||||
hret = hipz_h_query_port(shca->ipz_hca_handle, port, rblock);
|
||||
if (hret != H_SUCCESS) {
|
||||
ehca_err(&shca->ib_device, "Can't query port properties");
|
||||
ret = -EINVAL;
|
||||
goto modify_port2;
|
||||
@@ -337,7 +348,8 @@ int ehca_modify_port(struct ib_device *ibdev,
|
||||
hret = hipz_h_modify_port(shca->ipz_hca_handle, port,
|
||||
cap, props->init_type, port_modify_mask);
|
||||
if (hret != H_SUCCESS) {
|
||||
ehca_err(&shca->ib_device, "Modify port failed hret=%lx", hret);
|
||||
ehca_err(&shca->ib_device, "Modify port failed hret=%lx",
|
||||
hret);
|
||||
ret = -EINVAL;
|
||||
}
|
||||
|
||||
|
@@ -81,8 +81,9 @@ struct ib_mr *ehca_reg_phys_mr(struct ib_pd *pd,
|
||||
int num_phys_buf,
|
||||
int mr_access_flags, u64 *iova_start);
|
||||
|
||||
struct ib_mr *ehca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, u64 virt,
|
||||
int mr_access_flags, struct ib_udata *udata);
|
||||
struct ib_mr *ehca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
|
||||
u64 virt, int mr_access_flags,
|
||||
struct ib_udata *udata);
|
||||
|
||||
int ehca_rereg_phys_mr(struct ib_mr *mr,
|
||||
int mr_rereg_mask,
|
||||
|
@@ -107,7 +107,7 @@ static DEFINE_SPINLOCK(shca_list_lock);
|
||||
static struct timer_list poll_eqs_timer;
|
||||
|
||||
#ifdef CONFIG_PPC_64K_PAGES
|
||||
static struct kmem_cache *ctblk_cache = NULL;
|
||||
static struct kmem_cache *ctblk_cache;
|
||||
|
||||
void *ehca_alloc_fw_ctrlblock(gfp_t flags)
|
||||
{
|
||||
@@ -868,18 +868,21 @@ int __init ehca_module_init(void)
|
||||
printk(KERN_INFO "eHCA Infiniband Device Driver "
|
||||
"(Rel.: SVNEHCA_0023)\n");
|
||||
|
||||
if ((ret = ehca_create_comp_pool())) {
|
||||
ret = ehca_create_comp_pool();
|
||||
if (ret) {
|
||||
ehca_gen_err("Cannot create comp pool.");
|
||||
return ret;
|
||||
}
|
||||
|
||||
if ((ret = ehca_create_slab_caches())) {
|
||||
ret = ehca_create_slab_caches();
|
||||
if (ret) {
|
||||
ehca_gen_err("Cannot create SLAB caches");
|
||||
ret = -ENOMEM;
|
||||
goto module_init1;
|
||||
}
|
||||
|
||||
if ((ret = ibmebus_register_driver(&ehca_driver))) {
|
||||
ret = ibmebus_register_driver(&ehca_driver);
|
||||
if (ret) {
|
||||
ehca_gen_err("Cannot register eHCA device driver");
|
||||
ret = -EINVAL;
|
||||
goto module_init2;
|
||||
|
@@ -61,9 +61,9 @@ static struct ehca_mr *ehca_mr_new(void)
|
||||
struct ehca_mr *me;
|
||||
|
||||
me = kmem_cache_zalloc(mr_cache, GFP_KERNEL);
|
||||
if (me) {
|
||||
if (me)
|
||||
spin_lock_init(&me->mrlock);
|
||||
} else
|
||||
else
|
||||
ehca_gen_err("alloc failed");
|
||||
|
||||
return me;
|
||||
@@ -79,9 +79,9 @@ static struct ehca_mw *ehca_mw_new(void)
|
||||
struct ehca_mw *me;
|
||||
|
||||
me = kmem_cache_zalloc(mw_cache, GFP_KERNEL);
|
||||
if (me) {
|
||||
if (me)
|
||||
spin_lock_init(&me->mwlock);
|
||||
} else
|
||||
else
|
||||
ehca_gen_err("alloc failed");
|
||||
|
||||
return me;
|
||||
@@ -246,8 +246,9 @@ reg_phys_mr_exit0:
|
||||
|
||||
/*----------------------------------------------------------------------*/
|
||||
|
||||
struct ib_mr *ehca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, u64 virt,
|
||||
int mr_access_flags, struct ib_udata *udata)
|
||||
struct ib_mr *ehca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
|
||||
u64 virt, int mr_access_flags,
|
||||
struct ib_udata *udata)
|
||||
{
|
||||
struct ib_mr *ib_mr;
|
||||
struct ehca_mr *e_mr;
|
||||
@@ -322,8 +323,9 @@ struct ib_mr *ehca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, u64 virt
|
||||
(&e_mr->umem->chunk_list),
|
||||
list);
|
||||
|
||||
ret = ehca_reg_mr(shca, e_mr, (u64*) virt, length, mr_access_flags, e_pd,
|
||||
&pginfo, &e_mr->ib.ib_mr.lkey, &e_mr->ib.ib_mr.rkey);
|
||||
ret = ehca_reg_mr(shca, e_mr, (u64 *)virt, length, mr_access_flags,
|
||||
e_pd, &pginfo, &e_mr->ib.ib_mr.lkey,
|
||||
&e_mr->ib.ib_mr.rkey);
|
||||
if (ret) {
|
||||
ib_mr = ERR_PTR(ret);
|
||||
goto reg_user_mr_exit2;
|
||||
@@ -420,7 +422,7 @@ int ehca_rereg_phys_mr(struct ib_mr *mr,
|
||||
goto rereg_phys_mr_exit0;
|
||||
}
|
||||
if (!phys_buf_array || num_phys_buf <= 0) {
|
||||
ehca_err(mr->device, "bad input values: mr_rereg_mask=%x"
|
||||
ehca_err(mr->device, "bad input values mr_rereg_mask=%x"
|
||||
" phys_buf_array=%p num_phys_buf=%x",
|
||||
mr_rereg_mask, phys_buf_array, num_phys_buf);
|
||||
ret = -EINVAL;
|
||||
@@ -444,10 +446,10 @@ int ehca_rereg_phys_mr(struct ib_mr *mr,
|
||||
|
||||
/* set requested values dependent on rereg request */
|
||||
spin_lock_irqsave(&e_mr->mrlock, sl_flags);
|
||||
new_start = e_mr->start; /* new == old address */
|
||||
new_size = e_mr->size; /* new == old length */
|
||||
new_acl = e_mr->acl; /* new == old access control */
|
||||
new_pd = container_of(mr->pd,struct ehca_pd,ib_pd); /*new == old PD*/
|
||||
new_start = e_mr->start;
|
||||
new_size = e_mr->size;
|
||||
new_acl = e_mr->acl;
|
||||
new_pd = container_of(mr->pd, struct ehca_pd, ib_pd);
|
||||
|
||||
if (mr_rereg_mask & IB_MR_REREG_TRANS) {
|
||||
new_start = iova_start; /* change address */
|
||||
@@ -517,7 +519,7 @@ int ehca_query_mr(struct ib_mr *mr, struct ib_mr_attr *mr_attr)
|
||||
struct ehca_pd *my_pd = container_of(mr->pd, struct ehca_pd, ib_pd);
|
||||
u32 cur_pid = current->tgid;
|
||||
unsigned long sl_flags;
|
||||
struct ehca_mr_hipzout_parms hipzout = {{0},0,0,0,0,0};
|
||||
struct ehca_mr_hipzout_parms hipzout;
|
||||
|
||||
if (my_pd->ib_pd.uobject && my_pd->ib_pd.uobject->context &&
|
||||
(my_pd->ownpid != cur_pid)) {
|
||||
@@ -629,7 +631,7 @@ struct ib_mw *ehca_alloc_mw(struct ib_pd *pd)
|
||||
struct ehca_pd *e_pd = container_of(pd, struct ehca_pd, ib_pd);
|
||||
struct ehca_shca *shca =
|
||||
container_of(pd->device, struct ehca_shca, ib_device);
|
||||
struct ehca_mw_hipzout_parms hipzout = {{0},0};
|
||||
struct ehca_mw_hipzout_parms hipzout;
|
||||
|
||||
e_mw = ehca_mw_new();
|
||||
if (!e_mw) {
|
||||
@@ -841,8 +843,7 @@ int ehca_map_phys_fmr(struct ib_fmr *fmr,
|
||||
map_phys_fmr_exit0:
|
||||
if (ret)
|
||||
ehca_err(fmr->device, "ret=%x fmr=%p page_list=%p list_len=%x "
|
||||
"iova=%lx",
|
||||
ret, fmr, page_list, list_len, iova);
|
||||
"iova=%lx", ret, fmr, page_list, list_len, iova);
|
||||
return ret;
|
||||
} /* end ehca_map_phys_fmr() */
|
||||
|
||||
@@ -960,7 +961,7 @@ int ehca_reg_mr(struct ehca_shca *shca,
|
||||
int ret;
|
||||
u64 h_ret;
|
||||
u32 hipz_acl;
|
||||
struct ehca_mr_hipzout_parms hipzout = {{0},0,0,0,0,0};
|
||||
struct ehca_mr_hipzout_parms hipzout;
|
||||
|
||||
ehca_mrmw_map_acl(acl, &hipz_acl);
|
||||
ehca_mrmw_set_pgsize_hipz_acl(&hipz_acl);
|
||||
@@ -1127,7 +1128,7 @@ inline int ehca_rereg_mr_rereg1(struct ehca_shca *shca,
|
||||
u64 *kpage;
|
||||
u64 rpage;
|
||||
struct ehca_mr_pginfo pginfo_save;
|
||||
struct ehca_mr_hipzout_parms hipzout = {{0},0,0,0,0,0};
|
||||
struct ehca_mr_hipzout_parms hipzout;
|
||||
|
||||
ehca_mrmw_map_acl(acl, &hipz_acl);
|
||||
ehca_mrmw_set_pgsize_hipz_acl(&hipz_acl);
|
||||
@@ -1305,7 +1306,7 @@ int ehca_unmap_one_fmr(struct ehca_shca *shca,
|
||||
struct ehca_mr save_fmr;
|
||||
u32 tmp_lkey, tmp_rkey;
|
||||
struct ehca_mr_pginfo pginfo;
|
||||
struct ehca_mr_hipzout_parms hipzout = {{0},0,0,0,0,0};
|
||||
struct ehca_mr_hipzout_parms hipzout;
|
||||
struct ehca_mr save_mr;
|
||||
|
||||
if (e_fmr->fmr_max_pages <= MAX_RPAGES) {
|
||||
@@ -1397,7 +1398,7 @@ int ehca_reg_smr(struct ehca_shca *shca,
|
||||
int ret = 0;
|
||||
u64 h_ret;
|
||||
u32 hipz_acl;
|
||||
struct ehca_mr_hipzout_parms hipzout = {{0},0,0,0,0,0};
|
||||
struct ehca_mr_hipzout_parms hipzout;
|
||||
|
||||
ehca_mrmw_map_acl(acl, &hipz_acl);
|
||||
ehca_mrmw_set_pgsize_hipz_acl(&hipz_acl);
|
||||
@@ -1519,7 +1520,7 @@ int ehca_reg_maxmr(struct ehca_shca *shca,
|
||||
u64 h_ret;
|
||||
struct ehca_mr *e_origmr = shca->maxmr;
|
||||
u32 hipz_acl;
|
||||
struct ehca_mr_hipzout_parms hipzout = {{0},0,0,0,0,0};
|
||||
struct ehca_mr_hipzout_parms hipzout;
|
||||
|
||||
ehca_mrmw_map_acl(acl, &hipz_acl);
|
||||
ehca_mrmw_set_pgsize_hipz_acl(&hipz_acl);
|
||||
|
@@ -101,15 +101,10 @@ int ehca_fmr_check_page_list(struct ehca_mr *e_fmr,
|
||||
u64 *page_list,
|
||||
int list_len);
|
||||
|
||||
int ehca_set_pagebuf(struct ehca_mr *e_mr,
|
||||
struct ehca_mr_pginfo *pginfo,
|
||||
int ehca_set_pagebuf(struct ehca_mr_pginfo *pginfo,
|
||||
u32 number,
|
||||
u64 *kpage);
|
||||
|
||||
int ehca_set_pagebuf_1(struct ehca_mr *e_mr,
|
||||
struct ehca_mr_pginfo *pginfo,
|
||||
u64 *rpage);
|
||||
|
||||
int ehca_mr_is_maxmr(u64 size,
|
||||
u64 *iova_start);
|
||||
|
||||
|
@@ -602,10 +602,10 @@ struct ehca_qp *internal_create_qp(struct ib_pd *pd,
|
||||
/* UD circumvention */
|
||||
parms.act_nr_send_sges -= 2;
|
||||
parms.act_nr_recv_sges -= 2;
|
||||
swqe_size = offsetof(struct ehca_wqe,
|
||||
u.ud_av.sg_list[parms.act_nr_send_sges]);
|
||||
rwqe_size = offsetof(struct ehca_wqe,
|
||||
u.ud_av.sg_list[parms.act_nr_recv_sges]);
|
||||
swqe_size = offsetof(struct ehca_wqe, u.ud_av.sg_list[
|
||||
parms.act_nr_send_sges]);
|
||||
rwqe_size = offsetof(struct ehca_wqe, u.ud_av.sg_list[
|
||||
parms.act_nr_recv_sges]);
|
||||
}
|
||||
|
||||
if (IB_QPT_GSI == qp_type || IB_QPT_SMI == qp_type) {
|
||||
@@ -690,8 +690,8 @@ struct ehca_qp *internal_create_qp(struct ib_pd *pd,
|
||||
if (my_qp->send_cq) {
|
||||
ret = ehca_cq_assign_qp(my_qp->send_cq, my_qp);
|
||||
if (ret) {
|
||||
ehca_err(pd->device, "Couldn't assign qp to send_cq ret=%x",
|
||||
ret);
|
||||
ehca_err(pd->device,
|
||||
"Couldn't assign qp to send_cq ret=%x", ret);
|
||||
goto create_qp_exit4;
|
||||
}
|
||||
}
|
||||
@@ -1019,7 +1019,8 @@ static int internal_modify_qp(struct ib_qp *ibqp,
|
||||
goto modify_qp_exit1;
|
||||
}
|
||||
|
||||
if ((mqpcb->qp_state = ib2ehca_qp_state(qp_new_state)))
|
||||
mqpcb->qp_state = ib2ehca_qp_state(qp_new_state);
|
||||
if (mqpcb->qp_state)
|
||||
update_mask = EHCA_BMASK_SET(MQPCB_MASK_QP_STATE, 1);
|
||||
else {
|
||||
ret = -EINVAL;
|
||||
|
@@ -79,7 +79,8 @@ static inline int ehca_write_rwqe(struct ipz_queue *ipz_rqueue,
|
||||
}
|
||||
|
||||
if (ehca_debug_level) {
|
||||
ehca_gen_dbg("RECEIVE WQE written into ipz_rqueue=%p", ipz_rqueue);
|
||||
ehca_gen_dbg("RECEIVE WQE written into ipz_rqueue=%p",
|
||||
ipz_rqueue);
|
||||
ehca_dmp( wqe_p, 16*(6 + wqe_p->nr_of_data_seg), "recv wqe");
|
||||
}
|
||||
|
||||
@@ -534,9 +535,11 @@ poll_cq_one_read_cqe:
|
||||
|
||||
cqe_count++;
|
||||
if (unlikely(cqe->status & WC_STATUS_PURGE_BIT)) {
|
||||
struct ehca_qp *qp=ehca_cq_get_qp(my_cq, cqe->local_qp_number);
|
||||
struct ehca_qp *qp;
|
||||
int purgeflag;
|
||||
unsigned long flags;
|
||||
|
||||
qp = ehca_cq_get_qp(my_cq, cqe->local_qp_number);
|
||||
if (!qp) {
|
||||
ehca_err(cq->device, "cq_num=%x qp_num=%x "
|
||||
"could not find qp -> ignore cqe",
|
||||
@@ -551,8 +554,8 @@ poll_cq_one_read_cqe:
|
||||
spin_unlock_irqrestore(&qp->spinlock_s, flags);
|
||||
|
||||
if (purgeflag) {
|
||||
ehca_dbg(cq->device, "Got CQE with purged bit qp_num=%x "
|
||||
"src_qp=%x",
|
||||
ehca_dbg(cq->device,
|
||||
"Got CQE with purged bit qp_num=%x src_qp=%x",
|
||||
cqe->local_qp_number, cqe->remote_qp_number);
|
||||
if (ehca_debug_level)
|
||||
ehca_dmp(cqe, 64, "qp_num=%x src_qp=%x",
|
||||
|
@@ -119,7 +119,7 @@ extern int ehca_debug_level;
|
||||
unsigned int l = (unsigned int)(len); \
|
||||
unsigned char *deb = (unsigned char *)(adr); \
|
||||
for (x = 0; x < l; x += 16) { \
|
||||
printk("EHCA_DMP:%s " format \
|
||||
printk(KERN_INFO "EHCA_DMP:%s " format \
|
||||
" adr=%p ofs=%04x %016lx %016lx\n", \
|
||||
__FUNCTION__, ##args, deb, x, \
|
||||
*((u64 *)&deb[0]), *((u64 *)&deb[8])); \
|
||||
@@ -137,7 +137,7 @@ extern int ehca_debug_level;
|
||||
#define EHCA_BMASK_SHIFTPOS(mask) (((mask) >> 16) & 0xffff)
|
||||
|
||||
/* internal function, don't use */
|
||||
#define EHCA_BMASK_MASK(mask) (0xffffffffffffffffULL >> ((64-(mask))&0xffff))
|
||||
#define EHCA_BMASK_MASK(mask) (~0ULL >> ((64 - (mask)) & 0xffff))
|
||||
|
||||
/**
|
||||
* EHCA_BMASK_SET - return value shifted and masked by mask
|
||||
|
@@ -215,7 +215,8 @@ static int ehca_mmap_qp(struct vm_area_struct *vma, struct ehca_qp *qp,
|
||||
case 2: /* qp rqueue_addr */
|
||||
ehca_dbg(qp->ib_qp.device, "qp_num=%x rqueue",
|
||||
qp->ib_qp.qp_num);
|
||||
ret = ehca_mmap_queue(vma, &qp->ipz_rqueue, &qp->mm_count_rqueue);
|
||||
ret = ehca_mmap_queue(vma, &qp->ipz_rqueue,
|
||||
&qp->mm_count_rqueue);
|
||||
if (unlikely(ret)) {
|
||||
ehca_err(qp->ib_qp.device,
|
||||
"ehca_mmap_queue(rq) failed rc=%x qp_num=%x",
|
||||
@@ -227,7 +228,8 @@ static int ehca_mmap_qp(struct vm_area_struct *vma, struct ehca_qp *qp,
|
||||
case 3: /* qp squeue_addr */
|
||||
ehca_dbg(qp->ib_qp.device, "qp_num=%x squeue",
|
||||
qp->ib_qp.qp_num);
|
||||
ret = ehca_mmap_queue(vma, &qp->ipz_squeue, &qp->mm_count_squeue);
|
||||
ret = ehca_mmap_queue(vma, &qp->ipz_squeue,
|
||||
&qp->mm_count_squeue);
|
||||
if (unlikely(ret)) {
|
||||
ehca_err(qp->ib_qp.device,
|
||||
"ehca_mmap_queue(sq) failed rc=%x qp_num=%x",
|
||||
|
Reference in New Issue
Block a user