[PATCH rdma-next 8/8] RDMA/mana_ib: Skip firmware commands for invalidated handles
From: Long Li
Date: Fri Mar 06 2026 - 20:51:28 EST
After a service reset, firmware handles for PD, CQ, WQ, QP, and MR
are set to INVALID_MANA_HANDLE by the reset notification path.
Check for INVALID_MANA_HANDLE in each destroy callback before issuing
firmware destroy commands. When a handle is invalid, skip the firmware
call and proceed directly to kernel resource cleanup (umem, queues,
memory). This avoids sending stale handles to firmware after reset.
Affected callbacks:
- mana_ib_dealloc_pd: skip mana_ib_gd_destroy_pd
- mana_ib_destroy_cq: skip mana_ib_gd_destroy_cq and queue destroy
- mana_ib_destroy_wq: skip mana_ib_destroy_queue
- mana_ib_destroy_qp_rss: skip mana_destroy_wq_obj per WQ
- mana_ib_destroy_qp_raw: skip mana_destroy_wq_obj
- mana_ib_dereg_mr: skip mana_ib_gd_destroy_mr
Signed-off-by: Long Li <longli@xxxxxxxxxxxxx>
---
drivers/infiniband/hw/mana/cq.c | 10 ++++++----
drivers/infiniband/hw/mana/main.c | 12 +++++++++---
drivers/infiniband/hw/mana/mr.c | 8 +++++---
drivers/infiniband/hw/mana/qp.c | 9 ++++++---
4 files changed, 26 insertions(+), 13 deletions(-)
diff --git a/drivers/infiniband/hw/mana/cq.c b/drivers/infiniband/hw/mana/cq.c
index b054684b8de7..315301bccb97 100644
--- a/drivers/infiniband/hw/mana/cq.c
+++ b/drivers/infiniband/hw/mana/cq.c
@@ -143,10 +143,12 @@ int mana_ib_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata)
mana_ib_remove_cq_cb(mdev, cq);
- /* Ignore return code as there is not much we can do about it.
- * The error message is printed inside.
- */
- mana_ib_gd_destroy_cq(mdev, cq);
+ if (cq->cq_handle != INVALID_MANA_HANDLE) {
+ /* Ignore return code as there is not much we can do about it.
+ * The error message is printed inside.
+ */
+ mana_ib_gd_destroy_cq(mdev, cq);
+ }
mana_ib_destroy_queue(mdev, &cq->queue);
diff --git a/drivers/infiniband/hw/mana/main.c b/drivers/infiniband/hw/mana/main.c
index 61ce30aa9cb2..d60205184dba 100644
--- a/drivers/infiniband/hw/mana/main.c
+++ b/drivers/infiniband/hw/mana/main.c
@@ -147,6 +147,9 @@ int mana_ib_dealloc_pd(struct ib_pd *ibpd, struct ib_udata *udata)
mutex_unlock(&mana_ucontext->lock);
}
+ if (pd->pd_handle == INVALID_MANA_HANDLE)
+ return 0;
+
mana_gd_init_req_hdr(&req.hdr, GDMA_DESTROY_PD, sizeof(req),
sizeof(resp));
@@ -280,9 +283,12 @@ void mana_ib_dealloc_ucontext(struct ib_ucontext *ibcontext)
list_del_init(&mana_ucontext->dev_list);
mutex_unlock(&mdev->ucontext_lock);
- ret = mana_gd_destroy_doorbell_page(gc, mana_ucontext->doorbell);
- if (ret)
- ibdev_dbg(ibdev, "Failed to destroy doorbell page %d\n", ret);
+ if (mana_ucontext->doorbell != INVALID_DOORBELL) {
+ ret = mana_gd_destroy_doorbell_page(gc, mana_ucontext->doorbell);
+ if (ret)
+ ibdev_dbg(ibdev, "Failed to destroy doorbell page %d\n",
+ ret);
+ }
}
int mana_ib_create_kernel_queue(struct mana_ib_dev *mdev, u32 size, enum gdma_queue_type type,
diff --git a/drivers/infiniband/hw/mana/mr.c b/drivers/infiniband/hw/mana/mr.c
index 7189ccd41576..75bc2a9c366a 100644
--- a/drivers/infiniband/hw/mana/mr.c
+++ b/drivers/infiniband/hw/mana/mr.c
@@ -336,9 +336,11 @@ int mana_ib_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata)
mutex_unlock(&mana_ucontext->lock);
}
- err = mana_ib_gd_destroy_mr(dev, mr->mr_handle);
- if (err)
- return err;
+ if (mr->mr_handle != INVALID_MANA_HANDLE) {
+ err = mana_ib_gd_destroy_mr(dev, mr->mr_handle);
+ if (err)
+ return err;
+ }
if (mr->umem)
ib_umem_release(mr->umem);
diff --git a/drivers/infiniband/hw/mana/qp.c b/drivers/infiniband/hw/mana/qp.c
index d590aca9b93a..76d59addb645 100644
--- a/drivers/infiniband/hw/mana/qp.c
+++ b/drivers/infiniband/hw/mana/qp.c
@@ -846,9 +846,11 @@ static int mana_ib_destroy_qp_rss(struct mana_ib_qp *qp,
for (i = 0; i < (1 << ind_tbl->log_ind_tbl_size); i++) {
ibwq = ind_tbl->ind_tbl[i];
wq = container_of(ibwq, struct mana_ib_wq, ibwq);
- ibdev_dbg(&mdev->ib_dev, "destroying wq->rx_object %llu\n",
+ ibdev_dbg(&mdev->ib_dev,
+ "destroying wq->rx_object %llu\n",
wq->rx_object);
- mana_destroy_wq_obj(mpc, GDMA_RQ, wq->rx_object);
+ if (wq->rx_object != INVALID_MANA_HANDLE)
+ mana_destroy_wq_obj(mpc, GDMA_RQ, wq->rx_object);
}
return 0;
@@ -867,7 +869,8 @@ static int mana_ib_destroy_qp_raw(struct mana_ib_qp *qp, struct ib_udata *udata)
mpc = netdev_priv(ndev);
pd = container_of(ibpd, struct mana_ib_pd, ibpd);
- mana_destroy_wq_obj(mpc, GDMA_SQ, qp->qp_handle);
+ if (qp->qp_handle != INVALID_MANA_HANDLE)
+ mana_destroy_wq_obj(mpc, GDMA_SQ, qp->qp_handle);
mana_ib_destroy_queue(mdev, &qp->raw_sq);
--
2.43.0