Re: [PATCH RFC 3/3] nvme: delay failover by command quiesce timeout
From: Mohamed Khalfella
Date: Tue Apr 15 2025 - 20:17:51 EST
On 2025-04-15 14:17:48 +0200, Daniel Wagner wrote:
> On Thu, Apr 10, 2025 at 01:51:37AM -0700, Mohamed Khalfella wrote:
> > > +void nvme_schedule_failover(struct nvme_ctrl *ctrl)
> > > +{
> > > + unsigned long delay;
> > > +
> > > + if (ctrl->cqt)
> > > + delay = msecs_to_jiffies(ctrl->cqt);
> > > + else
> > > + delay = ctrl->kato * HZ;
> >
> > I thought that delay = m * ctrl->kato + ctrl->cqt
> > where m = ctrl->ctratt & NVME_CTRL_ATTR_TBKAS ? 3 : 2
> > no?
>
> The failover schedule delay is the additional amount of time we have to
> wait for the target to cleanup (CQT). If the CTQ is not valid I thought
> the spec said to wait for a KATO. Possible I got that wrong.
>
> The factor 3 or 2 is relavant for the timeout value for the KATO command
> we schedule. The failover schedule timeout is ontop of the command
> timeout value.
>
> > > --- a/drivers/nvme/host/multipath.c
> > > +++ b/drivers/nvme/host/multipath.c
> > > @@ -86,9 +86,11 @@ void nvme_mpath_start_freeze(struct nvme_subsystem *subsys)
> > > void nvme_failover_req(struct request *req)
> > > {
> > > struct nvme_ns *ns = req->q->queuedata;
> > > + struct nvme_ctrl *ctrl = nvme_req(req)->ctrl;
> > > u16 status = nvme_req(req)->status & NVME_SCT_SC_MASK;
> > > unsigned long flags;
> > > struct bio *bio;
> > > + enum nvme_ctrl_state state = nvme_ctrl_state(ctrl);
> > >
> > > nvme_mpath_clear_current_path(ns);
> > >
> > > @@ -121,9 +123,53 @@ void nvme_failover_req(struct request *req)
> > > blk_steal_bios(&ns->head->requeue_list, req);
> > > spin_unlock_irqrestore(&ns->head->requeue_lock, flags);
> > >
> > > - nvme_req(req)->status = 0;
> > > - nvme_end_req(req);
> > > - kblockd_schedule_work(&ns->head->requeue_work);
> > > + spin_lock_irqsave(&ctrl->lock, flags);
> > > + list_add_tail(&req->queuelist, &ctrl->failover_list);
> > > + spin_unlock_irqrestore(&ctrl->lock, flags);
> >
> > I see this is the only place where held requests are added to
> > failover_list.
> >
> > - Will this hold admin requests in failover_list?
>
> Yes.
Help me see this:
- nvme_failover_req() is the only place reqs are added to failover_list.
- nvme_decide_disposition() returns FAILOVER only if req has REQ_NVME_MPATH set.
How/where do admin requests get REQ_NVME_MPATH set?
>
> > - What about requests that do not go through nvme_failover_req(), like
> > passthrough requests, do we not want to hold these requests until it
> > is safe for them to be retried?
>
> Pasthrough commands should fail immediately. Userland is in charge here,
> not the kernel. At least this what should happen here.
>
> > - In case of controller reset or delete if nvme_disable_ctrl()
> > successfully disables the controller, then we do not want to add
> > canceled requests to failover_list, right? Does this implementation
> > consider this case?
>
> Not sure. I've tested a few things but I am pretty sure this RFC is far
> from being complete.
I think it does not, and maybe it should honor this. Otherwise every
controller reset/delete will end up holding requests unnecessarily.