Re: [PATCH 1/3] nvme: failover requests for inactive hctx

From: kernel test robot

Date: Thu Feb 26 2026 - 14:13:07 EST


Hi Daniel,

kernel test robot noticed the following build errors:

[auto build test ERROR on 6de23f81a5e08be8fbf5e8d7e9febc72a5b5f27f]

url: https://github.com/intel-lab-lkp/linux/commits/Daniel-Wagner/nvme-failover-requests-for-inactive-hctx/20260226-224213
base: 6de23f81a5e08be8fbf5e8d7e9febc72a5b5f27f
patch link: https://lore.kernel.org/r/20260226-revert-cpu-read-lock-v1-1-eb005072566e%40kernel.org
patch subject: [PATCH 1/3] nvme: failover requests for inactive hctx
config: riscv-defconfig (https://download.01.org/0day-ci/archive/20260227/202602270348.j0MMNhUj-lkp@xxxxxxxxx/config)
compiler: clang version 23.0.0git (https://github.com/llvm/llvm-project 9a109fbb6e184ec9bcce10615949f598f4c974a9)
reproduce (this is a W=1 build): (https://download.01.org/0day-ci/archive/20260227/202602270348.j0MMNhUj-lkp@xxxxxxxxx/reproduce)

If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <lkp@xxxxxxxxx>
| Closes: https://lore.kernel.org/oe-kbuild-all/202602270348.j0MMNhUj-lkp@xxxxxxxxx/

All errors (new ones prefixed by >>):

drivers/nvme/host/core.c:457:13: error: redefinition of 'nvme_failover_req'
457 | static void nvme_failover_req(struct request *req)
| ^
drivers/nvme/host/nvme.h:1020:20: note: previous definition is here
1020 | static inline void nvme_failover_req(struct request *req)
| ^
>> drivers/nvme/host/core.c:472:45: error: no member named 'ana_log_buf' in 'struct nvme_ctrl'
472 | if (nvme_is_ana_error(status) && ns->ctrl->ana_log_buf) {
| ~~~~~~~~ ^
>> drivers/nvme/host/core.c:474:34: error: no member named 'ana_work' in 'struct nvme_ctrl'
474 | queue_work(nvme_wq, &ns->ctrl->ana_work);
| ~~~~~~~~ ^
>> drivers/nvme/host/core.c:477:31: error: no member named 'requeue_lock' in 'struct nvme_ns_head'
477 | spin_lock_irqsave(&ns->head->requeue_lock, flags);
| ~~~~~~~~ ^
include/linux/spinlock.h:376:39: note: expanded from macro 'spin_lock_irqsave'
376 | raw_spin_lock_irqsave(spinlock_check(lock), flags); \
| ^~~~
include/linux/spinlock.h:244:34: note: expanded from macro 'raw_spin_lock_irqsave'
244 | flags = _raw_spin_lock_irqsave(lock); \
| ^~~~
>> drivers/nvme/host/core.c:494:28: error: no member named 'requeue_list' in 'struct nvme_ns_head'
494 | blk_steal_bios(&ns->head->requeue_list, req);
| ~~~~~~~~ ^
drivers/nvme/host/core.c:495:36: error: no member named 'requeue_lock' in 'struct nvme_ns_head'
495 | spin_unlock_irqrestore(&ns->head->requeue_lock, flags);
| ~~~~~~~~ ^
>> drivers/nvme/host/core.c:499:35: error: no member named 'requeue_work' in 'struct nvme_ns_head'
499 | kblockd_schedule_work(&ns->head->requeue_work);
| ~~~~~~~~ ^
7 errors generated.


vim +472 drivers/nvme/host/core.c

456
457 static void nvme_failover_req(struct request *req)
458 {
459 struct nvme_ns *ns = req->q->queuedata;
460 u16 status = nvme_req(req)->status & NVME_SCT_SC_MASK;
461 unsigned long flags;
462 struct bio *bio;
463
464 if (nvme_ns_head_multipath(ns->head))
465 nvme_mpath_clear_current_path(ns);
466
467 /*
468 * If we got back an ANA error, we know the controller is alive but not
469 * ready to serve this namespace. Kick of a re-read of the ANA
470 * information page, and just try any other available path for now.
471 */
> 472 if (nvme_is_ana_error(status) && ns->ctrl->ana_log_buf) {
473 set_bit(NVME_NS_ANA_PENDING, &ns->flags);
> 474 queue_work(nvme_wq, &ns->ctrl->ana_work);
475 }
476
> 477 spin_lock_irqsave(&ns->head->requeue_lock, flags);
478 for (bio = req->bio; bio; bio = bio->bi_next) {
479 if (nvme_ns_head_multipath(ns->head))
480 bio_set_dev(bio, ns->head->disk->part0);
481 if (bio->bi_opf & REQ_POLLED) {
482 bio->bi_opf &= ~REQ_POLLED;
483 bio->bi_cookie = BLK_QC_T_NONE;
484 }
485 /*
486 * The alternate request queue that we may end up submitting
487 * the bio to may be frozen temporarily, in this case REQ_NOWAIT
488 * will fail the I/O immediately with EAGAIN to the issuer.
489 * We are not in the issuer context which cannot block. Clear
490 * the flag to avoid spurious EAGAIN I/O failures.
491 */
492 bio->bi_opf &= ~REQ_NOWAIT;
493 }
> 494 blk_steal_bios(&ns->head->requeue_list, req);
495 spin_unlock_irqrestore(&ns->head->requeue_lock, flags);
496
497 nvme_req(req)->status = 0;
498 nvme_end_req(req);
> 499 kblockd_schedule_work(&ns->head->requeue_work);
500 }
501

--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki