Re: INFO: task hung in blkdev_open bug

From: yang lan
Date: Fri May 19 2023 - 03:13:00 EST


Hi,

./rqos/wbt/wb_background:4
./rqos/wbt/wb_normal:8
./rqos/wbt/unknown_cnt:0
./rqos/wbt/min_lat_nsec:2000000
./rqos/wbt/inflight:0: inflight 0
./rqos/wbt/inflight:1: inflight 0
./rqos/wbt/inflight:2: inflight 0
./rqos/wbt/id:0
./rqos/wbt/enabled:1
./rqos/wbt/curr_win_nsec:0
./hctx0/type:default
./hctx0/dispatch_busy:0
./hctx0/active:0
./hctx0/run:1
./hctx0/sched_tags_bitmap:00000000: 0100 0000 0000 0000 0000 0000 0000 0000
./hctx0/sched_tags_bitmap:00000010: 0000 0000 0000 0000 0000 0000 0000 0000
./hctx0/sched_tags:nr_tags=256
./hctx0/sched_tags:nr_reserved_tags=0
./hctx0/sched_tags:active_queues=0
./hctx0/sched_tags:bitmap_tags:
./hctx0/sched_tags:depth=256
./hctx0/sched_tags:busy=1
./hctx0/sched_tags:cleared=0
./hctx0/sched_tags:bits_per_word=64
./hctx0/sched_tags:map_nr=4
./hctx0/sched_tags:alloc_hint={245, 45}
./hctx0/sched_tags:wake_batch=8
./hctx0/sched_tags:wake_index=0
./hctx0/sched_tags:ws_active=0
./hctx0/sched_tags:ws={
./hctx0/sched_tags: {.wait=inactive},
./hctx0/sched_tags: {.wait=inactive},
./hctx0/sched_tags: {.wait=inactive},
./hctx0/sched_tags: {.wait=inactive},
./hctx0/sched_tags: {.wait=inactive},
./hctx0/sched_tags: {.wait=inactive},
./hctx0/sched_tags: {.wait=inactive},
./hctx0/sched_tags: {.wait=inactive},
./hctx0/sched_tags:}
./hctx0/sched_tags:round_robin=0
./hctx0/sched_tags:min_shallow_depth=192
./hctx0/tags_bitmap:00000000: 0000 0000 0100 0000 0000 0000 0000 0000
./hctx0/tags:nr_tags=128
./hctx0/tags:nr_reserved_tags=0
./hctx0/tags:active_queues=0
./hctx0/tags:bitmap_tags:
./hctx0/tags:depth=128
./hctx0/tags:busy=1
./hctx0/tags:cleared=0
./hctx0/tags:bits_per_word=32
./hctx0/tags:map_nr=4
./hctx0/tags:alloc_hint={123, 51}
./hctx0/tags:wake_batch=8
./hctx0/tags:wake_index=0
./hctx0/tags:ws_active=0
./hctx0/tags:ws={
./hctx0/tags: {.wait=inactive},
./hctx0/tags: {.wait=inactive},
./hctx0/tags: {.wait=inactive},
./hctx0/tags: {.wait=inactive},
./hctx0/tags: {.wait=inactive},
./hctx0/tags: {.wait=inactive},
./hctx0/tags: {.wait=inactive},
./hctx0/tags: {.wait=inactive},
./hctx0/tags:}
./hctx0/tags:round_robin=0
./hctx0/tags:min_shallow_depth=4294967295
./hctx0/ctx_map:00000000: 00
./hctx0/busy:ffff888016860000 {.op=READ, .cmd_flags=,
.rq_flags=STARTED|ELVPRIV|IO_STAT|STATS|ELV, .state=in_flight,
.tag=32, .internal_tag=0}
./hctx0/flags:alloc_policy=FIFO SHOULD_MERGE|BLOCKING
./sched/queued:0 1 0
./sched/owned_by_driver:0 1 0
./sched/async_depth:192
./sched/starved:0
./sched/batching:1
./state:SAME_COMP|NONROT|IO_STAT|INIT_DONE|STATS|REGISTERED|NOWAIT|30
./pm_only:0

So how can we know where the io is?

Regards,

Yang

Yu Kuai <yukuai1@xxxxxxxxxxxxxxx> 于2023年5月18日周四 11:30写道:
>
> Hi,
>
> 在 2023/05/18 0:27, yang lan 写道:
> > Hi,
> >
> > Thank you for your response.
> >
> >> Does this reproduce on current kernels, eg 6.4-rc2?
> >
> > Yeah, it can be reproduced on kernel 6.4-rc2.
> >
>
> Below log shows that io hang, can you collect following debugfs so
> that we can know where is the io now.
>
> cd /sys/kernel/debug/block/[test_device] && find . -type f -exec grep
> -aH . {} \;
>
> Thanks,
> Kuai
> > root@syzkaller:~# uname -a
> > Linux syzkaller 6.4.0-rc2 #1 SMP PREEMPT_DYNAMIC Wed May 17 22:58:52
> > CST 2023 x86_64 GNU/Linux
> > root@syzkaller:~# gcc poc_blkdev.c -o poc_blkdev
> > root@syzkaller:~# ./poc_blkdev
> > [ 128.718051][ T7121] nbd0: detected capacity change from 0 to 4
> > [ 158.917678][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 30 seconds
> > [ 188.997677][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 60 seconds
> > [ 219.077191][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 90 seconds
> > [ 249.157312][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 120 seconds
> > [ 279.237409][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 150 seconds
> > [ 309.317843][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 180 seconds
> > [ 339.397950][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 210 seconds
> > [ 369.478031][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 240 seconds
> > [ 399.558253][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 270 seconds
> > [ 429.638372][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 300 seconds
> > [ 459.718454][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 330 seconds
> > [ 489.798571][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 360 seconds
> > [ 519.878643][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 390 seconds
> > [ 549.958966][ T998] block nbd0: Possible stuck request
> > ffff888016f08000: control (read@0,2048B). Runtime 420 seconds
> > [ 571.719145][ T30] INFO: task systemd-udevd:7123 blocked for more
> > than 143 seconds.
> > [ 571.719652][ T30] Not tainted 6.4.0-rc2 #1
> > [ 571.719900][ T30] "echo 0 >
> > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
> > [ 571.720307][ T30] task:systemd-udevd state:D stack:26224
> > pid:7123 ppid:3998 flags:0x00004004
> > [ 571.720756][ T30] Call Trace:
> > [ 571.720923][ T30] <TASK>
> > [ 571.721073][ T30] __schedule+0x9ca/0x2630
> > [ 571.721348][ T30] ? firmware_map_remove+0x1e0/0x1e0
> > [ 571.721618][ T30] ? find_held_lock+0x33/0x1c0
> > [ 571.721866][ T30] ? lock_release+0x3b9/0x690
> > [ 571.722108][ T30] ? do_read_cache_folio+0x4ff/0xb20
> > [ 571.722447][ T30] ? lock_downgrade+0x6b0/0x6b0
> > [ 571.722785][ T30] ? mark_held_locks+0xb0/0x110
> > [ 571.723044][ T30] schedule+0xd3/0x1b0
> > [ 571.723264][ T30] io_schedule+0x1b/0x70
> > [ 571.723489][ T30] ? do_read_cache_folio+0x58c/0xb20
> > [ 571.723760][ T30] do_read_cache_folio+0x58c/0xb20
> > [ 571.724036][ T30] ? blkdev_readahead+0x20/0x20
> > [ 571.724319][ T30] ? __filemap_get_folio+0x8e0/0x8e0
> > [ 571.724588][ T30] ? __sanitizer_cov_trace_switch+0x53/0x90
> > [ 571.724885][ T30] ? __sanitizer_cov_trace_pc+0x1e/0x50
> > [ 571.725246][ T30] ? format_decode+0x1cf/0xb50
> > [ 571.725547][ T30] ? __sanitizer_cov_trace_pc+0x1e/0x50
> > [ 571.725837][ T30] ? fill_ptr_key+0x30/0x30
> > [ 571.726072][ T30] ? default_pointer+0x4a0/0x4a0
> > [ 571.726335][ T30] ? __isolate_free_page+0x220/0x220
> > [ 571.726608][ T30] ? filemap_fdatawrite_wbc+0x1c0/0x1c0
> > [ 571.726888][ T30] ? __sanitizer_cov_trace_pc+0x1e/0x50
> > [ 571.727172][ T30] ? read_part_sector+0x229/0x420
> > [ 571.727434][ T30] ? adfspart_check_ADFS+0x560/0x560
> > [ 571.727707][ T30] read_part_sector+0xfa/0x420
> > [ 571.727963][ T30] adfspart_check_POWERTEC+0x90/0x690
> > [ 571.728244][ T30] ? adfspart_check_ADFS+0x560/0x560
> > [ 571.728520][ T30] ? __kasan_slab_alloc+0x33/0x70
> > [ 571.728780][ T30] ? adfspart_check_ICS+0x8f0/0x8f0
> > [ 571.729889][ T30] ? snprintf+0xb2/0xe0
> > [ 571.730145][ T30] ? vsprintf+0x30/0x30
> > [ 571.730374][ T30] ? __sanitizer_cov_trace_pc+0x1e/0x50
> > [ 571.730659][ T30] ? adfspart_check_ICS+0x8f0/0x8f0
> > [ 571.730928][ T30] bdev_disk_changed+0x674/0x1260
> > [ 571.731189][ T30] ? write_comp_data+0x1f/0x70
> > [ 571.731439][ T30] ? iput+0xd0/0x780
> > [ 571.731646][ T30] blkdev_get_whole+0x186/0x260
> > [ 571.731886][ T30] blkdev_get_by_dev+0x4ce/0xae0
> > [ 571.732139][ T30] blkdev_open+0x140/0x2c0
> > [ 571.732366][ T30] do_dentry_open+0x6de/0x1450
> > [ 571.732612][ T30] ? blkdev_close+0x80/0x80
> > [ 571.732848][ T30] path_openat+0xd6d/0x26d0
> > [ 571.733084][ T30] ? lock_downgrade+0x6b0/0x6b0
> > [ 571.733336][ T30] ? vfs_path_lookup+0x110/0x110
> > [ 571.733591][ T30] do_filp_open+0x1bb/0x290
> > [ 571.733824][ T30] ? may_open_dev+0xf0/0xf0
> > [ 571.734061][ T30] ? __phys_addr_symbol+0x30/0x70
> > [ 571.734324][ T30] ? do_raw_spin_unlock+0x176/0x260
> > [ 571.734595][ T30] do_sys_openat2+0x5fd/0x980
> > [ 571.734837][ T30] ? file_open_root+0x3f0/0x3f0
> > [ 571.735087][ T30] ? seccomp_notify_ioctl+0xff0/0xff0
> > [ 571.735368][ T30] do_sys_open+0xce/0x140
> > [ 571.735596][ T30] ? filp_open+0x80/0x80
> > [ 571.735820][ T30] ? __secure_computing+0x1e3/0x340
> > [ 571.736090][ T30] do_syscall_64+0x38/0x80
> > [ 571.736325][ T30] entry_SYSCALL_64_after_hwframe+0x63/0xcd
> > [ 571.736626][ T30] RIP: 0033:0x7fb212210840
> > [ 571.736857][ T30] RSP: 002b:00007fffb37bbbe8 EFLAGS: 00000246
> > ORIG_RAX: 0000000000000002
> > [ 571.737269][ T30] RAX: ffffffffffffffda RBX: 0000560e09072e10
> > RCX: 00007fb212210840
> > [ 571.737651][ T30] RDX: 0000560e08e39fe3 RSI: 00000000000a0800
> > RDI: 0000560e090813b0
> > [ 571.738037][ T30] RBP: 00007fffb37bbd60 R08: 0000560e08e39670
> > R09: 0000000000000010
> > [ 571.738432][ T30] R10: 0000560e08e39d0c R11: 0000000000000246
> > R12: 00007fffb37bbcb0
> > [ 571.739563][ T30] R13: 0000560e09087a70 R14: 0000000000000003
> > R15: 000000000000000e
> > [ 571.739973][ T30] </TASK>
> > [ 571.740133][ T30]
> > [ 571.740133][ T30] Showing all locks held in the system:
> > [ 571.740495][ T30] 1 lock held by rcu_tasks_kthre/13:
> > [ 571.740758][ T30] #0: ffffffff8b6badd0
> > (rcu_tasks.tasks_gp_mutex){+.+.}-{3:3}, at:
> > rcu_tasks_one_gp+0x2b/0xdb0
> > [ 571.741301][ T30] 1 lock held by rcu_tasks_trace/14:
> > [ 571.741571][ T30] #0: ffffffff8b6baad0
> > (rcu_tasks_trace.tasks_gp_mutex){+.+.}-{3:3}, at:
> > rcu_tasks_one_gp+0x2b/0xdb0
> > [ 571.742134][ T30] 1 lock held by khungtaskd/30:
> > [ 571.742385][ T30] #0: ffffffff8b6bb960
> > (rcu_read_lock){....}-{1:2}, at: debug_show_all_locks+0x5b/0x300
> > [ 571.742947][ T30] 2 locks held by kworker/u8:0/50:
> > [ 571.743198][ T30] #0: ffff888016e7b138
> > ((wq_completion)nbd0-recv){+.+.}-{0:0}, at:
> > process_one_work+0x94b/0x17b0
> > [ 571.743809][ T30] #1: ffff888011e4fdd0
> > ((work_completion)(&args->work)){+.+.}-{0:0}, at:
> > process_one_work+0x984/0x17b0
> > [ 571.744393][ T30] 1 lock held by in:imklog/6784:
> > [ 571.744643][ T30] #0: ffff88801106e368
> > (&f->f_pos_lock){+.+.}-{3:3}, at: __fdget_pos+0xe9/0x100
> > [ 571.745122][ T30] 1 lock held by systemd-udevd/7123:
> > [ 571.745381][ T30] #0: ffff8880431854c8
> > (&disk->open_mutex){+.+.}-{3:3}, at: blkdev_get_by_dev+0x24b/0xae0
> > [ 571.745885][ T30]
> > [ 571.746008][ T30] =============================================
> > [ 571.746008][ T30]
> > [ 571.746424][ T30] NMI backtrace for cpu 1
> > [ 571.746642][ T30] CPU: 1 PID: 30 Comm: khungtaskd Not tainted 6.4.0-rc2 #1
> > [ 571.746989][ T30] Hardware name: QEMU Standard PC (i440FX + PIIX,
> > 1996), BIOS 1.12.0-1 04/01/2014
> > [ 571.747440][ T30] Call Trace:
> > [ 571.747606][ T30] <TASK>
> > [ 571.747764][ T30] dump_stack_lvl+0x91/0xf0
> > [ 571.747997][ T30] nmi_cpu_backtrace+0x21a/0x2b0
> > [ 571.748257][ T30] ? lapic_can_unplug_cpu+0xa0/0xa0
> > [ 571.748525][ T30] nmi_trigger_cpumask_backtrace+0x28c/0x2f0
> > [ 571.748830][ T30] watchdog+0xe4b/0x10c0
> > [ 571.749057][ T30] ? proc_dohung_task_timeout_secs+0x90/0x90
> > [ 571.749366][ T30] kthread+0x33b/0x430
> > [ 571.749596][ T30] ? kthread_complete_and_exit+0x40/0x40
> > [ 571.749891][ T30] ret_from_fork+0x1f/0x30
> > [ 571.750126][ T30] </TASK>
> > [ 571.750347][ T30] Sending NMI from CPU 1 to CPUs 0:
> > [ 571.750620][ C0] NMI backtrace for cpu 0
> > [ 571.750626][ C0] CPU: 0 PID: 3987 Comm: systemd-journal Not
> > tainted 6.4.0-rc2 #1
> > [ 571.750637][ C0] Hardware name: QEMU Standard PC (i440FX + PIIX,
> > 1996), BIOS 1.12.0-1 04/01/2014
> > [ 571.750643][ C0] RIP: 0033:0x7fb1d8c34bd1
> > [ 571.750652][ C0] Code: ed 4d 89 cf 75 a3 0f 1f 00 48 85 ed 75 4b
> > 48 8b 54 24 28 48 8b 44 24 18 48 8b 7c 24 20 48 29 da 48 8b 70 20 48
> > 0f af 54 24 08 <48> 83 c4 38 5b 5d 41 5c 41 5d 41 5e 41 5f e9 ac f2 04
> > 00 0f 1f 40
> > [ 571.750662][ C0] RSP: 002b:00007ffff9686c30 EFLAGS: 00000202
> > [ 571.750670][ C0] RAX: 00007ffff9686e50 RBX: 0000000000000002
> > RCX: 0000000000000010
> > [ 571.750677][ C0] RDX: 0000000000000010 RSI: 00007ffff9686d80
> > RDI: 00007ffff9686f20
> > [ 571.750683][ C0] RBP: 0000000000000000 R08: 0000000000000010
> > R09: 00007ffff9686d90
> > [ 571.750689][ C0] R10: 00007ffff9686fb0 R11: 00007fb1d8d6a060
> > R12: 00007ffff9686f30
> > [ 571.750696][ C0] R13: 00007fb1d9d20ee0 R14: 00007ffff9686f30
> > R15: 00007ffff9686d90
> > [ 571.750703][ C0] FS: 00007fb1da33d8c0 GS: 0000000000000000
> > [ 571.752358][ T30] Kernel panic - not syncing: hung_task: blocked tasks
> > [ 571.757337][ T30] CPU: 1 PID: 30 Comm: khungtaskd Not tainted 6.4.0-rc2 #1
> > [ 571.757686][ T30] Hardware name: QEMU Standard PC (i440FX + PIIX,
> > 1996), BIOS 1.12.0-1 04/01/2014
> > [ 571.758131][ T30] Call Trace:
> > [ 571.758302][ T30] <TASK>
> > [ 571.758462][ T30] dump_stack_lvl+0x91/0xf0
> > [ 571.758714][ T30] panic+0x62d/0x6a0
> > [ 571.758926][ T30] ? panic_smp_self_stop+0x90/0x90
> > [ 571.759188][ T30] ? preempt_schedule_common+0x1a/0xc0
> > [ 571.759486][ T30] ? preempt_schedule_thunk+0x1a/0x20
> > [ 571.759785][ T30] ? watchdog+0xc21/0x10c0
> > [ 571.760020][ T30] watchdog+0xc32/0x10c0
> > [ 571.760240][ T30] ? proc_dohung_task_timeout_secs+0x90/0x90
> > [ 571.760541][ T30] kthread+0x33b/0x430
> > [ 571.760753][ T30] ? kthread_complete_and_exit+0x40/0x40
> > [ 571.761052][ T30] ret_from_fork+0x1f/0x30
> > [ 571.761286][ T30] </TASK>
> > [ 571.761814][ T30] Kernel Offset: disabled
> > [ 571.762047][ T30] Rebooting in 86400 seconds..
> >
> >> You need to include poc_blkdev.c as part of your report.
> >
> > It's a little confusing and I'm sorry for that.
> > The poc_blkdev.c is exactly the C reproducer
> > (https://pastebin.com/raw/6mg7uF8W).
> >
> >> I suspect you've done something that is known to not work (as root,
> >> so we won't necessarily care). But I can't really say without seeing
> >> what you've done. Running syzkaller is an art, and most people aren't
> >> good at it. It takes a lot of work to submit good quality bug reports,
> >> see this article:
> >>
> >> https://blog.regehr.org/archives/2037
> >
> > I have read this article and thanks for your recommendations.
> > I'm not familiar with this module and I haven't figured out the root
> > cause of this bug yet.
> >
> > Regards,
> >
> > Yang
> >
> > Matthew Wilcox <willy@xxxxxxxxxxxxx> 于2023年5月17日周三 20:20写道:
> >>
> >> On Wed, May 17, 2023 at 07:12:23PM +0800, yang lan wrote:
> >>> root@syzkaller:~# uname -a
> >>> Linux syzkaller 5.10.179 #1 SMP PREEMPT Thu Apr 27 16:22:48 CST 2023
> >>
> >> Does this reproduce on current kernels, eg 6.4-rc2?
> >>
> >>> root@syzkaller:~# gcc poc_blkdev.c -o poc_blkdev
> >>
> >> You need to include poc_blkdev.c as part of your report.
> >>
> >>> Please let me know if I can provide any more information, and I hope I
> >>> didn't mess up this bug report.
> >>
> >> I suspect you've done something that is known to not work (as root,
> >> so we won't necessarily care). But I can't really say without seeing
> >> what you've done. Running syzkaller is an art, and most people aren't
> >> good at it. It takes a lot of work to submit good quality bug reports,
> >> see this article:
> >>
> >> https://blog.regehr.org/archives/2037
> >
> > Matthew Wilcox <willy@xxxxxxxxxxxxx> 于2023年5月17日周三 20:20写道:
> >>
> >> On Wed, May 17, 2023 at 07:12:23PM +0800, yang lan wrote:
> >>> root@syzkaller:~# uname -a
> >>> Linux syzkaller 5.10.179 #1 SMP PREEMPT Thu Apr 27 16:22:48 CST 2023
> >>
> >> Does this reproduce on current kernels, eg 6.4-rc2?
> >>
> >>> root@syzkaller:~# gcc poc_blkdev.c -o poc_blkdev
> >>
> >> You need to include poc_blkdev.c as part of your report.
> >>
> >>> Please let me know if I can provide any more information, and I hope I
> >>> didn't mess up this bug report.
> >>
> >> I suspect you've done something that is known to not work (as root,
> >> so we won't necessarily care). But I can't really say without seeing
> >> what you've done. Running syzkaller is an art, and most people aren't
> >> good at it. It takes a lot of work to submit good quality bug reports,
> >> see this article:
> >>
> >> https://blog.regehr.org/archives/2037
> > .
> >
>