Re: kvm+nouveau induced lockdep gripe
From: Mike Galbraith
Date: Tue Oct 27 2020 - 02:05:13 EST
On Mon, 2020-10-26 at 20:53 +0100, Sebastian Andrzej Siewior wrote:
>
> Could you try this, please?
Nogo, first call of sched_setscheduler() is via kthread_create(). I
confirmed that nuking that (gratuitous user foot saving override) call
on top of moving sched_set_fifo() does shut it up, but that won't fly.
> --- a/kernel/irq/manage.c
> +++ b/kernel/irq/manage.c
> @@ -1155,6 +1155,8 @@ static int irq_thread(void *data)
> irqreturn_t (*handler_fn)(struct irq_desc *desc,
> struct irqaction *action);
>
> + sched_set_fifo(current);
> +
> if (force_irqthreads && test_bit(IRQTF_FORCED_THREAD,
> &action->thread_flags))
> handler_fn = irq_forced_thread_fn;
> @@ -1320,8 +1322,6 @@ setup_irq_thread(struct irqaction *new,
> if (IS_ERR(t))
> return PTR_ERR(t);
>
> - sched_set_fifo(t);
> -
> /*
> * We keep the reference to the task struct even if
> * the thread dies to avoid that the interrupt code
>
[ 150.926954] ======================================================
[ 150.926967] WARNING: possible circular locking dependency detected
[ 150.926981] 5.10.0.g3650b22-master #13 Tainted: G S E
[ 150.926993] ------------------------------------------------------
[ 150.927005] libvirtd/1833 is trying to acquire lock:
[ 150.927016] ffff921a45ed55a8 (&mm->mmap_lock#2){++++}-{3:3}, at: mpol_rebind_mm+0x1e/0x50
[ 150.927052]
but task is already holding lock:
[ 150.927064] ffffffffad585410 (&cpuset_rwsem){++++}-{0:0}, at: cpuset_attach+0x38/0x390
[ 150.927094]
which lock already depends on the new lock.
[ 150.927108]
the existing dependency chain (in reverse order) is:
[ 150.927122]
-> #3 (&cpuset_rwsem){++++}-{0:0}:
[ 150.927145] cpuset_read_lock+0x39/0xd0
[ 150.927160] __sched_setscheduler+0x456/0xa90
[ 150.927173] _sched_setscheduler+0x69/0x70
[ 150.927187] __kthread_create_on_node+0x114/0x170
[ 150.927205] kthread_create_on_node+0x37/0x40
[ 150.927223] setup_irq_thread+0x33/0xb0
[ 150.927238] __setup_irq+0x4e0/0x7c0
[ 150.927254] request_threaded_irq+0xf8/0x160
[ 150.927393] nvkm_pci_oneinit+0x4c/0x70 [nouveau]
[ 150.927469] nvkm_subdev_init+0x60/0x1e0 [nouveau]
[ 150.927603] nvkm_device_init+0x10b/0x240 [nouveau]
[ 150.927733] nvkm_udevice_init+0x49/0x70 [nouveau]
[ 150.927804] nvkm_object_init+0x3d/0x180 [nouveau]
[ 150.927871] nvkm_ioctl_new+0x1a1/0x260 [nouveau]
[ 150.927938] nvkm_ioctl+0x10a/0x240 [nouveau]
[ 150.928001] nvif_object_ctor+0xeb/0x150 [nouveau]
[ 150.928069] nvif_device_ctor+0x1f/0x60 [nouveau]
[ 150.928201] nouveau_cli_init+0x1ac/0x590 [nouveau]
[ 150.928332] nouveau_drm_device_init+0x68/0x800 [nouveau]
[ 150.928462] nouveau_drm_probe+0xfb/0x200 [nouveau]
[ 150.928483] local_pci_probe+0x42/0x90
[ 150.928501] pci_device_probe+0xe7/0x1a0
[ 150.928519] really_probe+0xf7/0x4d0
[ 150.928536] driver_probe_device+0x5d/0x140
[ 150.928552] device_driver_attach+0x4f/0x60
[ 150.928569] __driver_attach+0xa4/0x140
[ 150.928584] bus_for_each_dev+0x67/0x90
[ 150.928600] bus_add_driver+0x18c/0x230
[ 150.928616] driver_register+0x5b/0xf0
[ 150.928633] do_one_initcall+0x54/0x2f0
[ 150.928650] do_init_module+0x5b/0x21b
[ 150.928667] load_module+0x1e40/0x2370
[ 150.928682] __do_sys_finit_module+0x98/0xe0
[ 150.928698] do_syscall_64+0x33/0x40
[ 150.928716] entry_SYSCALL_64_after_hwframe+0x44/0xa9
[ 150.928731]
-> #2 (&device->mutex){+.+.}-{3:3}:
[ 150.928761] __mutex_lock+0x90/0x9c0
[ 150.928895] nvkm_udevice_fini+0x23/0x70 [nouveau]
[ 150.928975] nvkm_object_fini+0xb8/0x210 [nouveau]
[ 150.929048] nvkm_object_fini+0x73/0x210 [nouveau]
[ 150.929119] nvkm_ioctl_del+0x7e/0xa0 [nouveau]
[ 150.929191] nvkm_ioctl+0x10a/0x240 [nouveau]
[ 150.929258] nvif_object_dtor+0x4a/0x60 [nouveau]
[ 150.929326] nvif_client_dtor+0xe/0x40 [nouveau]
[ 150.929455] nouveau_cli_fini+0x7a/0x90 [nouveau]
[ 150.929586] nouveau_drm_postclose+0xaa/0xe0 [nouveau]
[ 150.929638] drm_file_free.part.7+0x273/0x2c0 [drm]
[ 150.929680] drm_release+0x6e/0xf0 [drm]
[ 150.929697] __fput+0xb2/0x260
[ 150.929714] task_work_run+0x73/0xc0
[ 150.929732] exit_to_user_mode_prepare+0x1a5/0x1d0
[ 150.929749] syscall_exit_to_user_mode+0x46/0x2a0
[ 150.929767] entry_SYSCALL_64_after_hwframe+0x44/0xa9
[ 150.929782]
-> #1 (&cli->lock){+.+.}-{3:3}:
[ 150.929811] __mutex_lock+0x90/0x9c0
[ 150.929936] nouveau_mem_fini+0x4c/0x70 [nouveau]
[ 150.930062] nouveau_sgdma_destroy+0x20/0x50 [nouveau]
[ 150.930086] ttm_bo_cleanup_memtype_use+0x3e/0x60 [ttm]
[ 150.930109] ttm_bo_release+0x29c/0x600 [ttm]
[ 150.930130] ttm_bo_vm_close+0x15/0x30 [ttm]
[ 150.930150] remove_vma+0x3e/0x70
[ 150.930166] __do_munmap+0x2b7/0x4f0
[ 150.930181] __vm_munmap+0x5b/0xa0
[ 150.930195] __x64_sys_munmap+0x27/0x30
[ 150.930210] do_syscall_64+0x33/0x40
[ 150.930227] entry_SYSCALL_64_after_hwframe+0x44/0xa9
[ 150.930241]
-> #0 (&mm->mmap_lock#2){++++}-{3:3}:
[ 150.930275] __lock_acquire+0x149d/0x1a70
[ 150.930292] lock_acquire+0x1a7/0x3b0
[ 150.930307] down_write+0x38/0x70
[ 150.930323] mpol_rebind_mm+0x1e/0x50
[ 150.930340] cpuset_attach+0x229/0x390
[ 150.930355] cgroup_migrate_execute+0x46d/0x490
[ 150.930371] cgroup_attach_task+0x20c/0x3b0
[ 150.930388] __cgroup1_procs_write.constprop.21+0xf3/0x150
[ 150.930407] cgroup_file_write+0x64/0x210
[ 150.930423] kernfs_fop_write+0x117/0x1b0
[ 150.930440] vfs_write+0xe8/0x240
[ 150.930456] ksys_write+0x87/0xc0
[ 150.930471] do_syscall_64+0x33/0x40
[ 150.930487] entry_SYSCALL_64_after_hwframe+0x44/0xa9
[ 150.930501]
other info that might help us debug this:
[ 150.930522] Chain exists of:
&mm->mmap_lock#2 --> &device->mutex --> &cpuset_rwsem
[ 150.930561] Possible unsafe locking scenario:
[ 150.930577] CPU0 CPU1
[ 150.930589] ---- ----
[ 150.930602] lock(&cpuset_rwsem);
[ 150.930617] lock(&device->mutex);
[ 150.930635] lock(&cpuset_rwsem);
[ 150.930653] lock(&mm->mmap_lock#2);
[ 150.930671]
*** DEADLOCK ***
[ 150.930690] 6 locks held by libvirtd/1833:
[ 150.930703] #0: ffff921a6f1690f0 (&f->f_pos_lock){+.+.}-{3:3}, at: __fdget_pos+0x45/0x50
[ 150.930736] #1: ffff921c88c1d460 (sb_writers#7){.+.+}-{0:0}, at: vfs_write+0x1aa/0x240
[ 150.930771] #2: ffff921a48734488 (&of->mutex){+.+.}-{3:3}, at: kernfs_fop_write+0xe2/0x1b0
[ 150.930801] #3: ffffffffad581848 (cgroup_mutex){+.+.}-{3:3}, at: cgroup_kn_lock_live+0xea/0x1d0
[ 150.930833] #4: ffffffffad5816b0 (cgroup_threadgroup_rwsem){++++}-{0:0}, at: cgroup_procs_write_start+0x6e/0x200
[ 150.930866] #5: ffffffffad585410 (&cpuset_rwsem){++++}-{0:0}, at: cpuset_attach+0x38/0x390
[ 150.930899]
stack backtrace:
[ 150.930918] CPU: 3 PID: 1833 Comm: libvirtd Kdump: loaded Tainted: G S E 5.10.0.g3650b22-master #13
[ 150.930938] Hardware name: MEDION MS-7848/MS-7848, BIOS M7848W08.20C 09/23/2013
[ 150.930955] Call Trace:
[ 150.930974] dump_stack+0x77/0x97
[ 150.930993] check_noncircular+0xe7/0x100
[ 150.931012] ? stack_trace_save+0x3b/0x50
[ 150.931036] ? __lock_acquire+0x149d/0x1a70
[ 150.931053] __lock_acquire+0x149d/0x1a70
[ 150.931077] lock_acquire+0x1a7/0x3b0
[ 150.931093] ? mpol_rebind_mm+0x1e/0x50
[ 150.931114] down_write+0x38/0x70
[ 150.931129] ? mpol_rebind_mm+0x1e/0x50
[ 150.931144] mpol_rebind_mm+0x1e/0x50
[ 150.931162] cpuset_attach+0x229/0x390
[ 150.931180] cgroup_migrate_execute+0x46d/0x490
[ 150.931199] ? _raw_spin_unlock_irq+0x2f/0x50
[ 150.931217] cgroup_attach_task+0x20c/0x3b0
[ 150.931245] ? __cgroup1_procs_write.constprop.21+0xf3/0x150
[ 150.931263] __cgroup1_procs_write.constprop.21+0xf3/0x150
[ 150.931286] cgroup_file_write+0x64/0x210
[ 150.931304] kernfs_fop_write+0x117/0x1b0
[ 150.931323] vfs_write+0xe8/0x240
[ 150.931341] ksys_write+0x87/0xc0
[ 150.931357] ? lockdep_hardirqs_on+0x85/0x110
[ 150.931374] do_syscall_64+0x33/0x40
[ 150.931391] entry_SYSCALL_64_after_hwframe+0x44/0xa9
[ 150.931409] RIP: 0033:0x7fcdc56a6deb
[ 150.931425] Code: 53 48 89 d5 48 89 f3 48 83 ec 18 48 89 7c 24 08 e8 5a fd ff ff 48 89 ea 41 89 c0 48 89 de 48 8b 7c 24 08 b8 01 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 31 44 89 c7 48 89 44 24 08 e8 90 fd ff ff 48
[ 150.931456] RSP: 002b:00007fcdbcfdc2f0 EFLAGS: 00000293 ORIG_RAX: 0000000000000001
[ 150.931476] RAX: ffffffffffffffda RBX: 00007fcdb403ca00 RCX: 00007fcdc56a6deb
[ 150.931493] RDX: 0000000000000004 RSI: 00007fcdb403ca00 RDI: 000000000000001f
[ 150.931510] RBP: 0000000000000004 R08: 0000000000000000 R09: 0000000000000000
[ 150.931526] R10: 0000000000000000 R11: 0000000000000293 R12: 00007fcdb403ca00
[ 150.931543] R13: 0000000000000000 R14: 000000000000001f R15: 0000000000000214