Re: [PATCH v12 00/12]
From: Petr Mladek
Date: Thu Oct 11 2018 - 08:48:33 EST
On Thu 2018-08-30 13:58:15, Miroslav Benes wrote:
> On Tue, 28 Aug 2018, Petr Mladek wrote:
>
> > livepatch: Atomic replace feature
> >
> > The atomic replace allows to create cumulative patches. They
> > are useful when you maintain many livepatches and want to remove
> > one that is lower on the stack. In addition it is very useful when
> > more patches touch the same function and there are dependencies
> > between them.
> >
> > This version does another big refactoring based on feedback against
> > v11[*]. In particular, it removes the registration step, changes
> > the API and handling of livepatch dependencies. The aim is
> > to keep the number of possible variants on a sane level.
> > It helps the keep the feature "easy" to use and maintain.
> >
> > [*] https://lkml.kernel.org/r/20180323120028.31451-1-pmladek@xxxxxxxx
>
> Hi,
>
> I've started to review the patch set. Running selftests with lockdep
> enabled gives me...
>
> ======================================================
> WARNING: possible circular locking dependency detected
> 4.17.0-rc1-klp_replace_v12-117114-gfedb3eba611d #218 Tainted: G
> K
> ------------------------------------------------------
> kworker/1:1/49 is trying to acquire lock:
> 00000000bb88dc17 (kn->count#186){++++}, at: kernfs_remove+0x23/0x40
>
> but task is already holding lock:
> 0000000073632424 (klp_mutex){+.+.}, at: klp_transition_work_fn+0x17/0x40
>
> which lock already depends on the new lock.
>
>
> the existing dependency chain (in reverse order) is:
>
> -> #1 (klp_mutex){+.+.}:
> lock_acquire+0xd4/0x220
> __mutex_lock+0x75/0x920
> mutex_lock_nested+0x1b/0x20
> enabled_store+0x47/0x150
> kobj_attr_store+0x12/0x20
> sysfs_kf_write+0x4a/0x60
> kernfs_fop_write+0x123/0x1b0
> __vfs_write+0x2b/0x150
> vfs_write+0xc7/0x1c0
> ksys_write+0x49/0xa0
> __x64_sys_write+0x1a/0x20
> do_syscall_64+0x62/0x1b0
> entry_SYSCALL_64_after_hwframe+0x49/0xbe
>
> -> #0 (kn->count#186){++++}:
> __lock_acquire+0xe9d/0x1240
> lock_acquire+0xd4/0x220
> __kernfs_remove+0x23c/0x2c0
> kernfs_remove+0x23/0x40
> sysfs_remove_dir+0x51/0x60
> kobject_del+0x18/0x50
> kobject_cleanup+0x4b/0x180
> kobject_put+0x2a/0x50
> __klp_free_patch+0x5b/0x60
> klp_free_patch_nowait+0x12/0x30
> klp_try_complete_transition+0x13e/0x180
> klp_transition_work_fn+0x26/0x40
> process_one_work+0x1d8/0x5d0
> worker_thread+0x4d/0x3d0
> kthread+0x113/0x150
> ret_from_fork+0x3a/0x50
>
> other info that might help us debug this:
>
> Possible unsafe locking scenario:
>
> CPU0 CPU1
> ---- ----
> lock(klp_mutex);
> lock(kn->count#186);
> lock(klp_mutex);
> lock(kn->count#186);
Sigh, I overestimated the power of kobjects. I thought that this
must have been a false positive but it was not.
1. kernfs_fop_write() ignores kobj->kref. It takes care only
of its own reference count, see kernfs_get_active().
2. kobj_put() takes care only of kobj->kref. The following
code is called when the reference count reaches zero:
+ kobj_put()
+ kref_put()
+ kobject_release()
+ kobject_cleanup()
+ kobject_del()
+ sysfs_remove_dir()
+ kernfs_remove()
+ __kernfs_remove().
+ kernfs_drain()
, where kernfs_drain() waits until all opened files
are closed.
Now, we call kobj_put() under klp_mutex() when the sysfs
interface still exists. Files can be opened for
writing. As a result:
+ enabled_store() might wait for klp_mutex
+ kernfs_drain() would wait for enabled_store()
with klp_mutex() taken.
I have reproduced this with some extra sleeps.
I am going to work on another solution.
Best Regards,
Petr