Re: [PATCH v3 7/7] mm/gup: Allow VM_FAULT_RETRY for multiple times
From: Peter Xu
Date: Wed Sep 11 2019 - 23:05:49 EST
On Wed, Sep 11, 2019 at 10:47:59AM +0100, Linus Torvalds wrote:
> On Wed, Sep 11, 2019 at 8:11 AM Peter Xu <peterx@xxxxxxxxxx> wrote:
> >
> > This is the gup counterpart of the change that allows the
> > VM_FAULT_RETRY to happen for more than once. One thing to mention is
> > that we must check the fatal signal here before retry because the GUP
> > can be interrupted by that, otherwise we can loop forever.
>
> I still get nervous about the signal handling here.
>
> I'm not entirely sure we get it right even before your patch series.
>
> Right now, __get_user_pages() can return -ERESTARTSYS when it's killed:
>
> /*
> * If we have a pending SIGKILL, don't keep faulting pages and
> * potentially allocating memory.
> */
> if (fatal_signal_pending(current)) {
> ret = -ERESTARTSYS;
> goto out;
> }
>
> and I don't think your series changes that. And note how this is true
> _regardless_ of any FOLL_xyz flags (and we don't pass the
> FAULT_FLAG_xyz flags at all, they get generated deeper down if we
> actually end up faulting things in).
>
> So this part of the patch:
>
> + if (fatal_signal_pending(current))
> + goto out;
> +
> *locked = 1;
> - lock_dropped = true;
> down_read(&mm->mmap_sem);
> ret = __get_user_pages(tsk, mm, start, 1, flags | FOLL_TRIED,
> - pages, NULL, NULL);
> + pages, NULL, locked);
> + if (!*locked) {
> + /* Continue to retry until we succeeded */
> + BUG_ON(ret != 0);
> + goto retry;
>
> just makes me go "that can't be right". The fatal_signal_pending() is
> pointless and would probably better be something like
>
> if (down_read_killable(&mm->mmap_sem) < 0)
> goto out;
Thanks for noticing all these! I'd admit when I was working on the
series I didn't think & test very carefully with fatal signals but
mostly I'm making sure the normal signals should work especially for
processes like userfaultfd tracees so they won't hang death (I'm
always testing with GDB, and if without proper signal handle they do
hang death..).
I agree that we should probably replace the down_read() with the
killable version as you suggested. Though we might still want the
explicit check of fatal_signal_pending() to make sure we react even
faster because IMHO down_read_killable() does not really check signals
all the time but it just put us into killable state if we need to wait
for the mmap_sem. In other words, if we are always lucky that we get
the lock without even waiting anything then down_read_killable() will
still ignore the fatal signals forever.
>
> and then _after_ calling __get_user_pages(), the whole "negative error
> handling" should be more obvious.
>
> The BUG_ON(ret != 0) makes me nervous, but it might be fine (I guess
> the fatal signal handling has always been done before the lock is
> released?).
Yes it indeed looks nervous, though it's probably should be true in
all cases. Actually we already have checks like this, for example, in
current __get_user_pages_locked():
/* VM_FAULT_RETRY cannot return errors */
if (!*locked) {
BUG_ON(ret < 0);
BUG_ON(ret >= nr_pages);
}
And in the new retry path since we always pass in npages==1 so it must
be zero when VM_FAULT_RETRY.
While... When I'm looking into this more carefully I seem to have
found another bug that we might want to fix with hugetlbfs path:
diff --git a/mm/gup.c b/mm/gup.c
index 7230f60a68d6..29ee3de65fad 100644
--- a/mm/gup.c
+++ b/mm/gup.c
@@ -836,6 +836,16 @@ static long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
i = follow_hugetlb_page(mm, vma, pages, vmas,
&start, &nr_pages, i,
gup_flags, locked);
+ if (locked && *locked == 0) {
+ /*
+ * We've got a VM_FAULT_RETRY
+ * and we've lost mmap_sem.
+ * We must stop here.
+ */
+ BUG_ON(gup_flags & FOLL_NOWAIT);
+ BUG_ON(ret != 0);
+ goto out;
+ }
continue;
}
}
The problem is that if *locked==0 then we've lost the mmap_sem
already! Then we probably can't go any further before taking it back
again. With that, we should be able to keep the previous assumption
valid.
>
> But exactly *because* __get_user_pages() can already return on fatal
> signals, I think it should also set FAULT_FLAG_KILLABLE when faulting
> things in. I don't think it does right now, so it doesn't actually
> necessarily check fatal signals in a timely manner (not _during_ the
> fault, only _before_ it).
Probably correct again at least to me...
So for current gup we never use FAULT_FLAG_KILLABLE. However I can't
figure out a reason on why we should continue with anything if the
thread context is going to be destroyed after all. And since we're at
it, I also noticed that userfaultfd will react upon fatal signals even
without FAULT_FLAG_KILLABLE. So, here's the things that I think could
be good to have probably in my next post:
- Allow the gup code to always use FAULT_FLAG_KILLABLE as long as
FAULT_FLAG_ALLOW_RETRY && !FAULT_FLAG_NOWAIT (that should be when
"locked" parameter is passed into gup), and,
- With previous change, we touch up handle_userfault() to also respect
the fault flag, so instead of:
blocking_state = return_to_userland ? TASK_INTERRUPTIBLE :
TASK_KILLABLE;
We now let the blocking_state to be either (1) INTERRUPTIBLE, if
with the new FAULT_FLAG_INTERRUPTIBLE, or (2) KILLABLE, if with
FAULT_FLAG_KILLABLE, or (3) UNINTERRUPTIBLE. Though if we start to
use FAULT_FLAG_KILLABLE in gup codes then in most cases (both gup
and the default page fault flags that most archs use) the
userfaultfd code should also behave as before.
Does above make any sense?
There could also be considerations behind on the current model of
handling fatal signals that I totally overlooked. I would appreciate
if anyone can point that out if so.
>
> See what I'm reacting to?
>
> And maybe I'm wrong. Maybe I misread the code, and your changes. And
> at least some of these logic error predate your changes, I just was
> hoping that since this whole "react to signals" is very much what your
> patch series is working on, you'd look at this.
Yes, I'd be happy to (as long as they can be reviewed properly so I
can get a chance to fix all my potential mistakes :).
Thanks,
--
Peter Xu