Re: [patch] mm, memcg: add oom killer delay

From: Johannes Weiner
Date: Mon Jun 03 2013 - 12:49:09 EST


On Mon, Jun 03, 2013 at 05:34:32PM +0200, Michal Hocko wrote:
> On Sat 01-06-13 02:11:51, Johannes Weiner wrote:
> [...]
> > From: Johannes Weiner <hannes@xxxxxxxxxxx>
> > Subject: [PATCH] memcg: more robust oom handling
> >
> > The memcg OOM handling is incredibly fragile because once a memcg goes
> > OOM, one task (kernel or userspace) is responsible for resolving the
> > situation. Every other task that gets caught trying to charge memory
> > gets stuck in a waitqueue while potentially holding various filesystem
> > and mm locks on which the OOM handling task may now deadlock.
> >
> > Do two things to charge attempts under OOM:
> >
> > 1. Do not trap system calls (buffered IO and friends), just return
> > -ENOMEM. Userspace should be able to handle this... right?
> >
> > 2. Do not trap page faults directly in the charging context. Instead,
> > remember the OOMing memcg in the task struct and fully unwind the
> > page fault stack first. Then synchronize the memcg OOM from
> > pagefault_out_of_memory()
>
> I think this should work and I really like it! Nice work Johannes, I
> never dared to go that deep and my opposite approach was also much more
> fragile.
>
> I am just afraid about all the other archs that do not support (from
> quick grep it looks like: blackfin, c6x, h8300, metag, mn10300,
> openrisc, score and tile). What would be an alternative for them?
> #ifdefs for the old code (something like ARCH_HAS_FAULT_OOM_RETRY)? This
> would be acceptable for me.

blackfin is NOMMU but I guess the others should be converted to the
proper OOM protocol anyway and not just kill the faulting task. I can
update them in the next version of the patch (series).

> > Not-quite-signed-off-by: Johannes Weiner <hannes@xxxxxxxxxxx>
> > ---
> > arch/x86/mm/fault.c | 2 +
> > include/linux/memcontrol.h | 6 +++
> > include/linux/sched.h | 6 +++
> > mm/memcontrol.c | 104 +++++++++++++++++++++++++--------------------
> > mm/oom_kill.c | 7 ++-
> > 5 files changed, 78 insertions(+), 47 deletions(-)
> >
> [...]
> > diff --git a/include/linux/sched.h b/include/linux/sched.h
> > index e692a02..cf60aef 100644
> > --- a/include/linux/sched.h
> > +++ b/include/linux/sched.h
> > @@ -1282,6 +1282,8 @@ struct task_struct {
> > * execve */
> > unsigned in_iowait:1;
> >
> > + unsigned in_userfault:1;
> > +
>
> [This is more a nit pick but before I forget while I am reading through
> the rest of the patch.]
>
> OK there is a lot of room around those bit fields but as this is only
> for memcg and you are enlarging the structure by the pointer then you
> can reuse bottom bit of memcg pointer.

I just didn't want to put anything in the arch code that looks too
memcgish, even though it's the only user right now. But granted, it
will also probably remain the only user for a while.

> > @@ -2085,56 +2087,76 @@ static void memcg_oom_recover(struct mem_cgroup *memcg)
> > }
> >
> > /*
> > - * try to call OOM killer. returns false if we should exit memory-reclaim loop.
> > + * try to call OOM killer
> > */
> > -static bool mem_cgroup_handle_oom(struct mem_cgroup *memcg, gfp_t mask,
> > - int order)
> > +static void mem_cgroup_oom(struct mem_cgroup *memcg,
> > + gfp_t mask, int order,
> > + bool in_userfault)
> > {
> > - struct oom_wait_info owait;
> > - bool locked, need_to_kill;
> > -
> > - owait.memcg = memcg;
> > - owait.wait.flags = 0;
> > - owait.wait.func = memcg_oom_wake_function;
> > - owait.wait.private = current;
> > - INIT_LIST_HEAD(&owait.wait.task_list);
> > - need_to_kill = true;
> > - mem_cgroup_mark_under_oom(memcg);
> > + bool locked, need_to_kill = true;
> >
> > /* At first, try to OOM lock hierarchy under memcg.*/
> > spin_lock(&memcg_oom_lock);
> > locked = mem_cgroup_oom_lock(memcg);
> > - /*
> > - * Even if signal_pending(), we can't quit charge() loop without
> > - * accounting. So, UNINTERRUPTIBLE is appropriate. But SIGKILL
> > - * under OOM is always welcomed, use TASK_KILLABLE here.
> > - */
> > - prepare_to_wait(&memcg_oom_waitq, &owait.wait, TASK_KILLABLE);
> > - if (!locked || memcg->oom_kill_disable)
> > + if (!locked || memcg->oom_kill_disable) {
> > need_to_kill = false;
> > + if (in_userfault) {
> > + /*
> > + * We start sleeping on the OOM waitqueue only
> > + * after unwinding the page fault stack, so
> > + * make sure we detect wakeups that happen
> > + * between now and then.
> > + */
> > + mem_cgroup_mark_under_oom(memcg);
> > + current->memcg_oom.wakeups =
> > + atomic_read(&memcg->oom_wakeups);
> > + css_get(&memcg->css);
> > + current->memcg_oom.memcg = memcg;
> > + }
> > + }
> > if (locked)
> > mem_cgroup_oom_notify(memcg);
> > spin_unlock(&memcg_oom_lock);
> >
> > if (need_to_kill) {
> > - finish_wait(&memcg_oom_waitq, &owait.wait);
> > mem_cgroup_out_of_memory(memcg, mask, order);
> > - } else {
> > - schedule();
> > - finish_wait(&memcg_oom_waitq, &owait.wait);
> > + memcg_oom_recover(memcg);
>
> Why do we need to call memcg_oom_recover here? We do not know that any
> charges have been released. Say mem_cgroup_out_of_memory selected a task
> which migrated to our group (without its charges) so we would kill the
> poor guy and free no memory from this group.
> Now you wake up oom waiters to refault but they will end up in the same
> situation. I think it should be sufficient to wait for memcg_oom_recover
> until the memory is uncharged (which we do already).

It's a leftover from how it was before (see the memcg_wakeup_oom
below), but you are right, we can get rid of it.

> > }
> > - spin_lock(&memcg_oom_lock);
> > - if (locked)
> > +
> > + if (locked) {
> > + spin_lock(&memcg_oom_lock);
> > mem_cgroup_oom_unlock(memcg);
> > - memcg_wakeup_oom(memcg);
> > - spin_unlock(&memcg_oom_lock);
> > + spin_unlock(&memcg_oom_lock);
> > + }
> > +}
> >
> > - mem_cgroup_unmark_under_oom(memcg);
> [...]
> > @@ -2647,16 +2665,12 @@ again:
> > css_put(&memcg->css);
> > goto nomem;
> > case CHARGE_NOMEM: /* OOM routine works */
> > - if (!oom) {
> > + if (!oom || oom_check) {
>
> OK, this allows us to remove the confusing nr_oom_retries = MEM_CGROUP_RECLAIM_RETRIES
> from the branch where oom_check is set to true

Oops, will update.

Thanks for the review, Michal!
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/