Re: [PATCH 00/10] OOM Debug print selection and additional information
From: Edward Chron
Date: Thu Aug 29 2019 - 11:20:22 EST
On Thu, Aug 29, 2019 at 12:11 AM Michal Hocko <mhocko@xxxxxxxxxx> wrote:
> On Wed 28-08-19 12:46:20, Edward Chron wrote:
> > Our belief is if you really think eBPF is the preferred mechanism
> > then move OOM reporting to an eBPF.
> I've said that all this additional information has to be dynamically
> extensible rather than a part of the core kernel. Whether eBPF is the
> suitable tool, I do not know. I haven't explored that. There are other
> ways to inject code to the kernel. systemtap/kprobes, kernel modules and
> probably others.
For simple code injections eBPF or kprobe works and a tracepoint would
help with that. For example we could add our one line of task information
that we find very useful this way.
For adding controls to limit output for processes, slabs and vmalloc entries
it would be harder to inject code for that. Our solution was to use debugfs.
An alternate could to be add simple sysctl if using debugfs is not appropriate.
As our code illustrated this can be added without changing the existing report
in any substantive way. I think there is value in this and this is core to what
the OOM report should provide. Additional items may be add ons that are
environment specific but these are OOM reporting essentials IMHO.
> > I mentioned this before but I will reiterate this here.
> > So how do we get there? Let's look at the existing report which we know
> > has issues.
> > Other than a few essential OOM messages the OOM code should produce,
> > such as the Killed process message message sequence being included,
> > you could have the entire OOM report moved to an eBPF script and
> > therefore make it customizable, configurable or if you prefer programmable.
> I believe we should keep the current reporting in place and allow
> additional information via dynamic mechanism. Be it a registration
> mechanism that modules can hook into or other more dynamic way.
> The current reporting has proven to be useful in many typical oom
> situations in my past years of experience. It gives the rough state of
> the failing allocation, MM subsystem, tasks that are eligible and task
> that is killed so that you can understand why the event happened.
> I would argue that the eligible tasks should be printed on the opt-in
> bases because this is more of relict from the past when the victim
> selection was less deterministic. But that is another story.
> All the rest of dump_header should stay IMHO as a reasonable default and
> bare minimum.
> > Why? Because as we all agree, you'll never have a perfect OOM Report.
> > So if you believe this, than if you will, put your money where your mouth
> > is (so to speak) and make the entire OOM Report and eBPF script.
> > We'd be willing to help with this.
> > I'll give specific reasons why you want to do this.
> > - Don't want to maintain a lot of code in the kernel (eBPF code doesn't
> > count).
> > - Can't produce an ideal OOM report.
> > - Don't like configuring things but favor programmatic solutions.
> > - Agree the existing OOM report doesn't work for all environments.
> > - Want to allow flexibility but can't support everything people might
> > want.
> > - Then installing an eBPF for OOM Reporting isn't an option, it's
> > required.
> This is going into an extreme. We cannot serve all cases but that is
> true for any other heuristics/reporting in the kernel. We do care about
Unfortunately my argument for this is moot, this can't be done with
eBPF, at least not now.
> > The last reason is huge for people who live in a world with large data
> > centers. Data center managers are very conservative. They don't want to
> > deviate from standard operating procedure unless absolutely necessary.
> > If loading an OOM Report eBPF is standard to get OOM Reporting output,
> > then they'll accept that.
> I have already responded to this kind of argumentation elsewhere. This
> is not a relevant argument for any kernel implementation. This is a data
> process management process.
> Michal Hocko
> SUSE Labs