Re: [PATCH 00/10] OOM Debug print selection and additional information
From: Edward Chron
Date: Tue Aug 27 2019 - 21:13:35 EST
On Tue, Aug 27, 2019 at 5:50 PM Qian Cai <cai@xxxxxx> wrote:
> > On Aug 27, 2019, at 8:23 PM, Edward Chron <echron@xxxxxxxxxx> wrote:
> > On Tue, Aug 27, 2019 at 5:40 AM Qian Cai <cai@xxxxxx> wrote:
> > On Mon, 2019-08-26 at 12:36 -0700, Edward Chron wrote:
> > > This patch series provides code that works as a debug option through
> > > debugfs to provide additional controls to limit how much information
> > > gets printed when an OOM event occurs and or optionally print additional
> > > information about slab usage, vmalloc allocations, user process memory
> > > usage, the number of processes / tasks and some summary information
> > > about these tasks (number runable, i/o wait), system information
> > > (#CPUs, Kernel Version and other useful state of the system),
> > > ARP and ND Cache entry information.
> > >
> > > Linux OOM can optionally provide a lot of information, what's missing?
> > > ----------------------------------------------------------------------
> > > Linux provides a variety of detailed information when an OOM event occurs
> > > but has limited options to control how much output is produced. The
> > > system related information is produced unconditionally and limited per
> > > user process information is produced as a default enabled option. The
> > > per user process information may be disabled.
> > >
> > > Slab usage information was recently added and is output only if slab
> > > usage exceeds user memory usage.
> > >
> > > Many OOM events are due to user application memory usage sometimes in
> > > combination with the use of kernel resource usage that exceeds what is
> > > expected memory usage. Detailed information about how memory was being
> > > used when the event occurred may be required to identify the root cause
> > > of the OOM event.
> > >
> > > However, some environments are very large and printing all of the
> > > information about processes, slabs and or vmalloc allocations may
> > > not be feasible. For other environments printing as much information
> > > about these as possible may be needed to root cause OOM events.
> > >
> > For more in-depth analysis of OOM events, people could use kdump to save a
> > vmcore by setting "panic_on_oom", and then use the crash utility to analysis the
> > vmcore which contains pretty much all the information you need.
> > Certainly, this is the ideal. A full system dump would give you the maximum amount of
> > information.
> > Unfortunately some environments may lack space to store the dump,
> Kdump usually also support dumping to a remote target via NFS, SSH etc
> > let alone the time to dump the storage contents and restart the system. Some
> There is also âmakedumpfileâ that could compress and filter unwanted memory to reduce
> the vmcore size and speed up the dumping process by utilizing multi-threads.
> > systems can take many minutes to fully boot up, to reset and reinitialize all the
> > devices. So unfortunately this is not always an option, and we need an OOM Report.
> I am not sure how the system needs some minutes to reboot would be relevant for the
> discussion here. The idea is to save a vmcore and it can be analyzed offline even on
> another system as long as it having a matching âvmlinux.".
If selecting a dump on an OOM event doesn't reboot the system and if
it runs fast enough such
that it doesn't slow processing enough to appreciably effect the
system's responsiveness then
then it would be ideal solution. For some it would be over kill but
since it is an option it is a
choice to consider or not.