Re: [PATCH 00/10] OOM Debug print selection and additional information
From: Qian Cai
Date: Tue Aug 27 2019 - 21:32:47 EST
> On Aug 27, 2019, at 9:13 PM, Edward Chron <echron@xxxxxxxxxx> wrote:
>
> On Tue, Aug 27, 2019 at 5:50 PM Qian Cai <cai@xxxxxx> wrote:
>>
>>
>>
>>> On Aug 27, 2019, at 8:23 PM, Edward Chron <echron@xxxxxxxxxx> wrote:
>>>
>>>
>>>
>>> On Tue, Aug 27, 2019 at 5:40 AM Qian Cai <cai@xxxxxx> wrote:
>>> On Mon, 2019-08-26 at 12:36 -0700, Edward Chron wrote:
>>>> This patch series provides code that works as a debug option through
>>>> debugfs to provide additional controls to limit how much information
>>>> gets printed when an OOM event occurs and or optionally print additional
>>>> information about slab usage, vmalloc allocations, user process memory
>>>> usage, the number of processes / tasks and some summary information
>>>> about these tasks (number runable, i/o wait), system information
>>>> (#CPUs, Kernel Version and other useful state of the system),
>>>> ARP and ND Cache entry information.
>>>>
>>>> Linux OOM can optionally provide a lot of information, what's missing?
>>>> ----------------------------------------------------------------------
>>>> Linux provides a variety of detailed information when an OOM event occurs
>>>> but has limited options to control how much output is produced. The
>>>> system related information is produced unconditionally and limited per
>>>> user process information is produced as a default enabled option. The
>>>> per user process information may be disabled.
>>>>
>>>> Slab usage information was recently added and is output only if slab
>>>> usage exceeds user memory usage.
>>>>
>>>> Many OOM events are due to user application memory usage sometimes in
>>>> combination with the use of kernel resource usage that exceeds what is
>>>> expected memory usage. Detailed information about how memory was being
>>>> used when the event occurred may be required to identify the root cause
>>>> of the OOM event.
>>>>
>>>> However, some environments are very large and printing all of the
>>>> information about processes, slabs and or vmalloc allocations may
>>>> not be feasible. For other environments printing as much information
>>>> about these as possible may be needed to root cause OOM events.
>>>>
>>>
>>> For more in-depth analysis of OOM events, people could use kdump to save a
>>> vmcore by setting "panic_on_oom", and then use the crash utility to analysis the
>>> vmcore which contains pretty much all the information you need.
>>>
>>> Certainly, this is the ideal. A full system dump would give you the maximum amount of
>>> information.
>>>
>>> Unfortunately some environments may lack space to store the dump,
>>
>> Kdump usually also support dumping to a remote target via NFS, SSH etc
>>
>>> let alone the time to dump the storage contents and restart the system. Some
>>
>> There is also âmakedumpfileâ that could compress and filter unwanted memory to reduce
>> the vmcore size and speed up the dumping process by utilizing multi-threads.
>>
>>> systems can take many minutes to fully boot up, to reset and reinitialize all the
>>> devices. So unfortunately this is not always an option, and we need an OOM Report.
>>
>> I am not sure how the system needs some minutes to reboot would be relevant for the
>> discussion here. The idea is to save a vmcore and it can be analyzed offline even on
>> another system as long as it having a matching âvmlinux.".
>>
>>
>
> If selecting a dump on an OOM event doesn't reboot the system and if
> it runs fast enough such
> that it doesn't slow processing enough to appreciably effect the
> system's responsiveness then
> then it would be ideal solution. For some it would be over kill but
> since it is an option it is a
> choice to consider or not.
It sounds like you are looking for more of this,
https://github.com/iovisor/bcc/blob/master/tools/oomkill.py