Re: [RFC][PATCH 2/2] PM: Rework handling of interrupts duringsuspend-resume

From: Ingo Molnar
Date: Mon Feb 23 2009 - 10:07:29 EST



* Rafael J. Wysocki <rjw@xxxxxxx> wrote:

> On Monday 23 February 2009, Ingo Molnar wrote:
> >
> > * Eric W. Biederman <ebiederm@xxxxxxxxxxxx> wrote:
> >
> > > > What makes s2ram fragile is not human failure but the
> > > > combination of a handful of physical property:
> > > >
> > > > 1) Psychology: shutting the lid or pushing the suspend button is
> > > > a deceivingly 'simple' action to the user. But under the
> > > > hood, a ton of stuff happens: we deinitialize a lot of
> > > > things, we go through _all hardware state_, and we do so in a
> > > > serial fashion. If just one piece fails to do the right
> > > > thing, the box might not resume. Still, the user expects this
> > > > 'simple' thing to just work, all the time. No excuses
> > > > accepted.
> > > >
> > > > 2) Length of code: To get a successful s2ram sequence the kernel
> > > > runs through tens of thousands of lines of code. Code which
> > > > never gets executed on a normal box - only if we s2ram. If
> > > > just one step fails, we get a hung box.
> > > >
> > > > 3) Debuggability: a lot of s2ram code runs with the console off,
> > > > making any bugs hard to debug. Furthermore we have no
> > > > meaningful persistent storage either for kernel bug messages.
> > > > The RTC trick of PM_DEBUG works but is a very narrow channel
> > > > of information and it takes a lot of time to debug a bug via
> > > > that method.
> > >
> > > Yep that is an issue.
> >
> > I'd also like to add #4:
> >
> > 4) One more thing that makes s2ram special is that when the
> > resume path finds hardware often in an even more
> > deinitialized form than during normal bootup. During
> > normal bootup the BIOS/firmware has at least done some
> > minimal bootstrap (to get the kernel loaded), which
> > makes life easier for the kernel.
> >
> > At s2ram stage we've got a completely pure hardware
> > init state, with very minimal firmware activation.
>
> This is very true and at least in some cases done on purpose,
> AFAICS, due to some timing constraints forced on HW vendors by
> M$, for example.

IMHO i think it's the technically sane thing to do. Personally i
trust the quirks of bare metal much more than the combined
quirks of firmware _and_ bare metal.

> > So many of the init and deinit problems and bugs we
> > only hit in the s2ram path - which dynamics is again
> > not helpful.
>
> Plus ACPI requires us to do additional things during
> suspend-resume that are not done on boot-shutdown and which
> have their own ordering requirements (not necessarily stated
> directly, but such that we have do discover experimentally).
> That also change from one BIOS to another.

We could perhaps do a few things here to trigger bugs sooner.

For example at driver init, instead of executing just
->driver_open(), we could execute:

->driver_open()
->driver_suspend()
->driver_resume()

I.e. we'd simulate a suspend+resume mini-step. This makes it
sure that the basic driver callbacks are sane. It is also
supposed to work because the driver is just being initialized.

This way certain types of bugs would not show up as difficult to
debug s2ram regressions - but would show up as 'boot hang' or
'boot crash' bugs.

This does not simulate the "big picture" resume machinery (the
dependencies, etc.), nor does it trigger any of the "hardware
got really turned off" effects that true resume will trigger -
but at least it offloads a portion of the testing space from
's2ram' to 'bootup' testing.

What's your feeling - what percentage of all s2ram regressions
in the last year or so could have been triggered this way? Lets
assume we had 100 regressions in that timeframe - would it be in
the 10 bugs range? Or much lower or much higher?

Ingo
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/