Re: Hard Hang with __alloc_pages: 0-order allocation failed (gfp=0x20/1) - Not out of memory

From: Marcelo Tosatti
Date: Tue May 25 2004 - 17:53:20 EST


On Tue, May 25, 2004 at 02:47:59PM -0700, David S. Miller wrote:
> On Tue, 25 May 2004 14:20:23 -0700
> "Feldman, Scott" <scott.feldman@xxxxxxxxx> wrote:
>
> > Marcelo Tosatti wrote:
> >
> > > It seems we are calling alloc_skb(GFP_KERNEL) from inside an
> > > interrupt handler. Oops.
> >
> > We're calling dev_alloc_skb() from hard interrupt context, but it uses
> > GFP_ATOMIC, not GFP_KERNEL, so this is OK, right? I don't see the
> > problem with e1000.
>
> Neither do I, where is the detailed backtrace of this GFP_KERNEL
> allocation supposedly from interrupt context?

That was just a very wrong guess, I must admit.

Actually what seems to be happening is an e1000 IRQ while
trying to free memory (shrink_caches()) which tries to
allocate more memory.

Locked caused by extremely high load it seems.

You know better than me.
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/