Re: [PATCH 8/5] llist: Remove cpu_relax() usage in cmpxchg loops

From: Peter Zijlstra
Date: Mon Sep 12 2011 - 11:09:31 EST


On Mon, 2011-09-12 at 10:47 -0400, Mathieu Desnoyers wrote:
> * Andi Kleen (andi@xxxxxxxxxxxxxx) wrote:
> > On Mon, Sep 12, 2011 at 04:05:58PM +0200, Peter Zijlstra wrote:
> > > Subject: llist: Remove cpu_relax() usage in cmpxchg loops
> > > From: Peter Zijlstra <a.p.zijlstra@xxxxxxxxx>
> > > Date: Mon Sep 12 15:50:49 CEST 2011
> > >
> > > Initial benchmarks show they're a net loss (2 socket wsm):
> > >
> >
> > May still save power.
>
> Looking at kernel/spinlock.c:
>
> void __lockfunc __raw_##op##_lock(locktype##_t *lock) \
> {
> [...]
> while (!raw_##op##_can_lock(lock) && (lock)->break_lock)\
> arch_##op##_relax(&lock->raw_lock); \
>
> so basically, in typical locking primitives (spinlock), it looks like
> lower power consumption is preferred over getting the raw maximal
> performance in fully contented scenarios.

Who says its about power consumption? That was a baseless claim made by
Andi which you propagate as a truth. Typically PAUSE is too short to
really save any power and the only gain of having it in loops is to
provide a window where another core on the cache domain can have a go.

If power consumption would be the prime concern Intel should fix their
damn mwait implementation so we can use that for locks. Local spinners +
mwait would give a very power efficient 'spin'-lock.

> So what is the rationale for making those lock-less lists retry scheme
> different from spinlocks here ?

PAUSE should help on the medium contended case (it would be pointless on
heavy contention), but hurt the lightly contended case. We have all
kinds of contention spinlocks in the kernel, but we don't as of yet have
very contended llist users in the kernel.

Furthermore, I would argue we should avoid growing them, significantly
contended atomic ops are bad, use a different scheme.

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/