Re: High contention on the sk_buff_head.lock
From: Gregory Haskins
Date: Wed Mar 18 2009 - 23:46:50 EST
David Miller wrote:
> From: Gregory Haskins <ghaskins@xxxxxxxxxx>
> Date: Wed, 18 Mar 2009 17:54:04 -0400
>
>
>> Note that -rt doesnt typically context-switch under contention anymore
>> since we introduced adaptive-locks. Also note that the contention
>> against the lock is still contention, regardless of whether you have -rt
>> or not. Its just that the slow-path to handle the contended case for
>> -rt is more expensive than mainline. However, once you have the
>> contention as stated, you have already lost.
>>
>
> First, contention is not implicitly a bad thing.
>
However, when the contention in question is your top bottleneck, even
small improvements have the potential to yield large performance gains. :)
> Second, if the -rt kernel is doing adaptive spinning I see no
> reason why that adaptive spinning is not kicking in here
It does. Things would be *much* worse if it wasn't.
> to
> make this problem just go away.
>
Well, "go away" is a relative term. Before adaptive-locks, the box was
heavily context-switching under a network workload like we are
discussing, and that is where much of the performance was lost.
Adaptive-locks mitigate that particular problem so that spinlock_t
clients now more closely model mainline behavior (i.e. spin under
contention, at least when they can) and this brought -rt much closer to
what you expect for performance from mainline.
However, in -rt the contention path is, by necessity, still moderately
more heavy weight than a simple mainline spin (PI, etc), so any
contention will still hurt relatively more. And obviously
adaptive-locks do nothing to address the contention itself. Therefore,
as long as the contention remains it will have a higher impact in -rt.
But make no mistake: the contention can (and I believe does) affect both
trees.
To see this in action, try taking a moderately large smp system (8-way+)
and scaling the number of flows. At 1 flow the stack is generally quite
capable of maintaining line-rate (at least up to GigE). With two flows
this should result in each achieving roughly 50% of the line-rate, for a
sum total of line-rate. But as you add flows to the equation, the
aggregate bandwidth typically starts to drop off to be something less
than line-rate (many times its actually much less). And when this
happens, the contention in question is usually at the top of the charts
(thus all the interest in this thread).
> This lock is held for mere cycles, just to unlink an SKB from
> the networking qdisc, and then it is immediately released.
>
To clarify: I haven't been looking at the stack code since last summer,
so some things may have changed since then. However, the issue back
then from my perspective was the general backpressure in the qdisc
subsystem (e.g. dev->queue_lock), not just the skb head.lock per se.
This dev->queue_lock can be held for quite a long time, and the
mechanism in general scales poorly as the fabric speeds and core-counts
increase. (It is arguable whether you would even want another buffering
layer on any reasonably fast interconnect, anyway. It ultimately just
destabilizes the flows and is probably best left to the underlying
hardware which will typically have the proper facilities for
shaping/limiting, etc. However, that is a topic for another discussion ;).)
One approach that we found yielded some significant improvements here
was to bypass the qdisc outright and do per-cpu lockless queuing. One
thread/core could then simply aggregate the per-cpu buffers with only
minor contention between the egress (consuming) core and the producer
thread. I realize that this solution as-is may not be feasible for
production use. However, it did serve to prove the theory that backing
off cache-line pressures that scale with the core count *did* help the
stack make more forward progress per unit time (e.g. higher throughput)
as the number of flows increased. The idea was more or less to refactor
the egress/TX path to be more per-cpu, similar to the methodology of the
ingress/RX side.
You could probably adapt this same concept to work harmoniously with the
qdisc infrastructure (though as previously indicated, I am not sure if
we would *want* to ;)). Something to consider, anyway.
Regards,
-Greg
Attachment:
signature.asc
Description: OpenPGP digital signature