On Wed, Dec 13, 2017 at 11:25:13PM +0800, Quan Xu wrote:Â we still hit HLT.. we can use it with https://lkml.org/lkml/2017/8/29/279 ..
On Fri, Dec 8, 2017 at 11:10 PM, Konrad Rzeszutek Wilk <So:
konrad.wilk@xxxxxxxxxx> wrote:
On Fri, Dec 08, 2017 at 04:39:43PM +0800, Quan Xu wrote:one painful point is from VM idle path..
From: Ben Luo <bn0418@xxxxxxxxx>VM-exit
This patchset introduces a new paravirtualized mechanism to reduce
caused by guest timer accessing.And how bad is this blib in arming the timer?
And how often do you get this timer to be armed? OR better yet - what
are the workloads in which you found this VMExit to be painful?
Thanks. Or better yet - what
are the workloads in which you found this VMExit to be painful?
for some network req/resp services, or benchmark of process context
switches..
1) VM idle path and network req/resp services:
Does this go away if you don't hit the idle path? Meaning if you
loop without hitting HLT/MWAIT?
I am assuming the issue you are facingyes,
is the latency - that is first time the guest comes from HLT and
responds to the packet the latency is much higher than without?
And the arming of the timer?yes, it is also helpful to some timer-intensive services.
2) process context switches.
Is that related to the 1)? That is the 'schedule' call and the process
going to sleep waiting for an interrupt or timer?
This all sounds like issues with low-CPU usage workloads where you
need low latency responses?