Re: [PATCH] x86/kvm: Introduce boot parameter no-kvm-pvipi

From: zhenwei pi
Date: Sun Oct 24 2021 - 23:18:20 EST


On 10/21/21 3:17 PM, zhenwei pi wrote:
On 10/21/21 1:03 PM, Wanpeng Li wrote:
On Thu, 21 Oct 2021 at 11:05, zhenwei pi <pizhenwei@xxxxxxxxxxxxx> wrote:


On 10/21/21 4:12 AM, Sean Christopherson wrote:
On Wed, Oct 20, 2021, Wanpeng Li wrote:
On Wed, 20 Oct 2021 at 20:08, zhenwei pi <pizhenwei@xxxxxxxxxxxxx> wrote:

Although host side exposes KVM PV SEND IPI feature to guest side,
guest should still have a chance to disable it.

A typicall case of this parameter:
If the host AMD server enables AVIC feature, the flat mode of APIC
get better performance in the guest.

Hmm, I didn't find enough valuable information in your posting. We
observe AMD a lot before.
https://lore.kernel.org/all/CANRm+Cx597FNRUCyVz1D=B6Vs2GX3Sw57X7Muk+yMpi_hb+v1w@xxxxxxxxxxxxxx/T/#u

I too would like to see numbers.  I suspect the answer is going to be that
AVIC performs poorly in CPU overcommit scenarios because of the cost of managing
the tables and handling "failed delivery" exits, but that AVIC does quite well
when vCPUs are pinned 1:1 and IPIs rarely require an exit to the host.


Test env:
CPU: AMD EPYC 7642 48-Core Processor

Kmod args(enable avic and disable nested):
modprobe kvm-amd nested=0 avic=1 npt=1

QEMU args(disable x2apic):
... -cpu host,x2apic=off ...

Benchmark tool:
https://github.com/bytedance/kvm-utils/tree/master/microbenchmark/apic-ipi

~# insmod apic_ipi.ko options=5 && dmesg -c

   apic_ipi: 1 NUMA node(s)
   apic_ipi: apic [flat]
   apic_ipi: apic->send_IPI[default_send_IPI_single+0x0/0x40]
   apic_ipi: apic->send_IPI_mask[kvm_send_ipi_mask+0x0/0x10]
   apic_ipi:     IPI[kvm_send_ipi_mask] from CPU[0] to CPU[1]
   apic_ipi:             total cycles 375671259, avg 3756
   apic_ipi:     IPI[flat_send_IPI_mask] from CPU[0] to CPU[1]
   apic_ipi:             total cycles 221961822, avg 2219


apic->send_IPI_mask[kvm_send_ipi_mask+0x0/0x10]
    -> This line show current send_IPI_mask is kvm_send_ipi_mask(because
of PV SEND IPI FEATURE)

apic_ipi:       IPI[kvm_send_ipi_mask] from CPU[0] to CPU[1]
apic_ipi:               total cycles 375671259, avg 3756
    -->These lines show the average cycles of each kvm_send_ipi_mask: 3756

apic_ipi:       IPI[flat_send_IPI_mask] from CPU[0] to CPU[1]
apic_ipi:               total cycles 221961822, avg 2219
    -->These lines show the average cycles of each flat_send_IPI_mask: 2219

Just single target IPI is not eough.

     Wanpeng


Benchmark smp_call_function_single (https://github.com/bytedance/kvm-utils/blob/master/microbenchmark/ipi-bench/ipi_bench.c):

 Test env:
 CPU: AMD EPYC 7642 48-Core Processor

 Kmod args(enable avic and disable nested):
 modprobe kvm-amd nested=0 avic=1 npt=1

 QEMU args(disable x2apic):
 ... -cpu host,x2apic=off ...

1> without no-kvm-pvipi:
ipi_bench_single wait[1], CPU0[NODE0] -> CPU1[NODE0], loop = 100000
     elapsed =        424945631 cycles, average =     4249 cycles
     ipitime =        385246136 cycles, average =     3852 cycles
ipi_bench_single wait[0], CPU0[NODE0] -> CPU1[NODE0], loop = 100000
     elapsed =        419057953 cycles, average =     4190 cycles

2> with no-kvm-pvipi:
ipi_bench_single wait[1], CPU0[NODE0] -> CPU1[NODE0], loop = 100000
     elapsed =        321756407 cycles, average =     3217 cycles
     ipitime =        299433550 cycles, average =     2994 cycles
ipi_bench_single wait[0], CPU0[NODE0] -> CPU1[NODE0], loop = 100000
     elapsed =        295382146 cycles, average =     2953 cycles


Hi, Wanpeng & Sean

Also benchmark redis(by 127.0.0.1) in a guest(2vCPU), 'no-kvm-pvipi' gets better performance.

Test env:
Host side: pin 2vCPU on 2core in a die.
Guest side: run command:
taskset -c 1 ./redis-server --appendonly no
taskset -c 0 ./redis-benchmark -h 127.0.0.1 -d 1024 -n 10000000 -t get

1> without no-kvm-pvipi:
redis QPS: 193203.12 requests per second
kvm_pv_send_ipi exit: ~18K/s

2> with no-kvm-pvipi:
redis QPS: 196028.47 requests per second
avic_incomplete_ipi_interception exit: ~5K/s

--
zhenwei pi