Re: [PATCH v9 10/26] arm64: kvm: Unmask PMR before entering guest

From: Julien Thierry
Date: Wed Jan 30 2019 - 09:58:33 EST




On 30/01/2019 12:07, Christoffer Dall wrote:
> On Mon, Jan 21, 2019 at 03:33:29PM +0000, Julien Thierry wrote:
>> Interrupts masked by ICC_PMR_EL1 will not be signaled to the CPU. This
>> means that hypervisor will not receive masked interrupts while running a
>> guest.
>>
>
> You could add to the commit description how this works overall,
> something along the lines of:
>
> We need to make sure that all maskable interrupts are masked from the
> time we call local_irq_disable() in the main run loop, and remain so
> until we call local_irq_enable() after returning from the guest, and we
> need to ensure that we see no interrupts at all (including pseudo-NMIs)
> in the middle of the VM world-switch, while at the same time we need to
> ensure we exit the guest when there are interrupts for the host.
>
> We can accomplish this with pseudo-NMIs enabled by:
> (1) local_irq_disable: set the priority mask
> (2) enter guest: set PSTATE.I
> (3) clear the priority mask
> (4) eret to guest
> (5) exit guest: set the priotiy mask
> clear PSTATE.I (and restore other host PSTATE bits)
> (6) local_irq_enable: clear the priority mask.
>

Thanks for the suggestion. I'll add it to the commit.

> Also, took me a while to realize that when we come back from the guest,
> we call local_daif_restore with DAIF_PROCCTX_NOIRQ, which actually does
> both of the things in (5).
>

Yes, I can imagine this being no that obvious. I'll add a comment above
the call to local_daif_restore() in kvm_arm_vhe_guest_exit().

>> Avoid this by making sure ICC_PMR_EL1 is unmasked when we enter a guest.
>>
>> Signed-off-by: Julien Thierry <julien.thierry@xxxxxxx>
>> Acked-by: Catalin Marinas <catalin.marinas@xxxxxxx>
>> Cc: Christoffer Dall <christoffer.dall@xxxxxxx>
>> Cc: Marc Zyngier <marc.zyngier@xxxxxxx>
>> Cc: Catalin Marinas <catalin.marinas@xxxxxxx>
>> Cc: Will Deacon <will.deacon@xxxxxxx>
>> Cc: kvmarm@xxxxxxxxxxxxxxxxxxxxx
>> ---
>> arch/arm64/include/asm/kvm_host.h | 12 ++++++++++++
>> arch/arm64/kvm/hyp/switch.c | 16 ++++++++++++++++
>> 2 files changed, 28 insertions(+)
>>
>> diff --git a/arch/arm64/include/asm/kvm_host.h b/arch/arm64/include/asm/kvm_host.h
>> index 7732d0b..a1f9f55 100644
>> --- a/arch/arm64/include/asm/kvm_host.h
>> +++ b/arch/arm64/include/asm/kvm_host.h
>> @@ -24,6 +24,7 @@
>>
>> #include <linux/types.h>
>> #include <linux/kvm_types.h>
>> +#include <asm/arch_gicv3.h>
>> #include <asm/cpufeature.h>
>> #include <asm/daifflags.h>
>> #include <asm/fpsimd.h>
>> @@ -474,6 +475,17 @@ static inline int kvm_arch_vcpu_run_pid_change(struct kvm_vcpu *vcpu)
>> static inline void kvm_arm_vhe_guest_enter(void)
>> {
>> local_daif_mask();
>> +
>> + /*
>> + * Having IRQs masked via PMR when entering the guest means the GIC
>> + * will not signal the CPU of interrupts of lower priority, and the
>> + * only way to get out will be via guest exceptions.
>> + * Naturally, we want to avoid this.
>> + */
>> + if (system_uses_irq_prio_masking()) {
>> + gic_write_pmr(GIC_PRIO_IRQON);
>> + dsb(sy);
>> + }
>> }
>>
>> static inline void kvm_arm_vhe_guest_exit(void)
>> diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c
>> index b0b1478..6a4c2d6 100644
>> --- a/arch/arm64/kvm/hyp/switch.c
>> +++ b/arch/arm64/kvm/hyp/switch.c
>> @@ -22,6 +22,7 @@
>>
>> #include <kvm/arm_psci.h>
>>
>> +#include <asm/arch_gicv3.h>
>> #include <asm/cpufeature.h>
>> #include <asm/kvm_asm.h>
>> #include <asm/kvm_emulate.h>
>> @@ -521,6 +522,17 @@ int __hyp_text __kvm_vcpu_run_nvhe(struct kvm_vcpu *vcpu)
>> struct kvm_cpu_context *guest_ctxt;
>> u64 exit_code;
>>
>> + /*
>> + * Having IRQs masked via PMR when entering the guest means the GIC
>> + * will not signal the CPU of interrupts of lower priority, and the
>> + * only way to get out will be via guest exceptions.
>> + * Naturally, we want to avoid this.
>> + */
>> + if (system_uses_irq_prio_masking()) {
>> + gic_write_pmr(GIC_PRIO_IRQON);
>> + dsb(sy);
>> + }
>> +
>> vcpu = kern_hyp_va(vcpu);
>>
>> host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context);
>> @@ -573,6 +585,10 @@ int __hyp_text __kvm_vcpu_run_nvhe(struct kvm_vcpu *vcpu)
>> */
>> __debug_switch_to_host(vcpu);
>>
>> + /* Returning to host will clear PSR.I, remask PMR if needed */
>> + if (system_uses_irq_prio_masking())
>> + gic_write_pmr(GIC_PRIO_IRQOFF);
>> +
>> return exit_code;
>> }
>>
>
> nit: you could consider moving the non-vhe part into a new
> kvm_arm_nvhe_guest_enter, for symmetry with the vhe part.
>

The idea is tempting, however on the one hand for VHE we do the guest
enter/exit around the VHE vcpu run call, but in the non-VHE the guest
enter/exit would be called from within the vcpu run call since we rely
on the hvc happening to mask daif before unmasking PMR.

So, I'm not sure we would gain that much symmetry with the VHE side.
Unless we move the kvm_arm_vhe_guest_enter()/exit() to VHE vcpu run call
(they are empty for arch/arm and only called in one site anyway...).

> Otherwise looks good to me:
>
> Reviewed-by: Christoffer Dall <christoffer.dall@xxxxxxx>
>

Thanks,

--
Julien Thierry