On Mon, Aug 09, 2021 at 09:36:49PM +0800, Like Xu wrote:
On 6/8/2021 3:42 pm, Yang Weijiang wrote:
From: Like Xu <like.xu@xxxxxxxxxxxxxxx>
Arch LBRs are enabled by setting MSR_ARCH_LBR_CTL.LBREn to 1. A new guest
state field named "Guest IA32_LBR_CTL" is added to enhance guest LBR usage.
When guest Arch LBR is enabled, a guest LBR event will be created like the
model-specific LBR does. Clear guest LBR enable bit on host PMI handling so
guest can see expected config.
On processors that support Arch LBR, MSR_IA32_DEBUGCTLMSR[bit 0] has no
meaning. It can be written to 0 or 1, but reads will always return 0.
Like IA32_DEBUGCTL, IA32_ARCH_LBR_CTL msr is also preserved on INIT.
Regardless of the Arch LBR or legacy LBR, when the LBR_EN bit 0 of the
corresponding control MSR is set to 1, LBR recording will be enabled.
Signed-off-by: Like Xu <like.xu@xxxxxxxxxxxxxxx>
Signed-off-by: Yang Weijiang <weijiang.yang@xxxxxxxxx>
---
arch/x86/events/intel/lbr.c | 2 --
arch/x86/include/asm/msr-index.h | 1 +
arch/x86/include/asm/vmx.h | 2 ++
arch/x86/kvm/vmx/pmu_intel.c | 48 ++++++++++++++++++++++++++++----
arch/x86/kvm/vmx/vmx.c | 9 ++++++
5 files changed, 55 insertions(+), 7 deletions(-)
[...]
If KVM "trusts" user-space, then check with guest cpuid is OK.+static bool arch_lbr_ctl_is_valid(struct kvm_vcpu *vcpu, u64 ctl)
+{
+ unsigned int eax, ebx, ecx, edx;
+
+ if (!kvm_cpu_cap_has(X86_FEATURE_ARCH_LBR))
+ return false;
+
+ cpuid_count(0x1c, 0, &eax, &ebx, &ecx, &edx);
+ if (!(ebx & BIT(0)) && (ctl & ARCH_LBR_CTL_CPL))
+ return false;
+ if (!(ebx & BIT(2)) && (ctl & ARCH_LBR_CTL_STACK))
+ return false;
+ if (!(ebx & BIT(1)) && (ctl & ARCH_LBR_CTL_BRN_MASK))
+ return false;
+
+ return !(ctl & ~KVM_ARCH_LBR_CTL_MASK);
+}
Please check it with the *guest* cpuid entry.
But if user-space enable excessive controls, then check against guest
cpuid could make things mess.
Then I think it makes more sense to check against x86_pmu.lbr_xxx masks in above function
And it should remove the bits that are not supported by x86_pmu.lbr_ctl_mask before
vmcs_write64(...) if the guest value is a superset of the host value with
warning message.
for compatibility. What do you think of it?
+
static int intel_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
{
struct kvm_pmu *pmu = vcpu_to_pmu(vcpu);
@@ -392,6 +414,9 @@ static int intel_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
case MSR_ARCH_LBR_DEPTH:
msr_info->data = lbr_desc->records.nr;
return 0;
+ case MSR_ARCH_LBR_CTL:
+ msr_info->data = vmcs_read64(GUEST_IA32_LBR_CTL);
+ return 0;
default:
if ((pmc = get_gp_pmc(pmu, msr, MSR_IA32_PERFCTR0)) ||
(pmc = get_gp_pmc(pmu, msr, MSR_IA32_PMC0))) {
vmcs_write64(GUEST_IA32_DEBUGCTL, data);
[...]
OK, will add it.if (intel_pmu_lbr_is_enabled(vcpu) && !to_vmx(vcpu)->lbr_desc.event &&
(data & DEBUGCTLMSR_LBR))
@@ -4441,6 +4448,8 @@ static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
vmcs_writel(GUEST_SYSENTER_ESP, 0);
vmcs_writel(GUEST_SYSENTER_EIP, 0);
vmcs_write64(GUEST_IA32_DEBUGCTL, 0);
+ if (static_cpu_has(X86_FEATURE_ARCH_LBR))
+ vmcs_write64(GUEST_IA32_LBR_CTL, 0);
Please update dump_vmcs() to dump GUEST_IA32_LBR_CTL as well.
No, I didn't enable nested Arch LBR but unblocked some issues for nested case.
How about update the load_vmcs12_host_state() for GUEST_IA32_LBR_CTL
since you enabled the nested case in this patch set ?
}
kvm_set_rflags(vcpu, X86_EFLAGS_FIXED);