Re: [PATCH RFC 03/11] KVM: VMX: Split off vmx_onhyperv.{ch} from hyperv.{ch}

From: Maxim Levitsky
Date: Thu Oct 12 2023 - 15:37:24 EST


У вт, 2023-10-10 у 18:02 +0200, Vitaly Kuznetsov пише:
> hyperv.{ch} is currently a mix of stuff which is needed by both Hyper-V on
> KVM and KVM on Hyper-V. As a preparation to making Hyper-V emulation
> optional, put KVM-on-Hyper-V specific code into dedicated files.
>
> No functional change intended.
>
> Signed-off-by: Vitaly Kuznetsov <vkuznets@xxxxxxxxxx>
> ---
> arch/x86/kvm/Makefile | 4 +
> arch/x86/kvm/vmx/hyperv.c | 139 --------------------
> arch/x86/kvm/vmx/hyperv.h | 217 ++++++++++++++++----------------
> arch/x86/kvm/vmx/vmx.c | 1 +
> arch/x86/kvm/vmx/vmx_onhyperv.c | 36 ++++++
> arch/x86/kvm/vmx/vmx_onhyperv.h | 124 ++++++++++++++++++
> arch/x86/kvm/vmx/vmx_ops.h | 2 +-
> 7 files changed, 271 insertions(+), 252 deletions(-)
> create mode 100644 arch/x86/kvm/vmx/vmx_onhyperv.c
> create mode 100644 arch/x86/kvm/vmx/vmx_onhyperv.h
>
> diff --git a/arch/x86/kvm/Makefile b/arch/x86/kvm/Makefile
> index 80e3fe184d17..a99ffc3f3a3f 100644
> --- a/arch/x86/kvm/Makefile
> +++ b/arch/x86/kvm/Makefile
> @@ -26,6 +26,10 @@ kvm-intel-y += vmx/vmx.o vmx/vmenter.o vmx/pmu_intel.o vmx/vmcs12.o \
> vmx/hyperv.o vmx/nested.o vmx/posted_intr.o
> kvm-intel-$(CONFIG_X86_SGX_KVM) += vmx/sgx.o
>
> +ifdef CONFIG_HYPERV
> +kvm-intel-y += vmx/vmx_onhyperv.o
> +endif
> +
> kvm-amd-y += svm/svm.o svm/vmenter.o svm/pmu.o svm/nested.o svm/avic.o \
> svm/sev.o svm/hyperv.o
>
> diff --git a/arch/x86/kvm/vmx/hyperv.c b/arch/x86/kvm/vmx/hyperv.c
> index 313b8bb5b8a7..de13dc14fe1d 100644
> --- a/arch/x86/kvm/vmx/hyperv.c
> +++ b/arch/x86/kvm/vmx/hyperv.c
> @@ -13,111 +13,6 @@
>
> #define CC KVM_NESTED_VMENTER_CONSISTENCY_CHECK
>
> -/*
> - * Enlightened VMCSv1 doesn't support these:
> - *
> - * POSTED_INTR_NV = 0x00000002,
> - * GUEST_INTR_STATUS = 0x00000810,
> - * APIC_ACCESS_ADDR = 0x00002014,
> - * POSTED_INTR_DESC_ADDR = 0x00002016,
> - * EOI_EXIT_BITMAP0 = 0x0000201c,
> - * EOI_EXIT_BITMAP1 = 0x0000201e,
> - * EOI_EXIT_BITMAP2 = 0x00002020,
> - * EOI_EXIT_BITMAP3 = 0x00002022,
> - * GUEST_PML_INDEX = 0x00000812,
> - * PML_ADDRESS = 0x0000200e,
> - * VM_FUNCTION_CONTROL = 0x00002018,
> - * EPTP_LIST_ADDRESS = 0x00002024,
> - * VMREAD_BITMAP = 0x00002026,
> - * VMWRITE_BITMAP = 0x00002028,
> - *
> - * TSC_MULTIPLIER = 0x00002032,
> - * PLE_GAP = 0x00004020,
> - * PLE_WINDOW = 0x00004022,
> - * VMX_PREEMPTION_TIMER_VALUE = 0x0000482E,
> - *
> - * Currently unsupported in KVM:
> - * GUEST_IA32_RTIT_CTL = 0x00002814,
> - */
> -#define EVMCS1_SUPPORTED_PINCTRL \
> - (PIN_BASED_ALWAYSON_WITHOUT_TRUE_MSR | \
> - PIN_BASED_EXT_INTR_MASK | \
> - PIN_BASED_NMI_EXITING | \
> - PIN_BASED_VIRTUAL_NMIS)
> -
> -#define EVMCS1_SUPPORTED_EXEC_CTRL \
> - (CPU_BASED_ALWAYSON_WITHOUT_TRUE_MSR | \
> - CPU_BASED_HLT_EXITING | \
> - CPU_BASED_CR3_LOAD_EXITING | \
> - CPU_BASED_CR3_STORE_EXITING | \
> - CPU_BASED_UNCOND_IO_EXITING | \
> - CPU_BASED_MOV_DR_EXITING | \
> - CPU_BASED_USE_TSC_OFFSETTING | \
> - CPU_BASED_MWAIT_EXITING | \
> - CPU_BASED_MONITOR_EXITING | \
> - CPU_BASED_INVLPG_EXITING | \
> - CPU_BASED_RDPMC_EXITING | \
> - CPU_BASED_INTR_WINDOW_EXITING | \
> - CPU_BASED_CR8_LOAD_EXITING | \
> - CPU_BASED_CR8_STORE_EXITING | \
> - CPU_BASED_RDTSC_EXITING | \
> - CPU_BASED_TPR_SHADOW | \
> - CPU_BASED_USE_IO_BITMAPS | \
> - CPU_BASED_MONITOR_TRAP_FLAG | \
> - CPU_BASED_USE_MSR_BITMAPS | \
> - CPU_BASED_NMI_WINDOW_EXITING | \
> - CPU_BASED_PAUSE_EXITING | \
> - CPU_BASED_ACTIVATE_SECONDARY_CONTROLS)
> -
> -#define EVMCS1_SUPPORTED_2NDEXEC \
> - (SECONDARY_EXEC_VIRTUALIZE_X2APIC_MODE | \
> - SECONDARY_EXEC_WBINVD_EXITING | \
> - SECONDARY_EXEC_ENABLE_VPID | \
> - SECONDARY_EXEC_ENABLE_EPT | \
> - SECONDARY_EXEC_UNRESTRICTED_GUEST | \
> - SECONDARY_EXEC_DESC | \
> - SECONDARY_EXEC_ENABLE_RDTSCP | \
> - SECONDARY_EXEC_ENABLE_INVPCID | \
> - SECONDARY_EXEC_ENABLE_XSAVES | \
> - SECONDARY_EXEC_RDSEED_EXITING | \
> - SECONDARY_EXEC_RDRAND_EXITING | \
> - SECONDARY_EXEC_TSC_SCALING | \
> - SECONDARY_EXEC_ENABLE_USR_WAIT_PAUSE | \
> - SECONDARY_EXEC_PT_USE_GPA | \
> - SECONDARY_EXEC_PT_CONCEAL_VMX | \
> - SECONDARY_EXEC_BUS_LOCK_DETECTION | \
> - SECONDARY_EXEC_NOTIFY_VM_EXITING | \
> - SECONDARY_EXEC_ENCLS_EXITING)
> -
> -#define EVMCS1_SUPPORTED_3RDEXEC (0ULL)
> -
> -#define EVMCS1_SUPPORTED_VMEXIT_CTRL \
> - (VM_EXIT_ALWAYSON_WITHOUT_TRUE_MSR | \
> - VM_EXIT_SAVE_DEBUG_CONTROLS | \
> - VM_EXIT_ACK_INTR_ON_EXIT | \
> - VM_EXIT_HOST_ADDR_SPACE_SIZE | \
> - VM_EXIT_LOAD_IA32_PERF_GLOBAL_CTRL | \
> - VM_EXIT_SAVE_IA32_PAT | \
> - VM_EXIT_LOAD_IA32_PAT | \
> - VM_EXIT_SAVE_IA32_EFER | \
> - VM_EXIT_LOAD_IA32_EFER | \
> - VM_EXIT_CLEAR_BNDCFGS | \
> - VM_EXIT_PT_CONCEAL_PIP | \
> - VM_EXIT_CLEAR_IA32_RTIT_CTL)
> -
> -#define EVMCS1_SUPPORTED_VMENTRY_CTRL \
> - (VM_ENTRY_ALWAYSON_WITHOUT_TRUE_MSR | \
> - VM_ENTRY_LOAD_DEBUG_CONTROLS | \
> - VM_ENTRY_IA32E_MODE | \
> - VM_ENTRY_LOAD_IA32_PERF_GLOBAL_CTRL | \
> - VM_ENTRY_LOAD_IA32_PAT | \
> - VM_ENTRY_LOAD_IA32_EFER | \
> - VM_ENTRY_LOAD_BNDCFGS | \
> - VM_ENTRY_PT_CONCEAL_PIP | \
> - VM_ENTRY_LOAD_IA32_RTIT_CTL)
> -
> -#define EVMCS1_SUPPORTED_VMFUNC (0)
> -
> #define EVMCS1_OFFSET(x) offsetof(struct hv_enlightened_vmcs, x)
> #define EVMCS1_FIELD(number, name, clean_field)[ROL16(number, 6)] = \
> {EVMCS1_OFFSET(name), clean_field}
> @@ -608,40 +503,6 @@ int nested_evmcs_check_controls(struct vmcs12 *vmcs12)
> return 0;
> }
>
> -#if IS_ENABLED(CONFIG_HYPERV)
> -DEFINE_STATIC_KEY_FALSE(__kvm_is_using_evmcs);
> -
> -/*
> - * KVM on Hyper-V always uses the latest known eVMCSv1 revision, the assumption
> - * is: in case a feature has corresponding fields in eVMCS described and it was
> - * exposed in VMX feature MSRs, KVM is free to use it. Warn if KVM meets a
> - * feature which has no corresponding eVMCS field, this likely means that KVM
> - * needs to be updated.
> - */
> -#define evmcs_check_vmcs_conf(field, ctrl) \
> - do { \
> - typeof(vmcs_conf->field) unsupported; \
> - \
> - unsupported = vmcs_conf->field & ~EVMCS1_SUPPORTED_ ## ctrl; \
> - if (unsupported) { \
> - pr_warn_once(#field " unsupported with eVMCS: 0x%llx\n",\
> - (u64)unsupported); \
> - vmcs_conf->field &= EVMCS1_SUPPORTED_ ## ctrl; \
> - } \
> - } \
> - while (0)
> -
> -void evmcs_sanitize_exec_ctrls(struct vmcs_config *vmcs_conf)
> -{
> - evmcs_check_vmcs_conf(cpu_based_exec_ctrl, EXEC_CTRL);
> - evmcs_check_vmcs_conf(pin_based_exec_ctrl, PINCTRL);
> - evmcs_check_vmcs_conf(cpu_based_2nd_exec_ctrl, 2NDEXEC);
> - evmcs_check_vmcs_conf(cpu_based_3rd_exec_ctrl, 3RDEXEC);
> - evmcs_check_vmcs_conf(vmentry_ctrl, VMENTRY_CTRL);
> - evmcs_check_vmcs_conf(vmexit_ctrl, VMEXIT_CTRL);
> -}
> -#endif
> -
> int nested_enable_evmcs(struct kvm_vcpu *vcpu,
> uint16_t *vmcs_version)
> {
> diff --git a/arch/x86/kvm/vmx/hyperv.h b/arch/x86/kvm/vmx/hyperv.h
> index 9623fe1651c4..9401dbfaea7c 100644
> --- a/arch/x86/kvm/vmx/hyperv.h
> +++ b/arch/x86/kvm/vmx/hyperv.h
> @@ -14,12 +14,113 @@
> #include "vmcs.h"
> #include "vmcs12.h"
>
> -struct vmcs_config;
> -
> -#define current_evmcs ((struct hv_enlightened_vmcs *)this_cpu_read(current_vmcs))
> -
> #define KVM_EVMCS_VERSION 1
>
> +/*
> + * Enlightened VMCSv1 doesn't support these:
> + *
> + * POSTED_INTR_NV = 0x00000002,
> + * GUEST_INTR_STATUS = 0x00000810,
> + * APIC_ACCESS_ADDR = 0x00002014,
> + * POSTED_INTR_DESC_ADDR = 0x00002016,
> + * EOI_EXIT_BITMAP0 = 0x0000201c,
> + * EOI_EXIT_BITMAP1 = 0x0000201e,
> + * EOI_EXIT_BITMAP2 = 0x00002020,
> + * EOI_EXIT_BITMAP3 = 0x00002022,
> + * GUEST_PML_INDEX = 0x00000812,
> + * PML_ADDRESS = 0x0000200e,
> + * VM_FUNCTION_CONTROL = 0x00002018,
> + * EPTP_LIST_ADDRESS = 0x00002024,
> + * VMREAD_BITMAP = 0x00002026,
> + * VMWRITE_BITMAP = 0x00002028,
> + *
> + * TSC_MULTIPLIER = 0x00002032,
> + * PLE_GAP = 0x00004020,
> + * PLE_WINDOW = 0x00004022,
> + * VMX_PREEMPTION_TIMER_VALUE = 0x0000482E,
> + *
> + * Currently unsupported in KVM:
> + * GUEST_IA32_RTIT_CTL = 0x00002814,
> + */
> +#define EVMCS1_SUPPORTED_PINCTRL \
> + (PIN_BASED_ALWAYSON_WITHOUT_TRUE_MSR | \
> + PIN_BASED_EXT_INTR_MASK | \
> + PIN_BASED_NMI_EXITING | \
> + PIN_BASED_VIRTUAL_NMIS)
> +
> +#define EVMCS1_SUPPORTED_EXEC_CTRL \
> + (CPU_BASED_ALWAYSON_WITHOUT_TRUE_MSR | \
> + CPU_BASED_HLT_EXITING | \
> + CPU_BASED_CR3_LOAD_EXITING | \
> + CPU_BASED_CR3_STORE_EXITING | \
> + CPU_BASED_UNCOND_IO_EXITING | \
> + CPU_BASED_MOV_DR_EXITING | \
> + CPU_BASED_USE_TSC_OFFSETTING | \
> + CPU_BASED_MWAIT_EXITING | \
> + CPU_BASED_MONITOR_EXITING | \
> + CPU_BASED_INVLPG_EXITING | \
> + CPU_BASED_RDPMC_EXITING | \
> + CPU_BASED_INTR_WINDOW_EXITING | \
> + CPU_BASED_CR8_LOAD_EXITING | \
> + CPU_BASED_CR8_STORE_EXITING | \
> + CPU_BASED_RDTSC_EXITING | \
> + CPU_BASED_TPR_SHADOW | \
> + CPU_BASED_USE_IO_BITMAPS | \
> + CPU_BASED_MONITOR_TRAP_FLAG | \
> + CPU_BASED_USE_MSR_BITMAPS | \
> + CPU_BASED_NMI_WINDOW_EXITING | \
> + CPU_BASED_PAUSE_EXITING | \
> + CPU_BASED_ACTIVATE_SECONDARY_CONTROLS)
> +
> +#define EVMCS1_SUPPORTED_2NDEXEC \
> + (SECONDARY_EXEC_VIRTUALIZE_X2APIC_MODE | \
> + SECONDARY_EXEC_WBINVD_EXITING | \
> + SECONDARY_EXEC_ENABLE_VPID | \
> + SECONDARY_EXEC_ENABLE_EPT | \
> + SECONDARY_EXEC_UNRESTRICTED_GUEST | \
> + SECONDARY_EXEC_DESC | \
> + SECONDARY_EXEC_ENABLE_RDTSCP | \
> + SECONDARY_EXEC_ENABLE_INVPCID | \
> + SECONDARY_EXEC_ENABLE_XSAVES | \
> + SECONDARY_EXEC_RDSEED_EXITING | \
> + SECONDARY_EXEC_RDRAND_EXITING | \
> + SECONDARY_EXEC_TSC_SCALING | \
> + SECONDARY_EXEC_ENABLE_USR_WAIT_PAUSE | \
> + SECONDARY_EXEC_PT_USE_GPA | \
> + SECONDARY_EXEC_PT_CONCEAL_VMX | \
> + SECONDARY_EXEC_BUS_LOCK_DETECTION | \
> + SECONDARY_EXEC_NOTIFY_VM_EXITING | \
> + SECONDARY_EXEC_ENCLS_EXITING)
> +
> +#define EVMCS1_SUPPORTED_3RDEXEC (0ULL)
> +
> +#define EVMCS1_SUPPORTED_VMEXIT_CTRL \
> + (VM_EXIT_ALWAYSON_WITHOUT_TRUE_MSR | \
> + VM_EXIT_SAVE_DEBUG_CONTROLS | \
> + VM_EXIT_ACK_INTR_ON_EXIT | \
> + VM_EXIT_HOST_ADDR_SPACE_SIZE | \
> + VM_EXIT_LOAD_IA32_PERF_GLOBAL_CTRL | \
> + VM_EXIT_SAVE_IA32_PAT | \
> + VM_EXIT_LOAD_IA32_PAT | \
> + VM_EXIT_SAVE_IA32_EFER | \
> + VM_EXIT_LOAD_IA32_EFER | \
> + VM_EXIT_CLEAR_BNDCFGS | \
> + VM_EXIT_PT_CONCEAL_PIP | \
> + VM_EXIT_CLEAR_IA32_RTIT_CTL)
> +
> +#define EVMCS1_SUPPORTED_VMENTRY_CTRL \
> + (VM_ENTRY_ALWAYSON_WITHOUT_TRUE_MSR | \
> + VM_ENTRY_LOAD_DEBUG_CONTROLS | \
> + VM_ENTRY_IA32E_MODE | \
> + VM_ENTRY_LOAD_IA32_PERF_GLOBAL_CTRL | \
> + VM_ENTRY_LOAD_IA32_PAT | \
> + VM_ENTRY_LOAD_IA32_EFER | \
> + VM_ENTRY_LOAD_BNDCFGS | \
> + VM_ENTRY_PT_CONCEAL_PIP | \
> + VM_ENTRY_LOAD_IA32_RTIT_CTL)
> +
> +#define EVMCS1_SUPPORTED_VMFUNC (0)
> +
> struct evmcs_field {
> u16 offset;
> u16 clean_field;
> @@ -65,114 +166,6 @@ static inline u64 evmcs_read_any(struct hv_enlightened_vmcs *evmcs,
> return vmcs12_read_any((void *)evmcs, field, offset);
> }
>
> -#if IS_ENABLED(CONFIG_HYPERV)
> -
> -DECLARE_STATIC_KEY_FALSE(__kvm_is_using_evmcs);
> -
> -static __always_inline bool kvm_is_using_evmcs(void)
> -{
> - return static_branch_unlikely(&__kvm_is_using_evmcs);
> -}
> -
> -static __always_inline int get_evmcs_offset(unsigned long field,
> - u16 *clean_field)
> -{
> - int offset = evmcs_field_offset(field, clean_field);
> -
> - WARN_ONCE(offset < 0, "accessing unsupported EVMCS field %lx\n", field);
> - return offset;
> -}
> -
> -static __always_inline void evmcs_write64(unsigned long field, u64 value)
> -{
> - u16 clean_field;
> - int offset = get_evmcs_offset(field, &clean_field);
> -
> - if (offset < 0)
> - return;
> -
> - *(u64 *)((char *)current_evmcs + offset) = value;
> -
> - current_evmcs->hv_clean_fields &= ~clean_field;
> -}
> -
> -static __always_inline void evmcs_write32(unsigned long field, u32 value)
> -{
> - u16 clean_field;
> - int offset = get_evmcs_offset(field, &clean_field);
> -
> - if (offset < 0)
> - return;
> -
> - *(u32 *)((char *)current_evmcs + offset) = value;
> - current_evmcs->hv_clean_fields &= ~clean_field;
> -}
> -
> -static __always_inline void evmcs_write16(unsigned long field, u16 value)
> -{
> - u16 clean_field;
> - int offset = get_evmcs_offset(field, &clean_field);
> -
> - if (offset < 0)
> - return;
> -
> - *(u16 *)((char *)current_evmcs + offset) = value;
> - current_evmcs->hv_clean_fields &= ~clean_field;
> -}
> -
> -static __always_inline u64 evmcs_read64(unsigned long field)
> -{
> - int offset = get_evmcs_offset(field, NULL);
> -
> - if (offset < 0)
> - return 0;
> -
> - return *(u64 *)((char *)current_evmcs + offset);
> -}
> -
> -static __always_inline u32 evmcs_read32(unsigned long field)
> -{
> - int offset = get_evmcs_offset(field, NULL);
> -
> - if (offset < 0)
> - return 0;
> -
> - return *(u32 *)((char *)current_evmcs + offset);
> -}
> -
> -static __always_inline u16 evmcs_read16(unsigned long field)
> -{
> - int offset = get_evmcs_offset(field, NULL);
> -
> - if (offset < 0)
> - return 0;
> -
> - return *(u16 *)((char *)current_evmcs + offset);
> -}
> -
> -static inline void evmcs_load(u64 phys_addr)
> -{
> - struct hv_vp_assist_page *vp_ap =
> - hv_get_vp_assist_page(smp_processor_id());
> -
> - if (current_evmcs->hv_enlightenments_control.nested_flush_hypercall)
> - vp_ap->nested_control.features.directhypercall = 1;
> - vp_ap->current_nested_vmcs = phys_addr;
> - vp_ap->enlighten_vmentry = 1;
> -}
> -
> -void evmcs_sanitize_exec_ctrls(struct vmcs_config *vmcs_conf);
> -#else /* !IS_ENABLED(CONFIG_HYPERV) */
> -static __always_inline bool kvm_is_using_evmcs(void) { return false; }
> -static __always_inline void evmcs_write64(unsigned long field, u64 value) {}
> -static __always_inline void evmcs_write32(unsigned long field, u32 value) {}
> -static __always_inline void evmcs_write16(unsigned long field, u16 value) {}
> -static __always_inline u64 evmcs_read64(unsigned long field) { return 0; }
> -static __always_inline u32 evmcs_read32(unsigned long field) { return 0; }
> -static __always_inline u16 evmcs_read16(unsigned long field) { return 0; }
> -static inline void evmcs_load(u64 phys_addr) {}
> -#endif /* IS_ENABLED(CONFIG_HYPERV) */
> -
> #define EVMPTR_INVALID (-1ULL)
> #define EVMPTR_MAP_PENDING (-2ULL)
>
> diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c
> index b7dc7acf14be..04eb5d4d28bc 100644
> --- a/arch/x86/kvm/vmx/vmx.c
> +++ b/arch/x86/kvm/vmx/vmx.c
> @@ -66,6 +66,7 @@
> #include "vmx.h"
> #include "x86.h"
> #include "smm.h"
> +#include "vmx_onhyperv.h"
>
> MODULE_AUTHOR("Qumranet");
> MODULE_LICENSE("GPL");
> diff --git a/arch/x86/kvm/vmx/vmx_onhyperv.c b/arch/x86/kvm/vmx/vmx_onhyperv.c
> new file mode 100644
> index 000000000000..b9a8b91166d0
> --- /dev/null
> +++ b/arch/x86/kvm/vmx/vmx_onhyperv.c
> @@ -0,0 +1,36 @@
> +// SPDX-License-Identifier: GPL-2.0-only
> +
> +#include "capabilities.h"
> +#include "vmx_onhyperv.h"
> +
> +DEFINE_STATIC_KEY_FALSE(__kvm_is_using_evmcs);
> +
> +/*
> + * KVM on Hyper-V always uses the latest known eVMCSv1 revision, the assumption
> + * is: in case a feature has corresponding fields in eVMCS described and it was
> + * exposed in VMX feature MSRs, KVM is free to use it. Warn if KVM meets a
> + * feature which has no corresponding eVMCS field, this likely means that KVM
> + * needs to be updated.
> + */
> +#define evmcs_check_vmcs_conf(field, ctrl) \
> + do { \
> + typeof(vmcs_conf->field) unsupported; \
> + \
> + unsupported = vmcs_conf->field & ~EVMCS1_SUPPORTED_ ## ctrl; \
> + if (unsupported) { \
> + pr_warn_once(#field " unsupported with eVMCS: 0x%llx\n",\
> + (u64)unsupported); \
> + vmcs_conf->field &= EVMCS1_SUPPORTED_ ## ctrl; \
> + } \
> + } \
> + while (0)
> +
> +void evmcs_sanitize_exec_ctrls(struct vmcs_config *vmcs_conf)
> +{
> + evmcs_check_vmcs_conf(cpu_based_exec_ctrl, EXEC_CTRL);
> + evmcs_check_vmcs_conf(pin_based_exec_ctrl, PINCTRL);
> + evmcs_check_vmcs_conf(cpu_based_2nd_exec_ctrl, 2NDEXEC);
> + evmcs_check_vmcs_conf(cpu_based_3rd_exec_ctrl, 3RDEXEC);
> + evmcs_check_vmcs_conf(vmentry_ctrl, VMENTRY_CTRL);
> + evmcs_check_vmcs_conf(vmexit_ctrl, VMEXIT_CTRL);
> +}
> diff --git a/arch/x86/kvm/vmx/vmx_onhyperv.h b/arch/x86/kvm/vmx/vmx_onhyperv.h
> new file mode 100644
> index 000000000000..11541d272dbd
> --- /dev/null
> +++ b/arch/x86/kvm/vmx/vmx_onhyperv.h
> @@ -0,0 +1,124 @@
> +/* SPDX-License-Identifier: GPL-2.0-only */
> +
> +#ifndef __ARCH_X86_KVM_VMX_ONHYPERV_H__
> +#define __ARCH_X86_KVM_VMX_ONHYPERV_H__
> +
> +#include <asm/hyperv-tlfs.h>
> +
> +#include <linux/jump_label.h>
> +
> +#include "capabilities.h"
> +#include "hyperv.h"
> +#include "vmcs12.h"
> +
> +#define current_evmcs ((struct hv_enlightened_vmcs *)this_cpu_read(current_vmcs))
> +
> +#if IS_ENABLED(CONFIG_HYPERV)
> +
> +DECLARE_STATIC_KEY_FALSE(__kvm_is_using_evmcs);
> +
> +static __always_inline bool kvm_is_using_evmcs(void)
> +{
> + return static_branch_unlikely(&__kvm_is_using_evmcs);
> +}
> +
> +static __always_inline int get_evmcs_offset(unsigned long field,
> + u16 *clean_field)
> +{
> + int offset = evmcs_field_offset(field, clean_field);
> +
> + WARN_ONCE(offset < 0, "accessing unsupported EVMCS field %lx\n", field);
> + return offset;
> +}
> +
> +static __always_inline void evmcs_write64(unsigned long field, u64 value)
> +{
> + u16 clean_field;
> + int offset = get_evmcs_offset(field, &clean_field);
> +
> + if (offset < 0)
> + return;
> +
> + *(u64 *)((char *)current_evmcs + offset) = value;
> +
> + current_evmcs->hv_clean_fields &= ~clean_field;
> +}
> +
> +static __always_inline void evmcs_write32(unsigned long field, u32 value)
> +{
> + u16 clean_field;
> + int offset = get_evmcs_offset(field, &clean_field);
> +
> + if (offset < 0)
> + return;
> +
> + *(u32 *)((char *)current_evmcs + offset) = value;
> + current_evmcs->hv_clean_fields &= ~clean_field;
> +}
> +
> +static __always_inline void evmcs_write16(unsigned long field, u16 value)
> +{
> + u16 clean_field;
> + int offset = get_evmcs_offset(field, &clean_field);
> +
> + if (offset < 0)
> + return;
> +
> + *(u16 *)((char *)current_evmcs + offset) = value;
> + current_evmcs->hv_clean_fields &= ~clean_field;
> +}
> +
> +static __always_inline u64 evmcs_read64(unsigned long field)
> +{
> + int offset = get_evmcs_offset(field, NULL);
> +
> + if (offset < 0)
> + return 0;
> +
> + return *(u64 *)((char *)current_evmcs + offset);
> +}
> +
> +static __always_inline u32 evmcs_read32(unsigned long field)
> +{
> + int offset = get_evmcs_offset(field, NULL);
> +
> + if (offset < 0)
> + return 0;
> +
> + return *(u32 *)((char *)current_evmcs + offset);
> +}
> +
> +static __always_inline u16 evmcs_read16(unsigned long field)
> +{
> + int offset = get_evmcs_offset(field, NULL);
> +
> + if (offset < 0)
> + return 0;
> +
> + return *(u16 *)((char *)current_evmcs + offset);
> +}
> +
> +static inline void evmcs_load(u64 phys_addr)
> +{
> + struct hv_vp_assist_page *vp_ap =
> + hv_get_vp_assist_page(smp_processor_id());
> +
> + if (current_evmcs->hv_enlightenments_control.nested_flush_hypercall)
> + vp_ap->nested_control.features.directhypercall = 1;
> + vp_ap->current_nested_vmcs = phys_addr;
> + vp_ap->enlighten_vmentry = 1;
> +}
> +
> +void evmcs_sanitize_exec_ctrls(struct vmcs_config *vmcs_conf);
> +#else /* !IS_ENABLED(CONFIG_HYPERV) */
> +static __always_inline bool kvm_is_using_evmcs(void) { return false; }
> +static __always_inline void evmcs_write64(unsigned long field, u64 value) {}
> +static __always_inline void evmcs_write32(unsigned long field, u32 value) {}
> +static __always_inline void evmcs_write16(unsigned long field, u16 value) {}
> +static __always_inline u64 evmcs_read64(unsigned long field) { return 0; }
> +static __always_inline u32 evmcs_read32(unsigned long field) { return 0; }
> +static __always_inline u16 evmcs_read16(unsigned long field) { return 0; }
> +static inline void evmcs_load(u64 phys_addr) {}
> +#endif /* IS_ENABLED(CONFIG_HYPERV) */
> +
> +#endif /* __ARCH_X86_KVM_VMX_ONHYPERV_H__ */
> diff --git a/arch/x86/kvm/vmx/vmx_ops.h b/arch/x86/kvm/vmx/vmx_ops.h
> index 33af7b4c6eb4..f41ce3c24123 100644
> --- a/arch/x86/kvm/vmx/vmx_ops.h
> +++ b/arch/x86/kvm/vmx/vmx_ops.h
> @@ -6,7 +6,7 @@
>
> #include <asm/vmx.h>
>
> -#include "hyperv.h"
> +#include "vmx_onhyperv.h"
> #include "vmcs.h"
> #include "../x86.h"
>

I did an overall sanity check, including 'diff'ing the moved code,
and it looks good, but I might have missed something.

Reviewed-by: Maxim Levitsky <mlevitsk@xxxxxxxxxx>

Best regards,
Maxim Levitsky