[RFC patch 14/15] workpending: Provide infrastructure for work before entering a guest

From: Thomas Gleixner
Date: Thu Sep 19 2019 - 11:10:07 EST


Entering a guest is similar to exiting to user space. Pending work like
handling signals, rescheduling, task work etc. needs to be handled before
that.

Provide generic infrastructure to avoid duplication of the same handling code
all over the place.

Update ARM64 struct kvm_vcpu_stat with a signal_exit member so the generic
code compiles.

Signed-off-by: Thomas Gleixner <tglx@xxxxxxxxxxxxx>
---
arch/arm64/include/asm/kvm_host.h | 1
include/linux/entry-common.h | 66 ++++++++++++++++++++++++++++++++++++++
kernel/entry/common.c | 44 +++++++++++++++++++++++++
3 files changed, 111 insertions(+)

--- a/arch/arm64/include/asm/kvm_host.h
+++ b/arch/arm64/include/asm/kvm_host.h
@@ -409,6 +409,7 @@ struct kvm_vcpu_stat {
u64 wfi_exit_stat;
u64 mmio_exit_user;
u64 mmio_exit_kernel;
+ u64 signal_exits;
u64 exits;
};

--- a/include/linux/entry-common.h
+++ b/include/linux/entry-common.h
@@ -255,4 +255,70 @@ static inline void arch_syscall_exit_tra
/* Common syscall exit function */
void syscall_exit_to_usermode(struct pt_regs *regs, long syscall, long retval);

+#if IS_ENABLED(CONFIG_KVM)
+
+#include <linux/kvm_host.h>
+
+#ifndef ARCH_EXIT_TO_GUESTMODE_WORK
+# define ARCH_EXIT_TO_GUESTMODE_WORK (0)
+#endif
+
+#define EXIT_TO_GUESTMODE_WORK \
+ (_TIF_NEED_RESCHED | _TIF_SIGPENDING | _TIF_NOTIFY_RESUME | \
+ ARCH_EXIT_TO_GUESTMODE_WORK)
+
+int core_exit_to_guestmode_work(struct kvm *kvm, struct kvm_vcpu *vcpu,
+ unsigned long ti_work);
+
+/**
+ * arch_exit_to_guestmode - Architecture specific exit to guest mode function
+ * @kvm: Pointer to the guest instance
+ * @vcpu: Pointer to current's VCPU data
+ * @ti_work: Cached TIF flags gathered in exit_to_guestmode()
+ *
+ * Invoked from core_exit_to_guestmode_work(). Can be replaced by
+ * architecture specific code.
+ */
+static inline int arch_exit_to_guestmode(struct kvm *kvm, struct kvm_vcpu *vcpu,
+ unsigned long ti_work);
+
+#ifndef arch_exit_to_guestmode
+static inline int arch_exit_to_guestmode(struct kvm *kvm, struct kvm_vcpu *vcpu,
+ unsigned long ti_work)
+{
+ return 0;
+}
+#endif
+
+/**
+ * exit_to_guestmode - Check and handle pending work which needs to be
+ * handled before returning to guest mode
+ * @kvm: Pointer to the guest instance
+ * @vcpu: Pointer to current's VCPU data
+ *
+ * Returns: 0 or an error code
+ */
+static inline int exit_to_guestmode(struct kvm *kvm, struct kvm_vcpu *vcpu)
+{
+ unsigned long ti_work = READ_ONCE(current_thread_info()->flags);
+
+ if (unlikely(ti_work & EXIT_TO_GUESTMODE_WORK))
+ return core_exit_to_guestmode_work(kvm, vcpu, ti_work);
+ return 0;
+}
+
+
+/**
+ * _exit_to_guestmode_work_pending - Check if work is pending which needs to be
+ * handled before returning to guest mode
+ */
+static inline bool exit_to_guestmode_work_pending(void)
+{
+ unsigned long ti_work = READ_ONCE(current_thread_info()->flags);
+
+ return !!(ti_work & EXIT_TO_GUESTMODE_WORK);
+
+}
+#endif /* CONFIG_KVM */
+
#endif
--- a/kernel/entry/common.c
+++ b/kernel/entry/common.c
@@ -174,3 +174,47 @@ void syscall_exit_to_usermode(struct pt_
do_exit_to_usermode(regs);
#endif
}
+
+#if IS_ENABLED(CONFIG_KVM)
+int __weak arch_exit_to_guestmode_work(struct kvm *kvm, struct kvm_vcpu *vcpu,
+ unsigned long ti_work)
+{
+ return 0;
+}
+
+int core_exit_to_guestmode_work(struct kvm *kvm, struct kvm_vcpu *vcpu,
+ unsigned long ti_work)
+{
+ /*
+ * Before returning to guest mode handle all pending work
+ */
+ if (ti_work & _TIF_SIGPENDING) {
+ vcpu->run->exit_reason = KVM_EXIT_INTR;
+ vcpu->stat.signal_exits++;
+ return -EINTR;
+ }
+
+ if (ti_work & _TIF_NEED_RESCHED) {
+ srcu_read_unlock(&kvm->srcu, vcpu->srcu_idx);
+ schedule();
+ vcpu->srcu_idx = srcu_read_lock(&kvm->srcu);
+ }
+
+ if (ti_work & _TIF_PATCH_PENDING) {
+ srcu_read_unlock(&kvm->srcu, vcpu->srcu_idx);
+ klp_update_patch_state(current);
+ vcpu->srcu_idx = srcu_read_lock(&kvm->srcu);
+ }
+
+ if (ti_work & _TIF_NOTIFY_RESUME) {
+ srcu_read_unlock(&kvm->srcu, vcpu->srcu_idx);
+ clear_thread_flag(TIF_NOTIFY_RESUME);
+ tracehook_notify_resume(NULL);
+ vcpu->srcu_idx = srcu_read_lock(&kvm->srcu);
+ }
+
+ /* Any extra architecture specific work */
+ return arch_exit_to_guestmode_work(kvm, vcpu, ti_work);
+}
+EXPORT_SYMBOL_GPL(core_exit_to_guestmode_work);
+#endif