[PATCH v6 12/18] KVM: arm64: Support SDEI event injection, delivery
From: Gavin Shan
Date: Sun Apr 03 2022 - 11:41:25 EST
This supports event injection, delivery and cancellation. The event
is injected and cancelled by kvm_sdei_{inject, cancel}_event(). For
event delivery, vcpu request (KVM_REQ_SDEI) is introduced.
kvm_sdei_deliver_event() is called to accommodate the KVM_REQ_SDEI
request. The execution context is switched like below:
* x0 - x17 are saved. All of them are cleared except the following
registers:
x0: event number
x1: user argument associated with the SDEI event
x2: PC of the interrupted or preempted context
x3: PSTATE of the interrupted or preempted context
* PC is set to the event handler, which is provided when the event
is registered. PSTATE is modified according to the specification.
Signed-off-by: Gavin Shan <gshan@xxxxxxxxxx>
---
arch/arm64/include/asm/kvm_host.h | 1 +
arch/arm64/include/asm/kvm_sdei.h | 4 +
arch/arm64/kvm/arm.c | 3 +
arch/arm64/kvm/sdei.c | 260 ++++++++++++++++++++++++++++++
4 files changed, 268 insertions(+)
diff --git a/arch/arm64/include/asm/kvm_host.h b/arch/arm64/include/asm/kvm_host.h
index 7644a400c4a8..951264d4b64d 100644
--- a/arch/arm64/include/asm/kvm_host.h
+++ b/arch/arm64/include/asm/kvm_host.h
@@ -46,6 +46,7 @@
#define KVM_REQ_RECORD_STEAL KVM_ARCH_REQ(3)
#define KVM_REQ_RELOAD_GICv4 KVM_ARCH_REQ(4)
#define KVM_REQ_RELOAD_PMU KVM_ARCH_REQ(5)
+#define KVM_REQ_SDEI KVM_ARCH_REQ(6)
#define KVM_DIRTY_LOG_MANUAL_CAPS (KVM_DIRTY_LOG_MANUAL_PROTECT_ENABLE | \
KVM_DIRTY_LOG_INITIALLY_SET)
diff --git a/arch/arm64/include/asm/kvm_sdei.h b/arch/arm64/include/asm/kvm_sdei.h
index 2dbfb3ae0a48..f946d4ebdc14 100644
--- a/arch/arm64/include/asm/kvm_sdei.h
+++ b/arch/arm64/include/asm/kvm_sdei.h
@@ -142,6 +142,10 @@ KVM_SDEI_REGISTERED_EVENT_FUNC(unregister_pending, UNREGISTER_PENDING)
/* APIs */
int kvm_sdei_call(struct kvm_vcpu *vcpu);
+int kvm_sdei_inject_event(struct kvm_vcpu *vcpu, unsigned int num,
+ bool immediate);
+int kvm_sdei_cancel_event(struct kvm_vcpu *vcpu, unsigned int num);
+void kvm_sdei_deliver_event(struct kvm_vcpu *vcpu);
void kvm_sdei_create_vcpu(struct kvm_vcpu *vcpu);
void kvm_sdei_destroy_vcpu(struct kvm_vcpu *vcpu);
diff --git a/arch/arm64/kvm/arm.c b/arch/arm64/kvm/arm.c
index 227c0e390571..7e77d62aeab1 100644
--- a/arch/arm64/kvm/arm.c
+++ b/arch/arm64/kvm/arm.c
@@ -659,6 +659,9 @@ static void check_vcpu_requests(struct kvm_vcpu *vcpu)
if (kvm_check_request(KVM_REQ_VCPU_RESET, vcpu))
kvm_reset_vcpu(vcpu);
+ if (kvm_check_request(KVM_REQ_SDEI, vcpu))
+ kvm_sdei_deliver_event(vcpu);
+
/*
* Clear IRQ_PENDING requests that were made to guarantee
* that a VCPU sees new virtual interrupts.
diff --git a/arch/arm64/kvm/sdei.c b/arch/arm64/kvm/sdei.c
index 6716ed020aa2..9d18fee59751 100644
--- a/arch/arm64/kvm/sdei.c
+++ b/arch/arm64/kvm/sdei.c
@@ -87,6 +87,36 @@ static int reset_event(struct kvm_vcpu *vcpu,
return 0;
}
+static int inject_event(struct kvm_vcpu *vcpu,
+ struct kvm_sdei_event *event)
+{
+ struct kvm_sdei_vcpu *vsdei = vcpu->arch.sdei;
+ struct kvm_sdei_exposed_event *exposed_event;
+
+ /* The event should have been registered and enabled */
+ if (!kvm_sdei_is_registered(event) ||
+ !kvm_sdei_is_enabled(event))
+ return -ENOENT;
+
+ /*
+ * If the event is pending for unregistration, we shouldn't
+ * inject the event.
+ */
+ if (kvm_sdei_is_unregister_pending(event))
+ return -EAGAIN;
+
+ event->event_count++;
+ exposed_event = event->exposed_event;
+ if (kvm_sdei_is_critical(exposed_event->priority))
+ vsdei->critical_event_count++;
+ else
+ vsdei->normal_event_count++;
+
+ kvm_make_request(KVM_REQ_SDEI, vcpu);
+
+ return 0;
+}
+
static unsigned long hypercall_register(struct kvm_vcpu *vcpu)
{
struct kvm_sdei_vcpu *vsdei = vcpu->arch.sdei;
@@ -447,6 +477,236 @@ int kvm_sdei_call(struct kvm_vcpu *vcpu)
return 1;
}
+int kvm_sdei_inject_event(struct kvm_vcpu *vcpu,
+ unsigned int num,
+ bool immediate)
+{
+ struct kvm_sdei_vcpu *vsdei = vcpu->arch.sdei;
+ struct kvm_sdei_exposed_event *exposed_event;
+ struct kvm_sdei_event *event;
+ unsigned long event_count = 0;
+ int ret = 0;
+
+ if (!vsdei) {
+ ret = -EPERM;
+ goto out;
+ }
+
+ if (!kvm_sdei_is_supported(num)) {
+ ret = -EINVAL;
+ goto out;
+ }
+
+ spin_lock(&vsdei->lock);
+
+ /* Check if the vcpu has been masked */
+ if (vsdei->masked) {
+ ret = -EPERM;
+ goto unlock;
+ }
+
+ /* Check if the event exists */
+ event = find_event(vcpu, num);
+ if (!event) {
+ ret = -ENOENT;
+ goto unlock;
+ }
+
+ /*
+ * In some cases, the injected event is expected to be delivered
+ * immediately. However, there are two cases the injected event
+ * can't be delivered immediately: (a) the injected event is a
+ * critical one, but we already have pending critical events for
+ * delivery. (b) the injected event is a normal one, but we have
+ * pending events for delivery, regardless of their priorities.
+ */
+ if (immediate) {
+ exposed_event = event->exposed_event;
+ event_count = vsdei->critical_event_count;
+ if (kvm_sdei_is_normal(exposed_event->priority))
+ event_count += vsdei->normal_event_count;
+
+ if (event_count > 0)
+ return -ENOSPC;
+ }
+
+ ret = inject_event(vcpu, event);
+
+unlock:
+ spin_unlock(&vsdei->lock);
+out:
+ return ret;
+}
+
+int kvm_sdei_cancel_event(struct kvm_vcpu *vcpu, unsigned int num)
+{
+ struct kvm_sdei_vcpu *vsdei = vcpu->arch.sdei;
+ struct kvm_sdei_vcpu_context *context;
+ struct kvm_sdei_exposed_event *exposed_event;
+ struct kvm_sdei_event *event;
+ int ret = 0;
+
+ if (!vsdei) {
+ ret = -EPERM;
+ goto out;
+ }
+
+ if (!kvm_sdei_is_supported(num)) {
+ ret = -EINVAL;
+ goto out;
+ }
+
+ spin_lock(&vsdei->lock);
+
+ /* Check if the event exists */
+ event = find_event(vcpu, num);
+ if (!event) {
+ ret = -ENOENT;
+ goto unlock;
+ }
+
+ /* The event should have been registered and eanbled */
+ if (!kvm_sdei_is_registered(event) ||
+ !kvm_sdei_is_enabled(event)) {
+ ret = -EBUSY;
+ goto unlock;
+ }
+
+ /*
+ * If the event is pending for unregistration, we needn't
+ * cancel it because all the pending events will be cancelled
+ * after its handler is to finish.
+ */
+ if (kvm_sdei_is_unregister_pending(event)) {
+ ret = -EAGAIN;
+ goto unlock;
+ }
+
+ /* Return success if there is no pending events */
+ if (event->event_count <= 0)
+ goto unlock;
+
+ /* The event can't be cancelled if it's being handled */
+ exposed_event = event->exposed_event;
+ context = kvm_sdei_is_critical(exposed_event->priority) ?
+ &vsdei->context[SDEI_EVENT_PRIORITY_CRITICAL] :
+ &vsdei->context[SDEI_EVENT_PRIORITY_NORMAL];
+ if (event->event_count == 1 &&
+ context->event == event) {
+ ret = -EAGAIN;
+ goto unlock;
+ }
+
+ /* Cancel the event */
+ event->event_count--;
+ if (kvm_sdei_is_critical(exposed_event->priority))
+ vsdei->critical_event_count--;
+ else
+ vsdei->normal_event_count--;
+
+unlock:
+ spin_unlock(&vsdei->lock);
+out:
+ return ret;
+}
+
+void kvm_sdei_deliver_event(struct kvm_vcpu *vcpu)
+{
+ struct kvm_sdei_vcpu *vsdei = vcpu->arch.sdei;
+ struct kvm_sdei_vcpu_context *context;
+ struct kvm_sdei_exposed_event *exposed_event;
+ struct kvm_sdei_event *selected, *event;
+ unsigned long pstate;
+ unsigned int i;
+
+ if (!vsdei)
+ return;
+
+ spin_lock(&vsdei->lock);
+
+ /* Check if the vcpu has been masked */
+ if (vsdei->masked)
+ goto unlock;
+
+ /* The currently handled critical event can't be preempted */
+ context = &vsdei->context[SDEI_EVENT_PRIORITY_CRITICAL];
+ if (context->event)
+ goto unlock;
+
+ /*
+ * Select the event to be handled. The critical event is
+ * selected if we have one. Otherwise, the first normal
+ * event will be selected. Beside, the normal event can
+ * be preempted by the critical event. However, the normal
+ * event can't be preempted by another normal event.
+ */
+ selected = NULL;
+ context = &vsdei->context[SDEI_EVENT_PRIORITY_NORMAL];
+ kvm_sdei_for_each_event(vsdei, event, i) {
+ exposed_event = event->exposed_event;
+ if (event->event_count <= 0)
+ continue;
+
+ if (context->event &&
+ kvm_sdei_is_normal(exposed_event->priority))
+ continue;
+
+ if (kvm_sdei_is_critical(exposed_event->priority)) {
+ selected = event;
+ break;
+ }
+
+ selected = selected ? : event;
+ }
+
+ if (!selected)
+ goto unlock;
+
+ /*
+ * Save context: x0 -> x17, PC, PState. There might be pending
+ * exception or PC increment request in the last run on this vCPU.
+ * In this case, we need to save the site in advance. Otherwise,
+ * the passed entry point could be floated by 4 bytes in the
+ * subsequent call to __kvm_adjust_pc().
+ */
+ event = selected;
+ exposed_event = event->exposed_event;
+ context = kvm_sdei_is_critical(exposed_event->priority) ?
+ &vsdei->context[SDEI_EVENT_PRIORITY_CRITICAL] :
+ &vsdei->context[SDEI_EVENT_PRIORITY_NORMAL];
+ context->event = event;
+
+ __kvm_adjust_pc(vcpu);
+ context->pc = *vcpu_pc(vcpu);
+ context->pstate = *vcpu_cpsr(vcpu);
+ for (i = 0; i < ARRAY_SIZE(context->regs); i++)
+ context->regs[i] = vcpu_get_reg(vcpu, i);
+
+ /*
+ * Inject event: x0 -> x3, PC, PState. The other general
+ * purpose registers are cleared.
+ */
+ for (i = 0; i < ARRAY_SIZE(context->regs); i++)
+ vcpu_set_reg(vcpu, i, 0);
+
+ vcpu_set_reg(vcpu, 0, exposed_event->num);
+ vcpu_set_reg(vcpu, 1, event->ep_arg);
+ vcpu_set_reg(vcpu, 2, context->pc);
+ vcpu_set_reg(vcpu, 3, context->pstate);
+
+ pstate = context->pstate;
+ pstate |= (PSR_D_BIT | PSR_A_BIT | PSR_I_BIT | PSR_F_BIT);
+ pstate &= ~PSR_MODE_MASK;
+ pstate |= PSR_MODE_EL1h;
+ pstate &= ~PSR_MODE32_BIT;
+
+ *vcpu_cpsr(vcpu) = pstate;
+ *vcpu_pc(vcpu) = event->ep_address;
+
+unlock:
+ spin_unlock(&vsdei->lock);
+}
+
void kvm_sdei_create_vcpu(struct kvm_vcpu *vcpu)
{
struct kvm_sdei_vcpu *vsdei;
--
2.23.0