Re: [RFC PATCH v2] x86/fpu: make kernel-mode FPU reliably usable in softirqs
From: Ingo Molnar
Date: Wed Mar 05 2025 - 04:08:01 EST
* Eric Biggers <ebiggers@xxxxxxxxxx> wrote:
> From: Eric Biggers <ebiggers@xxxxxxxxxx>
>
> Currently kernel-mode FPU is not always usable in softirq context on
> x86, since softirqs can nest inside a kernel-mode FPU section in task
> context, and nested use of kernel-mode FPU is not supported.
>
> Therefore, x86 SIMD-optimized code that can be called in softirq context
> has to sometimes fall back to non-SIMD code. There are two options for
> the fallback, both of which are pretty terrible:
>
> (a) Use a scalar fallback. This can be 10-100x slower than vectorized
> code because it cannot use specialized instructions like AES, SHA,
> or carryless multiplication.
>
> (b) Execute the request asynchronously using a kworker. In other
> words, use the "crypto SIMD helper" in crypto/simd.c.
>
> Currently most of the x86 en/decryption code (skcipher and aead
> algorithms) uses option (b), since this avoids the slow scalar fallback
> and it is easier to wire up. But option (b) is still really bad for its
> own reasons:
>
> - Punting the request to a kworker is bad for performance too.
>
> - It forces the algorithm to be marked as asynchronous
> (CRYPTO_ALG_ASYNC), preventing it from being used by crypto API
> users who request a synchronous algorithm. That's another huge
> performance problem, which is especially unfortunate for users who
> don't even do en/decryption in softirq context.
>
> - It makes all en/decryption operations take a detour through
> crypto/simd.c. That involves additional checks and an additional
> indirect call, which slow down en/decryption for *everyone*.
>
> Fortunately, the skcipher and aead APIs are only usable in task and
> softirq context in the first place. Thus, if kernel-mode FPU were to
> be reliably usable in softirq context, no fallback would be needed.
> Indeed, other architectures such as arm, arm64, and riscv have
> already done this.
>
> Therefore, this patch updates x86 accordingly to reliably support
> kernel-mode FPU in softirqs.
>
> This is done by just disabling softirq processing in kernel-mode FPU
> sections (when hardirqs are not already disabled), as that prevents the
> nesting that was problematic.
>
> This will delay some softirqs slightly, but only ones that would have
> otherwise been nested inside a task context kernel-mode FPU section.
> Any such softirqs would have taken the slow fallback path before if they
> tried to do any en/decryption. Now these softirqs will just run at the
> end of the task context kernel-mode FPU section (since local_bh_enable()
> runs pending softirqs) and will no longer take the slow fallback path.
>
> Alternatives considered:
>
> - Make kernel-mode FPU sections fully preemptible. This would require
> growing task_struct by another struct fpstate which is more than 2K.
So that's something that will probably happen once the kernel is built
using APX anyway?
> - Make softirqs save/restore the kernel-mode FPU state to a per-CPU
> struct fpstate when nested use is detected. Somewhat interesting, but
> seems unnecessary when a simpler solution exists.
So:
> void kernel_fpu_begin_mask(unsigned int kfpu_mask)
> {
> - preempt_disable();
> + if (!irqs_disabled())
> + fpregs_lock();
> + if (!irqs_disabled())
> + fpregs_unlock();
So why is the irqs_disabled() check needed here? (On x86 it can be a
bit expensive at times, because the IRQ flag has to be loaded,
including all flags, so basically it's a soft synchronization point of
a sort.)
Ie. why cannot we simply do a local_bh_disable()/enable() pair (on
!RT), ie. fpregs_lock()/fpregs_unlock()?
local_bh_disable() is very similar in cost to preempt_disable(), both
are increasing the preempt_count.
Thanks,
Ingo