Re: [PATCH v3 01/11] riscv: Implement cmpxchg32/64() using Zacas

From: Alexandre Ghiti
Date: Fri Jul 19 2024 - 07:53:59 EST


On Fri, Jul 19, 2024 at 1:48 PM Alexandre Ghiti <alexghiti@xxxxxxxxxxxx> wrote:
>
> On Fri, Jul 19, 2024 at 2:45 AM Samuel Holland
> <samuel.holland@xxxxxxxxxx> wrote:
> >
> > Hi Alex,
> >
> > On 2024-07-17 1:19 AM, Alexandre Ghiti wrote:
> > > This adds runtime support for Zacas in cmpxchg operations.
> > >
> > > Signed-off-by: Alexandre Ghiti <alexghiti@xxxxxxxxxxxx>
> > > ---
> > > arch/riscv/Kconfig | 17 +++++++++++++++++
> > > arch/riscv/Makefile | 3 +++
> > > arch/riscv/include/asm/cmpxchg.h | 26 +++++++++++++++++++++++---
> > > 3 files changed, 43 insertions(+), 3 deletions(-)
> > >
> > > diff --git a/arch/riscv/Kconfig b/arch/riscv/Kconfig
> > > index 05ccba8ca33a..1caaedec88c7 100644
> > > --- a/arch/riscv/Kconfig
> > > +++ b/arch/riscv/Kconfig
> > > @@ -596,6 +596,23 @@ config RISCV_ISA_V_PREEMPTIVE
> > > preemption. Enabling this config will result in higher memory
> > > consumption due to the allocation of per-task's kernel Vector context.
> > >
> > > +config TOOLCHAIN_HAS_ZACAS
> > > + bool
> > > + default y
> > > + depends on !64BIT || $(cc-option,-mabi=lp64 -march=rv64ima_zacas)
> > > + depends on !32BIT || $(cc-option,-mabi=ilp32 -march=rv32ima_zacas)
> > > + depends on AS_HAS_OPTION_ARCH
> > > +
> > > +config RISCV_ISA_ZACAS
> > > + bool "Zacas extension support for atomic CAS"
> > > + depends on TOOLCHAIN_HAS_ZACAS
> > > + default y
> > > + help
> > > + Enable the use of the Zacas ISA-extension to implement kernel atomic
> > > + cmpxchg operations when it is detected at boot.
> > > +
> > > + If you don't know what to do here, say Y.
> > > +
> > > config TOOLCHAIN_HAS_ZBB
> > > bool
> > > default y
> > > diff --git a/arch/riscv/Makefile b/arch/riscv/Makefile
> > > index 06de9d365088..9fd13d7a9cc6 100644
> > > --- a/arch/riscv/Makefile
> > > +++ b/arch/riscv/Makefile
> > > @@ -85,6 +85,9 @@ endif
> > > # Check if the toolchain supports Zihintpause extension
> > > riscv-march-$(CONFIG_TOOLCHAIN_HAS_ZIHINTPAUSE) := $(riscv-march-y)_zihintpause
> > >
> > > +# Check if the toolchain supports Zacas
> > > +riscv-march-$(CONFIG_TOOLCHAIN_HAS_ZACAS) := $(riscv-march-y)_zacas
> > > +
> > > # Remove F,D,V from isa string for all. Keep extensions between "fd" and "v" by
> > > # matching non-v and non-multi-letter extensions out with the filter ([^v_]*)
> > > KBUILD_CFLAGS += -march=$(shell echo $(riscv-march-y) | sed -E 's/(rv32ima|rv64ima)fd([^v_]*)v?/\1\2/')
> > > diff --git a/arch/riscv/include/asm/cmpxchg.h b/arch/riscv/include/asm/cmpxchg.h
> > > index 808b4c78462e..5d38153e2f13 100644
> > > --- a/arch/riscv/include/asm/cmpxchg.h
> > > +++ b/arch/riscv/include/asm/cmpxchg.h
> > > @@ -9,6 +9,7 @@
> > > #include <linux/bug.h>
> > >
> > > #include <asm/fence.h>
> > > +#include <asm/alternative.h>
> > >
> > > #define __arch_xchg_masked(sc_sfx, prepend, append, r, p, n) \
> > > ({ \
> > > @@ -134,21 +135,40 @@
> > > r = (__typeof__(*(p)))((__retx & __mask) >> __s); \
> > > })
> > >
> > > -#define __arch_cmpxchg(lr_sfx, sc_sfx, prepend, append, r, p, co, o, n) \
> > > +#define __arch_cmpxchg(lr_sfx, sc_cas_sfx, prepend, append, r, p, co, o, n) \
> > > ({ \
> > > + __label__ no_zacas, end; \
> > > register unsigned int __rc; \
> > > \
> > > + if (IS_ENABLED(CONFIG_RISCV_ISA_ZACAS)) { \
> > > + asm goto(ALTERNATIVE("j %[no_zacas]", "nop", 0, \
> > > + RISCV_ISA_EXT_ZACAS, 1) \
> > > + : : : : no_zacas); \
> > > + \
> > > + __asm__ __volatile__ ( \
> > > + prepend \
> > > + " amocas" sc_cas_sfx " %0, %z2, %1\n" \
> > > + append \
> > > + : "+&r" (r), "+A" (*(p)) \
> > > + : "rJ" (n) \
> > > + : "memory"); \
> > > + goto end; \
> > > + } \
> > > + \
> > > +no_zacas: \
> > > __asm__ __volatile__ ( \
> > > prepend \
> > > "0: lr" lr_sfx " %0, %2\n" \
> > > " bne %0, %z3, 1f\n" \
> > > - " sc" sc_sfx " %1, %z4, %2\n" \
> > > + " sc" sc_cas_sfx " %1, %z4, %2\n" \
> > > " bnez %1, 0b\n" \
> > > append \
> >
> > This would probably be a good place to use inline ALTERNATIVE instead of an asm
> > goto. It saves overall code size, and a jump in the non-Zacas case, at the cost
> > of 3 nops in the Zacas case. (And all the nops can go after the amocas, where
> > they will likely be hidden by the amocas latency.)
>
> That's what Conor proposed indeed.
>
> I have just given it a try, but it does not work. The number of
> instructions in the zacas asm inline is different in the case of the
> fully-ordered version so I cannot set a unique number of nops. I could
> pass this information from arch_cmpxchg() down to __arch_cmpxchg() but
> those macros are already complex enough so I'd rather not add another
> parameter.
>
> This suggestion unfortunately cannot be applied to
> __arch_cmpxchg_masked(), nor __arch_xchg_masked().
>
> So unless you and Conor really insist, I'll drop the idea!

Or I can pass a nop when the full barrier is not needed, and it works!
I'll probably keep this version then since it avoids the introduction
of new macros or the use of a static branch to circumvent the llvm
bug.

>
> Thanks,
>
> Alex
>
>
> >
> > Regards,
> > Samuel
> >
> > > "1:\n" \
> > > : "=&r" (r), "=&r" (__rc), "+A" (*(p)) \
> > > : "rJ" (co o), "rJ" (n) \
> > > : "memory"); \
> > > + \
> > > +end:; \
> > > })
> > >
> > > #define _arch_cmpxchg(ptr, old, new, sc_sfx, prepend, append) \
> > > @@ -156,7 +176,7 @@
> > > __typeof__(ptr) __ptr = (ptr); \
> > > __typeof__(*(__ptr)) __old = (old); \
> > > __typeof__(*(__ptr)) __new = (new); \
> > > - __typeof__(*(__ptr)) __ret; \
> > > + __typeof__(*(__ptr)) __ret = (old); \
> > > \
> > > switch (sizeof(*__ptr)) { \
> > > case 1: \
> >