Re: [PATCH] riscv: locks: introduce ticket-based spinlock implementation

From: Guo Ren
Date: Wed Mar 24 2021 - 08:25:52 EST


On Wed, Mar 24, 2021 at 7:16 PM Vitaly Wool <vitaly.wool@xxxxxxxxxxxx> wrote:
>
>
>
> On Wed, Mar 24, 2021, 11:16 AM <guoren@xxxxxxxxxx> wrote:
>>
>> From: Guo Ren <guoren@xxxxxxxxxxxxxxxxx>
>>
>> This patch introduces a ticket lock implementation for riscv, along the
>> same lines as the implementation for arch/arm & arch/csky.
>
>
> Could you please provide a rationale for this? Like, what is wrong with the current implementation.
Ticket based spinlock's principle is here:
https://lwn.net/Articles/267968/

Current implementation will cause cache line bouncing when many harts
are acquiring the same spinlock.
I'm seeking a solution, maybe not fitting the current RISC-V base ISA.

I'll add more comments in the next version of patch.

>
> Thanks in advance,
>
> Best regards,
> Vitaly
>>
>>
>> Signed-off-by: Guo Ren <guoren@xxxxxxxxxxxxxxxxx>
>> Cc: Catalin Marinas <catalin.marinas@xxxxxxx>
>> Cc: Will Deacon <will.deacon@xxxxxxx>
>> Cc: Peter Zijlstra <peterz@xxxxxxxxxxxxx>
>> Cc: Palmer Dabbelt <palmerdabbelt@xxxxxxxxxx>
>> Cc: Anup Patel <anup@xxxxxxxxxxxxxx>
>> Cc: Arnd Bergmann <arnd@xxxxxxxx>
>> ---
>> arch/riscv/Kconfig | 1 +
>> arch/riscv/include/asm/Kbuild | 1 +
>> arch/riscv/include/asm/spinlock.h | 158 ++++++++++++--------------------
>> arch/riscv/include/asm/spinlock_types.h | 19 ++--
>> 4 files changed, 74 insertions(+), 105 deletions(-)
>>
>> diff --git a/arch/riscv/Kconfig b/arch/riscv/Kconfig
>> index 87d7b52..7c56a20 100644
>> --- a/arch/riscv/Kconfig
>> +++ b/arch/riscv/Kconfig
>> @@ -30,6 +30,7 @@ config RISCV
>> select ARCH_HAS_STRICT_KERNEL_RWX if MMU
>> select ARCH_OPTIONAL_KERNEL_RWX if ARCH_HAS_STRICT_KERNEL_RWX
>> select ARCH_OPTIONAL_KERNEL_RWX_DEFAULT
>> + select ARCH_USE_QUEUED_RWLOCKS
>> select ARCH_WANT_DEFAULT_TOPDOWN_MMAP_LAYOUT if MMU
>> select ARCH_WANT_FRAME_POINTERS
>> select ARCH_WANT_HUGE_PMD_SHARE if 64BIT
>> diff --git a/arch/riscv/include/asm/Kbuild b/arch/riscv/include/asm/Kbuild
>> index 445ccc9..e57ef80 100644
>> --- a/arch/riscv/include/asm/Kbuild
>> +++ b/arch/riscv/include/asm/Kbuild
>> @@ -3,5 +3,6 @@ generic-y += early_ioremap.h
>> generic-y += extable.h
>> generic-y += flat.h
>> generic-y += kvm_para.h
>> +generic-y += qrwlock.h
>> generic-y += user.h
>> generic-y += vmlinux.lds.h
>> diff --git a/arch/riscv/include/asm/spinlock.h b/arch/riscv/include/asm/spinlock.h
>> index f4f7fa1..2c81764 100644
>> --- a/arch/riscv/include/asm/spinlock.h
>> +++ b/arch/riscv/include/asm/spinlock.h
>> @@ -7,129 +7,91 @@
>> #ifndef _ASM_RISCV_SPINLOCK_H
>> #define _ASM_RISCV_SPINLOCK_H
>>
>> -#include <linux/kernel.h>
>> -#include <asm/current.h>
>> -#include <asm/fence.h>
>> -
>> /*
>> - * Simple spin lock operations. These provide no fairness guarantees.
>> + * Ticket-based spin-locking.
>> */
>> +static inline void arch_spin_lock(arch_spinlock_t *lock)
>> +{
>> + arch_spinlock_t lockval;
>> + u32 tmp;
>> +
>> + asm volatile (
>> + "1: lr.w %0, %2 \n"
>> + " mv %1, %0 \n"
>> + " addw %0, %0, %3 \n"
>> + " sc.w %0, %0, %2 \n"
>> + " bnez %0, 1b \n"
>> + : "=&r" (tmp), "=&r" (lockval), "+A" (lock->lock)
>> + : "r" (1 << TICKET_NEXT)
>> + : "memory");
>>
>> -/* FIXME: Replace this with a ticket lock, like MIPS. */
>> -
>> -#define arch_spin_is_locked(x) (READ_ONCE((x)->lock) != 0)
>> + while (lockval.tickets.next != lockval.tickets.owner) {
>> + /*
>> + * FIXME - we need wfi/wfe here to prevent:
>> + * - cache line bouncing
>> + * - saving cpu pipeline in multi-harts-per-core
>> + * processor
>> + */
>> + lockval.tickets.owner = READ_ONCE(lock->tickets.owner);
>> + }
>>
>> -static inline void arch_spin_unlock(arch_spinlock_t *lock)
>> -{
>> - smp_store_release(&lock->lock, 0);
>> + __atomic_acquire_fence();
>> }
>>
>> static inline int arch_spin_trylock(arch_spinlock_t *lock)
>> {
>> - int tmp = 1, busy;
>> -
>> - __asm__ __volatile__ (
>> - " amoswap.w %0, %2, %1\n"
>> - RISCV_ACQUIRE_BARRIER
>> - : "=r" (busy), "+A" (lock->lock)
>> - : "r" (tmp)
>> + u32 tmp, contended, res;
>> +
>> + do {
>> + asm volatile (
>> + " lr.w %0, %3 \n"
>> + " srliw %1, %0, %5 \n"
>> + " slliw %2, %0, %5 \n"
>> + " or %1, %2, %1 \n"
>> + " li %2, 0 \n"
>> + " sub %1, %1, %0 \n"
>> + " bnez %1, 1f \n"
>> + " addw %0, %0, %4 \n"
>> + " sc.w %2, %0, %3 \n"
>> + "1: \n"
>> + : "=&r" (tmp), "=&r" (contended), "=&r" (res),
>> + "+A" (lock->lock)
>> + : "r" (1 << TICKET_NEXT), "I" (TICKET_NEXT)
>> : "memory");
>> + } while (res);
>>
>> - return !busy;
>> -}
>> -
>> -static inline void arch_spin_lock(arch_spinlock_t *lock)
>> -{
>> - while (1) {
>> - if (arch_spin_is_locked(lock))
>> - continue;
>> -
>> - if (arch_spin_trylock(lock))
>> - break;
>> + if (!contended) {
>> + __atomic_acquire_fence();
>> + return 1;
>> + } else {
>> + return 0;
>> }
>> }
>>
>> -/***********************************************************/
>> -
>> -static inline void arch_read_lock(arch_rwlock_t *lock)
>> +static inline void arch_spin_unlock(arch_spinlock_t *lock)
>> {
>> - int tmp;
>> -
>> - __asm__ __volatile__(
>> - "1: lr.w %1, %0\n"
>> - " bltz %1, 1b\n"
>> - " addi %1, %1, 1\n"
>> - " sc.w %1, %1, %0\n"
>> - " bnez %1, 1b\n"
>> - RISCV_ACQUIRE_BARRIER
>> - : "+A" (lock->lock), "=&r" (tmp)
>> - :: "memory");
>> + smp_store_release(&lock->tickets.owner, lock->tickets.owner + 1);
>> + /* FIXME - we need ipi/sev here to notify above */
>> }
>>
>> -static inline void arch_write_lock(arch_rwlock_t *lock)
>> +static inline int arch_spin_value_unlocked(arch_spinlock_t lock)
>> {
>> - int tmp;
>> -
>> - __asm__ __volatile__(
>> - "1: lr.w %1, %0\n"
>> - " bnez %1, 1b\n"
>> - " li %1, -1\n"
>> - " sc.w %1, %1, %0\n"
>> - " bnez %1, 1b\n"
>> - RISCV_ACQUIRE_BARRIER
>> - : "+A" (lock->lock), "=&r" (tmp)
>> - :: "memory");
>> + return lock.tickets.owner == lock.tickets.next;
>> }
>>
>> -static inline int arch_read_trylock(arch_rwlock_t *lock)
>> +static inline int arch_spin_is_locked(arch_spinlock_t *lock)
>> {
>> - int busy;
>> -
>> - __asm__ __volatile__(
>> - "1: lr.w %1, %0\n"
>> - " bltz %1, 1f\n"
>> - " addi %1, %1, 1\n"
>> - " sc.w %1, %1, %0\n"
>> - " bnez %1, 1b\n"
>> - RISCV_ACQUIRE_BARRIER
>> - "1:\n"
>> - : "+A" (lock->lock), "=&r" (busy)
>> - :: "memory");
>> -
>> - return !busy;
>> + return !arch_spin_value_unlocked(READ_ONCE(*lock));
>> }
>>
>> -static inline int arch_write_trylock(arch_rwlock_t *lock)
>> +static inline int arch_spin_is_contended(arch_spinlock_t *lock)
>> {
>> - int busy;
>> -
>> - __asm__ __volatile__(
>> - "1: lr.w %1, %0\n"
>> - " bnez %1, 1f\n"
>> - " li %1, -1\n"
>> - " sc.w %1, %1, %0\n"
>> - " bnez %1, 1b\n"
>> - RISCV_ACQUIRE_BARRIER
>> - "1:\n"
>> - : "+A" (lock->lock), "=&r" (busy)
>> - :: "memory");
>> + struct __raw_tickets tickets = READ_ONCE(lock->tickets);
>>
>> - return !busy;
>> + return (tickets.next - tickets.owner) > 1;
>> }
>> +#define arch_spin_is_contended arch_spin_is_contended
>>
>> -static inline void arch_read_unlock(arch_rwlock_t *lock)
>> -{
>> - __asm__ __volatile__(
>> - RISCV_RELEASE_BARRIER
>> - " amoadd.w x0, %1, %0\n"
>> - : "+A" (lock->lock)
>> - : "r" (-1)
>> - : "memory");
>> -}
>> -
>> -static inline void arch_write_unlock(arch_rwlock_t *lock)
>> -{
>> - smp_store_release(&lock->lock, 0);
>> -}
>> +#include <asm/qrwlock.h>
>>
>> #endif /* _ASM_RISCV_SPINLOCK_H */
>> diff --git a/arch/riscv/include/asm/spinlock_types.h b/arch/riscv/include/asm/spinlock_types.h
>> index f398e76..d7b38bf 100644
>> --- a/arch/riscv/include/asm/spinlock_types.h
>> +++ b/arch/riscv/include/asm/spinlock_types.h
>> @@ -10,16 +10,21 @@
>> # error "please don't include this file directly"
>> #endif
>>
>> +#define TICKET_NEXT 16
>> +
>> typedef struct {
>> - volatile unsigned int lock;
>> + union {
>> + u32 lock;
>> + struct __raw_tickets {
>> + /* little endian */
>> + u16 owner;
>> + u16 next;
>> + } tickets;
>> + };
>> } arch_spinlock_t;
>>
>> -#define __ARCH_SPIN_LOCK_UNLOCKED { 0 }
>> -
>> -typedef struct {
>> - volatile unsigned int lock;
>> -} arch_rwlock_t;
>> +#define __ARCH_SPIN_LOCK_UNLOCKED { { 0 } }
>>
>> -#define __ARCH_RW_LOCK_UNLOCKED { 0 }
>> +#include <asm-generic/qrwlock_types.h>
>>
>> #endif /* _ASM_RISCV_SPINLOCK_TYPES_H */
>> --
>> 2.7.4
>>
>>
>> _______________________________________________
>> linux-riscv mailing list
>> linux-riscv@xxxxxxxxxxxxxxxxxxx
>> http://lists.infradead.org/mailman/listinfo/linux-riscv



--
Best Regards
Guo Ren

ML: https://lore.kernel.org/linux-csky/