Re: [PATCH] mm: slub: Ensure that slab_unlock() is atomic
From: Vineet Gupta
Date: Wed Mar 09 2016 - 08:23:52 EST
On Wednesday 09 March 2016 03:43 PM, Peter Zijlstra wrote:
>> There is clearly a problem in slub code that it is pairing a test_and_set_bit()
>> with a __clear_bit(). Latter can obviously clobber former if they are not a single
>> instruction each unlike x86 or they use llock/scond kind of instructions where the
>> interim store from other core is detected and causes a retry of whole llock/scond
>> sequence.
>
> Yes, test_and_set_bit() + __clear_bit() is broken.
But in SLUB: bit_spin_lock() + __bit_spin_unlock() is acceptable ? How so
(ignoring the performance thing for discussion sake, which is a side effect of
this implementation).
So despite the comment below in bit_spinlock.h I don't quite comprehend how this
is allowable. And if say, by deduction, this is fine for LLSC or lock prefixed
cases, then isn't this true in general for lot more cases in kernel, i.e. pairing
atomic lock with non-atomic unlock ? I'm missing something !
| /*
| * bit-based spin_unlock()
| * non-atomic version, which can be used eg. if the bit lock itself is
| * protecting the rest of the flags in the word.
| */
| static inline void __bit_spin_unlock(int bitnum, unsigned long *addr)