Re: [RFC PATCH] introduce sys_membarrier(): process-wide memorybarrier

From: Steven Rostedt
Date: Sun Jan 10 2010 - 16:04:09 EST


On Sun, 2010-01-10 at 12:10 -0500, Mathieu Desnoyers wrote:

> >
> > CPU 0 CPU 1
> > ---------- --------------
> > obj = list->obj;
> > <user space>
> > rcu_read_lock();
> > obj = rcu_dereference(list->obj);
> > obj->foo = bar;
> >
> > <preempt>
> > <kernel space>
> >
> > schedule();
> > cpumask_clear(mm_cpumask, cpu);
> >
> > sys_membarrier();
> > free(obj);
> >
> > <store to obj->foo goes to memory> <- corruption
> >
>
> Hrm, having a writer like this in a rcu read-side would be a bit weird.
> We have to look at the actual rcu_read_lock() implementation in urcu to
> see why load/stores are important on the rcu read-side.
>

No it is not weird, it is common. The read is on the link list that we
can access. Yes a write should be protected by other locks, so maybe
that is the weird part.

> (note: _STORE_SHARED is simply a volatile store)
>
> (Thread-local variable, shared with the thread doing synchronize_rcu())
> struct urcu_reader __thread urcu_reader;
>
> static inline void _rcu_read_lock(void)
> {
> long tmp;
>
> tmp = urcu_reader.ctr;
> if (likely(!(tmp & RCU_GP_CTR_NEST_MASK))) {
> _STORE_SHARED(urcu_reader.ctr, _LOAD_SHARED(urcu_gp_ctr));
> /*
> * Set active readers count for outermost nesting level before
> * accessing the pointer. See force_mb_all_threads().
> */
> barrier();
> } else {
> _STORE_SHARED(urcu_reader.ctr, tmp + RCU_GP_COUNT);
> }
> }
>
> So as you see here, we have to ensure that the store to urcu_reader.ctr
> is globally visible before entering the critical section (previous
> stores must complete before following loads). For rcu_read_unlock, it's
> the opposite:
>
> static inline void _rcu_read_unlock(void)
> {
> long tmp;
>
> tmp = urcu_reader.ctr;
> /*
> * Finish using rcu before decrementing the pointer.
> * See force_mb_all_threads().
> */
> if (likely((tmp & RCU_GP_CTR_NEST_MASK) == RCU_GP_COUNT)) {
> barrier();
> _STORE_SHARED(urcu_reader.ctr, urcu_reader.ctr - RCU_GP_COUNT);
> } else {
> _STORE_SHARED(urcu_reader.ctr, urcu_reader.ctr - RCU_GP_COUNT);
> }
> }

Thanks for the insight of the code. I need to get around and look at
your userspace implementation ;-)

>
> We need to ensure that previous loads complete before following stores.
>
> Therefore, the race with unlock showing that we need to order loads
> before stores:
>
> CPU 0 CPU 1
> -------------- --------------
> <user space> (already in read-side C.S.)
> obj = rcu_dereference(list->next);
> -> load list->next
> copy = obj->foo;
> rcu_read_unlock();
> -> store to urcu_reader.ctr
> <urcu_reader.ctr store is globally visible>
> list_del(obj);
> <preempt>
> <kernel space>
>
> schedule();
> cpumask_clear(mm_cpumask, cpu);

but here we are switching to a new task.

>
> sys_membarrier();
> set global g.p. (urcu_gp_ctr) phase to 1
> wait for all urcu_reader.ctr in phase 0
> set global g.p. (urcu_gp_ctr) phase to 0
> wait for all urcu_reader.ctr in phase 1
> sys_membarrier();
> free(obj);
> <list->next load hits memory>
> <obj->foo load hits memory> <- corruption

load of obj->foo is really load foo(obj) into some register. And for the
above to fail, that means that this load happened even after we switched
to kernel space, and that load of foo(obj) is still pending to get into
the thread stack that saved that register.

But I'm sure Paul will point me to some arch that does this ;-)

>
> >
> > So, if there's no smp_wmb() between the <preempt> and cpumask_clear()
> > then we have an issue?
>
> Considering the scenario above, we would need a full smp_mb() (or
> equivalent) rather than just smp_wmb() to be strictly correct.

I agree with Paul, we should just punt and grab the rq locks. That seems
to be the safest way without resorting to funny tricks to save 15% on a
slow path.

-- Steve


--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/