Re: [PATCH 08/11] membarrier: Provide core serializing command (v2)

From: Mathieu Desnoyers
Date: Mon Jan 29 2018 - 14:20:30 EST


----- On Jan 29, 2018, at 2:09 PM, Peter Zijlstra peterz@xxxxxxxxxxxxx wrote:

> On Mon, Jan 29, 2018 at 06:36:05PM +0000, Mathieu Desnoyers wrote:
>> ----- On Jan 29, 2018, at 1:15 PM, Peter Zijlstra peterz@xxxxxxxxxxxxx wrote:
>
>> > Aaah, its the case where we do not pass through switch_mm(), the partial
>> > comment got to me. I only realized after reading the next patch.
>>
>> Indeed, if we read the entire comment, it's made clear that this case is for
>> when switch_mm is not invoked, where the current mm is changed without going
>> through switch_mm(), when scheduling between uthread->kthread->uthread for
>> instance.
>>
>> /*
>> * When transitioning from a kernel thread to a userspace
>> * thread, mmdrop()'s implicit full barrier is required by the
>> * membarrier system call, because the current active_mm can
>> * become the current mm without going through switch_mm().
>> * membarrier also requires a core serializing instruction
>> * before going back to user-space after storing to rq->curr.
>> */
>>
>> Is there something I should improve in the wording of this added
>> sentence to make it clearer ?
>
> Can be improved I think, its got two unqualified "membarrier"s in and
> its a bit mixed up. I'm having a major case of the mondays (brain just
> won't start today), but maybe something like:
>
> When we switched through a kernel thread, the loop in
> membarrier_{private,global}_expedited() can have observed that
> kernel thread and not issued an IPI. We will also not pass
> through switch_mm(). Membarrier requires a barrier after writing
> rq->curr and returning to userspace, so provide them here:
>
> - a full memory barrier for {PRIVATE,GLOBAL}_EXPEDITED
> - a sync_core for SYNC_CORE

Editing to remove use of "we" and clarify, which ends up as:

/*
* When switching through a kernel thread, the loop in
* membarrier_{private,global}_expedited() may have observed that
* kernel thread and not issued an IPI. It is therefore possible to
* schedule between user->kernel->user threads without passing though
* switch_mm(). Membarrier requires a barrier after storing to
* rq->curr, before returning to userspace, so provide them here:
*
* - a full memory barrier for {PRIVATE,GLOBAL}_EXPEDITED, implicitly
* provided by mmdrop(),
* - a sync_core for SYNC_CORE.
*/

>
> Also I think changing the changlog to state where we need core-sync
> would be good. Currently the x86 patch does that, but not this one,
> while this introduces the feature.

Planning to add this:

Architectures selecting this feature need to either document that
they issue core serializing instructions when returning to user-space,
or implement their architecture-specific sync_core_before_usermode().

Thanks,

Mathieu

--
Mathieu Desnoyers
EfficiOS Inc.
http://www.efficios.com