[RFC PATCH] riscv/locking: Strengthen spin_lock() and spin_unlock()
Paul E. McKenney
paulmck at linux.vnet.ibm.com
Thu Feb 22 10:13:17 PST 2018
On Thu, Feb 22, 2018 at 09:27:09AM -0800, Daniel Lustig wrote:
> On 2/22/2018 6:12 AM, Andrea Parri wrote:
> > On Thu, Feb 22, 2018 at 02:40:04PM +0100, Peter Zijlstra wrote:
> >> On Thu, Feb 22, 2018 at 01:19:50PM +0100, Andrea Parri wrote:
> >>
> >>> C unlock-lock-read-ordering
> >>>
> >>> {}
> >>> /* s initially owned by P1 */
> >>>
> >>> P0(int *x, int *y)
> >>> {
> >>> WRITE_ONCE(*x, 1);
> >>> smp_wmb();
> >>> WRITE_ONCE(*y, 1);
> >>> }
> >>>
> >>> P1(int *x, int *y, spinlock_t *s)
> >>> {
> >>> int r0;
> >>> int r1;
> >>>
> >>> r0 = READ_ONCE(*y);
> >>> spin_unlock(s);
> >>> spin_lock(s);
> >>> r1 = READ_ONCE(*y);
> >>> }
> >>>
> >>> exists (1:r0=1 /\ 1:r1=0)
> >>>
> >>> RISCV RISCV-unlock-lock-read-ordering
> >>> {
> >>> 0:x2=x; 0:x4=y;
> >>> 1:x2=y; 1:x4=x; 1:x6=s;
> >>> s=1;
> >>> }
> >>> P0 | P1 ;
> >>> ori x1,x0,1 | lw x1,0(x2) ;
> >>> sw x1,0(x2) | amoswap.w.rl x0,x0,(x6) ;
> >>> fence w,w | ori x5,x0,1 ;
> >>> ori x3,x0,1 | amoswap.w.aq x0,x5,(x6) ;
> >>> sw x3,0(x4) | lw x3,0(x4) ;
> >>> exists
> >>> (1:x1=1 /\ 1:x3=0)
> >>
> >> So I would indeed expect this to be forbidden. Could someone please
> >> explain how this could be allowed?
> >
> > As mentioned in IRC, my understanding here is only based on the spec.
> > referred below and on its (available) formalizations. I expect that
> > RISC-V people will be able to provide more information.
>
> I think an RCpc interpretation of .aq and .rl would in fact allow
> the two normal loads in P1 to be reordered. Wouldn't it? Does that
> go against what the LKMM requires?
>
> The intuition would be that the amoswap.w.aq can forward from the
> amoswap.w.rl while that's still in the store buffer, and then the
> lw x3,0(x4) can also perform while the amoswap.w.rl is still in
> the store buffer, all before the l1 x1,0(x2) executes. That's
> not forbidden unless the amoswaps are RCsc, unless I'm missing
> something.
>
> Likewise even if the unlock()/lock() is between two stores. A
> control dependency might originate from the load part of the
> amoswap.w.aq, but there still would have to be something to
> ensure that this load part in fact performs after the store
> part of the amoswap.w.rl performs globally, and that's not
> automatic under RCpc.
>
> In any case, our concern that amoswap.w.rl and amoswap.w.aq might
> not actually be sufficient for spin_unlock() and spin_lock() was
> what prompted us to start our own internal discussion on this.
>
> FWIW we have made a few clarifications to our spec that haven't
> propagated to the public upstream yet, but that should be happening
> soon. In any case as it relates to the questions here, it's
> more a matter of us deciding what we should put into the spec in
> the end than it is a matter of checking the current text. In
> other words, we're still finalizing things, and there's still
> some time to tweak as necessary.
>
> >>
> >>> C unlock-lock-write-ordering
> >>>
> >>> {}
> >>> /* s initially owned by P0 */
> >>>
> >>> P0(int *x, int *y, spinlock_t *s)
> >>> {
> >>> WRITE_ONCE(*x, 1);
> >>> spin_unlock(s);
> >>> spin_lock(s);
> >>> WRITE_ONCE(*y, 1);
> >>> }
> >>>
> >>> P1(int *x, int *y)
> >>> {
> >>> int r0;
> >>> int r1;
> >>>
> >>> r0 = READ_ONCE(*y);
> >>> smp_rmb();
> >>> r1 = READ_ONCE(*y);
> >>> }
> >>>
> >>> exists (1:r0=1 /\ 1:r1=0)
> >>>
> >>> RISCV RISCV-unlock-lock-write-ordering
> >>> {
> >>> 0:x2=x; 0:x4=y; 0:x6=s;
> >>> 1:x2=y; 1:x4=x;
> >>> s=1;
> >>> }
> >>> P0 | P1 ;
> >>> ori x1,x0,1 | lw x1,0(x2) ;
> >>> sw x1,0(x2) | fence r,r ;
> >>> amoswap.w.rl x0,x0,(x6) | lw x3,0(x4) ;
> >>> ori x5,x0,1 | ;
> >>> amoswap.w.aq x0,x5,(x6) | ;
> >>> ori x3,x0,1 | ;
> >>> sw x3,0(x4) | ;
> >>> exists
> >>> (1:x1=1 /\ 1:x3=0)
> >>
> >> And here I think the RISCV conversion is flawed, there should be a ctrl
> >> dependency. The second store-word in P0 should depend on the result of
> >> amoswap.w.aq being 0.
> >
> > You're right: AFAICT, this can be remedied by inserting "beq x0,x5,FAIL00"
> > right after amoswap.w.aq (and this label at the end of P0); this does not
> > change the verdict of the available formalizations reported above however.
> >
> > (So, AFAICT, the above question remains valid/open.)
>
> This makes sense, but one other question: Paul mentioned earlier in the
> thread that
>
> > The PowerPC lock implementation's unlock-lock pair does not order writes
> > from the previous critical section against reads from the later critical
> > section, but it does order other combinations of reads and writes.
>
> My understanding is that the same logic about control dependencies applies
> here too, right? In other words, in spite of Peter's claim, the control
> dependency doesn't automatically fix this for RISC-V or for PowerPC?
For PowerPC, an lwsync instruction is used, which means that an external
obsserver can see a write from an earlier critical section being
reordered with a read from a later critical section, but no other
combination of reads and writes can be so reordered.
I will dig up PowerPC litmus tests demonstrating this and send them
along.
> >> (strictly speaking there should be a ctrl-dep in the read example too,
> >> except it'd be pointless for ordering reads, so I accept it being left
> >> out)
> >>
> >> Again, I cannot see how this could be allowed.
> >>
>
> Peter, in this test you mentioned earlier:
>
> P0()
> {
> WRITE_ONCE(x, 1);
> smp_store_release(&y, 1);
> r0 = smp_load_acquire(&y);
> WRITE_ONCE(z, 1);
> }
>
> P1()
> {
> r1 = READ_ONCE(z);
> smp_rmb();
> r2 = READ_ONCE(x);
> }
>
> exists: r0 == 1 /\ r1==1 /\ r2==0
>
> You expect this to be forbidden too, even if the release and acquire
> are RCpc? This part I don't follow. There's no conditional branch
> here to enforce a control dependency. I get and I agree that
> smp_store_release/smp_load_acquire are different than spin_unlock/
> spin_lock, but isn't that even more reason to allow this one if
> smp_store_release/smp_load_acquire are RCpc without question?
Here is the above test, reworked a bit to allow herd to accept it:
------------------------------------------------------------------------
C C-peterz
{
}
P0(int *x, int *y, int *z)
{
int r0;
WRITE_ONCE(*x, 1);
smp_store_release(y, 1);
r0 = smp_load_acquire(y);
WRITE_ONCE(*z, 1);
}
P1(int *x, int *y, int *z)
{
int r1;
int r2;
r1 = READ_ONCE(*z);
smp_rmb();
r2 = READ_ONCE(*x);
}
exists (0:r0 == 1 /\ 1:r1==1 /\ 1:r2==0)
------------------------------------------------------------------------
Please let me know if I messed up the conversion. On the perhaps unlikely
chance that I got it right, here is what the current version of the
Linux-kernel memory model says:
------------------------------------------------------------------------
$ herd7 -conf linux-kernel.cfg /tmp/C-peterz.litmus
Test C-peterz Allowed
States 4
0:r0=1; 1:r1=0; 1:r2=0;
0:r0=1; 1:r1=0; 1:r2=1;
0:r0=1; 1:r1=1; 1:r2=0;
0:r0=1; 1:r1=1; 1:r2=1;
Ok
Witnesses
Positive: 1 Negative: 3
Condition exists (0:r0=1 /\ 1:r1=1 /\ 1:r2=0)
Observation C-peterz Sometimes 1 3
Time C-peterz 0.01
Hash=6c4f7f8a8dc750fd5d706d6f3c104360
------------------------------------------------------------------------
So the current Linux-kernel memory model does in fact allow this, as
you say. But again, there is a proposal to tighten the Linux-kernel
memory model based on the pre-RISC-V set of architectures running the
Linux kernel, all of which would forbid this test.
So we have something that is not all that rare in the Linux kernel
community, namely two conflicting more-or-less concurrent changes.
This clearly needs to be resolved, either by us not strengthening the
Linux-kernel memory model in the way we were planning to or by you
strengthening RISC-V to be no weaker than PowerPC for these sorts of
externally viewed release-acquire situations.
Other thoughts?
Thanx, Paul
More information about the linux-riscv
mailing list