[PATCH v3 3/6] riscv: Add Zawrs support for spinlocks
Andrew Jones
ajones at ventanamicro.com
Mon Jul 29 07:01:02 PDT 2024
On Mon, Jul 29, 2024 at 03:43:30PM GMT, Alexandre Ghiti wrote:
...
> > +static __always_inline void __cmpwait(volatile void *ptr,
> > + unsigned long val,
> > + int size)
> > +{
> > + unsigned long tmp;
> > +
> > + asm goto(ALTERNATIVE("j %l[no_zawrs]", "nop",
> > + 0, RISCV_ISA_EXT_ZAWRS, 1)
> > + : : : : no_zawrs);
> > +
> > + switch (size) {
> > + case 4:
> > + asm volatile(
> > + " lr.w %0, %1\n"
> > + " xor %0, %0, %2\n"
> > + " bnez %0, 1f\n"
> > + ZAWRS_WRS_NTO "\n"
> > + "1:"
> > + : "=&r" (tmp), "+A" (*(u32 *)ptr)
> > + : "r" (val));
> > + break;
> > +#if __riscv_xlen == 64
> > + case 8:
> > + asm volatile(
> > + " lr.d %0, %1\n"
> > + " xor %0, %0, %2\n"
> > + " bnez %0, 1f\n"
> > + ZAWRS_WRS_NTO "\n"
> > + "1:"
> > + : "=&r" (tmp), "+A" (*(u64 *)ptr)
> > + : "r" (val));
> > + break;
> > +#endif
> > + default:
> > + BUILD_BUG();
> > + }
> > +
> > + return;
> > +
> > +no_zawrs:
> > + asm volatile(RISCV_PAUSE : : : "memory");
>
>
> Shouldn't we fallback to the previous implementation (cpu_relax()) here? Not
> sure this is really important, but I want to make sure it was not an
> oversight.
>
Hi Alex,
It was intentional. We can't easily call cpu_relax() from here because
asm/vdso/processor.h includes asm/barrier.h which includes asm/cmpxchg.h.
We've mostly reproduced cpu_relax() here since we're only skipping the
div, and, as __cmpwait will be used in loops which load memory for the
comparison, I didn't think we needed the extra div stalls.
Thanks,
drew
More information about the linux-riscv
mailing list