[PATCH 12/18] arm64: cmpxchg: avoid "cc" clobber in ll/sc routines
Will Deacon
will.deacon at arm.com
Tue Jul 21 10:32:12 PDT 2015
On Tue, Jul 21, 2015 at 06:16:07PM +0100, Catalin Marinas wrote:
> On Mon, Jul 13, 2015 at 10:25:13AM +0100, Will Deacon wrote:
> > We can perform the cmpxchg comparison using eor and cbnz which avoids
> > the "cc" clobber for the ll/sc case and consequently for the LSE case
> > where we may have to fall-back on the ll/sc code at runtime.
> >
> > Reviewed-by: Steve Capper <steve.capper at arm.com>
> > Signed-off-by: Will Deacon <will.deacon at arm.com>
> > ---
> > arch/arm64/include/asm/atomic_ll_sc.h | 14 ++++++--------
> > arch/arm64/include/asm/atomic_lse.h | 4 ++--
> > 2 files changed, 8 insertions(+), 10 deletions(-)
> >
> > diff --git a/arch/arm64/include/asm/atomic_ll_sc.h b/arch/arm64/include/asm/atomic_ll_sc.h
> > index 77d3aabf52ad..d21091bae901 100644
> > --- a/arch/arm64/include/asm/atomic_ll_sc.h
> > +++ b/arch/arm64/include/asm/atomic_ll_sc.h
> > @@ -96,14 +96,13 @@ __LL_SC_PREFIX(atomic_cmpxchg(atomic_t *ptr, int old, int new))
> >
> > asm volatile("// atomic_cmpxchg\n"
> > "1: ldxr %w1, %2\n"
> > -" cmp %w1, %w3\n"
> > -" b.ne 2f\n"
> > +" eor %w0, %w1, %w3\n"
> > +" cbnz %w0, 2f\n"
> > " stxr %w0, %w4, %2\n"
> > " cbnz %w0, 1b\n"
> > "2:"
> > : "=&r" (tmp), "=&r" (oldval), "+Q" (ptr->counter)
> > - : "Ir" (old), "r" (new)
> > - : "cc");
> > + : "Lr" (old), "r" (new));
>
> For the LL/SC case, does this make things any slower? We replace a cmp +
> b.ne with two arithmetic ops (eor and cbnz, unless the latter is somehow
> smarter). I don't think the condition flags usually need to be preserved
> across an asm statement, so the "cc" clobber probably didn't make much
> difference anyway.
I doubt you can measure it either way. The main reason for changing this
was for consistency with other, similar code and improved readability
(since otherwise we have a mystery "cc" clobber in the LSE version).
Will
More information about the linux-arm-kernel
mailing list