[PATCH 1/2] ARM: spinlock: use ticket algorithm for ARMv6+ locking implementation
Will Deacon
will.deacon at arm.com
Mon Jun 25 05:36:48 EDT 2012
Hi Nicolas,
Thanks for looking at this.
On Fri, Jun 22, 2012 at 09:08:47PM +0100, Nicolas Pitre wrote:
> On Fri, 22 Jun 2012, Will Deacon wrote:
>
> > Ticket spinlocks ensure locking fairness by reducing the thundering herd
> > effect when acquiring a lock.
>
> The conventional thundering herd effect is still there as all waiting
> CPUs are still woken up from WFE upon a spin_unlock with a SEV. Maybe
> the fact that the actual spinning is no longer banging on the exclusion
> monitor does count as reduction of the thundering herd effect at that
> level though, and if that is what you meant then this could be precised
> here.
Sure, I can clarify that. The thundering herd effect is reduced because (a)
the cacheline can remain shared and (b) the memory access time doesn't
determine the winner.
> > This is especially important on systems
> > where memory-access times are not necessarily uniform when accessing
> > the lock structure (for example, on a multi-cluster platform where the
> > lock is allocated into L1 when a CPU releases it).
>
> In which case it is more about fairness. This algorithm brings fairness
> due to its FIFO nature, despite possible memory access speed
> differences. And that is even more important than the thundering herd
> effect. Solving both at once is of course all good.
Precisely.
> > This patch implements the ticket spinlock algorithm for ARM, replacing
> > the simpler implementation for ARMv6+ processors.
> >
> > Cc: Nicolas Pitre <nico at fluxnic.net>
> > Signed-off-by: Will Deacon <will.deacon at arm.com>
>
> A minor remarks below. Otherwise...
>
> Reviewed-by: Nicolas Pitre <nico at linaro.org>
Thanks Nicolas.
> > static inline int arch_spin_trylock(arch_spinlock_t *lock)
> > {
> > unsigned long tmp;
> > + u32 slock;
> >
> > __asm__ __volatile__(
> > -" ldrex %0, [%1]\n"
> > -" teq %0, #0\n"
> > -" strexeq %0, %2, [%1]"
> > - : "=&r" (tmp)
> > - : "r" (&lock->lock), "r" (1)
> > +" ldrex %0, [%2]\n"
> > +" cmp %0, %0, ror #16\n"
> > +" movne %1, #1\n"
>
> You could replace the above 2 insns with:
>
> subs %1, %0, %0, ror #16
Wahey, that is extremely concise!
Cheers,
Will
More information about the linux-arm-kernel
mailing list