[PATCH] arm64: spinlock: clarify implementation details of arch_spin_lock
Vladimir Murzin
vladimir.murzin at arm.com
Thu Sep 1 03:47:00 PDT 2016
It seems to be quite confusing to see atomic load not being paired
with atomic store down to arch_spin_lock function. To prevent the same
questions/patches around this add a comment block explaining what is
going on there.
The comment has been stolen from Catalin's reply [1].
[1] https://lkml.org/lkml/2016/8/30/127
Signed-off-by: Vladimir Murzin <vladimir.murzin at arm.com>
---
arch/arm64/include/asm/spinlock.h | 12 ++++++++++++
1 file changed, 12 insertions(+)
diff --git a/arch/arm64/include/asm/spinlock.h b/arch/arm64/include/asm/spinlock.h
index e875a5a..9a2155c 100644
--- a/arch/arm64/include/asm/spinlock.h
+++ b/arch/arm64/include/asm/spinlock.h
@@ -113,6 +113,18 @@ static inline void arch_spin_lock(arch_spinlock_t *lock)
*/
" sevl\n"
"2: wfe\n"
+ /*
+ * Don't be confused with atomic load bellow not being paired
+ * with atomic store. This is needed because the
+ * arch_spin_unlock() code only uses an STLR without an
+ * explicit SEV (like we have on AArch32). An event is
+ * automatically generated when the exclusive monitor is
+ * cleared by STLR. But without setting it with a load
+ * exclusive in arch_spin_lock() (even though it does not
+ * acquire the lock), there won't be anything to clear, hence
+ * no event to be generated. In this case, the WFE would wait
+ * indefinitely.
+ */
" ldaxrh %w2, %4\n"
" eor %w1, %w2, %w0, lsr #16\n"
" cbnz %w1, 2b\n"
--
1.7.9.5
More information about the linux-arm-kernel
mailing list