[PATCH 3/5] arm64: atomics: lse: define ANDs in terms of ANDNOTs
Will Deacon
will at kernel.org
Mon Dec 13 11:29:36 PST 2021
On Fri, Dec 10, 2021 at 03:14:08PM +0000, Mark Rutland wrote:
> The FEAT_LSE atomic instructions include atomic bit-clear instructions
> (`ldclr*` and `stclr*`) which can be used to directly implement ANDNOT
> operations. Each AND op is implemented as a copy of the corresponding
> ANDNOT op with a leading `mvn` instruction to apply a bitwise NOT to the
> `i` argument.
>
> As the compiler has no visibility of the `mvn`, this leads to less than
> optimal code generation when generating `i` into a register. For
> example, __lse_atomic_fetch_and(0xf, v) can be compiled to:
>
> mov w1, #0xf
> mvn w1, w1
> ldclral w1, w1, [x2]
>
> This patch improves this by replacing the `mvn` with NOT in C before the
> inline assembly block, e.g.
>
> i = ~i;
>
> This allows the compiler to generate `i` into a register more optimally,
> e.g.
>
> mov w1, #0xfffffff0
> ldclral w1, w1, [x2]
>
> With this change the assembly for each AND op is identical to the
> corresponding ANDNOT op (including barriers and clobbers), so I've
> removed the inline assembly and rewritten each AND op in terms of the
> corresponding ANDNOT op, e.g.
>
> | static inline void __lse_atomic_and(int i, atomic_t *v)
> | {
> | return __lse_atomic_andnot(~i, v);
> | }
>
> This is intended as an optimization and cleanup.
> There should be no functional change as a result of this patch.
>
> Signed-off-by: Mark Rutland <mark.rutland at arm.com>
> Cc: Boqun Feng <boqun.feng at gmail.com>
> Cc: Catalin Marinas <catalin.marinas at arm.com>
> Cc: Peter Zijlstra <peterz at infradead.org>
> Cc: Will Deacon <will at kernel.org>
> ---
> arch/arm64/include/asm/atomic_lse.h | 34 ++++-------------------------
> 1 file changed, 4 insertions(+), 30 deletions(-)
Acked-by: Will Deacon <will at kernel.org>
Will
More information about the linux-arm-kernel
mailing list