[RFC v2 PATCH 4/7] arm64: tlbflush.h: add __tlbi() macro
Christoffer Dall
christoffer.dall at linaro.org
Mon Sep 5 23:38:30 PDT 2016
On Mon, Sep 05, 2016 at 05:31:34PM +0100, Punit Agrawal wrote:
> From: Mark Rutland <mark.rutland at arm.com>
>
> As with dsb() and isb(), add a __tlbi() helper so that we can avoid
> distracting asm boilerplate every time we want a TLBI. As some TLBI
> operations take an argument while others do not, some pre-processor is
> used to handle these two cases with different assembly blocks.
>
> The existing tlbflush.h code is moved over to use the helper.
>
> Signed-off-by: Mark Rutland <mark.rutland at arm.com>
> Cc: Catalin Marinas <catalin.marinas at arm.com>
> Cc: Marc Zyngier <marc.zyngier at arm.com>
> Cc: Will Deacon <will.deacon at arm.com>
> [ rename helper to __tlbi, update commit log ]
> Signed-off-by: Punit Agrawal <punit.agrawal at arm.com>
> ---
> arch/arm64/include/asm/tlbflush.h | 31 +++++++++++++++++++++++--------
> 1 file changed, 23 insertions(+), 8 deletions(-)
>
> diff --git a/arch/arm64/include/asm/tlbflush.h b/arch/arm64/include/asm/tlbflush.h
> index b460ae2..e9e42a7 100644
> --- a/arch/arm64/include/asm/tlbflush.h
> +++ b/arch/arm64/include/asm/tlbflush.h
> @@ -25,6 +25,21 @@
> #include <asm/cputype.h>
>
> /*
> + * Raw TLBI operations. Drivers and most kernel code should use the TLB
> + * management routines below in preference to these. Where necessary, these can
in preference to these?
> + * be used to avoid asm() boilerplate.
'these' seem to be a bit overloaded in this commentary.
> + *
> + * Can be used as __tlbi(op) or __tlbi(op, arg), depending on whether a
> + * particular TLBI op takes an argument or not. The macros below handle invoking
> + * the asm with or without the register argument as appropriate.
> + */
> +#define __TLBI_0(op, arg) asm ("tlbi " #op)
> +#define __TLBI_1(op, arg) asm ("tlbi " #op ", %0" : : "r" (arg))
> +#define __TLBI_N(op, arg, n, ...) __TLBI_##n(op, arg)
> +
> +#define __tlbi(op, ...) __TLBI_N(op, ##__VA_ARGS__, 1, 0)
> +
> +/*
> * TLB Management
> * ==============
> *
> @@ -66,7 +81,7 @@
> static inline void local_flush_tlb_all(void)
> {
> dsb(nshst);
> - asm("tlbi vmalle1");
> + __tlbi(vmalle1);
> dsb(nsh);
> isb();
> }
> @@ -74,7 +89,7 @@ static inline void local_flush_tlb_all(void)
> static inline void flush_tlb_all(void)
> {
> dsb(ishst);
> - asm("tlbi vmalle1is");
> + __tlbi(vmalle1is);
> dsb(ish);
> isb();
> }
> @@ -84,7 +99,7 @@ static inline void flush_tlb_mm(struct mm_struct *mm)
> unsigned long asid = ASID(mm) << 48;
>
> dsb(ishst);
> - asm("tlbi aside1is, %0" : : "r" (asid));
> + __tlbi(aside1is, asid);
> dsb(ish);
> }
>
> @@ -94,7 +109,7 @@ static inline void flush_tlb_page(struct vm_area_struct *vma,
> unsigned long addr = uaddr >> 12 | (ASID(vma->vm_mm) << 48);
>
> dsb(ishst);
> - asm("tlbi vale1is, %0" : : "r" (addr));
> + __tlbi(vale1is, addr);
> dsb(ish);
> }
>
> @@ -122,9 +137,9 @@ static inline void __flush_tlb_range(struct vm_area_struct *vma,
> dsb(ishst);
> for (addr = start; addr < end; addr += 1 << (PAGE_SHIFT - 12)) {
> if (last_level)
> - asm("tlbi vale1is, %0" : : "r"(addr));
> + __tlbi(vale1is, addr);
> else
> - asm("tlbi vae1is, %0" : : "r"(addr));
> + __tlbi(vae1is, addr);
> }
> dsb(ish);
> }
> @@ -149,7 +164,7 @@ static inline void flush_tlb_kernel_range(unsigned long start, unsigned long end
>
> dsb(ishst);
> for (addr = start; addr < end; addr += 1 << (PAGE_SHIFT - 12))
> - asm("tlbi vaae1is, %0" : : "r"(addr));
> + __tlbi(vaae1is, addr);
> dsb(ish);
> isb();
> }
> @@ -163,7 +178,7 @@ static inline void __flush_tlb_pgtable(struct mm_struct *mm,
> {
> unsigned long addr = uaddr >> 12 | (ASID(mm) << 48);
>
> - asm("tlbi vae1is, %0" : : "r" (addr));
> + __tlbi(vae1is, addr);
> dsb(ish);
> }
>
> --
> 2.8.1
>
More information about the linux-arm-kernel
mailing list