[PATCH] arm64: mte: avoid clearing PSTATE.TCO on entry unless necessary

Peter Collingbourne pcc at google.com
Wed Nov 10 14:07:46 PST 2021


On Mon, Nov 8, 2021 at 3:01 AM Mark Rutland <mark.rutland at arm.com> wrote:
>
> On Fri, Nov 05, 2021 at 01:18:38PM -0700, Peter Collingbourne wrote:
> > On Thu, Oct 7, 2021 at 3:20 AM Mark Rutland <mark.rutland at arm.com> wrote:
> > >
> > > On Tue, Oct 05, 2021 at 12:08:58PM -0700, Peter Collingbourne wrote:
> > > > On Tue, Oct 5, 2021 at 9:46 AM Catalin Marinas <catalin.marinas at arm.com> wrote:
> > > > > On Wed, Sep 29, 2021 at 12:45:24PM -0700, Peter Collingbourne wrote:
> > > > > > diff --git a/arch/arm64/kernel/entry.S b/arch/arm64/kernel/entry.S
> > > > > > index 2f69ae43941d..85ead6bbb38e 100644
> > > > > > --- a/arch/arm64/kernel/entry.S
> > > > > > +++ b/arch/arm64/kernel/entry.S
> > > > > > @@ -269,7 +269,28 @@ alternative_else_nop_endif
> > > > > >       .else
> > > > > >       add     x21, sp, #PT_REGS_SIZE
> > > > > >       get_current_task tsk
> > > > > > +     ldr     x0, [tsk, THREAD_SCTLR_USER]
> > > > > >       .endif /* \el == 0 */
> > > > > > +
> > > > > > +     /*
> > > > > > +      * Re-enable tag checking (TCO set on exception entry). This is only
> > > > > > +      * necessary if MTE is enabled in either the kernel or the userspace
> > > > > > +      * task in synchronous mode. With MTE disabled in the kernel and
> > > > > > +      * disabled or asynchronous in userspace, tag check faults (including in
> > > > > > +      * uaccesses) are not reported, therefore there is no need to re-enable
> > > > > > +      * checking. This is beneficial on microarchitectures where re-enabling
> > > > > > +      * TCO is expensive.
> > > > > > +      */
> > > > > > +#ifdef CONFIG_ARM64_MTE
> > > > > > +alternative_cb       kasan_hw_tags_enable
> > > > > > +     tbz     x0, #SCTLR_EL1_TCF0_SHIFT, 1f
> > > > > > +alternative_cb_end
> > > > > > +alternative_if ARM64_MTE
> > > > > > +     SET_PSTATE_TCO(0)
> > > > > > +alternative_else_nop_endif
> > > > > > +1:
> > > > > > +#endif
> > > > >
> > > > > I think we can get here from an interrupt as well. Can we guarantee that
> > > > > the sctlr_user is valid? We are not always in a user process context.
> > > >
> > > > Looking through the code in entry.S carefully it doesn't appear that
> > > > the tsk pointer is ever used when taking an exception from EL1. The
> > > > last user appears to have been removed in commit 3d2403fd10a1 ("arm64:
> > > > uaccess: remove set_fs()"). I just did a quick boot test on a couple
> > > > of platforms and things seem to work without the "get_current_task
> > > > tsk" line.
> > > >
> > > > So I can't be confident that tsk points to a valid task, but at least
> > > > it looks like it was the case prior to that commit.
> >
> > The EL1 entries unconditionally call enter_from_kernel_mode() which
> > (at least in some configurations) unconditionally accesses current, so
> > I'm reasonably confident that we have a valid current pointer here.
>
> That's the value in SP_EL0, which is valid so long as we never take an
> exception from the entry assembly itself. We restore that from __entry_task in
> the EL0 entry paths, and context-switch it when switching threads.
>
> I would like to get rid of `tsk` in the entry assembly, but that requires
> moving more of that assembly to C.
>
> > > > > Maybe only do the above checks if \el == 0, otherwise just bracket it
> > > > > with kasan_hw_tags_enable.
> > > >
> > > > Is it possible for us to do a uaccess inside the EL1 -> EL1 exception
> > > > handler? That was the one thing I was unsure about and it's why I
> > > > opted to do the same check regardless of EL.
> > >
> > > Today we can do so if we take a PMU IRQ from EL1 and try to do an EL0
> > > stack unwind or stack dump, and similar holds for some eBPF stuff.
> > >
> > > We also need to ensure that PSTATE.TCO is configured consistently so
> > > that context-switch works, otherwise where a CPU switches between tasks
> > > A and B, where A is preempted by an EL1 IRQ, and B is explicitly
> > > switching via a direct call to schedule(), the state of TCO will not be
> > > as expected (unless we track this per thread, and handle it in the
> > > context switch).
> >
> > Isn't this already context switched via storing the per-task SPSR?
> > Otherwise e.g. the TCO controls in __uaccess_disable_tco() and
> > __uaccess_enable_tco() would have the same problem.
>
> That's the case for pre-emptive scheduling off the back of an interrupt, but
> tasks can call into schedule without an intervening exception, e.g.  blocking
> on something like a mutex. Those blocking primitives aren't permitted to be
> used within a __uaccess_{disable,enable)_tco() critical section.

I see. I guess you wouldn't even need an EL1 IRQ to expose the bug
here -- you could schedule() directly from A to B where A is NONE and
B is SYNC. I think we can handle this by clearing TCO in
cpu_switch_to() with the same condition as for exception entry; done
in v2.

> > > I'd strongly prefer that the state of PSTATE.TCO upon taking an
> > > exception to EL1 is consistent (by the end of the early entry code),
> > > regardless of where that was taken from.
> > >
> > > Is it easier to manage this from within entry-common.c?
> >
> > It doesn't look like we have any common location to add code that runs
> > on every kernel entry in entry-common.c.
>
> I was thinking we could rework the existing mte_check_tfsr_*() callsites into
> mte_enter_from_{user,kernel}() hooks, which'd allow us to avoid redundant checks
> for MTE support.

We currently only check TFSR on entry from EL1 (which makes sense,
since we would not expect an entry from EL0 to have triggered an EL1
tag check fault earlier). So it seems like we'd want slightly
different code for the two ELs anyway.

> > I tried adding one (patch
> > below), but this ended up leading to a performance regression (versus
> > baseline) on some of the cores on the hardware that I have access to.
>
> How big a regression? On a real workload or a microbenchmark?

This was on the microbenchmark from:
https://patchwork.kernel.org/project/linux-arm-kernel/patch/20200801011152.39838-1-pcc@google.com/#23572981

I'm sorry, but I'm not sure that I can get into specifics of how large
the regression was.

> Is that true for a simple move of the current logic to C, i.e. for an
> unconditional SET_PSTATE_TCO(0) gated by cpus_have_final_cap(ARM64_MTE) ?

With the patch below I measured a performance regression relative to
baseline (this time on all cores).

Peter

diff --git a/arch/arm64/kernel/entry-common.c b/arch/arm64/kernel/entry-common.c
index 64cfe4a3798f..23dd72b0866c 100644
--- a/arch/arm64/kernel/entry-common.c
+++ b/arch/arm64/kernel/entry-common.c
@@ -17,6 +17,21 @@
 #include <asm/mmu.h>
 #include <asm/sysreg.h>

+static void mte_disable_tco_entry(void)
+{
+ /*
+ * Re-enable tag checking (TCO set on exception entry). This is only
+ * necessary if MTE is enabled in either the kernel or the userspace
+ * task in synchronous mode. With MTE disabled in the kernel and
+ * disabled or asynchronous in userspace, tag check faults (including in
+ * uaccesses) are not reported, therefore there is no need to re-enable
+ * checking. This is beneficial on microarchitectures where re-enabling
+ * TCO is expensive.
+ */
+ if (IS_ENABLED(CONFIG_ARM64_MTE))
+ asm volatile(SET_PSTATE_TCO(0));
+}
+
 /*
  * This is intended to match the logic in irqentry_enter(), handling the kernel
  * mode transitions only.
@@ -39,6 +54,7 @@ static void noinstr enter_from_kernel_mode(struct
pt_regs *regs)
  trace_hardirqs_off_finish();

  mte_check_tfsr_entry();
+ mte_disable_tco_entry();
 }

 /*
@@ -235,6 +251,7 @@ asmlinkage void noinstr enter_from_user_mode(void)


  CT_WARN_ON(ct_state() != CONTEXT_USER);
  user_exit_irqoff();
  trace_hardirqs_off_finish();
+ mte_disable_tco_entry();
 }

 asmlinkage void noinstr exit_to_user_mode(void)
diff --git a/arch/arm64/kernel/entry.S b/arch/arm64/kernel/entry.S
index 42c404d508ca..8c63279ffea7 100644
--- a/arch/arm64/kernel/entry.S
+++ b/arch/arm64/kernel/entry.S
@@ -339,13 +339,6 @@ alternative_if ARM64_HAS_IRQ_PRIO_MASKING
  msr_s SYS_ICC_PMR_EL1, x20
 alternative_else_nop_endif

- /* Re-enable tag checking (TCO set on exception entry) */
-#ifdef CONFIG_ARM64_MTE
-alternative_if ARM64_MTE
- SET_PSTATE_TCO(0)
-alternative_else_nop_endif
-#endif
-
  /*
  * Registers that may be useful after this macro is invoked:
  *



More information about the linux-arm-kernel mailing list