[RFC PATCH 04/14] arm64: kernel: suspend/resume registers save/restore
Lorenzo Pieralisi
lorenzo.pieralisi at arm.com
Mon Sep 2 05:57:28 EDT 2013
On Fri, Aug 30, 2013 at 06:23:10PM +0100, Catalin Marinas wrote:
> On Wed, Aug 28, 2013 at 12:35:56PM +0100, Lorenzo Pieralisi wrote:
> > Power management software requires the kernel to save and restore
> > CPU registers while going through suspend and resume operations
> > triggered by kernel subsystems like CPU idle and suspend to RAM.
> >
> > This patch implements code that provides save and restore mechanism
> > for the arm v8 implementation. Memory for the context is passed as
> > parameter to both cpu_do_suspend and cpu_do_resume functions, and allows
> > the callers to implement context allocation as they deem fit.
> >
> > The registers that are saved and restored correspond to the registers
> > actually required by the kernel to be up and running and is by no means
> > a complete save and restore of the entire v8 register set.
> >
> > Signed-off-by: Lorenzo Pieralisi <lorenzo.pieralisi at arm.com>
> > ---
> > arch/arm64/include/asm/proc-fns.h | 3 ++
> > arch/arm64/mm/proc.S | 64 +++++++++++++++++++++++++++++++++++++++
> > 2 files changed, 67 insertions(+)
> >
> > diff --git a/arch/arm64/include/asm/proc-fns.h b/arch/arm64/include/asm/proc-fns.h
> > index 7cdf466..0c657bb 100644
> > --- a/arch/arm64/include/asm/proc-fns.h
> > +++ b/arch/arm64/include/asm/proc-fns.h
> > @@ -26,11 +26,14 @@
> > #include <asm/page.h>
> >
> > struct mm_struct;
> > +struct cpu_suspend_ctx;
> >
> > extern void cpu_cache_off(void);
> > extern void cpu_do_idle(void);
> > extern void cpu_do_switch_mm(unsigned long pgd_phys, struct mm_struct *mm);
> > extern void cpu_reset(unsigned long addr) __attribute__((noreturn));
> > +extern void cpu_do_suspend(struct cpu_suspend_ctx *ptr);
> > +extern u64 cpu_do_resume(phys_addr_t ptr, u64 idmap_ttbr);
> >
> > #include <asm/memory.h>
> >
> > diff --git a/arch/arm64/mm/proc.S b/arch/arm64/mm/proc.S
> > index a82ae88..193bf98 100644
> > --- a/arch/arm64/mm/proc.S
> > +++ b/arch/arm64/mm/proc.S
> > @@ -80,6 +80,70 @@ ENTRY(cpu_do_idle)
> > ret
> > ENDPROC(cpu_do_idle)
> >
> > +#ifdef CONFIG_ARM_CPU_SUSPEND
> > +/**
> > + * cpu_do_suspend - save CPU registers context
> > + * x0: virtual address of context pointer
> > + */
> > +ENTRY(cpu_do_suspend)
> > + mrs x1, tpidr_el0
> > + str x1, [x0, #CPU_CTX_TPIDR_EL0]
> > + mrs x2, tpidrro_el0
> > + str x2, [x0, #CPU_CTX_TPIDRRO_EL0]
> > + mrs x3, contextidr_el1
> > + str x3, [x0, #CPU_CTX_CTXIDR_EL1]
> > + mrs x4, mair_el1
> > + str x4, [x0, #CPU_CTX_MAIR_EL1]
> > + mrs x5, cpacr_el1
> > + str x5, [x0, #CPU_CTX_CPACR_EL1]
> > + mrs x6, ttbr1_el1
> > + str x6, [x0, #CPU_CTX_TTBR1_EL1]
> > + mrs x7, tcr_el1
> > + str x7, [x0, #CPU_CTX_TCR_EL1]
> > + mrs x8, vbar_el1
> > + str x8, [x0, #CPU_CTX_VBAR_EL1]
> > + mrs x9, sctlr_el1
> > + str x9, [x0, #CPU_CTX_SCTLR_EL1]
> > + ret
> > +ENDPROC(cpu_do_suspend)
>
> Can you read all the registers a once and do some stp to save them?
Yes, absolutely. In a way the store pair will require an assumption on
the context structure layout - eg:
mrs x1, tpidr_el0
mrs x2, tpidrro_el0
stp x1, x2, [x0, #CPU_CTX_TPIDR_EL0]
implicitly assumes that the storage for TPIDR_EL0 and TPIDRRO_EL0 is
contiguous, we can't change the layout later, but I guess we can live
with that.
> > +
> > +/**
> > + * cpu_do_resume - registers layout should match the corresponding
> > + * cpu_do_suspend call
> > + *
> > + * x0: Physical address of context pointer
> > + * x1: Should contain the physical address of identity map page tables
> > + * used to turn on the MMU and complete context restore
> > + *
> > + * Returns:
> > + * sctlr value in x0
> > + */
> > +ENTRY(cpu_do_resume)
> > + tlbi vmalle1is // make sure tlb entries are invalid
> > + ldr x2, [x0, #CPU_CTX_TPIDR_EL0]
> > + msr tpidr_el0, x2
> > + ldr x3, [x0, #CPU_CTX_TPIDRRO_EL0]
> > + msr tpidrro_el0, x3
> > + ldr x4, [x0, #CPU_CTX_CTXIDR_EL1]
> > + msr contextidr_el1, x4
> > + ldr x5, [x0, #CPU_CTX_MAIR_EL1]
> > + msr mair_el1, x5
> > + ldr x6, [x0, #CPU_CTX_CPACR_EL1]
> > + msr cpacr_el1, x6
> > + msr ttbr0_el1, x1
> > + ldr x7, [x0, #CPU_CTX_TTBR1_EL1]
> > + msr ttbr1_el1, x7
> > + ldr x8, [x0, #CPU_CTX_TCR_EL1]
> > + msr tcr_el1, x8
> > + ldr x9, [x0, #CPU_CTX_VBAR_EL1]
> > + msr vbar_el1, x9
> > + ldr x0, [x0, #CPU_CTX_SCTLR_EL1]
> > + isb
> > + dsb sy
> > + ret
> > +ENDPROC(cpu_do_resume)
>
> Same here, use ldp.
>
> BTW, do we need the DSB here or just the ISB?
A dsb is required to ensure that the tlb invalidate has completed, I
think it is mandatory from an architectural standpoint but please
correct me if I am wrong.
Thanks for the review,
Lorenzo
More information about the linux-arm-kernel
mailing list