[PATCH 01/02] ARM: shmobile: Hook in MCPM and CCI to APMU code
Nicolas Pitre
nicolas.pitre at linaro.org
Thu Feb 27 23:08:59 EST 2014
On Thu, 27 Feb 2014, Magnus Damm wrote:
> From: Magnus Damm <damm at opensource.se>
>
> Add MCPM and CCI hooks to the shared APMU code to allow
> multicluster operation on r8a7790. Tested with SMP boot
> and CPU Hotplug on r8a7790 Lager.
>
> Signed-off-by: Magnus Damm <damm at opensource.se>
> ---
>
> arch/arm/mach-shmobile/headsmp.S | 7 +
> arch/arm/mach-shmobile/include/mach/common.h | 1
> arch/arm/mach-shmobile/platsmp-apmu.c | 111 ++++++++++++++++++++++++--
> arch/arm/mach-shmobile/platsmp.c | 4
> 4 files changed, 116 insertions(+), 7 deletions(-)
>
> --- 0001/arch/arm/mach-shmobile/headsmp.S
> +++ work/arch/arm/mach-shmobile/headsmp.S 2014-02-27 00:57:50.000000000 +0900
> @@ -19,6 +19,13 @@ ENTRY(shmobile_invalidate_start)
> b secondary_startup
> ENDPROC(shmobile_invalidate_start)
>
> +#ifdef CONFIG_MCPM
> +ENTRY(shmobile_invalidate_mcpm_entry)
> + bl v7_invalidate_l1
> + b mcpm_entry_point
> +ENDPROC(shmobile_invalidate_mcpm_entry)
> +#endif
You could get rid of this and shmobile_invalidate_start by simply adding
a shmobile specific entry in arch/arm/mm/proc-v7.S alongside the other
__v7_*_setup functions.
> +
> /*
> * Reset vector for secondary CPUs.
> * This will be mapped at address 0 by SBAR register.
> --- 0001/arch/arm/mach-shmobile/include/mach/common.h
> +++ work/arch/arm/mach-shmobile/include/mach/common.h 2014-02-27 00:57:50.000000000 +0900
> @@ -16,6 +16,7 @@ extern void shmobile_smp_hook(unsigned i
> unsigned long arg);
> extern int shmobile_smp_cpu_disable(unsigned int cpu);
> extern void shmobile_invalidate_start(void);
> +extern void shmobile_invalidate_mcpm_entry(void);
> extern void shmobile_boot_scu(vonid);
> extern void shmobile_smp_scu_prepare_cpus(unsigned int max_cpus);
> extern void shmobile_smp_scu_cpu_die(unsigned int cpu);
> --- 0001/arch/arm/mach-shmobile/platsmp-apmu.c
> +++ work/arch/arm/mach-shmobile/platsmp-apmu.c 2014-02-27 00:57:50.000000000 +0900
> @@ -7,22 +7,28 @@
> * it under the terms of the GNU General Public License version 2 as
> * published by the Free Software Foundation.
> */
> +#include <linux/arm-cci.h>
> #include <linux/delay.h>
> #include <linux/init.h>
> #include <linux/io.h>
> #include <linux/ioport.h>
> +#include <linux/of.h>
> #include <linux/of_address.h>
> #include <linux/smp.h>
> #include <asm/cacheflush.h>
> #include <asm/cp15.h>
> +#include <asm/mcpm.h>
> #include <asm/smp_plat.h>
> #include <mach/common.h>
>
> -static struct {
> +static struct apmu_cpu {
> void __iomem *iomem;
> int bit;
> } apmu_cpus[CONFIG_NR_CPUS];
>
> +#define MAX_NR_CLUSTERS 2
This is defined in mcpm.h already.
> +static struct apmu_cpu *apmu_clst2cpu[MAX_NR_CLUSTERS][CONFIG_NR_CPUS];
> +
> #define WUPCR_OFFS 0x10
> #define PSTR_OFFS 0x40
> #define CPUNCR_OFFS(n) (0x100 + (0x10 * (n)))
> @@ -69,14 +75,23 @@ static int apmu_wrap(int cpu, int (*fn)(
>
> static void apmu_init_cpu(struct resource *res, int cpu, int bit)
> {
> + u32 id;
> + int mcpm_cpu, mcpm_cluster;
> +
> if (apmu_cpus[cpu].iomem)
> return;
>
> apmu_cpus[cpu].iomem = ioremap_nocache(res->start, resource_size(res));
> apmu_cpus[cpu].bit = bit;
>
> - pr_debug("apmu ioremap %d %d 0x%08x 0x%08x\n", cpu, bit,
> - res->start, resource_size(res));
> + id = cpu_logical_map(cpu);
> + mcpm_cpu = MPIDR_AFFINITY_LEVEL(id, 0);
> + mcpm_cluster = MPIDR_AFFINITY_LEVEL(id, 1);
You should consider validating those values making sure they're lower
than CONFIG_NR_CPUS and MAX_NR_CLUSTERS respectively. And declaring
them as unsigned.
> +
> + pr_debug("apmu ioremap %d %d %pr %d %d\n",
> + cpu, bit, res, mcpm_cluster, mcpm_cpu);
> +
> + apmu_clst2cpu[mcpm_cluster][mcpm_cpu] = &apmu_cpus[cpu];
> }
>
> static struct {
> @@ -93,7 +108,8 @@ static struct {
> }
> };
>
> -static void apmu_parse_cfg(void (*fn)(struct resource *res, int cpu, int bit))
> +static void apmu_parse_cfg(void (*fn)(struct resource *res, int cpu, int bit),
> + bool allow_multicluster)
> {
> u32 id;
> int k;
> @@ -110,7 +126,8 @@ static void apmu_parse_cfg(void (*fn)(st
> is_allowed = true;
> }
> }
> - if (!is_allowed)
> +
> + if (!allow_multicluster && !is_allowed)
> continue;
>
> for (bit = 0; bit < ARRAY_SIZE(apmu_config[k].cpus); bit++) {
> @@ -124,14 +141,19 @@ static void apmu_parse_cfg(void (*fn)(st
> }
> }
>
> +static int __init shmobile_smp_apmu_cci_init(void);
> +
> void __init shmobile_smp_apmu_prepare_cpus(unsigned int max_cpus)
> {
> /* install boot code shared by all CPUs */
> shmobile_boot_fn = virt_to_phys(shmobile_smp_boot);
> shmobile_boot_arg = MPIDR_HWID_BITMASK;
>
> - /* perform per-cpu setup */
> - apmu_parse_cfg(apmu_init_cpu);
> + /* allow multi-cluster operation in case CCI is detected */
> + if (IS_ENABLED(CONFIG_ARM_CCI) && !shmobile_smp_apmu_cci_init())
> + apmu_parse_cfg(apmu_init_cpu, true);
> + else
> + apmu_parse_cfg(apmu_init_cpu, false);
> }
>
> int shmobile_smp_apmu_boot_secondary(unsigned int cpu, struct task_struct *idle)
> @@ -192,4 +214,79 @@ int shmobile_smp_apmu_cpu_kill(unsigned
> {
> return apmu_wrap(cpu, apmu_power_off_poll);
> }
> +#else
> +#define shmobile_smp_apmu_cpu_die() shmobile_smp_sleep()
> +static inline int shmobile_smp_apmu_cpu_kill(void) { return -ENOTSUPP; }
> #endif
> +
> +#if defined(CONFIG_MCPM) && defined(CONFIG_ARM_CCI)
> +static void __naked apmu_power_up_setup(unsigned int affinity_level)
> +{
> + asm volatile ("cmp r0, #1\n"
> + "bxne lr\n"
> + "b cci_enable_port_for_self ");
> +}
> +
> +static int apmu_power_up(unsigned int cpu, unsigned int cluster)
> +{
> + struct apmu_cpu *ac = apmu_clst2cpu[cluster][cpu];
> + if (!ac)
> + return -EINVAL;
> +
> + shmobile_smp_hook(ac - &apmu_cpus[0],
> + virt_to_phys(shmobile_invalidate_mcpm_entry), 0);
> +
> + return apmu_wrap(ac - &apmu_cpus[0], apmu_power_on);
> +}
It is possible for the power_up and power_down methods to be unordered
or called concurrently in some circumstances (please see comment inside
mcpm_cpu_power_down()). You should guard against that possibility with
a critical region and a count. Eventually we might move this
concurrency handling into the core but it is still up to backend code to
handle it for now until there is enough backend implementations to
factor away common issues. I'd suggest having a look at
arch/arm/mach-vexpress/dcscb.c for an example implementation.
> +static void apmu_power_down(void)
> +{
> + shmobile_smp_apmu_cpu_die(smp_processor_id());
> +}
This completely fails to handle the MCPM cluster race avoidance
protocol.
You should have a look at Documentation/arm/cluster-pm-race-avoidance.txt
and also arch/arm/mach-vexpress/dcscb.c again.
In particular, you need proper calls to __mcpm_cpu_going_down(),
__mcpm_outbound_enter_critical(), __mcpm_outbound_leave_critical() and
__mcpm_cpu_down(). The "last man" protocol must be used at least for
the disabling of the CCI port for the cluster becoming completely
unused, and ideally removing power to the whole cluster as well in that
case. Without the above calls, there is simply no benefits calling into
mcpm_entry_point as it is basically a huge passthrough.
Nicolas
More information about the linux-arm-kernel
mailing list