[PATCH v7 01/19] ARM64: Move PMU register related defines to asm/pmu.h
Marc Zyngier
marc.zyngier at arm.com
Tue Dec 15 03:34:14 PST 2015
On 15/12/15 08:49, Shannon Zhao wrote:
> From: Shannon Zhao <shannon.zhao at linaro.org>
>
> To use the ARMv8 PMU related register defines from the KVM code,
> we move the relevant definitions to asm/pmu.h header file.
>
> Signed-off-by: Anup Patel <anup.patel at linaro.org>
> Signed-off-by: Shannon Zhao <shannon.zhao at linaro.org>
> ---
> arch/arm64/include/asm/pmu.h | 64 ++++++++++++++++++++++++++++++++++++++++++
> arch/arm64/kernel/perf_event.c | 36 +-----------------------
> 2 files changed, 65 insertions(+), 35 deletions(-)
> create mode 100644 arch/arm64/include/asm/pmu.h
>
> diff --git a/arch/arm64/include/asm/pmu.h b/arch/arm64/include/asm/pmu.h
> new file mode 100644
> index 0000000..4264ea0
> --- /dev/null
> +++ b/arch/arm64/include/asm/pmu.h
> @@ -0,0 +1,64 @@
> +/*
> + * Copyright (C) 2015 Linaro Ltd, Shannon Zhao
Erm, not quite. You're simply moving existing code from one file to
another. That doesn't change the copyright of said code, which reads:
* PMU support
*
* Copyright (C) 2012 ARM Limited
* Author: Will Deacon <will.deacon at arm.com>
Please keep this mention in place.
> + *
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License version 2 as
> + * published by the Free Software Foundation.
> + *
> + * This program is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
> + * GNU General Public License for more details.
> + *
> + * You should have received a copy of the GNU General Public License
> + * along with this program. If not, see <http://www.gnu.org/licenses/>.
> + */
> +#ifndef __ASM_PMU_H
> +#define __ASM_PMU_H
> +
> +#define ARMV8_MAX_COUNTERS 32
> +#define ARMV8_COUNTER_MASK (ARMV8_MAX_COUNTERS - 1)
> +
> +/*
> + * Per-CPU PMCR: config reg
> + */
> +#define ARMV8_PMCR_E (1 << 0) /* Enable all counters */
> +#define ARMV8_PMCR_P (1 << 1) /* Reset all counters */
> +#define ARMV8_PMCR_C (1 << 2) /* Cycle counter reset */
> +#define ARMV8_PMCR_D (1 << 3) /* CCNT counts every 64th cpu cycle */
> +#define ARMV8_PMCR_X (1 << 4) /* Export to ETM */
> +#define ARMV8_PMCR_DP (1 << 5) /* Disable CCNT if non-invasive debug*/
> +#define ARMV8_PMCR_N_SHIFT 11 /* Number of counters supported */
> +#define ARMV8_PMCR_N_MASK 0x1f
> +#define ARMV8_PMCR_MASK 0x3f /* Mask for writable bits */
> +
> +/*
> + * PMCNTEN: counters enable reg
> + */
> +#define ARMV8_CNTEN_MASK 0xffffffff /* Mask for writable bits */
> +
> +/*
> + * PMINTEN: counters interrupt enable reg
> + */
> +#define ARMV8_INTEN_MASK 0xffffffff /* Mask for writable bits */
> +
> +/*
> + * PMOVSR: counters overflow flag status reg
> + */
> +#define ARMV8_OVSR_MASK 0xffffffff /* Mask for writable bits */
> +#define ARMV8_OVERFLOWED_MASK ARMV8_OVSR_MASK
> +
> +/*
> + * PMXEVTYPER: Event selection reg
> + */
> +#define ARMV8_EVTYPE_MASK 0xc80003ff /* Mask for writable bits */
> +#define ARMV8_EVTYPE_EVENT 0x3ff /* Mask for EVENT bits */
> +
> +/*
> + * Event filters for PMUv3
> + */
> +#define ARMV8_EXCLUDE_EL1 (1 << 31)
> +#define ARMV8_EXCLUDE_EL0 (1 << 30)
> +#define ARMV8_INCLUDE_EL2 (1 << 27)
> +
> +#endif /* __ASM_PMU_H */
> diff --git a/arch/arm64/kernel/perf_event.c b/arch/arm64/kernel/perf_event.c
> index 5b1897e..7eca5dc 100644
> --- a/arch/arm64/kernel/perf_event.c
> +++ b/arch/arm64/kernel/perf_event.c
> @@ -24,6 +24,7 @@
> #include <linux/of.h>
> #include <linux/perf/arm_pmu.h>
> #include <linux/platform_device.h>
> +#include <asm/pmu.h>
>
> /*
> * ARMv8 PMUv3 Performance Events handling code.
> @@ -187,9 +188,6 @@ static const unsigned armv8_a57_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
> #define ARMV8_IDX_COUNTER_LAST(cpu_pmu) \
> (ARMV8_IDX_CYCLE_COUNTER + cpu_pmu->num_events - 1)
>
> -#define ARMV8_MAX_COUNTERS 32
> -#define ARMV8_COUNTER_MASK (ARMV8_MAX_COUNTERS - 1)
> -
> /*
> * ARMv8 low level PMU access
> */
> @@ -200,38 +198,6 @@ static const unsigned armv8_a57_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
> #define ARMV8_IDX_TO_COUNTER(x) \
> (((x) - ARMV8_IDX_COUNTER0) & ARMV8_COUNTER_MASK)
>
> -/*
> - * Per-CPU PMCR: config reg
> - */
> -#define ARMV8_PMCR_E (1 << 0) /* Enable all counters */
> -#define ARMV8_PMCR_P (1 << 1) /* Reset all counters */
> -#define ARMV8_PMCR_C (1 << 2) /* Cycle counter reset */
> -#define ARMV8_PMCR_D (1 << 3) /* CCNT counts every 64th cpu cycle */
> -#define ARMV8_PMCR_X (1 << 4) /* Export to ETM */
> -#define ARMV8_PMCR_DP (1 << 5) /* Disable CCNT if non-invasive debug*/
> -#define ARMV8_PMCR_N_SHIFT 11 /* Number of counters supported */
> -#define ARMV8_PMCR_N_MASK 0x1f
> -#define ARMV8_PMCR_MASK 0x3f /* Mask for writable bits */
> -
> -/*
> - * PMOVSR: counters overflow flag status reg
> - */
> -#define ARMV8_OVSR_MASK 0xffffffff /* Mask for writable bits */
> -#define ARMV8_OVERFLOWED_MASK ARMV8_OVSR_MASK
> -
> -/*
> - * PMXEVTYPER: Event selection reg
> - */
> -#define ARMV8_EVTYPE_MASK 0xc80003ff /* Mask for writable bits */
> -#define ARMV8_EVTYPE_EVENT 0x3ff /* Mask for EVENT bits */
> -
> -/*
> - * Event filters for PMUv3
> - */
> -#define ARMV8_EXCLUDE_EL1 (1 << 31)
> -#define ARMV8_EXCLUDE_EL0 (1 << 30)
> -#define ARMV8_INCLUDE_EL2 (1 << 27)
> -
> static inline u32 armv8pmu_pmcr_read(void)
> {
> u32 val;
>
Thanks,
M.
--
Jazz is not dead. It just smells funny...
More information about the linux-arm-kernel
mailing list