[PATCH] arm64: perf: add support for percpu pmu interrupt
Vinayak Kale
vkale at apm.com
Tue Oct 15 02:33:25 EDT 2013
Hi Will,
Thanks for your comments. Please see inline below.
On Mon, Oct 14, 2013 at 6:04 PM, Will Deacon <will.deacon at arm.com> wrote:
> Hi Vinayak,
>
> On Mon, Oct 14, 2013 at 07:46:29AM +0100, Vinayak Kale wrote:
>> This patch adds support for irq registration when pmu interrupt type is PPI.
>> The patch also fixes ARMV8_EVTYPE_* macros since evtCount field width is
>> 10bits.
>>
>> Signed-off-by: Vinayak Kale <vkale at apm.com>
>> Signed-off-by: Tuan Phan <tphan at apm.com>
>> ---
>> arch/arm64/kernel/perf_event.c | 108 +++++++++++++++++++++++++++++-----------
>> 1 file changed, 78 insertions(+), 30 deletions(-)
>>
>> diff --git a/arch/arm64/kernel/perf_event.c b/arch/arm64/kernel/perf_event.c
>> index cea1594..ba3706d 100644
>> --- a/arch/arm64/kernel/perf_event.c
>> +++ b/arch/arm64/kernel/perf_event.c
>> @@ -363,22 +363,53 @@ validate_group(struct perf_event *event)
>> }
>
> [...]
>
>> static int
>> armpmu_reserve_hardware(struct arm_pmu *armpmu)
>> {
>> @@ -396,36 +427,53 @@ armpmu_reserve_hardware(struct arm_pmu *armpmu)
>> return -ENODEV;
>> }
>>
>> - for (i = 0; i < irqs; ++i) {
>> - err = 0;
>> - irq = platform_get_irq(pmu_device, i);
>> - if (irq < 0)
>> - continue;
>> + irq = platform_get_irq(pmu_device, 0);
>>
>> + if (irq >= 16 && irq <= 31) {
>
> This is horribly GIC specific and will break in the face of IRQ domains
> (since this is a virtual interrupt number).
>
hmm, yes. I had not considered irq domain mapping.
>> /*
>> - * If we have a single PMU interrupt that we can't shift,
>> - * assume that we're running on a uniprocessor machine and
>> - * continue. Otherwise, continue without this interrupt.
>> + * percpu PMU interrupt.
>> */
>> - if (irq_set_affinity(irq, cpumask_of(i)) && irqs > 1) {
>> - pr_warning("unable to set irq affinity (irq=%d, cpu=%u)\n",
>> - irq, i);
>> - continue;
>> - }
>> -
>> - err = request_irq(irq, armpmu->handle_irq,
>> - IRQF_NOBALANCING,
>> - "arm-pmu", armpmu);
>> + err = request_percpu_irq(irq, armpmu->handle_irq,
>> + "arm-pmu", armpmu);
>
> This is broken -- the dev_id *must* be a __percpu variable for percpu irqs.
>
Yes, this needs to be fixed.
>> if (err) {
>> pr_err("unable to request IRQ%d for ARM PMU counters\n",
>> - irq);
>> + irq);
>> armpmu_release_hardware(armpmu);
>> return err;
>> }
>>
>> - cpumask_set_cpu(i, &armpmu->active_irqs);
>> + on_each_cpu(armpmu_enable_percpu_irq, (void *)armpmu, 1);
>> + } else {
>> + for (i = 0; i < irqs; ++i) {
>> + err = 0;
>> + irq = platform_get_irq(pmu_device, i);
>> + if (irq < 0)
>> + continue;
>> +
>> + /*
>> + * If we have a single PMU interrupt that we can't shift,
>> + * assume that we're running on a uniprocessor machine and
>> + * continue. Otherwise, continue without this interrupt.
>> + */
>> + if (irq_set_affinity(irq, cpumask_of(i)) && irqs > 1) {
>> + pr_warning("unable to set irq affinity (irq=%d, cpu=%u)\n",
>> + irq, i);
>> + continue;
>> + }
>> +
>> + err = request_irq(irq, armpmu->handle_irq,
>> + IRQF_NOBALANCING,
>> + "arm-pmu", armpmu);
>
> A better way to do this is to try request_percpu_irq first. If that fails,
> then try request_irq. However, the error reporting out of request_percpu_irq
> could do with some cleanup (rather than just return -EINVAL) so we can
> detect the difference between `this interrupt isn't per-cpu' and `this
> per-cpu interrupt is invalid'. This can help us avoid the WARN_ON in
> request_irq when it is passed a per-cpu interrupt.
>
Trying request_percpu_irq first seems better. But if it fails then we
would straight away
assume it's not per-cpu interrupt and try request_irq. In this case we
may not be able to
detect 'this per-cpu interrupt is invalid' case.
>> @@ -784,8 +832,8 @@ static const unsigned armv8_pmuv3_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
>> /*
>> * PMXEVTYPER: Event selection reg
>> */
>> -#define ARMV8_EVTYPE_MASK 0xc80000ff /* Mask for writable bits */
>> -#define ARMV8_EVTYPE_EVENT 0xff /* Mask for EVENT bits */
>> +#define ARMV8_EVTYPE_MASK 0xc80003ff /* Mask for writable bits */
>> +#define ARMV8_EVTYPE_EVENT 0x3ff /* Mask for EVENT bits */
>>
>> /*
>> * Event filters for PMUv3
>> @@ -1175,7 +1223,7 @@ static void armv8pmu_reset(void *info)
>> static int armv8_pmuv3_map_event(struct perf_event *event)
>> {
>> return map_cpu_event(event, &armv8_pmuv3_perf_map,
>> - &armv8_pmuv3_perf_cache_map, 0xFF);
>> + &armv8_pmuv3_perf_cache_map, 0x3FF);
>> }
>
> What's all this?
>
The evtCount (event number) field width is 10bits in event selection register.
So need to fix ARMV8_EVTYPE_* macros and related mask value.
>From the subject of patch, one may think that the patch is specific
only to percpu irq changes (which is not true).
I had mentioned about fixing ARMV8_EVTYPE_* macros in patch description.
> Will
-Vinayak
More information about the linux-arm-kernel
mailing list