[PATCH v3 15/16] KVM: arm64: Drop perf.c and fold its tiny bits of code into arm.c / pmu.c

Marc Zyngier maz at kernel.org
Mon Oct 11 02:44:58 PDT 2021


On Wed, 22 Sep 2021 01:05:32 +0100,
Sean Christopherson <seanjc at google.com> wrote:
> 
> Call KVM's (un)register perf callbacks helpers directly from arm.c, and
> move the PMU bits into pmu.c and rename the related helper accordingly.
> 
> Signed-off-by: Sean Christopherson <seanjc at google.com>
> ---
>  arch/arm64/include/asm/kvm_host.h |  3 ---
>  arch/arm64/kvm/Makefile           |  2 +-
>  arch/arm64/kvm/arm.c              |  6 ++++--
>  arch/arm64/kvm/perf.c             | 27 ---------------------------
>  arch/arm64/kvm/pmu.c              |  8 ++++++++
>  include/kvm/arm_pmu.h             |  1 +
>  6 files changed, 14 insertions(+), 33 deletions(-)
>  delete mode 100644 arch/arm64/kvm/perf.c
> 
> diff --git a/arch/arm64/include/asm/kvm_host.h b/arch/arm64/include/asm/kvm_host.h
> index 828b6eaa2c56..f141ac65f4f1 100644
> --- a/arch/arm64/include/asm/kvm_host.h
> +++ b/arch/arm64/include/asm/kvm_host.h
> @@ -670,9 +670,6 @@ unsigned long kvm_mmio_read_buf(const void *buf, unsigned int len);
>  int kvm_handle_mmio_return(struct kvm_vcpu *vcpu);
>  int io_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa);
>  
> -void kvm_perf_init(void);
> -void kvm_perf_teardown(void);
> -
>  #ifdef CONFIG_GUEST_PERF_EVENTS
>  static inline bool kvm_arch_pmi_in_guest(struct kvm_vcpu *vcpu)
>  {
> diff --git a/arch/arm64/kvm/Makefile b/arch/arm64/kvm/Makefile
> index 989bb5dad2c8..0bcc378b7961 100644
> --- a/arch/arm64/kvm/Makefile
> +++ b/arch/arm64/kvm/Makefile
> @@ -12,7 +12,7 @@ obj-$(CONFIG_KVM) += hyp/
>  
>  kvm-y := $(KVM)/kvm_main.o $(KVM)/coalesced_mmio.o $(KVM)/eventfd.o \
>  	 $(KVM)/vfio.o $(KVM)/irqchip.o $(KVM)/binary_stats.o \
> -	 arm.o mmu.o mmio.o psci.o perf.o hypercalls.o pvtime.o \
> +	 arm.o mmu.o mmio.o psci.o hypercalls.o pvtime.o \
>  	 inject_fault.o va_layout.o handle_exit.o \
>  	 guest.o debug.o reset.o sys_regs.o \
>  	 vgic-sys-reg-v3.o fpsimd.o pmu.o \
> diff --git a/arch/arm64/kvm/arm.c b/arch/arm64/kvm/arm.c
> index 2b542fdc237e..48f89d80f464 100644
> --- a/arch/arm64/kvm/arm.c
> +++ b/arch/arm64/kvm/arm.c
> @@ -1744,7 +1744,9 @@ static int init_subsystems(void)
>  	if (err)
>  		goto out;
>  
> -	kvm_perf_init();
> +	kvm_pmu_init();
> +	kvm_register_perf_callbacks(NULL);
> +
>  	kvm_sys_reg_table_init();
>  
>  out:
> @@ -2160,7 +2162,7 @@ int kvm_arch_init(void *opaque)
>  /* NOP: Compiling as a module not supported */
>  void kvm_arch_exit(void)
>  {
> -	kvm_perf_teardown();
> +	kvm_unregister_perf_callbacks();
>  }
>  
>  static int __init early_kvm_mode_cfg(char *arg)
> diff --git a/arch/arm64/kvm/perf.c b/arch/arm64/kvm/perf.c
> deleted file mode 100644
> index 0b902e0d5b5d..000000000000
> --- a/arch/arm64/kvm/perf.c
> +++ /dev/null
> @@ -1,27 +0,0 @@
> -// SPDX-License-Identifier: GPL-2.0-only
> -/*
> - * Based on the x86 implementation.
> - *
> - * Copyright (C) 2012 ARM Ltd.
> - * Author: Marc Zyngier <marc.zyngier at arm.com>
> - */
> -
> -#include <linux/perf_event.h>
> -#include <linux/kvm_host.h>
> -
> -#include <asm/kvm_emulate.h>
> -
> -DEFINE_STATIC_KEY_FALSE(kvm_arm_pmu_available);
> -
> -void kvm_perf_init(void)
> -{
> -	if (kvm_pmu_probe_pmuver() != 0xf && !is_protected_kvm_enabled())
> -		static_branch_enable(&kvm_arm_pmu_available);
> -
> -	kvm_register_perf_callbacks(NULL);
> -}
> -
> -void kvm_perf_teardown(void)
> -{
> -	kvm_unregister_perf_callbacks();
> -}
> diff --git a/arch/arm64/kvm/pmu.c b/arch/arm64/kvm/pmu.c
> index 03a6c1f4a09a..d98b57a17043 100644
> --- a/arch/arm64/kvm/pmu.c
> +++ b/arch/arm64/kvm/pmu.c
> @@ -7,6 +7,14 @@
>  #include <linux/perf_event.h>
>  #include <asm/kvm_hyp.h>
>  
> +DEFINE_STATIC_KEY_FALSE(kvm_arm_pmu_available);
> +
> +void kvm_pmu_init(void)
> +{
> +	if (kvm_pmu_probe_pmuver() != 0xf && !is_protected_kvm_enabled())
> +		static_branch_enable(&kvm_arm_pmu_available);
> +}
> +
>  /*
>   * Given the perf event attributes and system type, determine
>   * if we are going to need to switch counters at guest entry/exit.
> diff --git a/include/kvm/arm_pmu.h b/include/kvm/arm_pmu.h
> index 864b9997efb2..42270676498d 100644
> --- a/include/kvm/arm_pmu.h
> +++ b/include/kvm/arm_pmu.h
> @@ -14,6 +14,7 @@
>  #define ARMV8_PMU_MAX_COUNTER_PAIRS	((ARMV8_PMU_MAX_COUNTERS + 1) >> 1)
>  
>  DECLARE_STATIC_KEY_FALSE(kvm_arm_pmu_available);
> +void kvm_pmu_init(void);
>  
>  static __always_inline bool kvm_arm_support_pmu_v3(void)
>  {

Note that this patch is now conflicting with e840f42a4992 ("KVM:
arm64: Fix PMU probe ordering"), which was merged in -rc4. Moving the
static key definition to arch/arm64/kvm/pmu-emul.c and getting rid of
kvm_pmu_init() altogether should be enough to resolve it.

With that,

Reviewed-by: Marc Zyngier <maz at kernel.org>

	M.

-- 
Without deviation from the norm, progress is not possible.



More information about the linux-riscv mailing list