[PATCH 7/8] RISC-V: KVM: Implement get event info function
Atish Kumar Patra
atishp at rivosinc.com
Mon Dec 9 17:16:29 PST 2024
On Mon, Dec 2, 2024 at 3:02 PM Samuel Holland <samuel.holland at sifive.com> wrote:
>
> Hi Atish,
>
> On 2024-11-19 2:29 PM, Atish Patra wrote:
> > The new get_event_info funciton allows the guest to query the presence
> > of multiple events with single SBI call. Currently, the perf driver
> > in linux guest invokes it for all the standard SBI PMU events. Support
> > the SBI function implementation in KVM as well.
> >
> > Signed-off-by: Atish Patra <atishp at rivosinc.com>
> > ---
> > arch/riscv/include/asm/kvm_vcpu_pmu.h | 3 ++
> > arch/riscv/kvm/vcpu_pmu.c | 67 +++++++++++++++++++++++++++++++++++
> > arch/riscv/kvm/vcpu_sbi_pmu.c | 3 ++
> > 3 files changed, 73 insertions(+)
> >
> > diff --git a/arch/riscv/include/asm/kvm_vcpu_pmu.h b/arch/riscv/include/asm/kvm_vcpu_pmu.h
> > index 1d85b6617508..9a930afc8f57 100644
> > --- a/arch/riscv/include/asm/kvm_vcpu_pmu.h
> > +++ b/arch/riscv/include/asm/kvm_vcpu_pmu.h
> > @@ -98,6 +98,9 @@ void kvm_riscv_vcpu_pmu_init(struct kvm_vcpu *vcpu);
> > int kvm_riscv_vcpu_pmu_snapshot_set_shmem(struct kvm_vcpu *vcpu, unsigned long saddr_low,
> > unsigned long saddr_high, unsigned long flags,
> > struct kvm_vcpu_sbi_return *retdata);
> > +int kvm_riscv_vcpu_pmu_event_info(struct kvm_vcpu *vcpu, unsigned long saddr_low,
> > + unsigned long saddr_high, unsigned long num_events,
> > + unsigned long flags, struct kvm_vcpu_sbi_return *retdata);
> > void kvm_riscv_vcpu_pmu_deinit(struct kvm_vcpu *vcpu);
> > void kvm_riscv_vcpu_pmu_reset(struct kvm_vcpu *vcpu);
> >
> > diff --git a/arch/riscv/kvm/vcpu_pmu.c b/arch/riscv/kvm/vcpu_pmu.c
> > index efd66835c2b8..a30f7ec31479 100644
> > --- a/arch/riscv/kvm/vcpu_pmu.c
> > +++ b/arch/riscv/kvm/vcpu_pmu.c
> > @@ -456,6 +456,73 @@ int kvm_riscv_vcpu_pmu_snapshot_set_shmem(struct kvm_vcpu *vcpu, unsigned long s
> > return 0;
> > }
> >
> > +int kvm_riscv_vcpu_pmu_event_info(struct kvm_vcpu *vcpu, unsigned long saddr_low,
> > + unsigned long saddr_high, unsigned long num_events,
> > + unsigned long flags, struct kvm_vcpu_sbi_return *retdata)
> > +{
> > + unsigned long hva;
> > + struct riscv_pmu_event_info *einfo;
> > + int shmem_size = num_events * sizeof(*einfo);
> > + bool writable;
> > + gpa_t shmem;
> > + u32 eidx, etype;
> > + u64 econfig;
> > + int ret;
> > +
> > + if (flags != 0 || (saddr_low & (SZ_16 - 1))) {
> > + ret = SBI_ERR_INVALID_PARAM;
> > + goto out;
> > + }
> > +
> > + shmem = saddr_low;
> > + if (saddr_high != 0) {
> > + if (IS_ENABLED(CONFIG_32BIT)) {
> > + shmem |= ((gpa_t)saddr_high << 32);
> > + } else {
> > + ret = SBI_ERR_INVALID_ADDRESS;
> > + goto out;
> > + }
> > + }
> > +
> > + hva = kvm_vcpu_gfn_to_hva_prot(vcpu, shmem >> PAGE_SHIFT, &writable);
> > + if (kvm_is_error_hva(hva) || !writable) {
> > + ret = SBI_ERR_INVALID_ADDRESS;
>
> This only checks the first page if the address crosses a page boundary. Maybe
> that is okay since kvm_vcpu_read_guest()/kvm_vcpu_write_guest() will fail if a
> later page is inaccessible?
>
That's my assumption as well.
We can invoke kvm_vcpu_gfn_to_hva_prot in a loop to validate all the
pages starting with shmem though.
The difference would be the error code returned
(SBI_ERR_INVALID_ADDRESS vs SBI_ERR_FAILURE)
which the specification allows anyways.
Let me know if you think KVM must validate the entire range. I can add in v2.
> > + goto out;
> > + }
> > +
> > + einfo = kzalloc(shmem_size, GFP_KERNEL);
> > + if (!einfo)
> > + return -ENOMEM;
> > +
> > + ret = kvm_vcpu_read_guest(vcpu, shmem, einfo, shmem_size);
> > + if (ret) {
> > + ret = SBI_ERR_FAILURE;
> > + goto free_mem;
> > + }
> > +
> > + for (int i = 0; i < num_events; i++) {
> > + eidx = einfo[i].event_idx;
> > + etype = kvm_pmu_get_perf_event_type(eidx);
> > + econfig = kvm_pmu_get_perf_event_config(eidx, einfo[i].event_data);
> > + ret = riscv_pmu_get_event_info(etype, econfig, NULL);
> > + if (ret > 0)
> > + einfo[i].output = 1;
>
> This also needs to write `output` in the else case to indicate that the event is
> not supported; the spec does not require the caller to initialize bit 0 of
> output to zero.
>
Sure. Added.
> Regards,
> Samuel
>
> > + }
> > +
> > + kvm_vcpu_write_guest(vcpu, shmem, einfo, shmem_size);
> > + if (ret) {
> > + ret = SBI_ERR_FAILURE;
> > + goto free_mem;
> > + }
> > +
> > +free_mem:
> > + kfree(einfo);
> > +out:
> > + retdata->err_val = ret;
> > +
> > + return 0;
> > +}
> > +
> > int kvm_riscv_vcpu_pmu_num_ctrs(struct kvm_vcpu *vcpu,
> > struct kvm_vcpu_sbi_return *retdata)
> > {
> > diff --git a/arch/riscv/kvm/vcpu_sbi_pmu.c b/arch/riscv/kvm/vcpu_sbi_pmu.c
> > index e4be34e03e83..a020d979d179 100644
> > --- a/arch/riscv/kvm/vcpu_sbi_pmu.c
> > +++ b/arch/riscv/kvm/vcpu_sbi_pmu.c
> > @@ -73,6 +73,9 @@ static int kvm_sbi_ext_pmu_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
> > case SBI_EXT_PMU_SNAPSHOT_SET_SHMEM:
> > ret = kvm_riscv_vcpu_pmu_snapshot_set_shmem(vcpu, cp->a0, cp->a1, cp->a2, retdata);
> > break;
> > + case SBI_EXT_PMU_EVENT_GET_INFO:
> > + ret = kvm_riscv_vcpu_pmu_event_info(vcpu, cp->a0, cp->a1, cp->a2, cp->a3, retdata);
> > + break;
> > default:
> > retdata->err_val = SBI_ERR_NOT_SUPPORTED;
> > }
> >
>
More information about the kvm-riscv
mailing list