[PATCH v2 2/8] KVM: arm64: selftest: Split arch_timer test code
Haibo Xu
xiaobo55x at gmail.com
Tue Sep 5 20:44:26 PDT 2023
On Wed, Sep 6, 2023 at 10:14 AM Haibo Xu <xiaobo55x at gmail.com> wrote:
>
> On Mon, Sep 4, 2023 at 9:24 PM Andrew Jones <ajones at ventanamicro.com> wrote:
> >
> > On Sat, Sep 02, 2023 at 08:59:24PM +0800, Haibo Xu wrote:
> > > Split the arch-neutral test code out of aarch64/arch_timer.c
> > > and put them into a common arch_timer.c. This is a preparation
> > > to share timer test codes in riscv.
> > >
> > > Suggested-by: Andrew Jones <ajones at ventanamicro.com>
> > > Signed-off-by: Haibo Xu <haibo1.xu at intel.com>
> > > ---
> > > tools/testing/selftests/kvm/Makefile | 9 +-
> > > .../selftests/kvm/aarch64/arch_timer.c | 288 +-----------------
> > > tools/testing/selftests/kvm/arch_timer.c | 252 +++++++++++++++
> > > .../selftests/kvm/include/timer_test.h | 52 ++++
> > > 4 files changed, 317 insertions(+), 284 deletions(-)
> > > create mode 100644 tools/testing/selftests/kvm/arch_timer.c
> > > create mode 100644 tools/testing/selftests/kvm/include/timer_test.h
> > >
> > > diff --git a/tools/testing/selftests/kvm/Makefile b/tools/testing/selftests/kvm/Makefile
> > > index 0b9c42fbce8c..fb8904e2c06a 100644
> > > --- a/tools/testing/selftests/kvm/Makefile
> > > +++ b/tools/testing/selftests/kvm/Makefile
> > > @@ -140,7 +140,6 @@ TEST_GEN_PROGS_x86_64 += system_counter_offset_test
> > > TEST_GEN_PROGS_EXTENDED_x86_64 += x86_64/nx_huge_pages_test
> > >
> > > TEST_GEN_PROGS_aarch64 += aarch64/aarch32_id_regs
> > > -TEST_GEN_PROGS_aarch64 += aarch64/arch_timer
> > > TEST_GEN_PROGS_aarch64 += aarch64/debug-exceptions
> > > TEST_GEN_PROGS_aarch64 += aarch64/hypercalls
> > > TEST_GEN_PROGS_aarch64 += aarch64/page_fault_test
> > > @@ -150,6 +149,7 @@ TEST_GEN_PROGS_aarch64 += aarch64/vcpu_width_config
> > > TEST_GEN_PROGS_aarch64 += aarch64/vgic_init
> > > TEST_GEN_PROGS_aarch64 += aarch64/vgic_irq
> > > TEST_GEN_PROGS_aarch64 += access_tracking_perf_test
> > > +TEST_GEN_PROGS_aarch64 += arch_timer
> > > TEST_GEN_PROGS_aarch64 += demand_paging_test
> > > TEST_GEN_PROGS_aarch64 += dirty_log_test
> > > TEST_GEN_PROGS_aarch64 += dirty_log_perf_test
> > > @@ -188,6 +188,7 @@ TEST_GEN_PROGS_riscv += set_memory_region_test
> > > TEST_GEN_PROGS_riscv += kvm_binary_stats_test
> > >
> > > SPLIT_TESTS += get-reg-list
> > > +SPLIT_TESTS += arch_timer
> > >
> > > TEST_PROGS += $(TEST_PROGS_$(ARCH_DIR))
> > > TEST_GEN_PROGS += $(TEST_GEN_PROGS_$(ARCH_DIR))
> > > @@ -248,13 +249,10 @@ TEST_DEP_FILES += $(patsubst %.o, %.d, $(SPLIT_TESTS_OBJS))
> > > -include $(TEST_DEP_FILES)
> > >
> > > $(TEST_GEN_PROGS) $(TEST_GEN_PROGS_EXTENDED): %: %.o
> > > - $(CC) $(CFLAGS) $(CPPFLAGS) $(LDFLAGS) $(TARGET_ARCH) $< $(LIBKVM_OBJS) $(LDLIBS) -o $@
> > > + $(CC) $(CFLAGS) $(CPPFLAGS) $(LDFLAGS) $(TARGET_ARCH) $^ $(LDLIBS) -o $@
> > > $(TEST_GEN_OBJ): $(OUTPUT)/%.o: %.c
> > > $(CC) $(CFLAGS) $(CPPFLAGS) $(TARGET_ARCH) -c $< -o $@
> > >
> > > -$(SPLIT_TESTS_TARGETS): %: %.o $(SPLIT_TESTS_OBJS)
> > > - $(CC) $(CFLAGS) $(CPPFLAGS) $(LDFLAGS) $(TARGET_ARCH) $^ $(LDLIBS) -o $@
> > > -
> > > EXTRA_CLEAN += $(LIBKVM_OBJS) $(TEST_DEP_FILES) $(TEST_GEN_OBJ) $(SPLIT_TESTS_OBJS) cscope.*
> > >
> > > x := $(shell mkdir -p $(sort $(dir $(LIBKVM_C_OBJ) $(LIBKVM_S_OBJ))))
> > > @@ -273,6 +271,7 @@ $(LIBKVM_STRING_OBJ): $(OUTPUT)/%.o: %.c
> > > x := $(shell mkdir -p $(sort $(dir $(TEST_GEN_PROGS))))
> > > $(TEST_GEN_PROGS): $(LIBKVM_OBJS)
> > > $(TEST_GEN_PROGS_EXTENDED): $(LIBKVM_OBJS)
> > > +$(SPLIT_TESTS_TARGETS): $(OUTPUT)/%: $(ARCH_DIR)/%.o
> >
> > The improvements to the Makefile to avoid SPLIT_TESTS_TARGETS needing its
> > own $(CC) line should preferably be done in a separate, preliminary patch.
> >
>
> Yes, Will move the change to a separate patch.
>
> > >
> > > cscope: include_paths = $(LINUX_TOOL_INCLUDE) $(LINUX_HDR_PATH) include lib ..
> > > cscope:
> > > diff --git a/tools/testing/selftests/kvm/aarch64/arch_timer.c b/tools/testing/selftests/kvm/aarch64/arch_timer.c
> > > index b63859829a96..ceb649548751 100644
> > > --- a/tools/testing/selftests/kvm/aarch64/arch_timer.c
> > > +++ b/tools/testing/selftests/kvm/aarch64/arch_timer.c
> > > @@ -1,91 +1,25 @@
> > > // SPDX-License-Identifier: GPL-2.0-only
> > > /*
> > > - * arch_timer.c - Tests the aarch64 timer IRQ functionality
> > > - *
> > > * The test validates both the virtual and physical timer IRQs using
> > > - * CVAL and TVAL registers. This consitutes the four stages in the test.
> > > - * The guest's main thread configures the timer interrupt for a stage
> > > - * and waits for it to fire, with a timeout equal to the timer period.
> > > - * It asserts that the timeout doesn't exceed the timer period.
> > > - *
> > > - * On the other hand, upon receipt of an interrupt, the guest's interrupt
> > > - * handler validates the interrupt by checking if the architectural state
> > > - * is in compliance with the specifications.
> > > - *
> > > - * The test provides command-line options to configure the timer's
> > > - * period (-p), number of vCPUs (-n), and iterations per stage (-i).
> > > - * To stress-test the timer stack even more, an option to migrate the
> > > - * vCPUs across pCPUs (-m), at a particular rate, is also provided.
> > > + * CVAL and TVAL registers.
> > > *
> > > * Copyright (c) 2021, Google LLC.
> > > */
> > > #define _GNU_SOURCE
> > >
> > > -#include <stdlib.h>
> > > -#include <pthread.h>
> > > -#include <linux/kvm.h>
> > > -#include <linux/sizes.h>
> > > -#include <linux/bitmap.h>
> > > -#include <sys/sysinfo.h>
> > > -
> > > -#include "kvm_util.h"
> > > -#include "processor.h"
> > > -#include "delay.h"
> > > #include "arch_timer.h"
> > > +#include "delay.h"
> > > #include "gic.h"
> > > +#include "processor.h"
> > > +#include "timer_test.h"
> > > #include "vgic.h"
> > >
> > > -#define NR_VCPUS_DEF 4
> > > -#define NR_TEST_ITERS_DEF 5
> > > -#define TIMER_TEST_PERIOD_MS_DEF 10
> > > -#define TIMER_TEST_ERR_MARGIN_US 100
> > > -#define TIMER_TEST_MIGRATION_FREQ_MS 2
> > > -
> > > -struct test_args {
> > > - int nr_vcpus;
> > > - int nr_iter;
> > > - int timer_period_ms;
> > > - int migration_freq_ms;
> > > - struct kvm_arm_counter_offset offset;
> > > -};
> > > -
> > > -static struct test_args test_args = {
> > > - .nr_vcpus = NR_VCPUS_DEF,
> > > - .nr_iter = NR_TEST_ITERS_DEF,
> > > - .timer_period_ms = TIMER_TEST_PERIOD_MS_DEF,
> > > - .migration_freq_ms = TIMER_TEST_MIGRATION_FREQ_MS,
> > > - .offset = { .reserved = 1 },
> > > -};
> > > -
> > > -#define msecs_to_usecs(msec) ((msec) * 1000LL)
> > > -
> > > -#define GICD_BASE_GPA 0x8000000ULL
> > > -#define GICR_BASE_GPA 0x80A0000ULL
> > > -
> > > -enum guest_stage {
> > > - GUEST_STAGE_VTIMER_CVAL = 1,
> > > - GUEST_STAGE_VTIMER_TVAL,
> > > - GUEST_STAGE_PTIMER_CVAL,
> > > - GUEST_STAGE_PTIMER_TVAL,
> > > - GUEST_STAGE_MAX,
> > > -};
> > > -
> > > -/* Shared variables between host and guest */
> > > -struct test_vcpu_shared_data {
> > > - int nr_iter;
> > > - enum guest_stage guest_stage;
> > > - uint64_t xcnt;
> > > -};
> > > -
> > > -static struct kvm_vcpu *vcpus[KVM_MAX_VCPUS];
> > > -static pthread_t pt_vcpu_run[KVM_MAX_VCPUS];
> > > -static struct test_vcpu_shared_data vcpu_shared_data[KVM_MAX_VCPUS];
> > > +extern struct test_args test_args;
> > > +extern struct kvm_vcpu *vcpus[];
> > > +extern struct test_vcpu_shared_data vcpu_shared_data[];
> > >
> > > static int vtimer_irq, ptimer_irq;
> > >
> > > -static unsigned long *vcpu_done_map;
> > > -static pthread_mutex_t vcpu_done_map_lock;
> > > -
> > > static void
> > > guest_configure_timer_action(struct test_vcpu_shared_data *shared_data)
> > > {
> > > @@ -222,137 +156,6 @@ static void guest_code(void)
> > > GUEST_DONE();
> > > }
> > >
> > > -static void *test_vcpu_run(void *arg)
> > > -{
> > > - unsigned int vcpu_idx = (unsigned long)arg;
> > > - struct ucall uc;
> > > - struct kvm_vcpu *vcpu = vcpus[vcpu_idx];
> > > - struct kvm_vm *vm = vcpu->vm;
> > > - struct test_vcpu_shared_data *shared_data = &vcpu_shared_data[vcpu_idx];
> > > -
> > > - vcpu_run(vcpu);
> > > -
> > > - /* Currently, any exit from guest is an indication of completion */
> > > - pthread_mutex_lock(&vcpu_done_map_lock);
> > > - __set_bit(vcpu_idx, vcpu_done_map);
> > > - pthread_mutex_unlock(&vcpu_done_map_lock);
> > > -
> > > - switch (get_ucall(vcpu, &uc)) {
> > > - case UCALL_SYNC:
> > > - case UCALL_DONE:
> > > - break;
> > > - case UCALL_ABORT:
> > > - sync_global_from_guest(vm, *shared_data);
> > > - fprintf(stderr, "Guest assert failed, vcpu %u; stage; %u; iter: %u\n",
> > > - vcpu_idx, shared_data->guest_stage, shared_data->nr_iter);
> > > - REPORT_GUEST_ASSERT(uc);
> > > - break;
> > > - default:
> > > - TEST_FAIL("Unexpected guest exit\n");
> > > - }
> > > -
> > > - return NULL;
> > > -}
> > > -
> > > -static uint32_t test_get_pcpu(void)
> > > -{
> > > - uint32_t pcpu;
> > > - unsigned int nproc_conf;
> > > - cpu_set_t online_cpuset;
> > > -
> > > - nproc_conf = get_nprocs_conf();
> > > - sched_getaffinity(0, sizeof(cpu_set_t), &online_cpuset);
> > > -
> > > - /* Randomly find an available pCPU to place a vCPU on */
> > > - do {
> > > - pcpu = rand() % nproc_conf;
> > > - } while (!CPU_ISSET(pcpu, &online_cpuset));
> > > -
> > > - return pcpu;
> > > -}
> > > -
> > > -static int test_migrate_vcpu(unsigned int vcpu_idx)
> > > -{
> > > - int ret;
> > > - cpu_set_t cpuset;
> > > - uint32_t new_pcpu = test_get_pcpu();
> > > -
> > > - CPU_ZERO(&cpuset);
> > > - CPU_SET(new_pcpu, &cpuset);
> > > -
> > > - pr_debug("Migrating vCPU: %u to pCPU: %u\n", vcpu_idx, new_pcpu);
> > > -
> > > - ret = pthread_setaffinity_np(pt_vcpu_run[vcpu_idx],
> > > - sizeof(cpuset), &cpuset);
> > > -
> > > - /* Allow the error where the vCPU thread is already finished */
> > > - TEST_ASSERT(ret == 0 || ret == ESRCH,
> > > - "Failed to migrate the vCPU:%u to pCPU: %u; ret: %d\n",
> > > - vcpu_idx, new_pcpu, ret);
> > > -
> > > - return ret;
> > > -}
> > > -
> > > -static void *test_vcpu_migration(void *arg)
> > > -{
> > > - unsigned int i, n_done;
> > > - bool vcpu_done;
> > > -
> > > - do {
> > > - usleep(msecs_to_usecs(test_args.migration_freq_ms));
> > > -
> > > - for (n_done = 0, i = 0; i < test_args.nr_vcpus; i++) {
> > > - pthread_mutex_lock(&vcpu_done_map_lock);
> > > - vcpu_done = test_bit(i, vcpu_done_map);
> > > - pthread_mutex_unlock(&vcpu_done_map_lock);
> > > -
> > > - if (vcpu_done) {
> > > - n_done++;
> > > - continue;
> > > - }
> > > -
> > > - test_migrate_vcpu(i);
> > > - }
> > > - } while (test_args.nr_vcpus != n_done);
> > > -
> > > - return NULL;
> > > -}
> > > -
> > > -static void test_run(struct kvm_vm *vm)
> > > -{
> > > - pthread_t pt_vcpu_migration;
> > > - unsigned int i;
> > > - int ret;
> > > -
> > > - pthread_mutex_init(&vcpu_done_map_lock, NULL);
> > > - vcpu_done_map = bitmap_zalloc(test_args.nr_vcpus);
> > > - TEST_ASSERT(vcpu_done_map, "Failed to allocate vcpu done bitmap\n");
> > > -
> > > - for (i = 0; i < (unsigned long)test_args.nr_vcpus; i++) {
> > > - ret = pthread_create(&pt_vcpu_run[i], NULL, test_vcpu_run,
> > > - (void *)(unsigned long)i);
> > > - TEST_ASSERT(!ret, "Failed to create vCPU-%d pthread\n", i);
> > > - }
> > > -
> > > - /* Spawn a thread to control the vCPU migrations */
> > > - if (test_args.migration_freq_ms) {
> > > - srand(time(NULL));
> > > -
> > > - ret = pthread_create(&pt_vcpu_migration, NULL,
> > > - test_vcpu_migration, NULL);
> > > - TEST_ASSERT(!ret, "Failed to create the migration pthread\n");
> > > - }
> > > -
> > > -
> > > - for (i = 0; i < test_args.nr_vcpus; i++)
> > > - pthread_join(pt_vcpu_run[i], NULL);
> > > -
> > > - if (test_args.migration_freq_ms)
> > > - pthread_join(pt_vcpu_migration, NULL);
> > > -
> > > - bitmap_free(vcpu_done_map);
> > > -}
> > > -
> > > static void test_init_timer_irq(struct kvm_vm *vm)
> > > {
> > > /* Timer initid should be same for all the vCPUs, so query only vCPU-0 */
> > > @@ -369,7 +172,7 @@ static void test_init_timer_irq(struct kvm_vm *vm)
> > >
> > > static int gic_fd;
> > >
> > > -static struct kvm_vm *test_vm_create(void)
> > > +struct kvm_vm *test_vm_create(void)
> > > {
> > > struct kvm_vm *vm;
> > > unsigned int i;
> > > @@ -400,81 +203,8 @@ static struct kvm_vm *test_vm_create(void)
> > > return vm;
> > > }
> > >
> > > -static void test_vm_cleanup(struct kvm_vm *vm)
> > > +void test_vm_cleanup(struct kvm_vm *vm)
> > > {
> > > close(gic_fd);
> > > kvm_vm_free(vm);
> > > }
> > > -
> > > -static void test_print_help(char *name)
> > > -{
> > > - pr_info("Usage: %s [-h] [-n nr_vcpus] [-i iterations] [-p timer_period_ms]\n",
> > > - name);
> > > - pr_info("\t-n: Number of vCPUs to configure (default: %u; max: %u)\n",
> > > - NR_VCPUS_DEF, KVM_MAX_VCPUS);
> > > - pr_info("\t-i: Number of iterations per stage (default: %u)\n",
> > > - NR_TEST_ITERS_DEF);
> > > - pr_info("\t-p: Periodicity (in ms) of the guest timer (default: %u)\n",
> > > - TIMER_TEST_PERIOD_MS_DEF);
> > > - pr_info("\t-m: Frequency (in ms) of vCPUs to migrate to different pCPU. 0 to turn off (default: %u)\n",
> > > - TIMER_TEST_MIGRATION_FREQ_MS);
> > > - pr_info("\t-o: Counter offset (in counter cycles, default: 0)\n");
> > > - pr_info("\t-h: print this help screen\n");
> > > -}
> > > -
> > > -static bool parse_args(int argc, char *argv[])
> > > -{
> > > - int opt;
> > > -
> > > - while ((opt = getopt(argc, argv, "hn:i:p:m:o:")) != -1) {
> > > - switch (opt) {
> > > - case 'n':
> > > - test_args.nr_vcpus = atoi_positive("Number of vCPUs", optarg);
> > > - if (test_args.nr_vcpus > KVM_MAX_VCPUS) {
> > > - pr_info("Max allowed vCPUs: %u\n",
> > > - KVM_MAX_VCPUS);
> > > - goto err;
> > > - }
> > > - break;
> > > - case 'i':
> > > - test_args.nr_iter = atoi_positive("Number of iterations", optarg);
> > > - break;
> > > - case 'p':
> > > - test_args.timer_period_ms = atoi_positive("Periodicity", optarg);
> > > - break;
> > > - case 'm':
> > > - test_args.migration_freq_ms = atoi_non_negative("Frequency", optarg);
> > > - break;
> > > - case 'o':
> > > - test_args.offset.counter_offset = strtol(optarg, NULL, 0);
> > > - test_args.offset.reserved = 0;
> > > - break;
> > > - case 'h':
> > > - default:
> > > - goto err;
> > > - }
> > > - }
> > > -
> > > - return true;
> > > -
> > > -err:
> > > - test_print_help(argv[0]);
> > > - return false;
> > > -}
> > > -
> > > -int main(int argc, char *argv[])
> > > -{
> > > - struct kvm_vm *vm;
> > > -
> > > - if (!parse_args(argc, argv))
> > > - exit(KSFT_SKIP);
> > > -
> > > - __TEST_REQUIRE(!test_args.migration_freq_ms || get_nprocs() >= 2,
> > > - "At least two physical CPUs needed for vCPU migration");
> > > -
> > > - vm = test_vm_create();
> > > - test_run(vm);
> > > - test_vm_cleanup(vm);
> > > -
> > > - return 0;
> > > -}
> > > diff --git a/tools/testing/selftests/kvm/arch_timer.c b/tools/testing/selftests/kvm/arch_timer.c
> > > new file mode 100644
> > > index 000000000000..529024f58c98
> > > --- /dev/null
> > > +++ b/tools/testing/selftests/kvm/arch_timer.c
> > > @@ -0,0 +1,252 @@
> > > +// SPDX-License-Identifier: GPL-2.0-only
> > > +/*
> > > + * arch_timer.c - Tests the arch timer IRQ functionality
> > > + *
> > > + * The guest's main thread configures the timer interrupt for and waits
> >
> > It looks like the text was edited to remove the 'stage' references, which
> > is fine by me, but the 'for' should have also been removed.
> >
> >
>
> Sure!
>
> > > + * for it to fire, with a timeout equal to the timer period.
> > > + * It asserts that the timeout doesn't exceed the timer period.
> > > + *
> > > + * On the other hand, upon receipt of an interrupt, the guest's interrupt
> > > + * handler validates the interrupt by checking if the architectural state
> > > + * is in compliance with the specifications.
> > > + *
> > > + * The test provides command-line options to configure the timer's
> > > + * period (-p), number of vCPUs (-n), and iterations per stage (-i).
> > > + * To stress-test the timer stack even more, an option to migrate the
> > > + * vCPUs across pCPUs (-m), at a particular rate, is also provided.
> > > + *
> > > + * Copyright (c) 2021, Google LLC.
> > > + */
> > > +
> > > +#define _GNU_SOURCE
> > > +
> > > +#include <stdlib.h>
> > > +#include <pthread.h>
> > > +#include <linux/sizes.h>
> > > +#include <linux/bitmap.h>
> > > +#include <sys/sysinfo.h>
> > > +
> > > +#include "timer_test.h"
> > > +
> > > +struct test_args test_args = {
> > > + .nr_vcpus = NR_VCPUS_DEF,
> > > + .nr_iter = NR_TEST_ITERS_DEF,
> > > + .timer_period_ms = TIMER_TEST_PERIOD_MS_DEF,
> > > + .migration_freq_ms = TIMER_TEST_MIGRATION_FREQ_MS,
> > > +#ifdef __aarch64__
> > > + .offset = { .reserved = 1 },
> > > +#endif
> >
> > Please run checkpatch, there are spaces instead of tabs in the struct.
> >
>
> Yes, the tabs were changed to spaces while copying. Will change it and
> run checkpatch to catch this kind of error in next version.
>
> > > +};
> > > +
> > > +struct kvm_vcpu *vcpus[KVM_MAX_VCPUS];
> > > +struct test_vcpu_shared_data vcpu_shared_data[KVM_MAX_VCPUS];
> > > +
> > > +static pthread_t pt_vcpu_run[KVM_MAX_VCPUS];
> > > +static unsigned long *vcpu_done_map;
> > > +static pthread_mutex_t vcpu_done_map_lock;
> > > +
> > > +static void *test_vcpu_run(void *arg)
> > > +{
> > > + unsigned int vcpu_idx = (unsigned long)arg;
> > > + struct ucall uc;
> > > + struct kvm_vcpu *vcpu = vcpus[vcpu_idx];
> > > + struct kvm_vm *vm = vcpu->vm;
> > > + struct test_vcpu_shared_data *shared_data = &vcpu_shared_data[vcpu_idx];
> > > +
> > > + vcpu_run(vcpu);
> > > +
> > > + /* Currently, any exit from guest is an indication of completion */
> > > + pthread_mutex_lock(&vcpu_done_map_lock);
> > > + __set_bit(vcpu_idx, vcpu_done_map);
> > > + pthread_mutex_unlock(&vcpu_done_map_lock);
> > > +
> > > + switch (get_ucall(vcpu, &uc)) {
> > > + case UCALL_SYNC:
> > > + case UCALL_DONE:
> > > + break;
> > > + case UCALL_ABORT:
> > > + sync_global_from_guest(vm, *shared_data);
> > > + fprintf(stderr, "Guest assert failed, vcpu %u; stage; %u; iter: %u\n",
> > > + vcpu_idx, shared_data->guest_stage, shared_data->nr_iter);
> > > + REPORT_GUEST_ASSERT(uc);
> > > + break;
> > > + default:
> > > + TEST_FAIL("Unexpected guest exit\n");
> > > + }
> > > +
> > > + pr_info("PASS(vCPU-%d).\n", vcpu_idx);
> >
> > This is new. I can live with it, but generally we don't want to modify
> > functions while moving them.
> >
>
> Yes, this change was supposed to go with patch 8/8.
>
> > > +
> > > + return NULL;
> > > +}
> > > +
> > > +static uint32_t test_get_pcpu(void)
> > > +{
> > > + uint32_t pcpu;
> > > + unsigned int nproc_conf;
> > > + cpu_set_t online_cpuset;
> > > +
> > > + nproc_conf = get_nprocs_conf();
> > > + sched_getaffinity(0, sizeof(cpu_set_t), &online_cpuset);
> > > +
> > > + /* Randomly find an available pCPU to place a vCPU on */
> > > + do {
> > > + pcpu = rand() % nproc_conf;
> > > + } while (!CPU_ISSET(pcpu, &online_cpuset));
> > > +
> > > + return pcpu;
> > > +}
> > > +
> > > +static int test_migrate_vcpu(unsigned int vcpu_idx)
> > > +{
> > > + int ret;
> > > + cpu_set_t cpuset;
> > > + uint32_t new_pcpu = test_get_pcpu();
> > > +
> > > + CPU_ZERO(&cpuset);
> > > + CPU_SET(new_pcpu, &cpuset);
> > > +
> > > + pr_debug("Migrating vCPU: %u to pCPU: %u\n", vcpu_idx, new_pcpu);
> > > +
> > > + ret = pthread_setaffinity_np(pt_vcpu_run[vcpu_idx],
> > > + sizeof(cpuset), &cpuset);
> > > +
> > > + /* Allow the error where the vCPU thread is already finished */
> > > + TEST_ASSERT(ret == 0 || ret == ESRCH,
> > > + "Failed to migrate the vCPU:%u to pCPU: %u; ret: %d\n",
> > > + vcpu_idx, new_pcpu, ret);
> > > +
> > > + return ret;
> > > +}
> > > +
> > > +static void *test_vcpu_migration(void *arg)
> > > +{
> > > + unsigned int i, n_done;
> > > + bool vcpu_done;
> > > +
> > > + do {
> > > + usleep(msecs_to_usecs(test_args.migration_freq_ms));
> > > +
> > > + for (n_done = 0, i = 0; i < test_args.nr_vcpus; i++) {
> > > + pthread_mutex_lock(&vcpu_done_map_lock);
> > > + vcpu_done = test_bit(i, vcpu_done_map);
> > > + pthread_mutex_unlock(&vcpu_done_map_lock);
> > > +
> > > + if (vcpu_done) {
> > > + n_done++;
> > > + continue;
> > > + }
> > > +
> > > + test_migrate_vcpu(i);
> > > + }
> > > + } while (test_args.nr_vcpus != n_done);
> > > +
> > > + return NULL;
> > > +}
> > > +
> > > +static void test_run(struct kvm_vm *vm)
> > > +{
> > > + pthread_t pt_vcpu_migration;
> > > + unsigned int i;
> > > + int ret;
> > > +
> > > + pthread_mutex_init(&vcpu_done_map_lock, NULL);
> > > + vcpu_done_map = bitmap_zalloc(test_args.nr_vcpus);
> > > + TEST_ASSERT(vcpu_done_map, "Failed to allocate vcpu done bitmap\n");
> > > +
> > > + for (i = 0; i < (unsigned long)test_args.nr_vcpus; i++) {
> > > + ret = pthread_create(&pt_vcpu_run[i], NULL, test_vcpu_run,
> > > + (void *)(unsigned long)i);
> > > + TEST_ASSERT(!ret, "Failed to create vCPU-%d pthread\n", i);
> > > + }
> > > +
> > > + /* Spawn a thread to control the vCPU migrations */
> > > + if (test_args.migration_freq_ms) {
> > > + srand(time(NULL));
> > > +
> > > + ret = pthread_create(&pt_vcpu_migration, NULL,
> > > + test_vcpu_migration, NULL);
> > > + TEST_ASSERT(!ret, "Failed to create the migration pthread\n");
> > > + }
> > > +
> > > +
> > > + for (i = 0; i < test_args.nr_vcpus; i++)
> > > + pthread_join(pt_vcpu_run[i], NULL);
> > > +
> > > + if (test_args.migration_freq_ms)
> > > + pthread_join(pt_vcpu_migration, NULL);
> > > +
> > > + bitmap_free(vcpu_done_map);
> > > +}
> > > +
> > > +static void test_print_help(char *name)
> > > +{
> > > + pr_info("Usage: %s [-h] [-n nr_vcpus] [-i iterations] [-p timer_period_ms]\n",
> > > + name);
> > > + pr_info("\t-n: Number of vCPUs to configure (default: %u; max: %u)\n",
> > > + NR_VCPUS_DEF, KVM_MAX_VCPUS);
> > > + pr_info("\t-i: Number of iterations per stage (default: %u)\n",
> > > + NR_TEST_ITERS_DEF);
> > > + pr_info("\t-p: Periodicity (in ms) of the guest timer (default: %u)\n",
> > > + TIMER_TEST_PERIOD_MS_DEF);
> > > + pr_info("\t-m: Frequency (in ms) of vCPUs to migrate to different pCPU. 0 to turn off (default: %u)\n",
> > > + TIMER_TEST_MIGRATION_FREQ_MS);
> > > + pr_info("\t-o: Counter offset (in counter cycles, default: 0)\n");
> > > + pr_info("\t-h: print this help screen\n");
> > > +}
> > > +
> > > +static bool parse_args(int argc, char *argv[])
> > > +{
> > > + int opt;
> > > +
> > > + while ((opt = getopt(argc, argv, "hn:i:p:m:o:")) != -1) {
> > > + switch (opt) {
> > > + case 'n':
> > > + test_args.nr_vcpus = atoi_positive("Number of vCPUs", optarg);
> > > + if (test_args.nr_vcpus > KVM_MAX_VCPUS) {
> > > + pr_info("Max allowed vCPUs: %u\n",
> > > + KVM_MAX_VCPUS);
> > > + goto err;
> > > + }
> > > + break;
> > > + case 'i':
> > > + test_args.nr_iter = atoi_positive("Number of iterations", optarg);
> > > + break;
> > > + case 'p':
> > > + test_args.timer_period_ms = atoi_positive("Periodicity", optarg);
> > > + break;
> > > + case 'm':
> > > + test_args.migration_freq_ms = atoi_non_negative("Frequency", optarg);
> > > + break;
> > > + case 'o':
> > > + test_args.offset.counter_offset = strtol(optarg, NULL, 0);
> > > + test_args.offset.reserved = 0;
> > > + break;
> > > + case 'h':
> > > + default:
> > > + goto err;
> > > + }
> > > + }
> > > +
> > > + return true;
> > > +
> > > +err:
> > > + test_print_help(argv[0]);
> > > + return false;
> > > +}
> > > +
> > > +int main(int argc, char *argv[])
> > > +{
> > > + struct kvm_vm *vm;
> > > +
> > > + if (!parse_args(argc, argv))
> > > + exit(KSFT_SKIP);
> > > +
> > > + __TEST_REQUIRE(!test_args.migration_freq_ms || get_nprocs() >= 2,
> > > + "At least two physical CPUs needed for vCPU migration");
> > > +
> > > + vm = test_vm_create();
> > > + test_run(vm);
> > > + test_vm_cleanup(vm);
> > > +
> > > + return 0;
> > > +}
> > > diff --git a/tools/testing/selftests/kvm/include/timer_test.h b/tools/testing/selftests/kvm/include/timer_test.h
> > > new file mode 100644
> > > index 000000000000..109e4d635627
> > > --- /dev/null
> > > +++ b/tools/testing/selftests/kvm/include/timer_test.h
> > > @@ -0,0 +1,52 @@
> > > +/* SPDX-License-Identifier: GPL-2.0-only */
> > > +/*
> > > + * tools/testing/selftests/kvm/include/timer_test.h
> > > + *
> > > + * Copyright (C) 2018, Google LLC
> > > + */
> > > +
> > > +#ifndef SELFTEST_KVM_TIMER_TEST_H
> > > +#define SELFTEST_KVM_TIMER_TEST_H
> > > +
> > > +#include "kvm_util.h"
> > > +
> > > +#define NR_VCPUS_DEF 4
> > > +#define NR_TEST_ITERS_DEF 5
> > > +#define TIMER_TEST_PERIOD_MS_DEF 10
> > > +#define TIMER_TEST_ERR_MARGIN_US 100
> > > +#define TIMER_TEST_MIGRATION_FREQ_MS 2
> > > +
> > > +#define msecs_to_usecs(msec) ((msec) * 1000LL)
> >
> > I'd move the above to include/test_util.h
> >
>
> Yes, msecs_to_usecs() macro should be a common API for all the tests.
>
> > > +
> > > +#define GICD_BASE_GPA 0x8000000ULL
> > > +#define GICR_BASE_GPA 0x80A0000ULL
> >
> > These defines belong in aarch64/arch_timer.c
> >
>
> These 2 defines were also defined in other test cases, shall we move them
> to an aarch64 specific header file? Maybe
> tools/testing/selftests/kvm/include/aarch64/gic.h?
>
> > > +
> > > +enum guest_stage {
> > > + GUEST_STAGE_VTIMER_CVAL=1,
> > > + GUEST_STAGE_VTIMER_TVAL,
> > > + GUEST_STAGE_PTIMER_CVAL,
> > > + GUEST_STAGE_PTIMER_TVAL,
> > > + GUEST_STAGE_MAX,
> > > +};
> >
> > This enum also belongs in aarch64/arch_timer.c
> >
>
> Yes, it should be in aarch64/arch_timer.c
>
After moving the above enum definition to aarch64/arch_timer.c, the
below errors was reported
while compiling kvm/arch_timer.o
include/timer_test.h:37:26: error: field ‘guest_stage’ has incomplete type
37 | enum guest_stage guest_stage;
| ^~~~~~~~~~~
Since kvm/arch_timer.c was independent of kvm/aarch64/arch_timer.c
during OBJ compiling,
I think it may be not possible to move the enum definition to
aarch64/arch_timer.c
If we keep the definition in this header file, we can enclose it with
#ifdef __aarch64__ for aarch64 only.
> > > +
> > > +/* Timer test cmdline parameters */
> > > +struct test_args
> > > +{
> > > + int nr_vcpus;
> > > + int nr_iter;
> > > + int timer_period_ms;
> > > + int migration_freq_ms;
> > > + struct kvm_arm_counter_offset offset;
> > > +};
> > > +
> > > +/* Shared variables between host and guest */
> > > +struct test_vcpu_shared_data {
> > > + int nr_iter;
> > > + enum guest_stage guest_stage;
> > > + uint64_t xcnt;
> > > +};
> > > +
> > > +struct kvm_vm* test_vm_create(void);
> >
> > Move * next to function name.
> >
>
> Sure, thanks!
>
> > > +void test_vm_cleanup(struct kvm_vm *vm);
> > > +
> > > +#endif /* SELFTEST_KVM_TIMER_TEST_H */
> > > --
> > > 2.34.1
> > >
> >
> > Thanks,
> > drew
More information about the linux-arm-kernel
mailing list