[PATCH v2 05/16] KVM: selftests: Allow independent execution of vCPUs in dirty_log_perf_test
Vipin Sharma
vipinsh at google.com
Fri Jun 2 09:09:03 PDT 2023
Give users command line option (-j) to execute vCPUs independently of
dirty log iterations after initialization is complete.
This change makes dirty_log_perf_test behave like real world workflows
where guest vCPUs keep on executing while VMM collects and clear dirty
logs. Total pages touched during execution of test will give good
estimate of how vCPUs are performing while dirty logging is enabled.
Signed-off-by: Vipin Sharma <vipinsh at google.com>
---
.../selftests/kvm/dirty_log_perf_test.c | 64 +++++++++++++------
1 file changed, 44 insertions(+), 20 deletions(-)
diff --git a/tools/testing/selftests/kvm/dirty_log_perf_test.c b/tools/testing/selftests/kvm/dirty_log_perf_test.c
index 14b012a0dcb1..fbf973d6cc66 100644
--- a/tools/testing/selftests/kvm/dirty_log_perf_test.c
+++ b/tools/testing/selftests/kvm/dirty_log_perf_test.c
@@ -69,6 +69,7 @@ static int iteration;
static int vcpu_last_completed_iteration[KVM_MAX_VCPUS];
static atomic_ullong total_reads;
static atomic_ullong total_writes;
+static bool lockstep_iterations;
static void vcpu_worker(struct memstress_vcpu_args *vcpu_args)
{
@@ -83,12 +84,16 @@ static void vcpu_worker(struct memstress_vcpu_args *vcpu_args)
struct timespec total = (struct timespec){0};
struct timespec avg;
struct ucall uc = {};
+ int current_iteration = -1;
int ret;
run = vcpu->run;
while (!READ_ONCE(host_quit)) {
- int current_iteration = READ_ONCE(iteration);
+ if (lockstep_iterations)
+ current_iteration = READ_ONCE(iteration);
+ else
+ current_iteration++;
clock_gettime(CLOCK_MONOTONIC, &start);
ret = _vcpu_run(vcpu);
@@ -118,13 +123,19 @@ static void vcpu_worker(struct memstress_vcpu_args *vcpu_args)
ts_diff.tv_nsec);
}
- /*
- * Keep running the guest while dirty logging is being disabled
- * (iteration is negative) so that vCPUs are accessing memory
- * for the entire duration of zapping collapsible SPTEs.
- */
- while (current_iteration == READ_ONCE(iteration) &&
- READ_ONCE(iteration) >= 0 && !READ_ONCE(host_quit)) {}
+ if (lockstep_iterations) {
+ /*
+ * Keep running the guest while dirty logging is being disabled
+ * (iteration is negative) so that vCPUs are accessing memory
+ * for the entire duration of zapping collapsible SPTEs.
+ */
+ while (current_iteration == READ_ONCE(iteration) &&
+ READ_ONCE(iteration) >= 0 && !READ_ONCE(host_quit))
+ ;
+ } else {
+ while (!READ_ONCE(iteration) && !READ_ONCE(host_quit))
+ ;
+ }
}
avg = timespec_div(total, vcpu_last_completed_iteration[vcpu_idx]);
@@ -332,18 +343,20 @@ static void run_test(enum vm_guest_mode mode, void *arg)
clock_gettime(CLOCK_MONOTONIC, &start);
iteration++;
- pr_debug("Starting iteration %d\n", iteration);
- for (i = 0; i < nr_vcpus; i++) {
- while (READ_ONCE(vcpu_last_completed_iteration[i])
- != iteration)
- ;
+ if (lockstep_iterations) {
+ pr_debug("Starting iteration %d\n", iteration);
+ for (i = 0; i < nr_vcpus; i++) {
+ while (READ_ONCE(vcpu_last_completed_iteration[i])
+ != iteration)
+ ;
+ }
+
+ ts_diff = timespec_elapsed(start);
+ vcpu_dirty_total = timespec_add(vcpu_dirty_total, ts_diff);
+ pr_info("Iteration %d dirty memory time: %ld.%.9lds\n",
+ iteration, ts_diff.tv_sec, ts_diff.tv_nsec);
}
- ts_diff = timespec_elapsed(start);
- vcpu_dirty_total = timespec_add(vcpu_dirty_total, ts_diff);
- pr_info("Iteration %d dirty memory time: %ld.%.9lds\n",
- iteration, ts_diff.tv_sec, ts_diff.tv_nsec);
-
clock_gettime(CLOCK_MONOTONIC, &start);
get_dirty_log(vm, bitmaps, p->slots);
ts_diff = timespec_elapsed(start);
@@ -365,6 +378,10 @@ static void run_test(enum vm_guest_mode mode, void *arg)
}
}
+ /* Block further vCPUs execution */
+ if (!lockstep_iterations)
+ WRITE_ONCE(iteration, 0);
+
/*
* Run vCPUs while dirty logging is being disabled to stress disabling
* in terms of both performance and correctness. Opt-in via command
@@ -458,6 +475,10 @@ static void help(char *name)
" To leave the application task unpinned, drop the final entry:\n\n"
" ./dirty_log_perf_test -v 3 -c 22,23,24\n\n"
" (default: no pinning)\n");
+ printf(" -j: Execute vCPUs independent of dirty log iterations\n"
+ " Independent vCPUs execution will allow them to continuously\n"
+ " dirty memory while main thread is collecting and clearing\n"
+ " dirty logs in each iteration.\n");
printf(" -k: Specify the chunk size in which dirty memory gets cleared\n"
" in memslots in each iteration. If the size is bigger than\n"
" the memslot size then whole memslot is cleared in one call.\n"
@@ -492,10 +513,10 @@ int main(int argc, char *argv[])
kvm_check_cap(KVM_CAP_MANUAL_DIRTY_LOG_PROTECT2);
dirty_log_manual_caps &= (KVM_DIRTY_LOG_MANUAL_PROTECT_ENABLE |
KVM_DIRTY_LOG_INITIALLY_SET);
-
+ lockstep_iterations = true;
guest_modes_append_default();
- while ((opt = getopt(argc, argv, "ab:c:eghi:k:l:m:nop:r:s:v:x:w:")) != -1) {
+ while ((opt = getopt(argc, argv, "ab:c:eghi:jk:l:m:nop:r:s:v:x:w:")) != -1) {
switch (opt) {
case 'a':
p.random_access = true;
@@ -519,6 +540,9 @@ int main(int argc, char *argv[])
case 'i':
p.iterations = atoi_positive("Number of iterations", optarg);
break;
+ case 'j':
+ lockstep_iterations = false;
+ break;
case 'k':
p.clear_chunk_size = parse_size(optarg);
break;
--
2.41.0.rc0.172.g3f132b7071-goog
More information about the linux-arm-kernel
mailing list