@@ -69,6 +69,7 @@ static int iteration;
static int vcpu_last_completed_iteration[KVM_MAX_VCPUS];
static atomic_ullong total_reads;
static atomic_ullong total_writes;
+static bool lockstep_iterations;
static void vcpu_worker(struct memstress_vcpu_args *vcpu_args)
{
@@ -83,12 +84,16 @@ static void vcpu_worker(struct memstress_vcpu_args *vcpu_args)
struct timespec total = (struct timespec){0};
struct timespec avg;
struct ucall uc = {};
+ int current_iteration = -1;
int ret;
run = vcpu->run;
while (!READ_ONCE(host_quit)) {
- int current_iteration = READ_ONCE(iteration);
+ if (lockstep_iterations)
+ current_iteration = READ_ONCE(iteration);
+ else
+ current_iteration++;
clock_gettime(CLOCK_MONOTONIC, &start);
ret = _vcpu_run(vcpu);
@@ -118,13 +123,19 @@ static void vcpu_worker(struct memstress_vcpu_args *vcpu_args)
ts_diff.tv_nsec);
}
- /*
- * Keep running the guest while dirty logging is being disabled
- * (iteration is negative) so that vCPUs are accessing memory
- * for the entire duration of zapping collapsible SPTEs.
- */
- while (current_iteration == READ_ONCE(iteration) &&
- READ_ONCE(iteration) >= 0 && !READ_ONCE(host_quit)) {}
+ if (lockstep_iterations) {
+ /*
+ * Keep running the guest while dirty logging is being disabled
+ * (iteration is negative) so that vCPUs are accessing memory
+ * for the entire duration of zapping collapsible SPTEs.
+ */
+ while (current_iteration == READ_ONCE(iteration) &&
+ READ_ONCE(iteration) >= 0 && !READ_ONCE(host_quit))
+ ;
+ } else {
+ while (!READ_ONCE(iteration))
+ ;
+ }
}
avg = timespec_div(total, vcpu_last_completed_iteration[vcpu_idx]);
@@ -238,17 +249,19 @@ static void run_test(enum vm_guest_mode mode, void *arg)
clock_gettime(CLOCK_MONOTONIC, &start);
iteration++;
- pr_debug("Starting iteration %d\n", iteration);
- for (i = 0; i < nr_vcpus; i++) {
- while (READ_ONCE(vcpu_last_completed_iteration[i])
- != iteration)
- ;
- }
+ if (lockstep_iterations) {
+ pr_debug("Starting iteration %d\n", iteration);
+ for (i = 0; i < nr_vcpus; i++) {
+ while (READ_ONCE(vcpu_last_completed_iteration[i])
+ != iteration)
+ ;
+ }
- ts_diff = timespec_elapsed(start);
- vcpu_dirty_total = timespec_add(vcpu_dirty_total, ts_diff);
- pr_info("Iteration %d dirty memory time: %ld.%.9lds\n",
- iteration, ts_diff.tv_sec, ts_diff.tv_nsec);
+ ts_diff = timespec_elapsed(start);
+ vcpu_dirty_total = timespec_add(vcpu_dirty_total, ts_diff);
+ pr_info("Iteration %d dirty memory time: %ld.%.9lds\n",
+ iteration, ts_diff.tv_sec, ts_diff.tv_nsec);
+ }
clock_gettime(CLOCK_MONOTONIC, &start);
memstress_get_dirty_log(vm, bitmaps, p->slots);
@@ -365,6 +378,10 @@ static void help(char *name)
" To leave the application task unpinned, drop the final entry:\n\n"
" ./dirty_log_perf_test -v 3 -c 22,23,24\n\n"
" (default: no pinning)\n");
+ printf(" -j: Execute vCPUs independent of dirty log iterations\n"
+ " Independent vCPUs execution will allow them to continuously\n"
+ " dirty memory while main thread is collecting and clearing\n"
+ " dirty logs in the main thread's iterations.\n");
printf(" -k: Specify the chunk size in which dirty memory gets cleared\n"
" in memslots in each iteration. If the size is bigger than\n"
" the memslot size then whole memslot is cleared in one call.\n"
@@ -399,10 +416,10 @@ int main(int argc, char *argv[])
kvm_check_cap(KVM_CAP_MANUAL_DIRTY_LOG_PROTECT2);
dirty_log_manual_caps &= (KVM_DIRTY_LOG_MANUAL_PROTECT_ENABLE |
KVM_DIRTY_LOG_INITIALLY_SET);
-
+ lockstep_iterations = true;
guest_modes_append_default();
- while ((opt = getopt(argc, argv, "ab:c:eghi:k:l:m:nop:r:s:v:x:w:")) != -1) {
+ while ((opt = getopt(argc, argv, "ab:c:eghi:jk:l:m:nop:r:s:v:x:w:")) != -1) {
switch (opt) {
case 'a':
p.random_access = true;
@@ -426,6 +443,9 @@ int main(int argc, char *argv[])
case 'i':
p.iterations = atoi_positive("Number of iterations", optarg);
break;
+ case 'j':
+ lockstep_iterations = false;
+ break;
case 'k':
p.clear_chunk_size = parse_size(optarg);
break;
Allow vCPUs to execute independent of dirty log iterations after initialization is complete. Hide this feature behind the new option "-j". This change makes dirty_log_perf_test execute like real world workflows where guest vCPUs keep on executing while VMM collects dirty logs. Total pages touched during execution of test will give good estimate of how vCPUs are performing while dirty logging is enabled. Signed-off-by: Vipin Sharma <vipinsh@google.com> --- .../selftests/kvm/dirty_log_perf_test.c | 60 ++++++++++++------- 1 file changed, 40 insertions(+), 20 deletions(-)