aboutsummaryrefslogtreecommitdiffstatshomepage
path: root/tools/testing/selftests/kvm/x86_64/tsc_scaling_sync.c
diff options
context:
space:
mode:
authorSean Christopherson <seanjc@google.com>2022-04-18 11:28:21 -0700
committerPaolo Bonzini <pbonzini@redhat.com>2022-06-11 11:46:59 -0400
commitee7f7d9e988e137f20a34b8c02dd28dd5312e3f1 (patch)
treebb8cfcd4c0b6cfe6b12e9e1d9067b98fbfd496c9 /tools/testing/selftests/kvm/x86_64/tsc_scaling_sync.c
parentKVM: selftests: Convert system_counter_offset_test away from VCPU_ID (diff)
downloadwireguard-linux-ee7f7d9e988e137f20a34b8c02dd28dd5312e3f1.tar.xz
wireguard-linux-ee7f7d9e988e137f20a34b8c02dd28dd5312e3f1.zip
KVM: selftests: Track kvm_vcpu object in tsc_scaling_sync
Track the added 'struct kvm_vcpu' object in tsc_scaling_sync instead of relying purely on the VM + vcpu_id combination. Ideally, the test wouldn't need to manually manage vCPUs, but the need to invoke a per-VM ioctl before creating vCPUs is not handled by the selftests framework, at least not yet... Signed-off-by: Sean Christopherson <seanjc@google.com> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
Diffstat (limited to 'tools/testing/selftests/kvm/x86_64/tsc_scaling_sync.c')
-rw-r--r--tools/testing/selftests/kvm/x86_64/tsc_scaling_sync.c17
1 files changed, 10 insertions, 7 deletions
diff --git a/tools/testing/selftests/kvm/x86_64/tsc_scaling_sync.c b/tools/testing/selftests/kvm/x86_64/tsc_scaling_sync.c
index f0083d8cfe98..b7cd5c47fc53 100644
--- a/tools/testing/selftests/kvm/x86_64/tsc_scaling_sync.c
+++ b/tools/testing/selftests/kvm/x86_64/tsc_scaling_sync.c
@@ -46,38 +46,41 @@ static void guest_code(void)
static void *run_vcpu(void *_cpu_nr)
{
- unsigned long cpu = (unsigned long)_cpu_nr;
+ unsigned long vcpu_id = (unsigned long)_cpu_nr;
unsigned long failures = 0;
static bool first_cpu_done;
+ struct kvm_vcpu *vcpu;
/* The kernel is fine, but vm_vcpu_add_default() needs locking */
pthread_spin_lock(&create_lock);
- vm_vcpu_add_default(vm, cpu, guest_code);
+ vm_vcpu_add_default(vm, vcpu_id, guest_code);
+ vcpu = vcpu_get(vm, vcpu_id);
if (!first_cpu_done) {
first_cpu_done = true;
- vcpu_set_msr(vm, cpu, MSR_IA32_TSC, TEST_TSC_OFFSET);
+ vcpu_set_msr(vm, vcpu->id, MSR_IA32_TSC, TEST_TSC_OFFSET);
}
pthread_spin_unlock(&create_lock);
for (;;) {
- volatile struct kvm_run *run = vcpu_state(vm, cpu);
+ volatile struct kvm_run *run = vcpu->run;
struct ucall uc;
- vcpu_run(vm, cpu);
+ vcpu_run(vm, vcpu->id);
TEST_ASSERT(run->exit_reason == KVM_EXIT_IO,
"Got exit_reason other than KVM_EXIT_IO: %u (%s)\n",
run->exit_reason,
exit_reason_str(run->exit_reason));
- switch (get_ucall(vm, cpu, &uc)) {
+ switch (get_ucall(vm, vcpu->id, &uc)) {
case UCALL_DONE:
goto out;
case UCALL_SYNC:
- printf("Guest %ld sync %lx %lx %ld\n", cpu, uc.args[2], uc.args[3], uc.args[2] - uc.args[3]);
+ printf("Guest %d sync %lx %lx %ld\n", vcpu->id,
+ uc.args[2], uc.args[3], uc.args[2] - uc.args[3]);
failures++;
break;