690 lines
17 KiB
C
690 lines
17 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (C) 2021, Red Hat, Inc.
|
|
*
|
|
* Tests for Hyper-V features enablement
|
|
*/
|
|
#include <asm/kvm_para.h>
|
|
#include <linux/kvm_para.h>
|
|
#include <stdint.h>
|
|
|
|
#include "test_util.h"
|
|
#include "kvm_util.h"
|
|
#include "processor.h"
|
|
#include "hyperv.h"
|
|
|
|
/*
|
|
* HYPERV_CPUID_ENLIGHTMENT_INFO.EBX is not a 'feature' CPUID leaf
|
|
* but to activate the feature it is sufficient to set it to a non-zero
|
|
* value. Use BIT(0) for that.
|
|
*/
|
|
#define HV_PV_SPINLOCKS_TEST \
|
|
KVM_X86_CPU_FEATURE(HYPERV_CPUID_ENLIGHTMENT_INFO, 0, EBX, 0)
|
|
|
|
struct msr_data {
|
|
uint32_t idx;
|
|
bool fault_expected;
|
|
bool write;
|
|
u64 write_val;
|
|
};
|
|
|
|
struct hcall_data {
|
|
uint64_t control;
|
|
uint64_t expect;
|
|
bool ud_expected;
|
|
};
|
|
|
|
static bool is_write_only_msr(uint32_t msr)
|
|
{
|
|
return msr == HV_X64_MSR_EOI;
|
|
}
|
|
|
|
static void guest_msr(struct msr_data *msr)
|
|
{
|
|
uint8_t vector = 0;
|
|
uint64_t msr_val = 0;
|
|
|
|
GUEST_ASSERT(msr->idx);
|
|
|
|
if (msr->write)
|
|
vector = wrmsr_safe(msr->idx, msr->write_val);
|
|
|
|
if (!vector && (!msr->write || !is_write_only_msr(msr->idx)))
|
|
vector = rdmsr_safe(msr->idx, &msr_val);
|
|
|
|
if (msr->fault_expected)
|
|
GUEST_ASSERT_3(vector == GP_VECTOR, msr->idx, vector, GP_VECTOR);
|
|
else
|
|
GUEST_ASSERT_3(!vector, msr->idx, vector, 0);
|
|
|
|
if (vector || is_write_only_msr(msr->idx))
|
|
goto done;
|
|
|
|
if (msr->write)
|
|
GUEST_ASSERT_3(msr_val == msr->write_val, msr->idx,
|
|
msr_val, msr->write_val);
|
|
|
|
/* Invariant TSC bit appears when TSC invariant control MSR is written to */
|
|
if (msr->idx == HV_X64_MSR_TSC_INVARIANT_CONTROL) {
|
|
if (!this_cpu_has(HV_ACCESS_TSC_INVARIANT))
|
|
GUEST_ASSERT(this_cpu_has(X86_FEATURE_INVTSC));
|
|
else
|
|
GUEST_ASSERT(this_cpu_has(X86_FEATURE_INVTSC) ==
|
|
!!(msr_val & HV_INVARIANT_TSC_EXPOSED));
|
|
}
|
|
|
|
done:
|
|
GUEST_DONE();
|
|
}
|
|
|
|
static void guest_hcall(vm_vaddr_t pgs_gpa, struct hcall_data *hcall)
|
|
{
|
|
u64 res, input, output;
|
|
uint8_t vector;
|
|
|
|
GUEST_ASSERT(hcall->control);
|
|
|
|
wrmsr(HV_X64_MSR_GUEST_OS_ID, HYPERV_LINUX_OS_ID);
|
|
wrmsr(HV_X64_MSR_HYPERCALL, pgs_gpa);
|
|
|
|
if (!(hcall->control & HV_HYPERCALL_FAST_BIT)) {
|
|
input = pgs_gpa;
|
|
output = pgs_gpa + 4096;
|
|
} else {
|
|
input = output = 0;
|
|
}
|
|
|
|
vector = __hyperv_hypercall(hcall->control, input, output, &res);
|
|
if (hcall->ud_expected) {
|
|
GUEST_ASSERT_2(vector == UD_VECTOR, hcall->control, vector);
|
|
} else {
|
|
GUEST_ASSERT_2(!vector, hcall->control, vector);
|
|
GUEST_ASSERT_2(res == hcall->expect, hcall->expect, res);
|
|
}
|
|
|
|
GUEST_DONE();
|
|
}
|
|
|
|
static void vcpu_reset_hv_cpuid(struct kvm_vcpu *vcpu)
|
|
{
|
|
/*
|
|
* Enable all supported Hyper-V features, then clear the leafs holding
|
|
* the features that will be tested one by one.
|
|
*/
|
|
vcpu_set_hv_cpuid(vcpu);
|
|
|
|
vcpu_clear_cpuid_entry(vcpu, HYPERV_CPUID_FEATURES);
|
|
vcpu_clear_cpuid_entry(vcpu, HYPERV_CPUID_ENLIGHTMENT_INFO);
|
|
vcpu_clear_cpuid_entry(vcpu, HYPERV_CPUID_SYNDBG_PLATFORM_CAPABILITIES);
|
|
}
|
|
|
|
static void guest_test_msrs_access(void)
|
|
{
|
|
struct kvm_cpuid2 *prev_cpuid = NULL;
|
|
struct kvm_vcpu *vcpu;
|
|
struct kvm_vm *vm;
|
|
struct ucall uc;
|
|
int stage = 0;
|
|
vm_vaddr_t msr_gva;
|
|
struct msr_data *msr;
|
|
bool has_invtsc = kvm_cpu_has(X86_FEATURE_INVTSC);
|
|
|
|
while (true) {
|
|
vm = vm_create_with_one_vcpu(&vcpu, guest_msr);
|
|
|
|
msr_gva = vm_vaddr_alloc_page(vm);
|
|
memset(addr_gva2hva(vm, msr_gva), 0x0, getpagesize());
|
|
msr = addr_gva2hva(vm, msr_gva);
|
|
|
|
vcpu_args_set(vcpu, 1, msr_gva);
|
|
vcpu_enable_cap(vcpu, KVM_CAP_HYPERV_ENFORCE_CPUID, 1);
|
|
|
|
if (!prev_cpuid) {
|
|
vcpu_reset_hv_cpuid(vcpu);
|
|
|
|
prev_cpuid = allocate_kvm_cpuid2(vcpu->cpuid->nent);
|
|
} else {
|
|
vcpu_init_cpuid(vcpu, prev_cpuid);
|
|
}
|
|
|
|
vm_init_descriptor_tables(vm);
|
|
vcpu_init_descriptor_tables(vcpu);
|
|
|
|
/* TODO: Make this entire test easier to maintain. */
|
|
if (stage >= 21)
|
|
vcpu_enable_cap(vcpu, KVM_CAP_HYPERV_SYNIC2, 0);
|
|
|
|
switch (stage) {
|
|
case 0:
|
|
/*
|
|
* Only available when Hyper-V identification is set
|
|
*/
|
|
msr->idx = HV_X64_MSR_GUEST_OS_ID;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 1:
|
|
msr->idx = HV_X64_MSR_HYPERCALL;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 2:
|
|
vcpu_set_cpuid_feature(vcpu, HV_MSR_HYPERCALL_AVAILABLE);
|
|
/*
|
|
* HV_X64_MSR_GUEST_OS_ID has to be written first to make
|
|
* HV_X64_MSR_HYPERCALL available.
|
|
*/
|
|
msr->idx = HV_X64_MSR_GUEST_OS_ID;
|
|
msr->write = true;
|
|
msr->write_val = HYPERV_LINUX_OS_ID;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 3:
|
|
msr->idx = HV_X64_MSR_GUEST_OS_ID;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 4:
|
|
msr->idx = HV_X64_MSR_HYPERCALL;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
|
|
case 5:
|
|
msr->idx = HV_X64_MSR_VP_RUNTIME;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 6:
|
|
vcpu_set_cpuid_feature(vcpu, HV_MSR_VP_RUNTIME_AVAILABLE);
|
|
msr->idx = HV_X64_MSR_VP_RUNTIME;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 7:
|
|
/* Read only */
|
|
msr->idx = HV_X64_MSR_VP_RUNTIME;
|
|
msr->write = true;
|
|
msr->write_val = 1;
|
|
msr->fault_expected = true;
|
|
break;
|
|
|
|
case 8:
|
|
msr->idx = HV_X64_MSR_TIME_REF_COUNT;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 9:
|
|
vcpu_set_cpuid_feature(vcpu, HV_MSR_TIME_REF_COUNT_AVAILABLE);
|
|
msr->idx = HV_X64_MSR_TIME_REF_COUNT;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 10:
|
|
/* Read only */
|
|
msr->idx = HV_X64_MSR_TIME_REF_COUNT;
|
|
msr->write = true;
|
|
msr->write_val = 1;
|
|
msr->fault_expected = true;
|
|
break;
|
|
|
|
case 11:
|
|
msr->idx = HV_X64_MSR_VP_INDEX;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 12:
|
|
vcpu_set_cpuid_feature(vcpu, HV_MSR_VP_INDEX_AVAILABLE);
|
|
msr->idx = HV_X64_MSR_VP_INDEX;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 13:
|
|
/* Read only */
|
|
msr->idx = HV_X64_MSR_VP_INDEX;
|
|
msr->write = true;
|
|
msr->write_val = 1;
|
|
msr->fault_expected = true;
|
|
break;
|
|
|
|
case 14:
|
|
msr->idx = HV_X64_MSR_RESET;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 15:
|
|
vcpu_set_cpuid_feature(vcpu, HV_MSR_RESET_AVAILABLE);
|
|
msr->idx = HV_X64_MSR_RESET;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 16:
|
|
msr->idx = HV_X64_MSR_RESET;
|
|
msr->write = true;
|
|
/*
|
|
* TODO: the test only writes '0' to HV_X64_MSR_RESET
|
|
* at the moment, writing some other value there will
|
|
* trigger real vCPU reset and the code is not prepared
|
|
* to handle it yet.
|
|
*/
|
|
msr->write_val = 0;
|
|
msr->fault_expected = false;
|
|
break;
|
|
|
|
case 17:
|
|
msr->idx = HV_X64_MSR_REFERENCE_TSC;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 18:
|
|
vcpu_set_cpuid_feature(vcpu, HV_MSR_REFERENCE_TSC_AVAILABLE);
|
|
msr->idx = HV_X64_MSR_REFERENCE_TSC;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 19:
|
|
msr->idx = HV_X64_MSR_REFERENCE_TSC;
|
|
msr->write = true;
|
|
msr->write_val = 0;
|
|
msr->fault_expected = false;
|
|
break;
|
|
|
|
case 20:
|
|
msr->idx = HV_X64_MSR_EOM;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 21:
|
|
/*
|
|
* Remains unavailable even with KVM_CAP_HYPERV_SYNIC2
|
|
* capability enabled and guest visible CPUID bit unset.
|
|
*/
|
|
msr->idx = HV_X64_MSR_EOM;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 22:
|
|
vcpu_set_cpuid_feature(vcpu, HV_MSR_SYNIC_AVAILABLE);
|
|
msr->idx = HV_X64_MSR_EOM;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 23:
|
|
msr->idx = HV_X64_MSR_EOM;
|
|
msr->write = true;
|
|
msr->write_val = 0;
|
|
msr->fault_expected = false;
|
|
break;
|
|
|
|
case 24:
|
|
msr->idx = HV_X64_MSR_STIMER0_CONFIG;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 25:
|
|
vcpu_set_cpuid_feature(vcpu, HV_MSR_SYNTIMER_AVAILABLE);
|
|
msr->idx = HV_X64_MSR_STIMER0_CONFIG;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 26:
|
|
msr->idx = HV_X64_MSR_STIMER0_CONFIG;
|
|
msr->write = true;
|
|
msr->write_val = 0;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 27:
|
|
/* Direct mode test */
|
|
msr->idx = HV_X64_MSR_STIMER0_CONFIG;
|
|
msr->write = true;
|
|
msr->write_val = 1 << 12;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 28:
|
|
vcpu_set_cpuid_feature(vcpu, HV_STIMER_DIRECT_MODE_AVAILABLE);
|
|
msr->idx = HV_X64_MSR_STIMER0_CONFIG;
|
|
msr->write = true;
|
|
msr->write_val = 1 << 12;
|
|
msr->fault_expected = false;
|
|
break;
|
|
|
|
case 29:
|
|
msr->idx = HV_X64_MSR_EOI;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 30:
|
|
vcpu_set_cpuid_feature(vcpu, HV_MSR_APIC_ACCESS_AVAILABLE);
|
|
msr->idx = HV_X64_MSR_EOI;
|
|
msr->write = true;
|
|
msr->write_val = 1;
|
|
msr->fault_expected = false;
|
|
break;
|
|
|
|
case 31:
|
|
msr->idx = HV_X64_MSR_TSC_FREQUENCY;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 32:
|
|
vcpu_set_cpuid_feature(vcpu, HV_ACCESS_FREQUENCY_MSRS);
|
|
msr->idx = HV_X64_MSR_TSC_FREQUENCY;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 33:
|
|
/* Read only */
|
|
msr->idx = HV_X64_MSR_TSC_FREQUENCY;
|
|
msr->write = true;
|
|
msr->write_val = 1;
|
|
msr->fault_expected = true;
|
|
break;
|
|
|
|
case 34:
|
|
msr->idx = HV_X64_MSR_REENLIGHTENMENT_CONTROL;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 35:
|
|
vcpu_set_cpuid_feature(vcpu, HV_ACCESS_REENLIGHTENMENT);
|
|
msr->idx = HV_X64_MSR_REENLIGHTENMENT_CONTROL;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 36:
|
|
msr->idx = HV_X64_MSR_REENLIGHTENMENT_CONTROL;
|
|
msr->write = true;
|
|
msr->write_val = 1;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 37:
|
|
/* Can only write '0' */
|
|
msr->idx = HV_X64_MSR_TSC_EMULATION_STATUS;
|
|
msr->write = true;
|
|
msr->write_val = 1;
|
|
msr->fault_expected = true;
|
|
break;
|
|
|
|
case 38:
|
|
msr->idx = HV_X64_MSR_CRASH_P0;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 39:
|
|
vcpu_set_cpuid_feature(vcpu, HV_FEATURE_GUEST_CRASH_MSR_AVAILABLE);
|
|
msr->idx = HV_X64_MSR_CRASH_P0;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 40:
|
|
msr->idx = HV_X64_MSR_CRASH_P0;
|
|
msr->write = true;
|
|
msr->write_val = 1;
|
|
msr->fault_expected = false;
|
|
break;
|
|
|
|
case 41:
|
|
msr->idx = HV_X64_MSR_SYNDBG_STATUS;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 42:
|
|
vcpu_set_cpuid_feature(vcpu, HV_FEATURE_DEBUG_MSRS_AVAILABLE);
|
|
vcpu_set_cpuid_feature(vcpu, HV_X64_SYNDBG_CAP_ALLOW_KERNEL_DEBUGGING);
|
|
msr->idx = HV_X64_MSR_SYNDBG_STATUS;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 43:
|
|
msr->idx = HV_X64_MSR_SYNDBG_STATUS;
|
|
msr->write = true;
|
|
msr->write_val = 0;
|
|
msr->fault_expected = false;
|
|
break;
|
|
|
|
case 44:
|
|
/* MSR is not available when CPUID feature bit is unset */
|
|
if (!has_invtsc)
|
|
continue;
|
|
msr->idx = HV_X64_MSR_TSC_INVARIANT_CONTROL;
|
|
msr->write = false;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 45:
|
|
/* MSR is vailable when CPUID feature bit is set */
|
|
if (!has_invtsc)
|
|
continue;
|
|
vcpu_set_cpuid_feature(vcpu, HV_ACCESS_TSC_INVARIANT);
|
|
msr->idx = HV_X64_MSR_TSC_INVARIANT_CONTROL;
|
|
msr->write = false;
|
|
msr->fault_expected = false;
|
|
break;
|
|
case 46:
|
|
/* Writing bits other than 0 is forbidden */
|
|
if (!has_invtsc)
|
|
continue;
|
|
msr->idx = HV_X64_MSR_TSC_INVARIANT_CONTROL;
|
|
msr->write = true;
|
|
msr->write_val = 0xdeadbeef;
|
|
msr->fault_expected = true;
|
|
break;
|
|
case 47:
|
|
/* Setting bit 0 enables the feature */
|
|
if (!has_invtsc)
|
|
continue;
|
|
msr->idx = HV_X64_MSR_TSC_INVARIANT_CONTROL;
|
|
msr->write = true;
|
|
msr->write_val = 1;
|
|
msr->fault_expected = false;
|
|
break;
|
|
|
|
default:
|
|
kvm_vm_free(vm);
|
|
return;
|
|
}
|
|
|
|
vcpu_set_cpuid(vcpu);
|
|
|
|
memcpy(prev_cpuid, vcpu->cpuid, kvm_cpuid2_size(vcpu->cpuid->nent));
|
|
|
|
pr_debug("Stage %d: testing msr: 0x%x for %s\n", stage,
|
|
msr->idx, msr->write ? "write" : "read");
|
|
|
|
vcpu_run(vcpu);
|
|
TEST_ASSERT_KVM_EXIT_REASON(vcpu, KVM_EXIT_IO);
|
|
|
|
switch (get_ucall(vcpu, &uc)) {
|
|
case UCALL_ABORT:
|
|
REPORT_GUEST_ASSERT_3(uc, "MSR = %lx, arg1 = %lx, arg2 = %lx");
|
|
return;
|
|
case UCALL_DONE:
|
|
break;
|
|
default:
|
|
TEST_FAIL("Unhandled ucall: %ld", uc.cmd);
|
|
return;
|
|
}
|
|
|
|
stage++;
|
|
kvm_vm_free(vm);
|
|
}
|
|
}
|
|
|
|
static void guest_test_hcalls_access(void)
|
|
{
|
|
struct kvm_cpuid2 *prev_cpuid = NULL;
|
|
struct kvm_vcpu *vcpu;
|
|
struct kvm_vm *vm;
|
|
struct ucall uc;
|
|
int stage = 0;
|
|
vm_vaddr_t hcall_page, hcall_params;
|
|
struct hcall_data *hcall;
|
|
|
|
while (true) {
|
|
vm = vm_create_with_one_vcpu(&vcpu, guest_hcall);
|
|
|
|
vm_init_descriptor_tables(vm);
|
|
vcpu_init_descriptor_tables(vcpu);
|
|
|
|
/* Hypercall input/output */
|
|
hcall_page = vm_vaddr_alloc_pages(vm, 2);
|
|
memset(addr_gva2hva(vm, hcall_page), 0x0, 2 * getpagesize());
|
|
|
|
hcall_params = vm_vaddr_alloc_page(vm);
|
|
memset(addr_gva2hva(vm, hcall_params), 0x0, getpagesize());
|
|
hcall = addr_gva2hva(vm, hcall_params);
|
|
|
|
vcpu_args_set(vcpu, 2, addr_gva2gpa(vm, hcall_page), hcall_params);
|
|
vcpu_enable_cap(vcpu, KVM_CAP_HYPERV_ENFORCE_CPUID, 1);
|
|
|
|
if (!prev_cpuid) {
|
|
vcpu_reset_hv_cpuid(vcpu);
|
|
|
|
prev_cpuid = allocate_kvm_cpuid2(vcpu->cpuid->nent);
|
|
} else {
|
|
vcpu_init_cpuid(vcpu, prev_cpuid);
|
|
}
|
|
|
|
switch (stage) {
|
|
case 0:
|
|
vcpu_set_cpuid_feature(vcpu, HV_MSR_HYPERCALL_AVAILABLE);
|
|
hcall->control = 0xbeef;
|
|
hcall->expect = HV_STATUS_INVALID_HYPERCALL_CODE;
|
|
break;
|
|
|
|
case 1:
|
|
hcall->control = HVCALL_POST_MESSAGE;
|
|
hcall->expect = HV_STATUS_ACCESS_DENIED;
|
|
break;
|
|
case 2:
|
|
vcpu_set_cpuid_feature(vcpu, HV_POST_MESSAGES);
|
|
hcall->control = HVCALL_POST_MESSAGE;
|
|
hcall->expect = HV_STATUS_INVALID_HYPERCALL_INPUT;
|
|
break;
|
|
|
|
case 3:
|
|
hcall->control = HVCALL_SIGNAL_EVENT;
|
|
hcall->expect = HV_STATUS_ACCESS_DENIED;
|
|
break;
|
|
case 4:
|
|
vcpu_set_cpuid_feature(vcpu, HV_SIGNAL_EVENTS);
|
|
hcall->control = HVCALL_SIGNAL_EVENT;
|
|
hcall->expect = HV_STATUS_INVALID_HYPERCALL_INPUT;
|
|
break;
|
|
|
|
case 5:
|
|
hcall->control = HVCALL_RESET_DEBUG_SESSION;
|
|
hcall->expect = HV_STATUS_INVALID_HYPERCALL_CODE;
|
|
break;
|
|
case 6:
|
|
vcpu_set_cpuid_feature(vcpu, HV_X64_SYNDBG_CAP_ALLOW_KERNEL_DEBUGGING);
|
|
hcall->control = HVCALL_RESET_DEBUG_SESSION;
|
|
hcall->expect = HV_STATUS_ACCESS_DENIED;
|
|
break;
|
|
case 7:
|
|
vcpu_set_cpuid_feature(vcpu, HV_DEBUGGING);
|
|
hcall->control = HVCALL_RESET_DEBUG_SESSION;
|
|
hcall->expect = HV_STATUS_OPERATION_DENIED;
|
|
break;
|
|
|
|
case 8:
|
|
hcall->control = HVCALL_FLUSH_VIRTUAL_ADDRESS_SPACE;
|
|
hcall->expect = HV_STATUS_ACCESS_DENIED;
|
|
break;
|
|
case 9:
|
|
vcpu_set_cpuid_feature(vcpu, HV_X64_REMOTE_TLB_FLUSH_RECOMMENDED);
|
|
hcall->control = HVCALL_FLUSH_VIRTUAL_ADDRESS_SPACE;
|
|
hcall->expect = HV_STATUS_SUCCESS;
|
|
break;
|
|
case 10:
|
|
hcall->control = HVCALL_FLUSH_VIRTUAL_ADDRESS_SPACE_EX;
|
|
hcall->expect = HV_STATUS_ACCESS_DENIED;
|
|
break;
|
|
case 11:
|
|
vcpu_set_cpuid_feature(vcpu, HV_X64_EX_PROCESSOR_MASKS_RECOMMENDED);
|
|
hcall->control = HVCALL_FLUSH_VIRTUAL_ADDRESS_SPACE_EX;
|
|
hcall->expect = HV_STATUS_SUCCESS;
|
|
break;
|
|
|
|
case 12:
|
|
hcall->control = HVCALL_SEND_IPI;
|
|
hcall->expect = HV_STATUS_ACCESS_DENIED;
|
|
break;
|
|
case 13:
|
|
vcpu_set_cpuid_feature(vcpu, HV_X64_CLUSTER_IPI_RECOMMENDED);
|
|
hcall->control = HVCALL_SEND_IPI;
|
|
hcall->expect = HV_STATUS_INVALID_HYPERCALL_INPUT;
|
|
break;
|
|
case 14:
|
|
/* Nothing in 'sparse banks' -> success */
|
|
hcall->control = HVCALL_SEND_IPI_EX;
|
|
hcall->expect = HV_STATUS_SUCCESS;
|
|
break;
|
|
|
|
case 15:
|
|
hcall->control = HVCALL_NOTIFY_LONG_SPIN_WAIT;
|
|
hcall->expect = HV_STATUS_ACCESS_DENIED;
|
|
break;
|
|
case 16:
|
|
vcpu_set_cpuid_feature(vcpu, HV_PV_SPINLOCKS_TEST);
|
|
hcall->control = HVCALL_NOTIFY_LONG_SPIN_WAIT;
|
|
hcall->expect = HV_STATUS_SUCCESS;
|
|
break;
|
|
case 17:
|
|
/* XMM fast hypercall */
|
|
hcall->control = HVCALL_FLUSH_VIRTUAL_ADDRESS_SPACE | HV_HYPERCALL_FAST_BIT;
|
|
hcall->ud_expected = true;
|
|
break;
|
|
case 18:
|
|
vcpu_set_cpuid_feature(vcpu, HV_X64_HYPERCALL_XMM_INPUT_AVAILABLE);
|
|
hcall->control = HVCALL_FLUSH_VIRTUAL_ADDRESS_SPACE | HV_HYPERCALL_FAST_BIT;
|
|
hcall->ud_expected = false;
|
|
hcall->expect = HV_STATUS_SUCCESS;
|
|
break;
|
|
case 19:
|
|
hcall->control = HV_EXT_CALL_QUERY_CAPABILITIES;
|
|
hcall->expect = HV_STATUS_ACCESS_DENIED;
|
|
break;
|
|
case 20:
|
|
vcpu_set_cpuid_feature(vcpu, HV_ENABLE_EXTENDED_HYPERCALLS);
|
|
hcall->control = HV_EXT_CALL_QUERY_CAPABILITIES | HV_HYPERCALL_FAST_BIT;
|
|
hcall->expect = HV_STATUS_INVALID_PARAMETER;
|
|
break;
|
|
case 21:
|
|
kvm_vm_free(vm);
|
|
return;
|
|
}
|
|
|
|
vcpu_set_cpuid(vcpu);
|
|
|
|
memcpy(prev_cpuid, vcpu->cpuid, kvm_cpuid2_size(vcpu->cpuid->nent));
|
|
|
|
pr_debug("Stage %d: testing hcall: 0x%lx\n", stage, hcall->control);
|
|
|
|
vcpu_run(vcpu);
|
|
TEST_ASSERT_KVM_EXIT_REASON(vcpu, KVM_EXIT_IO);
|
|
|
|
switch (get_ucall(vcpu, &uc)) {
|
|
case UCALL_ABORT:
|
|
REPORT_GUEST_ASSERT_2(uc, "arg1 = %lx, arg2 = %lx");
|
|
return;
|
|
case UCALL_DONE:
|
|
break;
|
|
default:
|
|
TEST_FAIL("Unhandled ucall: %ld", uc.cmd);
|
|
return;
|
|
}
|
|
|
|
stage++;
|
|
kvm_vm_free(vm);
|
|
}
|
|
}
|
|
|
|
int main(void)
|
|
{
|
|
pr_info("Testing access to Hyper-V specific MSRs\n");
|
|
guest_test_msrs_access();
|
|
|
|
pr_info("Testing access to Hyper-V hypercalls\n");
|
|
guest_test_hcalls_access();
|
|
}
|