Commit 86f5201d authored by Chao Peng's avatar Chao Peng Committed by Paolo Bonzini

KVM: x86: Add Intel Processor Trace cpuid emulation

Expose Intel Processor Trace to guest only when
the PT works in Host-Guest mode.
Signed-off-by: default avatarChao Peng <chao.p.peng@linux.intel.com>
Signed-off-by: default avatarLuwei Kang <luwei.kang@intel.com>
Signed-off-by: default avatarPaolo Bonzini <pbonzini@redhat.com>
parent f99e3daf
...@@ -1105,6 +1105,7 @@ struct kvm_x86_ops { ...@@ -1105,6 +1105,7 @@ struct kvm_x86_ops {
bool (*mpx_supported)(void); bool (*mpx_supported)(void);
bool (*xsaves_supported)(void); bool (*xsaves_supported)(void);
bool (*umip_emulated)(void); bool (*umip_emulated)(void);
bool (*pt_supported)(void);
int (*check_nested_events)(struct kvm_vcpu *vcpu, bool external_intr); int (*check_nested_events)(struct kvm_vcpu *vcpu, bool external_intr);
void (*request_immediate_exit)(struct kvm_vcpu *vcpu); void (*request_immediate_exit)(struct kvm_vcpu *vcpu);
......
...@@ -334,6 +334,7 @@ static inline int __do_cpuid_ent(struct kvm_cpuid_entry2 *entry, u32 function, ...@@ -334,6 +334,7 @@ static inline int __do_cpuid_ent(struct kvm_cpuid_entry2 *entry, u32 function,
unsigned f_mpx = kvm_mpx_supported() ? F(MPX) : 0; unsigned f_mpx = kvm_mpx_supported() ? F(MPX) : 0;
unsigned f_xsaves = kvm_x86_ops->xsaves_supported() ? F(XSAVES) : 0; unsigned f_xsaves = kvm_x86_ops->xsaves_supported() ? F(XSAVES) : 0;
unsigned f_umip = kvm_x86_ops->umip_emulated() ? F(UMIP) : 0; unsigned f_umip = kvm_x86_ops->umip_emulated() ? F(UMIP) : 0;
unsigned f_intel_pt = kvm_x86_ops->pt_supported() ? F(INTEL_PT) : 0;
/* cpuid 1.edx */ /* cpuid 1.edx */
const u32 kvm_cpuid_1_edx_x86_features = const u32 kvm_cpuid_1_edx_x86_features =
...@@ -392,7 +393,7 @@ static inline int __do_cpuid_ent(struct kvm_cpuid_entry2 *entry, u32 function, ...@@ -392,7 +393,7 @@ static inline int __do_cpuid_ent(struct kvm_cpuid_entry2 *entry, u32 function,
F(BMI2) | F(ERMS) | f_invpcid | F(RTM) | f_mpx | F(RDSEED) | F(BMI2) | F(ERMS) | f_invpcid | F(RTM) | f_mpx | F(RDSEED) |
F(ADX) | F(SMAP) | F(AVX512IFMA) | F(AVX512F) | F(AVX512PF) | F(ADX) | F(SMAP) | F(AVX512IFMA) | F(AVX512F) | F(AVX512PF) |
F(AVX512ER) | F(AVX512CD) | F(CLFLUSHOPT) | F(CLWB) | F(AVX512DQ) | F(AVX512ER) | F(AVX512CD) | F(CLFLUSHOPT) | F(CLWB) | F(AVX512DQ) |
F(SHA_NI) | F(AVX512BW) | F(AVX512VL); F(SHA_NI) | F(AVX512BW) | F(AVX512VL) | f_intel_pt;
/* cpuid 0xD.1.eax */ /* cpuid 0xD.1.eax */
const u32 kvm_cpuid_D_1_eax_x86_features = const u32 kvm_cpuid_D_1_eax_x86_features =
...@@ -423,7 +424,7 @@ static inline int __do_cpuid_ent(struct kvm_cpuid_entry2 *entry, u32 function, ...@@ -423,7 +424,7 @@ static inline int __do_cpuid_ent(struct kvm_cpuid_entry2 *entry, u32 function,
switch (function) { switch (function) {
case 0: case 0:
entry->eax = min(entry->eax, (u32)0xd); entry->eax = min(entry->eax, (u32)(f_intel_pt ? 0x14 : 0xd));
break; break;
case 1: case 1:
entry->edx &= kvm_cpuid_1_edx_x86_features; entry->edx &= kvm_cpuid_1_edx_x86_features;
...@@ -600,6 +601,23 @@ static inline int __do_cpuid_ent(struct kvm_cpuid_entry2 *entry, u32 function, ...@@ -600,6 +601,23 @@ static inline int __do_cpuid_ent(struct kvm_cpuid_entry2 *entry, u32 function,
} }
break; break;
} }
/* Intel PT */
case 0x14: {
int t, times = entry->eax;
if (!f_intel_pt)
break;
entry->flags |= KVM_CPUID_FLAG_SIGNIFCANT_INDEX;
for (t = 1; t <= times; ++t) {
if (*nent >= maxnent)
goto out;
do_cpuid_1_ent(&entry[t], function, t);
entry[t].flags |= KVM_CPUID_FLAG_SIGNIFCANT_INDEX;
++*nent;
}
break;
}
case KVM_CPUID_SIGNATURE: { case KVM_CPUID_SIGNATURE: {
static const char signature[12] = "KVMKVMKVM\0\0"; static const char signature[12] = "KVMKVMKVM\0\0";
const u32 *sigptr = (const u32 *)signature; const u32 *sigptr = (const u32 *)signature;
......
...@@ -5947,6 +5947,11 @@ static bool svm_umip_emulated(void) ...@@ -5947,6 +5947,11 @@ static bool svm_umip_emulated(void)
return false; return false;
} }
static bool svm_pt_supported(void)
{
return false;
}
static bool svm_has_wbinvd_exit(void) static bool svm_has_wbinvd_exit(void)
{ {
return true; return true;
...@@ -7188,6 +7193,7 @@ static struct kvm_x86_ops svm_x86_ops __ro_after_init = { ...@@ -7188,6 +7193,7 @@ static struct kvm_x86_ops svm_x86_ops __ro_after_init = {
.mpx_supported = svm_mpx_supported, .mpx_supported = svm_mpx_supported,
.xsaves_supported = svm_xsaves_supported, .xsaves_supported = svm_xsaves_supported,
.umip_emulated = svm_umip_emulated, .umip_emulated = svm_umip_emulated,
.pt_supported = svm_pt_supported,
.set_supported_cpuid = svm_set_supported_cpuid, .set_supported_cpuid = svm_set_supported_cpuid,
......
...@@ -5886,6 +5886,11 @@ static bool vmx_has_emulated_msr(int index) ...@@ -5886,6 +5886,11 @@ static bool vmx_has_emulated_msr(int index)
} }
} }
static bool vmx_pt_supported(void)
{
return pt_mode == PT_MODE_HOST_GUEST;
}
static void vmx_recover_nmi_blocking(struct vcpu_vmx *vmx) static void vmx_recover_nmi_blocking(struct vcpu_vmx *vmx)
{ {
u32 exit_intr_info; u32 exit_intr_info;
...@@ -7399,6 +7404,7 @@ static struct kvm_x86_ops vmx_x86_ops __ro_after_init = { ...@@ -7399,6 +7404,7 @@ static struct kvm_x86_ops vmx_x86_ops __ro_after_init = {
.mpx_supported = vmx_mpx_supported, .mpx_supported = vmx_mpx_supported,
.xsaves_supported = vmx_xsaves_supported, .xsaves_supported = vmx_xsaves_supported,
.umip_emulated = vmx_umip_emulated, .umip_emulated = vmx_umip_emulated,
.pt_supported = vmx_pt_supported,
.request_immediate_exit = vmx_request_immediate_exit, .request_immediate_exit = vmx_request_immediate_exit,
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment