Commit 40bbb9d0 authored by Sean Christopherson's avatar Sean Christopherson Committed by Paolo Bonzini

KVM: VMX: add struct kvm_vmx to hold VMX specific KVM vars

Add struct kvm_vmx, which wraps struct kvm, and a helper to_kvm_vmx()
that retrieves 'struct kvm_vmx *' from 'struct kvm *'.  Move the VMX
specific variables out of kvm_arch and into kvm_vmx.
Signed-off-by: default avatarSean Christopherson <sean.j.christopherson@intel.com>
Signed-off-by: default avatarPaolo Bonzini <pbonzini@redhat.com>
parent 2ac52ab8
...@@ -808,7 +808,6 @@ struct kvm_arch { ...@@ -808,7 +808,6 @@ struct kvm_arch {
struct mutex apic_map_lock; struct mutex apic_map_lock;
struct kvm_apic_map *apic_map; struct kvm_apic_map *apic_map;
unsigned int tss_addr;
bool apic_access_page_done; bool apic_access_page_done;
gpa_t wall_clock; gpa_t wall_clock;
...@@ -817,9 +816,6 @@ struct kvm_arch { ...@@ -817,9 +816,6 @@ struct kvm_arch {
bool hlt_in_guest; bool hlt_in_guest;
bool pause_in_guest; bool pause_in_guest;
bool ept_identity_pagetable_done;
gpa_t ept_identity_map_addr;
unsigned long irq_sources_bitmap; unsigned long irq_sources_bitmap;
s64 kvmclock_offset; s64 kvmclock_offset;
raw_spinlock_t tsc_write_lock; raw_spinlock_t tsc_write_lock;
......
...@@ -196,6 +196,14 @@ module_param(ple_window_max, int, S_IRUGO); ...@@ -196,6 +196,14 @@ module_param(ple_window_max, int, S_IRUGO);
extern const ulong vmx_return; extern const ulong vmx_return;
struct kvm_vmx {
struct kvm kvm;
unsigned int tss_addr;
bool ept_identity_pagetable_done;
gpa_t ept_identity_map_addr;
};
#define NR_AUTOLOAD_MSRS 8 #define NR_AUTOLOAD_MSRS 8
struct vmcs { struct vmcs {
...@@ -698,6 +706,11 @@ enum segment_cache_field { ...@@ -698,6 +706,11 @@ enum segment_cache_field {
SEG_FIELD_NR = 4 SEG_FIELD_NR = 4
}; };
static inline struct kvm_vmx *to_kvm_vmx(struct kvm *kvm)
{
return container_of(kvm, struct kvm_vmx, kvm);
}
static inline struct vcpu_vmx *to_vmx(struct kvm_vcpu *vcpu) static inline struct vcpu_vmx *to_vmx(struct kvm_vcpu *vcpu)
{ {
return container_of(vcpu, struct vcpu_vmx, vcpu); return container_of(vcpu, struct vcpu_vmx, vcpu);
...@@ -4219,6 +4232,7 @@ static void enter_rmode(struct kvm_vcpu *vcpu) ...@@ -4219,6 +4232,7 @@ static void enter_rmode(struct kvm_vcpu *vcpu)
{ {
unsigned long flags; unsigned long flags;
struct vcpu_vmx *vmx = to_vmx(vcpu); struct vcpu_vmx *vmx = to_vmx(vcpu);
struct kvm_vmx *kvm_vmx = to_kvm_vmx(vcpu->kvm);
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_TR], VCPU_SREG_TR); vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_TR], VCPU_SREG_TR);
vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_ES], VCPU_SREG_ES); vmx_get_segment(vcpu, &vmx->rmode.segs[VCPU_SREG_ES], VCPU_SREG_ES);
...@@ -4234,13 +4248,13 @@ static void enter_rmode(struct kvm_vcpu *vcpu) ...@@ -4234,13 +4248,13 @@ static void enter_rmode(struct kvm_vcpu *vcpu)
* Very old userspace does not call KVM_SET_TSS_ADDR before entering * Very old userspace does not call KVM_SET_TSS_ADDR before entering
* vcpu. Warn the user that an update is overdue. * vcpu. Warn the user that an update is overdue.
*/ */
if (!vcpu->kvm->arch.tss_addr) if (!kvm_vmx->tss_addr)
printk_once(KERN_WARNING "kvm: KVM_SET_TSS_ADDR need to be " printk_once(KERN_WARNING "kvm: KVM_SET_TSS_ADDR need to be "
"called before entering vcpu\n"); "called before entering vcpu\n");
vmx_segment_cache_clear(vmx); vmx_segment_cache_clear(vmx);
vmcs_writel(GUEST_TR_BASE, vcpu->kvm->arch.tss_addr); vmcs_writel(GUEST_TR_BASE, kvm_vmx->tss_addr);
vmcs_write32(GUEST_TR_LIMIT, RMODE_TSS_SIZE - 1); vmcs_write32(GUEST_TR_LIMIT, RMODE_TSS_SIZE - 1);
vmcs_write32(GUEST_TR_AR_BYTES, 0x008b); vmcs_write32(GUEST_TR_AR_BYTES, 0x008b);
...@@ -4530,7 +4544,7 @@ static void vmx_set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3) ...@@ -4530,7 +4544,7 @@ static void vmx_set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3)
is_guest_mode(vcpu)) is_guest_mode(vcpu))
guest_cr3 = kvm_read_cr3(vcpu); guest_cr3 = kvm_read_cr3(vcpu);
else else
guest_cr3 = vcpu->kvm->arch.ept_identity_map_addr; guest_cr3 = to_kvm_vmx(vcpu->kvm)->ept_identity_map_addr;
ept_load_pdptrs(vcpu); ept_load_pdptrs(vcpu);
} }
...@@ -4971,7 +4985,7 @@ static int init_rmode_tss(struct kvm *kvm) ...@@ -4971,7 +4985,7 @@ static int init_rmode_tss(struct kvm *kvm)
int idx, r; int idx, r;
idx = srcu_read_lock(&kvm->srcu); idx = srcu_read_lock(&kvm->srcu);
fn = kvm->arch.tss_addr >> PAGE_SHIFT; fn = to_kvm_vmx(kvm)->tss_addr >> PAGE_SHIFT;
r = kvm_clear_guest_page(kvm, fn, 0, PAGE_SIZE); r = kvm_clear_guest_page(kvm, fn, 0, PAGE_SIZE);
if (r < 0) if (r < 0)
goto out; goto out;
...@@ -4997,22 +5011,23 @@ static int init_rmode_tss(struct kvm *kvm) ...@@ -4997,22 +5011,23 @@ static int init_rmode_tss(struct kvm *kvm)
static int init_rmode_identity_map(struct kvm *kvm) static int init_rmode_identity_map(struct kvm *kvm)
{ {
struct kvm_vmx *kvm_vmx = to_kvm_vmx(kvm);
int i, idx, r = 0; int i, idx, r = 0;
kvm_pfn_t identity_map_pfn; kvm_pfn_t identity_map_pfn;
u32 tmp; u32 tmp;
/* Protect kvm->arch.ept_identity_pagetable_done. */ /* Protect kvm_vmx->ept_identity_pagetable_done. */
mutex_lock(&kvm->slots_lock); mutex_lock(&kvm->slots_lock);
if (likely(kvm->arch.ept_identity_pagetable_done)) if (likely(kvm_vmx->ept_identity_pagetable_done))
goto out2; goto out2;
if (!kvm->arch.ept_identity_map_addr) if (!kvm_vmx->ept_identity_map_addr)
kvm->arch.ept_identity_map_addr = VMX_EPT_IDENTITY_PAGETABLE_ADDR; kvm_vmx->ept_identity_map_addr = VMX_EPT_IDENTITY_PAGETABLE_ADDR;
identity_map_pfn = kvm->arch.ept_identity_map_addr >> PAGE_SHIFT; identity_map_pfn = kvm_vmx->ept_identity_map_addr >> PAGE_SHIFT;
r = __x86_set_memory_region(kvm, IDENTITY_PAGETABLE_PRIVATE_MEMSLOT, r = __x86_set_memory_region(kvm, IDENTITY_PAGETABLE_PRIVATE_MEMSLOT,
kvm->arch.ept_identity_map_addr, PAGE_SIZE); kvm_vmx->ept_identity_map_addr, PAGE_SIZE);
if (r < 0) if (r < 0)
goto out2; goto out2;
...@@ -5029,7 +5044,7 @@ static int init_rmode_identity_map(struct kvm *kvm) ...@@ -5029,7 +5044,7 @@ static int init_rmode_identity_map(struct kvm *kvm)
if (r < 0) if (r < 0)
goto out; goto out;
} }
kvm->arch.ept_identity_pagetable_done = true; kvm_vmx->ept_identity_pagetable_done = true;
out: out:
srcu_read_unlock(&kvm->srcu, idx); srcu_read_unlock(&kvm->srcu, idx);
...@@ -6106,13 +6121,13 @@ static int vmx_set_tss_addr(struct kvm *kvm, unsigned int addr) ...@@ -6106,13 +6121,13 @@ static int vmx_set_tss_addr(struct kvm *kvm, unsigned int addr)
PAGE_SIZE * 3); PAGE_SIZE * 3);
if (ret) if (ret)
return ret; return ret;
kvm->arch.tss_addr = addr; to_kvm_vmx(kvm)->tss_addr = addr;
return init_rmode_tss(kvm); return init_rmode_tss(kvm);
} }
static int vmx_set_identity_map_addr(struct kvm *kvm, u64 ident_addr) static int vmx_set_identity_map_addr(struct kvm *kvm, u64 ident_addr)
{ {
kvm->arch.ept_identity_map_addr = ident_addr; to_kvm_vmx(kvm)->ept_identity_map_addr = ident_addr;
return 0; return 0;
} }
...@@ -9771,12 +9786,13 @@ STACK_FRAME_NON_STANDARD(vmx_vcpu_run); ...@@ -9771,12 +9786,13 @@ STACK_FRAME_NON_STANDARD(vmx_vcpu_run);
static struct kvm *vmx_vm_alloc(void) static struct kvm *vmx_vm_alloc(void)
{ {
return kzalloc(sizeof(struct kvm), GFP_KERNEL); struct kvm_vmx *kvm_vmx = kzalloc(sizeof(struct kvm_vmx), GFP_KERNEL);
return &kvm_vmx->kvm;
} }
static void vmx_vm_free(struct kvm *kvm) static void vmx_vm_free(struct kvm *kvm)
{ {
kfree(kvm); kfree(to_kvm_vmx(kvm));
} }
static void vmx_switch_vmcs(struct kvm_vcpu *vcpu, struct loaded_vmcs *vmcs) static void vmx_switch_vmcs(struct kvm_vcpu *vcpu, struct loaded_vmcs *vmcs)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment