Commit 0f65dd70 authored by Avi Kivity's avatar Avi Kivity

KVM: x86 emulator: drop vcpu argument from memory read/write callbacks

Making the emulator caller agnostic.
Signed-off-by: default avatarAvi Kivity <avi@redhat.com>
parent 7295261c
...@@ -92,8 +92,9 @@ struct x86_emulate_ops { ...@@ -92,8 +92,9 @@ struct x86_emulate_ops {
* @val: [OUT] Value read from memory, zero-extended to 'u_long'. * @val: [OUT] Value read from memory, zero-extended to 'u_long'.
* @bytes: [IN ] Number of bytes to read from memory. * @bytes: [IN ] Number of bytes to read from memory.
*/ */
int (*read_std)(unsigned long addr, void *val, int (*read_std)(struct x86_emulate_ctxt *ctxt,
unsigned int bytes, struct kvm_vcpu *vcpu, unsigned long addr, void *val,
unsigned int bytes,
struct x86_exception *fault); struct x86_exception *fault);
/* /*
...@@ -103,8 +104,8 @@ struct x86_emulate_ops { ...@@ -103,8 +104,8 @@ struct x86_emulate_ops {
* @val: [OUT] Value write to memory, zero-extended to 'u_long'. * @val: [OUT] Value write to memory, zero-extended to 'u_long'.
* @bytes: [IN ] Number of bytes to write to memory. * @bytes: [IN ] Number of bytes to write to memory.
*/ */
int (*write_std)(unsigned long addr, void *val, int (*write_std)(struct x86_emulate_ctxt *ctxt,
unsigned int bytes, struct kvm_vcpu *vcpu, unsigned long addr, void *val, unsigned int bytes,
struct x86_exception *fault); struct x86_exception *fault);
/* /*
* fetch: Read bytes of standard (non-emulated/special) memory. * fetch: Read bytes of standard (non-emulated/special) memory.
...@@ -113,8 +114,8 @@ struct x86_emulate_ops { ...@@ -113,8 +114,8 @@ struct x86_emulate_ops {
* @val: [OUT] Value read from memory, zero-extended to 'u_long'. * @val: [OUT] Value read from memory, zero-extended to 'u_long'.
* @bytes: [IN ] Number of bytes to read from memory. * @bytes: [IN ] Number of bytes to read from memory.
*/ */
int (*fetch)(unsigned long addr, void *val, int (*fetch)(struct x86_emulate_ctxt *ctxt,
unsigned int bytes, struct kvm_vcpu *vcpu, unsigned long addr, void *val, unsigned int bytes,
struct x86_exception *fault); struct x86_exception *fault);
/* /*
...@@ -123,11 +124,9 @@ struct x86_emulate_ops { ...@@ -123,11 +124,9 @@ struct x86_emulate_ops {
* @val: [OUT] Value read from memory, zero-extended to 'u_long'. * @val: [OUT] Value read from memory, zero-extended to 'u_long'.
* @bytes: [IN ] Number of bytes to read from memory. * @bytes: [IN ] Number of bytes to read from memory.
*/ */
int (*read_emulated)(unsigned long addr, int (*read_emulated)(struct x86_emulate_ctxt *ctxt,
void *val, unsigned long addr, void *val, unsigned int bytes,
unsigned int bytes, struct x86_exception *fault);
struct x86_exception *fault,
struct kvm_vcpu *vcpu);
/* /*
* write_emulated: Write bytes to emulated/special memory area. * write_emulated: Write bytes to emulated/special memory area.
...@@ -136,11 +135,10 @@ struct x86_emulate_ops { ...@@ -136,11 +135,10 @@ struct x86_emulate_ops {
* required). * required).
* @bytes: [IN ] Number of bytes to write to memory. * @bytes: [IN ] Number of bytes to write to memory.
*/ */
int (*write_emulated)(unsigned long addr, int (*write_emulated)(struct x86_emulate_ctxt *ctxt,
const void *val, unsigned long addr, const void *val,
unsigned int bytes, unsigned int bytes,
struct x86_exception *fault, struct x86_exception *fault);
struct kvm_vcpu *vcpu);
/* /*
* cmpxchg_emulated: Emulate an atomic (LOCKed) CMPXCHG operation on an * cmpxchg_emulated: Emulate an atomic (LOCKed) CMPXCHG operation on an
...@@ -150,12 +148,12 @@ struct x86_emulate_ops { ...@@ -150,12 +148,12 @@ struct x86_emulate_ops {
* @new: [IN ] Value to write to @addr. * @new: [IN ] Value to write to @addr.
* @bytes: [IN ] Number of bytes to access using CMPXCHG. * @bytes: [IN ] Number of bytes to access using CMPXCHG.
*/ */
int (*cmpxchg_emulated)(unsigned long addr, int (*cmpxchg_emulated)(struct x86_emulate_ctxt *ctxt,
unsigned long addr,
const void *old, const void *old,
const void *new, const void *new,
unsigned int bytes, unsigned int bytes,
struct x86_exception *fault, struct x86_exception *fault);
struct kvm_vcpu *vcpu);
int (*pio_in_emulated)(int size, unsigned short port, void *val, int (*pio_in_emulated)(int size, unsigned short port, void *val,
unsigned int count, struct kvm_vcpu *vcpu); unsigned int count, struct kvm_vcpu *vcpu);
......
...@@ -645,8 +645,7 @@ static int segmented_read_std(struct x86_emulate_ctxt *ctxt, ...@@ -645,8 +645,7 @@ static int segmented_read_std(struct x86_emulate_ctxt *ctxt,
rc = linearize(ctxt, addr, size, false, &linear); rc = linearize(ctxt, addr, size, false, &linear);
if (rc != X86EMUL_CONTINUE) if (rc != X86EMUL_CONTINUE)
return rc; return rc;
return ctxt->ops->read_std(linear, data, size, ctxt->vcpu, return ctxt->ops->read_std(ctxt, linear, data, size, &ctxt->exception);
&ctxt->exception);
} }
static int do_fetch_insn_byte(struct x86_emulate_ctxt *ctxt, static int do_fetch_insn_byte(struct x86_emulate_ctxt *ctxt,
...@@ -665,8 +664,8 @@ static int do_fetch_insn_byte(struct x86_emulate_ctxt *ctxt, ...@@ -665,8 +664,8 @@ static int do_fetch_insn_byte(struct x86_emulate_ctxt *ctxt,
rc = __linearize(ctxt, addr, size, false, true, &linear); rc = __linearize(ctxt, addr, size, false, true, &linear);
if (rc != X86EMUL_CONTINUE) if (rc != X86EMUL_CONTINUE)
return rc; return rc;
rc = ops->fetch(linear, fc->data + cur_size, rc = ops->fetch(ctxt, linear, fc->data + cur_size,
size, ctxt->vcpu, &ctxt->exception); size, &ctxt->exception);
if (rc != X86EMUL_CONTINUE) if (rc != X86EMUL_CONTINUE)
return rc; return rc;
fc->end += size; fc->end += size;
...@@ -1047,8 +1046,8 @@ static int read_emulated(struct x86_emulate_ctxt *ctxt, ...@@ -1047,8 +1046,8 @@ static int read_emulated(struct x86_emulate_ctxt *ctxt,
if (mc->pos < mc->end) if (mc->pos < mc->end)
goto read_cached; goto read_cached;
rc = ops->read_emulated(addr, mc->data + mc->end, n, rc = ops->read_emulated(ctxt, addr, mc->data + mc->end, n,
&ctxt->exception, ctxt->vcpu); &ctxt->exception);
if (rc != X86EMUL_CONTINUE) if (rc != X86EMUL_CONTINUE)
return rc; return rc;
mc->end += n; mc->end += n;
...@@ -1087,8 +1086,8 @@ static int segmented_write(struct x86_emulate_ctxt *ctxt, ...@@ -1087,8 +1086,8 @@ static int segmented_write(struct x86_emulate_ctxt *ctxt,
rc = linearize(ctxt, addr, size, true, &linear); rc = linearize(ctxt, addr, size, true, &linear);
if (rc != X86EMUL_CONTINUE) if (rc != X86EMUL_CONTINUE)
return rc; return rc;
return ctxt->ops->write_emulated(linear, data, size, return ctxt->ops->write_emulated(ctxt, linear, data, size,
&ctxt->exception, ctxt->vcpu); &ctxt->exception);
} }
static int segmented_cmpxchg(struct x86_emulate_ctxt *ctxt, static int segmented_cmpxchg(struct x86_emulate_ctxt *ctxt,
...@@ -1102,8 +1101,8 @@ static int segmented_cmpxchg(struct x86_emulate_ctxt *ctxt, ...@@ -1102,8 +1101,8 @@ static int segmented_cmpxchg(struct x86_emulate_ctxt *ctxt,
rc = linearize(ctxt, addr, size, true, &linear); rc = linearize(ctxt, addr, size, true, &linear);
if (rc != X86EMUL_CONTINUE) if (rc != X86EMUL_CONTINUE)
return rc; return rc;
return ctxt->ops->cmpxchg_emulated(linear, orig_data, data, return ctxt->ops->cmpxchg_emulated(ctxt, linear, orig_data, data,
size, &ctxt->exception, ctxt->vcpu); size, &ctxt->exception);
} }
static int pio_in_emulated(struct x86_emulate_ctxt *ctxt, static int pio_in_emulated(struct x86_emulate_ctxt *ctxt,
...@@ -1168,8 +1167,7 @@ static int read_segment_descriptor(struct x86_emulate_ctxt *ctxt, ...@@ -1168,8 +1167,7 @@ static int read_segment_descriptor(struct x86_emulate_ctxt *ctxt,
if (dt.size < index * 8 + 7) if (dt.size < index * 8 + 7)
return emulate_gp(ctxt, selector & 0xfffc); return emulate_gp(ctxt, selector & 0xfffc);
addr = dt.address + index * 8; addr = dt.address + index * 8;
ret = ops->read_std(addr, desc, sizeof *desc, ctxt->vcpu, ret = ops->read_std(ctxt, addr, desc, sizeof *desc, &ctxt->exception);
&ctxt->exception);
return ret; return ret;
} }
...@@ -1190,8 +1188,7 @@ static int write_segment_descriptor(struct x86_emulate_ctxt *ctxt, ...@@ -1190,8 +1188,7 @@ static int write_segment_descriptor(struct x86_emulate_ctxt *ctxt,
return emulate_gp(ctxt, selector & 0xfffc); return emulate_gp(ctxt, selector & 0xfffc);
addr = dt.address + index * 8; addr = dt.address + index * 8;
ret = ops->write_std(addr, desc, sizeof *desc, ctxt->vcpu, ret = ops->write_std(ctxt, addr, desc, sizeof *desc, &ctxt->exception);
&ctxt->exception);
return ret; return ret;
} }
...@@ -1545,11 +1542,11 @@ int emulate_int_real(struct x86_emulate_ctxt *ctxt, ...@@ -1545,11 +1542,11 @@ int emulate_int_real(struct x86_emulate_ctxt *ctxt,
eip_addr = dt.address + (irq << 2); eip_addr = dt.address + (irq << 2);
cs_addr = dt.address + (irq << 2) + 2; cs_addr = dt.address + (irq << 2) + 2;
rc = ops->read_std(cs_addr, &cs, 2, ctxt->vcpu, &ctxt->exception); rc = ops->read_std(ctxt, cs_addr, &cs, 2, &ctxt->exception);
if (rc != X86EMUL_CONTINUE) if (rc != X86EMUL_CONTINUE)
return rc; return rc;
rc = ops->read_std(eip_addr, &eip, 2, ctxt->vcpu, &ctxt->exception); rc = ops->read_std(ctxt, eip_addr, &eip, 2, &ctxt->exception);
if (rc != X86EMUL_CONTINUE) if (rc != X86EMUL_CONTINUE)
return rc; return rc;
...@@ -2036,13 +2033,12 @@ static bool emulator_io_port_access_allowed(struct x86_emulate_ctxt *ctxt, ...@@ -2036,13 +2033,12 @@ static bool emulator_io_port_access_allowed(struct x86_emulate_ctxt *ctxt,
#ifdef CONFIG_X86_64 #ifdef CONFIG_X86_64
base |= ((u64)base3) << 32; base |= ((u64)base3) << 32;
#endif #endif
r = ops->read_std(base + 102, &io_bitmap_ptr, 2, ctxt->vcpu, NULL); r = ops->read_std(ctxt, base + 102, &io_bitmap_ptr, 2, NULL);
if (r != X86EMUL_CONTINUE) if (r != X86EMUL_CONTINUE)
return false; return false;
if (io_bitmap_ptr + port/8 > desc_limit_scaled(&tr_seg)) if (io_bitmap_ptr + port/8 > desc_limit_scaled(&tr_seg))
return false; return false;
r = ops->read_std(base + io_bitmap_ptr + port/8, &perm, 2, ctxt->vcpu, r = ops->read_std(ctxt, base + io_bitmap_ptr + port/8, &perm, 2, NULL);
NULL);
if (r != X86EMUL_CONTINUE) if (r != X86EMUL_CONTINUE)
return false; return false;
if ((perm >> bit_idx) & mask) if ((perm >> bit_idx) & mask)
...@@ -2150,7 +2146,7 @@ static int task_switch_16(struct x86_emulate_ctxt *ctxt, ...@@ -2150,7 +2146,7 @@ static int task_switch_16(struct x86_emulate_ctxt *ctxt,
int ret; int ret;
u32 new_tss_base = get_desc_base(new_desc); u32 new_tss_base = get_desc_base(new_desc);
ret = ops->read_std(old_tss_base, &tss_seg, sizeof tss_seg, ctxt->vcpu, ret = ops->read_std(ctxt, old_tss_base, &tss_seg, sizeof tss_seg,
&ctxt->exception); &ctxt->exception);
if (ret != X86EMUL_CONTINUE) if (ret != X86EMUL_CONTINUE)
/* FIXME: need to provide precise fault address */ /* FIXME: need to provide precise fault address */
...@@ -2158,13 +2154,13 @@ static int task_switch_16(struct x86_emulate_ctxt *ctxt, ...@@ -2158,13 +2154,13 @@ static int task_switch_16(struct x86_emulate_ctxt *ctxt,
save_state_to_tss16(ctxt, ops, &tss_seg); save_state_to_tss16(ctxt, ops, &tss_seg);
ret = ops->write_std(old_tss_base, &tss_seg, sizeof tss_seg, ctxt->vcpu, ret = ops->write_std(ctxt, old_tss_base, &tss_seg, sizeof tss_seg,
&ctxt->exception); &ctxt->exception);
if (ret != X86EMUL_CONTINUE) if (ret != X86EMUL_CONTINUE)
/* FIXME: need to provide precise fault address */ /* FIXME: need to provide precise fault address */
return ret; return ret;
ret = ops->read_std(new_tss_base, &tss_seg, sizeof tss_seg, ctxt->vcpu, ret = ops->read_std(ctxt, new_tss_base, &tss_seg, sizeof tss_seg,
&ctxt->exception); &ctxt->exception);
if (ret != X86EMUL_CONTINUE) if (ret != X86EMUL_CONTINUE)
/* FIXME: need to provide precise fault address */ /* FIXME: need to provide precise fault address */
...@@ -2173,10 +2169,10 @@ static int task_switch_16(struct x86_emulate_ctxt *ctxt, ...@@ -2173,10 +2169,10 @@ static int task_switch_16(struct x86_emulate_ctxt *ctxt,
if (old_tss_sel != 0xffff) { if (old_tss_sel != 0xffff) {
tss_seg.prev_task_link = old_tss_sel; tss_seg.prev_task_link = old_tss_sel;
ret = ops->write_std(new_tss_base, ret = ops->write_std(ctxt, new_tss_base,
&tss_seg.prev_task_link, &tss_seg.prev_task_link,
sizeof tss_seg.prev_task_link, sizeof tss_seg.prev_task_link,
ctxt->vcpu, &ctxt->exception); &ctxt->exception);
if (ret != X86EMUL_CONTINUE) if (ret != X86EMUL_CONTINUE)
/* FIXME: need to provide precise fault address */ /* FIXME: need to provide precise fault address */
return ret; return ret;
...@@ -2282,7 +2278,7 @@ static int task_switch_32(struct x86_emulate_ctxt *ctxt, ...@@ -2282,7 +2278,7 @@ static int task_switch_32(struct x86_emulate_ctxt *ctxt,
int ret; int ret;
u32 new_tss_base = get_desc_base(new_desc); u32 new_tss_base = get_desc_base(new_desc);
ret = ops->read_std(old_tss_base, &tss_seg, sizeof tss_seg, ctxt->vcpu, ret = ops->read_std(ctxt, old_tss_base, &tss_seg, sizeof tss_seg,
&ctxt->exception); &ctxt->exception);
if (ret != X86EMUL_CONTINUE) if (ret != X86EMUL_CONTINUE)
/* FIXME: need to provide precise fault address */ /* FIXME: need to provide precise fault address */
...@@ -2290,13 +2286,13 @@ static int task_switch_32(struct x86_emulate_ctxt *ctxt, ...@@ -2290,13 +2286,13 @@ static int task_switch_32(struct x86_emulate_ctxt *ctxt,
save_state_to_tss32(ctxt, ops, &tss_seg); save_state_to_tss32(ctxt, ops, &tss_seg);
ret = ops->write_std(old_tss_base, &tss_seg, sizeof tss_seg, ctxt->vcpu, ret = ops->write_std(ctxt, old_tss_base, &tss_seg, sizeof tss_seg,
&ctxt->exception); &ctxt->exception);
if (ret != X86EMUL_CONTINUE) if (ret != X86EMUL_CONTINUE)
/* FIXME: need to provide precise fault address */ /* FIXME: need to provide precise fault address */
return ret; return ret;
ret = ops->read_std(new_tss_base, &tss_seg, sizeof tss_seg, ctxt->vcpu, ret = ops->read_std(ctxt, new_tss_base, &tss_seg, sizeof tss_seg,
&ctxt->exception); &ctxt->exception);
if (ret != X86EMUL_CONTINUE) if (ret != X86EMUL_CONTINUE)
/* FIXME: need to provide precise fault address */ /* FIXME: need to provide precise fault address */
...@@ -2305,10 +2301,10 @@ static int task_switch_32(struct x86_emulate_ctxt *ctxt, ...@@ -2305,10 +2301,10 @@ static int task_switch_32(struct x86_emulate_ctxt *ctxt,
if (old_tss_sel != 0xffff) { if (old_tss_sel != 0xffff) {
tss_seg.prev_task_link = old_tss_sel; tss_seg.prev_task_link = old_tss_sel;
ret = ops->write_std(new_tss_base, ret = ops->write_std(ctxt, new_tss_base,
&tss_seg.prev_task_link, &tss_seg.prev_task_link,
sizeof tss_seg.prev_task_link, sizeof tss_seg.prev_task_link,
ctxt->vcpu, &ctxt->exception); &ctxt->exception);
if (ret != X86EMUL_CONTINUE) if (ret != X86EMUL_CONTINUE)
/* FIXME: need to provide precise fault address */ /* FIXME: need to provide precise fault address */
return ret; return ret;
......
...@@ -63,6 +63,9 @@ ...@@ -63,6 +63,9 @@
#define KVM_MAX_MCE_BANKS 32 #define KVM_MAX_MCE_BANKS 32
#define KVM_MCE_CAP_SUPPORTED (MCG_CTL_P | MCG_SER_P) #define KVM_MCE_CAP_SUPPORTED (MCG_CTL_P | MCG_SER_P)
#define emul_to_vcpu(ctxt) \
container_of(ctxt, struct kvm_vcpu, arch.emulate_ctxt)
/* EFER defaults: /* EFER defaults:
* - enable syscall per default because its emulated by KVM * - enable syscall per default because its emulated by KVM
* - enable LME and LMA per default on 64 bit KVM * - enable LME and LMA per default on 64 bit KVM
...@@ -3760,37 +3763,43 @@ static int kvm_read_guest_virt_helper(gva_t addr, void *val, unsigned int bytes, ...@@ -3760,37 +3763,43 @@ static int kvm_read_guest_virt_helper(gva_t addr, void *val, unsigned int bytes,
} }
/* used for instruction fetching */ /* used for instruction fetching */
static int kvm_fetch_guest_virt(gva_t addr, void *val, unsigned int bytes, static int kvm_fetch_guest_virt(struct x86_emulate_ctxt *ctxt,
struct kvm_vcpu *vcpu, gva_t addr, void *val, unsigned int bytes,
struct x86_exception *exception) struct x86_exception *exception)
{ {
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
u32 access = (kvm_x86_ops->get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0; u32 access = (kvm_x86_ops->get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0;
return kvm_read_guest_virt_helper(addr, val, bytes, vcpu, return kvm_read_guest_virt_helper(addr, val, bytes, vcpu,
access | PFERR_FETCH_MASK, access | PFERR_FETCH_MASK,
exception); exception);
} }
static int kvm_read_guest_virt(gva_t addr, void *val, unsigned int bytes, static int kvm_read_guest_virt(struct x86_emulate_ctxt *ctxt,
struct kvm_vcpu *vcpu, gva_t addr, void *val, unsigned int bytes,
struct x86_exception *exception) struct x86_exception *exception)
{ {
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
u32 access = (kvm_x86_ops->get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0; u32 access = (kvm_x86_ops->get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0;
return kvm_read_guest_virt_helper(addr, val, bytes, vcpu, access, return kvm_read_guest_virt_helper(addr, val, bytes, vcpu, access,
exception); exception);
} }
static int kvm_read_guest_virt_system(gva_t addr, void *val, unsigned int bytes, static int kvm_read_guest_virt_system(struct x86_emulate_ctxt *ctxt,
struct kvm_vcpu *vcpu, gva_t addr, void *val, unsigned int bytes,
struct x86_exception *exception) struct x86_exception *exception)
{ {
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
return kvm_read_guest_virt_helper(addr, val, bytes, vcpu, 0, exception); return kvm_read_guest_virt_helper(addr, val, bytes, vcpu, 0, exception);
} }
static int kvm_write_guest_virt_system(gva_t addr, void *val, static int kvm_write_guest_virt_system(struct x86_emulate_ctxt *ctxt,
gva_t addr, void *val,
unsigned int bytes, unsigned int bytes,
struct kvm_vcpu *vcpu,
struct x86_exception *exception) struct x86_exception *exception)
{ {
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
void *data = val; void *data = val;
int r = X86EMUL_CONTINUE; int r = X86EMUL_CONTINUE;
...@@ -3818,12 +3827,13 @@ static int kvm_write_guest_virt_system(gva_t addr, void *val, ...@@ -3818,12 +3827,13 @@ static int kvm_write_guest_virt_system(gva_t addr, void *val,
return r; return r;
} }
static int emulator_read_emulated(unsigned long addr, static int emulator_read_emulated(struct x86_emulate_ctxt *ctxt,
unsigned long addr,
void *val, void *val,
unsigned int bytes, unsigned int bytes,
struct x86_exception *exception, struct x86_exception *exception)
struct kvm_vcpu *vcpu)
{ {
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
gpa_t gpa; gpa_t gpa;
int handled; int handled;
...@@ -3844,7 +3854,7 @@ static int emulator_read_emulated(unsigned long addr, ...@@ -3844,7 +3854,7 @@ static int emulator_read_emulated(unsigned long addr,
if ((gpa & PAGE_MASK) == APIC_DEFAULT_PHYS_BASE) if ((gpa & PAGE_MASK) == APIC_DEFAULT_PHYS_BASE)
goto mmio; goto mmio;
if (kvm_read_guest_virt(addr, val, bytes, vcpu, exception) if (kvm_read_guest_virt(ctxt, addr, val, bytes, exception)
== X86EMUL_CONTINUE) == X86EMUL_CONTINUE)
return X86EMUL_CONTINUE; return X86EMUL_CONTINUE;
...@@ -3933,12 +3943,14 @@ static int emulator_write_emulated_onepage(unsigned long addr, ...@@ -3933,12 +3943,14 @@ static int emulator_write_emulated_onepage(unsigned long addr,
return X86EMUL_CONTINUE; return X86EMUL_CONTINUE;
} }
int emulator_write_emulated(unsigned long addr, int emulator_write_emulated(struct x86_emulate_ctxt *ctxt,
unsigned long addr,
const void *val, const void *val,
unsigned int bytes, unsigned int bytes,
struct x86_exception *exception, struct x86_exception *exception)
struct kvm_vcpu *vcpu)
{ {
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
/* Crossing a page boundary? */ /* Crossing a page boundary? */
if (((addr + bytes - 1) ^ addr) & PAGE_MASK) { if (((addr + bytes - 1) ^ addr) & PAGE_MASK) {
int rc, now; int rc, now;
...@@ -3966,13 +3978,14 @@ int emulator_write_emulated(unsigned long addr, ...@@ -3966,13 +3978,14 @@ int emulator_write_emulated(unsigned long addr,
(cmpxchg64((u64 *)(ptr), *(u64 *)(old), *(u64 *)(new)) == *(u64 *)(old)) (cmpxchg64((u64 *)(ptr), *(u64 *)(old), *(u64 *)(new)) == *(u64 *)(old))
#endif #endif
static int emulator_cmpxchg_emulated(unsigned long addr, static int emulator_cmpxchg_emulated(struct x86_emulate_ctxt *ctxt,
unsigned long addr,
const void *old, const void *old,
const void *new, const void *new,
unsigned int bytes, unsigned int bytes,
struct x86_exception *exception, struct x86_exception *exception)
struct kvm_vcpu *vcpu)
{ {
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
gpa_t gpa; gpa_t gpa;
struct page *page; struct page *page;
char *kaddr; char *kaddr;
...@@ -4028,7 +4041,7 @@ static int emulator_cmpxchg_emulated(unsigned long addr, ...@@ -4028,7 +4041,7 @@ static int emulator_cmpxchg_emulated(unsigned long addr,
emul_write: emul_write:
printk_once(KERN_WARNING "kvm: emulating exchange as write\n"); printk_once(KERN_WARNING "kvm: emulating exchange as write\n");
return emulator_write_emulated(addr, new, bytes, exception, vcpu); return emulator_write_emulated(ctxt, addr, new, bytes, exception);
} }
static int kernel_pio(struct kvm_vcpu *vcpu, void *pd) static int kernel_pio(struct kvm_vcpu *vcpu, void *pd)
...@@ -5009,7 +5022,8 @@ int kvm_fix_hypercall(struct kvm_vcpu *vcpu) ...@@ -5009,7 +5022,8 @@ int kvm_fix_hypercall(struct kvm_vcpu *vcpu)
kvm_x86_ops->patch_hypercall(vcpu, instruction); kvm_x86_ops->patch_hypercall(vcpu, instruction);
return emulator_write_emulated(rip, instruction, 3, NULL, vcpu); return emulator_write_emulated(&vcpu->arch.emulate_ctxt,
rip, instruction, 3, NULL);
} }
void realmode_lgdt(struct kvm_vcpu *vcpu, u16 limit, unsigned long base) void realmode_lgdt(struct kvm_vcpu *vcpu, u16 limit, unsigned long base)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment