Commit d77c26fc authored by Mike Day's avatar Mike Day Committed by Avi Kivity

KVM: CodingStyle cleanup

Signed-off-by: default avatarMike D. Day <ncmike@ncultra.org>
Signed-off-by: default avatarAvi Kivity <avi@qumranet.com>
parent 7e620d16
...@@ -528,7 +528,7 @@ extern struct kvm_x86_ops *kvm_x86_ops; ...@@ -528,7 +528,7 @@ extern struct kvm_x86_ops *kvm_x86_ops;
if (printk_ratelimit()) \ if (printk_ratelimit()) \
printk(KERN_ERR "kvm: %i: cpu%i " fmt, \ printk(KERN_ERR "kvm: %i: cpu%i " fmt, \
current->tgid, (vcpu)->vcpu_id , ## __VA_ARGS__); \ current->tgid, (vcpu)->vcpu_id , ## __VA_ARGS__); \
} while(0) } while (0)
#define kvm_printf(kvm, fmt ...) printk(KERN_DEBUG fmt) #define kvm_printf(kvm, fmt ...) printk(KERN_DEBUG fmt)
#define vcpu_printf(vcpu, fmt...) kvm_printf(vcpu->kvm, fmt) #define vcpu_printf(vcpu, fmt...) kvm_printf(vcpu->kvm, fmt)
...@@ -598,7 +598,7 @@ int kvm_set_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data); ...@@ -598,7 +598,7 @@ int kvm_set_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data);
struct x86_emulate_ctxt; struct x86_emulate_ctxt;
int kvm_emulate_pio (struct kvm_vcpu *vcpu, struct kvm_run *run, int in, int kvm_emulate_pio(struct kvm_vcpu *vcpu, struct kvm_run *run, int in,
int size, unsigned port); int size, unsigned port);
int kvm_emulate_pio_string(struct kvm_vcpu *vcpu, struct kvm_run *run, int in, int kvm_emulate_pio_string(struct kvm_vcpu *vcpu, struct kvm_run *run, int in,
int size, unsigned long count, int down, int size, unsigned long count, int down,
...@@ -607,7 +607,7 @@ void kvm_emulate_cpuid(struct kvm_vcpu *vcpu); ...@@ -607,7 +607,7 @@ void kvm_emulate_cpuid(struct kvm_vcpu *vcpu);
int kvm_emulate_halt(struct kvm_vcpu *vcpu); int kvm_emulate_halt(struct kvm_vcpu *vcpu);
int emulate_invlpg(struct kvm_vcpu *vcpu, gva_t address); int emulate_invlpg(struct kvm_vcpu *vcpu, gva_t address);
int emulate_clts(struct kvm_vcpu *vcpu); int emulate_clts(struct kvm_vcpu *vcpu);
int emulator_get_dr(struct x86_emulate_ctxt* ctxt, int dr, int emulator_get_dr(struct x86_emulate_ctxt *ctxt, int dr,
unsigned long *dest); unsigned long *dest);
int emulator_set_dr(struct x86_emulate_ctxt *ctxt, int dr, int emulator_set_dr(struct x86_emulate_ctxt *ctxt, int dr,
unsigned long value); unsigned long value);
...@@ -631,7 +631,7 @@ void kvm_put_guest_fpu(struct kvm_vcpu *vcpu); ...@@ -631,7 +631,7 @@ void kvm_put_guest_fpu(struct kvm_vcpu *vcpu);
void kvm_flush_remote_tlbs(struct kvm *kvm); void kvm_flush_remote_tlbs(struct kvm *kvm);
int emulator_read_std(unsigned long addr, int emulator_read_std(unsigned long addr,
void *val, void *val,
unsigned int bytes, unsigned int bytes,
struct kvm_vcpu *vcpu); struct kvm_vcpu *vcpu);
int emulator_write_emulated(unsigned long addr, int emulator_write_emulated(unsigned long addr,
...@@ -721,55 +721,55 @@ static inline struct kvm_mmu_page *page_header(hpa_t shadow_page) ...@@ -721,55 +721,55 @@ static inline struct kvm_mmu_page *page_header(hpa_t shadow_page)
static inline u16 read_fs(void) static inline u16 read_fs(void)
{ {
u16 seg; u16 seg;
asm ("mov %%fs, %0" : "=g"(seg)); asm("mov %%fs, %0" : "=g"(seg));
return seg; return seg;
} }
static inline u16 read_gs(void) static inline u16 read_gs(void)
{ {
u16 seg; u16 seg;
asm ("mov %%gs, %0" : "=g"(seg)); asm("mov %%gs, %0" : "=g"(seg));
return seg; return seg;
} }
static inline u16 read_ldt(void) static inline u16 read_ldt(void)
{ {
u16 ldt; u16 ldt;
asm ("sldt %0" : "=g"(ldt)); asm("sldt %0" : "=g"(ldt));
return ldt; return ldt;
} }
static inline void load_fs(u16 sel) static inline void load_fs(u16 sel)
{ {
asm ("mov %0, %%fs" : : "rm"(sel)); asm("mov %0, %%fs" : : "rm"(sel));
} }
static inline void load_gs(u16 sel) static inline void load_gs(u16 sel)
{ {
asm ("mov %0, %%gs" : : "rm"(sel)); asm("mov %0, %%gs" : : "rm"(sel));
} }
#ifndef load_ldt #ifndef load_ldt
static inline void load_ldt(u16 sel) static inline void load_ldt(u16 sel)
{ {
asm ("lldt %0" : : "rm"(sel)); asm("lldt %0" : : "rm"(sel));
} }
#endif #endif
static inline void get_idt(struct descriptor_table *table) static inline void get_idt(struct descriptor_table *table)
{ {
asm ("sidt %0" : "=m"(*table)); asm("sidt %0" : "=m"(*table));
} }
static inline void get_gdt(struct descriptor_table *table) static inline void get_gdt(struct descriptor_table *table)
{ {
asm ("sgdt %0" : "=m"(*table)); asm("sgdt %0" : "=m"(*table));
} }
static inline unsigned long read_tr_base(void) static inline unsigned long read_tr_base(void)
{ {
u16 tr; u16 tr;
asm ("str %0" : "=g"(tr)); asm("str %0" : "=g"(tr));
return segment_base(tr); return segment_base(tr);
} }
...@@ -785,17 +785,17 @@ static inline unsigned long read_msr(unsigned long msr) ...@@ -785,17 +785,17 @@ static inline unsigned long read_msr(unsigned long msr)
static inline void fx_save(struct i387_fxsave_struct *image) static inline void fx_save(struct i387_fxsave_struct *image)
{ {
asm ("fxsave (%0)":: "r" (image)); asm("fxsave (%0)":: "r" (image));
} }
static inline void fx_restore(struct i387_fxsave_struct *image) static inline void fx_restore(struct i387_fxsave_struct *image)
{ {
asm ("fxrstor (%0)":: "r" (image)); asm("fxrstor (%0)":: "r" (image));
} }
static inline void fpu_init(void) static inline void fpu_init(void)
{ {
asm ("finit"); asm("finit");
} }
static inline u32 get_rdx_init_val(void) static inline u32 get_rdx_init_val(void)
......
...@@ -104,7 +104,7 @@ static struct dentry *debugfs_dir; ...@@ -104,7 +104,7 @@ static struct dentry *debugfs_dir;
#define EFER_RESERVED_BITS 0xfffffffffffff2fe #define EFER_RESERVED_BITS 0xfffffffffffff2fe
#ifdef CONFIG_X86_64 #ifdef CONFIG_X86_64
// LDT or TSS descriptor in the GDT. 16 bytes. /* LDT or TSS descriptor in the GDT. 16 bytes. */
struct segment_descriptor_64 { struct segment_descriptor_64 {
struct segment_descriptor s; struct segment_descriptor s;
u32 base_higher; u32 base_higher;
...@@ -121,27 +121,27 @@ unsigned long segment_base(u16 selector) ...@@ -121,27 +121,27 @@ unsigned long segment_base(u16 selector)
struct descriptor_table gdt; struct descriptor_table gdt;
struct segment_descriptor *d; struct segment_descriptor *d;
unsigned long table_base; unsigned long table_base;
typedef unsigned long ul;
unsigned long v; unsigned long v;
if (selector == 0) if (selector == 0)
return 0; return 0;
asm ("sgdt %0" : "=m"(gdt)); asm("sgdt %0" : "=m"(gdt));
table_base = gdt.base; table_base = gdt.base;
if (selector & 4) { /* from ldt */ if (selector & 4) { /* from ldt */
u16 ldt_selector; u16 ldt_selector;
asm ("sldt %0" : "=g"(ldt_selector)); asm("sldt %0" : "=g"(ldt_selector));
table_base = segment_base(ldt_selector); table_base = segment_base(ldt_selector);
} }
d = (struct segment_descriptor *)(table_base + (selector & ~7)); d = (struct segment_descriptor *)(table_base + (selector & ~7));
v = d->base_low | ((ul)d->base_mid << 16) | ((ul)d->base_high << 24); v = d->base_low | ((unsigned long)d->base_mid << 16) |
((unsigned long)d->base_high << 24);
#ifdef CONFIG_X86_64 #ifdef CONFIG_X86_64
if (d->system == 0 if (d->system == 0 && (d->type == 2 || d->type == 9 || d->type == 11))
&& (d->type == 2 || d->type == 9 || d->type == 11)) v |= ((unsigned long) \
v |= ((ul)((struct segment_descriptor_64 *)d)->base_higher) << 32; ((struct segment_descriptor_64 *)d)->base_higher) << 32;
#endif #endif
return v; return v;
} }
...@@ -721,7 +721,7 @@ static int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, ...@@ -721,7 +721,7 @@ static int kvm_vm_ioctl_set_memory_region(struct kvm *kvm,
if (!new.phys_mem) if (!new.phys_mem)
goto out_unlock; goto out_unlock;
new.rmap = vmalloc(npages * sizeof(struct page*)); new.rmap = vmalloc(npages * sizeof(struct page *));
if (!new.rmap) if (!new.rmap)
goto out_unlock; goto out_unlock;
...@@ -904,17 +904,17 @@ static int kvm_vm_ioctl_get_irqchip(struct kvm *kvm, struct kvm_irqchip *chip) ...@@ -904,17 +904,17 @@ static int kvm_vm_ioctl_get_irqchip(struct kvm *kvm, struct kvm_irqchip *chip)
r = 0; r = 0;
switch (chip->chip_id) { switch (chip->chip_id) {
case KVM_IRQCHIP_PIC_MASTER: case KVM_IRQCHIP_PIC_MASTER:
memcpy (&chip->chip.pic, memcpy(&chip->chip.pic,
&pic_irqchip(kvm)->pics[0], &pic_irqchip(kvm)->pics[0],
sizeof(struct kvm_pic_state)); sizeof(struct kvm_pic_state));
break; break;
case KVM_IRQCHIP_PIC_SLAVE: case KVM_IRQCHIP_PIC_SLAVE:
memcpy (&chip->chip.pic, memcpy(&chip->chip.pic,
&pic_irqchip(kvm)->pics[1], &pic_irqchip(kvm)->pics[1],
sizeof(struct kvm_pic_state)); sizeof(struct kvm_pic_state));
break; break;
case KVM_IRQCHIP_IOAPIC: case KVM_IRQCHIP_IOAPIC:
memcpy (&chip->chip.ioapic, memcpy(&chip->chip.ioapic,
ioapic_irqchip(kvm), ioapic_irqchip(kvm),
sizeof(struct kvm_ioapic_state)); sizeof(struct kvm_ioapic_state));
break; break;
...@@ -932,17 +932,17 @@ static int kvm_vm_ioctl_set_irqchip(struct kvm *kvm, struct kvm_irqchip *chip) ...@@ -932,17 +932,17 @@ static int kvm_vm_ioctl_set_irqchip(struct kvm *kvm, struct kvm_irqchip *chip)
r = 0; r = 0;
switch (chip->chip_id) { switch (chip->chip_id) {
case KVM_IRQCHIP_PIC_MASTER: case KVM_IRQCHIP_PIC_MASTER:
memcpy (&pic_irqchip(kvm)->pics[0], memcpy(&pic_irqchip(kvm)->pics[0],
&chip->chip.pic, &chip->chip.pic,
sizeof(struct kvm_pic_state)); sizeof(struct kvm_pic_state));
break; break;
case KVM_IRQCHIP_PIC_SLAVE: case KVM_IRQCHIP_PIC_SLAVE:
memcpy (&pic_irqchip(kvm)->pics[1], memcpy(&pic_irqchip(kvm)->pics[1],
&chip->chip.pic, &chip->chip.pic,
sizeof(struct kvm_pic_state)); sizeof(struct kvm_pic_state));
break; break;
case KVM_IRQCHIP_IOAPIC: case KVM_IRQCHIP_IOAPIC:
memcpy (ioapic_irqchip(kvm), memcpy(ioapic_irqchip(kvm),
&chip->chip.ioapic, &chip->chip.ioapic,
sizeof(struct kvm_ioapic_state)); sizeof(struct kvm_ioapic_state));
break; break;
...@@ -1341,7 +1341,7 @@ int emulate_clts(struct kvm_vcpu *vcpu) ...@@ -1341,7 +1341,7 @@ int emulate_clts(struct kvm_vcpu *vcpu)
return X86EMUL_CONTINUE; return X86EMUL_CONTINUE;
} }
int emulator_get_dr(struct x86_emulate_ctxt* ctxt, int dr, unsigned long *dest) int emulator_get_dr(struct x86_emulate_ctxt *ctxt, int dr, unsigned long *dest)
{ {
struct kvm_vcpu *vcpu = ctxt->vcpu; struct kvm_vcpu *vcpu = ctxt->vcpu;
...@@ -1934,7 +1934,7 @@ static void pio_string_write(struct kvm_io_device *pio_dev, ...@@ -1934,7 +1934,7 @@ static void pio_string_write(struct kvm_io_device *pio_dev,
mutex_unlock(&vcpu->kvm->lock); mutex_unlock(&vcpu->kvm->lock);
} }
int kvm_emulate_pio (struct kvm_vcpu *vcpu, struct kvm_run *run, int in, int kvm_emulate_pio(struct kvm_vcpu *vcpu, struct kvm_run *run, int in,
int size, unsigned port) int size, unsigned port)
{ {
struct kvm_io_device *pio_dev; struct kvm_io_device *pio_dev;
...@@ -2089,7 +2089,7 @@ static int __vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -2089,7 +2089,7 @@ static int __vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
int r; int r;
if (unlikely(vcpu->mp_state == VCPU_MP_STATE_SIPI_RECEIVED)) { if (unlikely(vcpu->mp_state == VCPU_MP_STATE_SIPI_RECEIVED)) {
printk("vcpu %d received sipi with vector # %x\n", pr_debug("vcpu %d received sipi with vector # %x\n",
vcpu->vcpu_id, vcpu->sipi_vector); vcpu->vcpu_id, vcpu->sipi_vector);
kvm_lapic_reset(vcpu); kvm_lapic_reset(vcpu);
kvm_x86_ops->vcpu_reset(vcpu); kvm_x86_ops->vcpu_reset(vcpu);
...@@ -2363,7 +2363,8 @@ static int kvm_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu, ...@@ -2363,7 +2363,8 @@ static int kvm_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu,
sizeof sregs->interrupt_bitmap); sizeof sregs->interrupt_bitmap);
pending_vec = kvm_x86_ops->get_irq(vcpu); pending_vec = kvm_x86_ops->get_irq(vcpu);
if (pending_vec >= 0) if (pending_vec >= 0)
set_bit(pending_vec, (unsigned long *)sregs->interrupt_bitmap); set_bit(pending_vec,
(unsigned long *)sregs->interrupt_bitmap);
} else } else
memcpy(sregs->interrupt_bitmap, vcpu->irq_pending, memcpy(sregs->interrupt_bitmap, vcpu->irq_pending,
sizeof sregs->interrupt_bitmap); sizeof sregs->interrupt_bitmap);
...@@ -2436,7 +2437,8 @@ static int kvm_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu, ...@@ -2436,7 +2437,8 @@ static int kvm_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu,
/* Only pending external irq is handled here */ /* Only pending external irq is handled here */
if (pending_vec < max_bits) { if (pending_vec < max_bits) {
kvm_x86_ops->set_irq(vcpu, pending_vec); kvm_x86_ops->set_irq(vcpu, pending_vec);
printk("Set back pending irq %d\n", pending_vec); pr_debug("Set back pending irq %d\n",
pending_vec);
} }
} }
...@@ -3155,8 +3157,7 @@ static long kvm_vm_ioctl(struct file *filp, ...@@ -3155,8 +3157,7 @@ static long kvm_vm_ioctl(struct file *filp,
kvm->vpic = NULL; kvm->vpic = NULL;
goto out; goto out;
} }
} } else
else
goto out; goto out;
break; break;
case KVM_IRQ_LINE: { case KVM_IRQ_LINE: {
...@@ -3448,7 +3449,7 @@ static int kvm_cpu_hotplug(struct notifier_block *notifier, unsigned long val, ...@@ -3448,7 +3449,7 @@ static int kvm_cpu_hotplug(struct notifier_block *notifier, unsigned long val,
} }
static int kvm_reboot(struct notifier_block *notifier, unsigned long val, static int kvm_reboot(struct notifier_block *notifier, unsigned long val,
void *v) void *v)
{ {
if (val == SYS_RESTART) { if (val == SYS_RESTART) {
/* /*
...@@ -3655,7 +3656,7 @@ int kvm_init_x86(struct kvm_x86_ops *ops, unsigned int vcpu_size, ...@@ -3655,7 +3656,7 @@ int kvm_init_x86(struct kvm_x86_ops *ops, unsigned int vcpu_size,
r = misc_register(&kvm_dev); r = misc_register(&kvm_dev);
if (r) { if (r) {
printk (KERN_ERR "kvm: misc device register failed\n"); printk(KERN_ERR "kvm: misc device register failed\n");
goto out_free; goto out_free;
} }
...@@ -3683,6 +3684,7 @@ out: ...@@ -3683,6 +3684,7 @@ out:
kvm_x86_ops = NULL; kvm_x86_ops = NULL;
return r; return r;
} }
EXPORT_SYMBOL_GPL(kvm_init_x86);
void kvm_exit_x86(void) void kvm_exit_x86(void)
{ {
...@@ -3696,6 +3698,7 @@ void kvm_exit_x86(void) ...@@ -3696,6 +3698,7 @@ void kvm_exit_x86(void)
kvm_x86_ops->hardware_unsetup(); kvm_x86_ops->hardware_unsetup();
kvm_x86_ops = NULL; kvm_x86_ops = NULL;
} }
EXPORT_SYMBOL_GPL(kvm_exit_x86);
static __init int kvm_init(void) static __init int kvm_init(void)
{ {
...@@ -3710,7 +3713,9 @@ static __init int kvm_init(void) ...@@ -3710,7 +3713,9 @@ static __init int kvm_init(void)
kvm_init_msr_list(); kvm_init_msr_list();
if ((bad_page = alloc_page(GFP_KERNEL)) == NULL) { bad_page = alloc_page(GFP_KERNEL);
if (bad_page == NULL) {
r = -ENOMEM; r = -ENOMEM;
goto out; goto out;
} }
...@@ -3736,6 +3741,3 @@ static __exit void kvm_exit(void) ...@@ -3736,6 +3741,3 @@ static __exit void kvm_exit(void)
module_init(kvm_init) module_init(kvm_init)
module_exit(kvm_exit) module_exit(kvm_exit)
EXPORT_SYMBOL_GPL(kvm_init_x86);
EXPORT_SYMBOL_GPL(kvm_exit_x86);
...@@ -906,8 +906,7 @@ static int __apic_timer_fn(struct kvm_lapic *apic) ...@@ -906,8 +906,7 @@ static int __apic_timer_fn(struct kvm_lapic *apic)
wait_queue_head_t *q = &apic->vcpu->wq; wait_queue_head_t *q = &apic->vcpu->wq;
atomic_inc(&apic->timer.pending); atomic_inc(&apic->timer.pending);
if (waitqueue_active(q)) if (waitqueue_active(q)) {
{
apic->vcpu->mp_state = VCPU_MP_STATE_RUNNABLE; apic->vcpu->mp_state = VCPU_MP_STATE_RUNNABLE;
wake_up_interruptible(q); wake_up_interruptible(q);
} }
......
...@@ -90,7 +90,8 @@ static int dbg = 1; ...@@ -90,7 +90,8 @@ static int dbg = 1;
#define PT32_DIR_PSE36_SIZE 4 #define PT32_DIR_PSE36_SIZE 4
#define PT32_DIR_PSE36_SHIFT 13 #define PT32_DIR_PSE36_SHIFT 13
#define PT32_DIR_PSE36_MASK (((1ULL << PT32_DIR_PSE36_SIZE) - 1) << PT32_DIR_PSE36_SHIFT) #define PT32_DIR_PSE36_MASK \
(((1ULL << PT32_DIR_PSE36_SIZE) - 1) << PT32_DIR_PSE36_SHIFT)
#define PT_FIRST_AVAIL_BITS_SHIFT 9 #define PT_FIRST_AVAIL_BITS_SHIFT 9
...@@ -103,7 +104,7 @@ static int dbg = 1; ...@@ -103,7 +104,7 @@ static int dbg = 1;
#define PT64_LEVEL_BITS 9 #define PT64_LEVEL_BITS 9
#define PT64_LEVEL_SHIFT(level) \ #define PT64_LEVEL_SHIFT(level) \
( PAGE_SHIFT + (level - 1) * PT64_LEVEL_BITS ) (PAGE_SHIFT + (level - 1) * PT64_LEVEL_BITS)
#define PT64_LEVEL_MASK(level) \ #define PT64_LEVEL_MASK(level) \
(((1ULL << PT64_LEVEL_BITS) - 1) << PT64_LEVEL_SHIFT(level)) (((1ULL << PT64_LEVEL_BITS) - 1) << PT64_LEVEL_SHIFT(level))
...@@ -115,7 +116,7 @@ static int dbg = 1; ...@@ -115,7 +116,7 @@ static int dbg = 1;
#define PT32_LEVEL_BITS 10 #define PT32_LEVEL_BITS 10
#define PT32_LEVEL_SHIFT(level) \ #define PT32_LEVEL_SHIFT(level) \
( PAGE_SHIFT + (level - 1) * PT32_LEVEL_BITS ) (PAGE_SHIFT + (level - 1) * PT32_LEVEL_BITS)
#define PT32_LEVEL_MASK(level) \ #define PT32_LEVEL_MASK(level) \
(((1ULL << PT32_LEVEL_BITS) - 1) << PT32_LEVEL_SHIFT(level)) (((1ULL << PT32_LEVEL_BITS) - 1) << PT32_LEVEL_SHIFT(level))
...@@ -1489,7 +1490,8 @@ static void audit_mappings_page(struct kvm_vcpu *vcpu, u64 page_pte, ...@@ -1489,7 +1490,8 @@ static void audit_mappings_page(struct kvm_vcpu *vcpu, u64 page_pte,
printk(KERN_ERR "xx audit error: (%s) levels %d" printk(KERN_ERR "xx audit error: (%s) levels %d"
" gva %lx gpa %llx hpa %llx ent %llx %d\n", " gva %lx gpa %llx hpa %llx ent %llx %d\n",
audit_msg, vcpu->mmu.root_level, audit_msg, vcpu->mmu.root_level,
va, gpa, hpa, ent, is_shadow_present_pte(ent)); va, gpa, hpa, ent,
is_shadow_present_pte(ent));
else if (ent == shadow_notrap_nonpresent_pte else if (ent == shadow_notrap_nonpresent_pte
&& !is_error_hpa(hpa)) && !is_error_hpa(hpa))
printk(KERN_ERR "audit: (%s) notrap shadow," printk(KERN_ERR "audit: (%s) notrap shadow,"
......
...@@ -163,7 +163,7 @@ static int FNAME(walk_addr)(struct guest_walker *walker, ...@@ -163,7 +163,7 @@ static int FNAME(walk_addr)(struct guest_walker *walker,
walker->page = pfn_to_page(paddr >> PAGE_SHIFT); walker->page = pfn_to_page(paddr >> PAGE_SHIFT);
walker->table = kmap_atomic(walker->page, KM_USER0); walker->table = kmap_atomic(walker->page, KM_USER0);
--walker->level; --walker->level;
walker->table_gfn[walker->level - 1 ] = table_gfn; walker->table_gfn[walker->level - 1] = table_gfn;
pgprintk("%s: table_gfn[%d] %lx\n", __FUNCTION__, pgprintk("%s: table_gfn[%d] %lx\n", __FUNCTION__,
walker->level - 1, table_gfn); walker->level - 1, table_gfn);
} }
......
...@@ -229,12 +229,11 @@ static void skip_emulated_instruction(struct kvm_vcpu *vcpu) ...@@ -229,12 +229,11 @@ static void skip_emulated_instruction(struct kvm_vcpu *vcpu)
printk(KERN_DEBUG "%s: NOP\n", __FUNCTION__); printk(KERN_DEBUG "%s: NOP\n", __FUNCTION__);
return; return;
} }
if (svm->next_rip - svm->vmcb->save.rip > MAX_INST_SIZE) { if (svm->next_rip - svm->vmcb->save.rip > MAX_INST_SIZE)
printk(KERN_ERR "%s: ip 0x%llx next 0x%llx\n", printk(KERN_ERR "%s: ip 0x%llx next 0x%llx\n",
__FUNCTION__, __FUNCTION__,
svm->vmcb->save.rip, svm->vmcb->save.rip,
svm->next_rip); svm->next_rip);
}
vcpu->rip = svm->vmcb->save.rip = svm->next_rip; vcpu->rip = svm->vmcb->save.rip = svm->next_rip;
svm->vmcb->control.int_state &= ~SVM_INTERRUPT_SHADOW_MASK; svm->vmcb->control.int_state &= ~SVM_INTERRUPT_SHADOW_MASK;
...@@ -312,7 +311,7 @@ static void svm_hardware_enable(void *garbage) ...@@ -312,7 +311,7 @@ static void svm_hardware_enable(void *garbage)
svm_data->next_asid = svm_data->max_asid + 1; svm_data->next_asid = svm_data->max_asid + 1;
svm_features = cpuid_edx(SVM_CPUID_FUNC); svm_features = cpuid_edx(SVM_CPUID_FUNC);
asm volatile ( "sgdt %0" : "=m"(gdt_descr) ); asm volatile ("sgdt %0" : "=m"(gdt_descr));
gdt = (struct desc_struct *)gdt_descr.address; gdt = (struct desc_struct *)gdt_descr.address;
svm_data->tss_desc = (struct kvm_ldttss_desc *)(gdt + GDT_ENTRY_TSS); svm_data->tss_desc = (struct kvm_ldttss_desc *)(gdt + GDT_ENTRY_TSS);
...@@ -544,8 +543,7 @@ static void init_vmcb(struct vmcb *vmcb) ...@@ -544,8 +543,7 @@ static void init_vmcb(struct vmcb *vmcb)
init_sys_seg(&save->tr, SEG_TYPE_BUSY_TSS16); init_sys_seg(&save->tr, SEG_TYPE_BUSY_TSS16);
save->efer = MSR_EFER_SVME_MASK; save->efer = MSR_EFER_SVME_MASK;
save->dr6 = 0xffff0ff0;
save->dr6 = 0xffff0ff0;
save->dr7 = 0x400; save->dr7 = 0x400;
save->rflags = 2; save->rflags = 2;
save->rip = 0x0000fff0; save->rip = 0x0000fff0;
...@@ -783,7 +781,7 @@ static void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) ...@@ -783,7 +781,7 @@ static void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
svm->vmcb->save.efer |= KVM_EFER_LMA | KVM_EFER_LME; svm->vmcb->save.efer |= KVM_EFER_LMA | KVM_EFER_LME;
} }
if (is_paging(vcpu) && !(cr0 & X86_CR0_PG) ) { if (is_paging(vcpu) && !(cr0 & X86_CR0_PG)) {
vcpu->shadow_efer &= ~KVM_EFER_LMA; vcpu->shadow_efer &= ~KVM_EFER_LMA;
svm->vmcb->save.efer &= ~(KVM_EFER_LMA | KVM_EFER_LME); svm->vmcb->save.efer &= ~(KVM_EFER_LMA | KVM_EFER_LME);
} }
...@@ -1010,7 +1008,7 @@ static int shutdown_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run) ...@@ -1010,7 +1008,7 @@ static int shutdown_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
static int io_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run) static int io_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
{ {
u32 io_info = svm->vmcb->control.exit_info_1; //address size bug? u32 io_info = svm->vmcb->control.exit_info_1; /* address size bug? */
int size, down, in, string, rep; int size, down, in, string, rep;
unsigned port; unsigned port;
...@@ -1316,7 +1314,7 @@ static void reload_tss(struct kvm_vcpu *vcpu) ...@@ -1316,7 +1314,7 @@ static void reload_tss(struct kvm_vcpu *vcpu)
int cpu = raw_smp_processor_id(); int cpu = raw_smp_processor_id();
struct svm_cpu_data *svm_data = per_cpu(svm_data, cpu); struct svm_cpu_data *svm_data = per_cpu(svm_data, cpu);
svm_data->tss_desc->type = 9; //available 32/64-bit TSS svm_data->tss_desc->type = 9; /* available 32/64-bit TSS */
load_TR_desc(); load_TR_desc();
} }
...@@ -1434,9 +1432,9 @@ static void do_interrupt_requests(struct kvm_vcpu *vcpu, ...@@ -1434,9 +1432,9 @@ static void do_interrupt_requests(struct kvm_vcpu *vcpu,
* Interrupts blocked. Wait for unblock. * Interrupts blocked. Wait for unblock.
*/ */
if (!svm->vcpu.interrupt_window_open && if (!svm->vcpu.interrupt_window_open &&
(svm->vcpu.irq_summary || kvm_run->request_interrupt_window)) { (svm->vcpu.irq_summary || kvm_run->request_interrupt_window))
control->intercept |= 1ULL << INTERCEPT_VINTR; control->intercept |= 1ULL << INTERCEPT_VINTR;
} else else
control->intercept &= ~(1ULL << INTERCEPT_VINTR); control->intercept &= ~(1ULL << INTERCEPT_VINTR);
} }
...@@ -1581,23 +1579,23 @@ static void svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1581,23 +1579,23 @@ static void svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
: :
: [svm]"a"(svm), : [svm]"a"(svm),
[vmcb]"i"(offsetof(struct vcpu_svm, vmcb_pa)), [vmcb]"i"(offsetof(struct vcpu_svm, vmcb_pa)),
[rbx]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_RBX])), [rbx]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_RBX])),
[rcx]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_RCX])), [rcx]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_RCX])),
[rdx]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_RDX])), [rdx]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_RDX])),
[rsi]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_RSI])), [rsi]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_RSI])),
[rdi]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_RDI])), [rdi]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_RDI])),
[rbp]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_RBP])) [rbp]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_RBP]))
#ifdef CONFIG_X86_64 #ifdef CONFIG_X86_64
,[r8 ]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_R8])), , [r8]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_R8])),
[r9 ]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_R9 ])), [r9]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_R9])),
[r10]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_R10])), [r10]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_R10])),
[r11]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_R11])), [r11]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_R11])),
[r12]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_R12])), [r12]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_R12])),
[r13]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_R13])), [r13]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_R13])),
[r14]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_R14])), [r14]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_R14])),
[r15]"i"(offsetof(struct vcpu_svm,vcpu.regs[VCPU_REGS_R15])) [r15]"i"(offsetof(struct vcpu_svm, vcpu.regs[VCPU_REGS_R15]))
#endif #endif
: "cc", "memory" ); : "cc", "memory");
if ((svm->vmcb->save.dr7 & 0xff)) if ((svm->vmcb->save.dr7 & 0xff))
load_db_regs(svm->host_db_regs); load_db_regs(svm->host_db_regs);
......
...@@ -311,7 +311,7 @@ struct __attribute__ ((__packed__)) vmcb { ...@@ -311,7 +311,7 @@ struct __attribute__ ((__packed__)) vmcb {
#define SVM_EXIT_ERR -1 #define SVM_EXIT_ERR -1
#define SVM_CR0_SELECTIVE_MASK (1 << 3 | 1) // TS and MP #define SVM_CR0_SELECTIVE_MASK (1 << 3 | 1) /* TS and MP */
#define SVM_VMLOAD ".byte 0x0f, 0x01, 0xda" #define SVM_VMLOAD ".byte 0x0f, 0x01, 0xda"
#define SVM_VMRUN ".byte 0x0f, 0x01, 0xd8" #define SVM_VMRUN ".byte 0x0f, 0x01, 0xd8"
......
...@@ -62,7 +62,7 @@ struct vcpu_vmx { ...@@ -62,7 +62,7 @@ struct vcpu_vmx {
int gs_ldt_reload_needed; int gs_ldt_reload_needed;
int fs_reload_needed; int fs_reload_needed;
int guest_efer_loaded; int guest_efer_loaded;
}host_state; } host_state;
}; };
...@@ -271,7 +271,7 @@ static void vmcs_writel(unsigned long field, unsigned long value) ...@@ -271,7 +271,7 @@ static void vmcs_writel(unsigned long field, unsigned long value)
u8 error; u8 error;
asm volatile (ASM_VMX_VMWRITE_RAX_RDX "; setna %0" asm volatile (ASM_VMX_VMWRITE_RAX_RDX "; setna %0"
: "=q"(error) : "a"(value), "d"(field) : "cc" ); : "=q"(error) : "a"(value), "d"(field) : "cc");
if (unlikely(error)) if (unlikely(error))
vmwrite_error(field, value); vmwrite_error(field, value);
} }
...@@ -415,10 +415,10 @@ static void vmx_save_host_state(struct kvm_vcpu *vcpu) ...@@ -415,10 +415,10 @@ static void vmx_save_host_state(struct kvm_vcpu *vcpu)
#endif #endif
#ifdef CONFIG_X86_64 #ifdef CONFIG_X86_64
if (is_long_mode(&vmx->vcpu)) { if (is_long_mode(&vmx->vcpu))
save_msrs(vmx->host_msrs + save_msrs(vmx->host_msrs +
vmx->msr_offset_kernel_gs_base, 1); vmx->msr_offset_kernel_gs_base, 1);
}
#endif #endif
load_msrs(vmx->guest_msrs, vmx->save_nmsrs); load_msrs(vmx->guest_msrs, vmx->save_nmsrs);
load_transition_efer(vmx); load_transition_efer(vmx);
...@@ -845,7 +845,7 @@ static int vmx_get_irq(struct kvm_vcpu *vcpu) ...@@ -845,7 +845,7 @@ static int vmx_get_irq(struct kvm_vcpu *vcpu)
if (is_external_interrupt(idtv_info_field)) if (is_external_interrupt(idtv_info_field))
return idtv_info_field & VECTORING_INFO_VECTOR_MASK; return idtv_info_field & VECTORING_INFO_VECTOR_MASK;
else else
printk("pending exception: not handled yet\n"); printk(KERN_DEBUG "pending exception: not handled yet\n");
} }
return -1; return -1;
} }
...@@ -893,7 +893,7 @@ static void hardware_disable(void *garbage) ...@@ -893,7 +893,7 @@ static void hardware_disable(void *garbage)
} }
static __init int adjust_vmx_controls(u32 ctl_min, u32 ctl_opt, static __init int adjust_vmx_controls(u32 ctl_min, u32 ctl_opt,
u32 msr, u32* result) u32 msr, u32 *result)
{ {
u32 vmx_msr_low, vmx_msr_high; u32 vmx_msr_low, vmx_msr_high;
u32 ctl = ctl_min | ctl_opt; u32 ctl = ctl_min | ctl_opt;
...@@ -1102,7 +1102,7 @@ static void enter_pmode(struct kvm_vcpu *vcpu) ...@@ -1102,7 +1102,7 @@ static void enter_pmode(struct kvm_vcpu *vcpu)
vmcs_write32(GUEST_CS_AR_BYTES, 0x9b); vmcs_write32(GUEST_CS_AR_BYTES, 0x9b);
} }
static gva_t rmode_tss_base(struct kvm* kvm) static gva_t rmode_tss_base(struct kvm *kvm)
{ {
gfn_t base_gfn = kvm->memslots[0].base_gfn + kvm->memslots[0].npages - 3; gfn_t base_gfn = kvm->memslots[0].base_gfn + kvm->memslots[0].npages - 3;
return base_gfn << PAGE_SHIFT; return base_gfn << PAGE_SHIFT;
...@@ -1385,7 +1385,7 @@ static void vmx_set_gdt(struct kvm_vcpu *vcpu, struct descriptor_table *dt) ...@@ -1385,7 +1385,7 @@ static void vmx_set_gdt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
vmcs_writel(GUEST_GDTR_BASE, dt->base); vmcs_writel(GUEST_GDTR_BASE, dt->base);
} }
static int init_rmode_tss(struct kvm* kvm) static int init_rmode_tss(struct kvm *kvm)
{ {
gfn_t fn = rmode_tss_base(kvm) >> PAGE_SHIFT; gfn_t fn = rmode_tss_base(kvm) >> PAGE_SHIFT;
u16 data = 0; u16 data = 0;
...@@ -1494,7 +1494,7 @@ static int vmx_vcpu_setup(struct vcpu_vmx *vmx) ...@@ -1494,7 +1494,7 @@ static int vmx_vcpu_setup(struct vcpu_vmx *vmx)
vmcs_writel(GUEST_RIP, 0); vmcs_writel(GUEST_RIP, 0);
vmcs_writel(GUEST_RSP, 0); vmcs_writel(GUEST_RSP, 0);
//todo: dr0 = dr1 = dr2 = dr3 = 0; dr6 = 0xffff0ff0 /* todo: dr0 = dr1 = dr2 = dr3 = 0; dr6 = 0xffff0ff0 */
vmcs_writel(GUEST_DR7, 0x400); vmcs_writel(GUEST_DR7, 0x400);
vmcs_writel(GUEST_GDTR_BASE, 0); vmcs_writel(GUEST_GDTR_BASE, 0);
...@@ -1561,7 +1561,7 @@ static int vmx_vcpu_setup(struct vcpu_vmx *vmx) ...@@ -1561,7 +1561,7 @@ static int vmx_vcpu_setup(struct vcpu_vmx *vmx)
get_idt(&dt); get_idt(&dt);
vmcs_writel(HOST_IDTR_BASE, dt.base); /* 22.2.4 */ vmcs_writel(HOST_IDTR_BASE, dt.base); /* 22.2.4 */
asm ("mov $.Lkvm_vmx_return, %0" : "=r"(kvm_vmx_return)); asm("mov $.Lkvm_vmx_return, %0" : "=r"(kvm_vmx_return));
vmcs_writel(HOST_RIP, kvm_vmx_return); /* 22.2.5 */ vmcs_writel(HOST_RIP, kvm_vmx_return); /* 22.2.5 */
vmcs_write32(VM_EXIT_MSR_STORE_COUNT, 0); vmcs_write32(VM_EXIT_MSR_STORE_COUNT, 0);
vmcs_write32(VM_EXIT_MSR_LOAD_COUNT, 0); vmcs_write32(VM_EXIT_MSR_LOAD_COUNT, 0);
...@@ -1613,7 +1613,7 @@ static int vmx_vcpu_setup(struct vcpu_vmx *vmx) ...@@ -1613,7 +1613,7 @@ static int vmx_vcpu_setup(struct vcpu_vmx *vmx)
vmcs_writel(CR4_GUEST_HOST_MASK, KVM_GUEST_CR4_MASK); vmcs_writel(CR4_GUEST_HOST_MASK, KVM_GUEST_CR4_MASK);
vmx->vcpu.cr0 = 0x60000010; vmx->vcpu.cr0 = 0x60000010;
vmx_set_cr0(&vmx->vcpu, vmx->vcpu.cr0); // enter rmode vmx_set_cr0(&vmx->vcpu, vmx->vcpu.cr0); /* enter rmode */
vmx_set_cr4(&vmx->vcpu, 0); vmx_set_cr4(&vmx->vcpu, 0);
#ifdef CONFIG_X86_64 #ifdef CONFIG_X86_64
vmx_set_efer(&vmx->vcpu, 0); vmx_set_efer(&vmx->vcpu, 0);
...@@ -1644,7 +1644,7 @@ static void inject_rmode_irq(struct kvm_vcpu *vcpu, int irq) ...@@ -1644,7 +1644,7 @@ static void inject_rmode_irq(struct kvm_vcpu *vcpu, int irq)
u16 sp = vmcs_readl(GUEST_RSP); u16 sp = vmcs_readl(GUEST_RSP);
u32 ss_limit = vmcs_read32(GUEST_SS_LIMIT); u32 ss_limit = vmcs_read32(GUEST_SS_LIMIT);
if (sp > ss_limit || sp < 6 ) { if (sp > ss_limit || sp < 6) {
vcpu_printf(vcpu, "%s: #SS, rsp 0x%lx ss 0x%lx limit 0x%x\n", vcpu_printf(vcpu, "%s: #SS, rsp 0x%lx ss 0x%lx limit 0x%x\n",
__FUNCTION__, __FUNCTION__,
vmcs_readl(GUEST_RSP), vmcs_readl(GUEST_RSP),
...@@ -1664,15 +1664,18 @@ static void inject_rmode_irq(struct kvm_vcpu *vcpu, int irq) ...@@ -1664,15 +1664,18 @@ static void inject_rmode_irq(struct kvm_vcpu *vcpu, int irq)
ip = vmcs_readl(GUEST_RIP); ip = vmcs_readl(GUEST_RIP);
if (emulator_write_emulated(ss_base + sp - 2, &flags, 2, vcpu) != X86EMUL_CONTINUE || if (emulator_write_emulated(
emulator_write_emulated(ss_base + sp - 4, &cs, 2, vcpu) != X86EMUL_CONTINUE || ss_base + sp - 2, &flags, 2, vcpu) != X86EMUL_CONTINUE ||
emulator_write_emulated(ss_base + sp - 6, &ip, 2, vcpu) != X86EMUL_CONTINUE) { emulator_write_emulated(
ss_base + sp - 4, &cs, 2, vcpu) != X86EMUL_CONTINUE ||
emulator_write_emulated(
ss_base + sp - 6, &ip, 2, vcpu) != X86EMUL_CONTINUE) {
vcpu_printf(vcpu, "%s: write guest err\n", __FUNCTION__); vcpu_printf(vcpu, "%s: write guest err\n", __FUNCTION__);
return; return;
} }
vmcs_writel(GUEST_RFLAGS, flags & vmcs_writel(GUEST_RFLAGS, flags &
~( X86_EFLAGS_IF | X86_EFLAGS_AC | X86_EFLAGS_TF)); ~(X86_EFLAGS_IF | X86_EFLAGS_AC | X86_EFLAGS_TF));
vmcs_write16(GUEST_CS_SELECTOR, ent[1]) ; vmcs_write16(GUEST_CS_SELECTOR, ent[1]) ;
vmcs_writel(GUEST_CS_BASE, ent[1] << 4); vmcs_writel(GUEST_CS_BASE, ent[1] << 4);
vmcs_writel(GUEST_RIP, ent[0]); vmcs_writel(GUEST_RIP, ent[0]);
...@@ -1777,10 +1780,9 @@ static int handle_exception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1777,10 +1780,9 @@ static int handle_exception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
intr_info = vmcs_read32(VM_EXIT_INTR_INFO); intr_info = vmcs_read32(VM_EXIT_INTR_INFO);
if ((vect_info & VECTORING_INFO_VALID_MASK) && if ((vect_info & VECTORING_INFO_VALID_MASK) &&
!is_page_fault(intr_info)) { !is_page_fault(intr_info))
printk(KERN_ERR "%s: unexpected, vectoring info 0x%x " printk(KERN_ERR "%s: unexpected, vectoring info 0x%x "
"intr info 0x%x\n", __FUNCTION__, vect_info, intr_info); "intr info 0x%x\n", __FUNCTION__, vect_info, intr_info);
}
if (!irqchip_in_kernel(vcpu->kvm) && is_external_interrupt(vect_info)) { if (!irqchip_in_kernel(vcpu->kvm) && is_external_interrupt(vect_info)) {
int irq = vect_info & VECTORING_INFO_VECTOR_MASK; int irq = vect_info & VECTORING_INFO_VECTOR_MASK;
...@@ -1831,7 +1833,7 @@ static int handle_exception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1831,7 +1833,7 @@ static int handle_exception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
case EMULATE_DO_MMIO: case EMULATE_DO_MMIO:
++vcpu->stat.mmio_exits; ++vcpu->stat.mmio_exits;
return 0; return 0;
case EMULATE_FAIL: case EMULATE_FAIL:
kvm_report_emulation_failure(vcpu, "pagetable"); kvm_report_emulation_failure(vcpu, "pagetable");
break; break;
default: default:
...@@ -1849,7 +1851,8 @@ static int handle_exception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1849,7 +1851,8 @@ static int handle_exception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
return 1; return 1;
} }
if ((intr_info & (INTR_INFO_INTR_TYPE_MASK | INTR_INFO_VECTOR_MASK)) == (INTR_TYPE_EXCEPTION | 1)) { if ((intr_info & (INTR_INFO_INTR_TYPE_MASK | INTR_INFO_VECTOR_MASK)) ==
(INTR_TYPE_EXCEPTION | 1)) {
kvm_run->exit_reason = KVM_EXIT_DEBUG; kvm_run->exit_reason = KVM_EXIT_DEBUG;
return 0; return 0;
} }
...@@ -2138,8 +2141,8 @@ static int kvm_handle_exit(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu) ...@@ -2138,8 +2141,8 @@ static int kvm_handle_exit(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
return 0; return 0;
} }
if ( (vectoring_info & VECTORING_INFO_VALID_MASK) && if ((vectoring_info & VECTORING_INFO_VALID_MASK) &&
exit_reason != EXIT_REASON_EXCEPTION_NMI ) exit_reason != EXIT_REASON_EXCEPTION_NMI)
printk(KERN_WARNING "%s: unexpected, valid vectoring info and " printk(KERN_WARNING "%s: unexpected, valid vectoring info and "
"exit reason is 0x%x\n", __FUNCTION__, exit_reason); "exit reason is 0x%x\n", __FUNCTION__, exit_reason);
if (exit_reason < kvm_vmx_max_exit_handlers if (exit_reason < kvm_vmx_max_exit_handlers
...@@ -2238,7 +2241,7 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -2238,7 +2241,7 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
*/ */
vmcs_writel(HOST_CR0, read_cr0()); vmcs_writel(HOST_CR0, read_cr0());
asm ( asm(
/* Store host registers */ /* Store host registers */
#ifdef CONFIG_X86_64 #ifdef CONFIG_X86_64
"push %%rax; push %%rbx; push %%rdx;" "push %%rax; push %%rbx; push %%rdx;"
...@@ -2342,8 +2345,8 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -2342,8 +2345,8 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
[rdi]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RDI])), [rdi]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RDI])),
[rbp]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RBP])), [rbp]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RBP])),
#ifdef CONFIG_X86_64 #ifdef CONFIG_X86_64
[r8 ]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R8 ])), [r8]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R8])),
[r9 ]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R9 ])), [r9]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R9])),
[r10]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R10])), [r10]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R10])),
[r11]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R11])), [r11]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R11])),
[r12]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R12])), [r12]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R12])),
...@@ -2352,11 +2355,12 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -2352,11 +2355,12 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
[r15]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R15])), [r15]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R15])),
#endif #endif
[cr2]"i"(offsetof(struct kvm_vcpu, cr2)) [cr2]"i"(offsetof(struct kvm_vcpu, cr2))
: "cc", "memory" ); : "cc", "memory");
vcpu->interrupt_window_open = (vmcs_read32(GUEST_INTERRUPTIBILITY_INFO) & 3) == 0; vcpu->interrupt_window_open =
(vmcs_read32(GUEST_INTERRUPTIBILITY_INFO) & 3) == 0;
asm ("mov %0, %%ds; mov %0, %%es" : : "r"(__USER_DS)); asm("mov %0, %%ds; mov %0, %%es" : : "r"(__USER_DS));
vmx->launched = 1; vmx->launched = 1;
intr_info = vmcs_read32(VM_EXIT_INTR_INFO); intr_info = vmcs_read32(VM_EXIT_INTR_INFO);
......
...@@ -234,9 +234,9 @@ enum vmcs_field { ...@@ -234,9 +234,9 @@ enum vmcs_field {
/* /*
* Exit Qualifications for MOV for Control Register Access * Exit Qualifications for MOV for Control Register Access
*/ */
#define CONTROL_REG_ACCESS_NUM 0x7 /* 2:0, number of control register */ #define CONTROL_REG_ACCESS_NUM 0x7 /* 2:0, number of control reg.*/
#define CONTROL_REG_ACCESS_TYPE 0x30 /* 5:4, access type */ #define CONTROL_REG_ACCESS_TYPE 0x30 /* 5:4, access type */
#define CONTROL_REG_ACCESS_REG 0xf00 /* 10:8, general purpose register */ #define CONTROL_REG_ACCESS_REG 0xf00 /* 10:8, general purpose reg. */
#define LMSW_SOURCE_DATA_SHIFT 16 #define LMSW_SOURCE_DATA_SHIFT 16
#define LMSW_SOURCE_DATA (0xFFFF << LMSW_SOURCE_DATA_SHIFT) /* 16:31 lmsw source */ #define LMSW_SOURCE_DATA (0xFFFF << LMSW_SOURCE_DATA_SHIFT) /* 16:31 lmsw source */
#define REG_EAX (0 << 8) #define REG_EAX (0 << 8)
...@@ -259,11 +259,11 @@ enum vmcs_field { ...@@ -259,11 +259,11 @@ enum vmcs_field {
/* /*
* Exit Qualifications for MOV for Debug Register Access * Exit Qualifications for MOV for Debug Register Access
*/ */
#define DEBUG_REG_ACCESS_NUM 0x7 /* 2:0, number of debug register */ #define DEBUG_REG_ACCESS_NUM 0x7 /* 2:0, number of debug reg. */
#define DEBUG_REG_ACCESS_TYPE 0x10 /* 4, direction of access */ #define DEBUG_REG_ACCESS_TYPE 0x10 /* 4, direction of access */
#define TYPE_MOV_TO_DR (0 << 4) #define TYPE_MOV_TO_DR (0 << 4)
#define TYPE_MOV_FROM_DR (1 << 4) #define TYPE_MOV_FROM_DR (1 << 4)
#define DEBUG_REG_ACCESS_REG 0xf00 /* 11:8, general purpose register */ #define DEBUG_REG_ACCESS_REG 0xf00 /* 11:8, general purpose reg. */
/* segment AR */ /* segment AR */
......
...@@ -23,7 +23,7 @@ ...@@ -23,7 +23,7 @@
#include <stdio.h> #include <stdio.h>
#include <stdint.h> #include <stdint.h>
#include <public/xen.h> #include <public/xen.h>
#define DPRINTF(_f, _a ...) printf( _f , ## _a ) #define DPRINTF(_f, _a ...) printf(_f , ## _a)
#else #else
#include "kvm.h" #include "kvm.h"
#define DPRINTF(x...) do {} while (0) #define DPRINTF(x...) do {} while (0)
...@@ -285,21 +285,21 @@ static u16 twobyte_table[256] = { ...@@ -285,21 +285,21 @@ static u16 twobyte_table[256] = {
switch ((_dst).bytes) { \ switch ((_dst).bytes) { \
case 2: \ case 2: \
__asm__ __volatile__ ( \ __asm__ __volatile__ ( \
_PRE_EFLAGS("0","4","2") \ _PRE_EFLAGS("0", "4", "2") \
_op"w %"_wx"3,%1; " \ _op"w %"_wx"3,%1; " \
_POST_EFLAGS("0","4","2") \ _POST_EFLAGS("0", "4", "2") \
: "=m" (_eflags), "=m" ((_dst).val), \ : "=m" (_eflags), "=m" ((_dst).val), \
"=&r" (_tmp) \ "=&r" (_tmp) \
: _wy ((_src).val), "i" (EFLAGS_MASK) ); \ : _wy ((_src).val), "i" (EFLAGS_MASK)); \
break; \ break; \
case 4: \ case 4: \
__asm__ __volatile__ ( \ __asm__ __volatile__ ( \
_PRE_EFLAGS("0","4","2") \ _PRE_EFLAGS("0", "4", "2") \
_op"l %"_lx"3,%1; " \ _op"l %"_lx"3,%1; " \
_POST_EFLAGS("0","4","2") \ _POST_EFLAGS("0", "4", "2") \
: "=m" (_eflags), "=m" ((_dst).val), \ : "=m" (_eflags), "=m" ((_dst).val), \
"=&r" (_tmp) \ "=&r" (_tmp) \
: _ly ((_src).val), "i" (EFLAGS_MASK) ); \ : _ly ((_src).val), "i" (EFLAGS_MASK)); \
break; \ break; \
case 8: \ case 8: \
__emulate_2op_8byte(_op, _src, _dst, \ __emulate_2op_8byte(_op, _src, _dst, \
...@@ -311,16 +311,15 @@ static u16 twobyte_table[256] = { ...@@ -311,16 +311,15 @@ static u16 twobyte_table[256] = {
#define __emulate_2op(_op,_src,_dst,_eflags,_bx,_by,_wx,_wy,_lx,_ly,_qx,_qy) \ #define __emulate_2op(_op,_src,_dst,_eflags,_bx,_by,_wx,_wy,_lx,_ly,_qx,_qy) \
do { \ do { \
unsigned long _tmp; \ unsigned long _tmp; \
switch ( (_dst).bytes ) \ switch ((_dst).bytes) { \
{ \
case 1: \ case 1: \
__asm__ __volatile__ ( \ __asm__ __volatile__ ( \
_PRE_EFLAGS("0","4","2") \ _PRE_EFLAGS("0", "4", "2") \
_op"b %"_bx"3,%1; " \ _op"b %"_bx"3,%1; " \
_POST_EFLAGS("0","4","2") \ _POST_EFLAGS("0", "4", "2") \
: "=m" (_eflags), "=m" ((_dst).val), \ : "=m" (_eflags), "=m" ((_dst).val), \
"=&r" (_tmp) \ "=&r" (_tmp) \
: _by ((_src).val), "i" (EFLAGS_MASK) ); \ : _by ((_src).val), "i" (EFLAGS_MASK)); \
break; \ break; \
default: \ default: \
__emulate_2op_nobyte(_op, _src, _dst, _eflags, \ __emulate_2op_nobyte(_op, _src, _dst, _eflags, \
...@@ -349,34 +348,33 @@ static u16 twobyte_table[256] = { ...@@ -349,34 +348,33 @@ static u16 twobyte_table[256] = {
do { \ do { \
unsigned long _tmp; \ unsigned long _tmp; \
\ \
switch ( (_dst).bytes ) \ switch ((_dst).bytes) { \
{ \
case 1: \ case 1: \
__asm__ __volatile__ ( \ __asm__ __volatile__ ( \
_PRE_EFLAGS("0","3","2") \ _PRE_EFLAGS("0", "3", "2") \
_op"b %1; " \ _op"b %1; " \
_POST_EFLAGS("0","3","2") \ _POST_EFLAGS("0", "3", "2") \
: "=m" (_eflags), "=m" ((_dst).val), \ : "=m" (_eflags), "=m" ((_dst).val), \
"=&r" (_tmp) \ "=&r" (_tmp) \
: "i" (EFLAGS_MASK) ); \ : "i" (EFLAGS_MASK)); \
break; \ break; \
case 2: \ case 2: \
__asm__ __volatile__ ( \ __asm__ __volatile__ ( \
_PRE_EFLAGS("0","3","2") \ _PRE_EFLAGS("0", "3", "2") \
_op"w %1; " \ _op"w %1; " \
_POST_EFLAGS("0","3","2") \ _POST_EFLAGS("0", "3", "2") \
: "=m" (_eflags), "=m" ((_dst).val), \ : "=m" (_eflags), "=m" ((_dst).val), \
"=&r" (_tmp) \ "=&r" (_tmp) \
: "i" (EFLAGS_MASK) ); \ : "i" (EFLAGS_MASK)); \
break; \ break; \
case 4: \ case 4: \
__asm__ __volatile__ ( \ __asm__ __volatile__ ( \
_PRE_EFLAGS("0","3","2") \ _PRE_EFLAGS("0", "3", "2") \
_op"l %1; " \ _op"l %1; " \
_POST_EFLAGS("0","3","2") \ _POST_EFLAGS("0", "3", "2") \
: "=m" (_eflags), "=m" ((_dst).val), \ : "=m" (_eflags), "=m" ((_dst).val), \
"=&r" (_tmp) \ "=&r" (_tmp) \
: "i" (EFLAGS_MASK) ); \ : "i" (EFLAGS_MASK)); \
break; \ break; \
case 8: \ case 8: \
__emulate_1op_8byte(_op, _dst, _eflags); \ __emulate_1op_8byte(_op, _dst, _eflags); \
...@@ -389,21 +387,21 @@ static u16 twobyte_table[256] = { ...@@ -389,21 +387,21 @@ static u16 twobyte_table[256] = {
#define __emulate_2op_8byte(_op, _src, _dst, _eflags, _qx, _qy) \ #define __emulate_2op_8byte(_op, _src, _dst, _eflags, _qx, _qy) \
do { \ do { \
__asm__ __volatile__ ( \ __asm__ __volatile__ ( \
_PRE_EFLAGS("0","4","2") \ _PRE_EFLAGS("0", "4", "2") \
_op"q %"_qx"3,%1; " \ _op"q %"_qx"3,%1; " \
_POST_EFLAGS("0","4","2") \ _POST_EFLAGS("0", "4", "2") \
: "=m" (_eflags), "=m" ((_dst).val), "=&r" (_tmp) \ : "=m" (_eflags), "=m" ((_dst).val), "=&r" (_tmp) \
: _qy ((_src).val), "i" (EFLAGS_MASK) ); \ : _qy ((_src).val), "i" (EFLAGS_MASK)); \
} while (0) } while (0)
#define __emulate_1op_8byte(_op, _dst, _eflags) \ #define __emulate_1op_8byte(_op, _dst, _eflags) \
do { \ do { \
__asm__ __volatile__ ( \ __asm__ __volatile__ ( \
_PRE_EFLAGS("0","3","2") \ _PRE_EFLAGS("0", "3", "2") \
_op"q %1; " \ _op"q %1; " \
_POST_EFLAGS("0","3","2") \ _POST_EFLAGS("0", "3", "2") \
: "=m" (_eflags), "=m" ((_dst).val), "=&r" (_tmp) \ : "=m" (_eflags), "=m" ((_dst).val), "=&r" (_tmp) \
: "i" (EFLAGS_MASK) ); \ : "i" (EFLAGS_MASK)); \
} while (0) } while (0)
#elif defined(__i386__) #elif defined(__i386__)
...@@ -415,8 +413,8 @@ static u16 twobyte_table[256] = { ...@@ -415,8 +413,8 @@ static u16 twobyte_table[256] = {
#define insn_fetch(_type, _size, _eip) \ #define insn_fetch(_type, _size, _eip) \
({ unsigned long _x; \ ({ unsigned long _x; \
rc = ops->read_std((unsigned long)(_eip) + ctxt->cs_base, &_x, \ rc = ops->read_std((unsigned long)(_eip) + ctxt->cs_base, &_x, \
(_size), ctxt->vcpu); \ (_size), ctxt->vcpu); \
if ( rc != 0 ) \ if (rc != 0) \
goto done; \ goto done; \
(_eip) += (_size); \ (_eip) += (_size); \
(_type)_x; \ (_type)_x; \
...@@ -780,7 +778,7 @@ done_prefixes: ...@@ -780,7 +778,7 @@ done_prefixes:
} }
if (c->ad_bytes != 8) if (c->ad_bytes != 8)
c->modrm_ea = (u32)c->modrm_ea; c->modrm_ea = (u32)c->modrm_ea;
modrm_done: modrm_done:
; ;
} }
...@@ -828,10 +826,9 @@ done_prefixes: ...@@ -828,10 +826,9 @@ done_prefixes:
c->src.bytes = (c->d & ByteOp) ? 1 : c->src.bytes = (c->d & ByteOp) ? 1 :
c->op_bytes; c->op_bytes;
/* Don't fetch the address for invlpg: it could be unmapped. */ /* Don't fetch the address for invlpg: it could be unmapped. */
if (c->twobyte && c->b == 0x01 if (c->twobyte && c->b == 0x01 && c->modrm_reg == 7)
&& c->modrm_reg == 7)
break; break;
srcmem_common: srcmem_common:
/* /*
* For instructions with a ModR/M byte, switch to register * For instructions with a ModR/M byte, switch to register
* access if Mod = 3. * access if Mod = 3.
...@@ -1175,10 +1172,11 @@ x86_emulate_insn(struct x86_emulate_ctxt *ctxt, struct x86_emulate_ops *ops) ...@@ -1175,10 +1172,11 @@ x86_emulate_insn(struct x86_emulate_ctxt *ctxt, struct x86_emulate_ops *ops)
if (c->src.type == OP_MEM) { if (c->src.type == OP_MEM) {
c->src.ptr = (unsigned long *)cr2; c->src.ptr = (unsigned long *)cr2;
c->src.val = 0; c->src.val = 0;
if ((rc = ops->read_emulated((unsigned long)c->src.ptr, rc = ops->read_emulated((unsigned long)c->src.ptr,
&c->src.val, &c->src.val,
c->src.bytes, c->src.bytes,
ctxt->vcpu)) != 0) ctxt->vcpu);
if (rc != 0)
goto done; goto done;
c->src.orig_val = c->src.val; c->src.orig_val = c->src.val;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment