KVM: SVM: de-containization
container_of is wonderful, but not casting at all is better. This patch changes svm.c's internal functions to pass "struct vcpu_svm" instead of "struct kvm_vcpu" and using container_of. Signed-off-by: Rusty Russell <rusty@rustcorp.com.au> Signed-off-by: Avi Kivity <avi@qumranet.com>
This commit is contained in:
parent
3077c4513c
commit
e756fc626d
1 changed files with 108 additions and 124 deletions
|
@ -98,9 +98,9 @@ static inline u32 svm_has(u32 feat)
|
||||||
return svm_features & feat;
|
return svm_features & feat;
|
||||||
}
|
}
|
||||||
|
|
||||||
static unsigned get_addr_size(struct kvm_vcpu *vcpu)
|
static unsigned get_addr_size(struct vcpu_svm *svm)
|
||||||
{
|
{
|
||||||
struct vmcb_save_area *sa = &to_svm(vcpu)->vmcb->save;
|
struct vmcb_save_area *sa = &svm->vmcb->save;
|
||||||
u16 cs_attrib;
|
u16 cs_attrib;
|
||||||
|
|
||||||
if (!(sa->cr0 & X86_CR0_PE) || (sa->rflags & X86_EFLAGS_VM))
|
if (!(sa->cr0 & X86_CR0_PE) || (sa->rflags & X86_EFLAGS_VM))
|
||||||
|
@ -865,17 +865,15 @@ static void save_host_msrs(struct kvm_vcpu *vcpu)
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
static void new_asid(struct kvm_vcpu *vcpu, struct svm_cpu_data *svm_data)
|
static void new_asid(struct vcpu_svm *svm, struct svm_cpu_data *svm_data)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
|
||||||
|
|
||||||
if (svm_data->next_asid > svm_data->max_asid) {
|
if (svm_data->next_asid > svm_data->max_asid) {
|
||||||
++svm_data->asid_generation;
|
++svm_data->asid_generation;
|
||||||
svm_data->next_asid = 1;
|
svm_data->next_asid = 1;
|
||||||
svm->vmcb->control.tlb_ctl = TLB_CONTROL_FLUSH_ALL_ASID;
|
svm->vmcb->control.tlb_ctl = TLB_CONTROL_FLUSH_ALL_ASID;
|
||||||
}
|
}
|
||||||
|
|
||||||
vcpu->cpu = svm_data->cpu;
|
svm->vcpu.cpu = svm_data->cpu;
|
||||||
svm->asid_generation = svm_data->asid_generation;
|
svm->asid_generation = svm_data->asid_generation;
|
||||||
svm->vmcb->control.asid = svm_data->next_asid++;
|
svm->vmcb->control.asid = svm_data->next_asid++;
|
||||||
}
|
}
|
||||||
|
@ -929,42 +927,43 @@ static void svm_set_dr(struct kvm_vcpu *vcpu, int dr, unsigned long value,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static int pf_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int pf_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
|
||||||
u32 exit_int_info = svm->vmcb->control.exit_int_info;
|
u32 exit_int_info = svm->vmcb->control.exit_int_info;
|
||||||
|
struct kvm *kvm = svm->vcpu.kvm;
|
||||||
u64 fault_address;
|
u64 fault_address;
|
||||||
u32 error_code;
|
u32 error_code;
|
||||||
enum emulation_result er;
|
enum emulation_result er;
|
||||||
int r;
|
int r;
|
||||||
|
|
||||||
if (is_external_interrupt(exit_int_info))
|
if (is_external_interrupt(exit_int_info))
|
||||||
push_irq(vcpu, exit_int_info & SVM_EVTINJ_VEC_MASK);
|
push_irq(&svm->vcpu, exit_int_info & SVM_EVTINJ_VEC_MASK);
|
||||||
|
|
||||||
mutex_lock(&vcpu->kvm->lock);
|
mutex_lock(&kvm->lock);
|
||||||
|
|
||||||
fault_address = svm->vmcb->control.exit_info_2;
|
fault_address = svm->vmcb->control.exit_info_2;
|
||||||
error_code = svm->vmcb->control.exit_info_1;
|
error_code = svm->vmcb->control.exit_info_1;
|
||||||
r = kvm_mmu_page_fault(vcpu, fault_address, error_code);
|
r = kvm_mmu_page_fault(&svm->vcpu, fault_address, error_code);
|
||||||
if (r < 0) {
|
if (r < 0) {
|
||||||
mutex_unlock(&vcpu->kvm->lock);
|
mutex_unlock(&kvm->lock);
|
||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
if (!r) {
|
if (!r) {
|
||||||
mutex_unlock(&vcpu->kvm->lock);
|
mutex_unlock(&kvm->lock);
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
er = emulate_instruction(vcpu, kvm_run, fault_address, error_code);
|
er = emulate_instruction(&svm->vcpu, kvm_run, fault_address,
|
||||||
mutex_unlock(&vcpu->kvm->lock);
|
error_code);
|
||||||
|
mutex_unlock(&kvm->lock);
|
||||||
|
|
||||||
switch (er) {
|
switch (er) {
|
||||||
case EMULATE_DONE:
|
case EMULATE_DONE:
|
||||||
return 1;
|
return 1;
|
||||||
case EMULATE_DO_MMIO:
|
case EMULATE_DO_MMIO:
|
||||||
++vcpu->stat.mmio_exits;
|
++svm->vcpu.stat.mmio_exits;
|
||||||
return 0;
|
return 0;
|
||||||
case EMULATE_FAIL:
|
case EMULATE_FAIL:
|
||||||
vcpu_printf(vcpu, "%s: emulate fail\n", __FUNCTION__);
|
vcpu_printf(&svm->vcpu, "%s: emulate fail\n", __FUNCTION__);
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
BUG();
|
BUG();
|
||||||
|
@ -974,21 +973,18 @@ static int pf_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int nm_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int nm_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
|
||||||
|
|
||||||
svm->vmcb->control.intercept_exceptions &= ~(1 << NM_VECTOR);
|
svm->vmcb->control.intercept_exceptions &= ~(1 << NM_VECTOR);
|
||||||
if (!(vcpu->cr0 & X86_CR0_TS))
|
if (!(svm->vcpu.cr0 & X86_CR0_TS))
|
||||||
svm->vmcb->save.cr0 &= ~X86_CR0_TS;
|
svm->vmcb->save.cr0 &= ~X86_CR0_TS;
|
||||||
vcpu->fpu_active = 1;
|
svm->vcpu.fpu_active = 1;
|
||||||
|
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int shutdown_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int shutdown_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
|
||||||
/*
|
/*
|
||||||
* VMCB is undefined after a SHUTDOWN intercept
|
* VMCB is undefined after a SHUTDOWN intercept
|
||||||
* so reinitialize it.
|
* so reinitialize it.
|
||||||
|
@ -1000,11 +996,10 @@ static int shutdown_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int io_get_override(struct kvm_vcpu *vcpu,
|
static int io_get_override(struct vcpu_svm *svm,
|
||||||
struct vmcb_seg **seg,
|
struct vmcb_seg **seg,
|
||||||
int *addr_override)
|
int *addr_override)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
|
||||||
u8 inst[MAX_INST_SIZE];
|
u8 inst[MAX_INST_SIZE];
|
||||||
unsigned ins_length;
|
unsigned ins_length;
|
||||||
gva_t rip;
|
gva_t rip;
|
||||||
|
@ -1024,7 +1019,7 @@ static int io_get_override(struct kvm_vcpu *vcpu,
|
||||||
svm->vmcb->control.exit_info_2,
|
svm->vmcb->control.exit_info_2,
|
||||||
ins_length);
|
ins_length);
|
||||||
|
|
||||||
if (kvm_read_guest(vcpu, rip, ins_length, inst) != ins_length)
|
if (kvm_read_guest(&svm->vcpu, rip, ins_length, inst) != ins_length)
|
||||||
/* #PF */
|
/* #PF */
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
|
@ -1065,28 +1060,27 @@ static int io_get_override(struct kvm_vcpu *vcpu,
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static unsigned long io_adress(struct kvm_vcpu *vcpu, int ins, gva_t *address)
|
static unsigned long io_adress(struct vcpu_svm *svm, int ins, gva_t *address)
|
||||||
{
|
{
|
||||||
unsigned long addr_mask;
|
unsigned long addr_mask;
|
||||||
unsigned long *reg;
|
unsigned long *reg;
|
||||||
struct vmcb_seg *seg;
|
struct vmcb_seg *seg;
|
||||||
int addr_override;
|
int addr_override;
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
|
||||||
struct vmcb_save_area *save_area = &svm->vmcb->save;
|
struct vmcb_save_area *save_area = &svm->vmcb->save;
|
||||||
u16 cs_attrib = save_area->cs.attrib;
|
u16 cs_attrib = save_area->cs.attrib;
|
||||||
unsigned addr_size = get_addr_size(vcpu);
|
unsigned addr_size = get_addr_size(svm);
|
||||||
|
|
||||||
if (!io_get_override(vcpu, &seg, &addr_override))
|
if (!io_get_override(svm, &seg, &addr_override))
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
if (addr_override)
|
if (addr_override)
|
||||||
addr_size = (addr_size == 2) ? 4: (addr_size >> 1);
|
addr_size = (addr_size == 2) ? 4: (addr_size >> 1);
|
||||||
|
|
||||||
if (ins) {
|
if (ins) {
|
||||||
reg = &vcpu->regs[VCPU_REGS_RDI];
|
reg = &svm->vcpu.regs[VCPU_REGS_RDI];
|
||||||
seg = &svm->vmcb->save.es;
|
seg = &svm->vmcb->save.es;
|
||||||
} else {
|
} else {
|
||||||
reg = &vcpu->regs[VCPU_REGS_RSI];
|
reg = &svm->vcpu.regs[VCPU_REGS_RSI];
|
||||||
seg = (seg) ? seg : &svm->vmcb->save.ds;
|
seg = (seg) ? seg : &svm->vmcb->save.ds;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1099,7 +1093,7 @@ static unsigned long io_adress(struct kvm_vcpu *vcpu, int ins, gva_t *address)
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!(seg->attrib & SVM_SELECTOR_P_SHIFT)) {
|
if (!(seg->attrib & SVM_SELECTOR_P_SHIFT)) {
|
||||||
svm_inject_gp(vcpu, 0);
|
svm_inject_gp(&svm->vcpu, 0);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1107,16 +1101,15 @@ static unsigned long io_adress(struct kvm_vcpu *vcpu, int ins, gva_t *address)
|
||||||
return addr_mask;
|
return addr_mask;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int io_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int io_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
|
||||||
u32 io_info = svm->vmcb->control.exit_info_1; //address size bug?
|
u32 io_info = svm->vmcb->control.exit_info_1; //address size bug?
|
||||||
int size, down, in, string, rep;
|
int size, down, in, string, rep;
|
||||||
unsigned port;
|
unsigned port;
|
||||||
unsigned long count;
|
unsigned long count;
|
||||||
gva_t address = 0;
|
gva_t address = 0;
|
||||||
|
|
||||||
++vcpu->stat.io_exits;
|
++svm->vcpu.stat.io_exits;
|
||||||
|
|
||||||
svm->next_rip = svm->vmcb->control.exit_info_2;
|
svm->next_rip = svm->vmcb->control.exit_info_2;
|
||||||
|
|
||||||
|
@ -1131,7 +1124,7 @@ static int io_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
||||||
if (string) {
|
if (string) {
|
||||||
unsigned addr_mask;
|
unsigned addr_mask;
|
||||||
|
|
||||||
addr_mask = io_adress(vcpu, in, &address);
|
addr_mask = io_adress(svm, in, &address);
|
||||||
if (!addr_mask) {
|
if (!addr_mask) {
|
||||||
printk(KERN_DEBUG "%s: get io address failed\n",
|
printk(KERN_DEBUG "%s: get io address failed\n",
|
||||||
__FUNCTION__);
|
__FUNCTION__);
|
||||||
|
@ -1139,60 +1132,57 @@ static int io_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
||||||
}
|
}
|
||||||
|
|
||||||
if (rep)
|
if (rep)
|
||||||
count = vcpu->regs[VCPU_REGS_RCX] & addr_mask;
|
count = svm->vcpu.regs[VCPU_REGS_RCX] & addr_mask;
|
||||||
}
|
}
|
||||||
return kvm_setup_pio(vcpu, kvm_run, in, size, count, string, down,
|
return kvm_setup_pio(&svm->vcpu, kvm_run, in, size, count, string,
|
||||||
address, rep, port);
|
down, address, rep, port);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int nop_on_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int nop_on_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int halt_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int halt_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
|
||||||
|
|
||||||
svm->next_rip = svm->vmcb->save.rip + 1;
|
svm->next_rip = svm->vmcb->save.rip + 1;
|
||||||
skip_emulated_instruction(vcpu);
|
skip_emulated_instruction(&svm->vcpu);
|
||||||
return kvm_emulate_halt(vcpu);
|
return kvm_emulate_halt(&svm->vcpu);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int vmmcall_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int vmmcall_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
|
||||||
|
|
||||||
svm->next_rip = svm->vmcb->save.rip + 3;
|
svm->next_rip = svm->vmcb->save.rip + 3;
|
||||||
skip_emulated_instruction(vcpu);
|
skip_emulated_instruction(&svm->vcpu);
|
||||||
return kvm_hypercall(vcpu, kvm_run);
|
return kvm_hypercall(&svm->vcpu, kvm_run);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int invalid_op_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int invalid_op_interception(struct vcpu_svm *svm,
|
||||||
|
struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
inject_ud(vcpu);
|
inject_ud(&svm->vcpu);
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int task_switch_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int task_switch_interception(struct vcpu_svm *svm,
|
||||||
|
struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
printk(KERN_DEBUG "%s: task swiche is unsupported\n", __FUNCTION__);
|
printk(KERN_DEBUG "%s: task swiche is unsupported\n", __FUNCTION__);
|
||||||
kvm_run->exit_reason = KVM_EXIT_UNKNOWN;
|
kvm_run->exit_reason = KVM_EXIT_UNKNOWN;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int cpuid_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int cpuid_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
|
||||||
|
|
||||||
svm->next_rip = svm->vmcb->save.rip + 2;
|
svm->next_rip = svm->vmcb->save.rip + 2;
|
||||||
kvm_emulate_cpuid(vcpu);
|
kvm_emulate_cpuid(&svm->vcpu);
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int emulate_on_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int emulate_on_interception(struct vcpu_svm *svm,
|
||||||
|
struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
if (emulate_instruction(vcpu, NULL, 0, 0) != EMULATE_DONE)
|
if (emulate_instruction(&svm->vcpu, NULL, 0, 0) != EMULATE_DONE)
|
||||||
printk(KERN_ERR "%s: failed\n", __FUNCTION__);
|
printk(KERN_ERR "%s: failed\n", __FUNCTION__);
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
@ -1241,19 +1231,18 @@ static int svm_get_msr(struct kvm_vcpu *vcpu, unsigned ecx, u64 *data)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int rdmsr_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int rdmsr_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
u32 ecx = svm->vcpu.regs[VCPU_REGS_RCX];
|
||||||
u32 ecx = vcpu->regs[VCPU_REGS_RCX];
|
|
||||||
u64 data;
|
u64 data;
|
||||||
|
|
||||||
if (svm_get_msr(vcpu, ecx, &data))
|
if (svm_get_msr(&svm->vcpu, ecx, &data))
|
||||||
svm_inject_gp(vcpu, 0);
|
svm_inject_gp(&svm->vcpu, 0);
|
||||||
else {
|
else {
|
||||||
svm->vmcb->save.rax = data & 0xffffffff;
|
svm->vmcb->save.rax = data & 0xffffffff;
|
||||||
vcpu->regs[VCPU_REGS_RDX] = data >> 32;
|
svm->vcpu.regs[VCPU_REGS_RDX] = data >> 32;
|
||||||
svm->next_rip = svm->vmcb->save.rip + 2;
|
svm->next_rip = svm->vmcb->save.rip + 2;
|
||||||
skip_emulated_instruction(vcpu);
|
skip_emulated_instruction(&svm->vcpu);
|
||||||
}
|
}
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
@ -1302,29 +1291,28 @@ static int svm_set_msr(struct kvm_vcpu *vcpu, unsigned ecx, u64 data)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int wrmsr_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int wrmsr_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
u32 ecx = svm->vcpu.regs[VCPU_REGS_RCX];
|
||||||
u32 ecx = vcpu->regs[VCPU_REGS_RCX];
|
|
||||||
u64 data = (svm->vmcb->save.rax & -1u)
|
u64 data = (svm->vmcb->save.rax & -1u)
|
||||||
| ((u64)(vcpu->regs[VCPU_REGS_RDX] & -1u) << 32);
|
| ((u64)(svm->vcpu.regs[VCPU_REGS_RDX] & -1u) << 32);
|
||||||
svm->next_rip = svm->vmcb->save.rip + 2;
|
svm->next_rip = svm->vmcb->save.rip + 2;
|
||||||
if (svm_set_msr(vcpu, ecx, data))
|
if (svm_set_msr(&svm->vcpu, ecx, data))
|
||||||
svm_inject_gp(vcpu, 0);
|
svm_inject_gp(&svm->vcpu, 0);
|
||||||
else
|
else
|
||||||
skip_emulated_instruction(vcpu);
|
skip_emulated_instruction(&svm->vcpu);
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int msr_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int msr_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
if (to_svm(vcpu)->vmcb->control.exit_info_1)
|
if (svm->vmcb->control.exit_info_1)
|
||||||
return wrmsr_interception(vcpu, kvm_run);
|
return wrmsr_interception(svm, kvm_run);
|
||||||
else
|
else
|
||||||
return rdmsr_interception(vcpu, kvm_run);
|
return rdmsr_interception(svm, kvm_run);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int interrupt_window_interception(struct kvm_vcpu *vcpu,
|
static int interrupt_window_interception(struct vcpu_svm *svm,
|
||||||
struct kvm_run *kvm_run)
|
struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
/*
|
/*
|
||||||
|
@ -1332,8 +1320,8 @@ static int interrupt_window_interception(struct kvm_vcpu *vcpu,
|
||||||
* possible
|
* possible
|
||||||
*/
|
*/
|
||||||
if (kvm_run->request_interrupt_window &&
|
if (kvm_run->request_interrupt_window &&
|
||||||
!vcpu->irq_summary) {
|
!svm->vcpu.irq_summary) {
|
||||||
++vcpu->stat.irq_window_exits;
|
++svm->vcpu.stat.irq_window_exits;
|
||||||
kvm_run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN;
|
kvm_run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
@ -1341,7 +1329,7 @@ static int interrupt_window_interception(struct kvm_vcpu *vcpu,
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int (*svm_exit_handlers[])(struct kvm_vcpu *vcpu,
|
static int (*svm_exit_handlers[])(struct vcpu_svm *svm,
|
||||||
struct kvm_run *kvm_run) = {
|
struct kvm_run *kvm_run) = {
|
||||||
[SVM_EXIT_READ_CR0] = emulate_on_interception,
|
[SVM_EXIT_READ_CR0] = emulate_on_interception,
|
||||||
[SVM_EXIT_READ_CR3] = emulate_on_interception,
|
[SVM_EXIT_READ_CR3] = emulate_on_interception,
|
||||||
|
@ -1388,9 +1376,8 @@ static int (*svm_exit_handlers[])(struct kvm_vcpu *vcpu,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
static int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int handle_exit(struct vcpu_svm *svm, struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
|
||||||
u32 exit_code = svm->vmcb->control.exit_code;
|
u32 exit_code = svm->vmcb->control.exit_code;
|
||||||
|
|
||||||
if (is_external_interrupt(svm->vmcb->control.exit_int_info) &&
|
if (is_external_interrupt(svm->vmcb->control.exit_int_info) &&
|
||||||
|
@ -1407,7 +1394,7 @@ static int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
return svm_exit_handlers[exit_code](vcpu, kvm_run);
|
return svm_exit_handlers[exit_code](svm, kvm_run);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void reload_tss(struct kvm_vcpu *vcpu)
|
static void reload_tss(struct kvm_vcpu *vcpu)
|
||||||
|
@ -1419,80 +1406,77 @@ static void reload_tss(struct kvm_vcpu *vcpu)
|
||||||
load_TR_desc();
|
load_TR_desc();
|
||||||
}
|
}
|
||||||
|
|
||||||
static void pre_svm_run(struct kvm_vcpu *vcpu)
|
static void pre_svm_run(struct vcpu_svm *svm)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
|
||||||
int cpu = raw_smp_processor_id();
|
int cpu = raw_smp_processor_id();
|
||||||
|
|
||||||
struct svm_cpu_data *svm_data = per_cpu(svm_data, cpu);
|
struct svm_cpu_data *svm_data = per_cpu(svm_data, cpu);
|
||||||
|
|
||||||
svm->vmcb->control.tlb_ctl = TLB_CONTROL_DO_NOTHING;
|
svm->vmcb->control.tlb_ctl = TLB_CONTROL_DO_NOTHING;
|
||||||
if (vcpu->cpu != cpu ||
|
if (svm->vcpu.cpu != cpu ||
|
||||||
svm->asid_generation != svm_data->asid_generation)
|
svm->asid_generation != svm_data->asid_generation)
|
||||||
new_asid(vcpu, svm_data);
|
new_asid(svm, svm_data);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
static inline void kvm_do_inject_irq(struct kvm_vcpu *vcpu)
|
static inline void kvm_do_inject_irq(struct vcpu_svm *svm)
|
||||||
{
|
{
|
||||||
struct vmcb_control_area *control;
|
struct vmcb_control_area *control;
|
||||||
|
|
||||||
control = &to_svm(vcpu)->vmcb->control;
|
control = &svm->vmcb->control;
|
||||||
control->int_vector = pop_irq(vcpu);
|
control->int_vector = pop_irq(&svm->vcpu);
|
||||||
control->int_ctl &= ~V_INTR_PRIO_MASK;
|
control->int_ctl &= ~V_INTR_PRIO_MASK;
|
||||||
control->int_ctl |= V_IRQ_MASK |
|
control->int_ctl |= V_IRQ_MASK |
|
||||||
((/*control->int_vector >> 4*/ 0xf) << V_INTR_PRIO_SHIFT);
|
((/*control->int_vector >> 4*/ 0xf) << V_INTR_PRIO_SHIFT);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void kvm_reput_irq(struct kvm_vcpu *vcpu)
|
static void kvm_reput_irq(struct vcpu_svm *svm)
|
||||||
{
|
{
|
||||||
struct vmcb_control_area *control = &to_svm(vcpu)->vmcb->control;
|
struct vmcb_control_area *control = &svm->vmcb->control;
|
||||||
|
|
||||||
if (control->int_ctl & V_IRQ_MASK) {
|
if (control->int_ctl & V_IRQ_MASK) {
|
||||||
control->int_ctl &= ~V_IRQ_MASK;
|
control->int_ctl &= ~V_IRQ_MASK;
|
||||||
push_irq(vcpu, control->int_vector);
|
push_irq(&svm->vcpu, control->int_vector);
|
||||||
}
|
}
|
||||||
|
|
||||||
vcpu->interrupt_window_open =
|
svm->vcpu.interrupt_window_open =
|
||||||
!(control->int_state & SVM_INTERRUPT_SHADOW_MASK);
|
!(control->int_state & SVM_INTERRUPT_SHADOW_MASK);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void do_interrupt_requests(struct kvm_vcpu *vcpu,
|
static void do_interrupt_requests(struct vcpu_svm *svm,
|
||||||
struct kvm_run *kvm_run)
|
struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
|
||||||
struct vmcb_control_area *control = &svm->vmcb->control;
|
struct vmcb_control_area *control = &svm->vmcb->control;
|
||||||
|
|
||||||
vcpu->interrupt_window_open =
|
svm->vcpu.interrupt_window_open =
|
||||||
(!(control->int_state & SVM_INTERRUPT_SHADOW_MASK) &&
|
(!(control->int_state & SVM_INTERRUPT_SHADOW_MASK) &&
|
||||||
(svm->vmcb->save.rflags & X86_EFLAGS_IF));
|
(svm->vmcb->save.rflags & X86_EFLAGS_IF));
|
||||||
|
|
||||||
if (vcpu->interrupt_window_open && vcpu->irq_summary)
|
if (svm->vcpu.interrupt_window_open && svm->vcpu.irq_summary)
|
||||||
/*
|
/*
|
||||||
* If interrupts enabled, and not blocked by sti or mov ss. Good.
|
* If interrupts enabled, and not blocked by sti or mov ss. Good.
|
||||||
*/
|
*/
|
||||||
kvm_do_inject_irq(vcpu);
|
kvm_do_inject_irq(svm);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Interrupts blocked. Wait for unblock.
|
* Interrupts blocked. Wait for unblock.
|
||||||
*/
|
*/
|
||||||
if (!vcpu->interrupt_window_open &&
|
if (!svm->vcpu.interrupt_window_open &&
|
||||||
(vcpu->irq_summary || kvm_run->request_interrupt_window)) {
|
(svm->vcpu.irq_summary || kvm_run->request_interrupt_window)) {
|
||||||
control->intercept |= 1ULL << INTERCEPT_VINTR;
|
control->intercept |= 1ULL << INTERCEPT_VINTR;
|
||||||
} else
|
} else
|
||||||
control->intercept &= ~(1ULL << INTERCEPT_VINTR);
|
control->intercept &= ~(1ULL << INTERCEPT_VINTR);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void post_kvm_run_save(struct kvm_vcpu *vcpu,
|
static void post_kvm_run_save(struct vcpu_svm *svm,
|
||||||
struct kvm_run *kvm_run)
|
struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
struct vcpu_svm *svm = to_svm(vcpu);
|
kvm_run->ready_for_interrupt_injection
|
||||||
|
= (svm->vcpu.interrupt_window_open &&
|
||||||
kvm_run->ready_for_interrupt_injection = (vcpu->interrupt_window_open &&
|
svm->vcpu.irq_summary == 0);
|
||||||
vcpu->irq_summary == 0);
|
|
||||||
kvm_run->if_flag = (svm->vmcb->save.rflags & X86_EFLAGS_IF) != 0;
|
kvm_run->if_flag = (svm->vmcb->save.rflags & X86_EFLAGS_IF) != 0;
|
||||||
kvm_run->cr8 = vcpu->cr8;
|
kvm_run->cr8 = svm->vcpu.cr8;
|
||||||
kvm_run->apic_base = vcpu->apic_base;
|
kvm_run->apic_base = svm->vcpu.apic_base;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -1501,13 +1485,13 @@ static void post_kvm_run_save(struct kvm_vcpu *vcpu,
|
||||||
*
|
*
|
||||||
* No need to exit to userspace if we already have an interrupt queued.
|
* No need to exit to userspace if we already have an interrupt queued.
|
||||||
*/
|
*/
|
||||||
static int dm_request_for_irq_injection(struct kvm_vcpu *vcpu,
|
static int dm_request_for_irq_injection(struct vcpu_svm *svm,
|
||||||
struct kvm_run *kvm_run)
|
struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
return (!vcpu->irq_summary &&
|
return (!svm->vcpu.irq_summary &&
|
||||||
kvm_run->request_interrupt_window &&
|
kvm_run->request_interrupt_window &&
|
||||||
vcpu->interrupt_window_open &&
|
svm->vcpu.interrupt_window_open &&
|
||||||
(to_svm(vcpu)->vmcb->save.rflags & X86_EFLAGS_IF));
|
(svm->vmcb->save.rflags & X86_EFLAGS_IF));
|
||||||
}
|
}
|
||||||
|
|
||||||
static void save_db_regs(unsigned long *db_regs)
|
static void save_db_regs(unsigned long *db_regs)
|
||||||
|
@ -1545,7 +1529,7 @@ again:
|
||||||
return r;
|
return r;
|
||||||
|
|
||||||
if (!vcpu->mmio_read_completed)
|
if (!vcpu->mmio_read_completed)
|
||||||
do_interrupt_requests(vcpu, kvm_run);
|
do_interrupt_requests(svm, kvm_run);
|
||||||
|
|
||||||
clgi();
|
clgi();
|
||||||
|
|
||||||
|
@ -1554,7 +1538,7 @@ again:
|
||||||
if (test_and_clear_bit(KVM_TLB_FLUSH, &vcpu->requests))
|
if (test_and_clear_bit(KVM_TLB_FLUSH, &vcpu->requests))
|
||||||
svm_flush_tlb(vcpu);
|
svm_flush_tlb(vcpu);
|
||||||
|
|
||||||
pre_svm_run(vcpu);
|
pre_svm_run(svm);
|
||||||
|
|
||||||
save_host_msrs(vcpu);
|
save_host_msrs(vcpu);
|
||||||
fs_selector = read_fs();
|
fs_selector = read_fs();
|
||||||
|
@ -1714,7 +1698,7 @@ again:
|
||||||
|
|
||||||
stgi();
|
stgi();
|
||||||
|
|
||||||
kvm_reput_irq(vcpu);
|
kvm_reput_irq(svm);
|
||||||
|
|
||||||
svm->next_rip = 0;
|
svm->next_rip = 0;
|
||||||
|
|
||||||
|
@ -1722,29 +1706,29 @@ again:
|
||||||
kvm_run->exit_reason = KVM_EXIT_FAIL_ENTRY;
|
kvm_run->exit_reason = KVM_EXIT_FAIL_ENTRY;
|
||||||
kvm_run->fail_entry.hardware_entry_failure_reason
|
kvm_run->fail_entry.hardware_entry_failure_reason
|
||||||
= svm->vmcb->control.exit_code;
|
= svm->vmcb->control.exit_code;
|
||||||
post_kvm_run_save(vcpu, kvm_run);
|
post_kvm_run_save(svm, kvm_run);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
r = handle_exit(vcpu, kvm_run);
|
r = handle_exit(svm, kvm_run);
|
||||||
if (r > 0) {
|
if (r > 0) {
|
||||||
if (signal_pending(current)) {
|
if (signal_pending(current)) {
|
||||||
++vcpu->stat.signal_exits;
|
++vcpu->stat.signal_exits;
|
||||||
post_kvm_run_save(vcpu, kvm_run);
|
post_kvm_run_save(svm, kvm_run);
|
||||||
kvm_run->exit_reason = KVM_EXIT_INTR;
|
kvm_run->exit_reason = KVM_EXIT_INTR;
|
||||||
return -EINTR;
|
return -EINTR;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (dm_request_for_irq_injection(vcpu, kvm_run)) {
|
if (dm_request_for_irq_injection(svm, kvm_run)) {
|
||||||
++vcpu->stat.request_irq_exits;
|
++vcpu->stat.request_irq_exits;
|
||||||
post_kvm_run_save(vcpu, kvm_run);
|
post_kvm_run_save(svm, kvm_run);
|
||||||
kvm_run->exit_reason = KVM_EXIT_INTR;
|
kvm_run->exit_reason = KVM_EXIT_INTR;
|
||||||
return -EINTR;
|
return -EINTR;
|
||||||
}
|
}
|
||||||
kvm_resched(vcpu);
|
kvm_resched(vcpu);
|
||||||
goto again;
|
goto again;
|
||||||
}
|
}
|
||||||
post_kvm_run_save(vcpu, kvm_run);
|
post_kvm_run_save(svm, kvm_run);
|
||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Add table
Reference in a new issue