Commit 51aa01d1 authored by Avi Kivity's avatar Avi Kivity

KVM: VMX: Split up vmx_complete_interrupts()

vmx_complete_interrupts() does too much, split it up:
 - vmx_vcpu_run() gets the "cache important vmcs fields" part
 - a new vmx_complete_atomic_exit() gets the parts that must be done atomically
 - a new vmx_recover_nmi_blocking() does what its name says
 - vmx_complete_interrupts() retains the event injection recovery code

This helps in reducing the work done in atomic context.
Signed-off-by: default avatarAvi Kivity <avi@redhat.com>
parent 3842d135
...@@ -125,6 +125,7 @@ struct vcpu_vmx { ...@@ -125,6 +125,7 @@ struct vcpu_vmx {
unsigned long host_rsp; unsigned long host_rsp;
int launched; int launched;
u8 fail; u8 fail;
u32 exit_intr_info;
u32 idt_vectoring_info; u32 idt_vectoring_info;
struct shared_msr_entry *guest_msrs; struct shared_msr_entry *guest_msrs;
int nmsrs; int nmsrs;
...@@ -3775,18 +3776,9 @@ static void update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr) ...@@ -3775,18 +3776,9 @@ static void update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr)
vmcs_write32(TPR_THRESHOLD, irr); vmcs_write32(TPR_THRESHOLD, irr);
} }
static void vmx_complete_interrupts(struct vcpu_vmx *vmx) static void vmx_complete_atomic_exit(struct vcpu_vmx *vmx)
{ {
u32 exit_intr_info; u32 exit_intr_info = vmx->exit_intr_info;
u32 idt_vectoring_info = vmx->idt_vectoring_info;
bool unblock_nmi;
u8 vector;
int type;
bool idtv_info_valid;
exit_intr_info = vmcs_read32(VM_EXIT_INTR_INFO);
vmx->exit_reason = vmcs_read32(VM_EXIT_REASON);
/* Handle machine checks before interrupts are enabled */ /* Handle machine checks before interrupts are enabled */
if ((vmx->exit_reason == EXIT_REASON_MCE_DURING_VMENTRY) if ((vmx->exit_reason == EXIT_REASON_MCE_DURING_VMENTRY)
...@@ -3801,8 +3793,16 @@ static void vmx_complete_interrupts(struct vcpu_vmx *vmx) ...@@ -3801,8 +3793,16 @@ static void vmx_complete_interrupts(struct vcpu_vmx *vmx)
asm("int $2"); asm("int $2");
kvm_after_handle_nmi(&vmx->vcpu); kvm_after_handle_nmi(&vmx->vcpu);
} }
}
idtv_info_valid = idt_vectoring_info & VECTORING_INFO_VALID_MASK; static void vmx_recover_nmi_blocking(struct vcpu_vmx *vmx)
{
u32 exit_intr_info = vmx->exit_intr_info;
bool unblock_nmi;
u8 vector;
bool idtv_info_valid;
idtv_info_valid = vmx->idt_vectoring_info & VECTORING_INFO_VALID_MASK;
if (cpu_has_virtual_nmis()) { if (cpu_has_virtual_nmis()) {
unblock_nmi = (exit_intr_info & INTR_INFO_UNBLOCK_NMI) != 0; unblock_nmi = (exit_intr_info & INTR_INFO_UNBLOCK_NMI) != 0;
...@@ -3824,6 +3824,16 @@ static void vmx_complete_interrupts(struct vcpu_vmx *vmx) ...@@ -3824,6 +3824,16 @@ static void vmx_complete_interrupts(struct vcpu_vmx *vmx)
} else if (unlikely(vmx->soft_vnmi_blocked)) } else if (unlikely(vmx->soft_vnmi_blocked))
vmx->vnmi_blocked_time += vmx->vnmi_blocked_time +=
ktime_to_ns(ktime_sub(ktime_get(), vmx->entry_time)); ktime_to_ns(ktime_sub(ktime_get(), vmx->entry_time));
}
static void vmx_complete_interrupts(struct vcpu_vmx *vmx)
{
u32 idt_vectoring_info = vmx->idt_vectoring_info;
u8 vector;
int type;
bool idtv_info_valid;
idtv_info_valid = idt_vectoring_info & VECTORING_INFO_VALID_MASK;
vmx->vcpu.arch.nmi_injected = false; vmx->vcpu.arch.nmi_injected = false;
kvm_clear_exception_queue(&vmx->vcpu); kvm_clear_exception_queue(&vmx->vcpu);
...@@ -4036,6 +4046,11 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu) ...@@ -4036,6 +4046,11 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu)
asm("mov %0, %%ds; mov %0, %%es" : : "r"(__USER_DS)); asm("mov %0, %%ds; mov %0, %%es" : : "r"(__USER_DS));
vmx->launched = 1; vmx->launched = 1;
vmx->exit_reason = vmcs_read32(VM_EXIT_REASON);
vmx->exit_intr_info = vmcs_read32(VM_EXIT_INTR_INFO);
vmx_complete_atomic_exit(vmx);
vmx_recover_nmi_blocking(vmx);
vmx_complete_interrupts(vmx); vmx_complete_interrupts(vmx);
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment