Commit 22ccb142 authored by Xiantao Zhang's avatar Xiantao Zhang Committed by Avi Kivity

KVM: ia64: Code cleanup

Remove some unnecessary blank lines to accord with Kernel's coding style.
Also remove vcpu_get_itir_on_fault due to no reference to it.
Signed-off-by: default avatarXiantao Zhang <xiantao.zhang@intel.com>
Signed-off-by: default avatarAvi Kivity <avi@redhat.com>
parent 989c0f0e
...@@ -167,7 +167,6 @@ static u64 vcpu_get_itir_on_fault(struct kvm_vcpu *vcpu, u64 ifa) ...@@ -167,7 +167,6 @@ static u64 vcpu_get_itir_on_fault(struct kvm_vcpu *vcpu, u64 ifa)
return (rr1.val); return (rr1.val);
} }
/* /*
* Set vIFA & vITIR & vIHA, when vPSR.ic =1 * Set vIFA & vITIR & vIHA, when vPSR.ic =1
* Parameter: * Parameter:
...@@ -222,8 +221,6 @@ void itlb_fault(struct kvm_vcpu *vcpu, u64 vadr) ...@@ -222,8 +221,6 @@ void itlb_fault(struct kvm_vcpu *vcpu, u64 vadr)
inject_guest_interruption(vcpu, IA64_INST_TLB_VECTOR); inject_guest_interruption(vcpu, IA64_INST_TLB_VECTOR);
} }
/* /*
* Data Nested TLB Fault * Data Nested TLB Fault
* @ Data Nested TLB Vector * @ Data Nested TLB Vector
...@@ -245,7 +242,6 @@ void alt_dtlb(struct kvm_vcpu *vcpu, u64 vadr) ...@@ -245,7 +242,6 @@ void alt_dtlb(struct kvm_vcpu *vcpu, u64 vadr)
inject_guest_interruption(vcpu, IA64_ALT_DATA_TLB_VECTOR); inject_guest_interruption(vcpu, IA64_ALT_DATA_TLB_VECTOR);
} }
/* /*
* Data TLB Fault * Data TLB Fault
* @ Data TLB vector * @ Data TLB vector
...@@ -265,8 +261,6 @@ static void _vhpt_fault(struct kvm_vcpu *vcpu, u64 vadr) ...@@ -265,8 +261,6 @@ static void _vhpt_fault(struct kvm_vcpu *vcpu, u64 vadr)
/* If vPSR.ic, IFA, ITIR, IHA*/ /* If vPSR.ic, IFA, ITIR, IHA*/
set_ifa_itir_iha(vcpu, vadr, 1, 1, 1); set_ifa_itir_iha(vcpu, vadr, 1, 1, 1);
inject_guest_interruption(vcpu, IA64_VHPT_TRANS_VECTOR); inject_guest_interruption(vcpu, IA64_VHPT_TRANS_VECTOR);
} }
/* /*
...@@ -279,7 +273,6 @@ void ivhpt_fault(struct kvm_vcpu *vcpu, u64 vadr) ...@@ -279,7 +273,6 @@ void ivhpt_fault(struct kvm_vcpu *vcpu, u64 vadr)
_vhpt_fault(vcpu, vadr); _vhpt_fault(vcpu, vadr);
} }
/* /*
* VHPT Data Fault * VHPT Data Fault
* @ VHPT Translation vector * @ VHPT Translation vector
...@@ -290,8 +283,6 @@ void dvhpt_fault(struct kvm_vcpu *vcpu, u64 vadr) ...@@ -290,8 +283,6 @@ void dvhpt_fault(struct kvm_vcpu *vcpu, u64 vadr)
_vhpt_fault(vcpu, vadr); _vhpt_fault(vcpu, vadr);
} }
/* /*
* Deal with: * Deal with:
* General Exception vector * General Exception vector
...@@ -301,7 +292,6 @@ void _general_exception(struct kvm_vcpu *vcpu) ...@@ -301,7 +292,6 @@ void _general_exception(struct kvm_vcpu *vcpu)
inject_guest_interruption(vcpu, IA64_GENEX_VECTOR); inject_guest_interruption(vcpu, IA64_GENEX_VECTOR);
} }
/* /*
* Illegal Operation Fault * Illegal Operation Fault
* @ General Exception Vector * @ General Exception Vector
...@@ -419,19 +409,16 @@ static void __page_not_present(struct kvm_vcpu *vcpu, u64 vadr) ...@@ -419,19 +409,16 @@ static void __page_not_present(struct kvm_vcpu *vcpu, u64 vadr)
inject_guest_interruption(vcpu, IA64_PAGE_NOT_PRESENT_VECTOR); inject_guest_interruption(vcpu, IA64_PAGE_NOT_PRESENT_VECTOR);
} }
void data_page_not_present(struct kvm_vcpu *vcpu, u64 vadr) void data_page_not_present(struct kvm_vcpu *vcpu, u64 vadr)
{ {
__page_not_present(vcpu, vadr); __page_not_present(vcpu, vadr);
} }
void inst_page_not_present(struct kvm_vcpu *vcpu, u64 vadr) void inst_page_not_present(struct kvm_vcpu *vcpu, u64 vadr)
{ {
__page_not_present(vcpu, vadr); __page_not_present(vcpu, vadr);
} }
/* Deal with /* Deal with
* Data access rights vector * Data access rights vector
*/ */
...@@ -703,7 +690,6 @@ void vhpi_detection(struct kvm_vcpu *vcpu) ...@@ -703,7 +690,6 @@ void vhpi_detection(struct kvm_vcpu *vcpu)
} }
} }
void leave_hypervisor_tail(void) void leave_hypervisor_tail(void)
{ {
struct kvm_vcpu *v = current_vcpu; struct kvm_vcpu *v = current_vcpu;
...@@ -737,7 +723,6 @@ void leave_hypervisor_tail(void) ...@@ -737,7 +723,6 @@ void leave_hypervisor_tail(void)
} }
} }
static inline void handle_lds(struct kvm_pt_regs *regs) static inline void handle_lds(struct kvm_pt_regs *regs)
{ {
regs->cr_ipsr |= IA64_PSR_ED; regs->cr_ipsr |= IA64_PSR_ED;
......
...@@ -112,7 +112,6 @@ void switch_to_physical_rid(struct kvm_vcpu *vcpu) ...@@ -112,7 +112,6 @@ void switch_to_physical_rid(struct kvm_vcpu *vcpu)
return; return;
} }
void switch_to_virtual_rid(struct kvm_vcpu *vcpu) void switch_to_virtual_rid(struct kvm_vcpu *vcpu)
{ {
unsigned long psr; unsigned long psr;
...@@ -166,8 +165,6 @@ void switch_mm_mode(struct kvm_vcpu *vcpu, struct ia64_psr old_psr, ...@@ -166,8 +165,6 @@ void switch_mm_mode(struct kvm_vcpu *vcpu, struct ia64_psr old_psr,
return; return;
} }
/* /*
* In physical mode, insert tc/tr for region 0 and 4 uses * In physical mode, insert tc/tr for region 0 and 4 uses
* RID[0] and RID[4] which is for physical mode emulation. * RID[0] and RID[4] which is for physical mode emulation.
...@@ -269,7 +266,6 @@ static inline unsigned long fph_index(struct kvm_pt_regs *regs, ...@@ -269,7 +266,6 @@ static inline unsigned long fph_index(struct kvm_pt_regs *regs,
return rotate_reg(96, rrb_fr, (regnum - IA64_FIRST_ROTATING_FR)); return rotate_reg(96, rrb_fr, (regnum - IA64_FIRST_ROTATING_FR));
} }
/* /*
* The inverse of the above: given bspstore and the number of * The inverse of the above: given bspstore and the number of
* registers, calculate ar.bsp. * registers, calculate ar.bsp.
...@@ -1039,8 +1035,6 @@ u64 vcpu_tak(struct kvm_vcpu *vcpu, u64 vadr) ...@@ -1039,8 +1035,6 @@ u64 vcpu_tak(struct kvm_vcpu *vcpu, u64 vadr)
return key; return key;
} }
void kvm_thash(struct kvm_vcpu *vcpu, INST64 inst) void kvm_thash(struct kvm_vcpu *vcpu, INST64 inst)
{ {
unsigned long thash, vadr; unsigned long thash, vadr;
...@@ -1050,7 +1044,6 @@ void kvm_thash(struct kvm_vcpu *vcpu, INST64 inst) ...@@ -1050,7 +1044,6 @@ void kvm_thash(struct kvm_vcpu *vcpu, INST64 inst)
vcpu_set_gr(vcpu, inst.M46.r1, thash, 0); vcpu_set_gr(vcpu, inst.M46.r1, thash, 0);
} }
void kvm_ttag(struct kvm_vcpu *vcpu, INST64 inst) void kvm_ttag(struct kvm_vcpu *vcpu, INST64 inst)
{ {
unsigned long tag, vadr; unsigned long tag, vadr;
...@@ -1131,7 +1124,6 @@ int vcpu_tpa(struct kvm_vcpu *vcpu, u64 vadr, u64 *padr) ...@@ -1131,7 +1124,6 @@ int vcpu_tpa(struct kvm_vcpu *vcpu, u64 vadr, u64 *padr)
return IA64_NO_FAULT; return IA64_NO_FAULT;
} }
int kvm_tpa(struct kvm_vcpu *vcpu, INST64 inst) int kvm_tpa(struct kvm_vcpu *vcpu, INST64 inst)
{ {
unsigned long r1, r3; unsigned long r1, r3;
...@@ -1154,7 +1146,6 @@ void kvm_tak(struct kvm_vcpu *vcpu, INST64 inst) ...@@ -1154,7 +1146,6 @@ void kvm_tak(struct kvm_vcpu *vcpu, INST64 inst)
vcpu_set_gr(vcpu, inst.M46.r1, r1, 0); vcpu_set_gr(vcpu, inst.M46.r1, r1, 0);
} }
/************************************ /************************************
* Insert/Purge translation register/cache * Insert/Purge translation register/cache
************************************/ ************************************/
...@@ -1385,7 +1376,6 @@ void kvm_mov_to_ar_reg(struct kvm_vcpu *vcpu, INST64 inst) ...@@ -1385,7 +1376,6 @@ void kvm_mov_to_ar_reg(struct kvm_vcpu *vcpu, INST64 inst)
vcpu_set_itc(vcpu, r2); vcpu_set_itc(vcpu, r2);
} }
void kvm_mov_from_ar_reg(struct kvm_vcpu *vcpu, INST64 inst) void kvm_mov_from_ar_reg(struct kvm_vcpu *vcpu, INST64 inst)
{ {
unsigned long r1; unsigned long r1;
...@@ -1393,8 +1383,9 @@ void kvm_mov_from_ar_reg(struct kvm_vcpu *vcpu, INST64 inst) ...@@ -1393,8 +1383,9 @@ void kvm_mov_from_ar_reg(struct kvm_vcpu *vcpu, INST64 inst)
r1 = vcpu_get_itc(vcpu); r1 = vcpu_get_itc(vcpu);
vcpu_set_gr(vcpu, inst.M31.r1, r1, 0); vcpu_set_gr(vcpu, inst.M31.r1, r1, 0);
} }
/************************************************************************** /**************************************************************************
struct kvm_vcpu*protection key register access routines struct kvm_vcpu protection key register access routines
**************************************************************************/ **************************************************************************/
unsigned long vcpu_get_pkr(struct kvm_vcpu *vcpu, unsigned long reg) unsigned long vcpu_get_pkr(struct kvm_vcpu *vcpu, unsigned long reg)
...@@ -1407,20 +1398,6 @@ void vcpu_set_pkr(struct kvm_vcpu *vcpu, unsigned long reg, unsigned long val) ...@@ -1407,20 +1398,6 @@ void vcpu_set_pkr(struct kvm_vcpu *vcpu, unsigned long reg, unsigned long val)
ia64_set_pkr(reg, val); ia64_set_pkr(reg, val);
} }
unsigned long vcpu_get_itir_on_fault(struct kvm_vcpu *vcpu, unsigned long ifa)
{
union ia64_rr rr, rr1;
rr.val = vcpu_get_rr(vcpu, ifa);
rr1.val = 0;
rr1.ps = rr.ps;
rr1.rid = rr.rid;
return (rr1.val);
}
/******************************** /********************************
* Moves to privileged registers * Moves to privileged registers
********************************/ ********************************/
...@@ -1464,8 +1441,6 @@ unsigned long vcpu_set_rr(struct kvm_vcpu *vcpu, unsigned long reg, ...@@ -1464,8 +1441,6 @@ unsigned long vcpu_set_rr(struct kvm_vcpu *vcpu, unsigned long reg,
return (IA64_NO_FAULT); return (IA64_NO_FAULT);
} }
void kvm_mov_to_rr(struct kvm_vcpu *vcpu, INST64 inst) void kvm_mov_to_rr(struct kvm_vcpu *vcpu, INST64 inst)
{ {
unsigned long r3, r2; unsigned long r3, r2;
...@@ -1510,8 +1485,6 @@ void kvm_mov_to_pkr(struct kvm_vcpu *vcpu, INST64 inst) ...@@ -1510,8 +1485,6 @@ void kvm_mov_to_pkr(struct kvm_vcpu *vcpu, INST64 inst)
vcpu_set_pkr(vcpu, r3, r2); vcpu_set_pkr(vcpu, r3, r2);
} }
void kvm_mov_from_rr(struct kvm_vcpu *vcpu, INST64 inst) void kvm_mov_from_rr(struct kvm_vcpu *vcpu, INST64 inst)
{ {
unsigned long r3, r1; unsigned long r3, r1;
...@@ -1557,7 +1530,6 @@ void kvm_mov_from_pmc(struct kvm_vcpu *vcpu, INST64 inst) ...@@ -1557,7 +1530,6 @@ void kvm_mov_from_pmc(struct kvm_vcpu *vcpu, INST64 inst)
vcpu_set_gr(vcpu, inst.M43.r1, r1, 0); vcpu_set_gr(vcpu, inst.M43.r1, r1, 0);
} }
unsigned long vcpu_get_cpuid(struct kvm_vcpu *vcpu, unsigned long reg) unsigned long vcpu_get_cpuid(struct kvm_vcpu *vcpu, unsigned long reg)
{ {
/* FIXME: This could get called as a result of a rsvd-reg fault */ /* FIXME: This could get called as a result of a rsvd-reg fault */
...@@ -1609,7 +1581,6 @@ unsigned long kvm_mov_to_cr(struct kvm_vcpu *vcpu, INST64 inst) ...@@ -1609,7 +1581,6 @@ unsigned long kvm_mov_to_cr(struct kvm_vcpu *vcpu, INST64 inst)
return 0; return 0;
} }
unsigned long kvm_mov_from_cr(struct kvm_vcpu *vcpu, INST64 inst) unsigned long kvm_mov_from_cr(struct kvm_vcpu *vcpu, INST64 inst)
{ {
unsigned long tgt = inst.M33.r1; unsigned long tgt = inst.M33.r1;
...@@ -1633,8 +1604,6 @@ unsigned long kvm_mov_from_cr(struct kvm_vcpu *vcpu, INST64 inst) ...@@ -1633,8 +1604,6 @@ unsigned long kvm_mov_from_cr(struct kvm_vcpu *vcpu, INST64 inst)
return 0; return 0;
} }
void vcpu_set_psr(struct kvm_vcpu *vcpu, unsigned long val) void vcpu_set_psr(struct kvm_vcpu *vcpu, unsigned long val)
{ {
...@@ -1776,9 +1745,6 @@ void vcpu_bsw1(struct kvm_vcpu *vcpu) ...@@ -1776,9 +1745,6 @@ void vcpu_bsw1(struct kvm_vcpu *vcpu)
} }
} }
void vcpu_rfi(struct kvm_vcpu *vcpu) void vcpu_rfi(struct kvm_vcpu *vcpu)
{ {
unsigned long ifs, psr; unsigned long ifs, psr;
...@@ -1796,7 +1762,6 @@ void vcpu_rfi(struct kvm_vcpu *vcpu) ...@@ -1796,7 +1762,6 @@ void vcpu_rfi(struct kvm_vcpu *vcpu)
regs->cr_iip = VCPU(vcpu, iip); regs->cr_iip = VCPU(vcpu, iip);
} }
/* /*
VPSR can't keep track of below bits of guest PSR VPSR can't keep track of below bits of guest PSR
This function gets guest PSR This function gets guest PSR
......
...@@ -509,7 +509,6 @@ void thash_purge_all(struct kvm_vcpu *v) ...@@ -509,7 +509,6 @@ void thash_purge_all(struct kvm_vcpu *v)
local_flush_tlb_all(); local_flush_tlb_all();
} }
/* /*
* Lookup the hash table and its collision chain to find an entry * Lookup the hash table and its collision chain to find an entry
* covering this address rid:va or the entry. * covering this address rid:va or the entry.
...@@ -517,7 +516,6 @@ void thash_purge_all(struct kvm_vcpu *v) ...@@ -517,7 +516,6 @@ void thash_purge_all(struct kvm_vcpu *v)
* INPUT: * INPUT:
* in: TLB format for both VHPT & TLB. * in: TLB format for both VHPT & TLB.
*/ */
struct thash_data *vtlb_lookup(struct kvm_vcpu *v, u64 va, int is_data) struct thash_data *vtlb_lookup(struct kvm_vcpu *v, u64 va, int is_data)
{ {
struct thash_data *cch; struct thash_data *cch;
...@@ -547,7 +545,6 @@ struct thash_data *vtlb_lookup(struct kvm_vcpu *v, u64 va, int is_data) ...@@ -547,7 +545,6 @@ struct thash_data *vtlb_lookup(struct kvm_vcpu *v, u64 va, int is_data)
return NULL; return NULL;
} }
/* /*
* Initialize internal control data before service. * Initialize internal control data before service.
*/ */
...@@ -589,7 +586,6 @@ u64 kvm_gpa_to_mpa(u64 gpa) ...@@ -589,7 +586,6 @@ u64 kvm_gpa_to_mpa(u64 gpa)
return (pte >> PAGE_SHIFT << PAGE_SHIFT) | (gpa & ~PAGE_MASK); return (pte >> PAGE_SHIFT << PAGE_SHIFT) | (gpa & ~PAGE_MASK);
} }
/* /*
* Fetch guest bundle code. * Fetch guest bundle code.
* INPUT: * INPUT:
...@@ -631,7 +627,6 @@ int fetch_code(struct kvm_vcpu *vcpu, u64 gip, IA64_BUNDLE *pbundle) ...@@ -631,7 +627,6 @@ int fetch_code(struct kvm_vcpu *vcpu, u64 gip, IA64_BUNDLE *pbundle)
return IA64_NO_FAULT; return IA64_NO_FAULT;
} }
void kvm_init_vhpt(struct kvm_vcpu *v) void kvm_init_vhpt(struct kvm_vcpu *v)
{ {
v->arch.vhpt.num = VHPT_NUM_ENTRIES; v->arch.vhpt.num = VHPT_NUM_ENTRIES;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment