提交 8fe07367 编写于 作者: X Xiantao Zhang 提交者: Avi Kivity

KVM: ia64: Clean up vmm_ivt.S using tab to indent every line

Using tab for indentation for vmm_ivt.S.
Signed-off-by: NXiantao Zhang <xiantao.zhang@intel.com>
Signed-off-by: NAvi Kivity <avi@redhat.com>
上级 9f7d5bb5
/* /*
* /ia64/kvm_ivt.S * arch/ia64/kvm/vmm_ivt.S
* *
* Copyright (C) 1998-2001, 2003 Hewlett-Packard Co * Copyright (C) 1998-2001, 2003 Hewlett-Packard Co
* Stephane Eranian <eranian@hpl.hp.com> * Stephane Eranian <eranian@hpl.hp.com>
...@@ -82,7 +82,7 @@ ...@@ -82,7 +82,7 @@
mov r29=cr.ipsr; \ mov r29=cr.ipsr; \
;; \ ;; \
tbit.z p6,p7=r29,IA64_PSR_VM_BIT; \ tbit.z p6,p7=r29,IA64_PSR_VM_BIT; \
(p7)br.sptk.many kvm_dispatch_reflection; \ (p7) br.sptk.many kvm_dispatch_reflection; \
br.sptk.many kvm_vmm_panic; \ br.sptk.many kvm_vmm_panic; \
GLOBAL_ENTRY(kvm_vmm_panic) GLOBAL_ENTRY(kvm_vmm_panic)
...@@ -115,7 +115,6 @@ ENTRY(kvm_vhpt_miss) ...@@ -115,7 +115,6 @@ ENTRY(kvm_vhpt_miss)
KVM_FAULT(0) KVM_FAULT(0)
END(kvm_vhpt_miss) END(kvm_vhpt_miss)
.org kvm_ia64_ivt+0x400 .org kvm_ia64_ivt+0x400
//////////////////////////////////////////////////////////////// ////////////////////////////////////////////////////////////////
// 0x0400 Entry 1 (size 64 bundles) ITLB (21) // 0x0400 Entry 1 (size 64 bundles) ITLB (21)
...@@ -124,7 +123,7 @@ ENTRY(kvm_itlb_miss) ...@@ -124,7 +123,7 @@ ENTRY(kvm_itlb_miss)
mov r29=cr.ipsr; mov r29=cr.ipsr;
;; ;;
tbit.z p6,p7=r29,IA64_PSR_VM_BIT; tbit.z p6,p7=r29,IA64_PSR_VM_BIT;
(p6) br.sptk kvm_alt_itlb_miss (p6) br.sptk kvm_alt_itlb_miss
mov r19 = 1 mov r19 = 1
br.sptk kvm_itlb_miss_dispatch br.sptk kvm_itlb_miss_dispatch
KVM_FAULT(1); KVM_FAULT(1);
...@@ -138,7 +137,7 @@ ENTRY(kvm_dtlb_miss) ...@@ -138,7 +137,7 @@ ENTRY(kvm_dtlb_miss)
mov r29=cr.ipsr; mov r29=cr.ipsr;
;; ;;
tbit.z p6,p7=r29,IA64_PSR_VM_BIT; tbit.z p6,p7=r29,IA64_PSR_VM_BIT;
(p6)br.sptk kvm_alt_dtlb_miss (p6) br.sptk kvm_alt_dtlb_miss
br.sptk kvm_dtlb_miss_dispatch br.sptk kvm_dtlb_miss_dispatch
END(kvm_dtlb_miss) END(kvm_dtlb_miss)
...@@ -240,7 +239,7 @@ ENTRY(kvm_break_fault) ...@@ -240,7 +239,7 @@ ENTRY(kvm_break_fault)
;; ;;
KVM_SAVE_MIN_WITH_COVER_R19 KVM_SAVE_MIN_WITH_COVER_R19
;; ;;
alloc r14=ar.pfs,0,0,4,0 // now it's safe (must be first in insn group!) alloc r14=ar.pfs,0,0,4,0 //(must be first in insn group!)
mov out0=cr.ifa mov out0=cr.ifa
mov out2=cr.isr // FIXME: pity to make this slow access twice mov out2=cr.isr // FIXME: pity to make this slow access twice
mov out3=cr.iim // FIXME: pity to make this slow access twice mov out3=cr.iim // FIXME: pity to make this slow access twice
...@@ -428,9 +427,9 @@ ENTRY(kvm_virtual_exirq) ...@@ -428,9 +427,9 @@ ENTRY(kvm_virtual_exirq)
kvm_dispatch_vexirq: kvm_dispatch_vexirq:
cmp.eq p6,p0 = 1,r30 cmp.eq p6,p0 = 1,r30
;; ;;
(p6)add r29 = VMM_VCPU_SAVED_GP_OFFSET,r21 (p6) add r29 = VMM_VCPU_SAVED_GP_OFFSET,r21
;; ;;
(p6)ld8 r1 = [r29] (p6) ld8 r1 = [r29]
;; ;;
KVM_SAVE_MIN_WITH_COVER_R19 KVM_SAVE_MIN_WITH_COVER_R19
alloc r14=ar.pfs,0,0,1,0 alloc r14=ar.pfs,0,0,1,0
...@@ -456,13 +455,11 @@ END(kvm_virtual_exirq) ...@@ -456,13 +455,11 @@ END(kvm_virtual_exirq)
KVM_FAULT(14) KVM_FAULT(14)
// this code segment is from 2.6.16.13 // this code segment is from 2.6.16.13
.org kvm_ia64_ivt+0x3c00 .org kvm_ia64_ivt+0x3c00
/////////////////////////////////////////////////////////////////////// ///////////////////////////////////////////////////////////////////////
// 0x3c00 Entry 15 (size 64 bundles) Reserved // 0x3c00 Entry 15 (size 64 bundles) Reserved
KVM_FAULT(15) KVM_FAULT(15)
.org kvm_ia64_ivt+0x4000 .org kvm_ia64_ivt+0x4000
/////////////////////////////////////////////////////////////////////// ///////////////////////////////////////////////////////////////////////
// 0x4000 Entry 16 (size 64 bundles) Reserved // 0x4000 Entry 16 (size 64 bundles) Reserved
...@@ -619,13 +616,13 @@ ENTRY(kvm_virtualization_fault) ...@@ -619,13 +616,13 @@ ENTRY(kvm_virtualization_fault)
cmp.eq p10,p0=EVENT_SSM,r24 cmp.eq p10,p0=EVENT_SSM,r24
cmp.eq p11,p0=EVENT_MOV_TO_PSR,r24 cmp.eq p11,p0=EVENT_MOV_TO_PSR,r24
cmp.eq p12,p0=EVENT_THASH,r24 cmp.eq p12,p0=EVENT_THASH,r24
(p6) br.dptk.many kvm_asm_mov_from_ar (p6) br.dptk.many kvm_asm_mov_from_ar
(p7) br.dptk.many kvm_asm_mov_from_rr (p7) br.dptk.many kvm_asm_mov_from_rr
(p8) br.dptk.many kvm_asm_mov_to_rr (p8) br.dptk.many kvm_asm_mov_to_rr
(p9) br.dptk.many kvm_asm_rsm (p9) br.dptk.many kvm_asm_rsm
(p10) br.dptk.many kvm_asm_ssm (p10) br.dptk.many kvm_asm_ssm
(p11) br.dptk.many kvm_asm_mov_to_psr (p11) br.dptk.many kvm_asm_mov_to_psr
(p12) br.dptk.many kvm_asm_thash (p12) br.dptk.many kvm_asm_thash
;; ;;
kvm_virtualization_fault_back: kvm_virtualization_fault_back:
adds r16 = VMM_VCPU_SAVED_GP_OFFSET,r21 adds r16 = VMM_VCPU_SAVED_GP_OFFSET,r21
...@@ -640,7 +637,7 @@ kvm_virtualization_fault_back: ...@@ -640,7 +637,7 @@ kvm_virtualization_fault_back:
st8 [r17] = r25 st8 [r17] = r25
;; ;;
cmp.ne p6,p0=EVENT_RFI, r24 cmp.ne p6,p0=EVENT_RFI, r24
(p6) br.sptk kvm_dispatch_virtualization_fault (p6) br.sptk kvm_dispatch_virtualization_fault
;; ;;
adds r18=VMM_VPD_BASE_OFFSET,r21 adds r18=VMM_VPD_BASE_OFFSET,r21
;; ;;
...@@ -651,9 +648,9 @@ kvm_virtualization_fault_back: ...@@ -651,9 +648,9 @@ kvm_virtualization_fault_back:
ld8 r18=[r18] ld8 r18=[r18]
;; ;;
tbit.z p6,p0=r18,63 tbit.z p6,p0=r18,63
(p6) br.sptk kvm_dispatch_virtualization_fault (p6) br.sptk kvm_dispatch_virtualization_fault
;; ;;
//if vifs.v=1 desert current register frame //if vifs.v=1 desert current register frame
alloc r18=ar.pfs,0,0,0,0 alloc r18=ar.pfs,0,0,0,0
br.sptk kvm_dispatch_virtualization_fault br.sptk kvm_dispatch_virtualization_fault
END(kvm_virtualization_fault) END(kvm_virtualization_fault)
...@@ -856,7 +853,7 @@ ENTRY(kvm_itlb_miss_dispatch) ...@@ -856,7 +853,7 @@ ENTRY(kvm_itlb_miss_dispatch)
END(kvm_itlb_miss_dispatch) END(kvm_itlb_miss_dispatch)
ENTRY(kvm_dispatch_reflection) ENTRY(kvm_dispatch_reflection)
/* /*
* Input: * Input:
* psr.ic: off * psr.ic: off
* r19: intr type (offset into ivt, see ia64_int.h) * r19: intr type (offset into ivt, see ia64_int.h)
...@@ -893,7 +890,7 @@ ENTRY(kvm_dispatch_virtualization_fault) ...@@ -893,7 +890,7 @@ ENTRY(kvm_dispatch_virtualization_fault)
;; ;;
KVM_SAVE_MIN_WITH_COVER_R19 KVM_SAVE_MIN_WITH_COVER_R19
;; ;;
alloc r14=ar.pfs,0,0,2,0 // now it's safe (must be first in insn group!) alloc r14=ar.pfs,0,0,2,0 // (must be first in insn group!)
mov out0=r13 //vcpu mov out0=r13 //vcpu
adds r3=8,r2 // set up second base pointer adds r3=8,r2 // set up second base pointer
;; ;;
...@@ -917,7 +914,6 @@ ENTRY(kvm_dispatch_interrupt) ...@@ -917,7 +914,6 @@ ENTRY(kvm_dispatch_interrupt)
KVM_SAVE_MIN_WITH_COVER_R19 // uses r31; defines r2 and r3 KVM_SAVE_MIN_WITH_COVER_R19 // uses r31; defines r2 and r3
;; ;;
alloc r14=ar.pfs,0,0,1,0 // must be first in an insn group alloc r14=ar.pfs,0,0,1,0 // must be first in an insn group
//mov out0=cr.ivr // pass cr.ivr as first arg
adds r3=8,r2 // set up second base pointer for SAVE_REST adds r3=8,r2 // set up second base pointer for SAVE_REST
;; ;;
ssm psr.ic ssm psr.ic
...@@ -934,9 +930,6 @@ ENTRY(kvm_dispatch_interrupt) ...@@ -934,9 +930,6 @@ ENTRY(kvm_dispatch_interrupt)
br.call.sptk.many b6=kvm_ia64_handle_irq br.call.sptk.many b6=kvm_ia64_handle_irq
END(kvm_dispatch_interrupt) END(kvm_dispatch_interrupt)
GLOBAL_ENTRY(ia64_leave_nested) GLOBAL_ENTRY(ia64_leave_nested)
rsm psr.i rsm psr.i
;; ;;
...@@ -1065,13 +1058,10 @@ GLOBAL_ENTRY(ia64_leave_nested) ...@@ -1065,13 +1058,10 @@ GLOBAL_ENTRY(ia64_leave_nested)
rfi rfi
END(ia64_leave_nested) END(ia64_leave_nested)
GLOBAL_ENTRY(ia64_leave_hypervisor_prepare) GLOBAL_ENTRY(ia64_leave_hypervisor_prepare)
/* /*
* work.need_resched etc. mustn't get changed * work.need_resched etc. mustn't get changed
*by this CPU before it returns to *by this CPU before it returns to
;;
* user- or fsys-mode, hence we disable interrupts early on: * user- or fsys-mode, hence we disable interrupts early on:
*/ */
adds r2 = PT(R4)+16,r12 adds r2 = PT(R4)+16,r12
...@@ -1293,13 +1283,11 @@ GLOBAL_ENTRY(ia64_vmm_entry) ...@@ -1293,13 +1283,11 @@ GLOBAL_ENTRY(ia64_vmm_entry)
mov r24=r22 mov r24=r22
mov r25=r18 mov r25=r18
tbit.nz p1,p2 = r19,IA64_PSR_IC_BIT // p1=vpsr.ic tbit.nz p1,p2 = r19,IA64_PSR_IC_BIT // p1=vpsr.ic
(p1) br.cond.sptk.few kvm_vps_resume_normal (p1) br.cond.sptk.few kvm_vps_resume_normal
(p2) br.cond.sptk.many kvm_vps_resume_handler (p2) br.cond.sptk.many kvm_vps_resume_handler
;; ;;
END(ia64_vmm_entry) END(ia64_vmm_entry)
/* /*
* extern u64 ia64_call_vsa(u64 proc, u64 arg1, u64 arg2, * extern u64 ia64_call_vsa(u64 proc, u64 arg1, u64 arg2,
* u64 arg3, u64 arg4, u64 arg5, * u64 arg3, u64 arg4, u64 arg5,
...@@ -1340,7 +1328,7 @@ hostret = r24 ...@@ -1340,7 +1328,7 @@ hostret = r24
mov b6=entry mov b6=entry
br.cond.sptk b6 // call the service br.cond.sptk b6 // call the service
2: 2:
// Architectural sequence for enabling interrupts if necessary // Architectural sequence for enabling interrupts if necessary
(p7) ssm psr.ic (p7) ssm psr.ic
;; ;;
(p7) srlz.i (p7) srlz.i
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册