entry-armv.S 29.0 KB
Newer Older
L
Linus Torvalds 已提交
1 2 3 4 5
/*
 *  linux/arch/arm/kernel/entry-armv.S
 *
 *  Copyright (C) 1996,1997,1998 Russell King.
 *  ARM700 fix by Matthew Godbolt (linux-user@willothewisp.demon.co.uk)
6
 *  nommu support by Hyok S. Choi (hyok.choi@samsung.com)
L
Linus Torvalds 已提交
7 8 9 10 11 12 13
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License version 2 as
 * published by the Free Software Foundation.
 *
 *  Low-level vector interface routines
 *
14 15
 *  Note:  there is a StrongARM bug in the STMIA rn, {regs}^ instruction
 *  that causes it to save wrong values...  Be aware!
L
Linus Torvalds 已提交
16 17
 */

18
#include <asm/memory.h>
L
Linus Torvalds 已提交
19 20
#include <asm/glue.h>
#include <asm/vfpmacros.h>
21
#include <mach/entry-macro.S>
22
#include <asm/thread_notify.h>
23
#include <asm/unwind.h>
L
Linus Torvalds 已提交
24 25 26

#include "entry-header.S"

27 28 29 30
/*
 * Interrupt handling.  Preserves r7, r8, r9
 */
	.macro	irq_handler
31
	get_irqnr_preamble r5, lr
32 33 34 35 36
1:	get_irqnr_and_base r0, r6, r5, lr
	movne	r1, sp
	@
	@ routine called with r0 = irq number, r1 = struct pt_regs *
	@
37
	adrne	lr, BSYM(1b)
38
	bne	asm_do_IRQ
R
Russell King 已提交
39 40 41 42 43 44 45 46 47 48

#ifdef CONFIG_SMP
	/*
	 * XXX
	 *
	 * this macro assumes that irqstat (r6) and base (r5) are
	 * preserved from get_irqnr_and_base above
	 */
	test_for_ipi r0, r6, r5, lr
	movne	r0, sp
49
	adrne	lr, BSYM(1b)
R
Russell King 已提交
50
	bne	do_IPI
51 52 53 54

#ifdef CONFIG_LOCAL_TIMERS
	test_for_ltirq r0, r6, r5, lr
	movne	r0, sp
55
	adrne	lr, BSYM(1b)
56 57
	bne	do_local_timer
#endif
R
Russell King 已提交
58 59
#endif

60 61
	.endm

62 63 64 65 66 67
#ifdef CONFIG_KPROBES
	.section	.kprobes.text,"ax",%progbits
#else
	.text
#endif

L
Linus Torvalds 已提交
68 69 70
/*
 * Invalid mode handlers
 */
R
Russell King 已提交
71 72
	.macro	inv_entry, reason
	sub	sp, sp, #S_FRAME_SIZE
73 74 75 76
 ARM(	stmib	sp, {r1 - lr}		)
 THUMB(	stmia	sp, {r0 - r12}		)
 THUMB(	str	sp, [sp, #S_SP]		)
 THUMB(	str	lr, [sp, #S_LR]		)
L
Linus Torvalds 已提交
77 78 79 80
	mov	r1, #\reason
	.endm

__pabt_invalid:
R
Russell King 已提交
81 82
	inv_entry BAD_PREFETCH
	b	common_invalid
83
ENDPROC(__pabt_invalid)
L
Linus Torvalds 已提交
84 85

__dabt_invalid:
R
Russell King 已提交
86 87
	inv_entry BAD_DATA
	b	common_invalid
88
ENDPROC(__dabt_invalid)
L
Linus Torvalds 已提交
89 90

__irq_invalid:
R
Russell King 已提交
91 92
	inv_entry BAD_IRQ
	b	common_invalid
93
ENDPROC(__irq_invalid)
L
Linus Torvalds 已提交
94 95

__und_invalid:
R
Russell King 已提交
96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113
	inv_entry BAD_UNDEFINSTR

	@
	@ XXX fall through to common_invalid
	@

@
@ common_invalid - generic code for failed exception (re-entrant version of handlers)
@
common_invalid:
	zero_fp

	ldmia	r0, {r4 - r6}
	add	r0, sp, #S_PC		@ here for interlock avoidance
	mov	r7, #-1			@  ""   ""    ""        ""
	str	r4, [sp]		@ save preserved r0
	stmia	r0, {r5 - r7}		@ lr_<exception>,
					@ cpsr_<exception>, "old_r0"
L
Linus Torvalds 已提交
114 115 116

	mov	r0, sp
	b	bad_mode
117
ENDPROC(__und_invalid)
L
Linus Torvalds 已提交
118 119 120 121

/*
 * SVC mode handlers
 */
122 123 124 125 126 127 128

#if defined(CONFIG_AEABI) && (__LINUX_ARM_ARCH__ >= 5)
#define SPFIX(code...) code
#else
#define SPFIX(code...)
#endif

129
	.macro	svc_entry, stack_hole=0
130 131
 UNWIND(.fnstart		)
 UNWIND(.save {r0 - pc}		)
132 133 134 135 136 137 138
	sub	sp, sp, #(S_FRAME_SIZE + \stack_hole - 4)
#ifdef CONFIG_THUMB2_KERNEL
 SPFIX(	str	r0, [sp]	)	@ temporarily saved
 SPFIX(	mov	r0, sp		)
 SPFIX(	tst	r0, #4		)	@ test original stack alignment
 SPFIX(	ldr	r0, [sp]	)	@ restored
#else
139
 SPFIX(	tst	sp, #4		)
140 141 142
#endif
 SPFIX(	subeq	sp, sp, #4	)
	stmia	sp, {r1 - r12}
R
Russell King 已提交
143 144

	ldmia	r0, {r1 - r3}
145
	add	r5, sp, #S_SP - 4	@ here for interlock avoidance
R
Russell King 已提交
146
	mov	r4, #-1			@  ""  ""      ""       ""
147 148 149
	add	r0, sp, #(S_FRAME_SIZE + \stack_hole - 4)
 SPFIX(	addeq	r0, r0, #4	)
	str	r1, [sp, #-4]!		@ save the "real" r0 copied
R
Russell King 已提交
150 151
					@ from the exception stack

L
Linus Torvalds 已提交
152 153 154 155 156 157 158 159 160 161 162 163
	mov	r1, lr

	@
	@ We are now ready to fill in the remaining blanks on the stack:
	@
	@  r0 - sp_svc
	@  r1 - lr_svc
	@  r2 - lr_<exception>, already fixed up for correct return/restart
	@  r3 - spsr_<exception>
	@  r4 - orig_r0 (see pt_regs definition in ptrace.h)
	@
	stmia	r5, {r0 - r4}
164 165

	asm_trace_hardirqs_off
L
Linus Torvalds 已提交
166 167 168 169
	.endm

	.align	5
__dabt_svc:
R
Russell King 已提交
170
	svc_entry
L
Linus Torvalds 已提交
171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187

	@
	@ get ready to re-enable interrupts if appropriate
	@
	mrs	r9, cpsr
	tst	r3, #PSR_I_BIT
	biceq	r9, r9, #PSR_I_BIT

	@
	@ Call the processor-specific abort handler:
	@
	@  r2 - aborted context pc
	@  r3 - aborted context cpsr
	@
	@ The abort handler must return the aborted address in r0, and
	@ the fault status register in r1.  r9 must be preserved.
	@
P
Paul Brook 已提交
188
#ifdef MULTI_DABORT
L
Linus Torvalds 已提交
189 190
	ldr	r4, .LCprocfns
	mov	lr, pc
P
Paul Brook 已提交
191
	ldr	pc, [r4, #PROCESSOR_DABT_FUNC]
L
Linus Torvalds 已提交
192
#else
P
Paul Brook 已提交
193
	bl	CPU_DABORT_HANDLER
L
Linus Torvalds 已提交
194 195 196 197 198 199 200 201 202 203 204 205
#endif

	@
	@ set desired IRQ state, then call main handler
	@
	msr	cpsr_c, r9
	mov	r2, sp
	bl	do_DataAbort

	@
	@ IRQs off again before pulling preserved data off the stack
	@
206
	disable_irq
L
Linus Torvalds 已提交
207 208 209 210

	@
	@ restore SPSR and restart the instruction
	@
211 212
	ldr	r2, [sp, #S_PSR]
	svc_exit r2				@ return from exception
213
 UNWIND(.fnend		)
214
ENDPROC(__dabt_svc)
L
Linus Torvalds 已提交
215 216 217

	.align	5
__irq_svc:
R
Russell King 已提交
218 219
	svc_entry

L
Linus Torvalds 已提交
220
#ifdef CONFIG_PREEMPT
221 222 223 224
	get_thread_info tsk
	ldr	r8, [tsk, #TI_PREEMPT]		@ get preempt count
	add	r7, r8, #1			@ increment it
	str	r7, [tsk, #TI_PREEMPT]
L
Linus Torvalds 已提交
225
#endif
R
Russell King 已提交
226

227
	irq_handler
L
Linus Torvalds 已提交
228
#ifdef CONFIG_PREEMPT
R
Russell King 已提交
229
	str	r8, [tsk, #TI_PREEMPT]		@ restore preempt count
230
	ldr	r0, [tsk, #TI_FLAGS]		@ get flags
R
Russell King 已提交
231 232
	teq	r8, #0				@ if preempt count != 0
	movne	r0, #0				@ force flags to 0
L
Linus Torvalds 已提交
233 234 235
	tst	r0, #_TIF_NEED_RESCHED
	blne	svc_preempt
#endif
236
	ldr	r4, [sp, #S_PSR]		@ irqs are already disabled
R
Russell King 已提交
237
#ifdef CONFIG_TRACE_IRQFLAGS
238
	tst	r4, #PSR_I_BIT
R
Russell King 已提交
239 240
	bleq	trace_hardirqs_on
#endif
241
	svc_exit r4				@ return from exception
242
 UNWIND(.fnend		)
243
ENDPROC(__irq_svc)
L
Linus Torvalds 已提交
244 245 246 247 248

	.ltorg

#ifdef CONFIG_PREEMPT
svc_preempt:
R
Russell King 已提交
249
	mov	r8, lr
L
Linus Torvalds 已提交
250
1:	bl	preempt_schedule_irq		@ irq en/disable is done inside
251
	ldr	r0, [tsk, #TI_FLAGS]		@ get new tasks TI_FLAGS
L
Linus Torvalds 已提交
252
	tst	r0, #_TIF_NEED_RESCHED
R
Russell King 已提交
253
	moveq	pc, r8				@ go again
L
Linus Torvalds 已提交
254 255 256 257 258
	b	1b
#endif

	.align	5
__und_svc:
259 260 261 262 263 264
#ifdef CONFIG_KPROBES
	@ If a kprobe is about to simulate a "stmdb sp..." instruction,
	@ it obviously needs free stack space which then will belong to
	@ the saved context.
	svc_entry 64
#else
R
Russell King 已提交
265
	svc_entry
266
#endif
L
Linus Torvalds 已提交
267 268 269 270 271 272 273 274

	@
	@ call emulation code, which returns using r9 if it has emulated
	@ the instruction, or the more conventional lr if we are to treat
	@ this as a real undefined instruction
	@
	@  r0 - instruction
	@
275
#ifndef	CONFIG_THUMB2_KERNEL
L
Linus Torvalds 已提交
276
	ldr	r0, [r2, #-4]
277 278 279 280 281 282 283
#else
	ldrh	r0, [r2, #-2]			@ Thumb instruction at LR - 2
	and	r9, r0, #0xf800
	cmp	r9, #0xe800			@ 32-bit instruction if xx >= 0
	ldrhhs	r9, [r2]			@ bottom 16 bits
	orrhs	r0, r9, r0, lsl #16
#endif
284
	adr	r9, BSYM(1f)
L
Linus Torvalds 已提交
285 286 287 288 289 290 291 292
	bl	call_fpe

	mov	r0, sp				@ struct pt_regs *regs
	bl	do_undefinstr

	@
	@ IRQs off again before pulling preserved data off the stack
	@
293
1:	disable_irq
L
Linus Torvalds 已提交
294 295 296 297

	@
	@ restore SPSR and restart the instruction
	@
298 299
	ldr	r2, [sp, #S_PSR]		@ Get SVC cpsr
	svc_exit r2				@ return from exception
300
 UNWIND(.fnend		)
301
ENDPROC(__und_svc)
L
Linus Torvalds 已提交
302 303 304

	.align	5
__pabt_svc:
R
Russell King 已提交
305
	svc_entry
L
Linus Torvalds 已提交
306 307 308 309 310 311 312 313

	@
	@ re-enable interrupts if appropriate
	@
	mrs	r9, cpsr
	tst	r3, #PSR_I_BIT
	biceq	r9, r9, #PSR_I_BIT

P
Paul Brook 已提交
314
	mov	r0, r2			@ pass address of aborted instruction.
315
#ifdef MULTI_PABORT
P
Paul Brook 已提交
316 317 318 319
	ldr	r4, .LCprocfns
	mov	lr, pc
	ldr	pc, [r4, #PROCESSOR_PABT_FUNC]
#else
320
	bl	CPU_PABORT_HANDLER
P
Paul Brook 已提交
321 322
#endif
	msr	cpsr_c, r9			@ Maybe enable interrupts
323
	mov	r2, sp				@ regs
L
Linus Torvalds 已提交
324 325 326 327 328
	bl	do_PrefetchAbort		@ call abort handler

	@
	@ IRQs off again before pulling preserved data off the stack
	@
329
	disable_irq
L
Linus Torvalds 已提交
330 331 332 333

	@
	@ restore SPSR and restart the instruction
	@
334 335
	ldr	r2, [sp, #S_PSR]
	svc_exit r2				@ return from exception
336
 UNWIND(.fnend		)
337
ENDPROC(__pabt_svc)
L
Linus Torvalds 已提交
338 339

	.align	5
340 341
.LCcralign:
	.word	cr_alignment
P
Paul Brook 已提交
342
#ifdef MULTI_DABORT
L
Linus Torvalds 已提交
343 344 345 346 347 348 349 350
.LCprocfns:
	.word	processor
#endif
.LCfp:
	.word	fp_enter

/*
 * User mode handlers
351 352
 *
 * EABI note: sp_svc is always 64-bit aligned here, so should S_FRAME_SIZE
L
Linus Torvalds 已提交
353
 */
354 355 356 357 358

#if defined(CONFIG_AEABI) && (__LINUX_ARM_ARCH__ >= 5) && (S_FRAME_SIZE & 7)
#error "sizeof(struct pt_regs) must be a multiple of 8"
#endif

R
Russell King 已提交
359
	.macro	usr_entry
360 361
 UNWIND(.fnstart	)
 UNWIND(.cantunwind	)	@ don't unwind the user space
R
Russell King 已提交
362
	sub	sp, sp, #S_FRAME_SIZE
363 364
 ARM(	stmib	sp, {r1 - r12}	)
 THUMB(	stmia	sp, {r0 - r12}	)
R
Russell King 已提交
365 366 367 368 369 370 371

	ldmia	r0, {r1 - r3}
	add	r0, sp, #S_PC		@ here for interlock avoidance
	mov	r4, #-1			@  ""  ""     ""        ""

	str	r1, [sp]		@ save the "real" r0 copied
					@ from the exception stack
L
Linus Torvalds 已提交
372 373 374 375 376 377 378 379 380 381

	@
	@ We are now ready to fill in the remaining blanks on the stack:
	@
	@  r2 - lr_<exception>, already fixed up for correct return/restart
	@  r3 - spsr_<exception>
	@  r4 - orig_r0 (see pt_regs definition in ptrace.h)
	@
	@ Also, separately save sp_usr and lr_usr
	@
R
Russell King 已提交
382
	stmia	r0, {r2 - r4}
383 384
 ARM(	stmdb	r0, {sp, lr}^			)
 THUMB(	store_user_sp_lr r0, r1, S_SP - S_PC	)
L
Linus Torvalds 已提交
385 386 387 388

	@
	@ Enable the alignment trap while in kernel mode
	@
389
	alignment_trap r0
L
Linus Torvalds 已提交
390 391 392 393 394

	@
	@ Clear FP to mark the first stack frame
	@
	zero_fp
395 396

	asm_trace_hardirqs_off
L
Linus Torvalds 已提交
397 398
	.endm

399 400 401 402 403 404 405 406 407 408 409 410 411 412 413
	.macro	kuser_cmpxchg_check
#if __LINUX_ARM_ARCH__ < 6 && !defined(CONFIG_NEEDS_SYSCALL_FOR_CMPXCHG)
#ifndef CONFIG_MMU
#warning "NPTL on non MMU needs fixing"
#else
	@ Make sure our user space atomic helper is restarted
	@ if it was interrupted in a critical region.  Here we
	@ perform a quick test inline since it should be false
	@ 99.9999% of the time.  The rest is done out of line.
	cmp	r2, #TASK_SIZE
	blhs	kuser_cmpxchg_fixup
#endif
#endif
	.endm

L
Linus Torvalds 已提交
414 415
	.align	5
__dabt_usr:
R
Russell King 已提交
416
	usr_entry
417
	kuser_cmpxchg_check
L
Linus Torvalds 已提交
418 419 420 421 422 423 424 425 426 427

	@
	@ Call the processor-specific abort handler:
	@
	@  r2 - aborted context pc
	@  r3 - aborted context cpsr
	@
	@ The abort handler must return the aborted address in r0, and
	@ the fault status register in r1.
	@
P
Paul Brook 已提交
428
#ifdef MULTI_DABORT
L
Linus Torvalds 已提交
429 430
	ldr	r4, .LCprocfns
	mov	lr, pc
P
Paul Brook 已提交
431
	ldr	pc, [r4, #PROCESSOR_DABT_FUNC]
L
Linus Torvalds 已提交
432
#else
P
Paul Brook 已提交
433
	bl	CPU_DABORT_HANDLER
L
Linus Torvalds 已提交
434 435 436 437 438
#endif

	@
	@ IRQs on, then call the main handler
	@
439
	enable_irq
L
Linus Torvalds 已提交
440
	mov	r2, sp
441
	adr	lr, BSYM(ret_from_exception)
L
Linus Torvalds 已提交
442
	b	do_DataAbort
443
 UNWIND(.fnend		)
444
ENDPROC(__dabt_usr)
L
Linus Torvalds 已提交
445 446 447

	.align	5
__irq_usr:
R
Russell King 已提交
448
	usr_entry
449
	kuser_cmpxchg_check
L
Linus Torvalds 已提交
450

451
	get_thread_info tsk
L
Linus Torvalds 已提交
452
#ifdef CONFIG_PREEMPT
453 454 455
	ldr	r8, [tsk, #TI_PREEMPT]		@ get preempt count
	add	r7, r8, #1			@ increment it
	str	r7, [tsk, #TI_PREEMPT]
L
Linus Torvalds 已提交
456
#endif
R
Russell King 已提交
457

458
	irq_handler
L
Linus Torvalds 已提交
459
#ifdef CONFIG_PREEMPT
460 461
	ldr	r0, [tsk, #TI_PREEMPT]
	str	r8, [tsk, #TI_PREEMPT]
L
Linus Torvalds 已提交
462
	teq	r0, r7
463 464 465
 ARM(	strne	r0, [r0, -r0]	)
 THUMB(	movne	r0, #0		)
 THUMB(	strne	r0, [r0]	)
L
Linus Torvalds 已提交
466
#endif
R
Russell King 已提交
467 468 469
#ifdef CONFIG_TRACE_IRQFLAGS
	bl	trace_hardirqs_on
#endif
R
Russell King 已提交
470

L
Linus Torvalds 已提交
471 472
	mov	why, #0
	b	ret_to_user
473
 UNWIND(.fnend		)
474
ENDPROC(__irq_usr)
L
Linus Torvalds 已提交
475 476 477 478 479

	.ltorg

	.align	5
__und_usr:
R
Russell King 已提交
480
	usr_entry
L
Linus Torvalds 已提交
481 482 483 484 485 486 487 488

	@
	@ fall through to the emulation code, which returns using r9 if
	@ it has emulated the instruction, or the more conventional lr
	@ if we are to treat this as a real undefined instruction
	@
	@  r0 - instruction
	@
489 490
	adr	r9, BSYM(ret_from_exception)
	adr	lr, BSYM(__und_usr_unknown)
491
	tst	r3, #PSR_T_BIT			@ Thumb mode?
492
	itet	eq				@ explicit IT needed for the 1f label
493 494 495
	subeq	r4, r2, #4			@ ARM instr at LR - 4
	subne	r4, r2, #2			@ Thumb instr at LR - 2
1:	ldreqt	r0, [r4]
496 497 498
#ifdef CONFIG_CPU_ENDIAN_BE8
	reveq	r0, r0				@ little endian instruction
#endif
499 500 501
	beq	call_fpe
	@ Thumb instruction
#if __LINUX_ARM_ARCH__ >= 7
502 503 504 505
2:
 ARM(	ldrht	r5, [r4], #2	)
 THUMB(	ldrht	r5, [r4]	)
 THUMB(	add	r4, r4, #2	)
506 507 508 509 510 511 512 513 514
	and	r0, r5, #0xf800			@ mask bits 111x x... .... ....
	cmp	r0, #0xe800			@ 32bit instruction if xx != 0
	blo	__und_usr_unknown
3:	ldrht	r0, [r4]
	add	r2, r2, #2			@ r2 is PC + 2, make it PC + 4
	orr	r0, r0, r5, lsl #16
#else
	b	__und_usr_unknown
#endif
515
 UNWIND(.fnend		)
516
ENDPROC(__und_usr)
517

L
Linus Torvalds 已提交
518 519 520 521 522 523 524 525
	@
	@ fallthrough to call_fpe
	@

/*
 * The out of line fixup for the ldrt above.
 */
	.section .fixup, "ax"
526
4:	mov	pc, r9
L
Linus Torvalds 已提交
527 528
	.previous
	.section __ex_table,"a"
529 530 531 532 533
	.long	1b, 4b
#if __LINUX_ARM_ARCH__ >= 7
	.long	2b, 4b
	.long	3b, 4b
#endif
L
Linus Torvalds 已提交
534 535 536 537 538 539 540 541 542 543 544 545
	.previous

/*
 * Check whether the instruction is a co-processor instruction.
 * If yes, we need to call the relevant co-processor handler.
 *
 * Note that we don't do a full check here for the co-processor
 * instructions; all instructions with bit 27 set are well
 * defined.  The only instructions that should fault are the
 * co-processor instructions.  However, we have to watch out
 * for the ARM6/ARM7 SWI bug.
 *
546 547 548 549 550 551 552
 * NEON is a special case that has to be handled here. Not all
 * NEON instructions are co-processor instructions, so we have
 * to make a special case of checking for them. Plus, there's
 * five groups of them, so we have a table of mask/opcode pairs
 * to check against, and if any match then we branch off into the
 * NEON handler code.
 *
L
Linus Torvalds 已提交
553 554 555
 * Emulators may wish to make use of the following registers:
 *  r0  = instruction opcode.
 *  r2  = PC+4
556
 *  r9  = normal "successful" return address
L
Linus Torvalds 已提交
557
 *  r10 = this threads thread_info structure.
558
 *  lr  = unrecognised instruction return address
L
Linus Torvalds 已提交
559
 */
560 561 562 563 564 565 566
	@
	@ Fall-through from Thumb-2 __und_usr
	@
#ifdef CONFIG_NEON
	adr	r6, .LCneon_thumb_opcodes
	b	2f
#endif
L
Linus Torvalds 已提交
567
call_fpe:
568
#ifdef CONFIG_NEON
569
	adr	r6, .LCneon_arm_opcodes
570 571 572 573 574 575 576 577 578 579 580 581 582 583 584
2:
	ldr	r7, [r6], #4			@ mask value
	cmp	r7, #0				@ end mask?
	beq	1f
	and	r8, r0, r7
	ldr	r7, [r6], #4			@ opcode bits matching in mask
	cmp	r8, r7				@ NEON instruction?
	bne	2b
	get_thread_info r10
	mov	r7, #1
	strb	r7, [r10, #TI_USED_CP + 10]	@ mark CP#10 as used
	strb	r7, [r10, #TI_USED_CP + 11]	@ mark CP#11 as used
	b	do_vfp				@ let VFP handler handle this
1:
#endif
L
Linus Torvalds 已提交
585
	tst	r0, #0x08000000			@ only CDP/CPRT/LDC/STC have bit 27
586
	tstne	r0, #0x04000000			@ bit 26 set on both ARM and Thumb-2
L
Linus Torvalds 已提交
587 588 589 590 591 592 593
#if defined(CONFIG_CPU_ARM610) || defined(CONFIG_CPU_ARM710)
	and	r8, r0, #0x0f000000		@ mask out op-code bits
	teqne	r8, #0x0f000000			@ SWI (ARM6/7 bug)?
#endif
	moveq	pc, lr
	get_thread_info r10			@ get current thread
	and	r8, r0, #0x00000f00		@ mask out CP number
594
 THUMB(	lsr	r8, r8, #8		)
L
Linus Torvalds 已提交
595 596
	mov	r7, #1
	add	r6, r10, #TI_USED_CP
597 598
 ARM(	strb	r7, [r6, r8, lsr #8]	)	@ set appropriate used_cp[]
 THUMB(	strb	r7, [r6, r8]		)	@ set appropriate used_cp[]
L
Linus Torvalds 已提交
599 600 601 602 603 604 605
#ifdef CONFIG_IWMMXT
	@ Test if we need to give access to iWMMXt coprocessors
	ldr	r5, [r10, #TI_FLAGS]
	rsbs	r7, r8, #(1 << 8)		@ CP 0 or 1 only
	movcss	r7, r5, lsr #(TIF_USING_IWMMXT + 1)
	bcs	iwmmxt_task_enable
#endif
606 607 608 609 610 611 612 613 614
 ARM(	add	pc, pc, r8, lsr #6	)
 THUMB(	lsl	r8, r8, #2		)
 THUMB(	add	pc, r8			)
	nop

	W(mov)	pc, lr				@ CP#0
	W(b)	do_fpe				@ CP#1 (FPE)
	W(b)	do_fpe				@ CP#2 (FPE)
	W(mov)	pc, lr				@ CP#3
615 616 617 618 619
#ifdef CONFIG_CRUNCH
	b	crunch_task_enable		@ CP#4 (MaverickCrunch)
	b	crunch_task_enable		@ CP#5 (MaverickCrunch)
	b	crunch_task_enable		@ CP#6 (MaverickCrunch)
#else
620 621 622
	W(mov)	pc, lr				@ CP#4
	W(mov)	pc, lr				@ CP#5
	W(mov)	pc, lr				@ CP#6
623
#endif
624 625 626
	W(mov)	pc, lr				@ CP#7
	W(mov)	pc, lr				@ CP#8
	W(mov)	pc, lr				@ CP#9
L
Linus Torvalds 已提交
627
#ifdef CONFIG_VFP
628 629
	W(b)	do_vfp				@ CP#10 (VFP)
	W(b)	do_vfp				@ CP#11 (VFP)
L
Linus Torvalds 已提交
630
#else
631 632
	W(mov)	pc, lr				@ CP#10 (VFP)
	W(mov)	pc, lr				@ CP#11 (VFP)
L
Linus Torvalds 已提交
633
#endif
634 635 636 637
	W(mov)	pc, lr				@ CP#12
	W(mov)	pc, lr				@ CP#13
	W(mov)	pc, lr				@ CP#14 (Debug)
	W(mov)	pc, lr				@ CP#15 (Control)
L
Linus Torvalds 已提交
638

639 640 641
#ifdef CONFIG_NEON
	.align	6

642
.LCneon_arm_opcodes:
643 644 645 646 647 648
	.word	0xfe000000			@ mask
	.word	0xf2000000			@ opcode

	.word	0xff100000			@ mask
	.word	0xf4000000			@ opcode

649 650 651 652 653 654 655 656 657 658
	.word	0x00000000			@ mask
	.word	0x00000000			@ opcode

.LCneon_thumb_opcodes:
	.word	0xef000000			@ mask
	.word	0xef000000			@ opcode

	.word	0xff100000			@ mask
	.word	0xf9000000			@ opcode

659 660 661 662
	.word	0x00000000			@ mask
	.word	0x00000000			@ opcode
#endif

L
Linus Torvalds 已提交
663
do_fpe:
664
	enable_irq
L
Linus Torvalds 已提交
665 666 667 668 669 670 671 672 673 674 675 676 677 678 679
	ldr	r4, .LCfp
	add	r10, r10, #TI_FPSTATE		@ r10 = workspace
	ldr	pc, [r4]			@ Call FP module USR entry point

/*
 * The FP module is called with these registers set:
 *  r0  = instruction
 *  r2  = PC+4
 *  r9  = normal "successful" return address
 *  r10 = FP workspace
 *  lr  = unrecognised FP instruction return address
 */

	.data
ENTRY(fp_enter)
680
	.word	no_fp
681
	.previous
L
Linus Torvalds 已提交
682

683 684 685
ENTRY(no_fp)
	mov	pc, lr
ENDPROC(no_fp)
686 687

__und_usr_unknown:
688
	enable_irq
L
Linus Torvalds 已提交
689
	mov	r0, sp
690
	adr	lr, BSYM(ret_from_exception)
L
Linus Torvalds 已提交
691
	b	do_undefinstr
692
ENDPROC(__und_usr_unknown)
L
Linus Torvalds 已提交
693 694 695

	.align	5
__pabt_usr:
R
Russell King 已提交
696
	usr_entry
L
Linus Torvalds 已提交
697

P
Paul Brook 已提交
698
	mov	r0, r2			@ pass address of aborted instruction.
699
#ifdef MULTI_PABORT
P
Paul Brook 已提交
700 701 702 703
	ldr	r4, .LCprocfns
	mov	lr, pc
	ldr	pc, [r4, #PROCESSOR_PABT_FUNC]
#else
704
	bl	CPU_PABORT_HANDLER
P
Paul Brook 已提交
705
#endif
706
	enable_irq				@ Enable interrupts
707
	mov	r2, sp				@ regs
L
Linus Torvalds 已提交
708
	bl	do_PrefetchAbort		@ call abort handler
709
 UNWIND(.fnend		)
L
Linus Torvalds 已提交
710 711 712 713 714
	/* fall through */
/*
 * This is the return code to user mode for abort handlers
 */
ENTRY(ret_from_exception)
715 716
 UNWIND(.fnstart	)
 UNWIND(.cantunwind	)
L
Linus Torvalds 已提交
717 718 719
	get_thread_info tsk
	mov	why, #0
	b	ret_to_user
720
 UNWIND(.fnend		)
721 722
ENDPROC(__pabt_usr)
ENDPROC(ret_from_exception)
L
Linus Torvalds 已提交
723 724 725 726 727 728 729

/*
 * Register switch for ARMv3 and ARMv4 processors
 * r0 = previous task_struct, r1 = previous thread_info, r2 = next thread_info
 * previous and next are guaranteed not to be the same.
 */
ENTRY(__switch_to)
730 731
 UNWIND(.fnstart	)
 UNWIND(.cantunwind	)
L
Linus Torvalds 已提交
732 733
	add	ip, r1, #TI_CPU_SAVE
	ldr	r3, [r2, #TI_TP_VALUE]
734 735 736 737
 ARM(	stmia	ip!, {r4 - sl, fp, sp, lr} )	@ Store most regs on stack
 THUMB(	stmia	ip!, {r4 - sl, fp}	   )	@ Store most regs on stack
 THUMB(	str	sp, [ip], #4		   )
 THUMB(	str	lr, [ip], #4		   )
738 739
#ifdef CONFIG_MMU
	ldr	r6, [r2, #TI_CPU_DOMAIN]
740
#endif
741
#if defined(CONFIG_HAS_TLS_REG)
742
	mcr	p15, 0, r3, c13, c0, 3		@ set TLS register
743
#elif !defined(CONFIG_TLS_REG_EMUL)
L
Linus Torvalds 已提交
744
	mov	r4, #0xffff0fff
745 746
	str	r3, [r4, #-15]			@ TLS val at 0xffff0ff0
#endif
747
#ifdef CONFIG_MMU
L
Linus Torvalds 已提交
748 749
	mcr	p15, 0, r6, c3, c0, 0		@ Set domain register
#endif
750 751 752 753 754
	mov	r5, r0
	add	r4, r2, #TI_CPU_SAVE
	ldr	r0, =thread_notify_head
	mov	r1, #THREAD_NOTIFY_SWITCH
	bl	atomic_notifier_call_chain
755
 THUMB(	mov	ip, r4			   )
756
	mov	r0, r5
757 758 759 760
 ARM(	ldmia	r4, {r4 - sl, fp, sp, pc}  )	@ Load all regs saved previously
 THUMB(	ldmia	ip!, {r4 - sl, fp}	   )	@ Load all regs saved previously
 THUMB(	ldr	sp, [ip], #4		   )
 THUMB(	ldr	pc, [ip]		   )
761
 UNWIND(.fnend		)
762
ENDPROC(__switch_to)
L
Linus Torvalds 已提交
763 764

	__INIT
765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794

/*
 * User helpers.
 *
 * These are segment of kernel provided user code reachable from user space
 * at a fixed address in kernel memory.  This is used to provide user space
 * with some operations which require kernel help because of unimplemented
 * native feature and/or instructions in many ARM CPUs. The idea is for
 * this code to be executed directly in user mode for best efficiency but
 * which is too intimate with the kernel counter part to be left to user
 * libraries.  In fact this code might even differ from one CPU to another
 * depending on the available  instruction set and restrictions like on
 * SMP systems.  In other words, the kernel reserves the right to change
 * this code as needed without warning. Only the entry points and their
 * results are guaranteed to be stable.
 *
 * Each segment is 32-byte aligned and will be moved to the top of the high
 * vector page.  New segments (if ever needed) must be added in front of
 * existing ones.  This mechanism should be used only for things that are
 * really small and justified, and not be abused freely.
 *
 * User space is expected to implement those things inline when optimizing
 * for a processor that has the necessary native support, but only if such
 * resulting binaries are already to be incompatible with earlier ARM
 * processors due to the use of unsupported instructions other than what
 * is provided here.  In other words don't make binaries unable to run on
 * earlier processors just for the sake of not using these kernel helpers
 * if your compiled code is not going to use the new instructions for other
 * purpose.
 */
795
 THUMB(	.arm	)
796

797 798 799 800 801 802 803 804
	.macro	usr_ret, reg
#ifdef CONFIG_ARM_THUMB
	bx	\reg
#else
	mov	pc, \reg
#endif
	.endm

805 806 807 808
	.align	5
	.globl	__kuser_helper_start
__kuser_helper_start:

809 810 811 812 813 814 815 816 817 818 819 820 821 822 823
/*
 * Reference prototype:
 *
 *	void __kernel_memory_barrier(void)
 *
 * Input:
 *
 *	lr = return address
 *
 * Output:
 *
 *	none
 *
 * Clobbered:
 *
824
 *	none
825 826 827 828 829 830 831 832 833 834 835 836 837
 *
 * Definition and user space usage example:
 *
 *	typedef void (__kernel_dmb_t)(void);
 *	#define __kernel_dmb (*(__kernel_dmb_t *)0xffff0fa0)
 *
 * Apply any needed memory barrier to preserve consistency with data modified
 * manually and __kuser_cmpxchg usage.
 *
 * This could be used as follows:
 *
 * #define __kernel_dmb() \
 *         asm volatile ( "mov r0, #0xffff0fff; mov lr, pc; sub pc, r0, #95" \
838
 *	        : : : "r0", "lr","cc" )
839 840 841
 */

__kuser_memory_barrier:				@ 0xffff0fa0
842
	smp_dmb
843
	usr_ret	lr
844 845 846

	.align	5

847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877
/*
 * Reference prototype:
 *
 *	int __kernel_cmpxchg(int oldval, int newval, int *ptr)
 *
 * Input:
 *
 *	r0 = oldval
 *	r1 = newval
 *	r2 = ptr
 *	lr = return address
 *
 * Output:
 *
 *	r0 = returned value (zero or non-zero)
 *	C flag = set if r0 == 0, clear if r0 != 0
 *
 * Clobbered:
 *
 *	r3, ip, flags
 *
 * Definition and user space usage example:
 *
 *	typedef int (__kernel_cmpxchg_t)(int oldval, int newval, int *ptr);
 *	#define __kernel_cmpxchg (*(__kernel_cmpxchg_t *)0xffff0fc0)
 *
 * Atomically store newval in *ptr if *ptr is equal to oldval for user space.
 * Return zero if *ptr was changed or non-zero if no exchange happened.
 * The C flag is also set if *ptr was changed to allow for assembly
 * optimization in the calling code.
 *
878 879 880 881
 * Notes:
 *
 *    - This routine already includes memory barriers as needed.
 *
882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902
 * For example, a user space atomic_add implementation could look like this:
 *
 * #define atomic_add(ptr, val) \
 *	({ register unsigned int *__ptr asm("r2") = (ptr); \
 *	   register unsigned int __result asm("r1"); \
 *	   asm volatile ( \
 *	       "1: @ atomic_add\n\t" \
 *	       "ldr	r0, [r2]\n\t" \
 *	       "mov	r3, #0xffff0fff\n\t" \
 *	       "add	lr, pc, #4\n\t" \
 *	       "add	r1, r0, %2\n\t" \
 *	       "add	pc, r3, #(0xffff0fc0 - 0xffff0fff)\n\t" \
 *	       "bcc	1b" \
 *	       : "=&r" (__result) \
 *	       : "r" (__ptr), "rIL" (val) \
 *	       : "r0","r3","ip","lr","cc","memory" ); \
 *	   __result; })
 */

__kuser_cmpxchg:				@ 0xffff0fc0

903
#if defined(CONFIG_NEEDS_SYSCALL_FOR_CMPXCHG)
904

905 906 907 908 909
	/*
	 * Poor you.  No fast solution possible...
	 * The kernel itself must perform the operation.
	 * A special ghost syscall is used for that (see traps.c).
	 */
910 911 912
	stmfd	sp!, {r7, lr}
	mov	r7, #0xff00		@ 0xfff0 into r7 for EABI
	orr	r7, r7, #0xf0
913
	swi	#0x9ffff0
914
	ldmfd	sp!, {r7, pc}
915 916

#elif __LINUX_ARM_ARCH__ < 6
917

918 919
#ifdef CONFIG_MMU

920
	/*
921 922 923 924 925 926 927
	 * The only thing that can break atomicity in this cmpxchg
	 * implementation is either an IRQ or a data abort exception
	 * causing another process/thread to be scheduled in the middle
	 * of the critical sequence.  To prevent this, code is added to
	 * the IRQ and data abort exception handlers to set the pc back
	 * to the beginning of the critical section if it is found to be
	 * within that critical section (see kuser_cmpxchg_fixup).
928
	 */
929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949
1:	ldr	r3, [r2]			@ load current val
	subs	r3, r3, r0			@ compare with oldval
2:	streq	r1, [r2]			@ store newval if eq
	rsbs	r0, r3, #0			@ set return val and C flag
	usr_ret	lr

	.text
kuser_cmpxchg_fixup:
	@ Called from kuser_cmpxchg_check macro.
	@ r2 = address of interrupted insn (must be preserved).
	@ sp = saved regs. r7 and r8 are clobbered.
	@ 1b = first critical insn, 2b = last critical insn.
	@ If r2 >= 1b and r2 <= 2b then saved pc_usr is set to 1b.
	mov	r7, #0xffff0fff
	sub	r7, r7, #(0xffff0fff - (0xffff0fc0 + (1b - __kuser_cmpxchg)))
	subs	r8, r2, r7
	rsbcss	r8, r8, #(2b - 1b)
	strcs	r7, [sp, #S_PC]
	mov	pc, lr
	.previous

950 951 952 953
#else
#warning "NPTL on non MMU needs fixing"
	mov	r0, #-1
	adds	r0, r0, #0
954
	usr_ret	lr
955
#endif
956 957 958

#else

959 960 961
#ifdef CONFIG_SMP
	mcr	p15, 0, r0, c7, c10, 5	@ dmb
#endif
962
1:	ldrex	r3, [r2]
963 964
	subs	r3, r3, r0
	strexeq	r3, r1, [r2]
965 966
	teqeq	r3, #1
	beq	1b
967
	rsbs	r0, r3, #0
968
	/* beware -- each __kuser slot must be 8 instructions max */
969
#ifdef CONFIG_SMP
970 971
	b	__kuser_memory_barrier
#else
972
	usr_ret	lr
973
#endif
974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993

#endif

	.align	5

/*
 * Reference prototype:
 *
 *	int __kernel_get_tls(void)
 *
 * Input:
 *
 *	lr = return address
 *
 * Output:
 *
 *	r0 = TLS value
 *
 * Clobbered:
 *
994
 *	none
995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013
 *
 * Definition and user space usage example:
 *
 *	typedef int (__kernel_get_tls_t)(void);
 *	#define __kernel_get_tls (*(__kernel_get_tls_t *)0xffff0fe0)
 *
 * Get the TLS value as previously set via the __ARM_NR_set_tls syscall.
 *
 * This could be used as follows:
 *
 * #define __kernel_get_tls() \
 *	({ register unsigned int __val asm("r0"); \
 *         asm( "mov r0, #0xffff0fff; mov lr, pc; sub pc, r0, #31" \
 *	        : "=r" (__val) : : "lr","cc" ); \
 *	   __val; })
 */

__kuser_get_tls:				@ 0xffff0fe0

1014
#if !defined(CONFIG_HAS_TLS_REG) && !defined(CONFIG_TLS_REG_EMUL)
1015 1016 1017 1018
	ldr	r0, [pc, #(16 - 8)]		@ TLS stored at 0xffff0ff0
#else
	mrc	p15, 0, r0, c13, c0, 3		@ read TLS register
#endif
1019
	usr_ret	lr
1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043

	.rep	5
	.word	0			@ pad up to __kuser_helper_version
	.endr

/*
 * Reference declaration:
 *
 *	extern unsigned int __kernel_helper_version;
 *
 * Definition and user space usage example:
 *
 *	#define __kernel_helper_version (*(unsigned int *)0xffff0ffc)
 *
 * User space may read this to determine the curent number of helpers
 * available.
 */

__kuser_helper_version:				@ 0xffff0ffc
	.word	((__kuser_helper_end - __kuser_helper_start) >> 5)

	.globl	__kuser_helper_end
__kuser_helper_end:

1044
 THUMB(	.thumb	)
1045

L
Linus Torvalds 已提交
1046 1047 1048
/*
 * Vector stubs.
 *
1049 1050 1051
 * This code is copied to 0xffff0200 so we can use branches in the
 * vectors, rather than ldr's.  Note that this code must not
 * exceed 0x300 bytes.
L
Linus Torvalds 已提交
1052 1053 1054
 *
 * Common stub entry macro:
 *   Enter in IRQ mode, spsr = SVC/USR CPSR, lr = SVC/USR PC
R
Russell King 已提交
1055 1056 1057
 *
 * SP points to a minimal amount of processor-private memory, the address
 * of which is copied into r0 for the mode specific abort handler.
L
Linus Torvalds 已提交
1058
 */
1059
	.macro	vector_stub, name, mode, correction=0
L
Linus Torvalds 已提交
1060 1061 1062 1063 1064 1065
	.align	5

vector_\name:
	.if \correction
	sub	lr, lr, #\correction
	.endif
R
Russell King 已提交
1066 1067 1068 1069 1070 1071

	@
	@ Save r0, lr_<exception> (parent PC) and spsr_<exception>
	@ (parent CPSR)
	@
	stmia	sp, {r0, lr}		@ save r0, lr
L
Linus Torvalds 已提交
1072
	mrs	lr, spsr
R
Russell King 已提交
1073 1074
	str	lr, [sp, #8]		@ save spsr

L
Linus Torvalds 已提交
1075
	@
R
Russell King 已提交
1076
	@ Prepare for SVC32 mode.  IRQs remain disabled.
L
Linus Torvalds 已提交
1077
	@
R
Russell King 已提交
1078
	mrs	r0, cpsr
1079
	eor	r0, r0, #(\mode ^ SVC_MODE | PSR_ISETSTATE)
R
Russell King 已提交
1080
	msr	spsr_cxsf, r0
L
Linus Torvalds 已提交
1081

R
Russell King 已提交
1082 1083 1084 1085
	@
	@ the branch table must immediately follow this code
	@
	and	lr, lr, #0x0f
1086 1087
 THUMB(	adr	r0, 1f			)
 THUMB(	ldr	lr, [r0, lr, lsl #2]	)
1088
	mov	r0, sp
1089
 ARM(	ldr	lr, [pc, lr, lsl #2]	)
R
Russell King 已提交
1090
	movs	pc, lr			@ branch to handler in SVC mode
1091
ENDPROC(vector_\name)
1092 1093 1094 1095

	.align	2
	@ handler addresses follow this label
1:
L
Linus Torvalds 已提交
1096 1097
	.endm

1098
	.globl	__stubs_start
L
Linus Torvalds 已提交
1099 1100 1101 1102
__stubs_start:
/*
 * Interrupt dispatcher
 */
1103
	vector_stub	irq, IRQ_MODE, 4
L
Linus Torvalds 已提交
1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125

	.long	__irq_usr			@  0  (USR_26 / USR_32)
	.long	__irq_invalid			@  1  (FIQ_26 / FIQ_32)
	.long	__irq_invalid			@  2  (IRQ_26 / IRQ_32)
	.long	__irq_svc			@  3  (SVC_26 / SVC_32)
	.long	__irq_invalid			@  4
	.long	__irq_invalid			@  5
	.long	__irq_invalid			@  6
	.long	__irq_invalid			@  7
	.long	__irq_invalid			@  8
	.long	__irq_invalid			@  9
	.long	__irq_invalid			@  a
	.long	__irq_invalid			@  b
	.long	__irq_invalid			@  c
	.long	__irq_invalid			@  d
	.long	__irq_invalid			@  e
	.long	__irq_invalid			@  f

/*
 * Data abort dispatcher
 * Enter in ABT mode, spsr = USR CPSR, lr = USR PC
 */
1126
	vector_stub	dabt, ABT_MODE, 8
L
Linus Torvalds 已提交
1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148

	.long	__dabt_usr			@  0  (USR_26 / USR_32)
	.long	__dabt_invalid			@  1  (FIQ_26 / FIQ_32)
	.long	__dabt_invalid			@  2  (IRQ_26 / IRQ_32)
	.long	__dabt_svc			@  3  (SVC_26 / SVC_32)
	.long	__dabt_invalid			@  4
	.long	__dabt_invalid			@  5
	.long	__dabt_invalid			@  6
	.long	__dabt_invalid			@  7
	.long	__dabt_invalid			@  8
	.long	__dabt_invalid			@  9
	.long	__dabt_invalid			@  a
	.long	__dabt_invalid			@  b
	.long	__dabt_invalid			@  c
	.long	__dabt_invalid			@  d
	.long	__dabt_invalid			@  e
	.long	__dabt_invalid			@  f

/*
 * Prefetch abort dispatcher
 * Enter in ABT mode, spsr = USR CPSR, lr = USR PC
 */
1149
	vector_stub	pabt, ABT_MODE, 4
L
Linus Torvalds 已提交
1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171

	.long	__pabt_usr			@  0 (USR_26 / USR_32)
	.long	__pabt_invalid			@  1 (FIQ_26 / FIQ_32)
	.long	__pabt_invalid			@  2 (IRQ_26 / IRQ_32)
	.long	__pabt_svc			@  3 (SVC_26 / SVC_32)
	.long	__pabt_invalid			@  4
	.long	__pabt_invalid			@  5
	.long	__pabt_invalid			@  6
	.long	__pabt_invalid			@  7
	.long	__pabt_invalid			@  8
	.long	__pabt_invalid			@  9
	.long	__pabt_invalid			@  a
	.long	__pabt_invalid			@  b
	.long	__pabt_invalid			@  c
	.long	__pabt_invalid			@  d
	.long	__pabt_invalid			@  e
	.long	__pabt_invalid			@  f

/*
 * Undef instr entry dispatcher
 * Enter in UND mode, spsr = SVC/USR CPSR, lr = SVC/USR PC
 */
1172
	vector_stub	und, UND_MODE
L
Linus Torvalds 已提交
1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225

	.long	__und_usr			@  0 (USR_26 / USR_32)
	.long	__und_invalid			@  1 (FIQ_26 / FIQ_32)
	.long	__und_invalid			@  2 (IRQ_26 / IRQ_32)
	.long	__und_svc			@  3 (SVC_26 / SVC_32)
	.long	__und_invalid			@  4
	.long	__und_invalid			@  5
	.long	__und_invalid			@  6
	.long	__und_invalid			@  7
	.long	__und_invalid			@  8
	.long	__und_invalid			@  9
	.long	__und_invalid			@  a
	.long	__und_invalid			@  b
	.long	__und_invalid			@  c
	.long	__und_invalid			@  d
	.long	__und_invalid			@  e
	.long	__und_invalid			@  f

	.align	5

/*=============================================================================
 * Undefined FIQs
 *-----------------------------------------------------------------------------
 * Enter in FIQ mode, spsr = ANY CPSR, lr = ANY PC
 * MUST PRESERVE SVC SPSR, but need to switch to SVC mode to show our msg.
 * Basically to switch modes, we *HAVE* to clobber one register...  brain
 * damage alert!  I don't think that we can execute any code in here in any
 * other mode than FIQ...  Ok you can switch to another mode, but you can't
 * get out of that mode without clobbering one register.
 */
vector_fiq:
	disable_fiq
	subs	pc, lr, #4

/*=============================================================================
 * Address exception handler
 *-----------------------------------------------------------------------------
 * These aren't too critical.
 * (they're not supposed to happen, and won't happen in 32-bit data mode).
 */

vector_addrexcptn:
	b	vector_addrexcptn

/*
 * We group all the following data together to optimise
 * for CPUs with separate I & D caches.
 */
	.align	5

.LCvswi:
	.word	vector_swi

1226
	.globl	__stubs_end
L
Linus Torvalds 已提交
1227 1228
__stubs_end:

1229
	.equ	stubs_offset, __vectors_start + 0x200 - __stubs_start
L
Linus Torvalds 已提交
1230

1231 1232
	.globl	__vectors_start
__vectors_start:
1233 1234 1235 1236 1237 1238 1239 1240 1241 1242
 ARM(	swi	SYS_ERROR0	)
 THUMB(	svc	#0		)
 THUMB(	nop			)
	W(b)	vector_und + stubs_offset
	W(ldr)	pc, .LCvswi + stubs_offset
	W(b)	vector_pabt + stubs_offset
	W(b)	vector_dabt + stubs_offset
	W(b)	vector_addrexcptn + stubs_offset
	W(b)	vector_irq + stubs_offset
	W(b)	vector_fiq + stubs_offset
1243 1244 1245

	.globl	__vectors_end
__vectors_end:
L
Linus Torvalds 已提交
1246 1247 1248 1249 1250 1251 1252 1253 1254

	.data

	.globl	cr_alignment
	.globl	cr_no_alignment
cr_alignment:
	.space	4
cr_no_alignment:
	.space	4