tlbex.c 38.8 KB
Newer Older
L
Linus Torvalds 已提交
1 2 3 4 5 6 7
/*
 * This file is subject to the terms and conditions of the GNU General Public
 * License.  See the file "COPYING" in the main directory of this archive
 * for more details.
 *
 * Synthesize TLB refill handlers at runtime.
 *
8
 * Copyright (C) 2004, 2005, 2006, 2008  Thiemo Seufer
9
 * Copyright (C) 2005, 2007, 2008, 2009  Maciej W. Rozycki
10
 * Copyright (C) 2006  Ralf Baechle (ralf@linux-mips.org)
D
David Daney 已提交
11
 * Copyright (C) 2008, 2009 Cavium Networks, Inc.
12 13 14 15 16 17 18 19 20
 *
 * ... and the days got worse and worse and now you see
 * I've gone completly out of my mind.
 *
 * They're coming to take me a away haha
 * they're coming to take me a away hoho hihi haha
 * to the funny farm where code is beautiful all the time ...
 *
 * (Condolences to Napoleon XIV)
L
Linus Torvalds 已提交
21 22
 */

23
#include <linux/bug.h>
L
Linus Torvalds 已提交
24 25 26 27 28 29 30 31
#include <linux/kernel.h>
#include <linux/types.h>
#include <linux/string.h>
#include <linux/init.h>

#include <asm/mmu_context.h>
#include <asm/war.h>

32 33
#include "uasm.h"

34
static inline int r45k_bvahwbug(void)
L
Linus Torvalds 已提交
35 36 37 38 39
{
	/* XXX: We should probe for the presence of this bug, but we don't. */
	return 0;
}

40
static inline int r4k_250MHZhwbug(void)
L
Linus Torvalds 已提交
41 42 43 44 45
{
	/* XXX: We should probe for the presence of this bug, but we don't. */
	return 0;
}

46
static inline int __maybe_unused bcm1250_m3_war(void)
L
Linus Torvalds 已提交
47 48 49 50
{
	return BCM1250_M3_WAR;
}

51
static inline int __maybe_unused r10000_llsc_war(void)
L
Linus Torvalds 已提交
52 53 54 55
{
	return R10000_LLSC_WAR;
}

56 57 58 59 60 61 62 63 64
/*
 * Found by experiment: At least some revisions of the 4kc throw under
 * some circumstances a machine check exception, triggered by invalid
 * values in the index register.  Delaying the tlbp instruction until
 * after the next branch,  plus adding an additional nop in front of
 * tlbwi/tlbwr avoids the invalid index register values. Nobody knows
 * why; it's not an issue caused by the core RTL.
 *
 */
65
static int __cpuinit m4kc_tlbp_war(void)
66 67 68 69 70
{
	return (current_cpu_data.processor_id & 0xffff00) ==
	       (PRID_COMP_MIPS | PRID_IMP_4KC);
}

71
/* Handle labels (which must be positive integers). */
L
Linus Torvalds 已提交
72
enum label_id {
73
	label_second_part = 1,
L
Linus Torvalds 已提交
74
	label_leave,
75 76 77
#ifdef MODULE_START
	label_module_alloc,
#endif
L
Linus Torvalds 已提交
78 79 80 81 82 83 84 85 86
	label_vmalloc,
	label_vmalloc_done,
	label_tlbw_hazard,
	label_split,
	label_nopage_tlbl,
	label_nopage_tlbs,
	label_nopage_tlbm,
	label_smp_pgtable_change,
	label_r3000_write_probe_fail,
D
David Daney 已提交
87 88 89
#ifdef CONFIG_HUGETLB_PAGE
	label_tlb_huge_update,
#endif
L
Linus Torvalds 已提交
90 91
};

92 93
UASM_L_LA(_second_part)
UASM_L_LA(_leave)
94
#ifdef MODULE_START
95
UASM_L_LA(_module_alloc)
96
#endif
97 98 99 100 101 102 103 104 105
UASM_L_LA(_vmalloc)
UASM_L_LA(_vmalloc_done)
UASM_L_LA(_tlbw_hazard)
UASM_L_LA(_split)
UASM_L_LA(_nopage_tlbl)
UASM_L_LA(_nopage_tlbs)
UASM_L_LA(_nopage_tlbm)
UASM_L_LA(_smp_pgtable_change)
UASM_L_LA(_r3000_write_probe_fail)
D
David Daney 已提交
106 107 108
#ifdef CONFIG_HUGETLB_PAGE
UASM_L_LA(_tlb_huge_update)
#endif
109

110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125
/*
 * For debug purposes.
 */
static inline void dump_handler(const u32 *handler, int count)
{
	int i;

	pr_debug("\t.set push\n");
	pr_debug("\t.set noreorder\n");

	for (i = 0; i < count; i++)
		pr_debug("\t%p\t.word 0x%08x\n", &handler[i], handler[i]);

	pr_debug("\t.set pop\n");
}

L
Linus Torvalds 已提交
126 127 128 129 130
/* The only general purpose registers allowed in TLB handlers. */
#define K0		26
#define K1		27

/* Some CP0 registers */
131 132 133 134 135
#define C0_INDEX	0, 0
#define C0_ENTRYLO0	2, 0
#define C0_TCBIND	2, 2
#define C0_ENTRYLO1	3, 0
#define C0_CONTEXT	4, 0
D
David Daney 已提交
136
#define C0_PAGEMASK	5, 0
137 138 139 140
#define C0_BADVADDR	8, 0
#define C0_ENTRYHI	10, 0
#define C0_EPC		14, 0
#define C0_XCONTEXT	20, 0
L
Linus Torvalds 已提交
141

142
#ifdef CONFIG_64BIT
143
# define GET_CONTEXT(buf, reg) UASM_i_MFC0(buf, reg, C0_XCONTEXT)
L
Linus Torvalds 已提交
144
#else
145
# define GET_CONTEXT(buf, reg) UASM_i_MFC0(buf, reg, C0_CONTEXT)
L
Linus Torvalds 已提交
146 147 148 149 150 151 152 153 154 155
#endif

/* The worst case length of the handler is around 18 instructions for
 * R3000-style TLBs and up to 63 instructions for R4000-style TLBs.
 * Maximum space available is 32 instructions for R3000 and 64
 * instructions for R4000.
 *
 * We deliberately chose a buffer size of 128, so we won't scribble
 * over anything important on overflow before we panic.
 */
156
static u32 tlb_handler[128] __cpuinitdata;
L
Linus Torvalds 已提交
157 158

/* simply assume worst case size for labels and relocs */
159 160
static struct uasm_label labels[128] __cpuinitdata;
static struct uasm_reloc relocs[128] __cpuinitdata;
L
Linus Torvalds 已提交
161 162 163 164

/*
 * The R3000 TLB handler is simple.
 */
165
static void __cpuinit build_r3000_tlb_refill_handler(void)
L
Linus Torvalds 已提交
166 167 168 169 170 171 172
{
	long pgdc = (long)pgd_current;
	u32 *p;

	memset(tlb_handler, 0, sizeof(tlb_handler));
	p = tlb_handler;

173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189
	uasm_i_mfc0(&p, K0, C0_BADVADDR);
	uasm_i_lui(&p, K1, uasm_rel_hi(pgdc)); /* cp0 delay */
	uasm_i_lw(&p, K1, uasm_rel_lo(pgdc), K1);
	uasm_i_srl(&p, K0, K0, 22); /* load delay */
	uasm_i_sll(&p, K0, K0, 2);
	uasm_i_addu(&p, K1, K1, K0);
	uasm_i_mfc0(&p, K0, C0_CONTEXT);
	uasm_i_lw(&p, K1, 0, K1); /* cp0 delay */
	uasm_i_andi(&p, K0, K0, 0xffc); /* load delay */
	uasm_i_addu(&p, K1, K1, K0);
	uasm_i_lw(&p, K0, 0, K1);
	uasm_i_nop(&p); /* load delay */
	uasm_i_mtc0(&p, K0, C0_ENTRYLO0);
	uasm_i_mfc0(&p, K1, C0_EPC); /* cp0 delay */
	uasm_i_tlbwr(&p); /* cp0 delay */
	uasm_i_jr(&p, K1);
	uasm_i_rfe(&p); /* branch delay */
L
Linus Torvalds 已提交
190 191 192 193

	if (p > tlb_handler + 32)
		panic("TLB refill handler space exceeded");

194 195
	pr_debug("Wrote TLB refill handler (%u instructions).\n",
		 (unsigned int)(p - tlb_handler));
L
Linus Torvalds 已提交
196

197
	memcpy((void *)ebase, tlb_handler, 0x80);
198 199

	dump_handler((u32 *)ebase, 32);
L
Linus Torvalds 已提交
200 201 202 203 204 205 206 207 208
}

/*
 * The R4000 TLB handler is much more complicated. We have two
 * consecutive handler areas with 32 instructions space each.
 * Since they aren't used at the same time, we can overflow in the
 * other one.To keep things simple, we first assume linear space,
 * then we relocate it to the final handler layout as needed.
 */
209
static u32 final_handler[64] __cpuinitdata;
L
Linus Torvalds 已提交
210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232

/*
 * Hazards
 *
 * From the IDT errata for the QED RM5230 (Nevada), processor revision 1.0:
 * 2. A timing hazard exists for the TLBP instruction.
 *
 *      stalling_instruction
 *      TLBP
 *
 * The JTLB is being read for the TLBP throughout the stall generated by the
 * previous instruction. This is not really correct as the stalling instruction
 * can modify the address used to access the JTLB.  The failure symptom is that
 * the TLBP instruction will use an address created for the stalling instruction
 * and not the address held in C0_ENHI and thus report the wrong results.
 *
 * The software work-around is to not allow the instruction preceding the TLBP
 * to stall - make it an NOP or some other instruction guaranteed not to stall.
 *
 * Errata 2 will not be fixed.  This errata is also on the R5000.
 *
 * As if we MIPS hackers wouldn't know how to nop pipelines happy ...
 */
233
static void __cpuinit __maybe_unused build_tlb_probe_entry(u32 **p)
L
Linus Torvalds 已提交
234
{
235
	switch (current_cpu_type()) {
236
	/* Found by experiment: R4600 v2.0/R4700 needs this, too.  */
237
	case CPU_R4600:
238
	case CPU_R4700:
L
Linus Torvalds 已提交
239 240 241
	case CPU_R5000:
	case CPU_R5000A:
	case CPU_NEVADA:
242 243
		uasm_i_nop(p);
		uasm_i_tlbp(p);
L
Linus Torvalds 已提交
244 245 246
		break;

	default:
247
		uasm_i_tlbp(p);
L
Linus Torvalds 已提交
248 249 250 251 252 253 254 255 256 257
		break;
	}
}

/*
 * Write random or indexed TLB entry, and care about the hazards from
 * the preceeding mtc0 and for the following eret.
 */
enum tlb_write_entry { tlb_random, tlb_indexed };

258
static void __cpuinit build_tlb_write_entry(u32 **p, struct uasm_label **l,
259
					 struct uasm_reloc **r,
L
Linus Torvalds 已提交
260 261 262 263 264
					 enum tlb_write_entry wmode)
{
	void(*tlbw)(u32 **) = NULL;

	switch (wmode) {
265 266
	case tlb_random: tlbw = uasm_i_tlbwr; break;
	case tlb_indexed: tlbw = uasm_i_tlbwi; break;
L
Linus Torvalds 已提交
267 268
	}

269
	if (cpu_has_mips_r2) {
270 271
		if (cpu_has_mips_r2_exec_hazard)
			uasm_i_ehb(p);
272 273 274 275
		tlbw(p);
		return;
	}

276
	switch (current_cpu_type()) {
L
Linus Torvalds 已提交
277 278 279 280 281 282 283 284 285 286
	case CPU_R4000PC:
	case CPU_R4000SC:
	case CPU_R4000MC:
	case CPU_R4400PC:
	case CPU_R4400SC:
	case CPU_R4400MC:
		/*
		 * This branch uses up a mtc0 hazard nop slot and saves
		 * two nops after the tlbw instruction.
		 */
287
		uasm_il_bgezl(p, r, 0, label_tlbw_hazard);
L
Linus Torvalds 已提交
288
		tlbw(p);
289 290
		uasm_l_tlbw_hazard(l, *p);
		uasm_i_nop(p);
L
Linus Torvalds 已提交
291 292 293 294 295 296
		break;

	case CPU_R4600:
	case CPU_R4700:
	case CPU_R5000:
	case CPU_R5000A:
297
		uasm_i_nop(p);
298
		tlbw(p);
299
		uasm_i_nop(p);
300 301 302
		break;

	case CPU_R4300:
L
Linus Torvalds 已提交
303 304
	case CPU_5KC:
	case CPU_TX49XX:
305
	case CPU_PR4450:
306
		uasm_i_nop(p);
L
Linus Torvalds 已提交
307 308 309 310 311
		tlbw(p);
		break;

	case CPU_R10000:
	case CPU_R12000:
K
Kumba 已提交
312
	case CPU_R14000:
L
Linus Torvalds 已提交
313
	case CPU_4KC:
314
	case CPU_4KEC:
L
Linus Torvalds 已提交
315
	case CPU_SB1:
A
Andrew Isaacson 已提交
316
	case CPU_SB1A:
L
Linus Torvalds 已提交
317 318 319
	case CPU_4KSC:
	case CPU_20KC:
	case CPU_25KF:
320 321
	case CPU_BCM3302:
	case CPU_BCM4710:
322
	case CPU_LOONGSON2:
323
	case CPU_R5500:
324
		if (m4kc_tlbp_war())
325
			uasm_i_nop(p);
326
	case CPU_ALCHEMY:
L
Linus Torvalds 已提交
327 328 329 330
		tlbw(p);
		break;

	case CPU_NEVADA:
331
		uasm_i_nop(p); /* QED specifies 2 nops hazard */
L
Linus Torvalds 已提交
332 333 334 335
		/*
		 * This branch uses up a mtc0 hazard nop slot and saves
		 * a nop after the tlbw instruction.
		 */
336
		uasm_il_bgezl(p, r, 0, label_tlbw_hazard);
L
Linus Torvalds 已提交
337
		tlbw(p);
338
		uasm_l_tlbw_hazard(l, *p);
L
Linus Torvalds 已提交
339 340 341
		break;

	case CPU_RM7000:
342 343 344 345
		uasm_i_nop(p);
		uasm_i_nop(p);
		uasm_i_nop(p);
		uasm_i_nop(p);
L
Linus Torvalds 已提交
346 347 348 349 350 351 352 353 354 355
		tlbw(p);
		break;

	case CPU_RM9000:
		/*
		 * When the JTLB is updated by tlbwi or tlbwr, a subsequent
		 * use of the JTLB for instructions should not occur for 4
		 * cpu cycles and use for data translations should not occur
		 * for 3 cpu cycles.
		 */
356 357 358 359
		uasm_i_ssnop(p);
		uasm_i_ssnop(p);
		uasm_i_ssnop(p);
		uasm_i_ssnop(p);
L
Linus Torvalds 已提交
360
		tlbw(p);
361 362 363 364
		uasm_i_ssnop(p);
		uasm_i_ssnop(p);
		uasm_i_ssnop(p);
		uasm_i_ssnop(p);
L
Linus Torvalds 已提交
365 366 367 368 369 370 371
		break;

	case CPU_VR4111:
	case CPU_VR4121:
	case CPU_VR4122:
	case CPU_VR4181:
	case CPU_VR4181A:
372 373
		uasm_i_nop(p);
		uasm_i_nop(p);
L
Linus Torvalds 已提交
374
		tlbw(p);
375 376
		uasm_i_nop(p);
		uasm_i_nop(p);
L
Linus Torvalds 已提交
377 378 379 380
		break;

	case CPU_VR4131:
	case CPU_VR4133:
381
	case CPU_R5432:
382 383
		uasm_i_nop(p);
		uasm_i_nop(p);
L
Linus Torvalds 已提交
384 385 386 387 388 389 390 391 392 393
		tlbw(p);
		break;

	default:
		panic("No TLB refill handler yet (CPU type: %d)",
		      current_cpu_data.cputype);
		break;
	}
}

D
David Daney 已提交
394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485
#ifdef CONFIG_HUGETLB_PAGE
static __cpuinit void build_huge_tlb_write_entry(u32 **p,
						 struct uasm_label **l,
						 struct uasm_reloc **r,
						 unsigned int tmp,
						 enum tlb_write_entry wmode)
{
	/* Set huge page tlb entry size */
	uasm_i_lui(p, tmp, PM_HUGE_MASK >> 16);
	uasm_i_ori(p, tmp, tmp, PM_HUGE_MASK & 0xffff);
	uasm_i_mtc0(p, tmp, C0_PAGEMASK);

	build_tlb_write_entry(p, l, r, wmode);

	/* Reset default page size */
	if (PM_DEFAULT_MASK >> 16) {
		uasm_i_lui(p, tmp, PM_DEFAULT_MASK >> 16);
		uasm_i_ori(p, tmp, tmp, PM_DEFAULT_MASK & 0xffff);
		uasm_il_b(p, r, label_leave);
		uasm_i_mtc0(p, tmp, C0_PAGEMASK);
	} else if (PM_DEFAULT_MASK) {
		uasm_i_ori(p, tmp, 0, PM_DEFAULT_MASK);
		uasm_il_b(p, r, label_leave);
		uasm_i_mtc0(p, tmp, C0_PAGEMASK);
	} else {
		uasm_il_b(p, r, label_leave);
		uasm_i_mtc0(p, 0, C0_PAGEMASK);
	}
}

/*
 * Check if Huge PTE is present, if so then jump to LABEL.
 */
static void __cpuinit
build_is_huge_pte(u32 **p, struct uasm_reloc **r, unsigned int tmp,
		unsigned int pmd, int lid)
{
	UASM_i_LW(p, tmp, 0, pmd);
	uasm_i_andi(p, tmp, tmp, _PAGE_HUGE);
	uasm_il_bnez(p, r, tmp, lid);
}

static __cpuinit void build_huge_update_entries(u32 **p,
						unsigned int pte,
						unsigned int tmp)
{
	int small_sequence;

	/*
	 * A huge PTE describes an area the size of the
	 * configured huge page size. This is twice the
	 * of the large TLB entry size we intend to use.
	 * A TLB entry half the size of the configured
	 * huge page size is configured into entrylo0
	 * and entrylo1 to cover the contiguous huge PTE
	 * address space.
	 */
	small_sequence = (HPAGE_SIZE >> 7) < 0x10000;

	/* We can clobber tmp.  It isn't used after this.*/
	if (!small_sequence)
		uasm_i_lui(p, tmp, HPAGE_SIZE >> (7 + 16));

	UASM_i_SRL(p, pte, pte, 6); /* convert to entrylo */
	uasm_i_mtc0(p, pte, C0_ENTRYLO0); /* load it */
	/* convert to entrylo1 */
	if (small_sequence)
		UASM_i_ADDIU(p, pte, pte, HPAGE_SIZE >> 7);
	else
		UASM_i_ADDU(p, pte, pte, tmp);

	uasm_i_mtc0(p, pte, C0_ENTRYLO1); /* load it */
}

static __cpuinit void build_huge_handler_tail(u32 **p,
					      struct uasm_reloc **r,
					      struct uasm_label **l,
					      unsigned int pte,
					      unsigned int ptr)
{
#ifdef CONFIG_SMP
	UASM_i_SC(p, pte, 0, ptr);
	uasm_il_beqz(p, r, pte, label_tlb_huge_update);
	UASM_i_LW(p, pte, 0, ptr); /* Needed because SC killed our PTE */
#else
	UASM_i_SW(p, pte, 0, ptr);
#endif
	build_huge_update_entries(p, pte, ptr);
	build_huge_tlb_write_entry(p, l, r, pte, tlb_indexed);
}
#endif /* CONFIG_HUGETLB_PAGE */

486
#ifdef CONFIG_64BIT
L
Linus Torvalds 已提交
487 488 489 490
/*
 * TMP and PTR are scratch.
 * TMP will be clobbered, PTR will hold the pmd entry.
 */
491
static void __cpuinit
492
build_get_pmde64(u32 **p, struct uasm_label **l, struct uasm_reloc **r,
L
Linus Torvalds 已提交
493 494 495 496 497 498 499
		 unsigned int tmp, unsigned int ptr)
{
	long pgdc = (long)pgd_current;

	/*
	 * The vmalloc handling is not in the hotpath.
	 */
500
	uasm_i_dmfc0(p, tmp, C0_BADVADDR);
501
#ifdef MODULE_START
502
	uasm_il_bltz(p, r, tmp, label_module_alloc);
503
#else
504
	uasm_il_bltz(p, r, tmp, label_vmalloc);
505
#endif
506
	/* No uasm_i_nop needed here, since the next insn doesn't touch TMP. */
L
Linus Torvalds 已提交
507 508

#ifdef CONFIG_SMP
509 510 511 512
# ifdef  CONFIG_MIPS_MT_SMTC
	/*
	 * SMTC uses TCBind value as "CPU" index
	 */
513 514
	uasm_i_mfc0(p, ptr, C0_TCBIND);
	uasm_i_dsrl(p, ptr, ptr, 19);
515
# else
L
Linus Torvalds 已提交
516
	/*
517
	 * 64 bit SMP running in XKPHYS has smp_processor_id() << 3
L
Linus Torvalds 已提交
518 519
	 * stored in CONTEXT.
	 */
520 521
	uasm_i_dmfc0(p, ptr, C0_CONTEXT);
	uasm_i_dsrl(p, ptr, ptr, 23);
522
#endif
523 524 525 526
	UASM_i_LA_mostly(p, tmp, pgdc);
	uasm_i_daddu(p, ptr, ptr, tmp);
	uasm_i_dmfc0(p, tmp, C0_BADVADDR);
	uasm_i_ld(p, ptr, uasm_rel_lo(pgdc), ptr);
L
Linus Torvalds 已提交
527
#else
528 529
	UASM_i_LA_mostly(p, ptr, pgdc);
	uasm_i_ld(p, ptr, uasm_rel_lo(pgdc), ptr);
L
Linus Torvalds 已提交
530 531
#endif

532
	uasm_l_vmalloc_done(l, *p);
R
Ralf Baechle 已提交
533 534

	if (PGDIR_SHIFT - 3 < 32)		/* get pgd offset in bytes */
535
		uasm_i_dsrl(p, tmp, tmp, PGDIR_SHIFT-3);
R
Ralf Baechle 已提交
536
	else
537 538 539 540 541 542 543 544 545
		uasm_i_dsrl32(p, tmp, tmp, PGDIR_SHIFT - 3 - 32);

	uasm_i_andi(p, tmp, tmp, (PTRS_PER_PGD - 1)<<3);
	uasm_i_daddu(p, ptr, ptr, tmp); /* add in pgd offset */
	uasm_i_dmfc0(p, tmp, C0_BADVADDR); /* get faulting address */
	uasm_i_ld(p, ptr, 0, ptr); /* get pmd pointer */
	uasm_i_dsrl(p, tmp, tmp, PMD_SHIFT-3); /* get pmd offset in bytes */
	uasm_i_andi(p, tmp, tmp, (PTRS_PER_PMD - 1)<<3);
	uasm_i_daddu(p, ptr, ptr, tmp); /* add in pmd offset */
L
Linus Torvalds 已提交
546 547 548 549 550 551
}

/*
 * BVADDR is the faulting address, PTR is scratch.
 * PTR will hold the pgd for vmalloc.
 */
552
static void __cpuinit
553
build_get_pgd_vmalloc64(u32 **p, struct uasm_label **l, struct uasm_reloc **r,
L
Linus Torvalds 已提交
554 555 556 557
			unsigned int bvaddr, unsigned int ptr)
{
	long swpd = (long)swapper_pg_dir;

558 559 560
#ifdef MODULE_START
	long modd = (long)module_pg_dir;

561
	uasm_l_module_alloc(l, *p);
562 563 564 565 566
	/*
	 * Assumption:
	 * VMALLOC_START >= 0xc000000000000000UL
	 * MODULE_START >= 0xe000000000000000UL
	 */
567 568
	UASM_i_SLL(p, ptr, bvaddr, 2);
	uasm_il_bgez(p, r, ptr, label_vmalloc);
569

570 571 572
	if (uasm_in_compat_space_p(MODULE_START) &&
	    !uasm_rel_lo(MODULE_START)) {
		uasm_i_lui(p, ptr, uasm_rel_hi(MODULE_START)); /* delay slot */
573 574
	} else {
		/* unlikely configuration */
575 576
		uasm_i_nop(p); /* delay slot */
		UASM_i_LA(p, ptr, MODULE_START);
577
	}
578
	uasm_i_dsubu(p, bvaddr, bvaddr, ptr);
579

580 581 582
	if (uasm_in_compat_space_p(modd) && !uasm_rel_lo(modd)) {
		uasm_il_b(p, r, label_vmalloc_done);
		uasm_i_lui(p, ptr, uasm_rel_hi(modd));
583
	} else {
584 585 586 587
		UASM_i_LA_mostly(p, ptr, modd);
		uasm_il_b(p, r, label_vmalloc_done);
		if (uasm_in_compat_space_p(modd))
			uasm_i_addiu(p, ptr, ptr, uasm_rel_lo(modd));
588
		else
589
			uasm_i_daddiu(p, ptr, ptr, uasm_rel_lo(modd));
590 591
	}

592 593 594
	uasm_l_vmalloc(l, *p);
	if (uasm_in_compat_space_p(MODULE_START) &&
	    !uasm_rel_lo(MODULE_START) &&
595
	    MODULE_START << 32 == VMALLOC_START)
596
		uasm_i_dsll32(p, ptr, ptr, 0);	/* typical case */
597
	else
598
		UASM_i_LA(p, ptr, VMALLOC_START);
599
#else
600 601
	uasm_l_vmalloc(l, *p);
	UASM_i_LA(p, ptr, VMALLOC_START);
602
#endif
603
	uasm_i_dsubu(p, bvaddr, bvaddr, ptr);
L
Linus Torvalds 已提交
604

605 606 607
	if (uasm_in_compat_space_p(swpd) && !uasm_rel_lo(swpd)) {
		uasm_il_b(p, r, label_vmalloc_done);
		uasm_i_lui(p, ptr, uasm_rel_hi(swpd));
L
Linus Torvalds 已提交
608
	} else {
609 610 611 612
		UASM_i_LA_mostly(p, ptr, swpd);
		uasm_il_b(p, r, label_vmalloc_done);
		if (uasm_in_compat_space_p(swpd))
			uasm_i_addiu(p, ptr, ptr, uasm_rel_lo(swpd));
613
		else
614
			uasm_i_daddiu(p, ptr, ptr, uasm_rel_lo(swpd));
L
Linus Torvalds 已提交
615 616 617
	}
}

618
#else /* !CONFIG_64BIT */
L
Linus Torvalds 已提交
619 620 621 622 623

/*
 * TMP and PTR are scratch.
 * TMP will be clobbered, PTR will hold the pgd entry.
 */
624
static void __cpuinit __maybe_unused
L
Linus Torvalds 已提交
625 626 627 628 629 630
build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr)
{
	long pgdc = (long)pgd_current;

	/* 32 bit SMP has smp_processor_id() stored in CONTEXT. */
#ifdef CONFIG_SMP
631 632 633 634
#ifdef  CONFIG_MIPS_MT_SMTC
	/*
	 * SMTC uses TCBind value as "CPU" index
	 */
635 636 637
	uasm_i_mfc0(p, ptr, C0_TCBIND);
	UASM_i_LA_mostly(p, tmp, pgdc);
	uasm_i_srl(p, ptr, ptr, 19);
638 639 640 641
#else
	/*
	 * smp_processor_id() << 3 is stored in CONTEXT.
         */
642 643 644
	uasm_i_mfc0(p, ptr, C0_CONTEXT);
	UASM_i_LA_mostly(p, tmp, pgdc);
	uasm_i_srl(p, ptr, ptr, 23);
645
#endif
646
	uasm_i_addu(p, ptr, tmp, ptr);
L
Linus Torvalds 已提交
647
#else
648
	UASM_i_LA_mostly(p, ptr, pgdc);
L
Linus Torvalds 已提交
649
#endif
650 651 652 653 654
	uasm_i_mfc0(p, tmp, C0_BADVADDR); /* get faulting address */
	uasm_i_lw(p, ptr, uasm_rel_lo(pgdc), ptr);
	uasm_i_srl(p, tmp, tmp, PGDIR_SHIFT); /* get pgd only bits */
	uasm_i_sll(p, tmp, tmp, PGD_T_LOG2);
	uasm_i_addu(p, ptr, ptr, tmp); /* add in pgd offset */
L
Linus Torvalds 已提交
655 656
}

657
#endif /* !CONFIG_64BIT */
L
Linus Torvalds 已提交
658

659
static void __cpuinit build_adjust_context(u32 **p, unsigned int ctx)
L
Linus Torvalds 已提交
660
{
R
Ralf Baechle 已提交
661
	unsigned int shift = 4 - (PTE_T_LOG2 + 1) + PAGE_SHIFT - 12;
L
Linus Torvalds 已提交
662 663
	unsigned int mask = (PTRS_PER_PTE / 2 - 1) << (PTE_T_LOG2 + 1);

664
	switch (current_cpu_type()) {
L
Linus Torvalds 已提交
665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680
	case CPU_VR41XX:
	case CPU_VR4111:
	case CPU_VR4121:
	case CPU_VR4122:
	case CPU_VR4131:
	case CPU_VR4181:
	case CPU_VR4181A:
	case CPU_VR4133:
		shift += 2;
		break;

	default:
		break;
	}

	if (shift)
681 682
		UASM_i_SRL(p, ctx, ctx, shift);
	uasm_i_andi(p, ctx, ctx, mask);
L
Linus Torvalds 已提交
683 684
}

685
static void __cpuinit build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr)
L
Linus Torvalds 已提交
686 687 688 689 690 691 692 693
{
	/*
	 * Bug workaround for the Nevada. It seems as if under certain
	 * circumstances the move from cp0_context might produce a
	 * bogus result when the mfc0 instruction and its consumer are
	 * in a different cacheline or a load instruction, probably any
	 * memory reference, is between them.
	 */
694
	switch (current_cpu_type()) {
L
Linus Torvalds 已提交
695
	case CPU_NEVADA:
696
		UASM_i_LW(p, ptr, 0, ptr);
L
Linus Torvalds 已提交
697 698 699 700 701
		GET_CONTEXT(p, tmp); /* get context reg */
		break;

	default:
		GET_CONTEXT(p, tmp); /* get context reg */
702
		UASM_i_LW(p, ptr, 0, ptr);
L
Linus Torvalds 已提交
703 704 705 706
		break;
	}

	build_adjust_context(p, tmp);
707
	UASM_i_ADDU(p, ptr, ptr, tmp); /* add in offset */
L
Linus Torvalds 已提交
708 709
}

710
static void __cpuinit build_update_entries(u32 **p, unsigned int tmp,
L
Linus Torvalds 已提交
711 712 713 714 715 716 717 718
					unsigned int ptep)
{
	/*
	 * 64bit address support (36bit on a 32bit CPU) in a 32bit
	 * Kernel is a special case. Only a few CPUs use it.
	 */
#ifdef CONFIG_64BIT_PHYS_ADDR
	if (cpu_has_64bits) {
719 720 721 722 723 724
		uasm_i_ld(p, tmp, 0, ptep); /* get even pte */
		uasm_i_ld(p, ptep, sizeof(pte_t), ptep); /* get odd pte */
		uasm_i_dsrl(p, tmp, tmp, 6); /* convert to entrylo0 */
		uasm_i_mtc0(p, tmp, C0_ENTRYLO0); /* load it */
		uasm_i_dsrl(p, ptep, ptep, 6); /* convert to entrylo1 */
		uasm_i_mtc0(p, ptep, C0_ENTRYLO1); /* load it */
L
Linus Torvalds 已提交
725 726 727 728 729
	} else {
		int pte_off_even = sizeof(pte_t) / 2;
		int pte_off_odd = pte_off_even + sizeof(pte_t);

		/* The pte entries are pre-shifted */
730 731 732 733
		uasm_i_lw(p, tmp, pte_off_even, ptep); /* get even pte */
		uasm_i_mtc0(p, tmp, C0_ENTRYLO0); /* load it */
		uasm_i_lw(p, ptep, pte_off_odd, ptep); /* get odd pte */
		uasm_i_mtc0(p, ptep, C0_ENTRYLO1); /* load it */
L
Linus Torvalds 已提交
734 735
	}
#else
736 737
	UASM_i_LW(p, tmp, 0, ptep); /* get even pte */
	UASM_i_LW(p, ptep, sizeof(pte_t), ptep); /* get odd pte */
L
Linus Torvalds 已提交
738 739
	if (r45k_bvahwbug())
		build_tlb_probe_entry(p);
740
	UASM_i_SRL(p, tmp, tmp, 6); /* convert to entrylo0 */
L
Linus Torvalds 已提交
741
	if (r4k_250MHZhwbug())
742 743 744
		uasm_i_mtc0(p, 0, C0_ENTRYLO0);
	uasm_i_mtc0(p, tmp, C0_ENTRYLO0); /* load it */
	UASM_i_SRL(p, ptep, ptep, 6); /* convert to entrylo1 */
L
Linus Torvalds 已提交
745
	if (r45k_bvahwbug())
746
		uasm_i_mfc0(p, tmp, C0_INDEX);
L
Linus Torvalds 已提交
747
	if (r4k_250MHZhwbug())
748 749
		uasm_i_mtc0(p, 0, C0_ENTRYLO1);
	uasm_i_mtc0(p, ptep, C0_ENTRYLO1); /* load it */
L
Linus Torvalds 已提交
750 751 752
#endif
}

753 754 755 756 757 758 759 760
/*
 * For a 64-bit kernel, we are using the 64-bit XTLB refill exception
 * because EXL == 0.  If we wrap, we can also use the 32 instruction
 * slots before the XTLB refill exception handler which belong to the
 * unused TLB refill exception.
 */
#define MIPS64_REFILL_INSNS 32

761
static void __cpuinit build_r4000_tlb_refill_handler(void)
L
Linus Torvalds 已提交
762 763
{
	u32 *p = tlb_handler;
764 765
	struct uasm_label *l = labels;
	struct uasm_reloc *r = relocs;
L
Linus Torvalds 已提交
766 767 768 769 770 771 772 773 774 775 776 777
	u32 *f;
	unsigned int final_len;

	memset(tlb_handler, 0, sizeof(tlb_handler));
	memset(labels, 0, sizeof(labels));
	memset(relocs, 0, sizeof(relocs));
	memset(final_handler, 0, sizeof(final_handler));

	/*
	 * create the plain linear handler
	 */
	if (bcm1250_m3_war()) {
778 779 780 781 782 783
		UASM_i_MFC0(&p, K0, C0_BADVADDR);
		UASM_i_MFC0(&p, K1, C0_ENTRYHI);
		uasm_i_xor(&p, K0, K0, K1);
		UASM_i_SRL(&p, K0, K0, PAGE_SHIFT + 1);
		uasm_il_bnez(&p, &r, K0, label_leave);
		/* No need for uasm_i_nop */
L
Linus Torvalds 已提交
784 785
	}

786
#ifdef CONFIG_64BIT
L
Linus Torvalds 已提交
787 788 789 790 791
	build_get_pmde64(&p, &l, &r, K0, K1); /* get pmd in K1 */
#else
	build_get_pgde32(&p, K0, K1); /* get pgd in K1 */
#endif

D
David Daney 已提交
792 793 794 795
#ifdef CONFIG_HUGETLB_PAGE
	build_is_huge_pte(&p, &r, K0, K1, label_tlb_huge_update);
#endif

L
Linus Torvalds 已提交
796 797 798
	build_get_ptep(&p, K0, K1);
	build_update_entries(&p, K0, K1);
	build_tlb_write_entry(&p, &l, &r, tlb_random);
799 800
	uasm_l_leave(&l, p);
	uasm_i_eret(&p); /* return from trap */
L
Linus Torvalds 已提交
801

D
David Daney 已提交
802 803 804 805 806 807 808
#ifdef CONFIG_HUGETLB_PAGE
	uasm_l_tlb_huge_update(&l, p);
	UASM_i_LW(&p, K0, 0, K1);
	build_huge_update_entries(&p, K0, K1);
	build_huge_tlb_write_entry(&p, &l, &r, K0, tlb_random);
#endif

809
#ifdef CONFIG_64BIT
L
Linus Torvalds 已提交
810 811 812 813 814 815 816
	build_get_pgd_vmalloc64(&p, &l, &r, K0, K1);
#endif

	/*
	 * Overflow check: For the 64bit handler, we need at least one
	 * free instruction slot for the wrap-around branch. In worst
	 * case, if the intended insertion point is a delay slot, we
M
Matt LaPlante 已提交
817
	 * need three, with the second nop'ed and the third being
L
Linus Torvalds 已提交
818 819
	 * unused.
	 */
820 821
	/* Loongson2 ebase is different than r4k, we have more space */
#if defined(CONFIG_32BIT) || defined(CONFIG_CPU_LOONGSON2)
L
Linus Torvalds 已提交
822 823 824
	if ((p - tlb_handler) > 64)
		panic("TLB refill handler space exceeded");
#else
825 826 827 828
	if (((p - tlb_handler) > (MIPS64_REFILL_INSNS * 2) - 1)
	    || (((p - tlb_handler) > (MIPS64_REFILL_INSNS * 2) - 3)
		&& uasm_insn_has_bdelay(relocs,
					tlb_handler + MIPS64_REFILL_INSNS - 3)))
L
Linus Torvalds 已提交
829 830 831 832 833 834
		panic("TLB refill handler space exceeded");
#endif

	/*
	 * Now fold the handler in the TLB refill handler space.
	 */
835
#if defined(CONFIG_32BIT) || defined(CONFIG_CPU_LOONGSON2)
L
Linus Torvalds 已提交
836 837
	f = final_handler;
	/* Simplest case, just copy the handler. */
838
	uasm_copy_handler(relocs, labels, tlb_handler, p, f);
L
Linus Torvalds 已提交
839
	final_len = p - tlb_handler;
840
#else /* CONFIG_64BIT */
841 842
	f = final_handler + MIPS64_REFILL_INSNS;
	if ((p - tlb_handler) <= MIPS64_REFILL_INSNS) {
L
Linus Torvalds 已提交
843
		/* Just copy the handler. */
844
		uasm_copy_handler(relocs, labels, tlb_handler, p, f);
L
Linus Torvalds 已提交
845 846
		final_len = p - tlb_handler;
	} else {
D
David Daney 已提交
847 848 849
#if defined(CONFIG_HUGETLB_PAGE)
		const enum label_id ls = label_tlb_huge_update;
#elif defined(MODULE_START)
850 851 852 853 854 855 856 857 858 859 860 861
		const enum label_id ls = label_module_alloc;
#else
		const enum label_id ls = label_vmalloc;
#endif
		u32 *split;
		int ov = 0;
		int i;

		for (i = 0; i < ARRAY_SIZE(labels) && labels[i].lab != ls; i++)
			;
		BUG_ON(i == ARRAY_SIZE(labels));
		split = labels[i].addr;
L
Linus Torvalds 已提交
862 863

		/*
864
		 * See if we have overflown one way or the other.
L
Linus Torvalds 已提交
865
		 */
866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885
		if (split > tlb_handler + MIPS64_REFILL_INSNS ||
		    split < p - MIPS64_REFILL_INSNS)
			ov = 1;

		if (ov) {
			/*
			 * Split two instructions before the end.  One
			 * for the branch and one for the instruction
			 * in the delay slot.
			 */
			split = tlb_handler + MIPS64_REFILL_INSNS - 2;

			/*
			 * If the branch would fall in a delay slot,
			 * we must back up an additional instruction
			 * so that it is no longer in a delay slot.
			 */
			if (uasm_insn_has_bdelay(relocs, split - 1))
				split--;
		}
L
Linus Torvalds 已提交
886
		/* Copy first part of the handler. */
887
		uasm_copy_handler(relocs, labels, tlb_handler, split, f);
L
Linus Torvalds 已提交
888 889
		f += split - tlb_handler;

890 891 892 893 894 895 896 897 898 899 900 901 902
		if (ov) {
			/* Insert branch. */
			uasm_l_split(&l, final_handler);
			uasm_il_b(&f, &r, label_split);
			if (uasm_insn_has_bdelay(relocs, split))
				uasm_i_nop(&f);
			else {
				uasm_copy_handler(relocs, labels,
						  split, split + 1, f);
				uasm_move_labels(labels, f, f + 1, -1);
				f++;
				split++;
			}
L
Linus Torvalds 已提交
903 904 905
		}

		/* Copy the rest of the handler. */
906
		uasm_copy_handler(relocs, labels, split, p, final_handler);
907 908
		final_len = (f - (final_handler + MIPS64_REFILL_INSNS)) +
			    (p - split);
L
Linus Torvalds 已提交
909
	}
910
#endif /* CONFIG_64BIT */
L
Linus Torvalds 已提交
911

912 913 914
	uasm_resolve_relocs(relocs, labels);
	pr_debug("Wrote TLB refill handler (%u instructions).\n",
		 final_len);
L
Linus Torvalds 已提交
915

916
	memcpy((void *)ebase, final_handler, 0x100);
917 918

	dump_handler((u32 *)ebase, 64);
L
Linus Torvalds 已提交
919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935
}

/*
 * TLB load/store/modify handlers.
 *
 * Only the fastpath gets synthesized at runtime, the slowpath for
 * do_page_fault remains normal asm.
 */
extern void tlb_do_page_fault_0(void);
extern void tlb_do_page_fault_1(void);

/*
 * 128 instructions for the fastpath handler is generous and should
 * never be exceeded.
 */
#define FASTPATH_SIZE 128

936 937 938
u32 handle_tlbl[FASTPATH_SIZE] __cacheline_aligned;
u32 handle_tlbs[FASTPATH_SIZE] __cacheline_aligned;
u32 handle_tlbm[FASTPATH_SIZE] __cacheline_aligned;
L
Linus Torvalds 已提交
939

940
static void __cpuinit
941
iPTE_LW(u32 **p, unsigned int pte, unsigned int ptr)
L
Linus Torvalds 已提交
942 943 944 945
{
#ifdef CONFIG_SMP
# ifdef CONFIG_64BIT_PHYS_ADDR
	if (cpu_has_64bits)
946
		uasm_i_lld(p, pte, 0, ptr);
L
Linus Torvalds 已提交
947 948
	else
# endif
949
		UASM_i_LL(p, pte, 0, ptr);
L
Linus Torvalds 已提交
950 951 952
#else
# ifdef CONFIG_64BIT_PHYS_ADDR
	if (cpu_has_64bits)
953
		uasm_i_ld(p, pte, 0, ptr);
L
Linus Torvalds 已提交
954 955
	else
# endif
956
		UASM_i_LW(p, pte, 0, ptr);
L
Linus Torvalds 已提交
957 958 959
#endif
}

960
static void __cpuinit
961
iPTE_SW(u32 **p, struct uasm_reloc **r, unsigned int pte, unsigned int ptr,
962
	unsigned int mode)
L
Linus Torvalds 已提交
963
{
964 965 966 967
#ifdef CONFIG_64BIT_PHYS_ADDR
	unsigned int hwmode = mode & (_PAGE_VALID | _PAGE_DIRTY);
#endif

968
	uasm_i_ori(p, pte, pte, mode);
L
Linus Torvalds 已提交
969 970 971
#ifdef CONFIG_SMP
# ifdef CONFIG_64BIT_PHYS_ADDR
	if (cpu_has_64bits)
972
		uasm_i_scd(p, pte, 0, ptr);
L
Linus Torvalds 已提交
973 974
	else
# endif
975
		UASM_i_SC(p, pte, 0, ptr);
L
Linus Torvalds 已提交
976 977

	if (r10000_llsc_war())
978
		uasm_il_beqzl(p, r, pte, label_smp_pgtable_change);
L
Linus Torvalds 已提交
979
	else
980
		uasm_il_beqz(p, r, pte, label_smp_pgtable_change);
L
Linus Torvalds 已提交
981 982 983

# ifdef CONFIG_64BIT_PHYS_ADDR
	if (!cpu_has_64bits) {
984 985 986 987 988 989 990
		/* no uasm_i_nop needed */
		uasm_i_ll(p, pte, sizeof(pte_t) / 2, ptr);
		uasm_i_ori(p, pte, pte, hwmode);
		uasm_i_sc(p, pte, sizeof(pte_t) / 2, ptr);
		uasm_il_beqz(p, r, pte, label_smp_pgtable_change);
		/* no uasm_i_nop needed */
		uasm_i_lw(p, pte, 0, ptr);
L
Linus Torvalds 已提交
991
	} else
992
		uasm_i_nop(p);
L
Linus Torvalds 已提交
993
# else
994
	uasm_i_nop(p);
L
Linus Torvalds 已提交
995 996 997 998
# endif
#else
# ifdef CONFIG_64BIT_PHYS_ADDR
	if (cpu_has_64bits)
999
		uasm_i_sd(p, pte, 0, ptr);
L
Linus Torvalds 已提交
1000 1001
	else
# endif
1002
		UASM_i_SW(p, pte, 0, ptr);
L
Linus Torvalds 已提交
1003 1004 1005

# ifdef CONFIG_64BIT_PHYS_ADDR
	if (!cpu_has_64bits) {
1006 1007 1008 1009
		uasm_i_lw(p, pte, sizeof(pte_t) / 2, ptr);
		uasm_i_ori(p, pte, pte, hwmode);
		uasm_i_sw(p, pte, sizeof(pte_t) / 2, ptr);
		uasm_i_lw(p, pte, 0, ptr);
L
Linus Torvalds 已提交
1010 1011 1012 1013 1014 1015 1016 1017 1018 1019
	}
# endif
#endif
}

/*
 * Check if PTE is present, if not then jump to LABEL. PTR points to
 * the page table where this PTE is located, PTE will be re-loaded
 * with it's original value.
 */
1020
static void __cpuinit
1021
build_pte_present(u32 **p, struct uasm_reloc **r,
L
Linus Torvalds 已提交
1022 1023
		  unsigned int pte, unsigned int ptr, enum label_id lid)
{
1024 1025 1026
	uasm_i_andi(p, pte, pte, _PAGE_PRESENT | _PAGE_READ);
	uasm_i_xori(p, pte, pte, _PAGE_PRESENT | _PAGE_READ);
	uasm_il_bnez(p, r, pte, lid);
1027
	iPTE_LW(p, pte, ptr);
L
Linus Torvalds 已提交
1028 1029 1030
}

/* Make PTE valid, store result in PTR. */
1031
static void __cpuinit
1032
build_make_valid(u32 **p, struct uasm_reloc **r, unsigned int pte,
L
Linus Torvalds 已提交
1033 1034
		 unsigned int ptr)
{
1035 1036 1037
	unsigned int mode = _PAGE_VALID | _PAGE_ACCESSED;

	iPTE_SW(p, r, pte, ptr, mode);
L
Linus Torvalds 已提交
1038 1039 1040 1041 1042 1043
}

/*
 * Check if PTE can be written to, if not branch to LABEL. Regardless
 * restore PTE with value from PTR when done.
 */
1044
static void __cpuinit
1045
build_pte_writable(u32 **p, struct uasm_reloc **r,
L
Linus Torvalds 已提交
1046 1047
		   unsigned int pte, unsigned int ptr, enum label_id lid)
{
1048 1049 1050
	uasm_i_andi(p, pte, pte, _PAGE_PRESENT | _PAGE_WRITE);
	uasm_i_xori(p, pte, pte, _PAGE_PRESENT | _PAGE_WRITE);
	uasm_il_bnez(p, r, pte, lid);
1051
	iPTE_LW(p, pte, ptr);
L
Linus Torvalds 已提交
1052 1053 1054 1055 1056
}

/* Make PTE writable, update software status bits as well, then store
 * at PTR.
 */
1057
static void __cpuinit
1058
build_make_write(u32 **p, struct uasm_reloc **r, unsigned int pte,
L
Linus Torvalds 已提交
1059 1060
		 unsigned int ptr)
{
1061 1062 1063 1064
	unsigned int mode = (_PAGE_ACCESSED | _PAGE_MODIFIED | _PAGE_VALID
			     | _PAGE_DIRTY);

	iPTE_SW(p, r, pte, ptr, mode);
L
Linus Torvalds 已提交
1065 1066 1067 1068 1069 1070
}

/*
 * Check if PTE can be modified, if not branch to LABEL. Regardless
 * restore PTE with value from PTR when done.
 */
1071
static void __cpuinit
1072
build_pte_modifiable(u32 **p, struct uasm_reloc **r,
L
Linus Torvalds 已提交
1073 1074
		     unsigned int pte, unsigned int ptr, enum label_id lid)
{
1075 1076
	uasm_i_andi(p, pte, pte, _PAGE_WRITE);
	uasm_il_beqz(p, r, pte, lid);
1077
	iPTE_LW(p, pte, ptr);
L
Linus Torvalds 已提交
1078 1079 1080 1081 1082 1083
}

/*
 * R3000 style TLB load/store/modify handlers.
 */

1084 1085 1086 1087
/*
 * This places the pte into ENTRYLO0 and writes it with tlbwi.
 * Then it returns.
 */
1088
static void __cpuinit
1089
build_r3000_pte_reload_tlbwi(u32 **p, unsigned int pte, unsigned int tmp)
L
Linus Torvalds 已提交
1090
{
1091 1092 1093 1094 1095
	uasm_i_mtc0(p, pte, C0_ENTRYLO0); /* cp0 delay */
	uasm_i_mfc0(p, tmp, C0_EPC); /* cp0 delay */
	uasm_i_tlbwi(p);
	uasm_i_jr(p, tmp);
	uasm_i_rfe(p); /* branch delay */
L
Linus Torvalds 已提交
1096 1097 1098
}

/*
1099 1100 1101 1102
 * This places the pte into ENTRYLO0 and writes it with tlbwi
 * or tlbwr as appropriate.  This is because the index register
 * may have the probe fail bit set as a result of a trap on a
 * kseg2 access, i.e. without refill.  Then it returns.
L
Linus Torvalds 已提交
1103
 */
1104
static void __cpuinit
1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119
build_r3000_tlb_reload_write(u32 **p, struct uasm_label **l,
			     struct uasm_reloc **r, unsigned int pte,
			     unsigned int tmp)
{
	uasm_i_mfc0(p, tmp, C0_INDEX);
	uasm_i_mtc0(p, pte, C0_ENTRYLO0); /* cp0 delay */
	uasm_il_bltz(p, r, tmp, label_r3000_write_probe_fail); /* cp0 delay */
	uasm_i_mfc0(p, tmp, C0_EPC); /* branch delay */
	uasm_i_tlbwi(p); /* cp0 delay */
	uasm_i_jr(p, tmp);
	uasm_i_rfe(p); /* branch delay */
	uasm_l_r3000_write_probe_fail(l, *p);
	uasm_i_tlbwr(p); /* cp0 delay */
	uasm_i_jr(p, tmp);
	uasm_i_rfe(p); /* branch delay */
L
Linus Torvalds 已提交
1120 1121
}

1122
static void __cpuinit
L
Linus Torvalds 已提交
1123 1124 1125 1126 1127
build_r3000_tlbchange_handler_head(u32 **p, unsigned int pte,
				   unsigned int ptr)
{
	long pgdc = (long)pgd_current;

1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139
	uasm_i_mfc0(p, pte, C0_BADVADDR);
	uasm_i_lui(p, ptr, uasm_rel_hi(pgdc)); /* cp0 delay */
	uasm_i_lw(p, ptr, uasm_rel_lo(pgdc), ptr);
	uasm_i_srl(p, pte, pte, 22); /* load delay */
	uasm_i_sll(p, pte, pte, 2);
	uasm_i_addu(p, ptr, ptr, pte);
	uasm_i_mfc0(p, pte, C0_CONTEXT);
	uasm_i_lw(p, ptr, 0, ptr); /* cp0 delay */
	uasm_i_andi(p, pte, pte, 0xffc); /* load delay */
	uasm_i_addu(p, ptr, ptr, pte);
	uasm_i_lw(p, pte, 0, ptr);
	uasm_i_tlbp(p); /* load delay */
L
Linus Torvalds 已提交
1140 1141
}

1142
static void __cpuinit build_r3000_tlb_load_handler(void)
L
Linus Torvalds 已提交
1143 1144
{
	u32 *p = handle_tlbl;
1145 1146
	struct uasm_label *l = labels;
	struct uasm_reloc *r = relocs;
L
Linus Torvalds 已提交
1147 1148 1149 1150 1151 1152

	memset(handle_tlbl, 0, sizeof(handle_tlbl));
	memset(labels, 0, sizeof(labels));
	memset(relocs, 0, sizeof(relocs));

	build_r3000_tlbchange_handler_head(&p, K0, K1);
1153
	build_pte_present(&p, &r, K0, K1, label_nopage_tlbl);
1154
	uasm_i_nop(&p); /* load delay */
L
Linus Torvalds 已提交
1155
	build_make_valid(&p, &r, K0, K1);
1156
	build_r3000_tlb_reload_write(&p, &l, &r, K0, K1);
L
Linus Torvalds 已提交
1157

1158 1159 1160
	uasm_l_nopage_tlbl(&l, p);
	uasm_i_j(&p, (unsigned long)tlb_do_page_fault_0 & 0x0fffffff);
	uasm_i_nop(&p);
L
Linus Torvalds 已提交
1161 1162 1163 1164

	if ((p - handle_tlbl) > FASTPATH_SIZE)
		panic("TLB load handler fastpath space exceeded");

1165 1166 1167
	uasm_resolve_relocs(relocs, labels);
	pr_debug("Wrote TLB load handler fastpath (%u instructions).\n",
		 (unsigned int)(p - handle_tlbl));
L
Linus Torvalds 已提交
1168

1169
	dump_handler(handle_tlbl, ARRAY_SIZE(handle_tlbl));
L
Linus Torvalds 已提交
1170 1171
}

1172
static void __cpuinit build_r3000_tlb_store_handler(void)
L
Linus Torvalds 已提交
1173 1174
{
	u32 *p = handle_tlbs;
1175 1176
	struct uasm_label *l = labels;
	struct uasm_reloc *r = relocs;
L
Linus Torvalds 已提交
1177 1178 1179 1180 1181 1182

	memset(handle_tlbs, 0, sizeof(handle_tlbs));
	memset(labels, 0, sizeof(labels));
	memset(relocs, 0, sizeof(relocs));

	build_r3000_tlbchange_handler_head(&p, K0, K1);
1183
	build_pte_writable(&p, &r, K0, K1, label_nopage_tlbs);
1184
	uasm_i_nop(&p); /* load delay */
L
Linus Torvalds 已提交
1185
	build_make_write(&p, &r, K0, K1);
1186
	build_r3000_tlb_reload_write(&p, &l, &r, K0, K1);
L
Linus Torvalds 已提交
1187

1188 1189 1190
	uasm_l_nopage_tlbs(&l, p);
	uasm_i_j(&p, (unsigned long)tlb_do_page_fault_1 & 0x0fffffff);
	uasm_i_nop(&p);
L
Linus Torvalds 已提交
1191 1192 1193 1194

	if ((p - handle_tlbs) > FASTPATH_SIZE)
		panic("TLB store handler fastpath space exceeded");

1195 1196 1197
	uasm_resolve_relocs(relocs, labels);
	pr_debug("Wrote TLB store handler fastpath (%u instructions).\n",
		 (unsigned int)(p - handle_tlbs));
L
Linus Torvalds 已提交
1198

1199
	dump_handler(handle_tlbs, ARRAY_SIZE(handle_tlbs));
L
Linus Torvalds 已提交
1200 1201
}

1202
static void __cpuinit build_r3000_tlb_modify_handler(void)
L
Linus Torvalds 已提交
1203 1204
{
	u32 *p = handle_tlbm;
1205 1206
	struct uasm_label *l = labels;
	struct uasm_reloc *r = relocs;
L
Linus Torvalds 已提交
1207 1208 1209 1210 1211 1212

	memset(handle_tlbm, 0, sizeof(handle_tlbm));
	memset(labels, 0, sizeof(labels));
	memset(relocs, 0, sizeof(relocs));

	build_r3000_tlbchange_handler_head(&p, K0, K1);
1213
	build_pte_modifiable(&p, &r, K0, K1, label_nopage_tlbm);
1214
	uasm_i_nop(&p); /* load delay */
L
Linus Torvalds 已提交
1215
	build_make_write(&p, &r, K0, K1);
1216
	build_r3000_pte_reload_tlbwi(&p, K0, K1);
L
Linus Torvalds 已提交
1217

1218 1219 1220
	uasm_l_nopage_tlbm(&l, p);
	uasm_i_j(&p, (unsigned long)tlb_do_page_fault_1 & 0x0fffffff);
	uasm_i_nop(&p);
L
Linus Torvalds 已提交
1221 1222 1223 1224

	if ((p - handle_tlbm) > FASTPATH_SIZE)
		panic("TLB modify handler fastpath space exceeded");

1225 1226 1227
	uasm_resolve_relocs(relocs, labels);
	pr_debug("Wrote TLB modify handler fastpath (%u instructions).\n",
		 (unsigned int)(p - handle_tlbm));
L
Linus Torvalds 已提交
1228

1229
	dump_handler(handle_tlbm, ARRAY_SIZE(handle_tlbm));
L
Linus Torvalds 已提交
1230 1231 1232 1233 1234
}

/*
 * R4000 style TLB load/store/modify handlers.
 */
1235
static void __cpuinit
1236 1237
build_r4000_tlbchange_handler_head(u32 **p, struct uasm_label **l,
				   struct uasm_reloc **r, unsigned int pte,
L
Linus Torvalds 已提交
1238 1239
				   unsigned int ptr)
{
1240
#ifdef CONFIG_64BIT
L
Linus Torvalds 已提交
1241 1242 1243 1244 1245
	build_get_pmde64(p, l, r, pte, ptr); /* get pmd in ptr */
#else
	build_get_pgde32(p, pte, ptr); /* get pgd in ptr */
#endif

D
David Daney 已提交
1246 1247 1248 1249 1250 1251 1252 1253 1254
#ifdef CONFIG_HUGETLB_PAGE
	/*
	 * For huge tlb entries, pmd doesn't contain an address but
	 * instead contains the tlb pte. Check the PAGE_HUGE bit and
	 * see if we need to jump to huge tlb processing.
	 */
	build_is_huge_pte(p, r, pte, ptr, label_tlb_huge_update);
#endif

1255 1256 1257 1258 1259
	UASM_i_MFC0(p, pte, C0_BADVADDR);
	UASM_i_LW(p, ptr, 0, ptr);
	UASM_i_SRL(p, pte, pte, PAGE_SHIFT + PTE_ORDER - PTE_T_LOG2);
	uasm_i_andi(p, pte, pte, (PTRS_PER_PTE - 1) << PTE_T_LOG2);
	UASM_i_ADDU(p, ptr, ptr, pte);
L
Linus Torvalds 已提交
1260 1261

#ifdef CONFIG_SMP
1262 1263
	uasm_l_smp_pgtable_change(l, *p);
#endif
1264
	iPTE_LW(p, pte, ptr); /* get even pte */
1265 1266
	if (!m4kc_tlbp_war())
		build_tlb_probe_entry(p);
L
Linus Torvalds 已提交
1267 1268
}

1269
static void __cpuinit
1270 1271
build_r4000_tlbchange_handler_tail(u32 **p, struct uasm_label **l,
				   struct uasm_reloc **r, unsigned int tmp,
L
Linus Torvalds 已提交
1272 1273
				   unsigned int ptr)
{
1274 1275
	uasm_i_ori(p, ptr, ptr, sizeof(pte_t));
	uasm_i_xori(p, ptr, ptr, sizeof(pte_t));
L
Linus Torvalds 已提交
1276 1277
	build_update_entries(p, tmp, ptr);
	build_tlb_write_entry(p, l, r, tlb_indexed);
1278 1279
	uasm_l_leave(l, *p);
	uasm_i_eret(p); /* return from trap */
L
Linus Torvalds 已提交
1280

1281
#ifdef CONFIG_64BIT
L
Linus Torvalds 已提交
1282 1283 1284 1285
	build_get_pgd_vmalloc64(p, l, r, tmp, ptr);
#endif
}

1286
static void __cpuinit build_r4000_tlb_load_handler(void)
L
Linus Torvalds 已提交
1287 1288
{
	u32 *p = handle_tlbl;
1289 1290
	struct uasm_label *l = labels;
	struct uasm_reloc *r = relocs;
L
Linus Torvalds 已提交
1291 1292 1293 1294 1295 1296

	memset(handle_tlbl, 0, sizeof(handle_tlbl));
	memset(labels, 0, sizeof(labels));
	memset(relocs, 0, sizeof(relocs));

	if (bcm1250_m3_war()) {
1297 1298 1299 1300 1301 1302
		UASM_i_MFC0(&p, K0, C0_BADVADDR);
		UASM_i_MFC0(&p, K1, C0_ENTRYHI);
		uasm_i_xor(&p, K0, K0, K1);
		UASM_i_SRL(&p, K0, K0, PAGE_SHIFT + 1);
		uasm_il_bnez(&p, &r, K0, label_leave);
		/* No need for uasm_i_nop */
L
Linus Torvalds 已提交
1303 1304 1305
	}

	build_r4000_tlbchange_handler_head(&p, &l, &r, K0, K1);
1306
	build_pte_present(&p, &r, K0, K1, label_nopage_tlbl);
1307 1308
	if (m4kc_tlbp_war())
		build_tlb_probe_entry(&p);
L
Linus Torvalds 已提交
1309 1310 1311
	build_make_valid(&p, &r, K0, K1);
	build_r4000_tlbchange_handler_tail(&p, &l, &r, K0, K1);

D
David Daney 已提交
1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324
#ifdef CONFIG_HUGETLB_PAGE
	/*
	 * This is the entry point when build_r4000_tlbchange_handler_head
	 * spots a huge page.
	 */
	uasm_l_tlb_huge_update(&l, p);
	iPTE_LW(&p, K0, K1);
	build_pte_present(&p, &r, K0, K1, label_nopage_tlbl);
	build_tlb_probe_entry(&p);
	uasm_i_ori(&p, K0, K0, (_PAGE_ACCESSED | _PAGE_VALID));
	build_huge_handler_tail(&p, &r, &l, K0, K1);
#endif

1325 1326 1327
	uasm_l_nopage_tlbl(&l, p);
	uasm_i_j(&p, (unsigned long)tlb_do_page_fault_0 & 0x0fffffff);
	uasm_i_nop(&p);
L
Linus Torvalds 已提交
1328 1329 1330 1331

	if ((p - handle_tlbl) > FASTPATH_SIZE)
		panic("TLB load handler fastpath space exceeded");

1332 1333 1334
	uasm_resolve_relocs(relocs, labels);
	pr_debug("Wrote TLB load handler fastpath (%u instructions).\n",
		 (unsigned int)(p - handle_tlbl));
L
Linus Torvalds 已提交
1335

1336
	dump_handler(handle_tlbl, ARRAY_SIZE(handle_tlbl));
L
Linus Torvalds 已提交
1337 1338
}

1339
static void __cpuinit build_r4000_tlb_store_handler(void)
L
Linus Torvalds 已提交
1340 1341
{
	u32 *p = handle_tlbs;
1342 1343
	struct uasm_label *l = labels;
	struct uasm_reloc *r = relocs;
L
Linus Torvalds 已提交
1344 1345 1346 1347 1348 1349

	memset(handle_tlbs, 0, sizeof(handle_tlbs));
	memset(labels, 0, sizeof(labels));
	memset(relocs, 0, sizeof(relocs));

	build_r4000_tlbchange_handler_head(&p, &l, &r, K0, K1);
1350
	build_pte_writable(&p, &r, K0, K1, label_nopage_tlbs);
1351 1352
	if (m4kc_tlbp_war())
		build_tlb_probe_entry(&p);
L
Linus Torvalds 已提交
1353 1354 1355
	build_make_write(&p, &r, K0, K1);
	build_r4000_tlbchange_handler_tail(&p, &l, &r, K0, K1);

D
David Daney 已提交
1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369
#ifdef CONFIG_HUGETLB_PAGE
	/*
	 * This is the entry point when
	 * build_r4000_tlbchange_handler_head spots a huge page.
	 */
	uasm_l_tlb_huge_update(&l, p);
	iPTE_LW(&p, K0, K1);
	build_pte_writable(&p, &r, K0, K1, label_nopage_tlbs);
	build_tlb_probe_entry(&p);
	uasm_i_ori(&p, K0, K0,
		   _PAGE_ACCESSED | _PAGE_MODIFIED | _PAGE_VALID | _PAGE_DIRTY);
	build_huge_handler_tail(&p, &r, &l, K0, K1);
#endif

1370 1371 1372
	uasm_l_nopage_tlbs(&l, p);
	uasm_i_j(&p, (unsigned long)tlb_do_page_fault_1 & 0x0fffffff);
	uasm_i_nop(&p);
L
Linus Torvalds 已提交
1373 1374 1375 1376

	if ((p - handle_tlbs) > FASTPATH_SIZE)
		panic("TLB store handler fastpath space exceeded");

1377 1378 1379
	uasm_resolve_relocs(relocs, labels);
	pr_debug("Wrote TLB store handler fastpath (%u instructions).\n",
		 (unsigned int)(p - handle_tlbs));
L
Linus Torvalds 已提交
1380

1381
	dump_handler(handle_tlbs, ARRAY_SIZE(handle_tlbs));
L
Linus Torvalds 已提交
1382 1383
}

1384
static void __cpuinit build_r4000_tlb_modify_handler(void)
L
Linus Torvalds 已提交
1385 1386
{
	u32 *p = handle_tlbm;
1387 1388
	struct uasm_label *l = labels;
	struct uasm_reloc *r = relocs;
L
Linus Torvalds 已提交
1389 1390 1391 1392 1393 1394

	memset(handle_tlbm, 0, sizeof(handle_tlbm));
	memset(labels, 0, sizeof(labels));
	memset(relocs, 0, sizeof(relocs));

	build_r4000_tlbchange_handler_head(&p, &l, &r, K0, K1);
1395
	build_pte_modifiable(&p, &r, K0, K1, label_nopage_tlbm);
1396 1397
	if (m4kc_tlbp_war())
		build_tlb_probe_entry(&p);
L
Linus Torvalds 已提交
1398 1399 1400 1401
	/* Present and writable bits set, set accessed and dirty bits. */
	build_make_write(&p, &r, K0, K1);
	build_r4000_tlbchange_handler_tail(&p, &l, &r, K0, K1);

D
David Daney 已提交
1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413 1414 1415
#ifdef CONFIG_HUGETLB_PAGE
	/*
	 * This is the entry point when
	 * build_r4000_tlbchange_handler_head spots a huge page.
	 */
	uasm_l_tlb_huge_update(&l, p);
	iPTE_LW(&p, K0, K1);
	build_pte_modifiable(&p, &r, K0, K1, label_nopage_tlbm);
	build_tlb_probe_entry(&p);
	uasm_i_ori(&p, K0, K0,
		   _PAGE_ACCESSED | _PAGE_MODIFIED | _PAGE_VALID | _PAGE_DIRTY);
	build_huge_handler_tail(&p, &r, &l, K0, K1);
#endif

1416 1417 1418
	uasm_l_nopage_tlbm(&l, p);
	uasm_i_j(&p, (unsigned long)tlb_do_page_fault_1 & 0x0fffffff);
	uasm_i_nop(&p);
L
Linus Torvalds 已提交
1419 1420 1421 1422

	if ((p - handle_tlbm) > FASTPATH_SIZE)
		panic("TLB modify handler fastpath space exceeded");

1423 1424 1425
	uasm_resolve_relocs(relocs, labels);
	pr_debug("Wrote TLB modify handler fastpath (%u instructions).\n",
		 (unsigned int)(p - handle_tlbm));
1426

1427
	dump_handler(handle_tlbm, ARRAY_SIZE(handle_tlbm));
L
Linus Torvalds 已提交
1428 1429
}

1430
void __cpuinit build_tlb_refill_handler(void)
L
Linus Torvalds 已提交
1431 1432 1433 1434 1435 1436 1437 1438
{
	/*
	 * The refill handler is generated per-CPU, multi-node systems
	 * may have local storage for it. The other handlers are only
	 * needed once.
	 */
	static int run_once = 0;

1439
	switch (current_cpu_type()) {
L
Linus Torvalds 已提交
1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474
	case CPU_R2000:
	case CPU_R3000:
	case CPU_R3000A:
	case CPU_R3081E:
	case CPU_TX3912:
	case CPU_TX3922:
	case CPU_TX3927:
		build_r3000_tlb_refill_handler();
		if (!run_once) {
			build_r3000_tlb_load_handler();
			build_r3000_tlb_store_handler();
			build_r3000_tlb_modify_handler();
			run_once++;
		}
		break;

	case CPU_R6000:
	case CPU_R6000A:
		panic("No R6000 TLB refill handler yet");
		break;

	case CPU_R8000:
		panic("No R8000 TLB refill handler yet");
		break;

	default:
		build_r4000_tlb_refill_handler();
		if (!run_once) {
			build_r4000_tlb_load_handler();
			build_r4000_tlb_store_handler();
			build_r4000_tlb_modify_handler();
			run_once++;
		}
	}
}
1475

1476
void __cpuinit flush_tlb_handlers(void)
1477
{
1478
	local_flush_icache_range((unsigned long)handle_tlbl,
1479
			   (unsigned long)handle_tlbl + sizeof(handle_tlbl));
1480
	local_flush_icache_range((unsigned long)handle_tlbs,
1481
			   (unsigned long)handle_tlbs + sizeof(handle_tlbs));
1482
	local_flush_icache_range((unsigned long)handle_tlbm,
1483 1484
			   (unsigned long)handle_tlbm + sizeof(handle_tlbm));
}