irq-mips-gic.c 28.7 KB
Newer Older
1 2 3 4 5 6 7 8
/*
 * This file is subject to the terms and conditions of the GNU General Public
 * License.  See the file "COPYING" in the main directory of this archive
 * for more details.
 *
 * Copyright (C) 2008 Ralf Baechle (ralf@linux-mips.org)
 * Copyright (C) 2012 MIPS Technologies, Inc.  All rights reserved.
 */
9
#include <linux/bitmap.h>
10
#include <linux/clocksource.h>
11
#include <linux/init.h>
12
#include <linux/interrupt.h>
13
#include <linux/irq.h>
14
#include <linux/irqchip.h>
15
#include <linux/irqchip/mips-gic.h>
16
#include <linux/of_address.h>
17
#include <linux/sched.h>
18
#include <linux/smp.h>
19

20
#include <asm/mips-cm.h>
S
Steven J. Hill 已提交
21 22
#include <asm/setup.h>
#include <asm/traps.h>
23

24 25
#include <dt-bindings/interrupt-controller/mips-gic.h>

26
unsigned int gic_present;
S
Steven J. Hill 已提交
27

28
struct gic_pcpu_mask {
29
	DECLARE_BITMAP(pcpu_mask, GIC_MAX_INTRS);
30 31
};

32 33 34 35 36 37 38 39 40 41 42 43
struct gic_irq_spec {
	enum {
		GIC_DEVICE,
		GIC_IPI
	} type;

	union {
		struct cpumask *ipimask;
		unsigned int hwirq;
	};
};

44
static unsigned long __gic_base_addr;
45

46
static void __iomem *gic_base;
47
static struct gic_pcpu_mask pcpu_masks[NR_CPUS];
48
static DEFINE_SPINLOCK(gic_lock);
49
static struct irq_domain *gic_irq_domain;
50
static struct irq_domain *gic_dev_domain;
51
static struct irq_domain *gic_ipi_domain;
52
static int gic_shared_intrs;
53
static int gic_vpes;
54
static unsigned int gic_cpu_pin;
55
static unsigned int timer_cpu_pin;
56
static struct irq_chip gic_level_irq_controller, gic_edge_irq_controller;
57
DECLARE_BITMAP(ipi_resrv, GIC_MAX_INTRS);
58
DECLARE_BITMAP(ipi_available, GIC_MAX_INTRS);
59

60 61
static void __gic_irq_dispatch(void);

62
static inline u32 gic_read32(unsigned int reg)
63 64 65 66
{
	return __raw_readl(gic_base + reg);
}

67
static inline u64 gic_read64(unsigned int reg)
68
{
69
	return __raw_readq(gic_base + reg);
70 71
}

72
static inline unsigned long gic_read(unsigned int reg)
73
{
74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101
	if (!mips_cm_is64)
		return gic_read32(reg);
	else
		return gic_read64(reg);
}

static inline void gic_write32(unsigned int reg, u32 val)
{
	return __raw_writel(val, gic_base + reg);
}

static inline void gic_write64(unsigned int reg, u64 val)
{
	return __raw_writeq(val, gic_base + reg);
}

static inline void gic_write(unsigned int reg, unsigned long val)
{
	if (!mips_cm_is64)
		return gic_write32(reg, (u32)val);
	else
		return gic_write64(reg, (u64)val);
}

static inline void gic_update_bits(unsigned int reg, unsigned long mask,
				   unsigned long val)
{
	unsigned long regval;
102 103 104 105 106 107 108 109 110 111

	regval = gic_read(reg);
	regval &= ~mask;
	regval |= val;
	gic_write(reg, regval);
}

static inline void gic_reset_mask(unsigned int intr)
{
	gic_write(GIC_REG(SHARED, GIC_SH_RMASK) + GIC_INTR_OFS(intr),
112
		  1ul << GIC_INTR_BIT(intr));
113 114 115 116 117
}

static inline void gic_set_mask(unsigned int intr)
{
	gic_write(GIC_REG(SHARED, GIC_SH_SMASK) + GIC_INTR_OFS(intr),
118
		  1ul << GIC_INTR_BIT(intr));
119 120 121 122 123
}

static inline void gic_set_polarity(unsigned int intr, unsigned int pol)
{
	gic_update_bits(GIC_REG(SHARED, GIC_SH_SET_POLARITY) +
124 125
			GIC_INTR_OFS(intr), 1ul << GIC_INTR_BIT(intr),
			(unsigned long)pol << GIC_INTR_BIT(intr));
126 127 128 129 130
}

static inline void gic_set_trigger(unsigned int intr, unsigned int trig)
{
	gic_update_bits(GIC_REG(SHARED, GIC_SH_SET_TRIGGER) +
131 132
			GIC_INTR_OFS(intr), 1ul << GIC_INTR_BIT(intr),
			(unsigned long)trig << GIC_INTR_BIT(intr));
133 134 135 136 137
}

static inline void gic_set_dual_edge(unsigned int intr, unsigned int dual)
{
	gic_update_bits(GIC_REG(SHARED, GIC_SH_SET_DUAL) + GIC_INTR_OFS(intr),
138 139
			1ul << GIC_INTR_BIT(intr),
			(unsigned long)dual << GIC_INTR_BIT(intr));
140 141 142 143
}

static inline void gic_map_to_pin(unsigned int intr, unsigned int pin)
{
144 145
	gic_write32(GIC_REG(SHARED, GIC_SH_INTR_MAP_TO_PIN_BASE) +
		    GIC_SH_MAP_TO_PIN(intr), GIC_MAP_TO_PIN_MSK | pin);
146 147 148 149 150 151 152 153 154
}

static inline void gic_map_to_vpe(unsigned int intr, unsigned int vpe)
{
	gic_write(GIC_REG(SHARED, GIC_SH_INTR_MAP_TO_VPE_BASE) +
		  GIC_SH_MAP_TO_VPE_REG_OFF(intr, vpe),
		  GIC_SH_MAP_TO_VPE_REG_BIT(vpe));
}

155
#ifdef CONFIG_CLKSRC_MIPS_GIC
156
u64 gic_read_count(void)
157 158 159
{
	unsigned int hi, hi2, lo;

160
	if (mips_cm_is64)
161
		return (u64)gic_read(GIC_REG(SHARED, GIC_SH_COUNTER));
162

163
	do {
164 165 166
		hi = gic_read32(GIC_REG(SHARED, GIC_SH_COUNTER_63_32));
		lo = gic_read32(GIC_REG(SHARED, GIC_SH_COUNTER_31_00));
		hi2 = gic_read32(GIC_REG(SHARED, GIC_SH_COUNTER_63_32));
167 168
	} while (hi2 != hi);

169
	return (((u64) hi) << 32) + lo;
170
}
171

172 173 174 175
unsigned int gic_get_count_width(void)
{
	unsigned int bits, config;

176
	config = gic_read(GIC_REG(SHARED, GIC_SH_CONFIG));
177 178 179 180 181 182
	bits = 32 + 4 * ((config & GIC_SH_CONFIG_COUNTBITS_MSK) >>
			 GIC_SH_CONFIG_COUNTBITS_SHF);

	return bits;
}

183
void gic_write_compare(u64 cnt)
184
{
185 186 187 188 189 190 191 192
	if (mips_cm_is64) {
		gic_write(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE), cnt);
	} else {
		gic_write32(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE_HI),
					(int)(cnt >> 32));
		gic_write32(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE_LO),
					(int)(cnt & 0xffffffff));
	}
193 194
}

195
void gic_write_cpu_compare(u64 cnt, int cpu)
196 197 198 199 200
{
	unsigned long flags;

	local_irq_save(flags);

201
	gic_write(GIC_REG(VPE_LOCAL, GIC_VPE_OTHER_ADDR), mips_cm_vp_id(cpu));
202 203 204 205 206 207 208 209 210

	if (mips_cm_is64) {
		gic_write(GIC_REG(VPE_OTHER, GIC_VPE_COMPARE), cnt);
	} else {
		gic_write32(GIC_REG(VPE_OTHER, GIC_VPE_COMPARE_HI),
					(int)(cnt >> 32));
		gic_write32(GIC_REG(VPE_OTHER, GIC_VPE_COMPARE_LO),
					(int)(cnt & 0xffffffff));
	}
211 212 213 214

	local_irq_restore(flags);
}

215
u64 gic_read_compare(void)
216 217 218
{
	unsigned int hi, lo;

219
	if (mips_cm_is64)
220
		return (u64)gic_read(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE));
221

222 223
	hi = gic_read32(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE_HI));
	lo = gic_read32(GIC_REG(VPE_LOCAL, GIC_VPE_COMPARE_LO));
224

225
	return (((u64) hi) << 32) + lo;
226
}
227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247

void gic_start_count(void)
{
	u32 gicconfig;

	/* Start the counter */
	gicconfig = gic_read(GIC_REG(SHARED, GIC_SH_CONFIG));
	gicconfig &= ~(1 << GIC_SH_CONFIG_COUNTSTOP_SHF);
	gic_write(GIC_REG(SHARED, GIC_SH_CONFIG), gicconfig);
}

void gic_stop_count(void)
{
	u32 gicconfig;

	/* Stop the counter */
	gicconfig = gic_read(GIC_REG(SHARED, GIC_SH_CONFIG));
	gicconfig |= 1 << GIC_SH_CONFIG_COUNTSTOP_SHF;
	gic_write(GIC_REG(SHARED, GIC_SH_CONFIG), gicconfig);
}

248 249
#endif

250 251 252 253 254 255 256 257
unsigned gic_read_local_vp_id(void)
{
	unsigned long ident;

	ident = gic_read(GIC_REG(VPE_LOCAL, GIC_VP_IDENT));
	return ident & GIC_VP_IDENT_VCNUM_MSK;
}

258 259 260 261 262 263 264 265
static bool gic_local_irq_is_routable(int intr)
{
	u32 vpe_ctl;

	/* All local interrupts are routable in EIC mode. */
	if (cpu_has_veic)
		return true;

266
	vpe_ctl = gic_read32(GIC_REG(VPE_LOCAL, GIC_VPE_CTL));
267 268 269 270 271 272 273 274 275 276 277 278 279 280 281
	switch (intr) {
	case GIC_LOCAL_INT_TIMER:
		return vpe_ctl & GIC_VPE_CTL_TIMER_RTBL_MSK;
	case GIC_LOCAL_INT_PERFCTR:
		return vpe_ctl & GIC_VPE_CTL_PERFCNT_RTBL_MSK;
	case GIC_LOCAL_INT_FDC:
		return vpe_ctl & GIC_VPE_CTL_FDC_RTBL_MSK;
	case GIC_LOCAL_INT_SWINT0:
	case GIC_LOCAL_INT_SWINT1:
		return vpe_ctl & GIC_VPE_CTL_SWINT_RTBL_MSK;
	default:
		return true;
	}
}

282
static void gic_bind_eic_interrupt(int irq, int set)
S
Steven J. Hill 已提交
283 284 285 286 287
{
	/* Convert irq vector # to hw int # */
	irq -= GIC_PIN_TO_VEC_OFFSET;

	/* Set irq to use shadow set */
288 289
	gic_write(GIC_REG(VPE_LOCAL, GIC_VPE_EIC_SHADOW_SET_BASE) +
		  GIC_VPE_EIC_SS(irq), set);
S
Steven J. Hill 已提交
290 291
}

292
static void gic_send_ipi(struct irq_data *d, unsigned int cpu)
293
{
294 295 296
	irq_hw_number_t hwirq = GIC_HWIRQ_TO_SHARED(irqd_to_hwirq(d));

	gic_write(GIC_REG(SHARED, GIC_SH_WEDGE), GIC_SH_WEDGE_SET(hwirq));
297 298
}

299 300 301 302 303 304 305 306 307 308 309
int gic_get_c0_compare_int(void)
{
	if (!gic_local_irq_is_routable(GIC_LOCAL_INT_TIMER))
		return MIPS_CPU_IRQ_BASE + cp0_compare_irq;
	return irq_create_mapping(gic_irq_domain,
				  GIC_LOCAL_TO_HWIRQ(GIC_LOCAL_INT_TIMER));
}

int gic_get_c0_perfcount_int(void)
{
	if (!gic_local_irq_is_routable(GIC_LOCAL_INT_PERFCTR)) {
310
		/* Is the performance counter shared with the timer? */
311 312 313 314 315 316 317 318
		if (cp0_perfcount_irq < 0)
			return -1;
		return MIPS_CPU_IRQ_BASE + cp0_perfcount_irq;
	}
	return irq_create_mapping(gic_irq_domain,
				  GIC_LOCAL_TO_HWIRQ(GIC_LOCAL_INT_PERFCTR));
}

319 320 321 322 323 324 325 326 327 328 329 330 331
int gic_get_c0_fdc_int(void)
{
	if (!gic_local_irq_is_routable(GIC_LOCAL_INT_FDC)) {
		/* Is the FDC IRQ even present? */
		if (cp0_fdc_irq < 0)
			return -1;
		return MIPS_CPU_IRQ_BASE + cp0_fdc_irq;
	}

	return irq_create_mapping(gic_irq_domain,
				  GIC_LOCAL_TO_HWIRQ(GIC_LOCAL_INT_FDC));
}

332 333 334 335 336 337 338 339 340 341 342
int gic_get_usm_range(struct resource *gic_usm_res)
{
	if (!gic_present)
		return -1;

	gic_usm_res->start = __gic_base_addr + USM_VISIBLE_SECTION_OFS;
	gic_usm_res->end = gic_usm_res->start + (USM_VISIBLE_SECTION_SIZE - 1);

	return 0;
}

343
static void gic_handle_shared_int(bool chained)
344
{
345
	unsigned int i, intr, virq, gic_reg_step = mips_cm_is64 ? 8 : 4;
346
	unsigned long *pcpu_mask;
347
	unsigned long pending_reg, intrmask_reg;
348 349
	DECLARE_BITMAP(pending, GIC_MAX_INTRS);
	DECLARE_BITMAP(intrmask, GIC_MAX_INTRS);
350 351 352 353

	/* Get per-cpu bitmaps */
	pcpu_mask = pcpu_masks[smp_processor_id()].pcpu_mask;

354 355
	pending_reg = GIC_REG(SHARED, GIC_SH_PEND);
	intrmask_reg = GIC_REG(SHARED, GIC_SH_MASK);
356

357
	for (i = 0; i < BITS_TO_LONGS(gic_shared_intrs); i++) {
358 359
		pending[i] = gic_read(pending_reg);
		intrmask[i] = gic_read(intrmask_reg);
360 361
		pending_reg += gic_reg_step;
		intrmask_reg += gic_reg_step;
362

363
		if (!IS_ENABLED(CONFIG_64BIT) || mips_cm_is64)
364 365 366 367 368 369
			continue;

		pending[i] |= (u64)gic_read(pending_reg) << 32;
		intrmask[i] |= (u64)gic_read(intrmask_reg) << 32;
		pending_reg += gic_reg_step;
		intrmask_reg += gic_reg_step;
370 371
	}

372 373
	bitmap_and(pending, pending, intrmask, gic_shared_intrs);
	bitmap_and(pending, pending, pcpu_mask, gic_shared_intrs);
374

375
	for_each_set_bit(intr, pending, gic_shared_intrs) {
376 377
		virq = irq_linear_revmap(gic_irq_domain,
					 GIC_SHARED_TO_HWIRQ(intr));
378 379 380 381
		if (chained)
			generic_handle_irq(virq);
		else
			do_IRQ(virq);
382
	}
383 384
}

385
static void gic_mask_irq(struct irq_data *d)
386
{
387
	gic_reset_mask(GIC_HWIRQ_TO_SHARED(d->hwirq));
388 389
}

390
static void gic_unmask_irq(struct irq_data *d)
391
{
392
	gic_set_mask(GIC_HWIRQ_TO_SHARED(d->hwirq));
393 394
}

395 396
static void gic_ack_irq(struct irq_data *d)
{
397
	unsigned int irq = GIC_HWIRQ_TO_SHARED(d->hwirq);
398

399
	gic_write(GIC_REG(SHARED, GIC_SH_WEDGE), GIC_SH_WEDGE_CLR(irq));
400 401
}

402 403
static int gic_set_type(struct irq_data *d, unsigned int type)
{
404
	unsigned int irq = GIC_HWIRQ_TO_SHARED(d->hwirq);
405 406 407 408 409 410
	unsigned long flags;
	bool is_edge;

	spin_lock_irqsave(&gic_lock, flags);
	switch (type & IRQ_TYPE_SENSE_MASK) {
	case IRQ_TYPE_EDGE_FALLING:
411 412 413
		gic_set_polarity(irq, GIC_POL_NEG);
		gic_set_trigger(irq, GIC_TRIG_EDGE);
		gic_set_dual_edge(irq, GIC_TRIG_DUAL_DISABLE);
414 415 416
		is_edge = true;
		break;
	case IRQ_TYPE_EDGE_RISING:
417 418 419
		gic_set_polarity(irq, GIC_POL_POS);
		gic_set_trigger(irq, GIC_TRIG_EDGE);
		gic_set_dual_edge(irq, GIC_TRIG_DUAL_DISABLE);
420 421 422 423
		is_edge = true;
		break;
	case IRQ_TYPE_EDGE_BOTH:
		/* polarity is irrelevant in this case */
424 425
		gic_set_trigger(irq, GIC_TRIG_EDGE);
		gic_set_dual_edge(irq, GIC_TRIG_DUAL_ENABLE);
426 427 428
		is_edge = true;
		break;
	case IRQ_TYPE_LEVEL_LOW:
429 430 431
		gic_set_polarity(irq, GIC_POL_NEG);
		gic_set_trigger(irq, GIC_TRIG_LEVEL);
		gic_set_dual_edge(irq, GIC_TRIG_DUAL_DISABLE);
432 433 434 435
		is_edge = false;
		break;
	case IRQ_TYPE_LEVEL_HIGH:
	default:
436 437 438
		gic_set_polarity(irq, GIC_POL_POS);
		gic_set_trigger(irq, GIC_TRIG_LEVEL);
		gic_set_dual_edge(irq, GIC_TRIG_DUAL_DISABLE);
439 440 441 442
		is_edge = false;
		break;
	}

443 444 445 446 447 448
	if (is_edge)
		irq_set_chip_handler_name_locked(d, &gic_edge_irq_controller,
						 handle_edge_irq, NULL);
	else
		irq_set_chip_handler_name_locked(d, &gic_level_irq_controller,
						 handle_level_irq, NULL);
449
	spin_unlock_irqrestore(&gic_lock, flags);
450

451 452 453 454
	return 0;
}

#ifdef CONFIG_SMP
455 456
static int gic_set_affinity(struct irq_data *d, const struct cpumask *cpumask,
			    bool force)
457
{
458
	unsigned int irq = GIC_HWIRQ_TO_SHARED(d->hwirq);
459 460 461 462
	cpumask_t	tmp = CPU_MASK_NONE;
	unsigned long	flags;
	int		i;

463
	cpumask_and(&tmp, cpumask, cpu_online_mask);
464
	if (cpumask_empty(&tmp))
465
		return -EINVAL;
466 467 468 469

	/* Assumption : cpumask refers to a single CPU */
	spin_lock_irqsave(&gic_lock, flags);

470
	/* Re-route this IRQ */
471
	gic_map_to_vpe(irq, mips_cm_vp_id(cpumask_first(&tmp)));
472 473

	/* Update the pcpu_masks */
474
	for (i = 0; i < min(gic_vpes, NR_CPUS); i++)
475
		clear_bit(irq, pcpu_masks[i].pcpu_mask);
476
	set_bit(irq, pcpu_masks[cpumask_first(&tmp)].pcpu_mask);
477

478
	cpumask_copy(irq_data_get_affinity_mask(d), cpumask);
479 480
	spin_unlock_irqrestore(&gic_lock, flags);

481
	return IRQ_SET_MASK_OK_NOCOPY;
482 483 484
}
#endif

485 486 487 488 489 490 491 492 493 494 495
static struct irq_chip gic_level_irq_controller = {
	.name			=	"MIPS GIC",
	.irq_mask		=	gic_mask_irq,
	.irq_unmask		=	gic_unmask_irq,
	.irq_set_type		=	gic_set_type,
#ifdef CONFIG_SMP
	.irq_set_affinity	=	gic_set_affinity,
#endif
};

static struct irq_chip gic_edge_irq_controller = {
496
	.name			=	"MIPS GIC",
497
	.irq_ack		=	gic_ack_irq,
498 499
	.irq_mask		=	gic_mask_irq,
	.irq_unmask		=	gic_unmask_irq,
500
	.irq_set_type		=	gic_set_type,
501
#ifdef CONFIG_SMP
502
	.irq_set_affinity	=	gic_set_affinity,
503
#endif
504
	.ipi_send_single	=	gic_send_ipi,
505 506
};

507
static void gic_handle_local_int(bool chained)
508 509
{
	unsigned long pending, masked;
510
	unsigned int intr, virq;
511

512 513
	pending = gic_read32(GIC_REG(VPE_LOCAL, GIC_VPE_PEND));
	masked = gic_read32(GIC_REG(VPE_LOCAL, GIC_VPE_MASK));
514 515 516

	bitmap_and(&pending, &pending, &masked, GIC_NUM_LOCAL_INTRS);

517
	for_each_set_bit(intr, &pending, GIC_NUM_LOCAL_INTRS) {
518 519
		virq = irq_linear_revmap(gic_irq_domain,
					 GIC_LOCAL_TO_HWIRQ(intr));
520 521 522 523
		if (chained)
			generic_handle_irq(virq);
		else
			do_IRQ(virq);
524
	}
525 526 527 528 529 530
}

static void gic_mask_local_irq(struct irq_data *d)
{
	int intr = GIC_HWIRQ_TO_LOCAL(d->hwirq);

531
	gic_write32(GIC_REG(VPE_LOCAL, GIC_VPE_RMASK), 1 << intr);
532 533 534 535 536 537
}

static void gic_unmask_local_irq(struct irq_data *d)
{
	int intr = GIC_HWIRQ_TO_LOCAL(d->hwirq);

538
	gic_write32(GIC_REG(VPE_LOCAL, GIC_VPE_SMASK), 1 << intr);
539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554
}

static struct irq_chip gic_local_irq_controller = {
	.name			=	"MIPS GIC Local",
	.irq_mask		=	gic_mask_local_irq,
	.irq_unmask		=	gic_unmask_local_irq,
};

static void gic_mask_local_irq_all_vpes(struct irq_data *d)
{
	int intr = GIC_HWIRQ_TO_LOCAL(d->hwirq);
	int i;
	unsigned long flags;

	spin_lock_irqsave(&gic_lock, flags);
	for (i = 0; i < gic_vpes; i++) {
555 556
		gic_write(GIC_REG(VPE_LOCAL, GIC_VPE_OTHER_ADDR),
			  mips_cm_vp_id(i));
557
		gic_write32(GIC_REG(VPE_OTHER, GIC_VPE_RMASK), 1 << intr);
558 559 560 561 562 563 564 565 566 567 568 569
	}
	spin_unlock_irqrestore(&gic_lock, flags);
}

static void gic_unmask_local_irq_all_vpes(struct irq_data *d)
{
	int intr = GIC_HWIRQ_TO_LOCAL(d->hwirq);
	int i;
	unsigned long flags;

	spin_lock_irqsave(&gic_lock, flags);
	for (i = 0; i < gic_vpes; i++) {
570 571
		gic_write(GIC_REG(VPE_LOCAL, GIC_VPE_OTHER_ADDR),
			  mips_cm_vp_id(i));
572
		gic_write32(GIC_REG(VPE_OTHER, GIC_VPE_SMASK), 1 << intr);
573 574 575 576 577 578 579 580 581 582
	}
	spin_unlock_irqrestore(&gic_lock, flags);
}

static struct irq_chip gic_all_vpes_local_irq_controller = {
	.name			=	"MIPS GIC Local",
	.irq_mask		=	gic_mask_local_irq_all_vpes,
	.irq_unmask		=	gic_unmask_local_irq_all_vpes,
};

583
static void __gic_irq_dispatch(void)
584
{
585 586
	gic_handle_local_int(false);
	gic_handle_shared_int(false);
587
}
588

589
static void gic_irq_dispatch(struct irq_desc *desc)
590
{
591 592
	gic_handle_local_int(true);
	gic_handle_shared_int(true);
593 594
}

595
static void __init gic_basic_init(void)
596 597
{
	unsigned int i;
S
Steven J. Hill 已提交
598 599

	board_bind_eic_interrupt = &gic_bind_eic_interrupt;
600 601

	/* Setup defaults */
602
	for (i = 0; i < gic_shared_intrs; i++) {
603 604 605
		gic_set_polarity(i, GIC_POL_POS);
		gic_set_trigger(i, GIC_TRIG_LEVEL);
		gic_reset_mask(i);
606 607
	}

608 609 610
	for (i = 0; i < gic_vpes; i++) {
		unsigned int j;

611 612
		gic_write(GIC_REG(VPE_LOCAL, GIC_VPE_OTHER_ADDR),
			  mips_cm_vp_id(i));
613 614 615
		for (j = 0; j < GIC_NUM_LOCAL_INTRS; j++) {
			if (!gic_local_irq_is_routable(j))
				continue;
616
			gic_write32(GIC_REG(VPE_OTHER, GIC_VPE_RMASK), 1 << j);
617 618
		}
	}
619 620
}

621 622
static int gic_local_irq_domain_map(struct irq_domain *d, unsigned int virq,
				    irq_hw_number_t hw)
623
{
624 625 626 627 628 629 630 631 632 633 634 635
	int intr = GIC_HWIRQ_TO_LOCAL(hw);
	int ret = 0;
	int i;
	unsigned long flags;

	if (!gic_local_irq_is_routable(intr))
		return -EPERM;

	spin_lock_irqsave(&gic_lock, flags);
	for (i = 0; i < gic_vpes; i++) {
		u32 val = GIC_MAP_TO_PIN_MSK | gic_cpu_pin;

636 637
		gic_write(GIC_REG(VPE_LOCAL, GIC_VPE_OTHER_ADDR),
			  mips_cm_vp_id(i));
638 639 640

		switch (intr) {
		case GIC_LOCAL_INT_WD:
641
			gic_write32(GIC_REG(VPE_OTHER, GIC_VPE_WD_MAP), val);
642 643
			break;
		case GIC_LOCAL_INT_COMPARE:
644 645
			gic_write32(GIC_REG(VPE_OTHER, GIC_VPE_COMPARE_MAP),
				    val);
646 647
			break;
		case GIC_LOCAL_INT_TIMER:
648 649
			/* CONFIG_MIPS_CMP workaround (see __gic_init) */
			val = GIC_MAP_TO_PIN_MSK | timer_cpu_pin;
650 651
			gic_write32(GIC_REG(VPE_OTHER, GIC_VPE_TIMER_MAP),
				    val);
652 653
			break;
		case GIC_LOCAL_INT_PERFCTR:
654 655
			gic_write32(GIC_REG(VPE_OTHER, GIC_VPE_PERFCTR_MAP),
				    val);
656 657
			break;
		case GIC_LOCAL_INT_SWINT0:
658 659
			gic_write32(GIC_REG(VPE_OTHER, GIC_VPE_SWINT0_MAP),
				    val);
660 661
			break;
		case GIC_LOCAL_INT_SWINT1:
662 663
			gic_write32(GIC_REG(VPE_OTHER, GIC_VPE_SWINT1_MAP),
				    val);
664 665
			break;
		case GIC_LOCAL_INT_FDC:
666
			gic_write32(GIC_REG(VPE_OTHER, GIC_VPE_FDC_MAP), val);
667 668 669 670 671 672 673 674 675 676 677 678 679
			break;
		default:
			pr_err("Invalid local IRQ %d\n", intr);
			ret = -EINVAL;
			break;
		}
	}
	spin_unlock_irqrestore(&gic_lock, flags);

	return ret;
}

static int gic_shared_irq_domain_map(struct irq_domain *d, unsigned int virq,
680
				     irq_hw_number_t hw, unsigned int vpe)
681 682
{
	int intr = GIC_HWIRQ_TO_SHARED(hw);
683
	unsigned long flags;
684
	int i;
685 686

	spin_lock_irqsave(&gic_lock, flags);
687
	gic_map_to_pin(intr, gic_cpu_pin);
688
	gic_map_to_vpe(intr, mips_cm_vp_id(vpe));
689
	for (i = 0; i < min(gic_vpes, NR_CPUS); i++)
690
		clear_bit(intr, pcpu_masks[i].pcpu_mask);
691
	set_bit(intr, pcpu_masks[vpe].pcpu_mask);
692 693 694 695 696
	spin_unlock_irqrestore(&gic_lock, flags);

	return 0;
}

697 698
static int gic_setup_dev_chip(struct irq_domain *d, unsigned int virq,
			      unsigned int hwirq)
699
{
700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719
	struct irq_chip *chip;
	int err;

	if (hwirq >= GIC_SHARED_HWIRQ_BASE) {
		err = irq_domain_set_hwirq_and_chip(d, virq, hwirq,
						    &gic_level_irq_controller,
						    NULL);
	} else {
		switch (GIC_HWIRQ_TO_LOCAL(hwirq)) {
		case GIC_LOCAL_INT_TIMER:
		case GIC_LOCAL_INT_PERFCTR:
		case GIC_LOCAL_INT_FDC:
			/*
			 * HACK: These are all really percpu interrupts, but
			 * the rest of the MIPS kernel code does not use the
			 * percpu IRQ API for them.
			 */
			chip = &gic_all_vpes_local_irq_controller;
			irq_set_handler(virq, handle_percpu_irq);
			break;
720

721 722 723 724 725 726
		default:
			chip = &gic_local_irq_controller;
			irq_set_handler(virq, handle_percpu_devid_irq);
			irq_set_percpu_devid(virq);
			break;
		}
727

728 729 730 731 732
		err = irq_domain_set_hwirq_and_chip(d, virq, hwirq,
						    chip, NULL);
	}

	return err;
733 734
}

735 736 737 738 739 740 741 742
static int gic_irq_domain_alloc(struct irq_domain *d, unsigned int virq,
				unsigned int nr_irqs, void *arg)
{
	struct gic_irq_spec *spec = arg;
	irq_hw_number_t hwirq, base_hwirq;
	int cpu, ret, i;

	if (spec->type == GIC_DEVICE) {
743 744 745
		/* verify that shared irqs don't conflict with an IPI irq */
		if ((spec->hwirq >= GIC_SHARED_HWIRQ_BASE) &&
		    test_bit(GIC_HWIRQ_TO_SHARED(spec->hwirq), ipi_resrv))
746
			return -EBUSY;
747

748
		return gic_setup_dev_chip(d, virq, spec->hwirq);
749
	} else {
750
		base_hwirq = find_first_bit(ipi_available, gic_shared_intrs);
751 752 753 754 755 756
		if (base_hwirq == gic_shared_intrs) {
			return -ENOMEM;
		}

		/* check that we have enough space */
		for (i = base_hwirq; i < nr_irqs; i++) {
757
			if (!test_bit(i, ipi_available))
758 759
				return -EBUSY;
		}
760
		bitmap_clear(ipi_available, base_hwirq, nr_irqs);
761 762 763 764 765 766 767

		/* map the hwirq for each cpu consecutively */
		i = 0;
		for_each_cpu(cpu, spec->ipimask) {
			hwirq = GIC_SHARED_TO_HWIRQ(base_hwirq + i);

			ret = irq_domain_set_hwirq_and_chip(d, virq + i, hwirq,
768
							    &gic_level_irq_controller,
769 770 771 772
							    NULL);
			if (ret)
				goto error;

773 774
			irq_set_handler(virq + i, handle_level_irq);

775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790
			ret = gic_shared_irq_domain_map(d, virq + i, hwirq, cpu);
			if (ret)
				goto error;

			i++;
		}

		/*
		 * tell the parent about the base hwirq we allocated so it can
		 * set its own domain data
		 */
		spec->hwirq = base_hwirq;
	}

	return 0;
error:
791
	bitmap_set(ipi_available, base_hwirq, nr_irqs);
792 793 794 795 796 797 798 799 800 801 802 803 804 805
	return ret;
}

void gic_irq_domain_free(struct irq_domain *d, unsigned int virq,
			 unsigned int nr_irqs)
{
	irq_hw_number_t base_hwirq;
	struct irq_data *data;

	data = irq_get_irq_data(virq);
	if (!data)
		return;

	base_hwirq = GIC_HWIRQ_TO_SHARED(irqd_to_hwirq(data));
806
	bitmap_set(ipi_available, base_hwirq, nr_irqs);
807 808
}

809 810 811 812 813 814 815
int gic_irq_domain_match(struct irq_domain *d, struct device_node *node,
			 enum irq_domain_bus_token bus_token)
{
	/* this domain should'nt be accessed directly */
	return 0;
}

816
static const struct irq_domain_ops gic_irq_domain_ops = {
817 818
	.alloc = gic_irq_domain_alloc,
	.free = gic_irq_domain_free,
819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849
	.match = gic_irq_domain_match,
};

static int gic_dev_domain_xlate(struct irq_domain *d, struct device_node *ctrlr,
				const u32 *intspec, unsigned int intsize,
				irq_hw_number_t *out_hwirq,
				unsigned int *out_type)
{
	if (intsize != 3)
		return -EINVAL;

	if (intspec[0] == GIC_SHARED)
		*out_hwirq = GIC_SHARED_TO_HWIRQ(intspec[1]);
	else if (intspec[0] == GIC_LOCAL)
		*out_hwirq = GIC_LOCAL_TO_HWIRQ(intspec[1]);
	else
		return -EINVAL;
	*out_type = intspec[2] & IRQ_TYPE_SENSE_MASK;

	return 0;
}

static int gic_dev_domain_alloc(struct irq_domain *d, unsigned int virq,
				unsigned int nr_irqs, void *arg)
{
	struct irq_fwspec *fwspec = arg;
	struct gic_irq_spec spec = {
		.type = GIC_DEVICE,
	};
	int i, ret;

850 851 852 853
	if (fwspec->param[0] == GIC_SHARED)
		spec.hwirq = GIC_SHARED_TO_HWIRQ(fwspec->param[1]);
	else
		spec.hwirq = GIC_LOCAL_TO_HWIRQ(fwspec->param[1]);
854

855 856 857
	ret = irq_domain_alloc_irqs_parent(d, virq, nr_irqs, &spec);
	if (ret)
		return ret;
858

859 860
	for (i = 0; i < nr_irqs; i++) {
		ret = gic_setup_dev_chip(d, virq + i, spec.hwirq + i);
861
		if (ret)
862
			goto error;
863 864 865
	}

	return 0;
866 867 868 869

error:
	irq_domain_free_irqs_parent(d, virq, nr_irqs);
	return ret;
870 871 872 873 874 875 876 877 878
}

void gic_dev_domain_free(struct irq_domain *d, unsigned int virq,
			 unsigned int nr_irqs)
{
	/* no real allocation is done for dev irqs, so no need to free anything */
	return;
}

879 880 881
static void gic_dev_domain_activate(struct irq_domain *domain,
				    struct irq_data *d)
{
882 883 884 885
	if (GIC_HWIRQ_TO_LOCAL(d->hwirq) < GIC_NUM_LOCAL_INTRS)
		gic_local_irq_domain_map(domain, d->irq, d->hwirq);
	else
		gic_shared_irq_domain_map(domain, d->irq, d->hwirq, 0);
886 887
}

888 889 890 891
static struct irq_domain_ops gic_dev_domain_ops = {
	.xlate = gic_dev_domain_xlate,
	.alloc = gic_dev_domain_alloc,
	.free = gic_dev_domain_free,
892
	.activate = gic_dev_domain_activate,
893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957
};

static int gic_ipi_domain_xlate(struct irq_domain *d, struct device_node *ctrlr,
				const u32 *intspec, unsigned int intsize,
				irq_hw_number_t *out_hwirq,
				unsigned int *out_type)
{
	/*
	 * There's nothing to translate here. hwirq is dynamically allocated and
	 * the irq type is always edge triggered.
	 * */
	*out_hwirq = 0;
	*out_type = IRQ_TYPE_EDGE_RISING;

	return 0;
}

static int gic_ipi_domain_alloc(struct irq_domain *d, unsigned int virq,
				unsigned int nr_irqs, void *arg)
{
	struct cpumask *ipimask = arg;
	struct gic_irq_spec spec = {
		.type = GIC_IPI,
		.ipimask = ipimask
	};
	int ret, i;

	ret = irq_domain_alloc_irqs_parent(d, virq, nr_irqs, &spec);
	if (ret)
		return ret;

	/* the parent should have set spec.hwirq to the base_hwirq it allocated */
	for (i = 0; i < nr_irqs; i++) {
		ret = irq_domain_set_hwirq_and_chip(d, virq + i,
						    GIC_SHARED_TO_HWIRQ(spec.hwirq + i),
						    &gic_edge_irq_controller,
						    NULL);
		if (ret)
			goto error;

		ret = irq_set_irq_type(virq + i, IRQ_TYPE_EDGE_RISING);
		if (ret)
			goto error;
	}

	return 0;
error:
	irq_domain_free_irqs_parent(d, virq, nr_irqs);
	return ret;
}

void gic_ipi_domain_free(struct irq_domain *d, unsigned int virq,
			 unsigned int nr_irqs)
{
	irq_domain_free_irqs_parent(d, virq, nr_irqs);
}

int gic_ipi_domain_match(struct irq_domain *d, struct device_node *node,
			 enum irq_domain_bus_token bus_token)
{
	bool is_ipi;

	switch (bus_token) {
	case DOMAIN_BUS_IPI:
		is_ipi = d->bus_token == bus_token;
958
		return (!node || to_of_node(d->fwnode) == node) && is_ipi;
959 960 961 962 963 964 965 966 967 968 969
		break;
	default:
		return 0;
	}
}

static struct irq_domain_ops gic_ipi_domain_ops = {
	.xlate = gic_ipi_domain_xlate,
	.alloc = gic_ipi_domain_alloc,
	.free = gic_ipi_domain_free,
	.match = gic_ipi_domain_match,
970 971
};

972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994
static void __init gic_map_single_int(struct device_node *node,
				      unsigned int irq)
{
	unsigned int linux_irq;
	struct irq_fwspec local_int_fwspec = {
		.fwnode         = &node->fwnode,
		.param_count    = 3,
		.param          = {
			[0]     = GIC_LOCAL,
			[1]     = irq,
			[2]     = IRQ_TYPE_NONE,
		},
	};

	if (!gic_local_irq_is_routable(irq))
		return;

	linux_irq = irq_create_fwspec_mapping(&local_int_fwspec);
	WARN_ON(!linux_irq);
}

static void __init gic_map_interrupts(struct device_node *node)
{
995 996
	gic_map_single_int(node, GIC_LOCAL_INT_WD);
	gic_map_single_int(node, GIC_LOCAL_INT_COMPARE);
997 998
	gic_map_single_int(node, GIC_LOCAL_INT_TIMER);
	gic_map_single_int(node, GIC_LOCAL_INT_PERFCTR);
999 1000
	gic_map_single_int(node, GIC_LOCAL_INT_SWINT0);
	gic_map_single_int(node, GIC_LOCAL_INT_SWINT1);
1001 1002 1003
	gic_map_single_int(node, GIC_LOCAL_INT_FDC);
}

1004 1005 1006 1007
static void __init __gic_init(unsigned long gic_base_addr,
			      unsigned long gic_addrspace_size,
			      unsigned int cpu_vec, unsigned int irqbase,
			      struct device_node *node)
1008
{
1009
	unsigned int gicconfig, cpu;
1010
	unsigned int v[2];
1011

1012 1013
	__gic_base_addr = gic_base_addr;

1014
	gic_base = ioremap_nocache(gic_base_addr, gic_addrspace_size);
1015

1016
	gicconfig = gic_read(GIC_REG(SHARED, GIC_SH_CONFIG));
1017
	gic_shared_intrs = (gicconfig & GIC_SH_CONFIG_NUMINTRS_MSK) >>
1018
		   GIC_SH_CONFIG_NUMINTRS_SHF;
1019
	gic_shared_intrs = ((gic_shared_intrs + 1) * 8);
1020

1021
	gic_vpes = (gicconfig & GIC_SH_CONFIG_NUMVPES_MSK) >>
1022
		  GIC_SH_CONFIG_NUMVPES_SHF;
1023
	gic_vpes = gic_vpes + 1;
1024

1025
	if (cpu_has_veic) {
1026 1027 1028 1029 1030 1031 1032 1033
		/* Set EIC mode for all VPEs */
		for_each_present_cpu(cpu) {
			gic_write(GIC_REG(VPE_LOCAL, GIC_VPE_OTHER_ADDR),
				  mips_cm_vp_id(cpu));
			gic_write(GIC_REG(VPE_OTHER, GIC_VPE_CTL),
				  GIC_VPE_CTL_EIC_MODE_MSK);
		}

1034 1035
		/* Always use vector 1 in EIC mode */
		gic_cpu_pin = 0;
1036
		timer_cpu_pin = gic_cpu_pin;
1037 1038 1039 1040 1041 1042
		set_vi_handler(gic_cpu_pin + GIC_PIN_TO_VEC_OFFSET,
			       __gic_irq_dispatch);
	} else {
		gic_cpu_pin = cpu_vec - GIC_CPU_PIN_OFFSET;
		irq_set_chained_handler(MIPS_CPU_IRQ_BASE + cpu_vec,
					gic_irq_dispatch);
1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055
		/*
		 * With the CMP implementation of SMP (deprecated), other CPUs
		 * are started by the bootloader and put into a timer based
		 * waiting poll loop. We must not re-route those CPU's local
		 * timer interrupts as the wait instruction will never finish,
		 * so just handle whatever CPU interrupt it is routed to by
		 * default.
		 *
		 * This workaround should be removed when CMP support is
		 * dropped.
		 */
		if (IS_ENABLED(CONFIG_MIPS_CMP) &&
		    gic_local_irq_is_routable(GIC_LOCAL_INT_TIMER)) {
1056
			timer_cpu_pin = gic_read32(GIC_REG(VPE_LOCAL,
1057 1058 1059 1060 1061 1062 1063 1064 1065
							 GIC_VPE_TIMER_MAP)) &
					GIC_MAP_MSK;
			irq_set_chained_handler(MIPS_CPU_IRQ_BASE +
						GIC_CPU_PIN_OFFSET +
						timer_cpu_pin,
						gic_irq_dispatch);
		} else {
			timer_cpu_pin = gic_cpu_pin;
		}
1066 1067
	}

1068
	gic_irq_domain = irq_domain_add_simple(node, GIC_NUM_LOCAL_INTRS +
1069
					       gic_shared_intrs, irqbase,
1070 1071 1072
					       &gic_irq_domain_ops, NULL);
	if (!gic_irq_domain)
		panic("Failed to add GIC IRQ domain");
1073
	gic_irq_domain->name = "mips-gic-irq";
1074

1075 1076 1077 1078 1079
	gic_dev_domain = irq_domain_add_hierarchy(gic_irq_domain, 0,
						  GIC_NUM_LOCAL_INTRS + gic_shared_intrs,
						  node, &gic_dev_domain_ops, NULL);
	if (!gic_dev_domain)
		panic("Failed to add GIC DEV domain");
1080
	gic_dev_domain->name = "mips-gic-dev";
1081

1082 1083 1084 1085 1086 1087 1088
	gic_ipi_domain = irq_domain_add_hierarchy(gic_irq_domain,
						  IRQ_DOMAIN_FLAG_IPI_PER_CPU,
						  GIC_NUM_LOCAL_INTRS + gic_shared_intrs,
						  node, &gic_ipi_domain_ops, NULL);
	if (!gic_ipi_domain)
		panic("Failed to add GIC IPI domain");

1089
	gic_ipi_domain->name = "mips-gic-ipi";
1090 1091
	gic_ipi_domain->bus_token = DOMAIN_BUS_IPI;

1092 1093 1094 1095 1096 1097 1098 1099 1100
	if (node &&
	    !of_property_read_u32_array(node, "mti,reserved-ipi-vectors", v, 2)) {
		bitmap_set(ipi_resrv, v[0], v[1]);
	} else {
		/* Make the last 2 * gic_vpes available for IPIs */
		bitmap_set(ipi_resrv,
			   gic_shared_intrs - 2 * gic_vpes,
			   2 * gic_vpes);
	}
1101

1102
	bitmap_copy(ipi_available, ipi_resrv, GIC_MAX_INTRS);
1103
	gic_basic_init();
1104
	gic_map_interrupts(node);
1105
}
1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161

void __init gic_init(unsigned long gic_base_addr,
		     unsigned long gic_addrspace_size,
		     unsigned int cpu_vec, unsigned int irqbase)
{
	__gic_init(gic_base_addr, gic_addrspace_size, cpu_vec, irqbase, NULL);
}

static int __init gic_of_init(struct device_node *node,
			      struct device_node *parent)
{
	struct resource res;
	unsigned int cpu_vec, i = 0, reserved = 0;
	phys_addr_t gic_base;
	size_t gic_len;

	/* Find the first available CPU vector. */
	while (!of_property_read_u32_index(node, "mti,reserved-cpu-vectors",
					   i++, &cpu_vec))
		reserved |= BIT(cpu_vec);
	for (cpu_vec = 2; cpu_vec < 8; cpu_vec++) {
		if (!(reserved & BIT(cpu_vec)))
			break;
	}
	if (cpu_vec == 8) {
		pr_err("No CPU vectors available for GIC\n");
		return -ENODEV;
	}

	if (of_address_to_resource(node, 0, &res)) {
		/*
		 * Probe the CM for the GIC base address if not specified
		 * in the device-tree.
		 */
		if (mips_cm_present()) {
			gic_base = read_gcr_gic_base() &
				~CM_GCR_GIC_BASE_GICEN_MSK;
			gic_len = 0x20000;
		} else {
			pr_err("Failed to get GIC memory range\n");
			return -ENODEV;
		}
	} else {
		gic_base = res.start;
		gic_len = resource_size(&res);
	}

	if (mips_cm_present())
		write_gcr_gic_base(gic_base | CM_GCR_GIC_BASE_GICEN_MSK);
	gic_present = true;

	__gic_init(gic_base, gic_len, cpu_vec, 0, node);

	return 0;
}
IRQCHIP_DECLARE(mips_gic, "mti,gic", gic_of_init);