op_model_mipsxx.c 10.0 KB
Newer Older
1 2 3 4 5
/*
 * This file is subject to the terms and conditions of the GNU General Public
 * License.  See the file "COPYING" in the main directory of this archive
 * for more details.
 *
6
 * Copyright (C) 2004, 05, 06 by Ralf Baechle
7 8
 * Copyright (C) 2005 by MIPS Technologies, Inc.
 */
9
#include <linux/cpumask.h>
10 11 12
#include <linux/oprofile.h>
#include <linux/interrupt.h>
#include <linux/smp.h>
13
#include <asm/irq_regs.h>
14 15 16

#include "op_impl.h"

R
Ralf Baechle 已提交
17 18 19 20 21
#define M_PERFCTL_EXL			(1UL	  <<  0)
#define M_PERFCTL_KERNEL		(1UL	  <<  1)
#define M_PERFCTL_SUPERVISOR		(1UL	  <<  2)
#define M_PERFCTL_USER			(1UL	  <<  3)
#define M_PERFCTL_INTERRUPT_ENABLE	(1UL	  <<  4)
22
#define M_PERFCTL_EVENT(event)		(((event) & 0x3ff)  << 5)
R
Ralf Baechle 已提交
23
#define M_PERFCTL_VPEID(vpe)		((vpe)	  << 16)
24
#define M_PERFCTL_MT_EN(filter)		((filter) << 20)
R
Ralf Baechle 已提交
25 26 27 28 29 30
#define	   M_TC_EN_ALL			M_PERFCTL_MT_EN(0)
#define	   M_TC_EN_VPE			M_PERFCTL_MT_EN(1)
#define	   M_TC_EN_TC			M_PERFCTL_MT_EN(2)
#define M_PERFCTL_TCID(tcid)		((tcid)	  << 22)
#define M_PERFCTL_WIDE			(1UL	  << 30)
#define M_PERFCTL_MORE			(1UL	  << 31)
31

R
Ralf Baechle 已提交
32
#define M_COUNTER_OVERFLOW		(1UL	  << 31)
33

34
/* Netlogic XLR specific, count events in all threads in a core */
R
Ralf Baechle 已提交
35
#define M_PERFCTL_COUNT_ALL_THREADS	(1UL	  << 13)
36

37 38
static int (*save_perf_irq)(void);

39 40 41 42 43
/*
 * XLR has only one set of counters per core. Designate the
 * first hardware thread in the core for setup and init.
 * Skip CPUs with non-zero hardware thread id (4 hwt per core)
 */
44
#if defined(CONFIG_CPU_XLR) && defined(CONFIG_SMP)
45 46 47 48 49
#define oprofile_skip_cpu(c)	((cpu_logical_map(c) & 0x3) != 0)
#else
#define oprofile_skip_cpu(c)	0
#endif

50
#ifdef CONFIG_MIPS_MT_SMP
51 52 53 54 55
static int cpu_has_mipsmt_pertccounters;
#define WHAT		(M_TC_EN_VPE | \
			 M_PERFCTL_VPEID(cpu_data[smp_processor_id()].vpe_id))
#define vpe_id()	(cpu_has_mipsmt_pertccounters ? \
			0 : cpu_data[smp_processor_id()].vpe_id)
56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72

/*
 * The number of bits to shift to convert between counters per core and
 * counters per VPE.  There is no reasonable interface atm to obtain the
 * number of VPEs used by Linux and in the 34K this number is fixed to two
 * anyways so we hardcore a few things here for the moment.  The way it's
 * done here will ensure that oprofile VSMP kernel will run right on a lesser
 * core like a 24K also or with maxcpus=1.
 */
static inline unsigned int vpe_shift(void)
{
	if (num_possible_cpus() > 1)
		return 1;

	return 0;
}

73
#else
74

75
#define WHAT		0
76
#define vpe_id()	0
77 78 79 80 81 82

static inline unsigned int vpe_shift(void)
{
	return 0;
}

83
#endif
84

85 86 87 88 89 90 91 92 93 94
static inline unsigned int counters_total_to_per_cpu(unsigned int counters)
{
	return counters >> vpe_shift();
}

static inline unsigned int counters_per_cpu_to_total(unsigned int counters)
{
	return counters << vpe_shift();
}

95 96 97 98
#define __define_perf_accessors(r, n, np)				\
									\
static inline unsigned int r_c0_ ## r ## n(void)			\
{									\
99
	unsigned int cpu = vpe_id();					\
100 101 102 103 104 105 106 107 108
									\
	switch (cpu) {							\
	case 0:								\
		return read_c0_ ## r ## n();				\
	case 1:								\
		return read_c0_ ## r ## np();				\
	default:							\
		BUG();							\
	}								\
109
	return 0;							\
110 111 112 113
}									\
									\
static inline void w_c0_ ## r ## n(unsigned int value)			\
{									\
114
	unsigned int cpu = vpe_id();					\
115 116 117 118 119 120 121 122 123 124 125
									\
	switch (cpu) {							\
	case 0:								\
		write_c0_ ## r ## n(value);				\
		return;							\
	case 1:								\
		write_c0_ ## r ## np(value);				\
		return;							\
	default:							\
		BUG();							\
	}								\
126
	return;								\
127 128 129 130
}									\

__define_perf_accessors(perfcntr, 0, 2)
__define_perf_accessors(perfcntr, 1, 3)
131 132
__define_perf_accessors(perfcntr, 2, 0)
__define_perf_accessors(perfcntr, 3, 1)
133 134 135

__define_perf_accessors(perfctrl, 0, 2)
__define_perf_accessors(perfctrl, 1, 3)
136 137
__define_perf_accessors(perfctrl, 2, 0)
__define_perf_accessors(perfctrl, 3, 1)
138

139
struct op_mips_model op_model_mipsxx_ops;
140 141 142 143 144 145

static struct mipsxx_register_config {
	unsigned int control[4];
	unsigned int counter[4];
} reg;

R
Ralf Baechle 已提交
146
/* Compute all of the registers in preparation for enabling profiling.	*/
147 148 149

static void mipsxx_reg_setup(struct op_counter_config *ctr)
{
150
	unsigned int counters = op_model_mipsxx_ops.num_counters;
151 152 153 154 155 156 157 158 159 160 161
	int i;

	/* Compute the performance counter control word.  */
	for (i = 0; i < counters; i++) {
		reg.control[i] = 0;
		reg.counter[i] = 0;

		if (!ctr[i].enabled)
			continue;

		reg.control[i] = M_PERFCTL_EVENT(ctr[i].event) |
R
Ralf Baechle 已提交
162
				 M_PERFCTL_INTERRUPT_ENABLE;
163 164 165 166 167 168
		if (ctr[i].kernel)
			reg.control[i] |= M_PERFCTL_KERNEL;
		if (ctr[i].user)
			reg.control[i] |= M_PERFCTL_USER;
		if (ctr[i].exl)
			reg.control[i] |= M_PERFCTL_EXL;
169
		if (boot_cpu_type() == CPU_XLR)
170
			reg.control[i] |= M_PERFCTL_COUNT_ALL_THREADS;
171 172 173 174
		reg.counter[i] = 0x80000000 - ctr[i].count;
	}
}

R
Ralf Baechle 已提交
175
/* Program all of the registers in preparation for enabling profiling.	*/
176

177
static void mipsxx_cpu_setup(void *args)
178
{
179
	unsigned int counters = op_model_mipsxx_ops.num_counters;
180

181 182 183
	if (oprofile_skip_cpu(smp_processor_id()))
		return;

184 185
	switch (counters) {
	case 4:
186 187
		w_c0_perfctrl3(0);
		w_c0_perfcntr3(reg.counter[3]);
188
	case 3:
189 190
		w_c0_perfctrl2(0);
		w_c0_perfcntr2(reg.counter[2]);
191
	case 2:
192 193
		w_c0_perfctrl1(0);
		w_c0_perfcntr1(reg.counter[1]);
194
	case 1:
195 196
		w_c0_perfctrl0(0);
		w_c0_perfcntr0(reg.counter[0]);
197 198 199 200 201 202
	}
}

/* Start all counters on current CPU */
static void mipsxx_cpu_start(void *args)
{
203
	unsigned int counters = op_model_mipsxx_ops.num_counters;
204

205 206 207
	if (oprofile_skip_cpu(smp_processor_id()))
		return;

208 209
	switch (counters) {
	case 4:
210
		w_c0_perfctrl3(WHAT | reg.control[3]);
211
	case 3:
212
		w_c0_perfctrl2(WHAT | reg.control[2]);
213
	case 2:
214
		w_c0_perfctrl1(WHAT | reg.control[1]);
215
	case 1:
216
		w_c0_perfctrl0(WHAT | reg.control[0]);
217 218 219 220 221 222
	}
}

/* Stop all counters on current CPU */
static void mipsxx_cpu_stop(void *args)
{
223
	unsigned int counters = op_model_mipsxx_ops.num_counters;
224

225 226 227
	if (oprofile_skip_cpu(smp_processor_id()))
		return;

228 229
	switch (counters) {
	case 4:
230
		w_c0_perfctrl3(0);
231
	case 3:
232
		w_c0_perfctrl2(0);
233
	case 2:
234
		w_c0_perfctrl1(0);
235
	case 1:
236
		w_c0_perfctrl0(0);
237 238 239
	}
}

240
static int mipsxx_perfcount_handler(void)
241
{
242
	unsigned int counters = op_model_mipsxx_ops.num_counters;
243 244
	unsigned int control;
	unsigned int counter;
245 246 247 248
	int handled = IRQ_NONE;

	if (cpu_has_mips_r2 && !(read_c0_cause() & (1 << 26)))
		return handled;
249 250 251 252

	switch (counters) {
#define HANDLE_COUNTER(n)						\
	case n + 1:							\
253 254
		control = r_c0_perfctrl ## n();				\
		counter = r_c0_perfcntr ## n();				\
255 256
		if ((control & M_PERFCTL_INTERRUPT_ENABLE) &&		\
		    (counter & M_COUNTER_OVERFLOW)) {			\
257
			oprofile_add_sample(get_irq_regs(), n);		\
258
			w_c0_perfcntr ## n(reg.counter[n]);		\
259
			handled = IRQ_HANDLED;				\
260 261 262 263 264 265
		}
	HANDLE_COUNTER(3)
	HANDLE_COUNTER(2)
	HANDLE_COUNTER(1)
	HANDLE_COUNTER(0)
	}
266 267

	return handled;
268 269 270 271
}

#define M_CONFIG1_PC	(1 << 4)

272
static inline int __n_counters(void)
273 274 275
{
	if (!(read_c0_config1() & M_CONFIG1_PC))
		return 0;
276
	if (!(read_c0_perfctrl0() & M_PERFCTL_MORE))
277
		return 1;
278
	if (!(read_c0_perfctrl1() & M_PERFCTL_MORE))
279
		return 2;
280
	if (!(read_c0_perfctrl2() & M_PERFCTL_MORE))
281 282 283 284 285
		return 3;

	return 4;
}

286 287
static inline int n_counters(void)
{
288 289
	int counters;

290
	switch (current_cpu_type()) {
291 292
	case CPU_R10000:
		counters = 2;
293
		break;
294 295 296 297

	case CPU_R12000:
	case CPU_R14000:
		counters = 4;
298
		break;
299 300 301 302

	default:
		counters = __n_counters();
	}
303 304 305 306

	return counters;
}

307
static void reset_counters(void *arg)
308
{
309
	int counters = (int)(long)arg;
310 311
	switch (counters) {
	case 4:
312 313
		w_c0_perfctrl3(0);
		w_c0_perfcntr3(0);
314
	case 3:
315 316
		w_c0_perfctrl2(0);
		w_c0_perfcntr2(0);
317
	case 2:
318 319
		w_c0_perfctrl1(0);
		w_c0_perfcntr1(0);
320
	case 1:
321 322
		w_c0_perfctrl0(0);
		w_c0_perfcntr0(0);
323 324 325
	}
}

326 327 328 329 330
static irqreturn_t mipsxx_perfcount_int(int irq, void *dev_id)
{
	return mipsxx_perfcount_handler();
}

331 332 333 334 335
static int __init mipsxx_init(void)
{
	int counters;

	counters = n_counters();
336 337
	if (counters == 0) {
		printk(KERN_ERR "Oprofile: CPU has no performance counters\n");
338
		return -ENODEV;
339
	}
340

341 342 343 344 345
#ifdef CONFIG_MIPS_MT_SMP
	cpu_has_mipsmt_pertccounters = read_c0_config7() & (1<<19);
	if (!cpu_has_mipsmt_pertccounters)
		counters = counters_total_to_per_cpu(counters);
#endif
I
Ingo Molnar 已提交
346
	on_each_cpu(reset_counters, (void *)(long)counters, 1);
347

348
	op_model_mipsxx_ops.num_counters = counters;
349
	switch (current_cpu_type()) {
350 351 352 353
	case CPU_M14KC:
		op_model_mipsxx_ops.cpu_type = "mips/M14Kc";
		break;

354 355 356 357
	case CPU_M14KEC:
		op_model_mipsxx_ops.cpu_type = "mips/M14KEc";
		break;

358
	case CPU_20KC:
359
		op_model_mipsxx_ops.cpu_type = "mips/20K";
360 361
		break;

362
	case CPU_24K:
363
		op_model_mipsxx_ops.cpu_type = "mips/24K";
364 365
		break;

366
	case CPU_25KF:
367
		op_model_mipsxx_ops.cpu_type = "mips/25K";
368 369
		break;

370
	case CPU_1004K:
371
	case CPU_34K:
372
		op_model_mipsxx_ops.cpu_type = "mips/34K";
373
		break;
374 375

	case CPU_74K:
376
		op_model_mipsxx_ops.cpu_type = "mips/74K";
377
		break;
378

379
	case CPU_5KC:
380
		op_model_mipsxx_ops.cpu_type = "mips/5K";
381 382
		break;

383 384 385 386 387 388 389 390 391 392 393 394
	case CPU_R10000:
		if ((current_cpu_data.processor_id & 0xff) == 0x20)
			op_model_mipsxx_ops.cpu_type = "mips/r10000-v2.x";
		else
			op_model_mipsxx_ops.cpu_type = "mips/r10000";
		break;

	case CPU_R12000:
	case CPU_R14000:
		op_model_mipsxx_ops.cpu_type = "mips/r12000";
		break;

M
Mark Mason 已提交
395 396
	case CPU_SB1:
	case CPU_SB1A:
397
		op_model_mipsxx_ops.cpu_type = "mips/sb1";
M
Mark Mason 已提交
398 399
		break;

400 401 402 403
	case CPU_LOONGSON1:
		op_model_mipsxx_ops.cpu_type = "mips/loongson1";
		break;

404 405 406 407
	case CPU_XLR:
		op_model_mipsxx_ops.cpu_type = "mips/xlr";
		break;

408 409 410 411 412 413
	default:
		printk(KERN_ERR "Profiling unsupported for this CPU\n");

		return -ENODEV;
	}

414
	save_perf_irq = perf_irq;
415 416
	perf_irq = mipsxx_perfcount_handler;

417 418 419 420
	if ((cp0_perfcount_irq >= 0) && (cp0_compare_irq != cp0_perfcount_irq))
		return request_irq(cp0_perfcount_irq, mipsxx_perfcount_int,
			0, "Perfcounter", save_perf_irq);

421 422 423 424 425
	return 0;
}

static void mipsxx_exit(void)
{
426
	int counters = op_model_mipsxx_ops.num_counters;
427

428 429 430
	if ((cp0_perfcount_irq >= 0) && (cp0_compare_irq != cp0_perfcount_irq))
		free_irq(cp0_perfcount_irq, save_perf_irq);

431
	counters = counters_per_cpu_to_total(counters);
I
Ingo Molnar 已提交
432
	on_each_cpu(reset_counters, (void *)(long)counters, 1);
433

434
	perf_irq = save_perf_irq;
435 436
}

437
struct op_mips_model op_model_mipsxx_ops = {
438 439 440 441 442 443 444
	.reg_setup	= mipsxx_reg_setup,
	.cpu_setup	= mipsxx_cpu_setup,
	.init		= mipsxx_init,
	.exit		= mipsxx_exit,
	.cpu_start	= mipsxx_cpu_start,
	.cpu_stop	= mipsxx_cpu_stop,
};