intel_cacheinfo.c 25.9 KB
Newer Older
L
Linus Torvalds 已提交
1
/*
2
 *	Routines to indentify caches on Intel CPU.
L
Linus Torvalds 已提交
3
 *
4 5
 *	Changes:
 *	Venkatesh Pallipadi	: Adding cache identification through cpuid(4)
6
 *		Ashok Raj <ashok.raj@intel.com>: Work with CPU hotplug infrastructure.
7
 *	Andi Kleen / Andreas Herrmann	: CPUID4 emulation on AMD.
L
Linus Torvalds 已提交
8 9 10 11 12 13 14
 */

#include <linux/init.h>
#include <linux/slab.h>
#include <linux/device.h>
#include <linux/compiler.h>
#include <linux/cpu.h>
T
Tim Schmielau 已提交
15
#include <linux/sched.h>
16
#include <linux/pci.h>
L
Linus Torvalds 已提交
17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34

#include <asm/processor.h>
#include <asm/smp.h>

#define LVL_1_INST	1
#define LVL_1_DATA	2
#define LVL_2		3
#define LVL_3		4
#define LVL_TRACE	5

struct _cache_table
{
	unsigned char descriptor;
	char cache_type;
	short size;
};

/* all the cache descriptor types we care about (no TLB or trace cache entries) */
35
static struct _cache_table cache_table[] __cpuinitdata =
L
Linus Torvalds 已提交
36 37 38 39 40 41 42 43 44 45 46 47
{
	{ 0x06, LVL_1_INST, 8 },	/* 4-way set assoc, 32 byte line size */
	{ 0x08, LVL_1_INST, 16 },	/* 4-way set assoc, 32 byte line size */
	{ 0x0a, LVL_1_DATA, 8 },	/* 2 way set assoc, 32 byte line size */
	{ 0x0c, LVL_1_DATA, 16 },	/* 4-way set assoc, 32 byte line size */
	{ 0x22, LVL_3,      512 },	/* 4-way set assoc, sectored cache, 64 byte line size */
	{ 0x23, LVL_3,      1024 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x25, LVL_3,      2048 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x29, LVL_3,      4096 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x2c, LVL_1_DATA, 32 },	/* 8-way set assoc, 64 byte line size */
	{ 0x30, LVL_1_INST, 32 },	/* 8-way set assoc, 64 byte line size */
	{ 0x39, LVL_2,      128 },	/* 4-way set assoc, sectored cache, 64 byte line size */
48
	{ 0x3a, LVL_2,      192 },	/* 6-way set assoc, sectored cache, 64 byte line size */
L
Linus Torvalds 已提交
49 50
	{ 0x3b, LVL_2,      128 },	/* 2-way set assoc, sectored cache, 64 byte line size */
	{ 0x3c, LVL_2,      256 },	/* 4-way set assoc, sectored cache, 64 byte line size */
51 52
	{ 0x3d, LVL_2,      384 },	/* 6-way set assoc, sectored cache, 64 byte line size */
	{ 0x3e, LVL_2,      512 },	/* 4-way set assoc, sectored cache, 64 byte line size */
53
	{ 0x3f, LVL_2,      256 },	/* 2-way set assoc, 64 byte line size */
L
Linus Torvalds 已提交
54 55 56 57 58
	{ 0x41, LVL_2,      128 },	/* 4-way set assoc, 32 byte line size */
	{ 0x42, LVL_2,      256 },	/* 4-way set assoc, 32 byte line size */
	{ 0x43, LVL_2,      512 },	/* 4-way set assoc, 32 byte line size */
	{ 0x44, LVL_2,      1024 },	/* 4-way set assoc, 32 byte line size */
	{ 0x45, LVL_2,      2048 },	/* 4-way set assoc, 32 byte line size */
59 60 61 62 63 64 65
	{ 0x46, LVL_3,      4096 },	/* 4-way set assoc, 64 byte line size */
	{ 0x47, LVL_3,      8192 },	/* 8-way set assoc, 64 byte line size */
	{ 0x49, LVL_3,      4096 },	/* 16-way set assoc, 64 byte line size */
	{ 0x4a, LVL_3,      6144 },	/* 12-way set assoc, 64 byte line size */
	{ 0x4b, LVL_3,      8192 },	/* 16-way set assoc, 64 byte line size */
	{ 0x4c, LVL_3,     12288 },	/* 12-way set assoc, 64 byte line size */
	{ 0x4d, LVL_3,     16384 },	/* 16-way set assoc, 64 byte line size */
D
Dave Jones 已提交
66
	{ 0x4e, LVL_2,      6144 },	/* 24-way set assoc, 64 byte line size */
L
Linus Torvalds 已提交
67 68 69 70 71 72 73
	{ 0x60, LVL_1_DATA, 16 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x66, LVL_1_DATA, 8 },	/* 4-way set assoc, sectored cache, 64 byte line size */
	{ 0x67, LVL_1_DATA, 16 },	/* 4-way set assoc, sectored cache, 64 byte line size */
	{ 0x68, LVL_1_DATA, 32 },	/* 4-way set assoc, sectored cache, 64 byte line size */
	{ 0x70, LVL_TRACE,  12 },	/* 8-way set assoc */
	{ 0x71, LVL_TRACE,  16 },	/* 8-way set assoc */
	{ 0x72, LVL_TRACE,  32 },	/* 8-way set assoc */
74
	{ 0x73, LVL_TRACE,  64 },	/* 8-way set assoc */
L
Linus Torvalds 已提交
75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133
	{ 0x78, LVL_2,    1024 },	/* 4-way set assoc, 64 byte line size */
	{ 0x79, LVL_2,     128 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x7a, LVL_2,     256 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x7b, LVL_2,     512 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x7c, LVL_2,    1024 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x7d, LVL_2,    2048 },	/* 8-way set assoc, 64 byte line size */
	{ 0x7f, LVL_2,     512 },	/* 2-way set assoc, 64 byte line size */
	{ 0x82, LVL_2,     256 },	/* 8-way set assoc, 32 byte line size */
	{ 0x83, LVL_2,     512 },	/* 8-way set assoc, 32 byte line size */
	{ 0x84, LVL_2,    1024 },	/* 8-way set assoc, 32 byte line size */
	{ 0x85, LVL_2,    2048 },	/* 8-way set assoc, 32 byte line size */
	{ 0x86, LVL_2,     512 },	/* 4-way set assoc, 64 byte line size */
	{ 0x87, LVL_2,    1024 },	/* 8-way set assoc, 64 byte line size */
	{ 0x00, 0, 0}
};


enum _cache_type
{
	CACHE_TYPE_NULL	= 0,
	CACHE_TYPE_DATA = 1,
	CACHE_TYPE_INST = 2,
	CACHE_TYPE_UNIFIED = 3
};

union _cpuid4_leaf_eax {
	struct {
		enum _cache_type	type:5;
		unsigned int		level:3;
		unsigned int		is_self_initializing:1;
		unsigned int		is_fully_associative:1;
		unsigned int		reserved:4;
		unsigned int		num_threads_sharing:12;
		unsigned int		num_cores_on_die:6;
	} split;
	u32 full;
};

union _cpuid4_leaf_ebx {
	struct {
		unsigned int		coherency_line_size:12;
		unsigned int		physical_line_partition:10;
		unsigned int		ways_of_associativity:10;
	} split;
	u32 full;
};

union _cpuid4_leaf_ecx {
	struct {
		unsigned int		number_of_sets:32;
	} split;
	u32 full;
};

struct _cpuid4_info {
	union _cpuid4_leaf_eax eax;
	union _cpuid4_leaf_ebx ebx;
	union _cpuid4_leaf_ecx ecx;
	unsigned long size;
134
	unsigned long can_disable;
135 136 137 138 139 140 141 142 143 144
	DECLARE_BITMAP(shared_cpu_map, NR_CPUS);
};

/* subset of above _cpuid4_info w/o shared_cpu_map */
struct _cpuid4_info_regs {
	union _cpuid4_leaf_eax eax;
	union _cpuid4_leaf_ebx ebx;
	union _cpuid4_leaf_ecx ecx;
	unsigned long size;
	unsigned long can_disable;
L
Linus Torvalds 已提交
145 146
};

147
#ifdef CONFIG_PCI
148
static struct pci_device_id k8_nb_id[] = {
149 150 151
	{ PCI_DEVICE(PCI_VENDOR_ID_AMD, 0x1103) },
	{ PCI_DEVICE(PCI_VENDOR_ID_AMD, 0x1203) },
	{}
152
};
153
#endif
154

155 156 157 158
unsigned short			num_cache_leaves;

/* AMD doesn't have CPUID4. Emulate it here to report the same
   information to the user.  This makes some assumptions about the machine:
159
   L2 not shared, no SMT etc. that is currently true on AMD CPUs.
160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182

   In theory the TLBs could be reported as fake type (they are in "dummy").
   Maybe later */
union l1_cache {
	struct {
		unsigned line_size : 8;
		unsigned lines_per_tag : 8;
		unsigned assoc : 8;
		unsigned size_in_kb : 8;
	};
	unsigned val;
};

union l2_cache {
	struct {
		unsigned line_size : 8;
		unsigned lines_per_tag : 4;
		unsigned assoc : 4;
		unsigned size_in_kb : 16;
	};
	unsigned val;
};

183 184 185 186 187 188 189 190 191 192 193
union l3_cache {
	struct {
		unsigned line_size : 8;
		unsigned lines_per_tag : 4;
		unsigned assoc : 4;
		unsigned res : 2;
		unsigned size_encoded : 14;
	};
	unsigned val;
};

194
static unsigned short assocs[] __cpuinitdata = {
195
	[1] = 1, [2] = 2, [4] = 4, [6] = 8,
196 197
	[8] = 16, [0xa] = 32, [0xb] = 48,
	[0xc] = 64,
198
	[0xf] = 0xffff // ??
199 200
};

201 202
static unsigned char levels[] __cpuinitdata = { 1, 1, 2, 3 };
static unsigned char types[] __cpuinitdata = { 1, 2, 3, 3 };
203

204 205 206 207
static void __cpuinit
amd_cpuid4(int leaf, union _cpuid4_leaf_eax *eax,
		     union _cpuid4_leaf_ebx *ebx,
		     union _cpuid4_leaf_ecx *ecx)
208 209 210 211 212
{
	unsigned dummy;
	unsigned line_size, lines_per_tag, assoc, size_in_kb;
	union l1_cache l1i, l1d;
	union l2_cache l2;
213 214
	union l3_cache l3;
	union l1_cache *l1 = &l1d;
215 216 217 218 219 220

	eax->full = 0;
	ebx->full = 0;
	ecx->full = 0;

	cpuid(0x80000005, &dummy, &dummy, &l1d.val, &l1i.val);
221
	cpuid(0x80000006, &dummy, &dummy, &l2.val, &l3.val);
222

223 224 225 226 227 228
	switch (leaf) {
	case 1:
		l1 = &l1i;
	case 0:
		if (!l1->val)
			return;
229 230 231 232
		assoc = l1->assoc;
		line_size = l1->line_size;
		lines_per_tag = l1->lines_per_tag;
		size_in_kb = l1->size_in_kb;
233 234 235 236
		break;
	case 2:
		if (!l2.val)
			return;
237 238 239 240 241
		assoc = l2.assoc;
		line_size = l2.line_size;
		lines_per_tag = l2.lines_per_tag;
		/* cpu_data has errata corrections for K7 applied */
		size_in_kb = current_cpu_data.x86_cache_size;
242 243 244 245 246 247 248 249 250 251 252
		break;
	case 3:
		if (!l3.val)
			return;
		assoc = l3.assoc;
		line_size = l3.line_size;
		lines_per_tag = l3.lines_per_tag;
		size_in_kb = l3.size_encoded * 512;
		break;
	default:
		return;
253 254
	}

255 256 257 258 259 260 261 262 263 264
	eax->split.is_self_initializing = 1;
	eax->split.type = types[leaf];
	eax->split.level = levels[leaf];
	if (leaf == 3)
		eax->split.num_threads_sharing = current_cpu_data.x86_max_cores - 1;
	else
		eax->split.num_threads_sharing = 0;
	eax->split.num_cores_on_die = current_cpu_data.x86_max_cores - 1;


265 266 267 268 269 270 271 272
	if (assoc == 0xf)
		eax->split.is_fully_associative = 1;
	ebx->split.coherency_line_size = line_size - 1;
	ebx->split.ways_of_associativity = assocs[assoc] - 1;
	ebx->split.physical_line_partition = lines_per_tag - 1;
	ecx->split.number_of_sets = (size_in_kb * 1024) / line_size /
		(ebx->split.ways_of_associativity + 1) - 1;
}
L
Linus Torvalds 已提交
273

274
static void __cpuinit
275
amd_check_l3_disable(int index, struct _cpuid4_info_regs *this_leaf)
276 277 278
{
	if (index < 3)
		return;
279
	this_leaf->can_disable = 1;
280 281
}

282
static int
283 284
__cpuinit cpuid4_cache_lookup_regs(int index,
				   struct _cpuid4_info_regs *this_leaf)
L
Linus Torvalds 已提交
285
{
286 287 288 289
	union _cpuid4_leaf_eax 	eax;
	union _cpuid4_leaf_ebx 	ebx;
	union _cpuid4_leaf_ecx 	ecx;
	unsigned		edx;
L
Linus Torvalds 已提交
290

291
	if (boot_cpu_data.x86_vendor == X86_VENDOR_AMD) {
292
		amd_cpuid4(index, &eax, &ebx, &ecx);
293 294
		if (boot_cpu_data.x86 >= 0x10)
			amd_check_l3_disable(index, this_leaf);
295 296 297 298
	} else {
		cpuid_count(4, index, &eax.full, &ebx.full, &ecx.full, &edx);
	}

299
	if (eax.split.type == CACHE_TYPE_NULL)
300
		return -EIO; /* better error ? */
L
Linus Torvalds 已提交
301

302 303 304
	this_leaf->eax = eax;
	this_leaf->ebx = ebx;
	this_leaf->ecx = ecx;
305 306 307 308
	this_leaf->size = (ecx.split.number_of_sets          + 1) *
			  (ebx.split.coherency_line_size     + 1) *
			  (ebx.split.physical_line_partition + 1) *
			  (ebx.split.ways_of_associativity   + 1);
L
Linus Torvalds 已提交
309 310 311
	return 0;
}

312 313 314 315 316 317 318 319 320
static int
__cpuinit cpuid4_cache_lookup(int index, struct _cpuid4_info *this_leaf)
{
	struct _cpuid4_info_regs *leaf_regs =
		(struct _cpuid4_info_regs *)this_leaf;

	return cpuid4_cache_lookup_regs(index, leaf_regs);
}

321
static int __cpuinit find_num_cache_leaves(void)
L
Linus Torvalds 已提交
322 323 324
{
	unsigned int		eax, ebx, ecx, edx;
	union _cpuid4_leaf_eax	cache_eax;
325
	int 			i = -1;
L
Linus Torvalds 已提交
326

327 328 329
	do {
		++i;
		/* Do cpuid(4) loop to find out num_cache_leaves */
L
Linus Torvalds 已提交
330 331
		cpuid_count(4, i, &eax, &ebx, &ecx, &edx);
		cache_eax.full = eax;
332 333
	} while (cache_eax.split.type != CACHE_TYPE_NULL);
	return i;
L
Linus Torvalds 已提交
334 335
}

336
unsigned int __cpuinit init_intel_cacheinfo(struct cpuinfo_x86 *c)
L
Linus Torvalds 已提交
337 338 339 340
{
	unsigned int trace = 0, l1i = 0, l1d = 0, l2 = 0, l3 = 0; /* Cache sizes */
	unsigned int new_l1d = 0, new_l1i = 0; /* Cache sizes from cpuid(4) */
	unsigned int new_l2 = 0, new_l3 = 0, i; /* Cache sizes from cpuid(4) */
341
	unsigned int l2_id = 0, l3_id = 0, num_threads_sharing, index_msb;
342
#ifdef CONFIG_X86_HT
343
	unsigned int cpu = c->cpu_index;
344
#endif
L
Linus Torvalds 已提交
345

346
	if (c->cpuid_level > 3) {
L
Linus Torvalds 已提交
347 348 349 350 351 352 353 354 355 356 357 358 359
		static int is_initialized;

		if (is_initialized == 0) {
			/* Init num_cache_leaves from boot CPU */
			num_cache_leaves = find_num_cache_leaves();
			is_initialized++;
		}

		/*
		 * Whenever possible use cpuid(4), deterministic cache
		 * parameters cpuid leaf to find the cache details
		 */
		for (i = 0; i < num_cache_leaves; i++) {
360
			struct _cpuid4_info_regs this_leaf;
L
Linus Torvalds 已提交
361 362
			int retval;

363
			retval = cpuid4_cache_lookup_regs(i, &this_leaf);
L
Linus Torvalds 已提交
364 365 366 367 368 369 370 371 372 373 374 375
			if (retval >= 0) {
				switch(this_leaf.eax.split.level) {
				    case 1:
					if (this_leaf.eax.split.type ==
							CACHE_TYPE_DATA)
						new_l1d = this_leaf.size/1024;
					else if (this_leaf.eax.split.type ==
							CACHE_TYPE_INST)
						new_l1i = this_leaf.size/1024;
					break;
				    case 2:
					new_l2 = this_leaf.size/1024;
376 377 378
					num_threads_sharing = 1 + this_leaf.eax.split.num_threads_sharing;
					index_msb = get_count_order(num_threads_sharing);
					l2_id = c->apicid >> index_msb;
L
Linus Torvalds 已提交
379 380 381
					break;
				    case 3:
					new_l3 = this_leaf.size/1024;
382 383 384
					num_threads_sharing = 1 + this_leaf.eax.split.num_threads_sharing;
					index_msb = get_count_order(num_threads_sharing);
					l3_id = c->apicid >> index_msb;
L
Linus Torvalds 已提交
385 386 387 388 389 390 391
					break;
				    default:
					break;
				}
			}
		}
	}
392 393 394 395 396
	/*
	 * Don't use cpuid2 if cpuid4 is supported. For P4, we use cpuid2 for
	 * trace cache
	 */
	if ((num_cache_leaves == 0 || c->x86 == 15) && c->cpuid_level > 1) {
L
Linus Torvalds 已提交
397
		/* supports eax=2  call */
398 399
		int j, n;
		unsigned int regs[4];
L
Linus Torvalds 已提交
400
		unsigned char *dp = (unsigned char *)regs;
401 402 403 404
		int only_trace = 0;

		if (num_cache_leaves != 0 && c->x86 == 15)
			only_trace = 1;
L
Linus Torvalds 已提交
405 406 407 408 409 410 411 412 413

		/* Number of times to iterate */
		n = cpuid_eax(2) & 0xFF;

		for ( i = 0 ; i < n ; i++ ) {
			cpuid(2, &regs[0], &regs[1], &regs[2], &regs[3]);

			/* If bit 31 is set, this is an unknown format */
			for ( j = 0 ; j < 3 ; j++ ) {
414
				if (regs[j] & (1 << 31)) regs[j] = 0;
L
Linus Torvalds 已提交
415 416 417 418 419 420 421 422 423 424 425
			}

			/* Byte 0 is level count, not a descriptor */
			for ( j = 1 ; j < 16 ; j++ ) {
				unsigned char des = dp[j];
				unsigned char k = 0;

				/* look up this descriptor in the table */
				while (cache_table[k].descriptor != 0)
				{
					if (cache_table[k].descriptor == des) {
426 427
						if (only_trace && cache_table[k].cache_type != LVL_TRACE)
							break;
L
Linus Torvalds 已提交
428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452
						switch (cache_table[k].cache_type) {
						case LVL_1_INST:
							l1i += cache_table[k].size;
							break;
						case LVL_1_DATA:
							l1d += cache_table[k].size;
							break;
						case LVL_2:
							l2 += cache_table[k].size;
							break;
						case LVL_3:
							l3 += cache_table[k].size;
							break;
						case LVL_TRACE:
							trace += cache_table[k].size;
							break;
						}

						break;
					}

					k++;
				}
			}
		}
453
	}
L
Linus Torvalds 已提交
454

455 456
	if (new_l1d)
		l1d = new_l1d;
L
Linus Torvalds 已提交
457

458 459
	if (new_l1i)
		l1i = new_l1i;
L
Linus Torvalds 已提交
460

461 462
	if (new_l2) {
		l2 = new_l2;
463
#ifdef CONFIG_X86_HT
464
		per_cpu(cpu_llc_id, cpu) = l2_id;
465
#endif
466
	}
L
Linus Torvalds 已提交
467

468 469
	if (new_l3) {
		l3 = new_l3;
470
#ifdef CONFIG_X86_HT
471
		per_cpu(cpu_llc_id, cpu) = l3_id;
472
#endif
L
Linus Torvalds 已提交
473 474
	}

475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492
	if (trace)
		printk (KERN_INFO "CPU: Trace cache: %dK uops", trace);
	else if ( l1i )
		printk (KERN_INFO "CPU: L1 I cache: %dK", l1i);

	if (l1d)
		printk(", L1 D cache: %dK\n", l1d);
	else
		printk("\n");

	if (l2)
		printk(KERN_INFO "CPU: L2 cache: %dK\n", l2);

	if (l3)
		printk(KERN_INFO "CPU: L3 cache: %dK\n", l3);

	c->x86_cache_size = l3 ? l3 : (l2 ? l2 : (l1i+l1d));

L
Linus Torvalds 已提交
493 494 495 496
	return l2;
}

/* pointer to _cpuid4_info array (for each cache leaf) */
497
static DEFINE_PER_CPU(struct _cpuid4_info *, cpuid4_info);
498
#define CPUID4_INFO_IDX(x, y)	(&((per_cpu(cpuid4_info, x))[y]))
L
Linus Torvalds 已提交
499 500

#ifdef CONFIG_SMP
501
static void __cpuinit cache_shared_cpu_map_setup(unsigned int cpu, int index)
L
Linus Torvalds 已提交
502
{
503
	struct _cpuid4_info	*this_leaf, *sibling_leaf;
L
Linus Torvalds 已提交
504
	unsigned long num_threads_sharing;
505
	int index_msb, i;
506
	struct cpuinfo_x86 *c = &cpu_data(cpu);
L
Linus Torvalds 已提交
507 508 509 510 511

	this_leaf = CPUID4_INFO_IDX(cpu, index);
	num_threads_sharing = 1 + this_leaf->eax.split.num_threads_sharing;

	if (num_threads_sharing == 1)
512
		cpumask_set_cpu(cpu, to_cpumask(this_leaf->shared_cpu_map));
513 514 515 516
	else {
		index_msb = get_count_order(num_threads_sharing);

		for_each_online_cpu(i) {
517 518
			if (cpu_data(i).apicid >> index_msb ==
			    c->apicid >> index_msb) {
519 520
				cpumask_set_cpu(i,
					to_cpumask(this_leaf->shared_cpu_map));
521
				if (i != cpu && per_cpu(cpuid4_info, i))  {
522 523 524 525
					sibling_leaf =
						CPUID4_INFO_IDX(i, index);
					cpumask_set_cpu(cpu, to_cpumask(
						sibling_leaf->shared_cpu_map));
526 527 528 529 530
				}
			}
		}
	}
}
531
static void __cpuinit cache_remove_shared_cpu_map(unsigned int cpu, int index)
532 533 534 535 536
{
	struct _cpuid4_info	*this_leaf, *sibling_leaf;
	int sibling;

	this_leaf = CPUID4_INFO_IDX(cpu, index);
537
	for_each_cpu(sibling, to_cpumask(this_leaf->shared_cpu_map)) {
538
		sibling_leaf = CPUID4_INFO_IDX(sibling, index);
539 540
		cpumask_clear_cpu(cpu,
				  to_cpumask(sibling_leaf->shared_cpu_map));
541
	}
L
Linus Torvalds 已提交
542 543
}
#else
544 545
static void __cpuinit cache_shared_cpu_map_setup(unsigned int cpu, int index) {}
static void __cpuinit cache_remove_shared_cpu_map(unsigned int cpu, int index) {}
L
Linus Torvalds 已提交
546 547
#endif

548
static void __cpuinit free_cache_attributes(unsigned int cpu)
L
Linus Torvalds 已提交
549
{
550 551 552 553 554
	int i;

	for (i = 0; i < num_cache_leaves; i++)
		cache_remove_shared_cpu_map(cpu, i);

555 556
	kfree(per_cpu(cpuid4_info, cpu));
	per_cpu(cpuid4_info, cpu) = NULL;
L
Linus Torvalds 已提交
557 558
}

559
static void __cpuinit get_cpu_leaves(void *_retval)
L
Linus Torvalds 已提交
560
{
561
	int j, *retval = _retval, cpu = smp_processor_id();
562

L
Linus Torvalds 已提交
563 564
	/* Do cpuid and store the results */
	for (j = 0; j < num_cache_leaves; j++) {
565
		struct _cpuid4_info *this_leaf;
L
Linus Torvalds 已提交
566
		this_leaf = CPUID4_INFO_IDX(cpu, j);
567 568
		*retval = cpuid4_cache_lookup(j, this_leaf);
		if (unlikely(*retval < 0)) {
569 570 571 572
			int i;

			for (i = 0; i < j; i++)
				cache_remove_shared_cpu_map(cpu, i);
573
			break;
574
		}
L
Linus Torvalds 已提交
575 576
		cache_shared_cpu_map_setup(cpu, j);
	}
577 578 579 580 581 582 583 584 585 586 587 588 589
}

static int __cpuinit detect_cache_attributes(unsigned int cpu)
{
	int			retval;

	if (num_cache_leaves == 0)
		return -ENOENT;

	per_cpu(cpuid4_info, cpu) = kzalloc(
	    sizeof(struct _cpuid4_info) * num_cache_leaves, GFP_KERNEL);
	if (per_cpu(cpuid4_info, cpu) == NULL)
		return -ENOMEM;
L
Linus Torvalds 已提交
590

591
	smp_call_function_single(cpu, get_cpu_leaves, &retval, true);
592
	if (retval) {
593 594
		kfree(per_cpu(cpuid4_info, cpu));
		per_cpu(cpuid4_info, cpu) = NULL;
595 596
	}

597
	return retval;
L
Linus Torvalds 已提交
598 599 600 601 602 603 604 605 606 607
}

#ifdef CONFIG_SYSFS

#include <linux/kobject.h>
#include <linux/sysfs.h>

extern struct sysdev_class cpu_sysdev_class; /* from drivers/base/cpu.c */

/* pointer to kobject for cpuX/cache */
608
static DEFINE_PER_CPU(struct kobject *, cache_kobject);
L
Linus Torvalds 已提交
609 610 611 612 613 614 615 616

struct _index_kobject {
	struct kobject kobj;
	unsigned int cpu;
	unsigned short index;
};

/* pointer to array of kobjects for cpuX/cache/indexY */
617
static DEFINE_PER_CPU(struct _index_kobject *, index_kobject);
618
#define INDEX_KOBJECT_PTR(x, y)		(&((per_cpu(index_kobject, x))[y]))
L
Linus Torvalds 已提交
619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637

#define show_one_plus(file_name, object, val)				\
static ssize_t show_##file_name						\
			(struct _cpuid4_info *this_leaf, char *buf)	\
{									\
	return sprintf (buf, "%lu\n", (unsigned long)this_leaf->object + val); \
}

show_one_plus(level, eax.split.level, 0);
show_one_plus(coherency_line_size, ebx.split.coherency_line_size, 1);
show_one_plus(physical_line_partition, ebx.split.physical_line_partition, 1);
show_one_plus(ways_of_associativity, ebx.split.ways_of_associativity, 1);
show_one_plus(number_of_sets, ecx.split.number_of_sets, 1);

static ssize_t show_size(struct _cpuid4_info *this_leaf, char *buf)
{
	return sprintf (buf, "%luK\n", this_leaf->size / 1024);
}

638 639
static ssize_t show_shared_cpu_map_func(struct _cpuid4_info *this_leaf,
					int type, char *buf)
L
Linus Torvalds 已提交
640
{
641
	ptrdiff_t len = PTR_ALIGN(buf + PAGE_SIZE - 1, PAGE_SIZE) - buf;
642 643
	int n = 0;

644
	if (len > 1) {
645
		const struct cpumask *mask;
646

647
		mask = to_cpumask(this_leaf->shared_cpu_map);
648
		n = type?
649 650
			cpulist_scnprintf(buf, len-2, mask) :
			cpumask_scnprintf(buf, len-2, mask);
651 652
		buf[n++] = '\n';
		buf[n] = '\0';
653 654
	}
	return n;
L
Linus Torvalds 已提交
655 656
}

657 658 659 660 661 662 663 664 665 666
static inline ssize_t show_shared_cpu_map(struct _cpuid4_info *leaf, char *buf)
{
	return show_shared_cpu_map_func(leaf, 0, buf);
}

static inline ssize_t show_shared_cpu_list(struct _cpuid4_info *leaf, char *buf)
{
	return show_shared_cpu_map_func(leaf, 1, buf);
}

667 668 669 670
static ssize_t show_type(struct _cpuid4_info *this_leaf, char *buf)
{
	switch (this_leaf->eax.split.type) {
	case CACHE_TYPE_DATA:
L
Linus Torvalds 已提交
671
		return sprintf(buf, "Data\n");
672
	case CACHE_TYPE_INST:
L
Linus Torvalds 已提交
673
		return sprintf(buf, "Instruction\n");
674
	case CACHE_TYPE_UNIFIED:
L
Linus Torvalds 已提交
675
		return sprintf(buf, "Unified\n");
676
	default:
L
Linus Torvalds 已提交
677 678 679 680
		return sprintf(buf, "Unknown\n");
	}
}

681 682
#define to_object(k)	container_of(k, struct _index_kobject, kobj)
#define to_attr(a)	container_of(a, struct _cache_attr, attr)
683

684
#ifdef CONFIG_PCI
685 686 687 688 689 690 691 692 693 694 695 696 697 698
static struct pci_dev *get_k8_northbridge(int node)
{
	struct pci_dev *dev = NULL;
	int i;

	for (i = 0; i <= node; i++) {
		do {
			dev = pci_get_device(PCI_ANY_ID, PCI_ANY_ID, dev);
			if (!dev)
				break;
		} while (!pci_match_id(&k8_nb_id[0], dev));
		if (!dev)
			break;
	}
699
	return dev;
700
}
701 702 703 704 705 706
#else
static struct pci_dev *get_k8_northbridge(int node)
{
	return NULL;
}
#endif
707

708 709
static ssize_t show_cache_disable(struct _cpuid4_info *this_leaf, char *buf)
{
710 711
	const struct cpumask *mask = to_cpumask(this_leaf->shared_cpu_map);
	int node = cpu_to_node(cpumask_first(mask));
712
	struct pci_dev *dev = NULL;
713 714
	ssize_t ret = 0;
	int i;
715

716 717 718
	if (!this_leaf->can_disable)
		return sprintf(buf, "Feature not enabled\n");

719 720 721 722 723 724
	dev = get_k8_northbridge(node);
	if (!dev) {
		printk(KERN_ERR "Attempting AMD northbridge operation on a system with no northbridge\n");
		return -EINVAL;
	}

725 726 727 728 729 730 731 732 733 734 735 736
	for (i = 0; i < 2; i++) {
		unsigned int reg;

		pci_read_config_dword(dev, 0x1BC + i * 4, &reg);

		ret += sprintf(buf, "%sEntry: %d\n", buf, i);
		ret += sprintf(buf, "%sReads:  %s\tNew Entries: %s\n",  
			buf,
			reg & 0x80000000 ? "Disabled" : "Allowed",
			reg & 0x40000000 ? "Disabled" : "Allowed");
		ret += sprintf(buf, "%sSubCache: %x\tIndex: %x\n",
			buf, (reg & 0x30000) >> 16, reg & 0xfff);
737
	}
738
	return ret;
739 740
}

741 742 743
static ssize_t
store_cache_disable(struct _cpuid4_info *this_leaf, const char *buf,
		    size_t count)
744
{
745 746
	const struct cpumask *mask = to_cpumask(this_leaf->shared_cpu_map);
	int node = cpu_to_node(cpumask_first(mask));
747
	struct pci_dev *dev = NULL;
748 749 750 751 752 753 754 755 756 757 758 759 760 761 762
	unsigned int ret, index, val;

	if (!this_leaf->can_disable)
		return 0;

	if (strlen(buf) > 15)
		return -EINVAL;

	ret = sscanf(buf, "%x %x", &index, &val);
	if (ret != 2)
		return -EINVAL;
	if (index > 1)
		return -EINVAL;

	val |= 0xc0000000;
763 764 765 766 767
	dev = get_k8_northbridge(node);
	if (!dev) {
		printk(KERN_ERR "Attempting AMD northbridge operation on a system with no northbridge\n");
		return -EINVAL;
	}
768

769 770 771 772 773
	pci_write_config_dword(dev, 0x1BC + index * 4, val & ~0x40000000);
	wbinvd();
	pci_write_config_dword(dev, 0x1BC + index * 4, val);

	return 1;
774 775
}

L
Linus Torvalds 已提交
776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793
struct _cache_attr {
	struct attribute attr;
	ssize_t (*show)(struct _cpuid4_info *, char *);
	ssize_t (*store)(struct _cpuid4_info *, const char *, size_t count);
};

#define define_one_ro(_name) \
static struct _cache_attr _name = \
	__ATTR(_name, 0444, show_##_name, NULL)

define_one_ro(level);
define_one_ro(type);
define_one_ro(coherency_line_size);
define_one_ro(physical_line_partition);
define_one_ro(ways_of_associativity);
define_one_ro(number_of_sets);
define_one_ro(size);
define_one_ro(shared_cpu_map);
794
define_one_ro(shared_cpu_list);
L
Linus Torvalds 已提交
795

796 797
static struct _cache_attr cache_disable = __ATTR(cache_disable, 0644, show_cache_disable, store_cache_disable);

L
Linus Torvalds 已提交
798 799 800 801 802 803 804 805 806
static struct attribute * default_attrs[] = {
	&type.attr,
	&level.attr,
	&coherency_line_size.attr,
	&physical_line_partition.attr,
	&ways_of_associativity.attr,
	&number_of_sets.attr,
	&size.attr,
	&shared_cpu_map.attr,
807
	&shared_cpu_list.attr,
808
	&cache_disable.attr,
L
Linus Torvalds 已提交
809 810 811 812 813 814 815 816 817 818 819 820
	NULL
};

static ssize_t show(struct kobject * kobj, struct attribute * attr, char * buf)
{
	struct _cache_attr *fattr = to_attr(attr);
	struct _index_kobject *this_leaf = to_object(kobj);
	ssize_t ret;

	ret = fattr->show ?
		fattr->show(CPUID4_INFO_IDX(this_leaf->cpu, this_leaf->index),
			buf) :
821
		0;
L
Linus Torvalds 已提交
822 823 824 825 826 827
	return ret;
}

static ssize_t store(struct kobject * kobj, struct attribute * attr,
		     const char * buf, size_t count)
{
828 829 830 831
	struct _cache_attr *fattr = to_attr(attr);
	struct _index_kobject *this_leaf = to_object(kobj);
	ssize_t ret;

832 833 834
	ret = fattr->store ?
		fattr->store(CPUID4_INFO_IDX(this_leaf->cpu, this_leaf->index),
			buf, count) :
835 836
		0;
	return ret;
L
Linus Torvalds 已提交
837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852
}

static struct sysfs_ops sysfs_ops = {
	.show   = show,
	.store  = store,
};

static struct kobj_type ktype_cache = {
	.sysfs_ops	= &sysfs_ops,
	.default_attrs	= default_attrs,
};

static struct kobj_type ktype_percpu_entry = {
	.sysfs_ops	= &sysfs_ops,
};

853
static void __cpuinit cpuid4_cache_sysfs_exit(unsigned int cpu)
L
Linus Torvalds 已提交
854
{
855 856 857 858
	kfree(per_cpu(cache_kobject, cpu));
	kfree(per_cpu(index_kobject, cpu));
	per_cpu(cache_kobject, cpu) = NULL;
	per_cpu(index_kobject, cpu) = NULL;
L
Linus Torvalds 已提交
859 860 861
	free_cache_attributes(cpu);
}

862
static int __cpuinit cpuid4_cache_sysfs_init(unsigned int cpu)
L
Linus Torvalds 已提交
863
{
864
	int err;
L
Linus Torvalds 已提交
865 866 867 868

	if (num_cache_leaves == 0)
		return -ENOENT;

869 870 871
	err = detect_cache_attributes(cpu);
	if (err)
		return err;
L
Linus Torvalds 已提交
872 873

	/* Allocate all required memory */
874 875 876
	per_cpu(cache_kobject, cpu) =
		kzalloc(sizeof(struct kobject), GFP_KERNEL);
	if (unlikely(per_cpu(cache_kobject, cpu) == NULL))
L
Linus Torvalds 已提交
877 878
		goto err_out;

879
	per_cpu(index_kobject, cpu) = kzalloc(
L
Linus Torvalds 已提交
880
	    sizeof(struct _index_kobject ) * num_cache_leaves, GFP_KERNEL);
881
	if (unlikely(per_cpu(index_kobject, cpu) == NULL))
L
Linus Torvalds 已提交
882 883 884 885 886 887 888 889 890
		goto err_out;

	return 0;

err_out:
	cpuid4_cache_sysfs_exit(cpu);
	return -ENOMEM;
}

891
static DECLARE_BITMAP(cache_dev_map, NR_CPUS);
892

L
Linus Torvalds 已提交
893
/* Add/Remove cache interface for CPU device */
894
static int __cpuinit cache_add_dev(struct sys_device * sys_dev)
L
Linus Torvalds 已提交
895 896 897 898
{
	unsigned int cpu = sys_dev->id;
	unsigned long i, j;
	struct _index_kobject *this_object;
899
	int retval;
L
Linus Torvalds 已提交
900 901 902 903 904

	retval = cpuid4_cache_sysfs_init(cpu);
	if (unlikely(retval < 0))
		return retval;

905 906
	retval = kobject_init_and_add(per_cpu(cache_kobject, cpu),
				      &ktype_percpu_entry,
907
				      &sys_dev->kobj, "%s", "cache");
908 909 910 911
	if (retval < 0) {
		cpuid4_cache_sysfs_exit(cpu);
		return retval;
	}
L
Linus Torvalds 已提交
912 913 914 915 916

	for (i = 0; i < num_cache_leaves; i++) {
		this_object = INDEX_KOBJECT_PTR(cpu,i);
		this_object->cpu = cpu;
		this_object->index = i;
917
		retval = kobject_init_and_add(&(this_object->kobj),
918 919
					      &ktype_cache,
					      per_cpu(cache_kobject, cpu),
920
					      "index%1lu", i);
L
Linus Torvalds 已提交
921 922
		if (unlikely(retval)) {
			for (j = 0; j < i; j++) {
923
				kobject_put(&(INDEX_KOBJECT_PTR(cpu,j)->kobj));
L
Linus Torvalds 已提交
924
			}
925
			kobject_put(per_cpu(cache_kobject, cpu));
L
Linus Torvalds 已提交
926
			cpuid4_cache_sysfs_exit(cpu);
927
			return retval;
L
Linus Torvalds 已提交
928
		}
929
		kobject_uevent(&(this_object->kobj), KOBJ_ADD);
L
Linus Torvalds 已提交
930
	}
931
	cpumask_set_cpu(cpu, to_cpumask(cache_dev_map));
932

933
	kobject_uevent(per_cpu(cache_kobject, cpu), KOBJ_ADD);
934
	return 0;
L
Linus Torvalds 已提交
935 936
}

937
static void __cpuinit cache_remove_dev(struct sys_device * sys_dev)
L
Linus Torvalds 已提交
938 939 940 941
{
	unsigned int cpu = sys_dev->id;
	unsigned long i;

942
	if (per_cpu(cpuid4_info, cpu) == NULL)
943
		return;
944
	if (!cpumask_test_cpu(cpu, to_cpumask(cache_dev_map)))
945
		return;
946
	cpumask_clear_cpu(cpu, to_cpumask(cache_dev_map));
947 948

	for (i = 0; i < num_cache_leaves; i++)
949
		kobject_put(&(INDEX_KOBJECT_PTR(cpu,i)->kobj));
950
	kobject_put(per_cpu(cache_kobject, cpu));
L
Linus Torvalds 已提交
951
	cpuid4_cache_sysfs_exit(cpu);
952 953
}

954
static int __cpuinit cacheinfo_cpu_callback(struct notifier_block *nfb,
955 956 957 958 959 960 961 962
					unsigned long action, void *hcpu)
{
	unsigned int cpu = (unsigned long)hcpu;
	struct sys_device *sys_dev;

	sys_dev = get_cpu_sysdev(cpu);
	switch (action) {
	case CPU_ONLINE:
963
	case CPU_ONLINE_FROZEN:
964 965 966
		cache_add_dev(sys_dev);
		break;
	case CPU_DEAD:
967
	case CPU_DEAD_FROZEN:
968 969 970 971
		cache_remove_dev(sys_dev);
		break;
	}
	return NOTIFY_OK;
L
Linus Torvalds 已提交
972 973
}

974
static struct notifier_block __cpuinitdata cacheinfo_cpu_notifier =
975
{
976
	.notifier_call = cacheinfo_cpu_callback,
L
Linus Torvalds 已提交
977 978
};

979
static int __cpuinit cache_sysfs_init(void)
L
Linus Torvalds 已提交
980
{
981 982
	int i;

L
Linus Torvalds 已提交
983 984 985
	if (num_cache_leaves == 0)
		return 0;

986
	for_each_online_cpu(i) {
987 988
		int err;
		struct sys_device *sys_dev = get_cpu_sysdev(i);
989

990 991 992
		err = cache_add_dev(sys_dev);
		if (err)
			return err;
993
	}
994
	register_hotcpu_notifier(&cacheinfo_cpu_notifier);
995
	return 0;
L
Linus Torvalds 已提交
996 997
}

998
device_initcall(cache_sysfs_init);
L
Linus Torvalds 已提交
999 1000

#endif