intel_cacheinfo.c 32.0 KB
Newer Older
L
Linus Torvalds 已提交
1
/*
2
 *	Routines to indentify caches on Intel CPU.
L
Linus Torvalds 已提交
3
 *
4 5
 *	Changes:
 *	Venkatesh Pallipadi	: Adding cache identification through cpuid(4)
A
Alan Cox 已提交
6
 *	Ashok Raj <ashok.raj@intel.com>: Work with CPU hotplug infrastructure.
7
 *	Andi Kleen / Andreas Herrmann	: CPUID4 emulation on AMD.
L
Linus Torvalds 已提交
8 9 10 11 12 13 14
 */

#include <linux/init.h>
#include <linux/slab.h>
#include <linux/device.h>
#include <linux/compiler.h>
#include <linux/cpu.h>
T
Tim Schmielau 已提交
15
#include <linux/sched.h>
16
#include <linux/pci.h>
L
Linus Torvalds 已提交
17 18

#include <asm/processor.h>
A
Alan Cox 已提交
19
#include <linux/smp.h>
20
#include <asm/amd_nb.h>
21
#include <asm/smp.h>
L
Linus Torvalds 已提交
22 23 24 25 26 27 28

#define LVL_1_INST	1
#define LVL_1_DATA	2
#define LVL_2		3
#define LVL_3		4
#define LVL_TRACE	5

A
Alan Cox 已提交
29
struct _cache_table {
L
Linus Torvalds 已提交
30 31 32 33 34
	unsigned char descriptor;
	char cache_type;
	short size;
};

D
Dave Jones 已提交
35 36
#define MB(x)	((x) * 1024)

A
Alan Cox 已提交
37 38 39
/* All the cache descriptor types we care about (no TLB or
   trace cache entries) */

40
static const struct _cache_table __cpuinitconst cache_table[] =
L
Linus Torvalds 已提交
41 42 43
{
	{ 0x06, LVL_1_INST, 8 },	/* 4-way set assoc, 32 byte line size */
	{ 0x08, LVL_1_INST, 16 },	/* 4-way set assoc, 32 byte line size */
44
	{ 0x09, LVL_1_INST, 32 },	/* 4-way set assoc, 64 byte line size */
L
Linus Torvalds 已提交
45 46
	{ 0x0a, LVL_1_DATA, 8 },	/* 2 way set assoc, 32 byte line size */
	{ 0x0c, LVL_1_DATA, 16 },	/* 4-way set assoc, 32 byte line size */
47
	{ 0x0d, LVL_1_DATA, 16 },	/* 4-way set assoc, 64 byte line size */
48
	{ 0x0e, LVL_1_DATA, 24 },	/* 6-way set assoc, 64 byte line size */
49
	{ 0x21, LVL_2,      256 },	/* 8-way set assoc, 64 byte line size */
L
Linus Torvalds 已提交
50
	{ 0x22, LVL_3,      512 },	/* 4-way set assoc, sectored cache, 64 byte line size */
D
Dave Jones 已提交
51 52 53
	{ 0x23, LVL_3,      MB(1) },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x25, LVL_3,      MB(2) },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x29, LVL_3,      MB(4) },	/* 8-way set assoc, sectored cache, 64 byte line size */
L
Linus Torvalds 已提交
54 55 56
	{ 0x2c, LVL_1_DATA, 32 },	/* 8-way set assoc, 64 byte line size */
	{ 0x30, LVL_1_INST, 32 },	/* 8-way set assoc, 64 byte line size */
	{ 0x39, LVL_2,      128 },	/* 4-way set assoc, sectored cache, 64 byte line size */
57
	{ 0x3a, LVL_2,      192 },	/* 6-way set assoc, sectored cache, 64 byte line size */
L
Linus Torvalds 已提交
58 59
	{ 0x3b, LVL_2,      128 },	/* 2-way set assoc, sectored cache, 64 byte line size */
	{ 0x3c, LVL_2,      256 },	/* 4-way set assoc, sectored cache, 64 byte line size */
60 61
	{ 0x3d, LVL_2,      384 },	/* 6-way set assoc, sectored cache, 64 byte line size */
	{ 0x3e, LVL_2,      512 },	/* 4-way set assoc, sectored cache, 64 byte line size */
62
	{ 0x3f, LVL_2,      256 },	/* 2-way set assoc, 64 byte line size */
L
Linus Torvalds 已提交
63 64 65
	{ 0x41, LVL_2,      128 },	/* 4-way set assoc, 32 byte line size */
	{ 0x42, LVL_2,      256 },	/* 4-way set assoc, 32 byte line size */
	{ 0x43, LVL_2,      512 },	/* 4-way set assoc, 32 byte line size */
D
Dave Jones 已提交
66 67 68 69
	{ 0x44, LVL_2,      MB(1) },	/* 4-way set assoc, 32 byte line size */
	{ 0x45, LVL_2,      MB(2) },	/* 4-way set assoc, 32 byte line size */
	{ 0x46, LVL_3,      MB(4) },	/* 4-way set assoc, 64 byte line size */
	{ 0x47, LVL_3,      MB(8) },	/* 8-way set assoc, 64 byte line size */
70
	{ 0x48, LVL_2,      MB(3) },	/* 12-way set assoc, 64 byte line size */
D
Dave Jones 已提交
71 72 73 74 75 76
	{ 0x49, LVL_3,      MB(4) },	/* 16-way set assoc, 64 byte line size */
	{ 0x4a, LVL_3,      MB(6) },	/* 12-way set assoc, 64 byte line size */
	{ 0x4b, LVL_3,      MB(8) },	/* 16-way set assoc, 64 byte line size */
	{ 0x4c, LVL_3,      MB(12) },	/* 12-way set assoc, 64 byte line size */
	{ 0x4d, LVL_3,      MB(16) },	/* 16-way set assoc, 64 byte line size */
	{ 0x4e, LVL_2,      MB(6) },	/* 24-way set assoc, 64 byte line size */
L
Linus Torvalds 已提交
77 78 79 80 81 82 83
	{ 0x60, LVL_1_DATA, 16 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x66, LVL_1_DATA, 8 },	/* 4-way set assoc, sectored cache, 64 byte line size */
	{ 0x67, LVL_1_DATA, 16 },	/* 4-way set assoc, sectored cache, 64 byte line size */
	{ 0x68, LVL_1_DATA, 32 },	/* 4-way set assoc, sectored cache, 64 byte line size */
	{ 0x70, LVL_TRACE,  12 },	/* 8-way set assoc */
	{ 0x71, LVL_TRACE,  16 },	/* 8-way set assoc */
	{ 0x72, LVL_TRACE,  32 },	/* 8-way set assoc */
84
	{ 0x73, LVL_TRACE,  64 },	/* 8-way set assoc */
D
Dave Jones 已提交
85 86 87 88 89 90 91
	{ 0x78, LVL_2,      MB(1) },	/* 4-way set assoc, 64 byte line size */
	{ 0x79, LVL_2,      128 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x7a, LVL_2,      256 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x7b, LVL_2,      512 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x7c, LVL_2,      MB(1) },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x7d, LVL_2,      MB(2) },	/* 8-way set assoc, 64 byte line size */
	{ 0x7f, LVL_2,      512 },	/* 2-way set assoc, 64 byte line size */
92
	{ 0x80, LVL_2,      512 },	/* 8-way set assoc, 64 byte line size */
D
Dave Jones 已提交
93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113
	{ 0x82, LVL_2,      256 },	/* 8-way set assoc, 32 byte line size */
	{ 0x83, LVL_2,      512 },	/* 8-way set assoc, 32 byte line size */
	{ 0x84, LVL_2,      MB(1) },	/* 8-way set assoc, 32 byte line size */
	{ 0x85, LVL_2,      MB(2) },	/* 8-way set assoc, 32 byte line size */
	{ 0x86, LVL_2,      512 },	/* 4-way set assoc, 64 byte line size */
	{ 0x87, LVL_2,      MB(1) },	/* 8-way set assoc, 64 byte line size */
	{ 0xd0, LVL_3,      512 },	/* 4-way set assoc, 64 byte line size */
	{ 0xd1, LVL_3,      MB(1) },	/* 4-way set assoc, 64 byte line size */
	{ 0xd2, LVL_3,      MB(2) },	/* 4-way set assoc, 64 byte line size */
	{ 0xd6, LVL_3,      MB(1) },	/* 8-way set assoc, 64 byte line size */
	{ 0xd7, LVL_3,      MB(2) },	/* 8-way set assoc, 64 byte line size */
	{ 0xd8, LVL_3,      MB(4) },	/* 12-way set assoc, 64 byte line size */
	{ 0xdc, LVL_3,      MB(2) },	/* 12-way set assoc, 64 byte line size */
	{ 0xdd, LVL_3,      MB(4) },	/* 12-way set assoc, 64 byte line size */
	{ 0xde, LVL_3,      MB(8) },	/* 12-way set assoc, 64 byte line size */
	{ 0xe2, LVL_3,      MB(2) },	/* 16-way set assoc, 64 byte line size */
	{ 0xe3, LVL_3,      MB(4) },	/* 16-way set assoc, 64 byte line size */
	{ 0xe4, LVL_3,      MB(8) },	/* 16-way set assoc, 64 byte line size */
	{ 0xea, LVL_3,      MB(12) },	/* 24-way set assoc, 64 byte line size */
	{ 0xeb, LVL_3,      MB(18) },	/* 24-way set assoc, 64 byte line size */
	{ 0xec, LVL_3,      MB(24) },	/* 24-way set assoc, 64 byte line size */
L
Linus Torvalds 已提交
114 115 116 117
	{ 0x00, 0, 0}
};


A
Alan Cox 已提交
118
enum _cache_type {
L
Linus Torvalds 已提交
119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153
	CACHE_TYPE_NULL	= 0,
	CACHE_TYPE_DATA = 1,
	CACHE_TYPE_INST = 2,
	CACHE_TYPE_UNIFIED = 3
};

union _cpuid4_leaf_eax {
	struct {
		enum _cache_type	type:5;
		unsigned int		level:3;
		unsigned int		is_self_initializing:1;
		unsigned int		is_fully_associative:1;
		unsigned int		reserved:4;
		unsigned int		num_threads_sharing:12;
		unsigned int		num_cores_on_die:6;
	} split;
	u32 full;
};

union _cpuid4_leaf_ebx {
	struct {
		unsigned int		coherency_line_size:12;
		unsigned int		physical_line_partition:10;
		unsigned int		ways_of_associativity:10;
	} split;
	u32 full;
};

union _cpuid4_leaf_ecx {
	struct {
		unsigned int		number_of_sets:32;
	} split;
	u32 full;
};

154
struct _cpuid4_info_regs {
L
Linus Torvalds 已提交
155 156 157 158
	union _cpuid4_leaf_eax eax;
	union _cpuid4_leaf_ebx ebx;
	union _cpuid4_leaf_ecx ecx;
	unsigned long size;
159
	struct amd_northbridge *nb;
160 161
};

162 163 164
struct _cpuid4_info {
	struct _cpuid4_info_regs base;
	DECLARE_BITMAP(shared_cpu_map, NR_CPUS);
L
Linus Torvalds 已提交
165 166
};

167 168 169 170
unsigned short			num_cache_leaves;

/* AMD doesn't have CPUID4. Emulate it here to report the same
   information to the user.  This makes some assumptions about the machine:
171
   L2 not shared, no SMT etc. that is currently true on AMD CPUs.
172 173 174 175 176

   In theory the TLBs could be reported as fake type (they are in "dummy").
   Maybe later */
union l1_cache {
	struct {
A
Alan Cox 已提交
177 178 179 180
		unsigned line_size:8;
		unsigned lines_per_tag:8;
		unsigned assoc:8;
		unsigned size_in_kb:8;
181 182 183 184 185 186
	};
	unsigned val;
};

union l2_cache {
	struct {
A
Alan Cox 已提交
187 188 189 190
		unsigned line_size:8;
		unsigned lines_per_tag:4;
		unsigned assoc:4;
		unsigned size_in_kb:16;
191 192 193 194
	};
	unsigned val;
};

195 196
union l3_cache {
	struct {
A
Alan Cox 已提交
197 198 199 200 201
		unsigned line_size:8;
		unsigned lines_per_tag:4;
		unsigned assoc:4;
		unsigned res:2;
		unsigned size_encoded:14;
202 203 204 205
	};
	unsigned val;
};

206
static const unsigned short __cpuinitconst assocs[] = {
207 208 209 210 211 212 213
	[1] = 1,
	[2] = 2,
	[4] = 4,
	[6] = 8,
	[8] = 16,
	[0xa] = 32,
	[0xb] = 48,
214
	[0xc] = 64,
215 216 217
	[0xd] = 96,
	[0xe] = 128,
	[0xf] = 0xffff /* fully associative - no way to show this currently */
218 219
};

220 221
static const unsigned char __cpuinitconst levels[] = { 1, 1, 2, 3 };
static const unsigned char __cpuinitconst types[] = { 1, 2, 3, 3 };
222

223 224 225 226
static void __cpuinit
amd_cpuid4(int leaf, union _cpuid4_leaf_eax *eax,
		     union _cpuid4_leaf_ebx *ebx,
		     union _cpuid4_leaf_ecx *ecx)
227 228 229 230 231
{
	unsigned dummy;
	unsigned line_size, lines_per_tag, assoc, size_in_kb;
	union l1_cache l1i, l1d;
	union l2_cache l2;
232 233
	union l3_cache l3;
	union l1_cache *l1 = &l1d;
234 235 236 237 238 239

	eax->full = 0;
	ebx->full = 0;
	ecx->full = 0;

	cpuid(0x80000005, &dummy, &dummy, &l1d.val, &l1i.val);
240
	cpuid(0x80000006, &dummy, &dummy, &l2.val, &l3.val);
241

242 243 244 245 246 247
	switch (leaf) {
	case 1:
		l1 = &l1i;
	case 0:
		if (!l1->val)
			return;
248
		assoc = assocs[l1->assoc];
249 250 251
		line_size = l1->line_size;
		lines_per_tag = l1->lines_per_tag;
		size_in_kb = l1->size_in_kb;
252 253 254 255
		break;
	case 2:
		if (!l2.val)
			return;
256
		assoc = assocs[l2.assoc];
257 258 259
		line_size = l2.line_size;
		lines_per_tag = l2.lines_per_tag;
		/* cpu_data has errata corrections for K7 applied */
260
		size_in_kb = __this_cpu_read(cpu_info.x86_cache_size);
261 262 263 264
		break;
	case 3:
		if (!l3.val)
			return;
265
		assoc = assocs[l3.assoc];
266 267 268
		line_size = l3.line_size;
		lines_per_tag = l3.lines_per_tag;
		size_in_kb = l3.size_encoded * 512;
269 270 271 272
		if (boot_cpu_has(X86_FEATURE_AMD_DCM)) {
			size_in_kb = size_in_kb >> 1;
			assoc = assoc >> 1;
		}
273 274 275
		break;
	default:
		return;
276 277
	}

278 279 280
	eax->split.is_self_initializing = 1;
	eax->split.type = types[leaf];
	eax->split.level = levels[leaf];
281
	eax->split.num_threads_sharing = 0;
282
	eax->split.num_cores_on_die = __this_cpu_read(cpu_info.x86_max_cores) - 1;
283 284


285
	if (assoc == 0xffff)
286 287
		eax->split.is_fully_associative = 1;
	ebx->split.coherency_line_size = line_size - 1;
288
	ebx->split.ways_of_associativity = assoc - 1;
289 290 291 292
	ebx->split.physical_line_partition = lines_per_tag - 1;
	ecx->split.number_of_sets = (size_in_kb * 1024) / line_size /
		(ebx->split.ways_of_associativity + 1) - 1;
}
L
Linus Torvalds 已提交
293

294 295
struct _cache_attr {
	struct attribute attr;
296 297 298
	ssize_t (*show)(struct _cpuid4_info *, char *, unsigned int);
	ssize_t (*store)(struct _cpuid4_info *, const char *, size_t count,
			 unsigned int);
299 300
};

301
#ifdef CONFIG_AMD_NB
302 303 304 305

/*
 * L3 cache descriptors
 */
306
static void __cpuinit amd_calc_l3_indices(struct amd_northbridge *nb)
307
{
308
	struct amd_l3_cache *l3 = &nb->l3_cache;
309
	unsigned int sc0, sc1, sc2, sc3;
310
	u32 val = 0;
311

312
	pci_read_config_dword(nb->misc, 0x1C4, &val);
313 314

	/* calculate subcache sizes */
315 316
	l3->subcaches[0] = sc0 = !(val & BIT(0));
	l3->subcaches[1] = sc1 = !(val & BIT(4));
317 318 319 320 321 322

	if (boot_cpu_data.x86 == 0x15) {
		l3->subcaches[0] = sc0 += !(val & BIT(1));
		l3->subcaches[1] = sc1 += !(val & BIT(5));
	}

323 324 325
	l3->subcaches[2] = sc2 = !(val & BIT(8))  + !(val & BIT(9));
	l3->subcaches[3] = sc3 = !(val & BIT(12)) + !(val & BIT(13));

326
	l3->indices = (max(max3(sc0, sc1, sc2), sc3) << 10) - 1;
327 328
}

329
static void __cpuinit amd_init_l3_cache(struct _cpuid4_info_regs *this_leaf, int index)
330
{
331 332
	int node;

333
	/* only for L3, and not in virtualized environments */
334
	if (index < 3)
335 336
		return;

337
	node = amd_get_nb_id(smp_processor_id());
338 339 340
	this_leaf->nb = node_to_amd_nb(node);
	if (this_leaf->nb && !this_leaf->nb->l3_cache.indices)
		amd_calc_l3_indices(this_leaf->nb);
341 342
}

343 344 345 346 347 348 349
/*
 * check whether a slot used for disabling an L3 index is occupied.
 * @l3: L3 cache descriptor
 * @slot: slot number (0..1)
 *
 * @returns: the disabled index if used or negative value if slot free.
 */
350
int amd_get_l3_disable_slot(struct amd_northbridge *nb, unsigned slot)
351 352 353
{
	unsigned int reg = 0;

354
	pci_read_config_dword(nb->misc, 0x1BC + slot * 4, &reg);
355 356 357 358 359 360 361 362

	/* check whether this slot is activated already */
	if (reg & (3UL << 30))
		return reg & 0xfff;

	return -1;
}

363
static ssize_t show_cache_disable(struct _cpuid4_info *this_leaf, char *buf,
364
				  unsigned int slot)
365
{
366
	int index;
367

368
	if (!this_leaf->base.nb || !amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE))
369 370
		return -EINVAL;

371
	index = amd_get_l3_disable_slot(this_leaf->base.nb, slot);
372 373
	if (index >= 0)
		return sprintf(buf, "%d\n", index);
374

375
	return sprintf(buf, "FREE\n");
376 377
}

378
#define SHOW_CACHE_DISABLE(slot)					\
379
static ssize_t								\
380 381
show_cache_disable_##slot(struct _cpuid4_info *this_leaf, char *buf,	\
			  unsigned int cpu)				\
382
{									\
383
	return show_cache_disable(this_leaf, buf, slot);		\
384 385 386 387
}
SHOW_CACHE_DISABLE(0)
SHOW_CACHE_DISABLE(1)

388
static void amd_l3_disable_index(struct amd_northbridge *nb, int cpu,
389 390 391 392 393 394 395 396 397 398 399 400
				 unsigned slot, unsigned long idx)
{
	int i;

	idx |= BIT(30);

	/*
	 *  disable index in all 4 subcaches
	 */
	for (i = 0; i < 4; i++) {
		u32 reg = idx | (i << 20);

401
		if (!nb->l3_cache.subcaches[i])
402 403
			continue;

404
		pci_write_config_dword(nb->misc, 0x1BC + slot * 4, reg);
405 406 407 408 409 410 411 412 413

		/*
		 * We need to WBINVD on a core on the node containing the L3
		 * cache which indices we disable therefore a simple wbinvd()
		 * is not sufficient.
		 */
		wbinvd_on_cpu(cpu);

		reg |= BIT(31);
414
		pci_write_config_dword(nb->misc, 0x1BC + slot * 4, reg);
415 416 417
	}
}

418 419 420 421 422 423 424 425 426 427
/*
 * disable a L3 cache index by using a disable-slot
 *
 * @l3:    L3 cache descriptor
 * @cpu:   A CPU on the node containing the L3 cache
 * @slot:  slot number (0..1)
 * @index: index to disable
 *
 * @return: 0 on success, error status on failure
 */
428
int amd_set_l3_disable_slot(struct amd_northbridge *nb, int cpu, unsigned slot,
429
			    unsigned long index)
430
{
431
	int ret = 0;
432

433
	/*  check if @slot is already used or the index is already disabled */
434
	ret = amd_get_l3_disable_slot(nb, slot);
435
	if (ret >= 0)
436
		return -EEXIST;
437

438
	if (index > nb->l3_cache.indices)
439 440
		return -EINVAL;

441
	/* check whether the other slot has disabled the same index already */
442
	if (index == amd_get_l3_disable_slot(nb, !slot))
443
		return -EEXIST;
444

445
	amd_l3_disable_index(nb, cpu, slot, index);
446 447 448 449 450 451 452 453 454 455 456

	return 0;
}

static ssize_t store_cache_disable(struct _cpuid4_info *this_leaf,
				  const char *buf, size_t count,
				  unsigned int slot)
{
	unsigned long val = 0;
	int cpu, err = 0;

457 458 459
	if (!capable(CAP_SYS_ADMIN))
		return -EPERM;

460
	if (!this_leaf->base.nb || !amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE))
461 462
		return -EINVAL;

463
	cpu = cpumask_first(to_cpumask(this_leaf->shared_cpu_map));
464

465
	if (strict_strtoul(buf, 10, &val) < 0)
466 467
		return -EINVAL;

468
	err = amd_set_l3_disable_slot(this_leaf->base.nb, cpu, slot, val);
469 470
	if (err) {
		if (err == -EEXIST)
471 472
			pr_warning("L3 slot %d in use/index already disabled!\n",
				   slot);
473 474
		return err;
	}
475 476 477
	return count;
}

478
#define STORE_CACHE_DISABLE(slot)					\
479
static ssize_t								\
480
store_cache_disable_##slot(struct _cpuid4_info *this_leaf,		\
481 482
			   const char *buf, size_t count,		\
			   unsigned int cpu)				\
483
{									\
484
	return store_cache_disable(this_leaf, buf, count, slot);	\
485
}
486 487 488 489 490 491 492 493
STORE_CACHE_DISABLE(0)
STORE_CACHE_DISABLE(1)

static struct _cache_attr cache_disable_0 = __ATTR(cache_disable_0, 0644,
		show_cache_disable_0, store_cache_disable_0);
static struct _cache_attr cache_disable_1 = __ATTR(cache_disable_1, 0644,
		show_cache_disable_1, store_cache_disable_1);

494 495 496
static ssize_t
show_subcaches(struct _cpuid4_info *this_leaf, char *buf, unsigned int cpu)
{
497
	if (!this_leaf->base.nb || !amd_nb_has_feature(AMD_NB_L3_PARTITIONING))
498 499 500 501 502 503 504 505 506 507 508 509 510 511
		return -EINVAL;

	return sprintf(buf, "%x\n", amd_get_subcaches(cpu));
}

static ssize_t
store_subcaches(struct _cpuid4_info *this_leaf, const char *buf, size_t count,
		unsigned int cpu)
{
	unsigned long val;

	if (!capable(CAP_SYS_ADMIN))
		return -EPERM;

512
	if (!this_leaf->base.nb || !amd_nb_has_feature(AMD_NB_L3_PARTITIONING))
513 514 515 516 517 518 519 520 521 522 523 524 525 526
		return -EINVAL;

	if (strict_strtoul(buf, 16, &val) < 0)
		return -EINVAL;

	if (amd_set_subcaches(cpu, val))
		return -EINVAL;

	return count;
}

static struct _cache_attr subcaches =
	__ATTR(subcaches, 0644, show_subcaches, store_subcaches);

527
#else	/* CONFIG_AMD_NB */
528
#define amd_init_l3_cache(x, y)
529
#endif /* CONFIG_AMD_NB */
530

531
static int
532 533
__cpuinit cpuid4_cache_lookup_regs(int index,
				   struct _cpuid4_info_regs *this_leaf)
L
Linus Torvalds 已提交
534
{
535 536 537
	union _cpuid4_leaf_eax	eax;
	union _cpuid4_leaf_ebx	ebx;
	union _cpuid4_leaf_ecx	ecx;
538
	unsigned		edx;
L
Linus Torvalds 已提交
539

540
	if (boot_cpu_data.x86_vendor == X86_VENDOR_AMD) {
541
		amd_cpuid4(index, &eax, &ebx, &ecx);
542
		amd_init_l3_cache(this_leaf, index);
543 544 545 546
	} else {
		cpuid_count(4, index, &eax.full, &ebx.full, &ecx.full, &edx);
	}

547
	if (eax.split.type == CACHE_TYPE_NULL)
548
		return -EIO; /* better error ? */
L
Linus Torvalds 已提交
549

550 551 552
	this_leaf->eax = eax;
	this_leaf->ebx = ebx;
	this_leaf->ecx = ecx;
553 554 555 556
	this_leaf->size = (ecx.split.number_of_sets          + 1) *
			  (ebx.split.coherency_line_size     + 1) *
			  (ebx.split.physical_line_partition + 1) *
			  (ebx.split.ways_of_associativity   + 1);
L
Linus Torvalds 已提交
557 558 559
	return 0;
}

560
static int __cpuinit find_num_cache_leaves(void)
L
Linus Torvalds 已提交
561 562 563
{
	unsigned int		eax, ebx, ecx, edx;
	union _cpuid4_leaf_eax	cache_eax;
564
	int 			i = -1;
L
Linus Torvalds 已提交
565

566 567 568
	do {
		++i;
		/* Do cpuid(4) loop to find out num_cache_leaves */
L
Linus Torvalds 已提交
569 570
		cpuid_count(4, i, &eax, &ebx, &ecx, &edx);
		cache_eax.full = eax;
571 572
	} while (cache_eax.split.type != CACHE_TYPE_NULL);
	return i;
L
Linus Torvalds 已提交
573 574
}

575
unsigned int __cpuinit init_intel_cacheinfo(struct cpuinfo_x86 *c)
L
Linus Torvalds 已提交
576
{
A
Alan Cox 已提交
577 578
	/* Cache sizes */
	unsigned int trace = 0, l1i = 0, l1d = 0, l2 = 0, l3 = 0;
L
Linus Torvalds 已提交
579 580
	unsigned int new_l1d = 0, new_l1i = 0; /* Cache sizes from cpuid(4) */
	unsigned int new_l2 = 0, new_l3 = 0, i; /* Cache sizes from cpuid(4) */
581
	unsigned int l2_id = 0, l3_id = 0, num_threads_sharing, index_msb;
582
#ifdef CONFIG_X86_HT
583
	unsigned int cpu = c->cpu_index;
584
#endif
L
Linus Torvalds 已提交
585

586
	if (c->cpuid_level > 3) {
L
Linus Torvalds 已提交
587 588 589 590 591 592 593 594 595 596 597 598 599
		static int is_initialized;

		if (is_initialized == 0) {
			/* Init num_cache_leaves from boot CPU */
			num_cache_leaves = find_num_cache_leaves();
			is_initialized++;
		}

		/*
		 * Whenever possible use cpuid(4), deterministic cache
		 * parameters cpuid leaf to find the cache details
		 */
		for (i = 0; i < num_cache_leaves; i++) {
600
			struct _cpuid4_info_regs this_leaf;
L
Linus Torvalds 已提交
601 602
			int retval;

603
			retval = cpuid4_cache_lookup_regs(i, &this_leaf);
L
Linus Torvalds 已提交
604
			if (retval >= 0) {
A
Alan Cox 已提交
605 606
				switch (this_leaf.eax.split.level) {
				case 1:
L
Linus Torvalds 已提交
607 608 609 610 611 612 613
					if (this_leaf.eax.split.type ==
							CACHE_TYPE_DATA)
						new_l1d = this_leaf.size/1024;
					else if (this_leaf.eax.split.type ==
							CACHE_TYPE_INST)
						new_l1i = this_leaf.size/1024;
					break;
A
Alan Cox 已提交
614
				case 2:
L
Linus Torvalds 已提交
615
					new_l2 = this_leaf.size/1024;
616 617 618
					num_threads_sharing = 1 + this_leaf.eax.split.num_threads_sharing;
					index_msb = get_count_order(num_threads_sharing);
					l2_id = c->apicid >> index_msb;
L
Linus Torvalds 已提交
619
					break;
A
Alan Cox 已提交
620
				case 3:
L
Linus Torvalds 已提交
621
					new_l3 = this_leaf.size/1024;
622
					num_threads_sharing = 1 + this_leaf.eax.split.num_threads_sharing;
A
Alan Cox 已提交
623 624
					index_msb = get_count_order(
							num_threads_sharing);
625
					l3_id = c->apicid >> index_msb;
L
Linus Torvalds 已提交
626
					break;
A
Alan Cox 已提交
627
				default:
L
Linus Torvalds 已提交
628 629 630 631 632
					break;
				}
			}
		}
	}
633 634 635 636 637
	/*
	 * Don't use cpuid2 if cpuid4 is supported. For P4, we use cpuid2 for
	 * trace cache
	 */
	if ((num_cache_leaves == 0 || c->x86 == 15) && c->cpuid_level > 1) {
L
Linus Torvalds 已提交
638
		/* supports eax=2  call */
639 640
		int j, n;
		unsigned int regs[4];
L
Linus Torvalds 已提交
641
		unsigned char *dp = (unsigned char *)regs;
642 643 644 645
		int only_trace = 0;

		if (num_cache_leaves != 0 && c->x86 == 15)
			only_trace = 1;
L
Linus Torvalds 已提交
646 647 648 649

		/* Number of times to iterate */
		n = cpuid_eax(2) & 0xFF;

A
Alan Cox 已提交
650
		for (i = 0 ; i < n ; i++) {
L
Linus Torvalds 已提交
651 652 653
			cpuid(2, &regs[0], &regs[1], &regs[2], &regs[3]);

			/* If bit 31 is set, this is an unknown format */
A
Alan Cox 已提交
654 655 656
			for (j = 0 ; j < 3 ; j++)
				if (regs[j] & (1 << 31))
					regs[j] = 0;
L
Linus Torvalds 已提交
657 658

			/* Byte 0 is level count, not a descriptor */
A
Alan Cox 已提交
659
			for (j = 1 ; j < 16 ; j++) {
L
Linus Torvalds 已提交
660 661 662 663
				unsigned char des = dp[j];
				unsigned char k = 0;

				/* look up this descriptor in the table */
A
Alan Cox 已提交
664
				while (cache_table[k].descriptor != 0) {
L
Linus Torvalds 已提交
665
					if (cache_table[k].descriptor == des) {
666 667
						if (only_trace && cache_table[k].cache_type != LVL_TRACE)
							break;
L
Linus Torvalds 已提交
668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692
						switch (cache_table[k].cache_type) {
						case LVL_1_INST:
							l1i += cache_table[k].size;
							break;
						case LVL_1_DATA:
							l1d += cache_table[k].size;
							break;
						case LVL_2:
							l2 += cache_table[k].size;
							break;
						case LVL_3:
							l3 += cache_table[k].size;
							break;
						case LVL_TRACE:
							trace += cache_table[k].size;
							break;
						}

						break;
					}

					k++;
				}
			}
		}
693
	}
L
Linus Torvalds 已提交
694

695 696
	if (new_l1d)
		l1d = new_l1d;
L
Linus Torvalds 已提交
697

698 699
	if (new_l1i)
		l1i = new_l1i;
L
Linus Torvalds 已提交
700

701 702
	if (new_l2) {
		l2 = new_l2;
703
#ifdef CONFIG_X86_HT
704
		per_cpu(cpu_llc_id, cpu) = l2_id;
705
#endif
706
	}
L
Linus Torvalds 已提交
707

708 709
	if (new_l3) {
		l3 = new_l3;
710
#ifdef CONFIG_X86_HT
711
		per_cpu(cpu_llc_id, cpu) = l3_id;
712
#endif
L
Linus Torvalds 已提交
713 714
	}

715 716
	c->x86_cache_size = l3 ? l3 : (l2 ? l2 : (l1i+l1d));

L
Linus Torvalds 已提交
717 718 719
	return l2;
}

720 721
#ifdef CONFIG_SYSFS

L
Linus Torvalds 已提交
722
/* pointer to _cpuid4_info array (for each cache leaf) */
723 724
static DEFINE_PER_CPU(struct _cpuid4_info *, ici_cpuid4_info);
#define CPUID4_INFO_IDX(x, y)	(&((per_cpu(ici_cpuid4_info, x))[y]))
L
Linus Torvalds 已提交
725 726

#ifdef CONFIG_SMP
727 728

static int __cpuinit cache_shared_amd_cpu_map_setup(unsigned int cpu, int index)
L
Linus Torvalds 已提交
729
{
730 731
	struct _cpuid4_info *this_leaf;
	int ret, i, sibling;
732
	struct cpuinfo_x86 *c = &cpu_data(cpu);
L
Linus Torvalds 已提交
733

734 735 736
	ret = 0;
	if (index == 3) {
		ret = 1;
737
		for_each_cpu(i, cpu_llc_shared_mask(cpu)) {
738
			if (!per_cpu(ici_cpuid4_info, i))
739 740
				continue;
			this_leaf = CPUID4_INFO_IDX(i, index);
741
			for_each_cpu(sibling, cpu_llc_shared_mask(cpu)) {
742 743 744 745
				if (!cpu_online(sibling))
					continue;
				set_bit(sibling, this_leaf->shared_cpu_map);
			}
746
		}
747 748 749 750 751 752 753 754 755 756 757 758
	} else if ((c->x86 == 0x15) && ((index == 1) || (index == 2))) {
		ret = 1;
		for_each_cpu(i, cpu_sibling_mask(cpu)) {
			if (!per_cpu(ici_cpuid4_info, i))
				continue;
			this_leaf = CPUID4_INFO_IDX(i, index);
			for_each_cpu(sibling, cpu_sibling_mask(cpu)) {
				if (!cpu_online(sibling))
					continue;
				set_bit(sibling, this_leaf->shared_cpu_map);
			}
		}
759
	}
760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775

	return ret;
}

static void __cpuinit cache_shared_cpu_map_setup(unsigned int cpu, int index)
{
	struct _cpuid4_info *this_leaf, *sibling_leaf;
	unsigned long num_threads_sharing;
	int index_msb, i;
	struct cpuinfo_x86 *c = &cpu_data(cpu);

	if (c->x86_vendor == X86_VENDOR_AMD) {
		if (cache_shared_amd_cpu_map_setup(cpu, index))
			return;
	}

L
Linus Torvalds 已提交
776
	this_leaf = CPUID4_INFO_IDX(cpu, index);
777
	num_threads_sharing = 1 + this_leaf->base.eax.split.num_threads_sharing;
L
Linus Torvalds 已提交
778 779

	if (num_threads_sharing == 1)
780
		cpumask_set_cpu(cpu, to_cpumask(this_leaf->shared_cpu_map));
781 782 783 784
	else {
		index_msb = get_count_order(num_threads_sharing);

		for_each_online_cpu(i) {
785 786
			if (cpu_data(i).apicid >> index_msb ==
			    c->apicid >> index_msb) {
787 788
				cpumask_set_cpu(i,
					to_cpumask(this_leaf->shared_cpu_map));
789
				if (i != cpu && per_cpu(ici_cpuid4_info, i))  {
790 791 792 793
					sibling_leaf =
						CPUID4_INFO_IDX(i, index);
					cpumask_set_cpu(cpu, to_cpumask(
						sibling_leaf->shared_cpu_map));
794 795 796 797 798
				}
			}
		}
	}
}
799
static void __cpuinit cache_remove_shared_cpu_map(unsigned int cpu, int index)
800 801 802 803 804
{
	struct _cpuid4_info	*this_leaf, *sibling_leaf;
	int sibling;

	this_leaf = CPUID4_INFO_IDX(cpu, index);
805
	for_each_cpu(sibling, to_cpumask(this_leaf->shared_cpu_map)) {
806
		sibling_leaf = CPUID4_INFO_IDX(sibling, index);
807 808
		cpumask_clear_cpu(cpu,
				  to_cpumask(sibling_leaf->shared_cpu_map));
809
	}
L
Linus Torvalds 已提交
810 811
}
#else
A
Alan Cox 已提交
812 813 814 815 816 817 818
static void __cpuinit cache_shared_cpu_map_setup(unsigned int cpu, int index)
{
}

static void __cpuinit cache_remove_shared_cpu_map(unsigned int cpu, int index)
{
}
L
Linus Torvalds 已提交
819 820
#endif

821
static void __cpuinit free_cache_attributes(unsigned int cpu)
L
Linus Torvalds 已提交
822
{
823 824 825 826 827
	int i;

	for (i = 0; i < num_cache_leaves; i++)
		cache_remove_shared_cpu_map(cpu, i);

828 829
	kfree(per_cpu(ici_cpuid4_info, cpu));
	per_cpu(ici_cpuid4_info, cpu) = NULL;
L
Linus Torvalds 已提交
830 831
}

832
static void __cpuinit get_cpu_leaves(void *_retval)
L
Linus Torvalds 已提交
833
{
834
	int j, *retval = _retval, cpu = smp_processor_id();
835

L
Linus Torvalds 已提交
836 837
	/* Do cpuid and store the results */
	for (j = 0; j < num_cache_leaves; j++) {
838 839 840
		struct _cpuid4_info *this_leaf = CPUID4_INFO_IDX(cpu, j);

		*retval = cpuid4_cache_lookup_regs(j, &this_leaf->base);
841
		if (unlikely(*retval < 0)) {
842 843 844 845
			int i;

			for (i = 0; i < j; i++)
				cache_remove_shared_cpu_map(cpu, i);
846
			break;
847
		}
L
Linus Torvalds 已提交
848 849
		cache_shared_cpu_map_setup(cpu, j);
	}
850 851 852 853 854 855 856 857 858
}

static int __cpuinit detect_cache_attributes(unsigned int cpu)
{
	int			retval;

	if (num_cache_leaves == 0)
		return -ENOENT;

859
	per_cpu(ici_cpuid4_info, cpu) = kzalloc(
860
	    sizeof(struct _cpuid4_info) * num_cache_leaves, GFP_KERNEL);
861
	if (per_cpu(ici_cpuid4_info, cpu) == NULL)
862
		return -ENOMEM;
L
Linus Torvalds 已提交
863

864
	smp_call_function_single(cpu, get_cpu_leaves, &retval, true);
865
	if (retval) {
866 867
		kfree(per_cpu(ici_cpuid4_info, cpu));
		per_cpu(ici_cpuid4_info, cpu) = NULL;
868 869
	}

870
	return retval;
L
Linus Torvalds 已提交
871 872 873 874
}

#include <linux/kobject.h>
#include <linux/sysfs.h>
875
#include <linux/cpu.h>
L
Linus Torvalds 已提交
876 877

/* pointer to kobject for cpuX/cache */
878
static DEFINE_PER_CPU(struct kobject *, ici_cache_kobject);
L
Linus Torvalds 已提交
879 880 881 882 883 884 885 886

struct _index_kobject {
	struct kobject kobj;
	unsigned int cpu;
	unsigned short index;
};

/* pointer to array of kobjects for cpuX/cache/indexY */
887 888
static DEFINE_PER_CPU(struct _index_kobject *, ici_index_kobject);
#define INDEX_KOBJECT_PTR(x, y)		(&((per_cpu(ici_index_kobject, x))[y]))
L
Linus Torvalds 已提交
889 890

#define show_one_plus(file_name, object, val)				\
891 892
static ssize_t show_##file_name(struct _cpuid4_info *this_leaf, char *buf, \
				unsigned int cpu)			\
L
Linus Torvalds 已提交
893
{									\
A
Alan Cox 已提交
894
	return sprintf(buf, "%lu\n", (unsigned long)this_leaf->object + val); \
L
Linus Torvalds 已提交
895 896
}

897 898 899 900 901
show_one_plus(level, base.eax.split.level, 0);
show_one_plus(coherency_line_size, base.ebx.split.coherency_line_size, 1);
show_one_plus(physical_line_partition, base.ebx.split.physical_line_partition, 1);
show_one_plus(ways_of_associativity, base.ebx.split.ways_of_associativity, 1);
show_one_plus(number_of_sets, base.ecx.split.number_of_sets, 1);
L
Linus Torvalds 已提交
902

903 904
static ssize_t show_size(struct _cpuid4_info *this_leaf, char *buf,
			 unsigned int cpu)
L
Linus Torvalds 已提交
905
{
906
	return sprintf(buf, "%luK\n", this_leaf->base.size / 1024);
L
Linus Torvalds 已提交
907 908
}

909 910
static ssize_t show_shared_cpu_map_func(struct _cpuid4_info *this_leaf,
					int type, char *buf)
L
Linus Torvalds 已提交
911
{
912
	ptrdiff_t len = PTR_ALIGN(buf + PAGE_SIZE - 1, PAGE_SIZE) - buf;
913 914
	int n = 0;

915
	if (len > 1) {
916
		const struct cpumask *mask;
917

918
		mask = to_cpumask(this_leaf->shared_cpu_map);
A
Alan Cox 已提交
919
		n = type ?
920 921
			cpulist_scnprintf(buf, len-2, mask) :
			cpumask_scnprintf(buf, len-2, mask);
922 923
		buf[n++] = '\n';
		buf[n] = '\0';
924 925
	}
	return n;
L
Linus Torvalds 已提交
926 927
}

928 929
static inline ssize_t show_shared_cpu_map(struct _cpuid4_info *leaf, char *buf,
					  unsigned int cpu)
930 931 932 933
{
	return show_shared_cpu_map_func(leaf, 0, buf);
}

934 935
static inline ssize_t show_shared_cpu_list(struct _cpuid4_info *leaf, char *buf,
					   unsigned int cpu)
936 937 938 939
{
	return show_shared_cpu_map_func(leaf, 1, buf);
}

940 941
static ssize_t show_type(struct _cpuid4_info *this_leaf, char *buf,
			 unsigned int cpu)
942
{
943
	switch (this_leaf->base.eax.split.type) {
944
	case CACHE_TYPE_DATA:
L
Linus Torvalds 已提交
945
		return sprintf(buf, "Data\n");
946
	case CACHE_TYPE_INST:
L
Linus Torvalds 已提交
947
		return sprintf(buf, "Instruction\n");
948
	case CACHE_TYPE_UNIFIED:
L
Linus Torvalds 已提交
949
		return sprintf(buf, "Unified\n");
950
	default:
L
Linus Torvalds 已提交
951 952 953 954
		return sprintf(buf, "Unknown\n");
	}
}

955 956
#define to_object(k)	container_of(k, struct _index_kobject, kobj)
#define to_attr(a)	container_of(a, struct _cache_attr, attr)
957

L
Linus Torvalds 已提交
958 959 960 961 962 963 964 965 966 967 968 969
#define define_one_ro(_name) \
static struct _cache_attr _name = \
	__ATTR(_name, 0444, show_##_name, NULL)

define_one_ro(level);
define_one_ro(type);
define_one_ro(coherency_line_size);
define_one_ro(physical_line_partition);
define_one_ro(ways_of_associativity);
define_one_ro(number_of_sets);
define_one_ro(size);
define_one_ro(shared_cpu_map);
970
define_one_ro(shared_cpu_list);
L
Linus Torvalds 已提交
971

A
Alan Cox 已提交
972
static struct attribute *default_attrs[] = {
973 974 975 976 977 978 979 980 981
	&type.attr,
	&level.attr,
	&coherency_line_size.attr,
	&physical_line_partition.attr,
	&ways_of_associativity.attr,
	&number_of_sets.attr,
	&size.attr,
	&shared_cpu_map.attr,
	&shared_cpu_list.attr,
982 983 984
	NULL
};

985
#ifdef CONFIG_AMD_NB
986 987 988 989 990 991 992 993 994 995 996 997 998
static struct attribute ** __cpuinit amd_l3_attrs(void)
{
	static struct attribute **attrs;
	int n;

	if (attrs)
		return attrs;

	n = sizeof (default_attrs) / sizeof (struct attribute *);

	if (amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE))
		n += 2;

999 1000 1001
	if (amd_nb_has_feature(AMD_NB_L3_PARTITIONING))
		n += 1;

1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013
	attrs = kzalloc(n * sizeof (struct attribute *), GFP_KERNEL);
	if (attrs == NULL)
		return attrs = default_attrs;

	for (n = 0; default_attrs[n]; n++)
		attrs[n] = default_attrs[n];

	if (amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE)) {
		attrs[n++] = &cache_disable_0.attr;
		attrs[n++] = &cache_disable_1.attr;
	}

1014 1015 1016
	if (amd_nb_has_feature(AMD_NB_L3_PARTITIONING))
		attrs[n++] = &subcaches.attr;

1017 1018
	return attrs;
}
1019
#endif
L
Linus Torvalds 已提交
1020

A
Alan Cox 已提交
1021
static ssize_t show(struct kobject *kobj, struct attribute *attr, char *buf)
L
Linus Torvalds 已提交
1022 1023 1024 1025 1026 1027 1028
{
	struct _cache_attr *fattr = to_attr(attr);
	struct _index_kobject *this_leaf = to_object(kobj);
	ssize_t ret;

	ret = fattr->show ?
		fattr->show(CPUID4_INFO_IDX(this_leaf->cpu, this_leaf->index),
1029
			buf, this_leaf->cpu) :
1030
		0;
L
Linus Torvalds 已提交
1031 1032 1033
	return ret;
}

A
Alan Cox 已提交
1034 1035
static ssize_t store(struct kobject *kobj, struct attribute *attr,
		     const char *buf, size_t count)
L
Linus Torvalds 已提交
1036
{
1037 1038 1039 1040
	struct _cache_attr *fattr = to_attr(attr);
	struct _index_kobject *this_leaf = to_object(kobj);
	ssize_t ret;

1041 1042
	ret = fattr->store ?
		fattr->store(CPUID4_INFO_IDX(this_leaf->cpu, this_leaf->index),
1043
			buf, count, this_leaf->cpu) :
1044 1045
		0;
	return ret;
L
Linus Torvalds 已提交
1046 1047
}

1048
static const struct sysfs_ops sysfs_ops = {
L
Linus Torvalds 已提交
1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061
	.show   = show,
	.store  = store,
};

static struct kobj_type ktype_cache = {
	.sysfs_ops	= &sysfs_ops,
	.default_attrs	= default_attrs,
};

static struct kobj_type ktype_percpu_entry = {
	.sysfs_ops	= &sysfs_ops,
};

1062
static void __cpuinit cpuid4_cache_sysfs_exit(unsigned int cpu)
L
Linus Torvalds 已提交
1063
{
1064 1065 1066 1067
	kfree(per_cpu(ici_cache_kobject, cpu));
	kfree(per_cpu(ici_index_kobject, cpu));
	per_cpu(ici_cache_kobject, cpu) = NULL;
	per_cpu(ici_index_kobject, cpu) = NULL;
L
Linus Torvalds 已提交
1068 1069 1070
	free_cache_attributes(cpu);
}

1071
static int __cpuinit cpuid4_cache_sysfs_init(unsigned int cpu)
L
Linus Torvalds 已提交
1072
{
1073
	int err;
L
Linus Torvalds 已提交
1074 1075 1076 1077

	if (num_cache_leaves == 0)
		return -ENOENT;

1078 1079 1080
	err = detect_cache_attributes(cpu);
	if (err)
		return err;
L
Linus Torvalds 已提交
1081 1082

	/* Allocate all required memory */
1083
	per_cpu(ici_cache_kobject, cpu) =
1084
		kzalloc(sizeof(struct kobject), GFP_KERNEL);
1085
	if (unlikely(per_cpu(ici_cache_kobject, cpu) == NULL))
L
Linus Torvalds 已提交
1086 1087
		goto err_out;

1088
	per_cpu(ici_index_kobject, cpu) = kzalloc(
A
Alan Cox 已提交
1089
	    sizeof(struct _index_kobject) * num_cache_leaves, GFP_KERNEL);
1090
	if (unlikely(per_cpu(ici_index_kobject, cpu) == NULL))
L
Linus Torvalds 已提交
1091 1092 1093 1094 1095 1096 1097 1098 1099
		goto err_out;

	return 0;

err_out:
	cpuid4_cache_sysfs_exit(cpu);
	return -ENOMEM;
}

1100
static DECLARE_BITMAP(cache_dev_map, NR_CPUS);
1101

L
Linus Torvalds 已提交
1102
/* Add/Remove cache interface for CPU device */
1103
static int __cpuinit cache_add_dev(struct device *dev)
L
Linus Torvalds 已提交
1104
{
1105
	unsigned int cpu = dev->id;
L
Linus Torvalds 已提交
1106 1107
	unsigned long i, j;
	struct _index_kobject *this_object;
1108
	struct _cpuid4_info   *this_leaf;
1109
	int retval;
L
Linus Torvalds 已提交
1110 1111 1112 1113 1114

	retval = cpuid4_cache_sysfs_init(cpu);
	if (unlikely(retval < 0))
		return retval;

1115
	retval = kobject_init_and_add(per_cpu(ici_cache_kobject, cpu),
1116
				      &ktype_percpu_entry,
1117
				      &dev->kobj, "%s", "cache");
1118 1119 1120 1121
	if (retval < 0) {
		cpuid4_cache_sysfs_exit(cpu);
		return retval;
	}
L
Linus Torvalds 已提交
1122 1123

	for (i = 0; i < num_cache_leaves; i++) {
A
Alan Cox 已提交
1124
		this_object = INDEX_KOBJECT_PTR(cpu, i);
L
Linus Torvalds 已提交
1125 1126
		this_object->cpu = cpu;
		this_object->index = i;
1127 1128 1129

		this_leaf = CPUID4_INFO_IDX(cpu, i);

1130 1131
		ktype_cache.default_attrs = default_attrs;
#ifdef CONFIG_AMD_NB
1132
		if (this_leaf->base.nb)
1133 1134
			ktype_cache.default_attrs = amd_l3_attrs();
#endif
1135
		retval = kobject_init_and_add(&(this_object->kobj),
1136
					      &ktype_cache,
1137
					      per_cpu(ici_cache_kobject, cpu),
1138
					      "index%1lu", i);
L
Linus Torvalds 已提交
1139
		if (unlikely(retval)) {
A
Alan Cox 已提交
1140 1141
			for (j = 0; j < i; j++)
				kobject_put(&(INDEX_KOBJECT_PTR(cpu, j)->kobj));
1142
			kobject_put(per_cpu(ici_cache_kobject, cpu));
L
Linus Torvalds 已提交
1143
			cpuid4_cache_sysfs_exit(cpu);
1144
			return retval;
L
Linus Torvalds 已提交
1145
		}
1146
		kobject_uevent(&(this_object->kobj), KOBJ_ADD);
L
Linus Torvalds 已提交
1147
	}
1148
	cpumask_set_cpu(cpu, to_cpumask(cache_dev_map));
1149

1150
	kobject_uevent(per_cpu(ici_cache_kobject, cpu), KOBJ_ADD);
1151
	return 0;
L
Linus Torvalds 已提交
1152 1153
}

1154
static void __cpuinit cache_remove_dev(struct device *dev)
L
Linus Torvalds 已提交
1155
{
1156
	unsigned int cpu = dev->id;
L
Linus Torvalds 已提交
1157 1158
	unsigned long i;

1159
	if (per_cpu(ici_cpuid4_info, cpu) == NULL)
1160
		return;
1161
	if (!cpumask_test_cpu(cpu, to_cpumask(cache_dev_map)))
1162
		return;
1163
	cpumask_clear_cpu(cpu, to_cpumask(cache_dev_map));
1164 1165

	for (i = 0; i < num_cache_leaves; i++)
A
Alan Cox 已提交
1166
		kobject_put(&(INDEX_KOBJECT_PTR(cpu, i)->kobj));
1167
	kobject_put(per_cpu(ici_cache_kobject, cpu));
L
Linus Torvalds 已提交
1168
	cpuid4_cache_sysfs_exit(cpu);
1169 1170
}

1171
static int __cpuinit cacheinfo_cpu_callback(struct notifier_block *nfb,
1172 1173 1174
					unsigned long action, void *hcpu)
{
	unsigned int cpu = (unsigned long)hcpu;
1175
	struct device *dev;
1176

1177
	dev = get_cpu_device(cpu);
1178 1179
	switch (action) {
	case CPU_ONLINE:
1180
	case CPU_ONLINE_FROZEN:
1181
		cache_add_dev(dev);
1182 1183
		break;
	case CPU_DEAD:
1184
	case CPU_DEAD_FROZEN:
1185
		cache_remove_dev(dev);
1186 1187 1188
		break;
	}
	return NOTIFY_OK;
L
Linus Torvalds 已提交
1189 1190
}

A
Alan Cox 已提交
1191
static struct notifier_block __cpuinitdata cacheinfo_cpu_notifier = {
1192
	.notifier_call = cacheinfo_cpu_callback,
L
Linus Torvalds 已提交
1193 1194
};

1195
static int __cpuinit cache_sysfs_init(void)
L
Linus Torvalds 已提交
1196
{
1197 1198
	int i;

L
Linus Torvalds 已提交
1199 1200 1201
	if (num_cache_leaves == 0)
		return 0;

1202
	for_each_online_cpu(i) {
1203
		int err;
1204
		struct device *dev = get_cpu_device(i);
1205

1206
		err = cache_add_dev(dev);
1207 1208
		if (err)
			return err;
1209
	}
1210
	register_hotcpu_notifier(&cacheinfo_cpu_notifier);
1211
	return 0;
L
Linus Torvalds 已提交
1212 1213
}

1214
device_initcall(cache_sysfs_init);
L
Linus Torvalds 已提交
1215 1216

#endif