intel_cacheinfo.c 33.0 KB
Newer Older
L
Linus Torvalds 已提交
1
/*
2
 *	Routines to identify caches on Intel CPU.
L
Linus Torvalds 已提交
3
 *
4 5
 *	Changes:
 *	Venkatesh Pallipadi	: Adding cache identification through cpuid(4)
A
Alan Cox 已提交
6
 *	Ashok Raj <ashok.raj@intel.com>: Work with CPU hotplug infrastructure.
7
 *	Andi Kleen / Andreas Herrmann	: CPUID4 emulation on AMD.
L
Linus Torvalds 已提交
8 9 10 11 12 13 14
 */

#include <linux/init.h>
#include <linux/slab.h>
#include <linux/device.h>
#include <linux/compiler.h>
#include <linux/cpu.h>
T
Tim Schmielau 已提交
15
#include <linux/sched.h>
16
#include <linux/pci.h>
L
Linus Torvalds 已提交
17 18

#include <asm/processor.h>
A
Alan Cox 已提交
19
#include <linux/smp.h>
20
#include <asm/amd_nb.h>
21
#include <asm/smp.h>
L
Linus Torvalds 已提交
22 23 24 25 26 27 28

#define LVL_1_INST	1
#define LVL_1_DATA	2
#define LVL_2		3
#define LVL_3		4
#define LVL_TRACE	5

A
Alan Cox 已提交
29
struct _cache_table {
L
Linus Torvalds 已提交
30 31 32 33 34
	unsigned char descriptor;
	char cache_type;
	short size;
};

D
Dave Jones 已提交
35 36
#define MB(x)	((x) * 1024)

A
Alan Cox 已提交
37 38 39
/* All the cache descriptor types we care about (no TLB or
   trace cache entries) */

40
static const struct _cache_table cache_table[] =
L
Linus Torvalds 已提交
41 42 43
{
	{ 0x06, LVL_1_INST, 8 },	/* 4-way set assoc, 32 byte line size */
	{ 0x08, LVL_1_INST, 16 },	/* 4-way set assoc, 32 byte line size */
44
	{ 0x09, LVL_1_INST, 32 },	/* 4-way set assoc, 64 byte line size */
L
Linus Torvalds 已提交
45 46
	{ 0x0a, LVL_1_DATA, 8 },	/* 2 way set assoc, 32 byte line size */
	{ 0x0c, LVL_1_DATA, 16 },	/* 4-way set assoc, 32 byte line size */
47
	{ 0x0d, LVL_1_DATA, 16 },	/* 4-way set assoc, 64 byte line size */
48
	{ 0x0e, LVL_1_DATA, 24 },	/* 6-way set assoc, 64 byte line size */
49
	{ 0x21, LVL_2,      256 },	/* 8-way set assoc, 64 byte line size */
L
Linus Torvalds 已提交
50
	{ 0x22, LVL_3,      512 },	/* 4-way set assoc, sectored cache, 64 byte line size */
D
Dave Jones 已提交
51 52 53
	{ 0x23, LVL_3,      MB(1) },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x25, LVL_3,      MB(2) },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x29, LVL_3,      MB(4) },	/* 8-way set assoc, sectored cache, 64 byte line size */
L
Linus Torvalds 已提交
54 55 56
	{ 0x2c, LVL_1_DATA, 32 },	/* 8-way set assoc, 64 byte line size */
	{ 0x30, LVL_1_INST, 32 },	/* 8-way set assoc, 64 byte line size */
	{ 0x39, LVL_2,      128 },	/* 4-way set assoc, sectored cache, 64 byte line size */
57
	{ 0x3a, LVL_2,      192 },	/* 6-way set assoc, sectored cache, 64 byte line size */
L
Linus Torvalds 已提交
58 59
	{ 0x3b, LVL_2,      128 },	/* 2-way set assoc, sectored cache, 64 byte line size */
	{ 0x3c, LVL_2,      256 },	/* 4-way set assoc, sectored cache, 64 byte line size */
60 61
	{ 0x3d, LVL_2,      384 },	/* 6-way set assoc, sectored cache, 64 byte line size */
	{ 0x3e, LVL_2,      512 },	/* 4-way set assoc, sectored cache, 64 byte line size */
62
	{ 0x3f, LVL_2,      256 },	/* 2-way set assoc, 64 byte line size */
L
Linus Torvalds 已提交
63 64 65
	{ 0x41, LVL_2,      128 },	/* 4-way set assoc, 32 byte line size */
	{ 0x42, LVL_2,      256 },	/* 4-way set assoc, 32 byte line size */
	{ 0x43, LVL_2,      512 },	/* 4-way set assoc, 32 byte line size */
D
Dave Jones 已提交
66 67 68 69
	{ 0x44, LVL_2,      MB(1) },	/* 4-way set assoc, 32 byte line size */
	{ 0x45, LVL_2,      MB(2) },	/* 4-way set assoc, 32 byte line size */
	{ 0x46, LVL_3,      MB(4) },	/* 4-way set assoc, 64 byte line size */
	{ 0x47, LVL_3,      MB(8) },	/* 8-way set assoc, 64 byte line size */
70
	{ 0x48, LVL_2,      MB(3) },	/* 12-way set assoc, 64 byte line size */
D
Dave Jones 已提交
71 72 73 74 75 76
	{ 0x49, LVL_3,      MB(4) },	/* 16-way set assoc, 64 byte line size */
	{ 0x4a, LVL_3,      MB(6) },	/* 12-way set assoc, 64 byte line size */
	{ 0x4b, LVL_3,      MB(8) },	/* 16-way set assoc, 64 byte line size */
	{ 0x4c, LVL_3,      MB(12) },	/* 12-way set assoc, 64 byte line size */
	{ 0x4d, LVL_3,      MB(16) },	/* 16-way set assoc, 64 byte line size */
	{ 0x4e, LVL_2,      MB(6) },	/* 24-way set assoc, 64 byte line size */
L
Linus Torvalds 已提交
77 78 79 80 81 82 83
	{ 0x60, LVL_1_DATA, 16 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x66, LVL_1_DATA, 8 },	/* 4-way set assoc, sectored cache, 64 byte line size */
	{ 0x67, LVL_1_DATA, 16 },	/* 4-way set assoc, sectored cache, 64 byte line size */
	{ 0x68, LVL_1_DATA, 32 },	/* 4-way set assoc, sectored cache, 64 byte line size */
	{ 0x70, LVL_TRACE,  12 },	/* 8-way set assoc */
	{ 0x71, LVL_TRACE,  16 },	/* 8-way set assoc */
	{ 0x72, LVL_TRACE,  32 },	/* 8-way set assoc */
84
	{ 0x73, LVL_TRACE,  64 },	/* 8-way set assoc */
D
Dave Jones 已提交
85 86 87 88 89 90 91
	{ 0x78, LVL_2,      MB(1) },	/* 4-way set assoc, 64 byte line size */
	{ 0x79, LVL_2,      128 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x7a, LVL_2,      256 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x7b, LVL_2,      512 },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x7c, LVL_2,      MB(1) },	/* 8-way set assoc, sectored cache, 64 byte line size */
	{ 0x7d, LVL_2,      MB(2) },	/* 8-way set assoc, 64 byte line size */
	{ 0x7f, LVL_2,      512 },	/* 2-way set assoc, 64 byte line size */
92
	{ 0x80, LVL_2,      512 },	/* 8-way set assoc, 64 byte line size */
D
Dave Jones 已提交
93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113
	{ 0x82, LVL_2,      256 },	/* 8-way set assoc, 32 byte line size */
	{ 0x83, LVL_2,      512 },	/* 8-way set assoc, 32 byte line size */
	{ 0x84, LVL_2,      MB(1) },	/* 8-way set assoc, 32 byte line size */
	{ 0x85, LVL_2,      MB(2) },	/* 8-way set assoc, 32 byte line size */
	{ 0x86, LVL_2,      512 },	/* 4-way set assoc, 64 byte line size */
	{ 0x87, LVL_2,      MB(1) },	/* 8-way set assoc, 64 byte line size */
	{ 0xd0, LVL_3,      512 },	/* 4-way set assoc, 64 byte line size */
	{ 0xd1, LVL_3,      MB(1) },	/* 4-way set assoc, 64 byte line size */
	{ 0xd2, LVL_3,      MB(2) },	/* 4-way set assoc, 64 byte line size */
	{ 0xd6, LVL_3,      MB(1) },	/* 8-way set assoc, 64 byte line size */
	{ 0xd7, LVL_3,      MB(2) },	/* 8-way set assoc, 64 byte line size */
	{ 0xd8, LVL_3,      MB(4) },	/* 12-way set assoc, 64 byte line size */
	{ 0xdc, LVL_3,      MB(2) },	/* 12-way set assoc, 64 byte line size */
	{ 0xdd, LVL_3,      MB(4) },	/* 12-way set assoc, 64 byte line size */
	{ 0xde, LVL_3,      MB(8) },	/* 12-way set assoc, 64 byte line size */
	{ 0xe2, LVL_3,      MB(2) },	/* 16-way set assoc, 64 byte line size */
	{ 0xe3, LVL_3,      MB(4) },	/* 16-way set assoc, 64 byte line size */
	{ 0xe4, LVL_3,      MB(8) },	/* 16-way set assoc, 64 byte line size */
	{ 0xea, LVL_3,      MB(12) },	/* 24-way set assoc, 64 byte line size */
	{ 0xeb, LVL_3,      MB(18) },	/* 24-way set assoc, 64 byte line size */
	{ 0xec, LVL_3,      MB(24) },	/* 24-way set assoc, 64 byte line size */
L
Linus Torvalds 已提交
114 115 116 117
	{ 0x00, 0, 0}
};


A
Alan Cox 已提交
118
enum _cache_type {
L
Linus Torvalds 已提交
119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153
	CACHE_TYPE_NULL	= 0,
	CACHE_TYPE_DATA = 1,
	CACHE_TYPE_INST = 2,
	CACHE_TYPE_UNIFIED = 3
};

union _cpuid4_leaf_eax {
	struct {
		enum _cache_type	type:5;
		unsigned int		level:3;
		unsigned int		is_self_initializing:1;
		unsigned int		is_fully_associative:1;
		unsigned int		reserved:4;
		unsigned int		num_threads_sharing:12;
		unsigned int		num_cores_on_die:6;
	} split;
	u32 full;
};

union _cpuid4_leaf_ebx {
	struct {
		unsigned int		coherency_line_size:12;
		unsigned int		physical_line_partition:10;
		unsigned int		ways_of_associativity:10;
	} split;
	u32 full;
};

union _cpuid4_leaf_ecx {
	struct {
		unsigned int		number_of_sets:32;
	} split;
	u32 full;
};

154
struct _cpuid4_info_regs {
L
Linus Torvalds 已提交
155 156 157 158
	union _cpuid4_leaf_eax eax;
	union _cpuid4_leaf_ebx ebx;
	union _cpuid4_leaf_ecx ecx;
	unsigned long size;
159
	struct amd_northbridge *nb;
160 161
};

162 163 164
struct _cpuid4_info {
	struct _cpuid4_info_regs base;
	DECLARE_BITMAP(shared_cpu_map, NR_CPUS);
L
Linus Torvalds 已提交
165 166
};

167 168 169 170
unsigned short			num_cache_leaves;

/* AMD doesn't have CPUID4. Emulate it here to report the same
   information to the user.  This makes some assumptions about the machine:
171
   L2 not shared, no SMT etc. that is currently true on AMD CPUs.
172 173 174 175 176

   In theory the TLBs could be reported as fake type (they are in "dummy").
   Maybe later */
union l1_cache {
	struct {
A
Alan Cox 已提交
177 178 179 180
		unsigned line_size:8;
		unsigned lines_per_tag:8;
		unsigned assoc:8;
		unsigned size_in_kb:8;
181 182 183 184 185 186
	};
	unsigned val;
};

union l2_cache {
	struct {
A
Alan Cox 已提交
187 188 189 190
		unsigned line_size:8;
		unsigned lines_per_tag:4;
		unsigned assoc:4;
		unsigned size_in_kb:16;
191 192 193 194
	};
	unsigned val;
};

195 196
union l3_cache {
	struct {
A
Alan Cox 已提交
197 198 199 200 201
		unsigned line_size:8;
		unsigned lines_per_tag:4;
		unsigned assoc:4;
		unsigned res:2;
		unsigned size_encoded:14;
202 203 204 205
	};
	unsigned val;
};

206
static const unsigned short assocs[] = {
207 208 209 210 211 212 213
	[1] = 1,
	[2] = 2,
	[4] = 4,
	[6] = 8,
	[8] = 16,
	[0xa] = 32,
	[0xb] = 48,
214
	[0xc] = 64,
215 216 217
	[0xd] = 96,
	[0xe] = 128,
	[0xf] = 0xffff /* fully associative - no way to show this currently */
218 219
};

220 221
static const unsigned char levels[] = { 1, 1, 2, 3 };
static const unsigned char types[] = { 1, 2, 3, 3 };
222

223
static void
224 225 226
amd_cpuid4(int leaf, union _cpuid4_leaf_eax *eax,
		     union _cpuid4_leaf_ebx *ebx,
		     union _cpuid4_leaf_ecx *ecx)
227 228 229 230 231
{
	unsigned dummy;
	unsigned line_size, lines_per_tag, assoc, size_in_kb;
	union l1_cache l1i, l1d;
	union l2_cache l2;
232 233
	union l3_cache l3;
	union l1_cache *l1 = &l1d;
234 235 236 237 238 239

	eax->full = 0;
	ebx->full = 0;
	ecx->full = 0;

	cpuid(0x80000005, &dummy, &dummy, &l1d.val, &l1i.val);
240
	cpuid(0x80000006, &dummy, &dummy, &l2.val, &l3.val);
241

242 243 244 245 246 247
	switch (leaf) {
	case 1:
		l1 = &l1i;
	case 0:
		if (!l1->val)
			return;
248
		assoc = assocs[l1->assoc];
249 250 251
		line_size = l1->line_size;
		lines_per_tag = l1->lines_per_tag;
		size_in_kb = l1->size_in_kb;
252 253 254 255
		break;
	case 2:
		if (!l2.val)
			return;
256
		assoc = assocs[l2.assoc];
257 258 259
		line_size = l2.line_size;
		lines_per_tag = l2.lines_per_tag;
		/* cpu_data has errata corrections for K7 applied */
260
		size_in_kb = __this_cpu_read(cpu_info.x86_cache_size);
261 262 263 264
		break;
	case 3:
		if (!l3.val)
			return;
265
		assoc = assocs[l3.assoc];
266 267 268
		line_size = l3.line_size;
		lines_per_tag = l3.lines_per_tag;
		size_in_kb = l3.size_encoded * 512;
269 270 271 272
		if (boot_cpu_has(X86_FEATURE_AMD_DCM)) {
			size_in_kb = size_in_kb >> 1;
			assoc = assoc >> 1;
		}
273 274 275
		break;
	default:
		return;
276 277
	}

278 279 280
	eax->split.is_self_initializing = 1;
	eax->split.type = types[leaf];
	eax->split.level = levels[leaf];
281
	eax->split.num_threads_sharing = 0;
282
	eax->split.num_cores_on_die = __this_cpu_read(cpu_info.x86_max_cores) - 1;
283 284


285
	if (assoc == 0xffff)
286 287
		eax->split.is_fully_associative = 1;
	ebx->split.coherency_line_size = line_size - 1;
288
	ebx->split.ways_of_associativity = assoc - 1;
289 290 291 292
	ebx->split.physical_line_partition = lines_per_tag - 1;
	ecx->split.number_of_sets = (size_in_kb * 1024) / line_size /
		(ebx->split.ways_of_associativity + 1) - 1;
}
L
Linus Torvalds 已提交
293

294 295
struct _cache_attr {
	struct attribute attr;
296 297 298
	ssize_t (*show)(struct _cpuid4_info *, char *, unsigned int);
	ssize_t (*store)(struct _cpuid4_info *, const char *, size_t count,
			 unsigned int);
299 300
};

301
#if defined(CONFIG_AMD_NB) && defined(CONFIG_SYSFS)
302 303 304
/*
 * L3 cache descriptors
 */
305
static void amd_calc_l3_indices(struct amd_northbridge *nb)
306
{
307
	struct amd_l3_cache *l3 = &nb->l3_cache;
308
	unsigned int sc0, sc1, sc2, sc3;
309
	u32 val = 0;
310

311
	pci_read_config_dword(nb->misc, 0x1C4, &val);
312 313

	/* calculate subcache sizes */
314 315
	l3->subcaches[0] = sc0 = !(val & BIT(0));
	l3->subcaches[1] = sc1 = !(val & BIT(4));
316 317 318 319 320 321

	if (boot_cpu_data.x86 == 0x15) {
		l3->subcaches[0] = sc0 += !(val & BIT(1));
		l3->subcaches[1] = sc1 += !(val & BIT(5));
	}

322 323 324
	l3->subcaches[2] = sc2 = !(val & BIT(8))  + !(val & BIT(9));
	l3->subcaches[3] = sc3 = !(val & BIT(12)) + !(val & BIT(13));

325
	l3->indices = (max(max3(sc0, sc1, sc2), sc3) << 10) - 1;
326 327
}

328
static void amd_init_l3_cache(struct _cpuid4_info_regs *this_leaf, int index)
329
{
330 331
	int node;

332
	/* only for L3, and not in virtualized environments */
333
	if (index < 3)
334 335
		return;

336
	node = amd_get_nb_id(smp_processor_id());
337 338 339
	this_leaf->nb = node_to_amd_nb(node);
	if (this_leaf->nb && !this_leaf->nb->l3_cache.indices)
		amd_calc_l3_indices(this_leaf->nb);
340 341
}

342 343 344 345 346 347 348
/*
 * check whether a slot used for disabling an L3 index is occupied.
 * @l3: L3 cache descriptor
 * @slot: slot number (0..1)
 *
 * @returns: the disabled index if used or negative value if slot free.
 */
349
int amd_get_l3_disable_slot(struct amd_northbridge *nb, unsigned slot)
350 351 352
{
	unsigned int reg = 0;

353
	pci_read_config_dword(nb->misc, 0x1BC + slot * 4, &reg);
354 355 356 357 358 359 360 361

	/* check whether this slot is activated already */
	if (reg & (3UL << 30))
		return reg & 0xfff;

	return -1;
}

362
static ssize_t show_cache_disable(struct _cpuid4_info *this_leaf, char *buf,
363
				  unsigned int slot)
364
{
365
	int index;
366

367
	if (!this_leaf->base.nb || !amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE))
368 369
		return -EINVAL;

370
	index = amd_get_l3_disable_slot(this_leaf->base.nb, slot);
371 372
	if (index >= 0)
		return sprintf(buf, "%d\n", index);
373

374
	return sprintf(buf, "FREE\n");
375 376
}

377
#define SHOW_CACHE_DISABLE(slot)					\
378
static ssize_t								\
379 380
show_cache_disable_##slot(struct _cpuid4_info *this_leaf, char *buf,	\
			  unsigned int cpu)				\
381
{									\
382
	return show_cache_disable(this_leaf, buf, slot);		\
383 384 385 386
}
SHOW_CACHE_DISABLE(0)
SHOW_CACHE_DISABLE(1)

387
static void amd_l3_disable_index(struct amd_northbridge *nb, int cpu,
388 389 390 391 392 393 394 395 396 397 398 399
				 unsigned slot, unsigned long idx)
{
	int i;

	idx |= BIT(30);

	/*
	 *  disable index in all 4 subcaches
	 */
	for (i = 0; i < 4; i++) {
		u32 reg = idx | (i << 20);

400
		if (!nb->l3_cache.subcaches[i])
401 402
			continue;

403
		pci_write_config_dword(nb->misc, 0x1BC + slot * 4, reg);
404 405 406 407 408 409 410 411 412

		/*
		 * We need to WBINVD on a core on the node containing the L3
		 * cache which indices we disable therefore a simple wbinvd()
		 * is not sufficient.
		 */
		wbinvd_on_cpu(cpu);

		reg |= BIT(31);
413
		pci_write_config_dword(nb->misc, 0x1BC + slot * 4, reg);
414 415 416
	}
}

417 418 419 420 421 422 423 424 425 426
/*
 * disable a L3 cache index by using a disable-slot
 *
 * @l3:    L3 cache descriptor
 * @cpu:   A CPU on the node containing the L3 cache
 * @slot:  slot number (0..1)
 * @index: index to disable
 *
 * @return: 0 on success, error status on failure
 */
427
int amd_set_l3_disable_slot(struct amd_northbridge *nb, int cpu, unsigned slot,
428
			    unsigned long index)
429
{
430
	int ret = 0;
431

432
	/*  check if @slot is already used or the index is already disabled */
433
	ret = amd_get_l3_disable_slot(nb, slot);
434
	if (ret >= 0)
435
		return -EEXIST;
436

437
	if (index > nb->l3_cache.indices)
438 439
		return -EINVAL;

440
	/* check whether the other slot has disabled the same index already */
441
	if (index == amd_get_l3_disable_slot(nb, !slot))
442
		return -EEXIST;
443

444
	amd_l3_disable_index(nb, cpu, slot, index);
445 446 447 448 449 450 451 452 453 454 455

	return 0;
}

static ssize_t store_cache_disable(struct _cpuid4_info *this_leaf,
				  const char *buf, size_t count,
				  unsigned int slot)
{
	unsigned long val = 0;
	int cpu, err = 0;

456 457 458
	if (!capable(CAP_SYS_ADMIN))
		return -EPERM;

459
	if (!this_leaf->base.nb || !amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE))
460 461
		return -EINVAL;

462
	cpu = cpumask_first(to_cpumask(this_leaf->shared_cpu_map));
463

464
	if (strict_strtoul(buf, 10, &val) < 0)
465 466
		return -EINVAL;

467
	err = amd_set_l3_disable_slot(this_leaf->base.nb, cpu, slot, val);
468 469
	if (err) {
		if (err == -EEXIST)
470 471
			pr_warning("L3 slot %d in use/index already disabled!\n",
				   slot);
472 473
		return err;
	}
474 475 476
	return count;
}

477
#define STORE_CACHE_DISABLE(slot)					\
478
static ssize_t								\
479
store_cache_disable_##slot(struct _cpuid4_info *this_leaf,		\
480 481
			   const char *buf, size_t count,		\
			   unsigned int cpu)				\
482
{									\
483
	return store_cache_disable(this_leaf, buf, count, slot);	\
484
}
485 486 487 488 489 490 491 492
STORE_CACHE_DISABLE(0)
STORE_CACHE_DISABLE(1)

static struct _cache_attr cache_disable_0 = __ATTR(cache_disable_0, 0644,
		show_cache_disable_0, store_cache_disable_0);
static struct _cache_attr cache_disable_1 = __ATTR(cache_disable_1, 0644,
		show_cache_disable_1, store_cache_disable_1);

493 494 495
static ssize_t
show_subcaches(struct _cpuid4_info *this_leaf, char *buf, unsigned int cpu)
{
496
	if (!this_leaf->base.nb || !amd_nb_has_feature(AMD_NB_L3_PARTITIONING))
497 498 499 500 501 502 503 504 505 506 507 508 509 510
		return -EINVAL;

	return sprintf(buf, "%x\n", amd_get_subcaches(cpu));
}

static ssize_t
store_subcaches(struct _cpuid4_info *this_leaf, const char *buf, size_t count,
		unsigned int cpu)
{
	unsigned long val;

	if (!capable(CAP_SYS_ADMIN))
		return -EPERM;

511
	if (!this_leaf->base.nb || !amd_nb_has_feature(AMD_NB_L3_PARTITIONING))
512 513 514 515 516 517 518 519 520 521 522 523 524 525
		return -EINVAL;

	if (strict_strtoul(buf, 16, &val) < 0)
		return -EINVAL;

	if (amd_set_subcaches(cpu, val))
		return -EINVAL;

	return count;
}

static struct _cache_attr subcaches =
	__ATTR(subcaches, 0644, show_subcaches, store_subcaches);

526
#else
527
#define amd_init_l3_cache(x, y)
528
#endif  /* CONFIG_AMD_NB && CONFIG_SYSFS */
529

530
static int
531
cpuid4_cache_lookup_regs(int index, struct _cpuid4_info_regs *this_leaf)
L
Linus Torvalds 已提交
532
{
533 534 535
	union _cpuid4_leaf_eax	eax;
	union _cpuid4_leaf_ebx	ebx;
	union _cpuid4_leaf_ecx	ecx;
536
	unsigned		edx;
L
Linus Torvalds 已提交
537

538
	if (boot_cpu_data.x86_vendor == X86_VENDOR_AMD) {
539 540 541 542 543
		if (cpu_has_topoext)
			cpuid_count(0x8000001d, index, &eax.full,
				    &ebx.full, &ecx.full, &edx);
		else
			amd_cpuid4(index, &eax, &ebx, &ecx);
544
		amd_init_l3_cache(this_leaf, index);
545 546 547 548
	} else {
		cpuid_count(4, index, &eax.full, &ebx.full, &ecx.full, &edx);
	}

549
	if (eax.split.type == CACHE_TYPE_NULL)
550
		return -EIO; /* better error ? */
L
Linus Torvalds 已提交
551

552 553 554
	this_leaf->eax = eax;
	this_leaf->ebx = ebx;
	this_leaf->ecx = ecx;
555 556 557 558
	this_leaf->size = (ecx.split.number_of_sets          + 1) *
			  (ebx.split.coherency_line_size     + 1) *
			  (ebx.split.physical_line_partition + 1) *
			  (ebx.split.ways_of_associativity   + 1);
L
Linus Torvalds 已提交
559 560 561
	return 0;
}

562
static int find_num_cache_leaves(struct cpuinfo_x86 *c)
L
Linus Torvalds 已提交
563
{
564
	unsigned int		eax, ebx, ecx, edx, op;
L
Linus Torvalds 已提交
565
	union _cpuid4_leaf_eax	cache_eax;
566
	int 			i = -1;
L
Linus Torvalds 已提交
567

568 569 570 571 572
	if (c->x86_vendor == X86_VENDOR_AMD)
		op = 0x8000001d;
	else
		op = 4;

573 574
	do {
		++i;
575 576
		/* Do cpuid(op) loop to find out num_cache_leaves */
		cpuid_count(op, i, &eax, &ebx, &ecx, &edx);
L
Linus Torvalds 已提交
577
		cache_eax.full = eax;
578 579
	} while (cache_eax.split.type != CACHE_TYPE_NULL);
	return i;
L
Linus Torvalds 已提交
580 581
}

582
void init_amd_cacheinfo(struct cpuinfo_x86 *c)
583 584 585 586 587 588 589 590 591 592 593 594
{

	if (cpu_has_topoext) {
		num_cache_leaves = find_num_cache_leaves(c);
	} else if (c->extended_cpuid_level >= 0x80000006) {
		if (cpuid_edx(0x80000006) & 0xf000)
			num_cache_leaves = 4;
		else
			num_cache_leaves = 3;
	}
}

595
unsigned int init_intel_cacheinfo(struct cpuinfo_x86 *c)
L
Linus Torvalds 已提交
596
{
A
Alan Cox 已提交
597 598
	/* Cache sizes */
	unsigned int trace = 0, l1i = 0, l1d = 0, l2 = 0, l3 = 0;
L
Linus Torvalds 已提交
599 600
	unsigned int new_l1d = 0, new_l1i = 0; /* Cache sizes from cpuid(4) */
	unsigned int new_l2 = 0, new_l3 = 0, i; /* Cache sizes from cpuid(4) */
601
	unsigned int l2_id = 0, l3_id = 0, num_threads_sharing, index_msb;
602
#ifdef CONFIG_X86_HT
603
	unsigned int cpu = c->cpu_index;
604
#endif
L
Linus Torvalds 已提交
605

606
	if (c->cpuid_level > 3) {
L
Linus Torvalds 已提交
607 608 609 610
		static int is_initialized;

		if (is_initialized == 0) {
			/* Init num_cache_leaves from boot CPU */
611
			num_cache_leaves = find_num_cache_leaves(c);
L
Linus Torvalds 已提交
612 613 614 615 616 617 618 619
			is_initialized++;
		}

		/*
		 * Whenever possible use cpuid(4), deterministic cache
		 * parameters cpuid leaf to find the cache details
		 */
		for (i = 0; i < num_cache_leaves; i++) {
620
			struct _cpuid4_info_regs this_leaf = {};
L
Linus Torvalds 已提交
621 622
			int retval;

623
			retval = cpuid4_cache_lookup_regs(i, &this_leaf);
624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647
			if (retval < 0)
				continue;

			switch (this_leaf.eax.split.level) {
			case 1:
				if (this_leaf.eax.split.type == CACHE_TYPE_DATA)
					new_l1d = this_leaf.size/1024;
				else if (this_leaf.eax.split.type == CACHE_TYPE_INST)
					new_l1i = this_leaf.size/1024;
				break;
			case 2:
				new_l2 = this_leaf.size/1024;
				num_threads_sharing = 1 + this_leaf.eax.split.num_threads_sharing;
				index_msb = get_count_order(num_threads_sharing);
				l2_id = c->apicid & ~((1 << index_msb) - 1);
				break;
			case 3:
				new_l3 = this_leaf.size/1024;
				num_threads_sharing = 1 + this_leaf.eax.split.num_threads_sharing;
				index_msb = get_count_order(num_threads_sharing);
				l3_id = c->apicid & ~((1 << index_msb) - 1);
				break;
			default:
				break;
L
Linus Torvalds 已提交
648 649 650
			}
		}
	}
651 652 653 654 655
	/*
	 * Don't use cpuid2 if cpuid4 is supported. For P4, we use cpuid2 for
	 * trace cache
	 */
	if ((num_cache_leaves == 0 || c->x86 == 15) && c->cpuid_level > 1) {
L
Linus Torvalds 已提交
656
		/* supports eax=2  call */
657 658
		int j, n;
		unsigned int regs[4];
L
Linus Torvalds 已提交
659
		unsigned char *dp = (unsigned char *)regs;
660 661 662 663
		int only_trace = 0;

		if (num_cache_leaves != 0 && c->x86 == 15)
			only_trace = 1;
L
Linus Torvalds 已提交
664 665 666 667

		/* Number of times to iterate */
		n = cpuid_eax(2) & 0xFF;

A
Alan Cox 已提交
668
		for (i = 0 ; i < n ; i++) {
L
Linus Torvalds 已提交
669 670 671
			cpuid(2, &regs[0], &regs[1], &regs[2], &regs[3]);

			/* If bit 31 is set, this is an unknown format */
A
Alan Cox 已提交
672 673 674
			for (j = 0 ; j < 3 ; j++)
				if (regs[j] & (1 << 31))
					regs[j] = 0;
L
Linus Torvalds 已提交
675 676

			/* Byte 0 is level count, not a descriptor */
A
Alan Cox 已提交
677
			for (j = 1 ; j < 16 ; j++) {
L
Linus Torvalds 已提交
678 679 680 681
				unsigned char des = dp[j];
				unsigned char k = 0;

				/* look up this descriptor in the table */
A
Alan Cox 已提交
682
				while (cache_table[k].descriptor != 0) {
L
Linus Torvalds 已提交
683
					if (cache_table[k].descriptor == des) {
684 685
						if (only_trace && cache_table[k].cache_type != LVL_TRACE)
							break;
L
Linus Torvalds 已提交
686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710
						switch (cache_table[k].cache_type) {
						case LVL_1_INST:
							l1i += cache_table[k].size;
							break;
						case LVL_1_DATA:
							l1d += cache_table[k].size;
							break;
						case LVL_2:
							l2 += cache_table[k].size;
							break;
						case LVL_3:
							l3 += cache_table[k].size;
							break;
						case LVL_TRACE:
							trace += cache_table[k].size;
							break;
						}

						break;
					}

					k++;
				}
			}
		}
711
	}
L
Linus Torvalds 已提交
712

713 714
	if (new_l1d)
		l1d = new_l1d;
L
Linus Torvalds 已提交
715

716 717
	if (new_l1i)
		l1i = new_l1i;
L
Linus Torvalds 已提交
718

719 720
	if (new_l2) {
		l2 = new_l2;
721
#ifdef CONFIG_X86_HT
722
		per_cpu(cpu_llc_id, cpu) = l2_id;
723
#endif
724
	}
L
Linus Torvalds 已提交
725

726 727
	if (new_l3) {
		l3 = new_l3;
728
#ifdef CONFIG_X86_HT
729
		per_cpu(cpu_llc_id, cpu) = l3_id;
730
#endif
L
Linus Torvalds 已提交
731 732
	}

733 734 735 736 737 738 739 740 741 742 743 744
#ifdef CONFIG_X86_HT
	/*
	 * If cpu_llc_id is not yet set, this means cpuid_level < 4 which in
	 * turns means that the only possibility is SMT (as indicated in
	 * cpuid1). Since cpuid2 doesn't specify shared caches, and we know
	 * that SMT shares all caches, we can unconditionally set cpu_llc_id to
	 * c->phys_proc_id.
	 */
	if (per_cpu(cpu_llc_id, cpu) == BAD_APICID)
		per_cpu(cpu_llc_id, cpu) = c->phys_proc_id;
#endif

745 746
	c->x86_cache_size = l3 ? l3 : (l2 ? l2 : (l1i+l1d));

L
Linus Torvalds 已提交
747 748 749
	return l2;
}

750 751
#ifdef CONFIG_SYSFS

L
Linus Torvalds 已提交
752
/* pointer to _cpuid4_info array (for each cache leaf) */
753 754
static DEFINE_PER_CPU(struct _cpuid4_info *, ici_cpuid4_info);
#define CPUID4_INFO_IDX(x, y)	(&((per_cpu(ici_cpuid4_info, x))[y]))
L
Linus Torvalds 已提交
755 756

#ifdef CONFIG_SMP
757

758
static int cache_shared_amd_cpu_map_setup(unsigned int cpu, int index)
L
Linus Torvalds 已提交
759
{
760
	struct _cpuid4_info *this_leaf;
761
	int i, sibling;
L
Linus Torvalds 已提交
762

763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778
	if (cpu_has_topoext) {
		unsigned int apicid, nshared, first, last;

		if (!per_cpu(ici_cpuid4_info, cpu))
			return 0;

		this_leaf = CPUID4_INFO_IDX(cpu, index);
		nshared = this_leaf->base.eax.split.num_threads_sharing + 1;
		apicid = cpu_data(cpu).apicid;
		first = apicid - (apicid % nshared);
		last = first + nshared - 1;

		for_each_online_cpu(i) {
			apicid = cpu_data(i).apicid;
			if ((apicid < first) || (apicid > last))
				continue;
779
			if (!per_cpu(ici_cpuid4_info, i))
780 781
				continue;
			this_leaf = CPUID4_INFO_IDX(i, index);
782 783 784 785

			for_each_online_cpu(sibling) {
				apicid = cpu_data(sibling).apicid;
				if ((apicid < first) || (apicid > last))
786 787 788
					continue;
				set_bit(sibling, this_leaf->shared_cpu_map);
			}
789
		}
790 791
	} else if (index == 3) {
		for_each_cpu(i, cpu_llc_shared_mask(cpu)) {
792 793 794
			if (!per_cpu(ici_cpuid4_info, i))
				continue;
			this_leaf = CPUID4_INFO_IDX(i, index);
795
			for_each_cpu(sibling, cpu_llc_shared_mask(cpu)) {
796 797 798 799 800
				if (!cpu_online(sibling))
					continue;
				set_bit(sibling, this_leaf->shared_cpu_map);
			}
		}
801 802
	} else
		return 0;
803

804
	return 1;
805 806
}

807
static void cache_shared_cpu_map_setup(unsigned int cpu, int index)
808 809 810 811 812 813 814 815 816 817 818
{
	struct _cpuid4_info *this_leaf, *sibling_leaf;
	unsigned long num_threads_sharing;
	int index_msb, i;
	struct cpuinfo_x86 *c = &cpu_data(cpu);

	if (c->x86_vendor == X86_VENDOR_AMD) {
		if (cache_shared_amd_cpu_map_setup(cpu, index))
			return;
	}

L
Linus Torvalds 已提交
819
	this_leaf = CPUID4_INFO_IDX(cpu, index);
820
	num_threads_sharing = 1 + this_leaf->base.eax.split.num_threads_sharing;
L
Linus Torvalds 已提交
821 822

	if (num_threads_sharing == 1)
823
		cpumask_set_cpu(cpu, to_cpumask(this_leaf->shared_cpu_map));
824 825 826 827
	else {
		index_msb = get_count_order(num_threads_sharing);

		for_each_online_cpu(i) {
828 829
			if (cpu_data(i).apicid >> index_msb ==
			    c->apicid >> index_msb) {
830 831
				cpumask_set_cpu(i,
					to_cpumask(this_leaf->shared_cpu_map));
832
				if (i != cpu && per_cpu(ici_cpuid4_info, i))  {
833 834 835 836
					sibling_leaf =
						CPUID4_INFO_IDX(i, index);
					cpumask_set_cpu(cpu, to_cpumask(
						sibling_leaf->shared_cpu_map));
837 838 839 840 841
				}
			}
		}
	}
}
842
static void cache_remove_shared_cpu_map(unsigned int cpu, int index)
843 844 845 846 847
{
	struct _cpuid4_info	*this_leaf, *sibling_leaf;
	int sibling;

	this_leaf = CPUID4_INFO_IDX(cpu, index);
848
	for_each_cpu(sibling, to_cpumask(this_leaf->shared_cpu_map)) {
849
		sibling_leaf = CPUID4_INFO_IDX(sibling, index);
850 851
		cpumask_clear_cpu(cpu,
				  to_cpumask(sibling_leaf->shared_cpu_map));
852
	}
L
Linus Torvalds 已提交
853 854
}
#else
855
static void cache_shared_cpu_map_setup(unsigned int cpu, int index)
A
Alan Cox 已提交
856 857 858
{
}

859
static void cache_remove_shared_cpu_map(unsigned int cpu, int index)
A
Alan Cox 已提交
860 861
{
}
L
Linus Torvalds 已提交
862 863
#endif

864
static void free_cache_attributes(unsigned int cpu)
L
Linus Torvalds 已提交
865
{
866 867 868 869 870
	int i;

	for (i = 0; i < num_cache_leaves; i++)
		cache_remove_shared_cpu_map(cpu, i);

871 872
	kfree(per_cpu(ici_cpuid4_info, cpu));
	per_cpu(ici_cpuid4_info, cpu) = NULL;
L
Linus Torvalds 已提交
873 874
}

875
static void get_cpu_leaves(void *_retval)
L
Linus Torvalds 已提交
876
{
877
	int j, *retval = _retval, cpu = smp_processor_id();
878

L
Linus Torvalds 已提交
879 880
	/* Do cpuid and store the results */
	for (j = 0; j < num_cache_leaves; j++) {
881 882 883
		struct _cpuid4_info *this_leaf = CPUID4_INFO_IDX(cpu, j);

		*retval = cpuid4_cache_lookup_regs(j, &this_leaf->base);
884
		if (unlikely(*retval < 0)) {
885 886 887 888
			int i;

			for (i = 0; i < j; i++)
				cache_remove_shared_cpu_map(cpu, i);
889
			break;
890
		}
L
Linus Torvalds 已提交
891 892
		cache_shared_cpu_map_setup(cpu, j);
	}
893 894
}

895
static int detect_cache_attributes(unsigned int cpu)
896 897 898 899 900 901
{
	int			retval;

	if (num_cache_leaves == 0)
		return -ENOENT;

902
	per_cpu(ici_cpuid4_info, cpu) = kzalloc(
903
	    sizeof(struct _cpuid4_info) * num_cache_leaves, GFP_KERNEL);
904
	if (per_cpu(ici_cpuid4_info, cpu) == NULL)
905
		return -ENOMEM;
L
Linus Torvalds 已提交
906

907
	smp_call_function_single(cpu, get_cpu_leaves, &retval, true);
908
	if (retval) {
909 910
		kfree(per_cpu(ici_cpuid4_info, cpu));
		per_cpu(ici_cpuid4_info, cpu) = NULL;
911 912
	}

913
	return retval;
L
Linus Torvalds 已提交
914 915 916 917
}

#include <linux/kobject.h>
#include <linux/sysfs.h>
918
#include <linux/cpu.h>
L
Linus Torvalds 已提交
919 920

/* pointer to kobject for cpuX/cache */
921
static DEFINE_PER_CPU(struct kobject *, ici_cache_kobject);
L
Linus Torvalds 已提交
922 923 924 925 926 927 928 929

struct _index_kobject {
	struct kobject kobj;
	unsigned int cpu;
	unsigned short index;
};

/* pointer to array of kobjects for cpuX/cache/indexY */
930 931
static DEFINE_PER_CPU(struct _index_kobject *, ici_index_kobject);
#define INDEX_KOBJECT_PTR(x, y)		(&((per_cpu(ici_index_kobject, x))[y]))
L
Linus Torvalds 已提交
932 933

#define show_one_plus(file_name, object, val)				\
934 935
static ssize_t show_##file_name(struct _cpuid4_info *this_leaf, char *buf, \
				unsigned int cpu)			\
L
Linus Torvalds 已提交
936
{									\
A
Alan Cox 已提交
937
	return sprintf(buf, "%lu\n", (unsigned long)this_leaf->object + val); \
L
Linus Torvalds 已提交
938 939
}

940 941 942 943 944
show_one_plus(level, base.eax.split.level, 0);
show_one_plus(coherency_line_size, base.ebx.split.coherency_line_size, 1);
show_one_plus(physical_line_partition, base.ebx.split.physical_line_partition, 1);
show_one_plus(ways_of_associativity, base.ebx.split.ways_of_associativity, 1);
show_one_plus(number_of_sets, base.ecx.split.number_of_sets, 1);
L
Linus Torvalds 已提交
945

946 947
static ssize_t show_size(struct _cpuid4_info *this_leaf, char *buf,
			 unsigned int cpu)
L
Linus Torvalds 已提交
948
{
949
	return sprintf(buf, "%luK\n", this_leaf->base.size / 1024);
L
Linus Torvalds 已提交
950 951
}

952 953
static ssize_t show_shared_cpu_map_func(struct _cpuid4_info *this_leaf,
					int type, char *buf)
L
Linus Torvalds 已提交
954
{
955
	ptrdiff_t len = PTR_ALIGN(buf + PAGE_SIZE - 1, PAGE_SIZE) - buf;
956 957
	int n = 0;

958
	if (len > 1) {
959
		const struct cpumask *mask;
960

961
		mask = to_cpumask(this_leaf->shared_cpu_map);
A
Alan Cox 已提交
962
		n = type ?
963 964
			cpulist_scnprintf(buf, len-2, mask) :
			cpumask_scnprintf(buf, len-2, mask);
965 966
		buf[n++] = '\n';
		buf[n] = '\0';
967 968
	}
	return n;
L
Linus Torvalds 已提交
969 970
}

971 972
static inline ssize_t show_shared_cpu_map(struct _cpuid4_info *leaf, char *buf,
					  unsigned int cpu)
973 974 975 976
{
	return show_shared_cpu_map_func(leaf, 0, buf);
}

977 978
static inline ssize_t show_shared_cpu_list(struct _cpuid4_info *leaf, char *buf,
					   unsigned int cpu)
979 980 981 982
{
	return show_shared_cpu_map_func(leaf, 1, buf);
}

983 984
static ssize_t show_type(struct _cpuid4_info *this_leaf, char *buf,
			 unsigned int cpu)
985
{
986
	switch (this_leaf->base.eax.split.type) {
987
	case CACHE_TYPE_DATA:
L
Linus Torvalds 已提交
988
		return sprintf(buf, "Data\n");
989
	case CACHE_TYPE_INST:
L
Linus Torvalds 已提交
990
		return sprintf(buf, "Instruction\n");
991
	case CACHE_TYPE_UNIFIED:
L
Linus Torvalds 已提交
992
		return sprintf(buf, "Unified\n");
993
	default:
L
Linus Torvalds 已提交
994 995 996 997
		return sprintf(buf, "Unknown\n");
	}
}

998 999
#define to_object(k)	container_of(k, struct _index_kobject, kobj)
#define to_attr(a)	container_of(a, struct _cache_attr, attr)
1000

L
Linus Torvalds 已提交
1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012
#define define_one_ro(_name) \
static struct _cache_attr _name = \
	__ATTR(_name, 0444, show_##_name, NULL)

define_one_ro(level);
define_one_ro(type);
define_one_ro(coherency_line_size);
define_one_ro(physical_line_partition);
define_one_ro(ways_of_associativity);
define_one_ro(number_of_sets);
define_one_ro(size);
define_one_ro(shared_cpu_map);
1013
define_one_ro(shared_cpu_list);
L
Linus Torvalds 已提交
1014

A
Alan Cox 已提交
1015
static struct attribute *default_attrs[] = {
1016 1017 1018 1019 1020 1021 1022 1023 1024
	&type.attr,
	&level.attr,
	&coherency_line_size.attr,
	&physical_line_partition.attr,
	&ways_of_associativity.attr,
	&number_of_sets.attr,
	&size.attr,
	&shared_cpu_map.attr,
	&shared_cpu_list.attr,
1025 1026 1027
	NULL
};

1028
#ifdef CONFIG_AMD_NB
1029
static struct attribute **amd_l3_attrs(void)
1030 1031 1032 1033 1034 1035 1036
{
	static struct attribute **attrs;
	int n;

	if (attrs)
		return attrs;

1037
	n = ARRAY_SIZE(default_attrs);
1038 1039 1040 1041

	if (amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE))
		n += 2;

1042 1043 1044
	if (amd_nb_has_feature(AMD_NB_L3_PARTITIONING))
		n += 1;

1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056
	attrs = kzalloc(n * sizeof (struct attribute *), GFP_KERNEL);
	if (attrs == NULL)
		return attrs = default_attrs;

	for (n = 0; default_attrs[n]; n++)
		attrs[n] = default_attrs[n];

	if (amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE)) {
		attrs[n++] = &cache_disable_0.attr;
		attrs[n++] = &cache_disable_1.attr;
	}

1057 1058 1059
	if (amd_nb_has_feature(AMD_NB_L3_PARTITIONING))
		attrs[n++] = &subcaches.attr;

1060 1061
	return attrs;
}
1062
#endif
L
Linus Torvalds 已提交
1063

A
Alan Cox 已提交
1064
static ssize_t show(struct kobject *kobj, struct attribute *attr, char *buf)
L
Linus Torvalds 已提交
1065 1066 1067 1068 1069 1070 1071
{
	struct _cache_attr *fattr = to_attr(attr);
	struct _index_kobject *this_leaf = to_object(kobj);
	ssize_t ret;

	ret = fattr->show ?
		fattr->show(CPUID4_INFO_IDX(this_leaf->cpu, this_leaf->index),
1072
			buf, this_leaf->cpu) :
1073
		0;
L
Linus Torvalds 已提交
1074 1075 1076
	return ret;
}

A
Alan Cox 已提交
1077 1078
static ssize_t store(struct kobject *kobj, struct attribute *attr,
		     const char *buf, size_t count)
L
Linus Torvalds 已提交
1079
{
1080 1081 1082 1083
	struct _cache_attr *fattr = to_attr(attr);
	struct _index_kobject *this_leaf = to_object(kobj);
	ssize_t ret;

1084 1085
	ret = fattr->store ?
		fattr->store(CPUID4_INFO_IDX(this_leaf->cpu, this_leaf->index),
1086
			buf, count, this_leaf->cpu) :
1087 1088
		0;
	return ret;
L
Linus Torvalds 已提交
1089 1090
}

1091
static const struct sysfs_ops sysfs_ops = {
L
Linus Torvalds 已提交
1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104
	.show   = show,
	.store  = store,
};

static struct kobj_type ktype_cache = {
	.sysfs_ops	= &sysfs_ops,
	.default_attrs	= default_attrs,
};

static struct kobj_type ktype_percpu_entry = {
	.sysfs_ops	= &sysfs_ops,
};

1105
static void cpuid4_cache_sysfs_exit(unsigned int cpu)
L
Linus Torvalds 已提交
1106
{
1107 1108 1109 1110
	kfree(per_cpu(ici_cache_kobject, cpu));
	kfree(per_cpu(ici_index_kobject, cpu));
	per_cpu(ici_cache_kobject, cpu) = NULL;
	per_cpu(ici_index_kobject, cpu) = NULL;
L
Linus Torvalds 已提交
1111 1112 1113
	free_cache_attributes(cpu);
}

1114
static int cpuid4_cache_sysfs_init(unsigned int cpu)
L
Linus Torvalds 已提交
1115
{
1116
	int err;
L
Linus Torvalds 已提交
1117 1118 1119 1120

	if (num_cache_leaves == 0)
		return -ENOENT;

1121 1122 1123
	err = detect_cache_attributes(cpu);
	if (err)
		return err;
L
Linus Torvalds 已提交
1124 1125

	/* Allocate all required memory */
1126
	per_cpu(ici_cache_kobject, cpu) =
1127
		kzalloc(sizeof(struct kobject), GFP_KERNEL);
1128
	if (unlikely(per_cpu(ici_cache_kobject, cpu) == NULL))
L
Linus Torvalds 已提交
1129 1130
		goto err_out;

1131
	per_cpu(ici_index_kobject, cpu) = kzalloc(
A
Alan Cox 已提交
1132
	    sizeof(struct _index_kobject) * num_cache_leaves, GFP_KERNEL);
1133
	if (unlikely(per_cpu(ici_index_kobject, cpu) == NULL))
L
Linus Torvalds 已提交
1134 1135 1136 1137 1138 1139 1140 1141 1142
		goto err_out;

	return 0;

err_out:
	cpuid4_cache_sysfs_exit(cpu);
	return -ENOMEM;
}

1143
static DECLARE_BITMAP(cache_dev_map, NR_CPUS);
1144

L
Linus Torvalds 已提交
1145
/* Add/Remove cache interface for CPU device */
1146
static int cache_add_dev(struct device *dev)
L
Linus Torvalds 已提交
1147
{
1148
	unsigned int cpu = dev->id;
L
Linus Torvalds 已提交
1149 1150
	unsigned long i, j;
	struct _index_kobject *this_object;
1151
	struct _cpuid4_info   *this_leaf;
1152
	int retval;
L
Linus Torvalds 已提交
1153 1154 1155 1156 1157

	retval = cpuid4_cache_sysfs_init(cpu);
	if (unlikely(retval < 0))
		return retval;

1158
	retval = kobject_init_and_add(per_cpu(ici_cache_kobject, cpu),
1159
				      &ktype_percpu_entry,
1160
				      &dev->kobj, "%s", "cache");
1161 1162 1163 1164
	if (retval < 0) {
		cpuid4_cache_sysfs_exit(cpu);
		return retval;
	}
L
Linus Torvalds 已提交
1165 1166

	for (i = 0; i < num_cache_leaves; i++) {
A
Alan Cox 已提交
1167
		this_object = INDEX_KOBJECT_PTR(cpu, i);
L
Linus Torvalds 已提交
1168 1169
		this_object->cpu = cpu;
		this_object->index = i;
1170 1171 1172

		this_leaf = CPUID4_INFO_IDX(cpu, i);

1173 1174
		ktype_cache.default_attrs = default_attrs;
#ifdef CONFIG_AMD_NB
1175
		if (this_leaf->base.nb)
1176 1177
			ktype_cache.default_attrs = amd_l3_attrs();
#endif
1178
		retval = kobject_init_and_add(&(this_object->kobj),
1179
					      &ktype_cache,
1180
					      per_cpu(ici_cache_kobject, cpu),
1181
					      "index%1lu", i);
L
Linus Torvalds 已提交
1182
		if (unlikely(retval)) {
A
Alan Cox 已提交
1183 1184
			for (j = 0; j < i; j++)
				kobject_put(&(INDEX_KOBJECT_PTR(cpu, j)->kobj));
1185
			kobject_put(per_cpu(ici_cache_kobject, cpu));
L
Linus Torvalds 已提交
1186
			cpuid4_cache_sysfs_exit(cpu);
1187
			return retval;
L
Linus Torvalds 已提交
1188
		}
1189
		kobject_uevent(&(this_object->kobj), KOBJ_ADD);
L
Linus Torvalds 已提交
1190
	}
1191
	cpumask_set_cpu(cpu, to_cpumask(cache_dev_map));
1192

1193
	kobject_uevent(per_cpu(ici_cache_kobject, cpu), KOBJ_ADD);
1194
	return 0;
L
Linus Torvalds 已提交
1195 1196
}

1197
static void cache_remove_dev(struct device *dev)
L
Linus Torvalds 已提交
1198
{
1199
	unsigned int cpu = dev->id;
L
Linus Torvalds 已提交
1200 1201
	unsigned long i;

1202
	if (per_cpu(ici_cpuid4_info, cpu) == NULL)
1203
		return;
1204
	if (!cpumask_test_cpu(cpu, to_cpumask(cache_dev_map)))
1205
		return;
1206
	cpumask_clear_cpu(cpu, to_cpumask(cache_dev_map));
1207 1208

	for (i = 0; i < num_cache_leaves; i++)
A
Alan Cox 已提交
1209
		kobject_put(&(INDEX_KOBJECT_PTR(cpu, i)->kobj));
1210
	kobject_put(per_cpu(ici_cache_kobject, cpu));
L
Linus Torvalds 已提交
1211
	cpuid4_cache_sysfs_exit(cpu);
1212 1213
}

1214 1215
static int cacheinfo_cpu_callback(struct notifier_block *nfb,
				  unsigned long action, void *hcpu)
1216 1217
{
	unsigned int cpu = (unsigned long)hcpu;
1218
	struct device *dev;
1219

1220
	dev = get_cpu_device(cpu);
1221 1222
	switch (action) {
	case CPU_ONLINE:
1223
	case CPU_ONLINE_FROZEN:
1224
		cache_add_dev(dev);
1225 1226
		break;
	case CPU_DEAD:
1227
	case CPU_DEAD_FROZEN:
1228
		cache_remove_dev(dev);
1229 1230 1231
		break;
	}
	return NOTIFY_OK;
L
Linus Torvalds 已提交
1232 1233
}

1234
static struct notifier_block cacheinfo_cpu_notifier = {
1235
	.notifier_call = cacheinfo_cpu_callback,
L
Linus Torvalds 已提交
1236 1237
};

1238
static int __init cache_sysfs_init(void)
L
Linus Torvalds 已提交
1239
{
1240
	int i, err = 0;
1241

L
Linus Torvalds 已提交
1242 1243 1244
	if (num_cache_leaves == 0)
		return 0;

1245
	cpu_notifier_register_begin();
1246
	for_each_online_cpu(i) {
1247
		struct device *dev = get_cpu_device(i);
1248

1249
		err = cache_add_dev(dev);
1250
		if (err)
1251
			goto out;
1252
	}
1253 1254 1255 1256 1257
	__register_hotcpu_notifier(&cacheinfo_cpu_notifier);

out:
	cpu_notifier_register_done();
	return err;
L
Linus Torvalds 已提交
1258 1259
}

1260
device_initcall(cache_sysfs_init);
L
Linus Torvalds 已提交
1261 1262

#endif