cpufeature.h 10.7 KB
Newer Older
H
H. Peter Anvin 已提交
1 2
#ifndef _ASM_X86_CPUFEATURE_H
#define _ASM_X86_CPUFEATURE_H
H
H. Peter Anvin 已提交
3

4
#include <asm/processor.h>
5

6 7
#if defined(__KERNEL__) && !defined(__ASSEMBLY__)

8
#include <asm/asm.h>
9 10
#include <linux/bitops.h>

11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30
enum cpuid_leafs
{
	CPUID_1_EDX		= 0,
	CPUID_8000_0001_EDX,
	CPUID_8086_0001_EDX,
	CPUID_LNX_1,
	CPUID_1_ECX,
	CPUID_C000_0001_EDX,
	CPUID_8000_0001_ECX,
	CPUID_LNX_2,
	CPUID_LNX_3,
	CPUID_7_0_EBX,
	CPUID_D_1_EAX,
	CPUID_F_0_EDX,
	CPUID_F_1_EDX,
	CPUID_8000_0008_EBX,
	CPUID_6_EAX,
	CPUID_8000_000A_EDX,
};

31
#ifdef CONFIG_X86_FEATURE_NAMES
32 33
extern const char * const x86_cap_flags[NCAPINTS*32];
extern const char * const x86_power_flags[32];
34 35 36 37 38 39
#define X86_CAP_FMT "%s"
#define x86_cap_flag(flag) x86_cap_flags[flag]
#else
#define X86_CAP_FMT "%d:%d"
#define x86_cap_flag(flag) ((flag) >> 5), ((flag) & 31)
#endif
40

41 42 43 44 45 46
/*
 * In order to save room, we index into this array by doing
 * X86_BUG_<name> - NCAPINTS*32.
 */
extern const char * const x86_bug_flags[NBUGINTS*32];

47 48 49
#define test_cpu_cap(c, bit)						\
	 test_bit(bit, (unsigned long *)((c)->x86_capability))

50
#define REQUIRED_MASK_BIT_SET(bit)					\
H
H. Peter Anvin 已提交
51 52 53 54 55 56 57
	 ( (((bit)>>5)==0 && (1UL<<((bit)&31) & REQUIRED_MASK0)) ||	\
	   (((bit)>>5)==1 && (1UL<<((bit)&31) & REQUIRED_MASK1)) ||	\
	   (((bit)>>5)==2 && (1UL<<((bit)&31) & REQUIRED_MASK2)) ||	\
	   (((bit)>>5)==3 && (1UL<<((bit)&31) & REQUIRED_MASK3)) ||	\
	   (((bit)>>5)==4 && (1UL<<((bit)&31) & REQUIRED_MASK4)) ||	\
	   (((bit)>>5)==5 && (1UL<<((bit)&31) & REQUIRED_MASK5)) ||	\
	   (((bit)>>5)==6 && (1UL<<((bit)&31) & REQUIRED_MASK6)) ||	\
58 59
	   (((bit)>>5)==7 && (1UL<<((bit)&31) & REQUIRED_MASK7)) ||	\
	   (((bit)>>5)==8 && (1UL<<((bit)&31) & REQUIRED_MASK8)) ||	\
60 61
	   (((bit)>>5)==9 && (1UL<<((bit)&31) & REQUIRED_MASK9)) )

D
Dave Hansen 已提交
62 63 64 65 66 67 68 69 70 71 72 73
#define DISABLED_MASK_BIT_SET(bit)					\
	 ( (((bit)>>5)==0 && (1UL<<((bit)&31) & DISABLED_MASK0)) ||	\
	   (((bit)>>5)==1 && (1UL<<((bit)&31) & DISABLED_MASK1)) ||	\
	   (((bit)>>5)==2 && (1UL<<((bit)&31) & DISABLED_MASK2)) ||	\
	   (((bit)>>5)==3 && (1UL<<((bit)&31) & DISABLED_MASK3)) ||	\
	   (((bit)>>5)==4 && (1UL<<((bit)&31) & DISABLED_MASK4)) ||	\
	   (((bit)>>5)==5 && (1UL<<((bit)&31) & DISABLED_MASK5)) ||	\
	   (((bit)>>5)==6 && (1UL<<((bit)&31) & DISABLED_MASK6)) ||	\
	   (((bit)>>5)==7 && (1UL<<((bit)&31) & DISABLED_MASK7)) ||	\
	   (((bit)>>5)==8 && (1UL<<((bit)&31) & DISABLED_MASK8)) ||	\
	   (((bit)>>5)==9 && (1UL<<((bit)&31) & DISABLED_MASK9)) )

74 75
#define cpu_has(c, bit)							\
	(__builtin_constant_p(bit) && REQUIRED_MASK_BIT_SET(bit) ? 1 :	\
76 77
	 test_cpu_cap(c, bit))

78 79 80 81
#define this_cpu_has(bit)						\
	(__builtin_constant_p(bit) && REQUIRED_MASK_BIT_SET(bit) ? 1 : 	\
	 x86_this_cpu_test_bit(bit, (unsigned long *)&cpu_info.x86_capability))

D
Dave Hansen 已提交
82 83 84 85 86 87 88 89 90 91 92 93
/*
 * This macro is for detection of features which need kernel
 * infrastructure to be used.  It may *not* directly test the CPU
 * itself.  Use the cpu_has() family if you want true runtime
 * testing of CPU features, like in hypervisor code where you are
 * supporting a possible guest feature where host support for it
 * is not relevant.
 */
#define cpu_feature_enabled(bit)	\
	(__builtin_constant_p(bit) && DISABLED_MASK_BIT_SET(bit) ? 0 :	\
	 cpu_has(&boot_cpu_data, bit))

H
H. Peter Anvin 已提交
94 95
#define boot_cpu_has(bit)	cpu_has(&boot_cpu_data, bit)

96 97
#define set_cpu_cap(c, bit)	set_bit(bit, (unsigned long *)((c)->x86_capability))
#define clear_cpu_cap(c, bit)	clear_bit(bit, (unsigned long *)((c)->x86_capability))
98 99
#define setup_clear_cpu_cap(bit) do { \
	clear_cpu_cap(&boot_cpu_data, bit);	\
100
	set_bit(bit, (unsigned long *)cpu_caps_cleared); \
101
} while (0)
102 103
#define setup_force_cpu_cap(bit) do { \
	set_cpu_cap(&boot_cpu_data, bit);	\
104
	set_bit(bit, (unsigned long *)cpu_caps_set);	\
105
} while (0)
106

H
H. Peter Anvin 已提交
107 108 109 110 111 112 113 114
#define cpu_has_fpu		boot_cpu_has(X86_FEATURE_FPU)
#define cpu_has_pse		boot_cpu_has(X86_FEATURE_PSE)
#define cpu_has_tsc		boot_cpu_has(X86_FEATURE_TSC)
#define cpu_has_pge		boot_cpu_has(X86_FEATURE_PGE)
#define cpu_has_apic		boot_cpu_has(X86_FEATURE_APIC)
#define cpu_has_fxsr		boot_cpu_has(X86_FEATURE_FXSR)
#define cpu_has_xmm		boot_cpu_has(X86_FEATURE_XMM)
#define cpu_has_xmm2		boot_cpu_has(X86_FEATURE_XMM2)
115
#define cpu_has_aes		boot_cpu_has(X86_FEATURE_AES)
116
#define cpu_has_avx		boot_cpu_has(X86_FEATURE_AVX)
117
#define cpu_has_avx2		boot_cpu_has(X86_FEATURE_AVX2)
118
#define cpu_has_clflush		boot_cpu_has(X86_FEATURE_CLFLUSH)
119
#define cpu_has_gbpages		boot_cpu_has(X86_FEATURE_GBPAGES)
S
stephane eranian 已提交
120
#define cpu_has_arch_perfmon	boot_cpu_has(X86_FEATURE_ARCH_PERFMON)
121
#define cpu_has_pat		boot_cpu_has(X86_FEATURE_PAT)
122
#define cpu_has_x2apic		boot_cpu_has(X86_FEATURE_X2APIC)
123
#define cpu_has_xsave		boot_cpu_has(X86_FEATURE_XSAVE)
124
#define cpu_has_xsaves		boot_cpu_has(X86_FEATURE_XSAVES)
125
#define cpu_has_osxsave		boot_cpu_has(X86_FEATURE_OSXSAVE)
126
#define cpu_has_hypervisor	boot_cpu_has(X86_FEATURE_HYPERVISOR)
127 128 129 130
/*
 * Do not add any more of those clumsy macros - use static_cpu_has_safe() for
 * fast paths and boot_cpu_has() otherwise!
 */
H
H. Peter Anvin 已提交
131

132
#if __GNUC__ >= 4 && defined(CONFIG_X86_FAST_FEATURE_TESTS)
133
extern void warn_pre_alternatives(void);
134
extern bool __static_cpu_has_safe(u16 bit);
135

136 137 138 139 140
/*
 * Static testing of CPU features.  Used the same as boot_cpu_has().
 * These are only valid after alternatives have run, but will statically
 * patch the target code for additional performance.
 */
141
static __always_inline __pure bool __static_cpu_has(u16 bit)
142
{
143
#ifdef CC_HAVE_ASM_GOTO
144 145

#ifdef CONFIG_X86_DEBUG_STATIC_CPU_HAS
146

147 148 149 150
		/*
		 * Catch too early usage of this before alternatives
		 * have run.
		 */
151
		asm_volatile_goto("1: jmp %l[t_warn]\n"
152 153 154 155 156 157 158
			 "2:\n"
			 ".section .altinstructions,\"a\"\n"
			 " .long 1b - .\n"
			 " .long 0\n"		/* no replacement */
			 " .word %P0\n"		/* 1: do replace */
			 " .byte 2b - 1b\n"	/* source len */
			 " .byte 0\n"		/* replacement len */
159
			 " .byte 0\n"		/* pad len */
160 161 162
			 ".previous\n"
			 /* skipping size check since replacement size = 0 */
			 : : "i" (X86_FEATURE_ALWAYS) : : t_warn);
163

164 165
#endif

166
		asm_volatile_goto("1: jmp %l[t_no]\n"
167 168
			 "2:\n"
			 ".section .altinstructions,\"a\"\n"
169 170
			 " .long 1b - .\n"
			 " .long 0\n"		/* no replacement */
171
			 " .word %P0\n"		/* feature bit */
172 173
			 " .byte 2b - 1b\n"	/* source len */
			 " .byte 0\n"		/* replacement len */
174
			 " .byte 0\n"		/* pad len */
175
			 ".previous\n"
176
			 /* skipping size check since replacement size = 0 */
177 178 179 180
			 : : "i" (bit) : : t_no);
		return true;
	t_no:
		return false;
181 182 183 184 185 186

#ifdef CONFIG_X86_DEBUG_STATIC_CPU_HAS
	t_warn:
		warn_pre_alternatives();
		return false;
#endif
187 188 189

#else /* CC_HAVE_ASM_GOTO */

190 191 192 193 194
		u8 flag;
		/* Open-coded due to __stringify() in ALTERNATIVE() */
		asm volatile("1: movb $0,%0\n"
			     "2:\n"
			     ".section .altinstructions,\"a\"\n"
195 196
			     " .long 1b - .\n"
			     " .long 3f - .\n"
197
			     " .word %P1\n"		/* feature bit */
198 199
			     " .byte 2b - 1b\n"		/* source len */
			     " .byte 4f - 3f\n"		/* replacement len */
200
			     " .byte 0\n"		/* pad len */
201 202 203
			     ".previous\n"
			     ".section .discard,\"aw\",@progbits\n"
			     " .byte 0xff + (4f-3f) - (2b-1b)\n" /* size check */
204 205 206 207 208 209 210
			     ".previous\n"
			     ".section .altinstr_replacement,\"ax\"\n"
			     "3: movb $1,%0\n"
			     "4:\n"
			     ".previous\n"
			     : "=qm" (flag) : "i" (bit));
		return flag;
211 212

#endif /* CC_HAVE_ASM_GOTO */
213 214 215 216 217 218
}

#define static_cpu_has(bit)					\
(								\
	__builtin_constant_p(boot_cpu_has(bit)) ?		\
		boot_cpu_has(bit) :				\
219
	__builtin_constant_p(bit) ?				\
220 221 222
		__static_cpu_has(bit) :				\
		boot_cpu_has(bit)				\
)
223 224 225

static __always_inline __pure bool _static_cpu_has_safe(u16 bit)
{
226
#ifdef CC_HAVE_ASM_GOTO
227
		asm_volatile_goto("1: jmp %l[t_dynamic]\n"
228
			 "2:\n"
229 230 231
			 ".skip -(((5f-4f) - (2b-1b)) > 0) * "
			         "((5f-4f) - (2b-1b)),0x90\n"
			 "3:\n"
232 233
			 ".section .altinstructions,\"a\"\n"
			 " .long 1b - .\n"		/* src offset */
234
			 " .long 4f - .\n"		/* repl offset */
235
			 " .word %P1\n"			/* always replace */
236 237 238
			 " .byte 3b - 1b\n"		/* src len */
			 " .byte 5f - 4f\n"		/* repl len */
			 " .byte 3b - 2b\n"		/* pad len */
239 240
			 ".previous\n"
			 ".section .altinstr_replacement,\"ax\"\n"
241
			 "4: jmp %l[t_no]\n"
242
			 "5:\n"
243 244 245 246 247
			 ".previous\n"
			 ".section .altinstructions,\"a\"\n"
			 " .long 1b - .\n"		/* src offset */
			 " .long 0\n"			/* no replacement */
			 " .word %P0\n"			/* feature bit */
248
			 " .byte 3b - 1b\n"		/* src len */
249
			 " .byte 0\n"			/* repl len */
250
			 " .byte 0\n"			/* pad len */
251 252 253 254 255 256 257 258
			 ".previous\n"
			 : : "i" (bit), "i" (X86_FEATURE_ALWAYS)
			 : : t_dynamic, t_no);
		return true;
	t_no:
		return false;
	t_dynamic:
		return __static_cpu_has_safe(bit);
259
#else
260 261 262 263 264 265 266 267 268 269
		u8 flag;
		/* Open-coded due to __stringify() in ALTERNATIVE() */
		asm volatile("1: movb $2,%0\n"
			     "2:\n"
			     ".section .altinstructions,\"a\"\n"
			     " .long 1b - .\n"		/* src offset */
			     " .long 3f - .\n"		/* repl offset */
			     " .word %P2\n"		/* always replace */
			     " .byte 2b - 1b\n"		/* source len */
			     " .byte 4f - 3f\n"		/* replacement len */
270
			     " .byte 0\n"		/* pad len */
271 272 273 274 275 276 277 278 279 280 281 282 283 284
			     ".previous\n"
			     ".section .discard,\"aw\",@progbits\n"
			     " .byte 0xff + (4f-3f) - (2b-1b)\n" /* size check */
			     ".previous\n"
			     ".section .altinstr_replacement,\"ax\"\n"
			     "3: movb $0,%0\n"
			     "4:\n"
			     ".previous\n"
			     ".section .altinstructions,\"a\"\n"
			     " .long 1b - .\n"		/* src offset */
			     " .long 5f - .\n"		/* repl offset */
			     " .word %P1\n"		/* feature bit */
			     " .byte 4b - 3b\n"		/* src len */
			     " .byte 6f - 5f\n"		/* repl len */
285
			     " .byte 0\n"		/* pad len */
286 287 288 289 290 291 292 293 294 295 296
			     ".previous\n"
			     ".section .discard,\"aw\",@progbits\n"
			     " .byte 0xff + (6f-5f) - (4b-3b)\n" /* size check */
			     ".previous\n"
			     ".section .altinstr_replacement,\"ax\"\n"
			     "5: movb $1,%0\n"
			     "6:\n"
			     ".previous\n"
			     : "=qm" (flag)
			     : "i" (bit), "i" (X86_FEATURE_ALWAYS));
		return (flag == 2 ? __static_cpu_has_safe(bit) : flag);
297
#endif /* CC_HAVE_ASM_GOTO */
298 299 300 301 302 303 304 305
}

#define static_cpu_has_safe(bit)				\
(								\
	__builtin_constant_p(boot_cpu_has(bit)) ?		\
		boot_cpu_has(bit) :				\
		_static_cpu_has_safe(bit)			\
)
306 307 308 309
#else
/*
 * gcc 3.x is too stupid to do the static test; fall back to dynamic.
 */
310 311
#define static_cpu_has(bit)		boot_cpu_has(bit)
#define static_cpu_has_safe(bit)	boot_cpu_has(bit)
312
#endif
313

314 315 316
#define cpu_has_bug(c, bit)		cpu_has(c, (bit))
#define set_cpu_bug(c, bit)		set_cpu_cap(c, (bit))
#define clear_cpu_bug(c, bit)		clear_cpu_cap(c, (bit))
317

318 319 320
#define static_cpu_has_bug(bit)		static_cpu_has((bit))
#define static_cpu_has_bug_safe(bit)	static_cpu_has_safe((bit))
#define boot_cpu_has_bug(bit)		cpu_has_bug(&boot_cpu_data, (bit))
321

322 323
#define MAX_CPU_FEATURES		(NCAPINTS * 32)
#define cpu_have_feature		boot_cpu_has
324

325 326 327
#define CPU_FEATURE_TYPEFMT		"x86,ven%04Xfam%04Xmod%04X"
#define CPU_FEATURE_TYPEVAL		boot_cpu_data.x86_vendor, boot_cpu_data.x86, \
					boot_cpu_data.x86_model
328

329
#endif /* defined(__KERNEL__) && !defined(__ASSEMBLY__) */
H
H. Peter Anvin 已提交
330
#endif /* _ASM_X86_CPUFEATURE_H */