xstate.c 33.4 KB
Newer Older
1 2 3 4 5 6
/*
 * xsave/xrstor support.
 *
 * Author: Suresh Siddha <suresh.b.siddha@intel.com>
 */
#include <linux/compat.h>
F
Fenghua Yu 已提交
7
#include <linux/cpu.h>
8
#include <linux/mman.h>
9
#include <linux/pkeys.h>
10

11
#include <asm/fpu/api.h>
12
#include <asm/fpu/internal.h>
13
#include <asm/fpu/signal.h>
14
#include <asm/fpu/regset.h>
15
#include <asm/fpu/xstate.h>
I
Ingo Molnar 已提交
16

A
Andy Lutomirski 已提交
17
#include <asm/tlbflush.h>
18
#include <asm/cpufeature.h>
19

20 21 22 23 24
/*
 * Although we spell it out in here, the Processor Trace
 * xfeature is completely unused.  We use other mechanisms
 * to save/restore PT state in Linux.
 */
25 26 27 28 29 30 31 32 33 34
static const char *xfeature_names[] =
{
	"x87 floating point registers"	,
	"SSE registers"			,
	"AVX registers"			,
	"MPX bounds registers"		,
	"MPX CSR"			,
	"AVX-512 opmask"		,
	"AVX-512 Hi256"			,
	"AVX-512 ZMM_Hi256"		,
35
	"Processor Trace (unused)"	,
36
	"Protection Keys User registers",
37 38 39
	"unknown xstate feature"	,
};

40 41 42 43 44 45 46 47 48 49 50 51 52
static short xsave_cpuid_features[] __initdata = {
	X86_FEATURE_FPU,
	X86_FEATURE_XMM,
	X86_FEATURE_AVX,
	X86_FEATURE_MPX,
	X86_FEATURE_MPX,
	X86_FEATURE_AVX512F,
	X86_FEATURE_AVX512F,
	X86_FEATURE_AVX512F,
	X86_FEATURE_INTEL_PT,
	X86_FEATURE_PKU,
};

53
/*
54
 * Mask of xstate features supported by the CPU and the kernel:
55
 */
56
u64 xfeatures_mask __read_mostly;
57

D
Dave Hansen 已提交
58 59
static unsigned int xstate_offsets[XFEATURE_MAX] = { [ 0 ... XFEATURE_MAX - 1] = -1};
static unsigned int xstate_sizes[XFEATURE_MAX]   = { [ 0 ... XFEATURE_MAX - 1] = -1};
60
static unsigned int xstate_comp_offsets[sizeof(xfeatures_mask)*8];
61

62 63 64 65 66 67 68
/*
 * The XSAVE area of kernel can be in standard or compacted format;
 * it is always in standard format for user mode. This is the user
 * mode standard format size used for signal and ptrace frames.
 */
unsigned int fpu_user_xstate_size;

69 70 71 72 73 74 75 76 77 78 79 80 81
/*
 * Clear all of the X86_FEATURE_* bits that are unavailable
 * when the CPU has no XSAVE support.
 */
void fpu__xstate_clear_all_cpu_caps(void)
{
	setup_clear_cpu_cap(X86_FEATURE_XSAVE);
	setup_clear_cpu_cap(X86_FEATURE_XSAVEOPT);
	setup_clear_cpu_cap(X86_FEATURE_XSAVEC);
	setup_clear_cpu_cap(X86_FEATURE_XSAVES);
	setup_clear_cpu_cap(X86_FEATURE_AVX);
	setup_clear_cpu_cap(X86_FEATURE_AVX2);
	setup_clear_cpu_cap(X86_FEATURE_AVX512F);
82
	setup_clear_cpu_cap(X86_FEATURE_AVX512IFMA);
83 84 85
	setup_clear_cpu_cap(X86_FEATURE_AVX512PF);
	setup_clear_cpu_cap(X86_FEATURE_AVX512ER);
	setup_clear_cpu_cap(X86_FEATURE_AVX512CD);
86 87 88
	setup_clear_cpu_cap(X86_FEATURE_AVX512DQ);
	setup_clear_cpu_cap(X86_FEATURE_AVX512BW);
	setup_clear_cpu_cap(X86_FEATURE_AVX512VL);
89
	setup_clear_cpu_cap(X86_FEATURE_MPX);
90
	setup_clear_cpu_cap(X86_FEATURE_XGETBV1);
91
	setup_clear_cpu_cap(X86_FEATURE_AVX512VBMI);
92
	setup_clear_cpu_cap(X86_FEATURE_PKU);
93 94
	setup_clear_cpu_cap(X86_FEATURE_AVX512_4VNNIW);
	setup_clear_cpu_cap(X86_FEATURE_AVX512_4FMAPS);
95
	setup_clear_cpu_cap(X86_FEATURE_AVX512_VPOPCNTDQ);
96 97
}

98 99 100 101 102 103 104 105 106 107 108 109 110 111 112
/*
 * Return whether the system supports a given xfeature.
 *
 * Also return the name of the (most advanced) feature that the caller requested:
 */
int cpu_has_xfeatures(u64 xfeatures_needed, const char **feature_name)
{
	u64 xfeatures_missing = xfeatures_needed & ~xfeatures_mask;

	if (unlikely(feature_name)) {
		long xfeature_idx, max_idx;
		u64 xfeatures_print;
		/*
		 * So we use FLS here to be able to print the most advanced
		 * feature that was requested but is missing. So if a driver
D
Dave Hansen 已提交
113
		 * asks about "XFEATURE_MASK_SSE | XFEATURE_MASK_YMM" we'll print the
114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135
		 * missing AVX feature - this is the most informative message
		 * to users:
		 */
		if (xfeatures_missing)
			xfeatures_print = xfeatures_missing;
		else
			xfeatures_print = xfeatures_needed;

		xfeature_idx = fls64(xfeatures_print)-1;
		max_idx = ARRAY_SIZE(xfeature_names)-1;
		xfeature_idx = min(xfeature_idx, max_idx);

		*feature_name = xfeature_names[xfeature_idx];
	}

	if (xfeatures_missing)
		return 0;

	return 1;
}
EXPORT_SYMBOL_GPL(cpu_has_xfeatures);

136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156
static int xfeature_is_supervisor(int xfeature_nr)
{
	/*
	 * We currently do not support supervisor states, but if
	 * we did, we could find out like this.
	 *
	 * SDM says: If state component 'i' is a user state component,
	 * ECX[0] return 0; if state component i is a supervisor
	 * state component, ECX[0] returns 1.
	 */
	u32 eax, ebx, ecx, edx;

	cpuid_count(XSTATE_CPUID, xfeature_nr, &eax, &ebx, &ecx, &edx);
	return !!(ecx & 1);
}

static int xfeature_is_user(int xfeature_nr)
{
	return !xfeature_is_supervisor(xfeature_nr);
}

157
/*
158 159 160 161 162
 * When executing XSAVEOPT (or other optimized XSAVE instructions), if
 * a processor implementation detects that an FPU state component is still
 * (or is again) in its initialized state, it may clear the corresponding
 * bit in the header.xfeatures field, and can skip the writeout of registers
 * to the corresponding memory layout.
163 164 165 166 167 168 169 170
 *
 * This means that when the bit is zero, the state component might still contain
 * some previous - non-initialized register state.
 *
 * Before writing xstate information to user-space we sanitize those components,
 * to always ensure that the memory layout of a feature will be in the init state
 * if the corresponding header bit is zero. This is to ensure that user-space doesn't
 * see some stale state in the memory layout during signal handling, debugging etc.
171
 */
172
void fpstate_sanitize_xstate(struct fpu *fpu)
173
{
174
	struct fxregs_state *fx = &fpu->state.fxsave;
175
	int feature_bit;
176
	u64 xfeatures;
177

178
	if (!use_xsaveopt())
179 180
		return;

181
	xfeatures = fpu->state.xsave.header.xfeatures;
182 183 184

	/*
	 * None of the feature bits are in init state. So nothing else
L
Lucas De Marchi 已提交
185
	 * to do for us, as the memory layout is up to date.
186
	 */
187
	if ((xfeatures & xfeatures_mask) == xfeatures_mask)
188 189 190 191 192
		return;

	/*
	 * FP is in init state
	 */
D
Dave Hansen 已提交
193
	if (!(xfeatures & XFEATURE_MASK_FP)) {
194 195 196 197 198 199 200 201 202 203 204 205
		fx->cwd = 0x37f;
		fx->swd = 0;
		fx->twd = 0;
		fx->fop = 0;
		fx->rip = 0;
		fx->rdp = 0;
		memset(&fx->st_space[0], 0, 128);
	}

	/*
	 * SSE is in init state
	 */
D
Dave Hansen 已提交
206
	if (!(xfeatures & XFEATURE_MASK_SSE))
207 208
		memset(&fx->xmm_space[0], 0, 256);

209 210 211 212 213
	/*
	 * First two features are FPU and SSE, which above we handled
	 * in a special way already:
	 */
	feature_bit = 0x2;
214
	xfeatures = (xfeatures_mask & ~xfeatures) >> 2;
215 216

	/*
217 218 219
	 * Update all the remaining memory layouts according to their
	 * standard xstate layout, if their header bit is in the init
	 * state:
220
	 */
221 222
	while (xfeatures) {
		if (xfeatures & 0x1) {
223
			int offset = xstate_comp_offsets[feature_bit];
224 225
			int size = xstate_sizes[feature_bit];

226
			memcpy((void *)fx + offset,
227
			       (void *)&init_fpstate.xsave + offset,
228 229 230
			       size);
		}

231
		xfeatures >>= 1;
232 233 234 235
		feature_bit++;
	}
}

236
/*
237 238
 * Enable the extended processor state save/restore feature.
 * Called once per CPU onlining.
239
 */
240
void fpu__init_cpu_xstate(void)
241
{
242
	if (!boot_cpu_has(X86_FEATURE_XSAVE) || !xfeatures_mask)
243
		return;
Y
Yu-cheng Yu 已提交
244 245 246 247 248 249 250 251 252
	/*
	 * Make it clear that XSAVES supervisor states are not yet
	 * implemented should anyone expect it to work by changing
	 * bits in XFEATURE_MASK_* macros and XCR0.
	 */
	WARN_ONCE((xfeatures_mask & XFEATURE_MASK_SUPERVISOR),
		"x86/fpu: XSAVES supervisor states are not yet implemented.\n");

	xfeatures_mask &= ~XFEATURE_MASK_SUPERVISOR;
253

A
Andy Lutomirski 已提交
254
	cr4_set_bits(X86_CR4_OSXSAVE);
255
	xsetbv(XCR_XFEATURE_ENABLED_MASK, xfeatures_mask);
256 257
}

258 259 260 261 262 263 264 265 266 267
/*
 * Note that in the future we will likely need a pair of
 * functions here: one for user xstates and the other for
 * system xstates.  For now, they are the same.
 */
static int xfeature_enabled(enum xfeature xfeature)
{
	return !!(xfeatures_mask & (1UL << xfeature));
}

268
/*
269 270
 * Record the offsets and sizes of various xstates contained
 * in the XSAVE state memory layout.
271
 */
272
static void __init setup_xstate_features(void)
273
{
D
Dave Hansen 已提交
274
	u32 eax, ebx, ecx, edx, i;
275 276 277
	/* start at the beginnning of the "extended state" */
	unsigned int last_good_offset = offsetof(struct xregs_state,
						 extended_state_area);
278 279 280 281 282 283 284 285 286
	/*
	 * The FP xstates and SSE xstates are legacy states. They are always
	 * in the fixed offsets in the xsave area in either compacted form
	 * or standard form.
	 */
	xstate_offsets[0] = 0;
	xstate_sizes[0] = offsetof(struct fxregs_state, xmm_space);
	xstate_offsets[1] = xstate_sizes[0];
	xstate_sizes[1] = FIELD_SIZEOF(struct fxregs_state, xmm_space);
287

D
Dave Hansen 已提交
288
	for (i = FIRST_EXTENDED_XFEATURE; i < XFEATURE_MAX; i++) {
289 290
		if (!xfeature_enabled(i))
			continue;
291

292
		cpuid_count(XSTATE_CPUID, i, &eax, &ebx, &ecx, &edx);
293 294 295 296 297 298 299 300

		/*
		 * If an xfeature is supervisor state, the offset
		 * in EBX is invalid. We leave it to -1.
		 */
		if (xfeature_is_user(i))
			xstate_offsets[i] = ebx;

D
Dave Hansen 已提交
301
		xstate_sizes[i] = eax;
302 303 304 305 306 307 308 309
		/*
		 * In our xstate size checks, we assume that the
		 * highest-numbered xstate feature has the
		 * highest offset in the buffer.  Ensure it does.
		 */
		WARN_ONCE(last_good_offset > xstate_offsets[i],
			"x86/fpu: misordered xstate at %d\n", last_good_offset);
		last_good_offset = xstate_offsets[i];
310
	}
311 312
}

313
static void __init print_xstate_feature(u64 xstate_mask)
314
{
315
	const char *feature_name;
316

317
	if (cpu_has_xfeatures(xstate_mask, &feature_name))
318
		pr_info("x86/fpu: Supporting XSAVE feature 0x%03Lx: '%s'\n", xstate_mask, feature_name);
319 320 321 322 323
}

/*
 * Print out all the supported xstate features:
 */
324
static void __init print_xstate_features(void)
325
{
D
Dave Hansen 已提交
326 327 328 329 330 331 332 333
	print_xstate_feature(XFEATURE_MASK_FP);
	print_xstate_feature(XFEATURE_MASK_SSE);
	print_xstate_feature(XFEATURE_MASK_YMM);
	print_xstate_feature(XFEATURE_MASK_BNDREGS);
	print_xstate_feature(XFEATURE_MASK_BNDCSR);
	print_xstate_feature(XFEATURE_MASK_OPMASK);
	print_xstate_feature(XFEATURE_MASK_ZMM_Hi256);
	print_xstate_feature(XFEATURE_MASK_Hi16_ZMM);
334
	print_xstate_feature(XFEATURE_MASK_PKRU);
335 336
}

337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363
/*
 * This check is important because it is easy to get XSTATE_*
 * confused with XSTATE_BIT_*.
 */
#define CHECK_XFEATURE(nr) do {		\
	WARN_ON(nr < FIRST_EXTENDED_XFEATURE);	\
	WARN_ON(nr >= XFEATURE_MAX);	\
} while (0)

/*
 * We could cache this like xstate_size[], but we only use
 * it here, so it would be a waste of space.
 */
static int xfeature_is_aligned(int xfeature_nr)
{
	u32 eax, ebx, ecx, edx;

	CHECK_XFEATURE(xfeature_nr);
	cpuid_count(XSTATE_CPUID, xfeature_nr, &eax, &ebx, &ecx, &edx);
	/*
	 * The value returned by ECX[1] indicates the alignment
	 * of state component 'i' when the compacted format
	 * of the extended region of an XSAVE area is used:
	 */
	return !!(ecx & 2);
}

364 365 366 367 368
/*
 * This function sets up offsets and sizes of all extended states in
 * xsave area. This supports both standard format and compacted format
 * of the xsave aread.
 */
369
static void __init setup_xstate_comp(void)
370
{
371
	unsigned int xstate_comp_sizes[sizeof(xfeatures_mask)*8];
372 373
	int i;

374 375 376 377 378 379
	/*
	 * The FP xstates and SSE xstates are legacy states. They are always
	 * in the fixed offsets in the xsave area in either compacted form
	 * or standard form.
	 */
	xstate_comp_offsets[0] = 0;
380
	xstate_comp_offsets[1] = offsetof(struct fxregs_state, xmm_space);
381

382
	if (!boot_cpu_has(X86_FEATURE_XSAVES)) {
D
Dave Hansen 已提交
383
		for (i = FIRST_EXTENDED_XFEATURE; i < XFEATURE_MAX; i++) {
384
			if (xfeature_enabled(i)) {
385 386 387 388 389 390 391
				xstate_comp_offsets[i] = xstate_offsets[i];
				xstate_comp_sizes[i] = xstate_sizes[i];
			}
		}
		return;
	}

392 393
	xstate_comp_offsets[FIRST_EXTENDED_XFEATURE] =
		FXSAVE_SIZE + XSAVE_HDR_SIZE;
394

D
Dave Hansen 已提交
395
	for (i = FIRST_EXTENDED_XFEATURE; i < XFEATURE_MAX; i++) {
396
		if (xfeature_enabled(i))
397 398 399 400
			xstate_comp_sizes[i] = xstate_sizes[i];
		else
			xstate_comp_sizes[i] = 0;

401
		if (i > FIRST_EXTENDED_XFEATURE) {
402 403 404
			xstate_comp_offsets[i] = xstate_comp_offsets[i-1]
					+ xstate_comp_sizes[i-1];

405 406 407 408
			if (xfeature_is_aligned(i))
				xstate_comp_offsets[i] =
					ALIGN(xstate_comp_offsets[i], 64);
		}
409 410 411
	}
}

412 413 414 415 416 417 418 419 420 421 422 423 424 425 426
/*
 * Print out xstate component offsets and sizes
 */
static void __init print_xstate_offset_size(void)
{
	int i;

	for (i = FIRST_EXTENDED_XFEATURE; i < XFEATURE_MAX; i++) {
		if (!xfeature_enabled(i))
			continue;
		pr_info("x86/fpu: xstate_offset[%d]: %4d, xstate_sizes[%d]: %4d\n",
			 i, xstate_comp_offsets[i], i, xstate_sizes[i]);
	}
}

427 428 429
/*
 * setup the xstate image representing the init state
 */
430
static void __init setup_init_fpu_buf(void)
431
{
432
	static int on_boot_cpu __initdata = 1;
433 434 435 436

	WARN_ON_FPU(!on_boot_cpu);
	on_boot_cpu = 0;

437
	if (!boot_cpu_has(X86_FEATURE_XSAVE))
438 439 440
		return;

	setup_xstate_features();
441
	print_xstate_features();
442

443
	if (boot_cpu_has(X86_FEATURE_XSAVES))
444
		init_fpstate.xsave.header.xcomp_bv = (u64)1 << 63 | xfeatures_mask;
445

446
	/*
447
	 * Init all the features state with header.xfeatures being 0x0
448
	 */
449
	copy_kernel_to_xregs_booting(&init_fpstate.xsave);
450

451 452 453 454
	/*
	 * Dump the init state again. This is to identify the init state
	 * of any feature which is not represented by all zero's.
	 */
455
	copy_xregs_to_kernel_booting(&init_fpstate.xsave);
456 457
}

458 459 460 461
static int xfeature_uncompacted_offset(int xfeature_nr)
{
	u32 eax, ebx, ecx, edx;

462 463 464 465 466 467 468 469 470 471
	/*
	 * Only XSAVES supports supervisor states and it uses compacted
	 * format. Checking a supervisor state's uncompacted offset is
	 * an error.
	 */
	if (XFEATURE_MASK_SUPERVISOR & (1 << xfeature_nr)) {
		WARN_ONCE(1, "No fixed offset for xstate %d\n", xfeature_nr);
		return -1;
	}

472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494
	CHECK_XFEATURE(xfeature_nr);
	cpuid_count(XSTATE_CPUID, xfeature_nr, &eax, &ebx, &ecx, &edx);
	return ebx;
}

static int xfeature_size(int xfeature_nr)
{
	u32 eax, ebx, ecx, edx;

	CHECK_XFEATURE(xfeature_nr);
	cpuid_count(XSTATE_CPUID, xfeature_nr, &eax, &ebx, &ecx, &edx);
	return eax;
}

/*
 * 'XSAVES' implies two different things:
 * 1. saving of supervisor/system state
 * 2. using the compacted format
 *
 * Use this function when dealing with the compacted format so
 * that it is obvious which aspect of 'XSAVES' is being handled
 * by the calling code.
 */
495
int using_compacted_format(void)
496
{
497
	return boot_cpu_has(X86_FEATURE_XSAVES);
498 499
}

500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523
/* Validate an xstate header supplied by userspace (ptrace or sigreturn) */
int validate_xstate_header(const struct xstate_header *hdr)
{
	/* No unknown or supervisor features may be set */
	if (hdr->xfeatures & (~xfeatures_mask | XFEATURE_MASK_SUPERVISOR))
		return -EINVAL;

	/* Userspace must use the uncompacted format */
	if (hdr->xcomp_bv)
		return -EINVAL;

	/*
	 * If 'reserved' is shrunken to add a new field, make sure to validate
	 * that new field here!
	 */
	BUILD_BUG_ON(sizeof(hdr->reserved) != 48);

	/* No reserved bits may be set */
	if (memchr_inv(hdr->reserved, 0, sizeof(hdr->reserved)))
		return -EINVAL;

	return 0;
}

524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549
static void __xstate_dump_leaves(void)
{
	int i;
	u32 eax, ebx, ecx, edx;
	static int should_dump = 1;

	if (!should_dump)
		return;
	should_dump = 0;
	/*
	 * Dump out a few leaves past the ones that we support
	 * just in case there are some goodies up there
	 */
	for (i = 0; i < XFEATURE_MAX + 10; i++) {
		cpuid_count(XSTATE_CPUID, i, &eax, &ebx, &ecx, &edx);
		pr_warn("CPUID[%02x, %02x]: eax=%08x ebx=%08x ecx=%08x edx=%08x\n",
			XSTATE_CPUID, i, eax, ebx, ecx, edx);
	}
}

#define XSTATE_WARN_ON(x) do {							\
	if (WARN_ONCE(x, "XSAVE consistency problem, dumping leaves")) {	\
		__xstate_dump_leaves();						\
	}									\
} while (0)

550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579
#define XCHECK_SZ(sz, nr, nr_macro, __struct) do {			\
	if ((nr == nr_macro) &&						\
	    WARN_ONCE(sz != sizeof(__struct),				\
		"%s: struct is %zu bytes, cpu state %d bytes\n",	\
		__stringify(nr_macro), sizeof(__struct), sz)) {		\
		__xstate_dump_leaves();					\
	}								\
} while (0)

/*
 * We have a C struct for each 'xstate'.  We need to ensure
 * that our software representation matches what the CPU
 * tells us about the state's size.
 */
static void check_xstate_against_struct(int nr)
{
	/*
	 * Ask the CPU for the size of the state.
	 */
	int sz = xfeature_size(nr);
	/*
	 * Match each CPU state with the corresponding software
	 * structure.
	 */
	XCHECK_SZ(sz, nr, XFEATURE_YMM,       struct ymmh_struct);
	XCHECK_SZ(sz, nr, XFEATURE_BNDREGS,   struct mpx_bndreg_state);
	XCHECK_SZ(sz, nr, XFEATURE_BNDCSR,    struct mpx_bndcsr_state);
	XCHECK_SZ(sz, nr, XFEATURE_OPMASK,    struct avx_512_opmask_state);
	XCHECK_SZ(sz, nr, XFEATURE_ZMM_Hi256, struct avx_512_zmm_uppers_state);
	XCHECK_SZ(sz, nr, XFEATURE_Hi16_ZMM,  struct avx_512_hi16_state);
580
	XCHECK_SZ(sz, nr, XFEATURE_PKRU,      struct pkru_state);
581 582 583 584 585 586 587

	/*
	 * Make *SURE* to add any feature numbers in below if
	 * there are "holes" in the xsave state component
	 * numbers.
	 */
	if ((nr < XFEATURE_YMM) ||
588 589
	    (nr >= XFEATURE_MAX) ||
	    (nr == XFEATURE_PT_UNIMPLEMENTED_SO_FAR)) {
590 591 592 593 594
		WARN_ONCE(1, "no structure for xstate: %d\n", nr);
		XSTATE_WARN_ON(1);
	}
}

595 596 597 598 599 600 601 602 603 604 605 606 607
/*
 * This essentially double-checks what the cpu told us about
 * how large the XSAVE buffer needs to be.  We are recalculating
 * it to be safe.
 */
static void do_extra_xstate_size_checks(void)
{
	int paranoid_xstate_size = FXSAVE_SIZE + XSAVE_HDR_SIZE;
	int i;

	for (i = FIRST_EXTENDED_XFEATURE; i < XFEATURE_MAX; i++) {
		if (!xfeature_enabled(i))
			continue;
608 609

		check_xstate_against_struct(i);
610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633
		/*
		 * Supervisor state components can be managed only by
		 * XSAVES, which is compacted-format only.
		 */
		if (!using_compacted_format())
			XSTATE_WARN_ON(xfeature_is_supervisor(i));

		/* Align from the end of the previous feature */
		if (xfeature_is_aligned(i))
			paranoid_xstate_size = ALIGN(paranoid_xstate_size, 64);
		/*
		 * The offset of a given state in the non-compacted
		 * format is given to us in a CPUID leaf.  We check
		 * them for being ordered (increasing offsets) in
		 * setup_xstate_features().
		 */
		if (!using_compacted_format())
			paranoid_xstate_size = xfeature_uncompacted_offset(i);
		/*
		 * The compacted-format offset always depends on where
		 * the previous state ended.
		 */
		paranoid_xstate_size += xfeature_size(i);
	}
634
	XSTATE_WARN_ON(paranoid_xstate_size != fpu_kernel_xstate_size);
635 636
}

637

F
Fenghua Yu 已提交
638
/*
639
 * Get total size of enabled xstates in XCR0/xfeatures_mask.
640 641 642 643 644 645 646 647
 *
 * Note the SDM's wording here.  "sub-function 0" only enumerates
 * the size of the *user* states.  If we use it to size a buffer
 * that we use 'XSAVES' on, we could potentially overflow the
 * buffer because 'XSAVES' saves system states too.
 *
 * Note that we do not currently set any bits on IA32_XSS so
 * 'XCR0 | IA32_XSS == XCR0' for now.
F
Fenghua Yu 已提交
648
 */
649
static unsigned int __init get_xsaves_size(void)
F
Fenghua Yu 已提交
650 651
{
	unsigned int eax, ebx, ecx, edx;
652 653 654 655 656 657 658 659 660 661 662
	/*
	 * - CPUID function 0DH, sub-function 1:
	 *    EBX enumerates the size (in bytes) required by
	 *    the XSAVES instruction for an XSAVE area
	 *    containing all the state components
	 *    corresponding to bits currently set in
	 *    XCR0 | IA32_XSS.
	 */
	cpuid_count(XSTATE_CPUID, 1, &eax, &ebx, &ecx, &edx);
	return ebx;
}
F
Fenghua Yu 已提交
663

664 665 666 667 668 669 670 671 672 673 674 675
static unsigned int __init get_xsave_size(void)
{
	unsigned int eax, ebx, ecx, edx;
	/*
	 * - CPUID function 0DH, sub-function 0:
	 *    EBX enumerates the size (in bytes) required by
	 *    the XSAVE instruction for an XSAVE area
	 *    containing all the *user* state components
	 *    corresponding to bits currently set in XCR0.
	 */
	cpuid_count(XSTATE_CPUID, 0, &eax, &ebx, &ecx, &edx);
	return ebx;
D
Dave Hansen 已提交
676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694
}

/*
 * Will the runtime-enumerated 'xstate_size' fit in the init
 * task's statically-allocated buffer?
 */
static bool is_supported_xstate_size(unsigned int test_xstate_size)
{
	if (test_xstate_size <= sizeof(union fpregs_state))
		return true;

	pr_warn("x86/fpu: xstate buffer too small (%zu < %d), disabling xsave\n",
			sizeof(union fpregs_state), test_xstate_size);
	return false;
}

static int init_xstate_size(void)
{
	/* Recompute the context size for enabled features: */
695 696 697 698 699 700 701 702 703
	unsigned int possible_xstate_size;
	unsigned int xsave_size;

	xsave_size = get_xsave_size();

	if (boot_cpu_has(X86_FEATURE_XSAVES))
		possible_xstate_size = get_xsaves_size();
	else
		possible_xstate_size = xsave_size;
D
Dave Hansen 已提交
704 705 706 707 708 709 710 711 712

	/* Ensure we have the space to store all enabled: */
	if (!is_supported_xstate_size(possible_xstate_size))
		return -EINVAL;

	/*
	 * The size is OK, we are definitely going to use xsave,
	 * make it known to the world that we need more space.
	 */
713
	fpu_kernel_xstate_size = possible_xstate_size;
714
	do_extra_xstate_size_checks();
715 716 717 718 719

	/*
	 * User space is always in standard format.
	 */
	fpu_user_xstate_size = xsave_size;
D
Dave Hansen 已提交
720 721 722
	return 0;
}

D
Dave Hansen 已提交
723 724 725 726 727
/*
 * We enabled the XSAVE hardware, but something went wrong and
 * we can not use it.  Disable it.
 */
static void fpu__init_disable_system_xstate(void)
D
Dave Hansen 已提交
728 729 730 731
{
	xfeatures_mask = 0;
	cr4_clear_bits(X86_CR4_OSXSAVE);
	fpu__xstate_clear_all_cpu_caps();
F
Fenghua Yu 已提交
732 733
}

734 735
/*
 * Enable and initialize the xsave feature.
736
 * Called once per system bootup.
737
 */
738
void __init fpu__init_system_xstate(void)
739 740
{
	unsigned int eax, ebx, ecx, edx;
741
	static int on_boot_cpu __initdata = 1;
D
Dave Hansen 已提交
742
	int err;
743
	int i;
744 745 746

	WARN_ON_FPU(!on_boot_cpu);
	on_boot_cpu = 0;
747

748 749 750 751 752
	if (!boot_cpu_has(X86_FEATURE_FPU)) {
		pr_info("x86/fpu: No FPU detected\n");
		return;
	}

753
	if (!boot_cpu_has(X86_FEATURE_XSAVE)) {
754 755
		pr_info("x86/fpu: x87 FPU will use %s\n",
			boot_cpu_has(X86_FEATURE_FXSR) ? "FXSAVE" : "FSAVE");
756 757 758
		return;
	}

759
	if (boot_cpu_data.cpuid_level < XSTATE_CPUID) {
760
		WARN_ON_FPU(1);
761 762 763 764
		return;
	}

	cpuid_count(XSTATE_CPUID, 0, &eax, &ebx, &ecx, &edx);
765
	xfeatures_mask = eax + ((u64)edx << 32);
766

D
Dave Hansen 已提交
767
	if ((xfeatures_mask & XFEATURE_MASK_FPSSE) != XFEATURE_MASK_FPSSE) {
768 769 770 771 772
		/*
		 * This indicates that something really unexpected happened
		 * with the enumeration.  Disable XSAVE and try to continue
		 * booting without it.  This is too early to BUG().
		 */
773
		pr_err("x86/fpu: FP/SSE not present amongst the CPU's xstate features: 0x%llx.\n", xfeatures_mask);
774
		goto out_disable;
775 776
	}

777 778 779 780 781 782 783 784
	/*
	 * Clear XSAVE features that are disabled in the normal CPUID.
	 */
	for (i = 0; i < ARRAY_SIZE(xsave_cpuid_features); i++) {
		if (!boot_cpu_has(xsave_cpuid_features[i]))
			xfeatures_mask &= ~BIT(i);
	}

785
	xfeatures_mask &= fpu__get_supported_xfeatures_mask();
786

787 788
	/* Enable xstate instructions to be able to continue with initialization: */
	fpu__init_cpu_xstate();
D
Dave Hansen 已提交
789
	err = init_xstate_size();
790 791
	if (err)
		goto out_disable;
792

793 794 795 796 797 798
	/*
	 * Update info used for ptrace frames; use standard-format size and no
	 * supervisor xstates:
	 */
	update_regset_xstate_info(fpu_user_xstate_size,	xfeatures_mask & ~XFEATURE_MASK_SUPERVISOR);

I
Ingo Molnar 已提交
799
	fpu__init_prepare_fx_sw_frame();
800
	setup_init_fpu_buf();
801
	setup_xstate_comp();
802
	print_xstate_offset_size();
803

804
	pr_info("x86/fpu: Enabled xstate features 0x%llx, context size is %d bytes, using '%s' format.\n",
805
		xfeatures_mask,
806
		fpu_kernel_xstate_size,
807
		boot_cpu_has(X86_FEATURE_XSAVES) ? "compacted" : "standard");
808 809 810 811 812
	return;

out_disable:
	/* something went wrong, try to boot without any XSAVE support */
	fpu__init_disable_system_xstate();
813
}
814

815 816 817 818 819 820 821 822
/*
 * Restore minimal FPU state after suspend:
 */
void fpu__resume_cpu(void)
{
	/*
	 * Restore XCR0 on xsave capable CPUs:
	 */
823
	if (boot_cpu_has(X86_FEATURE_XSAVE))
824 825 826
		xsetbv(XCR_XFEATURE_ENABLED_MASK, xfeatures_mask);
}

827 828 829 830 831 832 833 834 835 836 837
/*
 * Given an xstate feature mask, calculate where in the xsave
 * buffer the state is.  Callers should ensure that the buffer
 * is valid.
 *
 * Note: does not work for compacted buffers.
 */
void *__raw_xsave_addr(struct xregs_state *xsave, int xstate_feature_mask)
{
	int feature_nr = fls64(xstate_feature_mask) - 1;

838 839 840 841 842
	if (!xfeature_enabled(feature_nr)) {
		WARN_ON_FPU(1);
		return NULL;
	}

843 844
	return (void *)xsave + xstate_comp_offsets[feature_nr];
}
845 846 847 848 849 850 851
/*
 * Given the xsave area and a state inside, this function returns the
 * address of the state.
 *
 * This is the API that is called to get xstate address in either
 * standard format or compacted format of xsave area.
 *
852 853 854
 * Note that if there is no data for the field in the xsave buffer
 * this will return NULL.
 *
855
 * Inputs:
856 857
 *	xstate: the thread's storage area for all FPU data
 *	xstate_feature: state which is defined in xsave.h (e.g.
D
Dave Hansen 已提交
858
 *	XFEATURE_MASK_FP, XFEATURE_MASK_SSE, etc...)
859
 * Output:
860 861
 *	address of the state in the xsave area, or NULL if the
 *	field is not present in the xsave buffer.
862
 */
863
void *get_xsave_addr(struct xregs_state *xsave, int xstate_feature)
864
{
865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889
	/*
	 * Do we even *have* xsave state?
	 */
	if (!boot_cpu_has(X86_FEATURE_XSAVE))
		return NULL;

	/*
	 * We should not ever be requesting features that we
	 * have not enabled.  Remember that pcntxt_mask is
	 * what we write to the XCR0 register.
	 */
	WARN_ONCE(!(xfeatures_mask & xstate_feature),
		  "get of unsupported state");
	/*
	 * This assumes the last 'xsave*' instruction to
	 * have requested that 'xstate_feature' be saved.
	 * If it did not, we might be seeing and old value
	 * of the field in the buffer.
	 *
	 * This can happen because the last 'xsave' did not
	 * request that this feature be saved (unlikely)
	 * or because the "init optimization" caused it
	 * to not be saved.
	 */
	if (!(xsave->header.xfeatures & xstate_feature))
890 891
		return NULL;

892
	return __raw_xsave_addr(xsave, xstate_feature);
893
}
P
Paolo Bonzini 已提交
894
EXPORT_SYMBOL_GPL(get_xsave_addr);
895 896 897 898 899 900 901 902 903 904 905 906

/*
 * This wraps up the common operations that need to occur when retrieving
 * data from xsave state.  It first ensures that the current task was
 * using the FPU and retrieves the data in to a buffer.  It then calculates
 * the offset of the requested field in the buffer.
 *
 * This function is safe to call whether the FPU is in use or not.
 *
 * Note that this only works on the current task.
 *
 * Inputs:
D
Dave Hansen 已提交
907 908
 *	@xsave_state: state which is defined in xsave.h (e.g. XFEATURE_MASK_FP,
 *	XFEATURE_MASK_SSE, etc...)
909 910 911 912 913 914 915 916
 * Output:
 *	address of the state in the xsave area or NULL if the state
 *	is not present or is in its 'init state'.
 */
const void *get_xsave_field_ptr(int xsave_state)
{
	struct fpu *fpu = &current->thread.fpu;

917
	if (!fpu->initialized)
918 919 920 921 922 923 924 925 926
		return NULL;
	/*
	 * fpu__save() takes the CPU's xstate registers
	 * and saves them off to the 'fpu memory buffer.
	 */
	fpu__save(fpu);

	return get_xsave_addr(&fpu->state.xsave, xsave_state);
}
927

928 929
#ifdef CONFIG_ARCH_HAS_PKEYS

930 931 932
#define NR_VALID_PKRU_BITS (CONFIG_NR_PROTECTION_KEYS * 2)
#define PKRU_VALID_MASK (NR_VALID_PKRU_BITS - 1)
/*
933 934
 * This will go out and modify PKRU register to set the access
 * rights for @pkey to @init_val.
935 936 937 938
 */
int arch_set_user_pkey_access(struct task_struct *tsk, int pkey,
		unsigned long init_val)
{
939
	u32 old_pkru;
940 941 942 943 944 945 946 947 948 949
	int pkey_shift = (pkey * PKRU_BITS_PER_PKEY);
	u32 new_pkru_bits = 0;

	/*
	 * This check implies XSAVE support.  OSPKE only gets
	 * set if we enable XSAVE and we enable PKU in XCR0.
	 */
	if (!boot_cpu_has(X86_FEATURE_OSPKE))
		return -EINVAL;

950
	/* Set the bits we need in PKRU:  */
951 952 953 954 955
	if (init_val & PKEY_DISABLE_ACCESS)
		new_pkru_bits |= PKRU_AD_BIT;
	if (init_val & PKEY_DISABLE_WRITE)
		new_pkru_bits |= PKRU_WD_BIT;

956
	/* Shift the bits in to the correct place in PKRU for pkey: */
957 958
	new_pkru_bits <<= pkey_shift;

959 960 961
	/* Get old PKRU and mask off any old bits in place: */
	old_pkru = read_pkru();
	old_pkru &= ~((PKRU_AD_BIT|PKRU_WD_BIT) << pkey_shift);
962

963 964
	/* Write old part along with new part: */
	write_pkru(old_pkru | new_pkru_bits);
965 966 967

	return 0;
}
968
#endif /* ! CONFIG_ARCH_HAS_PKEYS */
969

970 971 972 973 974 975 976 977 978
/*
 * Weird legacy quirk: SSE and YMM states store information in the
 * MXCSR and MXCSR_FLAGS fields of the FP area. That means if the FP
 * area is marked as unused in the xfeatures header, we need to copy
 * MXCSR and MXCSR_FLAGS if either SSE or YMM are in use.
 */
static inline bool xfeatures_mxcsr_quirk(u64 xfeatures)
{
	if (!(xfeatures & (XFEATURE_MASK_SSE|XFEATURE_MASK_YMM)))
979
		return false;
980 981

	if (xfeatures & XFEATURE_MASK_FP)
982
		return false;
983

984
	return true;
985 986
}

987 988 989 990
/*
 * This is similar to user_regset_copyout(), but will not add offset to
 * the source data pointer or increment pos, count, kbuf, and ubuf.
 */
991
static inline void
992 993
__copy_xstate_to_kernel(void *kbuf, const void *data,
			unsigned int offset, unsigned int size, unsigned int size_total)
994
{
995 996
	if (offset < size_total) {
		unsigned int copy = min(size, size_total - offset);
997

998
		memcpy(kbuf + offset, data, copy);
999 1000 1001 1002 1003 1004 1005 1006 1007 1008
	}
}

/*
 * Convert from kernel XSAVES compacted format to standard format and copy
 * to a kernel-space ptrace buffer.
 *
 * It supports partial copy but pos always starts from zero. This is called
 * from xstateregs_get() and there we check the CPU has XSAVES.
 */
1009
int copy_xstate_to_kernel(void *kbuf, struct xregs_state *xsave, unsigned int offset_start, unsigned int size_total)
1010 1011 1012
{
	unsigned int offset, size;
	struct xstate_header header;
1013
	int i;
1014 1015 1016 1017

	/*
	 * Currently copy_regset_to_user() starts from pos 0:
	 */
1018
	if (unlikely(offset_start != 0))
1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033
		return -EFAULT;

	/*
	 * The destination is a ptrace buffer; we put in only user xstates:
	 */
	memset(&header, 0, sizeof(header));
	header.xfeatures = xsave->header.xfeatures;
	header.xfeatures &= ~XFEATURE_MASK_SUPERVISOR;

	/*
	 * Copy xregs_state->header:
	 */
	offset = offsetof(struct xregs_state, header);
	size = sizeof(header);

1034
	__copy_xstate_to_kernel(kbuf, &header, offset, size, size_total);
1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045

	for (i = 0; i < XFEATURE_MAX; i++) {
		/*
		 * Copy only in-use xstates:
		 */
		if ((header.xfeatures >> i) & 1) {
			void *src = __raw_xsave_addr(xsave, 1 << i);

			offset = xstate_offsets[i];
			size = xstate_sizes[i];

1046 1047 1048 1049
			/* The next component has to fit fully into the output buffer: */
			if (offset + size > size_total)
				break;

1050
			__copy_xstate_to_kernel(kbuf, src, offset, size, size_total);
1051 1052 1053 1054
		}

	}

1055 1056 1057 1058 1059 1060
	if (xfeatures_mxcsr_quirk(header.xfeatures)) {
		offset = offsetof(struct fxregs_state, mxcsr);
		size = MXCSR_AND_FLAGS_SIZE;
		__copy_xstate_to_kernel(kbuf, &xsave->i387.mxcsr, offset, size, size_total);
	}

1061 1062 1063 1064 1065 1066
	/*
	 * Fill xsave->i387.sw_reserved value for ptrace frame:
	 */
	offset = offsetof(struct fxregs_state, sw_reserved);
	size = sizeof(xstate_fx_sw_bytes);

1067
	__copy_xstate_to_kernel(kbuf, xstate_fx_sw_bytes, offset, size, size_total);
1068 1069 1070 1071 1072

	return 0;
}

static inline int
1073
__copy_xstate_to_user(void __user *ubuf, const void *data, unsigned int offset, unsigned int size, unsigned int size_total)
1074
{
1075
	if (!size)
1076 1077
		return 0;

1078 1079
	if (offset < size_total) {
		unsigned int copy = min(size, size_total - offset);
1080

1081
		if (__copy_to_user(ubuf + offset, data, copy))
1082
			return -EFAULT;
1083 1084 1085 1086 1087 1088
	}
	return 0;
}

/*
 * Convert from kernel XSAVES compacted format to standard format and copy
1089
 * to a user-space buffer. It supports partial copy but pos always starts from
1090 1091 1092
 * zero. This is called from xstateregs_get() and there we check the CPU
 * has XSAVES.
 */
1093
int copy_xstate_to_user(void __user *ubuf, struct xregs_state *xsave, unsigned int offset_start, unsigned int size_total)
1094 1095 1096 1097 1098 1099 1100 1101
{
	unsigned int offset, size;
	int ret, i;
	struct xstate_header header;

	/*
	 * Currently copy_regset_to_user() starts from pos 0:
	 */
1102
	if (unlikely(offset_start != 0))
1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117
		return -EFAULT;

	/*
	 * The destination is a ptrace buffer; we put in only user xstates:
	 */
	memset(&header, 0, sizeof(header));
	header.xfeatures = xsave->header.xfeatures;
	header.xfeatures &= ~XFEATURE_MASK_SUPERVISOR;

	/*
	 * Copy xregs_state->header:
	 */
	offset = offsetof(struct xregs_state, header);
	size = sizeof(header);

1118
	ret = __copy_xstate_to_user(ubuf, &header, offset, size, size_total);
1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131
	if (ret)
		return ret;

	for (i = 0; i < XFEATURE_MAX; i++) {
		/*
		 * Copy only in-use xstates:
		 */
		if ((header.xfeatures >> i) & 1) {
			void *src = __raw_xsave_addr(xsave, 1 << i);

			offset = xstate_offsets[i];
			size = xstate_sizes[i];

1132 1133 1134 1135
			/* The next component has to fit fully into the output buffer: */
			if (offset + size > size_total)
				break;

1136
			ret = __copy_xstate_to_user(ubuf, src, offset, size, size_total);
1137 1138 1139 1140 1141 1142
			if (ret)
				return ret;
		}

	}

1143 1144 1145 1146 1147 1148
	if (xfeatures_mxcsr_quirk(header.xfeatures)) {
		offset = offsetof(struct fxregs_state, mxcsr);
		size = MXCSR_AND_FLAGS_SIZE;
		__copy_xstate_to_user(ubuf, &xsave->i387.mxcsr, offset, size, size_total);
	}

1149 1150 1151 1152 1153 1154
	/*
	 * Fill xsave->i387.sw_reserved value for ptrace frame:
	 */
	offset = offsetof(struct fxregs_state, sw_reserved);
	size = sizeof(xstate_fx_sw_bytes);

1155
	ret = __copy_xstate_to_user(ubuf, xstate_fx_sw_bytes, offset, size, size_total);
1156 1157 1158 1159 1160 1161 1162
	if (ret)
		return ret;

	return 0;
}

/*
1163
 * Convert from a ptrace standard-format kernel buffer to kernel XSAVES format
1164
 * and copy to the target thread. This is called from xstateregs_set().
1165
 */
1166
int copy_kernel_to_xstate(struct xregs_state *xsave, const void *kbuf)
1167 1168 1169
{
	unsigned int offset, size;
	int i;
1170
	struct xstate_header hdr;
1171 1172

	offset = offsetof(struct xregs_state, header);
1173
	size = sizeof(hdr);
1174

1175
	memcpy(&hdr, kbuf + offset, size);
1176

1177
	if (validate_xstate_header(&hdr))
1178 1179 1180 1181 1182
		return -EINVAL;

	for (i = 0; i < XFEATURE_MAX; i++) {
		u64 mask = ((u64)1 << i);

1183
		if (hdr.xfeatures & mask) {
1184 1185 1186 1187 1188
			void *dst = __raw_xsave_addr(xsave, 1 << i);

			offset = xstate_offsets[i];
			size = xstate_sizes[i];

1189
			memcpy(dst, kbuf + offset, size);
1190 1191 1192
		}
	}

1193
	if (xfeatures_mxcsr_quirk(hdr.xfeatures)) {
1194 1195 1196 1197 1198
		offset = offsetof(struct fxregs_state, mxcsr);
		size = MXCSR_AND_FLAGS_SIZE;
		memcpy(&xsave->i387.mxcsr, kbuf + offset, size);
	}

1199 1200 1201 1202 1203 1204 1205 1206 1207
	/*
	 * The state that came in from userspace was user-state only.
	 * Mask all the user states out of 'xfeatures':
	 */
	xsave->header.xfeatures &= XFEATURE_MASK_SUPERVISOR;

	/*
	 * Add back in the features that came in from userspace:
	 */
1208
	xsave->header.xfeatures |= hdr.xfeatures;
1209 1210 1211 1212 1213

	return 0;
}

/*
1214 1215 1216 1217
 * Convert from a ptrace or sigreturn standard-format user-space buffer to
 * kernel XSAVES format and copy to the target thread. This is called from
 * xstateregs_set(), as well as potentially from the sigreturn() and
 * rt_sigreturn() system calls.
1218
 */
1219
int copy_user_to_xstate(struct xregs_state *xsave, const void __user *ubuf)
1220 1221 1222
{
	unsigned int offset, size;
	int i;
1223
	struct xstate_header hdr;
1224 1225

	offset = offsetof(struct xregs_state, header);
1226
	size = sizeof(hdr);
1227

1228
	if (__copy_from_user(&hdr, ubuf + offset, size))
1229
		return -EFAULT;
1230

1231
	if (validate_xstate_header(&hdr))
1232 1233 1234 1235 1236
		return -EINVAL;

	for (i = 0; i < XFEATURE_MAX; i++) {
		u64 mask = ((u64)1 << i);

1237
		if (hdr.xfeatures & mask) {
1238 1239 1240 1241 1242
			void *dst = __raw_xsave_addr(xsave, 1 << i);

			offset = xstate_offsets[i];
			size = xstate_sizes[i];

1243 1244
			if (__copy_from_user(dst, ubuf + offset, size))
				return -EFAULT;
1245 1246 1247
		}
	}

1248
	if (xfeatures_mxcsr_quirk(hdr.xfeatures)) {
1249 1250 1251 1252 1253 1254
		offset = offsetof(struct fxregs_state, mxcsr);
		size = MXCSR_AND_FLAGS_SIZE;
		if (__copy_from_user(&xsave->i387.mxcsr, ubuf + offset, size))
			return -EFAULT;
	}

1255 1256 1257 1258 1259 1260 1261 1262 1263
	/*
	 * The state that came in from userspace was user-state only.
	 * Mask all the user states out of 'xfeatures':
	 */
	xsave->header.xfeatures &= XFEATURE_MASK_SUPERVISOR;

	/*
	 * Add back in the features that came in from userspace:
	 */
1264
	xsave->header.xfeatures |= hdr.xfeatures;
1265 1266 1267

	return 0;
}