fpsimd.c 12.6 KB
Newer Older
C
Catalin Marinas 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
/*
 * FP/SIMD context switching and fault handling
 *
 * Copyright (C) 2012 ARM Ltd.
 * Author: Catalin Marinas <catalin.marinas@arm.com>
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License version 2 as
 * published by the Free Software Foundation.
 *
 * This program is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program.  If not, see <http://www.gnu.org/licenses/>.
 */

20
#include <linux/bottom_half.h>
21
#include <linux/cpu.h>
22
#include <linux/cpu_pm.h>
C
Catalin Marinas 已提交
23
#include <linux/kernel.h>
24
#include <linux/linkage.h>
C
Catalin Marinas 已提交
25
#include <linux/init.h>
26
#include <linux/percpu.h>
27
#include <linux/preempt.h>
28
#include <linux/sched/signal.h>
C
Catalin Marinas 已提交
29 30 31 32
#include <linux/signal.h>

#include <asm/fpsimd.h>
#include <asm/cputype.h>
33
#include <asm/simd.h>
C
Catalin Marinas 已提交
34 35 36 37 38 39 40 41

#define FPEXC_IOF	(1 << 0)
#define FPEXC_DZF	(1 << 1)
#define FPEXC_OFF	(1 << 2)
#define FPEXC_UFF	(1 << 3)
#define FPEXC_IXF	(1 << 4)
#define FPEXC_IDF	(1 << 7)

42 43 44 45 46 47 48 49 50 51
/*
 * In order to reduce the number of times the FPSIMD state is needlessly saved
 * and restored, we need to keep track of two things:
 * (a) for each task, we need to remember which CPU was the last one to have
 *     the task's FPSIMD state loaded into its FPSIMD registers;
 * (b) for each CPU, we need to remember which task's userland FPSIMD state has
 *     been loaded into its FPSIMD registers most recently, or whether it has
 *     been used to perform kernel mode NEON in the meantime.
 *
 * For (a), we add a 'cpu' field to struct fpsimd_state, which gets updated to
52
 * the id of the current CPU every time the state is loaded onto a CPU. For (b),
53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68
 * we add the per-cpu variable 'fpsimd_last_state' (below), which contains the
 * address of the userland FPSIMD state of the task that was loaded onto the CPU
 * the most recently, or NULL if kernel mode NEON has been performed after that.
 *
 * With this in place, we no longer have to restore the next FPSIMD state right
 * when switching between tasks. Instead, we can defer this check to userland
 * resume, at which time we verify whether the CPU's fpsimd_last_state and the
 * task's fpsimd_state.cpu are still mutually in sync. If this is the case, we
 * can omit the FPSIMD restore.
 *
 * As an optimization, we use the thread_info flag TIF_FOREIGN_FPSTATE to
 * indicate whether or not the userland FPSIMD state of the current task is
 * present in the registers. The flag is set unless the FPSIMD registers of this
 * CPU currently contain the most recent userland FPSIMD state of the current
 * task.
 *
69 70 71 72 73 74 75
 * In order to allow softirq handlers to use FPSIMD, kernel_neon_begin() may
 * save the task's FPSIMD context back to task_struct from softirq context.
 * To prevent this from racing with the manipulation of the task's FPSIMD state
 * from task context and thereby corrupting the state, it is necessary to
 * protect any manipulation of a task's fpsimd_state or TIF_FOREIGN_FPSTATE
 * flag with local_bh_disable() unless softirqs are already masked.
 *
76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102
 * For a certain task, the sequence may look something like this:
 * - the task gets scheduled in; if both the task's fpsimd_state.cpu field
 *   contains the id of the current CPU, and the CPU's fpsimd_last_state per-cpu
 *   variable points to the task's fpsimd_state, the TIF_FOREIGN_FPSTATE flag is
 *   cleared, otherwise it is set;
 *
 * - the task returns to userland; if TIF_FOREIGN_FPSTATE is set, the task's
 *   userland FPSIMD state is copied from memory to the registers, the task's
 *   fpsimd_state.cpu field is set to the id of the current CPU, the current
 *   CPU's fpsimd_last_state pointer is set to this task's fpsimd_state and the
 *   TIF_FOREIGN_FPSTATE flag is cleared;
 *
 * - the task executes an ordinary syscall; upon return to userland, the
 *   TIF_FOREIGN_FPSTATE flag will still be cleared, so no FPSIMD state is
 *   restored;
 *
 * - the task executes a syscall which executes some NEON instructions; this is
 *   preceded by a call to kernel_neon_begin(), which copies the task's FPSIMD
 *   register contents to memory, clears the fpsimd_last_state per-cpu variable
 *   and sets the TIF_FOREIGN_FPSTATE flag;
 *
 * - the task gets preempted after kernel_neon_end() is called; as we have not
 *   returned from the 2nd syscall yet, TIF_FOREIGN_FPSTATE is still set so
 *   whatever is in the FPSIMD registers is not saved to memory, but discarded.
 */
static DEFINE_PER_CPU(struct fpsimd_state *, fpsimd_last_state);

C
Catalin Marinas 已提交
103 104 105
/*
 * Trapped FP/ASIMD access.
 */
106
asmlinkage void do_fpsimd_acc(unsigned int esr, struct pt_regs *regs)
C
Catalin Marinas 已提交
107 108 109 110 111 112 113 114
{
	/* TODO: implement lazy context saving/restoring */
	WARN_ON(1);
}

/*
 * Raise a SIGFPE for the current process.
 */
115
asmlinkage void do_fpsimd_exc(unsigned int esr, struct pt_regs *regs)
C
Catalin Marinas 已提交
116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140
{
	siginfo_t info;
	unsigned int si_code = 0;

	if (esr & FPEXC_IOF)
		si_code = FPE_FLTINV;
	else if (esr & FPEXC_DZF)
		si_code = FPE_FLTDIV;
	else if (esr & FPEXC_OFF)
		si_code = FPE_FLTOVF;
	else if (esr & FPEXC_UFF)
		si_code = FPE_FLTUND;
	else if (esr & FPEXC_IXF)
		si_code = FPE_FLTRES;

	memset(&info, 0, sizeof(info));
	info.si_signo = SIGFPE;
	info.si_code = si_code;
	info.si_addr = (void __user *)instruction_pointer(regs);

	send_sig_info(SIGFPE, &info, current);
}

void fpsimd_thread_switch(struct task_struct *next)
{
141 142
	if (!system_supports_fpsimd())
		return;
143 144 145 146 147 148
	/*
	 * Save the current FPSIMD state to memory, but only if whatever is in
	 * the registers is in fact the most recent userland FPSIMD state of
	 * 'current'.
	 */
	if (current->mm && !test_thread_flag(TIF_FOREIGN_FPSTATE))
C
Catalin Marinas 已提交
149
		fpsimd_save_state(&current->thread.fpsimd_state);
150 151 152 153 154 155 156 157 158 159 160 161 162

	if (next->mm) {
		/*
		 * If we are switching to a task whose most recent userland
		 * FPSIMD state is already in the registers of *this* cpu,
		 * we can skip loading the state from memory. Otherwise, set
		 * the TIF_FOREIGN_FPSTATE flag so the state will be loaded
		 * upon the next return to userland.
		 */
		struct fpsimd_state *st = &next->thread.fpsimd_state;

		if (__this_cpu_read(fpsimd_last_state) == st
		    && st->cpu == smp_processor_id())
163
			clear_tsk_thread_flag(next, TIF_FOREIGN_FPSTATE);
164
		else
165
			set_tsk_thread_flag(next, TIF_FOREIGN_FPSTATE);
166
	}
C
Catalin Marinas 已提交
167 168 169 170
}

void fpsimd_flush_thread(void)
{
171 172
	if (!system_supports_fpsimd())
		return;
173 174 175

	local_bh_disable();

C
Catalin Marinas 已提交
176
	memset(&current->thread.fpsimd_state, 0, sizeof(struct fpsimd_state));
177
	fpsimd_flush_task_state(current);
178
	set_thread_flag(TIF_FOREIGN_FPSTATE);
179 180

	local_bh_enable();
C
Catalin Marinas 已提交
181 182
}

183
/*
184 185
 * Save the userland FPSIMD state of 'current' to memory, but only if the state
 * currently held in the registers does in fact belong to 'current'
186 187 188
 */
void fpsimd_preserve_current_state(void)
{
189 190
	if (!system_supports_fpsimd())
		return;
191 192 193

	local_bh_disable();

194 195
	if (!test_thread_flag(TIF_FOREIGN_FPSTATE))
		fpsimd_save_state(&current->thread.fpsimd_state);
196 197

	local_bh_enable();
198 199 200
}

/*
201 202 203 204 205 206
 * Load the userland FPSIMD state of 'current' from memory, but only if the
 * FPSIMD state already held in the registers is /not/ the most recent FPSIMD
 * state of 'current'
 */
void fpsimd_restore_current_state(void)
{
207 208
	if (!system_supports_fpsimd())
		return;
209 210 211

	local_bh_disable();

212 213 214 215
	if (test_and_clear_thread_flag(TIF_FOREIGN_FPSTATE)) {
		struct fpsimd_state *st = &current->thread.fpsimd_state;

		fpsimd_load_state(st);
216
		__this_cpu_write(fpsimd_last_state, st);
217 218
		st->cpu = smp_processor_id();
	}
219 220

	local_bh_enable();
221 222 223 224 225 226
}

/*
 * Load an updated userland FPSIMD state for 'current' from memory and set the
 * flag that indicates that the FPSIMD register contents are the most recent
 * FPSIMD state of 'current'
227 228 229
 */
void fpsimd_update_current_state(struct fpsimd_state *state)
{
230 231
	if (!system_supports_fpsimd())
		return;
232 233 234

	local_bh_disable();

235
	fpsimd_load_state(state);
236 237 238
	if (test_and_clear_thread_flag(TIF_FOREIGN_FPSTATE)) {
		struct fpsimd_state *st = &current->thread.fpsimd_state;

239
		__this_cpu_write(fpsimd_last_state, st);
240 241
		st->cpu = smp_processor_id();
	}
242 243

	local_bh_enable();
244 245
}

246 247 248 249 250 251 252 253
/*
 * Invalidate live CPU copies of task t's FPSIMD state
 */
void fpsimd_flush_task_state(struct task_struct *t)
{
	t->thread.fpsimd_state.cpu = NR_CPUS;
}

254 255
#ifdef CONFIG_KERNEL_MODE_NEON

256
DEFINE_PER_CPU(bool, kernel_neon_busy);
257
EXPORT_PER_CPU_SYMBOL(kernel_neon_busy);
258

259 260 261
/*
 * Kernel-side NEON support functions
 */
262 263 264 265 266 267 268 269 270 271 272 273 274 275 276

/*
 * kernel_neon_begin(): obtain the CPU FPSIMD registers for use by the calling
 * context
 *
 * Must not be called unless may_use_simd() returns true.
 * Task context in the FPSIMD registers is saved back to memory as necessary.
 *
 * A matching call to kernel_neon_end() must be made before returning from the
 * calling context.
 *
 * The caller may freely use the FPSIMD registers until kernel_neon_end() is
 * called.
 */
void kernel_neon_begin(void)
277
{
278 279
	if (WARN_ON(!system_supports_fpsimd()))
		return;
280

281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296
	BUG_ON(!may_use_simd());

	local_bh_disable();

	__this_cpu_write(kernel_neon_busy, true);

	/* Save unsaved task fpsimd state, if any: */
	if (current->mm && !test_and_set_thread_flag(TIF_FOREIGN_FPSTATE))
		fpsimd_save_state(&current->thread.fpsimd_state);

	/* Invalidate any task state remaining in the fpsimd regs: */
	__this_cpu_write(fpsimd_last_state, NULL);

	preempt_disable();

	local_bh_enable();
297
}
298
EXPORT_SYMBOL(kernel_neon_begin);
299

300 301 302 303 304 305 306 307 308
/*
 * kernel_neon_end(): give the CPU FPSIMD registers back to the current task
 *
 * Must be called from a context in which kernel_neon_begin() was previously
 * called, with no call to kernel_neon_end() in the meantime.
 *
 * The caller must not use the FPSIMD registers after this function is called,
 * unless kernel_neon_begin() is called again in the meantime.
 */
309 310
void kernel_neon_end(void)
{
311 312
	bool busy;

313 314
	if (!system_supports_fpsimd())
		return;
315 316 317 318 319

	busy = __this_cpu_xchg(kernel_neon_busy, false);
	WARN_ON(!busy);	/* No matching kernel_neon_begin()? */

	preempt_enable();
320 321 322
}
EXPORT_SYMBOL(kernel_neon_end);

323 324
#ifdef CONFIG_EFI

325 326
static DEFINE_PER_CPU(struct fpsimd_state, efi_fpsimd_state);
static DEFINE_PER_CPU(bool, efi_fpsimd_state_used);
327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373

/*
 * EFI runtime services support functions
 *
 * The ABI for EFI runtime services allows EFI to use FPSIMD during the call.
 * This means that for EFI (and only for EFI), we have to assume that FPSIMD
 * is always used rather than being an optional accelerator.
 *
 * These functions provide the necessary support for ensuring FPSIMD
 * save/restore in the contexts from which EFI is used.
 *
 * Do not use them for any other purpose -- if tempted to do so, you are
 * either doing something wrong or you need to propose some refactoring.
 */

/*
 * __efi_fpsimd_begin(): prepare FPSIMD for making an EFI runtime services call
 */
void __efi_fpsimd_begin(void)
{
	if (!system_supports_fpsimd())
		return;

	WARN_ON(preemptible());

	if (may_use_simd())
		kernel_neon_begin();
	else {
		fpsimd_save_state(this_cpu_ptr(&efi_fpsimd_state));
		__this_cpu_write(efi_fpsimd_state_used, true);
	}
}

/*
 * __efi_fpsimd_end(): clean up FPSIMD after an EFI runtime services call
 */
void __efi_fpsimd_end(void)
{
	if (!system_supports_fpsimd())
		return;

	if (__this_cpu_xchg(efi_fpsimd_state_used, false))
		fpsimd_load_state(this_cpu_ptr(&efi_fpsimd_state));
	else
		kernel_neon_end();
}

374 375
#endif /* CONFIG_EFI */

376 377
#endif /* CONFIG_KERNEL_MODE_NEON */

378 379 380 381 382 383
#ifdef CONFIG_CPU_PM
static int fpsimd_cpu_pm_notifier(struct notifier_block *self,
				  unsigned long cmd, void *v)
{
	switch (cmd) {
	case CPU_PM_ENTER:
384
		if (current->mm && !test_thread_flag(TIF_FOREIGN_FPSTATE))
385
			fpsimd_save_state(&current->thread.fpsimd_state);
386
		this_cpu_write(fpsimd_last_state, NULL);
387 388 389
		break;
	case CPU_PM_EXIT:
		if (current->mm)
390
			set_thread_flag(TIF_FOREIGN_FPSTATE);
391 392 393 394 395 396 397 398 399 400 401 402
		break;
	case CPU_PM_ENTER_FAILED:
	default:
		return NOTIFY_DONE;
	}
	return NOTIFY_OK;
}

static struct notifier_block fpsimd_cpu_pm_notifier_block = {
	.notifier_call = fpsimd_cpu_pm_notifier,
};

403
static void __init fpsimd_pm_init(void)
404 405 406 407 408 409 410 411
{
	cpu_pm_register_notifier(&fpsimd_cpu_pm_notifier_block);
}

#else
static inline void fpsimd_pm_init(void) { }
#endif /* CONFIG_CPU_PM */

412
#ifdef CONFIG_HOTPLUG_CPU
413
static int fpsimd_cpu_dead(unsigned int cpu)
414
{
415 416
	per_cpu(fpsimd_last_state, cpu) = NULL;
	return 0;
417 418 419 420
}

static inline void fpsimd_hotplug_init(void)
{
421 422
	cpuhp_setup_state_nocalls(CPUHP_ARM64_FPSIMD_DEAD, "arm64/fpsimd:dead",
				  NULL, fpsimd_cpu_dead);
423 424 425 426 427 428
}

#else
static inline void fpsimd_hotplug_init(void) { }
#endif

C
Catalin Marinas 已提交
429 430 431 432 433
/*
 * FP/SIMD support code initialisation.
 */
static int __init fpsimd_init(void)
{
434 435 436 437
	if (elf_hwcap & HWCAP_FP) {
		fpsimd_pm_init();
		fpsimd_hotplug_init();
	} else {
C
Catalin Marinas 已提交
438 439 440
		pr_notice("Floating-point is not implemented\n");
	}

441
	if (!(elf_hwcap & HWCAP_ASIMD))
C
Catalin Marinas 已提交
442
		pr_notice("Advanced SIMD is not implemented\n");
443

C
Catalin Marinas 已提交
444 445 446
	return 0;
}
late_initcall(fpsimd_init);