process.c 4.3 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13
/* MN10300  Process handling code
 *
 * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved.
 * Written by David Howells (dhowells@redhat.com)
 *
 * This program is free software; you can redistribute it and/or
 * modify it under the terms of the GNU General Public Licence
 * as published by the Free Software Foundation; either version
 * 2 of the Licence, or (at your option) any later version.
 */
#include <linux/module.h>
#include <linux/errno.h>
#include <linux/sched.h>
14
#include <linux/sched/debug.h>
15
#include <linux/sched/task.h>
16 17 18 19 20 21 22 23 24 25 26 27 28
#include <linux/kernel.h>
#include <linux/mm.h>
#include <linux/smp.h>
#include <linux/stddef.h>
#include <linux/unistd.h>
#include <linux/ptrace.h>
#include <linux/user.h>
#include <linux/interrupt.h>
#include <linux/delay.h>
#include <linux/reboot.h>
#include <linux/percpu.h>
#include <linux/err.h>
#include <linux/fs.h>
29
#include <linux/slab.h>
30
#include <linux/rcupdate.h>
31
#include <linux/uaccess.h>
32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54
#include <asm/pgtable.h>
#include <asm/io.h>
#include <asm/processor.h>
#include <asm/mmu_context.h>
#include <asm/fpu.h>
#include <asm/reset-regs.h>
#include <asm/gdb-stub.h>
#include "internal.h"

/*
 * return saved PC of a blocked thread.
 */
unsigned long thread_saved_pc(struct task_struct *tsk)
{
	return ((unsigned long *) tsk->thread.sp)[3];
}

/*
 * power off function, if any
 */
void (*pm_power_off)(void);
EXPORT_SYMBOL(pm_power_off);

55 56 57 58
/*
 * On SMP it's slightly faster (but much more power-consuming!)
 * to poll the ->work.need_resched flag instead of waiting for the
 * cross-CPU IPI to arrive. Use this option with caution.
T
Thomas Gleixner 已提交
59 60
 *
 * tglx: No idea why this depends on HOTPLUG_CPU !?!
61
 */
T
Thomas Gleixner 已提交
62 63
#if !defined(CONFIG_SMP) || defined(CONFIG_HOTPLUG_CPU)
void arch_cpu_idle(void)
64
{
T
Thomas Gleixner 已提交
65
	safe_halt();
66
}
T
Thomas Gleixner 已提交
67
#endif
68 69 70 71 72 73 74

void release_segments(struct mm_struct *mm)
{
}

void machine_restart(char *cmd)
{
75
#ifdef CONFIG_KERNEL_DEBUGGER
76 77 78 79 80 81 82 83 84 85 86 87
	gdbstub_exit(0);
#endif

#ifdef mn10300_unit_hard_reset
	mn10300_unit_hard_reset();
#else
	mn10300_proc_hard_reset();
#endif
}

void machine_halt(void)
{
88
#ifdef CONFIG_KERNEL_DEBUGGER
89 90 91 92 93 94
	gdbstub_exit(0);
#endif
}

void machine_power_off(void)
{
95
#ifdef CONFIG_KERNEL_DEBUGGER
96 97 98 99 100 101
	gdbstub_exit(0);
#endif
}

void show_regs(struct pt_regs *regs)
{
102
	show_regs_print_info(KERN_DEFAULT);
103 104 105 106 107
}

/*
 * free current thread data structures etc..
 */
108
void exit_thread(struct task_struct *tsk)
109
{
110
	exit_fpu(tsk);
111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130
}

void flush_thread(void)
{
	flush_fpu();
}

void release_thread(struct task_struct *dead_task)
{
}

/*
 * we do not have to muck with descriptors here, that is
 * done in switch_mm() as needed.
 */
void copy_segments(struct task_struct *p, struct mm_struct *new_mm)
{
}

/*
131 132
 * this gets called so that we can store lazy state into memory and copy the
 * current task into the new thread.
133
 */
134
int arch_dup_task_struct(struct task_struct *dst, struct task_struct *src)
135
{
136 137 138
	unlazy_fpu(src);
	*dst = *src;
	return 0;
139 140 141 142 143 144
}

/*
 * set up the kernel stack for a new thread and copy arch-specific thread
 * control information
 */
A
Alexey Dobriyan 已提交
145
int copy_thread(unsigned long clone_flags,
146
		unsigned long c_usp, unsigned long ustk_size,
147
		struct task_struct *p)
148
{
149
	struct thread_info *ti = task_thread_info(p);
150
	struct pt_regs *c_regs;
151 152 153 154 155 156
	unsigned long c_ksp;

	c_ksp = (unsigned long) task_stack_page(p) + THREAD_SIZE;

	/* allocate the userspace exception frame and set it up */
	c_ksp -= sizeof(struct pt_regs);
157
	c_regs = (struct pt_regs *) c_ksp;
158
	c_ksp -= 12; /* allocate function call ABI slack */
159

160
	/* set up things up so the scheduler can start the new task */
161
	p->thread.uregs = c_regs;
162 163 164 165 166
	ti->frame	= c_regs;
	p->thread.a3	= (unsigned long) c_regs;
	p->thread.sp	= c_ksp;
	p->thread.wchan	= p->thread.pc;
	p->thread.usp	= c_usp;
167

168
	if (unlikely(p->flags & PF_KTHREAD)) {
169 170 171 172 173 174 175 176
		memset(c_regs, 0, sizeof(struct pt_regs));
		c_regs->a0 = c_usp; /* function */
		c_regs->d0 = ustk_size; /* argument */
		local_save_flags(c_regs->epsw);
		c_regs->epsw |= EPSW_IE | EPSW_IM_7;
		p->thread.pc	= (unsigned long) ret_from_kernel_thread;
		return 0;
	}
177 178 179
	*c_regs = *current_pt_regs();
	if (c_usp)
		c_regs->sp = c_usp;
180
	c_regs->epsw &= ~EPSW_FE; /* my FPU */
181 182 183

	/* the new TLS pointer is passed in as arg #5 to sys_clone() */
	if (clone_flags & CLONE_SETTLS)
184
		c_regs->e2 = current_frame()->d3;
185 186 187 188 189 190 191 192 193 194

	p->thread.pc	= (unsigned long) ret_from_fork;

	return 0;
}

unsigned long get_wchan(struct task_struct *p)
{
	return p->thread.wchan;
}