smp.c 10.5 KB
Newer Older
L
Linus Torvalds 已提交
1 2 3 4 5 6 7 8 9 10
/*
** SMP Support
**
** Copyright (C) 1999 Walt Drummond <drummond@valinux.com>
** Copyright (C) 1999 David Mosberger-Tang <davidm@hpl.hp.com>
** Copyright (C) 2001,2004 Grant Grundler <grundler@parisc-linux.org>
** 
** Lots of stuff stolen from arch/alpha/kernel/smp.c
** ...and then parisc stole from arch/ia64/kernel/smp.c. Thanks David! :^)
**
11
** Thanks to John Curry and Ullas Ponnadi. I learned a lot from their work.
L
Linus Torvalds 已提交
12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30
** -grant (1/12/2001)
**
**	This program is free software; you can redistribute it and/or modify
**	it under the terms of the GNU General Public License as published by
**      the Free Software Foundation; either version 2 of the License, or
**      (at your option) any later version.
*/
#include <linux/types.h>
#include <linux/spinlock.h>
#include <linux/slab.h>

#include <linux/kernel.h>
#include <linux/module.h>
#include <linux/sched.h>
#include <linux/init.h>
#include <linux/interrupt.h>
#include <linux/smp.h>
#include <linux/kernel_stat.h>
#include <linux/mm.h>
A
Alexey Dobriyan 已提交
31
#include <linux/err.h>
L
Linus Torvalds 已提交
32 33 34 35 36 37 38
#include <linux/delay.h>
#include <linux/bitops.h>

#include <asm/system.h>
#include <asm/atomic.h>
#include <asm/current.h>
#include <asm/delay.h>
39
#include <asm/tlbflush.h>
L
Linus Torvalds 已提交
40 41 42 43 44 45 46 47 48 49 50 51

#include <asm/io.h>
#include <asm/irq.h>		/* for CPU_IRQ_REGION and friends */
#include <asm/mmu_context.h>
#include <asm/page.h>
#include <asm/pgtable.h>
#include <asm/pgalloc.h>
#include <asm/processor.h>
#include <asm/ptrace.h>
#include <asm/unistd.h>
#include <asm/cacheflush.h>

52 53 54 55 56 57 58 59 60
#undef DEBUG_SMP
#ifdef DEBUG_SMP
static int smp_debug_lvl = 0;
#define smp_debug(lvl, printargs...)		\
		if (lvl >= smp_debug_lvl)	\
			printk(printargs);
#else
#define smp_debug(lvl, ...)
#endif /* DEBUG_SMP */
L
Linus Torvalds 已提交
61 62 63 64 65

DEFINE_SPINLOCK(smp_lock);

volatile struct task_struct *smp_init_current_idle_task;

66
static volatile int cpu_now_booting __read_mostly = 0;	/* track which CPU is booting */
L
Linus Torvalds 已提交
67

68
static int parisc_max_cpus __read_mostly = 1;
L
Linus Torvalds 已提交
69

70
DEFINE_PER_CPU(spinlock_t, ipi_lock) = SPIN_LOCK_UNLOCKED;
L
Linus Torvalds 已提交
71 72 73 74 75

enum ipi_message_type {
	IPI_NOP=0,
	IPI_RESCHEDULE=1,
	IPI_CALL_FUNC,
76
	IPI_CALL_FUNC_SINGLE,
L
Linus Torvalds 已提交
77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122
	IPI_CPU_START,
	IPI_CPU_STOP,
	IPI_CPU_TEST
};


/********** SMP inter processor interrupt and communication routines */

#undef PER_CPU_IRQ_REGION
#ifdef PER_CPU_IRQ_REGION
/* XXX REVISIT Ignore for now.
**    *May* need this "hook" to register IPI handler
**    once we have perCPU ExtIntr switch tables.
*/
static void
ipi_init(int cpuid)
{
#error verify IRQ_OFFSET(IPI_IRQ) is ipi_interrupt() in new IRQ region

	if(cpu_online(cpuid) )
	{
		switch_to_idle_task(current);
	}

	return;
}
#endif


/*
** Yoink this CPU from the runnable list... 
**
*/
static void
halt_processor(void) 
{
	/* REVISIT : redirect I/O Interrupts to another CPU? */
	/* REVISIT : does PM *know* this CPU isn't available? */
	cpu_clear(smp_processor_id(), cpu_online_map);
	local_irq_disable();
	for (;;)
		;
}


irqreturn_t
M
Matthew Wilcox 已提交
123
ipi_interrupt(int irq, void *dev_id) 
L
Linus Torvalds 已提交
124 125 126 127 128 129 130 131 132 133 134 135
{
	int this_cpu = smp_processor_id();
	struct cpuinfo_parisc *p = &cpu_data[this_cpu];
	unsigned long ops;
	unsigned long flags;

	/* Count this now; we may make a call that never returns. */
	p->ipi_count++;

	mb();	/* Order interrupt and bit testing. */

	for (;;) {
136 137
		spinlock_t *lock = &per_cpu(ipi_lock, this_cpu);
		spin_lock_irqsave(lock, flags);
L
Linus Torvalds 已提交
138 139
		ops = p->pending_ipi;
		p->pending_ipi = 0;
140
		spin_unlock_irqrestore(lock, flags);
L
Linus Torvalds 已提交
141 142 143 144 145 146 147 148 149

		mb(); /* Order bit clearing and data access. */

		if (!ops)
		    break;

		while (ops) {
			unsigned long which = ffz(~ops);

150 151
			ops &= ~(1 << which);

L
Linus Torvalds 已提交
152
			switch (which) {
153
			case IPI_NOP:
154
				smp_debug(100, KERN_DEBUG "CPU%d IPI_NOP\n", this_cpu);
155 156
				break;
				
L
Linus Torvalds 已提交
157
			case IPI_RESCHEDULE:
158
				smp_debug(100, KERN_DEBUG "CPU%d IPI_RESCHEDULE\n", this_cpu);
L
Linus Torvalds 已提交
159 160 161 162 163 164 165
				/*
				 * Reschedule callback.  Everything to be
				 * done is done by the interrupt return path.
				 */
				break;

			case IPI_CALL_FUNC:
166
				smp_debug(100, KERN_DEBUG "CPU%d IPI_CALL_FUNC\n", this_cpu);
167 168 169 170 171 172
				generic_smp_call_function_interrupt();
				break;

			case IPI_CALL_FUNC_SINGLE:
				smp_debug(100, KERN_DEBUG "CPU%d IPI_CALL_FUNC_SINGLE\n", this_cpu);
				generic_smp_call_function_single_interrupt();
L
Linus Torvalds 已提交
173 174 175
				break;

			case IPI_CPU_START:
176
				smp_debug(100, KERN_DEBUG "CPU%d IPI_CPU_START\n", this_cpu);
L
Linus Torvalds 已提交
177 178 179
				break;

			case IPI_CPU_STOP:
180
				smp_debug(100, KERN_DEBUG "CPU%d IPI_CPU_STOP\n", this_cpu);
L
Linus Torvalds 已提交
181 182 183 184
				halt_processor();
				break;

			case IPI_CPU_TEST:
185
				smp_debug(100, KERN_DEBUG "CPU%d is alive!\n", this_cpu);
L
Linus Torvalds 已提交
186 187 188 189 190 191 192
				break;

			default:
				printk(KERN_CRIT "Unknown IPI num on CPU%d: %lu\n",
					this_cpu, which);
				return IRQ_NONE;
			} /* Switch */
J
James Bottomley 已提交
193 194 195
		/* let in any pending interrupts */
		local_irq_enable();
		local_irq_disable();
L
Linus Torvalds 已提交
196 197 198 199 200 201 202 203 204 205
		} /* while (ops) */
	}
	return IRQ_HANDLED;
}


static inline void
ipi_send(int cpu, enum ipi_message_type op)
{
	struct cpuinfo_parisc *p = &cpu_data[cpu];
206
	spinlock_t *lock = &per_cpu(ipi_lock, cpu);
L
Linus Torvalds 已提交
207 208
	unsigned long flags;

209
	spin_lock_irqsave(lock, flags);
L
Linus Torvalds 已提交
210 211
	p->pending_ipi |= 1 << op;
	gsc_writel(IPI_IRQ - CPU_IRQ_BASE, cpu_data[cpu].hpa);
212
	spin_unlock_irqrestore(lock, flags);
L
Linus Torvalds 已提交
213 214
}

215 216 217 218 219 220 221 222
static void
send_IPI_mask(cpumask_t mask, enum ipi_message_type op)
{
	int cpu;

	for_each_cpu_mask(cpu, mask)
		ipi_send(cpu, op);
}
L
Linus Torvalds 已提交
223 224 225 226

static inline void
send_IPI_single(int dest_cpu, enum ipi_message_type op)
{
H
Helge Deller 已提交
227
	BUG_ON(dest_cpu == NO_PROC_ID);
L
Linus Torvalds 已提交
228 229 230 231 232 233 234 235 236

	ipi_send(dest_cpu, op);
}

static inline void
send_IPI_allbutself(enum ipi_message_type op)
{
	int i;
	
237 238
	for_each_online_cpu(i) {
		if (i != smp_processor_id())
L
Linus Torvalds 已提交
239 240 241 242 243 244 245 246 247 248 249 250 251 252
			send_IPI_single(i, op);
	}
}


inline void 
smp_send_stop(void)	{ send_IPI_allbutself(IPI_CPU_STOP); }

static inline void
smp_send_start(void)	{ send_IPI_allbutself(IPI_CPU_START); }

void 
smp_send_reschedule(int cpu) { send_IPI_single(cpu, IPI_RESCHEDULE); }

253 254 255 256 257 258
void
smp_send_all_nop(void)
{
	send_IPI_allbutself(IPI_NOP);
}

259
void arch_send_call_function_ipi(cpumask_t mask)
L
Linus Torvalds 已提交
260
{
261
	send_IPI_mask(mask, IPI_CALL_FUNC);
L
Linus Torvalds 已提交
262 263
}

264 265 266 267
void arch_send_call_function_single_ipi(int cpu)
{
	send_IPI_single(cpu, IPI_CALL_FUNC_SINGLE);
}
L
Linus Torvalds 已提交
268 269 270 271 272 273 274 275 276

/*
 * Flush all other CPU's tlb and then mine.  Do this with on_each_cpu()
 * as we want to ensure all TLB's flushed before proceeding.
 */

void
smp_flush_tlb_all(void)
{
277
	on_each_cpu(flush_tlb_all_local, NULL, 1);
L
Linus Torvalds 已提交
278 279 280 281 282 283 284 285
}

/*
 * Called by secondaries to update state and initialize CPU registers.
 */
static void __init
smp_cpu_init(int cpunum)
{
286
	extern int init_per_cpu(int);  /* arch/parisc/kernel/processor.c */
L
Linus Torvalds 已提交
287
	extern void init_IRQ(void);    /* arch/parisc/kernel/irq.c */
288
	extern void start_cpu_itimer(void); /* arch/parisc/kernel/time.c */
L
Linus Torvalds 已提交
289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308

	/* Set modes and Enable floating point coprocessor */
	(void) init_per_cpu(cpunum);

	disable_sr_hashing();

	mb();

	/* Well, support 2.4 linux scheme as well. */
	if (cpu_test_and_set(cpunum, cpu_online_map))
	{
		extern void machine_halt(void); /* arch/parisc.../process.c */

		printk(KERN_CRIT "CPU#%d already initialized!\n", cpunum);
		machine_halt();
	}  

	/* Initialise the idle task for this CPU */
	atomic_inc(&init_mm.mm_count);
	current->active_mm = &init_mm;
H
Helge Deller 已提交
309
	BUG_ON(current->mm);
L
Linus Torvalds 已提交
310 311
	enter_lazy_tlb(&init_mm, current);

312
	init_IRQ();   /* make sure no IRQs are enabled or pending */
313
	start_cpu_itimer();
L
Linus Torvalds 已提交
314 315 316 317 318 319 320 321 322 323 324 325
}


/*
 * Slaves start using C here. Indirectly called from smp_slave_stext.
 * Do what start_kernel() and main() do for boot strap processor (aka monarch)
 */
void __init smp_callin(void)
{
	int slave_id = cpu_now_booting;

	smp_cpu_init(slave_id);
326
	preempt_disable();
L
Linus Torvalds 已提交
327 328

	flush_cache_all_local(); /* start with known state */
329
	flush_tlb_all_local(NULL);
L
Linus Torvalds 已提交
330 331 332 333 334 335 336 337 338 339 340 341

	local_irq_enable();  /* Interrupts have been off until now */

	cpu_idle();      /* Wait for timer to schedule some work */

	/* NOTREACHED */
	panic("smp_callin() AAAAaaaaahhhh....\n");
}

/*
 * Bring one cpu online.
 */
342
int __cpuinit smp_boot_one_cpu(int cpuid)
L
Linus Torvalds 已提交
343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360
{
	struct task_struct *idle;
	long timeout;

	/* 
	 * Create an idle task for this CPU.  Note the address wed* give 
	 * to kernel_thread is irrelevant -- it's going to start
	 * where OS_BOOT_RENDEVZ vector in SAL says to start.  But
	 * this gets all the other task-y sort of data structures set
	 * up like we wish.   We need to pull the just created idle task 
	 * off the run queue and stuff it into the init_tasks[] array.  
	 * Sheesh . . .
	 */

	idle = fork_idle(cpuid);
	if (IS_ERR(idle))
		panic("SMP: fork failed for CPU:%d", cpuid);

361
	task_thread_info(idle)->cpu = cpuid;
L
Linus Torvalds 已提交
362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412

	/* Let _start know what logical CPU we're booting
	** (offset into init_tasks[],cpu_data[])
	*/
	cpu_now_booting = cpuid;

	/* 
	** boot strap code needs to know the task address since
	** it also contains the process stack.
	*/
	smp_init_current_idle_task = idle ;
	mb();

	printk("Releasing cpu %d now, hpa=%lx\n", cpuid, cpu_data[cpuid].hpa);

	/*
	** This gets PDC to release the CPU from a very tight loop.
	**
	** From the PA-RISC 2.0 Firmware Architecture Reference Specification:
	** "The MEM_RENDEZ vector specifies the location of OS_RENDEZ which 
	** is executed after receiving the rendezvous signal (an interrupt to 
	** EIR{0}). MEM_RENDEZ is valid only when it is nonzero and the 
	** contents of memory are valid."
	*/
	gsc_writel(TIMER_IRQ - CPU_IRQ_BASE, cpu_data[cpuid].hpa);
	mb();

	/* 
	 * OK, wait a bit for that CPU to finish staggering about. 
	 * Slave will set a bit when it reaches smp_cpu_init().
	 * Once the "monarch CPU" sees the bit change, it can move on.
	 */
	for (timeout = 0; timeout < 10000; timeout++) {
		if(cpu_online(cpuid)) {
			/* Which implies Slave has started up */
			cpu_now_booting = 0;
			smp_init_current_idle_task = NULL;
			goto alive ;
		}
		udelay(100);
		barrier();
	}

	put_task_struct(idle);
	idle = NULL;

	printk(KERN_CRIT "SMP: CPU:%d is stuck.\n", cpuid);
	return -1;

alive:
	/* Remember the Slave data */
413
	smp_debug(100, KERN_DEBUG "SMP: CPU:%d came alive after %ld _us\n",
L
Linus Torvalds 已提交
414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432
		cpuid, timeout * 100);
	return 0;
}

void __devinit smp_prepare_boot_cpu(void)
{
	int bootstrap_processor=cpu_data[0].cpuid;	/* CPU ID of BSP */

	/* Setup BSP mappings */
	printk("SMP: bootstrap CPU ID is %d\n",bootstrap_processor);

	cpu_set(bootstrap_processor, cpu_online_map);
	cpu_set(bootstrap_processor, cpu_present_map);
}



/*
** inventory.c:do_inventory() hasn't yet been run and thus we
433
** don't 'discover' the additional CPUs until later.
L
Linus Torvalds 已提交
434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451
*/
void __init smp_prepare_cpus(unsigned int max_cpus)
{
	cpus_clear(cpu_present_map);
	cpu_set(0, cpu_present_map);

	parisc_max_cpus = max_cpus;
	if (!max_cpus)
		printk(KERN_INFO "SMP mode deactivated.\n");
}


void smp_cpus_done(unsigned int cpu_max)
{
	return;
}


452
int __cpuinit __cpu_up(unsigned int cpu)
L
Linus Torvalds 已提交
453 454 455 456 457 458 459 460 461 462 463 464 465 466
{
	if (cpu != 0 && cpu < parisc_max_cpus)
		smp_boot_one_cpu(cpu);

	return cpu_online(cpu) ? 0 : -ENOSYS;
}

#ifdef CONFIG_PROC_FS
int __init
setup_profiling_timer(unsigned int multiplier)
{
	return -EINVAL;
}
#endif