smp.c 7.3 KB
Newer Older
L
Linus Torvalds 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32
/*
 * SMP Support
 *
 * Copyright (C) 1999 Walt Drummond <drummond@valinux.com>
 * Copyright (C) 1999, 2001, 2003 David Mosberger-Tang <davidm@hpl.hp.com>
 *
 * Lots of stuff stolen from arch/alpha/kernel/smp.c
 *
 * 01/05/16 Rohit Seth <rohit.seth@intel.com>  IA64-SMP functions. Reorganized
 * the existing code (on the lines of x86 port).
 * 00/09/11 David Mosberger <davidm@hpl.hp.com> Do loops_per_jiffy
 * calibration on each CPU.
 * 00/08/23 Asit Mallick <asit.k.mallick@intel.com> fixed logical processor id
 * 00/03/31 Rohit Seth <rohit.seth@intel.com>	Fixes for Bootstrap Processor
 * & cpu_online_map now gets done here (instead of setup.c)
 * 99/10/05 davidm	Update to bring it in sync with new command-line processing
 *  scheme.
 * 10/13/00 Goutham Rao <goutham.rao@intel.com> Updated smp_call_function and
 *		smp_call_function_single to resend IPI on timeouts
 */
#include <linux/module.h>
#include <linux/kernel.h>
#include <linux/sched.h>
#include <linux/init.h>
#include <linux/interrupt.h>
#include <linux/smp.h>
#include <linux/kernel_stat.h>
#include <linux/mm.h>
#include <linux/cache.h>
#include <linux/delay.h>
#include <linux/efi.h>
#include <linux/bitops.h>
Z
Zou Nan hai 已提交
33
#include <linux/kexec.h>
L
Linus Torvalds 已提交
34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51

#include <asm/atomic.h>
#include <asm/current.h>
#include <asm/delay.h>
#include <asm/machvec.h>
#include <asm/io.h>
#include <asm/irq.h>
#include <asm/page.h>
#include <asm/pgalloc.h>
#include <asm/pgtable.h>
#include <asm/processor.h>
#include <asm/ptrace.h>
#include <asm/sal.h>
#include <asm/system.h>
#include <asm/tlbflush.h>
#include <asm/unistd.h>
#include <asm/mca.h>

52 53 54 55 56 57 58 59 60 61 62
/*
 * Note: alignment of 4 entries/cacheline was empirically determined
 * to be a good tradeoff between hot cachelines & spreading the array
 * across too many cacheline.
 */
static struct local_tlb_flush_counts {
	unsigned int count;
} __attribute__((__aligned__(32))) local_tlb_flush_counts[NR_CPUS];

static DEFINE_PER_CPU(unsigned int, shadow_flush_counts[NR_CPUS]) ____cacheline_aligned;

L
Linus Torvalds 已提交
63 64
#define IPI_CALL_FUNC		0
#define IPI_CPU_STOP		1
65
#define IPI_CALL_FUNC_SINGLE	2
Z
Zou Nan hai 已提交
66
#define IPI_KDUMP_CPU_STOP	3
L
Linus Torvalds 已提交
67 68

/* This needs to be cacheline aligned because it is written to by *other* CPUs.  */
69
static DEFINE_PER_CPU_SHARED_ALIGNED(u64, ipi_operation);
L
Linus Torvalds 已提交
70 71 72 73

extern void cpu_halt (void);

static void
H
Hidetoshi Seto 已提交
74
stop_this_cpu(void)
L
Linus Torvalds 已提交
75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96
{
	/*
	 * Remove this CPU:
	 */
	cpu_clear(smp_processor_id(), cpu_online_map);
	max_xtp();
	local_irq_disable();
	cpu_halt();
}

void
cpu_die(void)
{
	max_xtp();
	local_irq_disable();
	cpu_halt();
	/* Should never be here */
	BUG();
	for (;;);
}

irqreturn_t
97
handle_IPI (int irq, void *dev_id)
L
Linus Torvalds 已提交
98 99 100 101 102 103 104 105 106 107 108 109 110 111 112
{
	int this_cpu = get_cpu();
	unsigned long *pending_ipis = &__ia64_per_cpu_var(ipi_operation);
	unsigned long ops;

	mb();	/* Order interrupt and bit testing. */
	while ((ops = xchg(pending_ipis, 0)) != 0) {
		mb();	/* Order bit clearing and data access. */
		do {
			unsigned long which;

			which = ffz(~ops);
			ops &= ~(1 << which);

			switch (which) {
H
Hidetoshi Seto 已提交
113
			case IPI_CPU_STOP:
L
Linus Torvalds 已提交
114 115
				stop_this_cpu();
				break;
116 117 118 119 120 121
			case IPI_CALL_FUNC:
				generic_smp_call_function_interrupt();
				break;
			case IPI_CALL_FUNC_SINGLE:
				generic_smp_call_function_single_interrupt();
				break;
122
#ifdef CONFIG_KEXEC
H
Hidetoshi Seto 已提交
123
			case IPI_KDUMP_CPU_STOP:
Z
Zou Nan hai 已提交
124 125 126
				unw_init_running(kdump_cpu_freeze, NULL);
				break;
#endif
H
Hidetoshi Seto 已提交
127 128 129
			default:
				printk(KERN_CRIT "Unknown IPI on CPU %d: %lu\n",
						this_cpu, which);
L
Linus Torvalds 已提交
130 131 132 133 134 135 136 137 138
				break;
			}
		} while (ops);
		mb();	/* Order data access and bit testing. */
	}
	put_cpu();
	return IRQ_HANDLED;
}

139 140


L
Linus Torvalds 已提交
141
/*
S
Simon Arlott 已提交
142
 * Called with preemption disabled.
L
Linus Torvalds 已提交
143 144 145 146 147 148 149 150 151
 */
static inline void
send_IPI_single (int dest_cpu, int op)
{
	set_bit(op, &per_cpu(ipi_operation, dest_cpu));
	platform_send_ipi(dest_cpu, IA64_IPI_VECTOR, IA64_IPI_DM_INT, 0);
}

/*
S
Simon Arlott 已提交
152
 * Called with preemption disabled.
L
Linus Torvalds 已提交
153 154 155 156 157 158
 */
static inline void
send_IPI_allbutself (int op)
{
	unsigned int i;

159 160
	for_each_online_cpu(i) {
		if (i != smp_processor_id())
L
Linus Torvalds 已提交
161 162 163 164
			send_IPI_single(i, op);
	}
}

165 166 167 168 169 170 171 172 173 174 175 176 177
/*
 * Called with preemption disabled.
 */
static inline void
send_IPI_mask(cpumask_t mask, int op)
{
	unsigned int cpu;

	for_each_cpu_mask(cpu, mask) {
			send_IPI_single(cpu, op);
	}
}

L
Linus Torvalds 已提交
178
/*
S
Simon Arlott 已提交
179
 * Called with preemption disabled.
L
Linus Torvalds 已提交
180 181 182 183 184 185
 */
static inline void
send_IPI_all (int op)
{
	int i;

186 187 188
	for_each_online_cpu(i) {
		send_IPI_single(i, op);
	}
L
Linus Torvalds 已提交
189 190 191
}

/*
S
Simon Arlott 已提交
192
 * Called with preemption disabled.
L
Linus Torvalds 已提交
193 194 195 196 197 198 199
 */
static inline void
send_IPI_self (int op)
{
	send_IPI_single(smp_processor_id(), op);
}

200
#ifdef CONFIG_KEXEC
Z
Zou Nan hai 已提交
201
void
A
Al Viro 已提交
202
kdump_smp_send_stop(void)
Z
Zou Nan hai 已提交
203 204 205 206 207
{
 	send_IPI_allbutself(IPI_KDUMP_CPU_STOP);
}

void
A
Al Viro 已提交
208
kdump_smp_send_init(void)
Z
Zou Nan hai 已提交
209 210 211 212 213 214 215 216 217 218 219
{
	unsigned int cpu, self_cpu;
	self_cpu = smp_processor_id();
	for_each_online_cpu(cpu) {
		if (cpu != self_cpu) {
			if(kdump_status[cpu] == 0)
				platform_send_ipi(cpu, 0, IA64_IPI_DM_INIT, 0);
		}
	}
}
#endif
L
Linus Torvalds 已提交
220
/*
S
Simon Arlott 已提交
221
 * Called with preemption disabled.
L
Linus Torvalds 已提交
222 223 224 225 226 227 228
 */
void
smp_send_reschedule (int cpu)
{
	platform_send_ipi(cpu, IA64_IPI_RESCHEDULE, IA64_IPI_DM_INT, 0);
}

229
/*
S
Simon Arlott 已提交
230
 * Called with preemption disabled.
231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284
 */
static void
smp_send_local_flush_tlb (int cpu)
{
	platform_send_ipi(cpu, IA64_IPI_LOCAL_TLB_FLUSH, IA64_IPI_DM_INT, 0);
}

void
smp_local_flush_tlb(void)
{
	/*
	 * Use atomic ops. Otherwise, the load/increment/store sequence from
	 * a "++" operation can have the line stolen between the load & store.
	 * The overhead of the atomic op in negligible in this case & offers
	 * significant benefit for the brief periods where lots of cpus
	 * are simultaneously flushing TLBs.
	 */
	ia64_fetchadd(1, &local_tlb_flush_counts[smp_processor_id()].count, acq);
	local_flush_tlb_all();
}

#define FLUSH_DELAY	5 /* Usec backoff to eliminate excessive cacheline bouncing */

void
smp_flush_tlb_cpumask(cpumask_t xcpumask)
{
	unsigned int *counts = __ia64_per_cpu_var(shadow_flush_counts);
	cpumask_t cpumask = xcpumask;
	int mycpu, cpu, flush_mycpu = 0;

	preempt_disable();
	mycpu = smp_processor_id();

	for_each_cpu_mask(cpu, cpumask)
		counts[cpu] = local_tlb_flush_counts[cpu].count;

	mb();
	for_each_cpu_mask(cpu, cpumask) {
		if (cpu == mycpu)
			flush_mycpu = 1;
		else
			smp_send_local_flush_tlb(cpu);
	}

	if (flush_mycpu)
		smp_local_flush_tlb();

	for_each_cpu_mask(cpu, cpumask)
		while(counts[cpu] == local_tlb_flush_counts[cpu].count)
			udelay(FLUSH_DELAY);

	preempt_enable();
}

L
Linus Torvalds 已提交
285 286 287 288 289 290 291 292 293
void
smp_flush_tlb_all (void)
{
	on_each_cpu((void (*)(void *))local_flush_tlb_all, NULL, 1, 1);
}

void
smp_flush_tlb_mm (struct mm_struct *mm)
{
294
	preempt_disable();
L
Linus Torvalds 已提交
295 296 297 298
	/* this happens for the common case of a single-threaded fork():  */
	if (likely(mm == current->active_mm && atomic_read(&mm->mm_users) == 1))
	{
		local_finish_flush_tlb_mm(mm);
299
		preempt_enable();
L
Linus Torvalds 已提交
300 301 302
		return;
	}

303
	preempt_enable();
L
Linus Torvalds 已提交
304 305 306 307 308 309 310 311 312 313
	/*
	 * We could optimize this further by using mm->cpu_vm_mask to track which CPUs
	 * have been running in the address space.  It's not clear that this is worth the
	 * trouble though: to avoid races, we have to raise the IPI on the target CPU
	 * anyhow, and once a CPU is interrupted, the cost of local_flush_tlb_all() is
	 * rather trivial.
	 */
	on_each_cpu((void (*)(void *))local_finish_flush_tlb_mm, mm, 1, 1);
}

314
void arch_send_call_function_single_ipi(int cpu)
315
{
316
	send_IPI_single(cpu, IPI_CALL_FUNC_SINGLE);
317 318
}

319
void arch_send_call_function_ipi(cpumask_t mask)
L
Linus Torvalds 已提交
320
{
321
	send_IPI_mask(mask, IPI_CALL_FUNC);
L
Linus Torvalds 已提交
322 323 324 325 326 327 328 329 330 331 332
}

/*
 * this function calls the 'stop' function on all other CPUs in the system.
 */
void
smp_send_stop (void)
{
	send_IPI_allbutself(IPI_CPU_STOP);
}

333
int
L
Linus Torvalds 已提交
334 335 336 337
setup_profiling_timer (unsigned int multiplier)
{
	return -EINVAL;
}