setup.c 14.0 KB
Newer Older
1 2 3 4 5 6 7 8 9 10
/*
 * Machine specific setup for xen
 *
 * Jeremy Fitzhardinge <jeremy@xensource.com>, XenSource Inc, 2007
 */

#include <linux/module.h>
#include <linux/sched.h>
#include <linux/mm.h>
#include <linux/pm.h>
11
#include <linux/memblock.h>
12
#include <linux/cpuidle.h>
13
#include <linux/cpufreq.h>
14 15

#include <asm/elf.h>
R
Roland McGrath 已提交
16
#include <asm/vdso.h>
17 18
#include <asm/e820.h>
#include <asm/setup.h>
19
#include <asm/acpi.h>
20 21 22
#include <asm/xen/hypervisor.h>
#include <asm/xen/hypercall.h>

23
#include <xen/xen.h>
24
#include <xen/page.h>
25
#include <xen/interface/callback.h>
I
Ian Campbell 已提交
26
#include <xen/interface/memory.h>
27 28 29
#include <xen/interface/physdev.h>
#include <xen/features.h>
#include "xen-ops.h"
30
#include "vdso.h"
31 32 33 34

/* These are code, but not functions.  Defined in entry.S */
extern const char xen_hypervisor_callback[];
extern const char xen_failsafe_callback[];
T
Tej 已提交
35 36 37
extern void xen_sysenter_target(void);
extern void xen_syscall_target(void);
extern void xen_syscall32_target(void);
38

39
/* Amount of extra memory space we add to the e820 ranges */
40
struct xen_memory_region xen_extra_mem[XEN_EXTRA_MEM_MAX_REGIONS] __initdata;
41

42 43 44
/* Number of pages released from the initial allocation. */
unsigned long xen_released_pages;

45 46 47 48 49 50 51 52 53 54 55 56
/* 
 * The maximum amount of extra memory compared to the base size.  The
 * main scaling factor is the size of struct page.  At extreme ratios
 * of base:extra, all the base memory can be filled with page
 * structures for the extra memory, leaving no space for anything
 * else.
 * 
 * 10x seems like a reasonable balance between scaling flexibility and
 * leaving a practically usable system.
 */
#define EXTRA_MEM_RATIO		(10)

57
static void __init xen_add_extra_mem(u64 start, u64 size)
58
{
59
	unsigned long pfn;
60
	int i;
61

62 63 64 65 66 67 68 69 70 71 72 73 74 75 76
	for (i = 0; i < XEN_EXTRA_MEM_MAX_REGIONS; i++) {
		/* Add new region. */
		if (xen_extra_mem[i].size == 0) {
			xen_extra_mem[i].start = start;
			xen_extra_mem[i].size  = size;
			break;
		}
		/* Append to existing region. */
		if (xen_extra_mem[i].start + xen_extra_mem[i].size == start) {
			xen_extra_mem[i].size += size;
			break;
		}
	}
	if (i == XEN_EXTRA_MEM_MAX_REGIONS)
		printk(KERN_WARNING "Warning: not enough extra memory regions\n");
77

78
	memblock_reserve(start, size);
79

80
	xen_max_p2m_pfn = PFN_DOWN(start + size);
81

82
	for (pfn = PFN_DOWN(start); pfn <= xen_max_p2m_pfn; pfn++)
83
		__set_phys_to_machine(pfn, INVALID_P2M_ENTRY);
84 85
}

86 87
static unsigned long __init xen_release_chunk(unsigned long start,
					      unsigned long end)
88 89 90 91 92 93
{
	struct xen_memory_reservation reservation = {
		.address_bits = 0,
		.extent_order = 0,
		.domid        = DOMID_SELF
	};
94
	unsigned long len = 0;
95 96 97
	unsigned long pfn;
	int ret;

98 99 100 101 102 103 104 105 106 107 108 109
	for(pfn = start; pfn < end; pfn++) {
		unsigned long mfn = pfn_to_mfn(pfn);

		/* Make sure pfn exists to start with */
		if (mfn == INVALID_P2M_ENTRY || mfn_to_pfn(mfn) != pfn)
			continue;

		set_xen_guest_handle(reservation.extent_start, &mfn);
		reservation.nr_extents = 1;

		ret = HYPERVISOR_memory_op(XENMEM_decrease_reservation,
					   &reservation);
110
		WARN(ret != 1, "Failed to release pfn %lx err=%d\n", pfn, ret);
111
		if (ret == 1) {
112
			__set_phys_to_machine(pfn, INVALID_P2M_ENTRY);
113 114 115
			len++;
		}
	}
116 117 118
	if (len)
		printk(KERN_INFO "Freeing  %lx-%lx pfn range: %lu pages freed\n",
		       start, end, len);
119 120 121

	return len;
}
122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138
static unsigned long __init xen_populate_physmap(unsigned long start,
						 unsigned long end)
{
	struct xen_memory_reservation reservation = {
		.address_bits = 0,
		.extent_order = 0,
		.domid        = DOMID_SELF
	};
	unsigned long len = 0;
	int ret;

	for (pfn = start; pfn < end; pfn++) {
		unsigned long frame;

		/* Make sure pfn does not exists to start with */
		if (pfn_to_mfn(pfn) != INVALID_P2M_ENTRY)
			continue;
139

140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220
		frame = pfn;
		set_xen_guest_handle(reservation.extent_start, &frame);
		reservation.nr_extents = 1;

		ret = HYPERVISOR_memory_op(XENMEM_populate_physmap, &reservation);
		WARN(ret != 1, "Failed to populate pfn %lx err=%d\n", pfn, ret);
		if (ret == 1) {
			if (!early_set_phys_to_machine(pfn, frame)) {
				set_xen_guest_handle(reservation.extent_start, &frame);
				reservation.nr_extents = 1;
				ret = HYPERVISOR_memory_op(XENMEM_decrease_reservation,
							&reservation);
				break;
			}
			len++;
		} else
			break;
	}
	if (len)
		printk(KERN_INFO "Populated %lx-%lx pfn range: %lu pages added\n",
		       start, end, len);
	return len;
}
static unsigned long __init xen_populate_chunk(
	const struct e820entry *list, size_t map_size,
	unsigned long max_pfn, unsigned long *last_pfn,
	unsigned long credits_left)
{
	const struct e820entry *entry;
	unsigned int i;
	unsigned long done = 0;
	unsigned long dest_pfn;

	for (i = 0, entry = list; i < map_size; i++, entry++) {
		unsigned long credits = credits_left;
		unsigned long s_pfn;
		unsigned long e_pfn;
		unsigned long pfns;
		long capacity;

		if (credits <= 0)
			break;

		if (entry->type != E820_RAM)
			continue;

		e_pfn = PFN_UP(entry->addr + entry->size);

		/* We only care about E820 after the xen_start_info->nr_pages */
		if (e_pfn <= max_pfn)
			continue;

		s_pfn = PFN_DOWN(entry->addr);
		/* If the E820 falls within the nr_pages, we want to start
		 * at the nr_pages PFN.
		 * If that would mean going past the E820 entry, skip it
		 */
		if (s_pfn <= max_pfn) {
			capacity = e_pfn - max_pfn;
			dest_pfn = max_pfn;
		} else {
			/* last_pfn MUST be within E820_RAM regions */
			if (*last_pfn && e_pfn >= *last_pfn)
				s_pfn = *last_pfn;
			capacity = e_pfn - s_pfn;
			dest_pfn = s_pfn;
		}
		/* If we had filled this E820_RAM entry, go to the next one. */
		if (capacity <= 0)
			continue;

		if (credits > capacity)
			credits = capacity;

		pfns = xen_populate_physmap(dest_pfn, dest_pfn + credits);
		done += pfns;
		credits_left -= pfns;
		*last_pfn = (dest_pfn + pfns);
	}
	return done;
}
221 222
static unsigned long __init xen_set_identity_and_release(
	const struct e820entry *list, size_t map_size, unsigned long nr_pages)
223
{
224
	phys_addr_t start = 0;
225
	unsigned long released = 0;
226
	unsigned long identity = 0;
227
	const struct e820entry *entry;
228 229
	int i;

230 231 232 233 234 235 236 237 238 239 240
	/*
	 * Combine non-RAM regions and gaps until a RAM region (or the
	 * end of the map) is reached, then set the 1:1 map and
	 * release the pages (if available) in those non-RAM regions.
	 *
	 * The combined non-RAM regions are rounded to a whole number
	 * of pages so any partial pages are accessible via the 1:1
	 * mapping.  This is needed for some BIOSes that put (for
	 * example) the DMI tables in a reserved region that begins on
	 * a non-page boundary.
	 */
241
	for (i = 0, entry = list; i < map_size; i++, entry++) {
242 243 244 245
		phys_addr_t end = entry->addr + entry->size;
		if (entry->type == E820_RAM || i == map_size - 1) {
			unsigned long start_pfn = PFN_DOWN(start);
			unsigned long end_pfn = PFN_UP(end);
246

247 248
			if (entry->type == E820_RAM)
				end_pfn = PFN_UP(entry->addr);
249

250 251 252 253
			if (start_pfn < end_pfn) {
				if (start_pfn < nr_pages)
					released += xen_release_chunk(
						start_pfn, min(end_pfn, nr_pages));
254 255

				identity += set_phys_range_identity(
256 257 258
					start_pfn, end_pfn);
			}
			start = end;
259 260
		}
	}
261

262 263 264 265
	if (released)
		printk(KERN_INFO "Released %lu pages of unused memory\n", released);
	if (identity)
		printk(KERN_INFO "Set %ld page(s) to 1-1 mapping\n", identity);
266 267

	return released;
268
}
269 270 271 272 273 274 275

static unsigned long __init xen_get_max_pages(void)
{
	unsigned long max_pages = MAX_DOMAIN_PAGES;
	domid_t domid = DOMID_SELF;
	int ret;

276 277 278 279 280 281 282 283 284 285 286 287 288 289 290
	/*
	 * For the initial domain we use the maximum reservation as
	 * the maximum page.
	 *
	 * For guest domains the current maximum reservation reflects
	 * the current maximum rather than the static maximum. In this
	 * case the e820 map provided to us will cover the static
	 * maximum region.
	 */
	if (xen_initial_domain()) {
		ret = HYPERVISOR_memory_op(XENMEM_maximum_reservation, &domid);
		if (ret > 0)
			max_pages = ret;
	}

291 292 293
	return min(max_pages, MAX_DOMAIN_PAGES);
}

294 295 296 297 298 299 300 301 302 303 304 305 306
static void xen_align_and_add_e820_region(u64 start, u64 size, int type)
{
	u64 end = start + size;

	/* Align RAM regions to page boundaries. */
	if (type == E820_RAM) {
		start = PAGE_ALIGN(start);
		end &= ~((u64)PAGE_SIZE - 1);
	}

	e820_add_region(start, end - start, type);
}

307 308 309 310 311
/**
 * machine_specific_memory_setup - Hook for machine specific memory setup.
 **/
char * __init xen_memory_setup(void)
{
I
Ian Campbell 已提交
312 313
	static struct e820entry map[E820MAX] __initdata;

314
	unsigned long max_pfn = xen_start_info->nr_pages;
I
Ian Campbell 已提交
315 316 317
	unsigned long long mem_end;
	int rc;
	struct xen_memory_map memmap;
318
	unsigned long max_pages;
319
	unsigned long last_pfn = 0;
320
	unsigned long extra_pages = 0;
321
	unsigned long populated;
I
Ian Campbell 已提交
322
	int i;
I
Ian Campbell 已提交
323
	int op;
324

325
	max_pfn = min(MAX_DOMAIN_PAGES, max_pfn);
I
Ian Campbell 已提交
326 327 328 329 330
	mem_end = PFN_PHYS(max_pfn);

	memmap.nr_entries = E820MAX;
	set_xen_guest_handle(memmap.buffer, map);

I
Ian Campbell 已提交
331 332 333 334
	op = xen_initial_domain() ?
		XENMEM_machine_memory_map :
		XENMEM_memory_map;
	rc = HYPERVISOR_memory_op(op, &memmap);
I
Ian Campbell 已提交
335
	if (rc == -ENOSYS) {
336
		BUG_ON(xen_initial_domain());
I
Ian Campbell 已提交
337 338 339 340 341 342 343 344 345
		memmap.nr_entries = 1;
		map[0].addr = 0ULL;
		map[0].size = mem_end;
		/* 8MB slack (to balance backend allocations). */
		map[0].size += 8ULL << 20;
		map[0].type = E820_RAM;
		rc = 0;
	}
	BUG_ON(rc);
346

347 348 349 350 351 352 353
	/* Make sure the Xen-supplied memory map is well-ordered. */
	sanitize_e820_map(map, memmap.nr_entries, &memmap.nr_entries);

	max_pages = xen_get_max_pages();
	if (max_pages > max_pfn)
		extra_pages += max_pages - max_pfn;

354 355 356 357 358 359 360
	/*
	 * Set P2M for all non-RAM pages and E820 gaps to be identity
	 * type PFNs.  Any RAM pages that would be made inaccesible by
	 * this are first released.
	 */
	xen_released_pages = xen_set_identity_and_release(
		map, memmap.nr_entries, max_pfn);
361

362 363 364 365 366 367 368 369 370 371 372 373
	/*
	 * Populate back the non-RAM pages and E820 gaps that had been
	 * released. */
	populated = xen_populate_chunk(map, memmap.nr_entries,
			max_pfn, &last_pfn, xen_released_pages);

	extra_pages += (xen_released_pages - populated);

	if (last_pfn > max_pfn) {
		max_pfn = min(MAX_DOMAIN_PAGES, last_pfn);
		mem_end = PFN_PHYS(max_pfn);
	}
374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401
	/*
	 * Clamp the amount of extra memory to a EXTRA_MEM_RATIO
	 * factor the base size.  On non-highmem systems, the base
	 * size is the full initial memory allocation; on highmem it
	 * is limited to the max size of lowmem, so that it doesn't
	 * get completely filled.
	 *
	 * In principle there could be a problem in lowmem systems if
	 * the initial memory is also very large with respect to
	 * lowmem, but we won't try to deal with that here.
	 */
	extra_pages = min(EXTRA_MEM_RATIO * min(max_pfn, PFN_DOWN(MAXMEM)),
			  extra_pages);
	i = 0;
	while (i < memmap.nr_entries) {
		u64 addr = map[i].addr;
		u64 size = map[i].size;
		u32 type = map[i].type;

		if (type == E820_RAM) {
			if (addr < mem_end) {
				size = min(size, mem_end - addr);
			} else if (extra_pages) {
				size = min(size, (u64)extra_pages * PAGE_SIZE);
				extra_pages -= size / PAGE_SIZE;
				xen_add_extra_mem(addr, size);
			} else
				type = E820_UNUSABLE;
402 403
		}

404
		xen_align_and_add_e820_region(addr, size, type);
405

406 407 408 409
		map[i].addr += size;
		map[i].size -= size;
		if (map[i].size == 0)
			i++;
I
Ian Campbell 已提交
410
	}
411 412

	/*
413 414
	 * In domU, the ISA region is normal, usable memory, but we
	 * reserve ISA memory anyway because too many things poke
415 416 417 418
	 * about in there.
	 */
	e820_add_region(ISA_START_ADDRESS, ISA_END_ADDRESS - ISA_START_ADDRESS,
			E820_RESERVED);
419

420 421 422 423 424 425
	/*
	 * Reserve Xen bits:
	 *  - mfn_list
	 *  - xen_start_info
	 * See comment above "struct start_info" in <xen/interface/xen.h>
	 */
426 427
	memblock_reserve(__pa(xen_start_info->mfn_list),
			 xen_start_info->pt_base - xen_start_info->mfn_list);
428 429 430

	sanitize_e820_map(e820.map, ARRAY_SIZE(e820.map), &e820.nr_map);

431 432 433
	return "Xen";
}

434 435
/*
 * Set the bit indicating "nosegneg" library variants should be used.
436 437
 * We only need to bother in pure 32-bit mode; compat 32-bit processes
 * can have un-truncated segments, so wrapping around is allowed.
438
 */
439
static void __init fiddle_vdso(void)
440
{
441 442 443 444 445
#ifdef CONFIG_X86_32
	u32 *mask;
	mask = VDSO32_SYMBOL(&vdso32_int80_start, NOTE_MASK);
	*mask |= 1 << VDSO_NOTE_NONEGSEG_BIT;
	mask = VDSO32_SYMBOL(&vdso32_sysenter_start, NOTE_MASK);
446
	*mask |= 1 << VDSO_NOTE_NONEGSEG_BIT;
447
#endif
448 449
}

450
static int __cpuinit register_callback(unsigned type, const void *func)
451
{
452 453 454
	struct callback_register callback = {
		.type = type,
		.address = XEN_CALLBACK(__KERNEL_CS, func),
455 456 457
		.flags = CALLBACKF_mask_events,
	};

458 459 460 461 462
	return HYPERVISOR_callback_op(CALLBACKOP_register, &callback);
}

void __cpuinit xen_enable_sysenter(void)
{
463
	int ret;
464
	unsigned sysenter_feature;
465 466

#ifdef CONFIG_X86_32
467
	sysenter_feature = X86_FEATURE_SEP;
468
#else
469
	sysenter_feature = X86_FEATURE_SYSENTER32;
470
#endif
471

472 473 474
	if (!boot_cpu_has(sysenter_feature))
		return;

475
	ret = register_callback(CALLBACKTYPE_sysenter, xen_sysenter_target);
476 477
	if(ret != 0)
		setup_clear_cpu_cap(sysenter_feature);
478 479
}

480 481 482 483 484 485 486
void __cpuinit xen_enable_syscall(void)
{
#ifdef CONFIG_X86_64
	int ret;

	ret = register_callback(CALLBACKTYPE_syscall, xen_syscall_target);
	if (ret != 0) {
487
		printk(KERN_ERR "Failed to set syscall callback: %d\n", ret);
488 489 490 491 492
		/* Pretty fatal; 64-bit userspace has no other
		   mechanism for syscalls. */
	}

	if (boot_cpu_has(X86_FEATURE_SYSCALL32)) {
493 494
		ret = register_callback(CALLBACKTYPE_syscall32,
					xen_syscall32_target);
495
		if (ret != 0)
496
			setup_clear_cpu_cap(X86_FEATURE_SYSCALL32);
497 498 499 500
	}
#endif /* CONFIG_X86_64 */
}

501 502
void __init xen_arch_setup(void)
{
503 504
	xen_panic_handler_init();

505 506 507 508
	HYPERVISOR_vm_assist(VMASST_CMD_enable, VMASST_TYPE_4gb_segments);
	HYPERVISOR_vm_assist(VMASST_CMD_enable, VMASST_TYPE_writable_pagetables);

	if (!xen_feature(XENFEAT_auto_translated_physmap))
T
Tej 已提交
509 510
		HYPERVISOR_vm_assist(VMASST_CMD_enable,
				     VMASST_TYPE_pae_extended_cr3);
511

512 513 514
	if (register_callback(CALLBACKTYPE_event, xen_hypervisor_callback) ||
	    register_callback(CALLBACKTYPE_failsafe, xen_failsafe_callback))
		BUG();
515

516
	xen_enable_sysenter();
517
	xen_enable_syscall();
518

519 520 521 522 523 524 525 526 527 528 529
#ifdef CONFIG_ACPI
	if (!(xen_start_info->flags & SIF_INITDOMAIN)) {
		printk(KERN_INFO "ACPI in unprivileged domain disabled\n");
		disable_acpi();
	}
#endif

	memcpy(boot_command_line, xen_start_info->cmd_line,
	       MAX_GUEST_CMDLINE > COMMAND_LINE_SIZE ?
	       COMMAND_LINE_SIZE : MAX_GUEST_CMDLINE);

J
Jeremy Fitzhardinge 已提交
530 531 532 533
	/* Set up idle, making sure it calls safe_halt() pvop */
#ifdef CONFIG_X86_32
	boot_cpu_data.hlt_works_ok = 1;
#endif
534
	disable_cpuidle();
535
	disable_cpufreq();
536
	WARN_ON(set_pm_idle_to_default());
537
	fiddle_vdso();
538
}