page_tables.c 36.2 KB
Newer Older
R
Rusty Russell 已提交
1 2
/*P:700
 * The pagetable code, on the other hand, still shows the scars of
3 4 5
 * previous encounters.  It's functional, and as neat as it can be in the
 * circumstances, but be wary, for these things are subtle and break easily.
 * The Guest provides a virtual to physical mapping, but we can neither trust
6
 * it nor use it: we verify and convert it here then point the CPU to the
R
Rusty Russell 已提交
7 8
 * converted Guest pages when running the Guest.
:*/
9 10

/* Copyright (C) Rusty Russell IBM Corporation 2006.
R
Rusty Russell 已提交
11 12
 * GPL v2 and any later version */
#include <linux/mm.h>
13
#include <linux/gfp.h>
R
Rusty Russell 已提交
14 15 16 17 18
#include <linux/types.h>
#include <linux/spinlock.h>
#include <linux/random.h>
#include <linux/percpu.h>
#include <asm/tlbflush.h>
19
#include <asm/uaccess.h>
R
Rusty Russell 已提交
20 21
#include "lg.h"

R
Rusty Russell 已提交
22 23
/*M:008
 * We hold reference to pages, which prevents them from being swapped.
24 25
 * It'd be nice to have a callback in the "struct mm_struct" when Linux wants
 * to swap out.  If we had this, and a shrinker callback to trim PTE pages, we
R
Rusty Russell 已提交
26 27
 * could probably consider launching Guests as non-root.
:*/
28

R
Rusty Russell 已提交
29 30 31
/*H:300
 * The Page Table Code
 *
R
Rusty Russell 已提交
32 33 34 35
 * We use two-level page tables for the Guest, or three-level with PAE.  If
 * you're not entirely comfortable with virtual addresses, physical addresses
 * and page tables then I recommend you review arch/x86/lguest/boot.c's "Page
 * Table Handling" (with diagrams!).
R
Rusty Russell 已提交
36 37 38 39 40 41 42 43 44
 *
 * The Guest keeps page tables, but we maintain the actual ones here: these are
 * called "shadow" page tables.  Which is a very Guest-centric name: these are
 * the real page tables the CPU uses, although we keep them up to date to
 * reflect the Guest's.  (See what I mean about weird naming?  Since when do
 * shadows reflect anything?)
 *
 * Anyway, this is the most complicated part of the Host code.  There are seven
 * parts to this:
R
Rusty Russell 已提交
45 46 47
 *  (i) Looking up a page table entry when the Guest faults,
 *  (ii) Making sure the Guest stack is mapped,
 *  (iii) Setting up a page table entry when the Guest tells us one has changed,
R
Rusty Russell 已提交
48
 *  (iv) Switching page tables,
R
Rusty Russell 已提交
49
 *  (v) Flushing (throwing away) page tables,
R
Rusty Russell 已提交
50 51
 *  (vi) Mapping the Switcher when the Guest is about to run,
 *  (vii) Setting up the page tables initially.
R
Rusty Russell 已提交
52
:*/
R
Rusty Russell 已提交
53

R
Rusty Russell 已提交
54
/*
R
Rusty Russell 已提交
55 56
 * The Switcher uses the complete top PTE page.  That's 1024 PTE entries (4MB)
 * or 512 PTE entries with PAE (2MB).
R
Rusty Russell 已提交
57
 */
58
#define SWITCHER_PGD_INDEX (PTRS_PER_PGD - 1)
R
Rusty Russell 已提交
59

R
Rusty Russell 已提交
60 61 62 63
/*
 * For PAE we need the PMD index as well. We use the last 2MB, so we
 * will need the last pmd entry of the last pmd page.
 */
M
Matias Zabaljauregui 已提交
64 65 66 67 68 69 70
#ifdef CONFIG_X86_PAE
#define SWITCHER_PMD_INDEX 	(PTRS_PER_PMD - 1)
#define CHECK_GPGD_MASK		_PAGE_PRESENT
#else
#define CHECK_GPGD_MASK		_PAGE_TABLE
#endif

R
Rusty Russell 已提交
71 72
/*
 * We actually need a separate PTE page for each CPU.  Remember that after the
R
Rusty Russell 已提交
73
 * Switcher code itself comes two pages for each CPU, and we don't want this
R
Rusty Russell 已提交
74 75
 * CPU's guest to see the pages of any other CPU.
 */
76
static DEFINE_PER_CPU(pte_t *, switcher_pte_pages);
R
Rusty Russell 已提交
77 78
#define switcher_pte_page(cpu) per_cpu(switcher_pte_pages, cpu)

R
Rusty Russell 已提交
79 80
/*H:320
 * The page table code is curly enough to need helper functions to keep it
R
Rusty Russell 已提交
81 82
 * clear and clean.  The kernel itself provides many of them; one advantage
 * of insisting that the Guest and Host use the same CONFIG_PAE setting.
R
Rusty Russell 已提交
83
 *
84
 * There are two functions which return pointers to the shadow (aka "real")
R
Rusty Russell 已提交
85 86 87
 * page tables.
 *
 * spgd_addr() takes the virtual address and returns a pointer to the top-level
R
Rusty Russell 已提交
88 89
 * page directory entry (PGD) for that address.  Since we keep track of several
 * page tables, the "i" argument tells us which one we're interested in (it's
R
Rusty Russell 已提交
90 91
 * usually the current one).
 */
92
static pgd_t *spgd_addr(struct lg_cpu *cpu, u32 i, unsigned long vaddr)
R
Rusty Russell 已提交
93
{
94
	unsigned int index = pgd_index(vaddr);
R
Rusty Russell 已提交
95

R
Rusty Russell 已提交
96
	/* Return a pointer index'th pgd entry for the i'th page table. */
97
	return &cpu->lg->pgdirs[i].pgdir[index];
R
Rusty Russell 已提交
98 99
}

M
Matias Zabaljauregui 已提交
100
#ifdef CONFIG_X86_PAE
R
Rusty Russell 已提交
101 102
/*
 * This routine then takes the PGD entry given above, which contains the
M
Matias Zabaljauregui 已提交
103
 * address of the PMD page.  It then returns a pointer to the PMD entry for the
R
Rusty Russell 已提交
104 105
 * given address.
 */
M
Matias Zabaljauregui 已提交
106 107 108 109 110 111 112 113 114 115 116 117 118
static pmd_t *spmd_addr(struct lg_cpu *cpu, pgd_t spgd, unsigned long vaddr)
{
	unsigned int index = pmd_index(vaddr);
	pmd_t *page;

	/* You should never call this if the PGD entry wasn't valid */
	BUG_ON(!(pgd_flags(spgd) & _PAGE_PRESENT));
	page = __va(pgd_pfn(spgd) << PAGE_SHIFT);

	return &page[index];
}
#endif

R
Rusty Russell 已提交
119 120
/*
 * This routine then takes the page directory entry returned above, which
R
Rusty Russell 已提交
121
 * contains the address of the page table entry (PTE) page.  It then returns a
R
Rusty Russell 已提交
122 123
 * pointer to the PTE entry for the given address.
 */
M
Matias Zabaljauregui 已提交
124
static pte_t *spte_addr(struct lg_cpu *cpu, pgd_t spgd, unsigned long vaddr)
R
Rusty Russell 已提交
125
{
M
Matias Zabaljauregui 已提交
126 127 128 129 130 131 132
#ifdef CONFIG_X86_PAE
	pmd_t *pmd = spmd_addr(cpu, spgd, vaddr);
	pte_t *page = __va(pmd_pfn(*pmd) << PAGE_SHIFT);

	/* You should never call this if the PMD entry wasn't valid */
	BUG_ON(!(pmd_flags(*pmd) & _PAGE_PRESENT));
#else
133
	pte_t *page = __va(pgd_pfn(spgd) << PAGE_SHIFT);
R
Rusty Russell 已提交
134
	/* You should never call this if the PGD entry wasn't valid */
135
	BUG_ON(!(pgd_flags(spgd) & _PAGE_PRESENT));
M
Matias Zabaljauregui 已提交
136 137
#endif

138
	return &page[pte_index(vaddr)];
R
Rusty Russell 已提交
139 140
}

R
Rusty Russell 已提交
141
/*
R
Rusty Russell 已提交
142
 * These functions are just like the above, except they access the Guest
R
Rusty Russell 已提交
143 144
 * page tables.  Hence they return a Guest address.
 */
145
static unsigned long gpgd_addr(struct lg_cpu *cpu, unsigned long vaddr)
R
Rusty Russell 已提交
146
{
147
	unsigned int index = vaddr >> (PGDIR_SHIFT);
148
	return cpu->lg->pgdirs[cpu->cpu_pgd].gpgdir + index * sizeof(pgd_t);
R
Rusty Russell 已提交
149 150
}

M
Matias Zabaljauregui 已提交
151
#ifdef CONFIG_X86_PAE
R
Rusty Russell 已提交
152
/* Follow the PGD to the PMD. */
M
Matias Zabaljauregui 已提交
153
static unsigned long gpmd_addr(pgd_t gpgd, unsigned long vaddr)
R
Rusty Russell 已提交
154
{
155 156
	unsigned long gpage = pgd_pfn(gpgd) << PAGE_SHIFT;
	BUG_ON(!(pgd_flags(gpgd) & _PAGE_PRESENT));
M
Matias Zabaljauregui 已提交
157 158 159
	return gpage + pmd_index(vaddr) * sizeof(pmd_t);
}

R
Rusty Russell 已提交
160
/* Follow the PMD to the PTE. */
M
Matias Zabaljauregui 已提交
161
static unsigned long gpte_addr(struct lg_cpu *cpu,
R
Rusty Russell 已提交
162
			       pmd_t gpmd, unsigned long vaddr)
M
Matias Zabaljauregui 已提交
163
{
R
Rusty Russell 已提交
164
	unsigned long gpage = pmd_pfn(gpmd) << PAGE_SHIFT;
M
Matias Zabaljauregui 已提交
165 166

	BUG_ON(!(pmd_flags(gpmd) & _PAGE_PRESENT));
R
Rusty Russell 已提交
167 168
	return gpage + pte_index(vaddr) * sizeof(pte_t);
}
M
Matias Zabaljauregui 已提交
169
#else
R
Rusty Russell 已提交
170
/* Follow the PGD to the PTE (no mid-level for !PAE). */
R
Rusty Russell 已提交
171 172 173 174 175 176
static unsigned long gpte_addr(struct lg_cpu *cpu,
				pgd_t gpgd, unsigned long vaddr)
{
	unsigned long gpage = pgd_pfn(gpgd) << PAGE_SHIFT;

	BUG_ON(!(pgd_flags(gpgd) & _PAGE_PRESENT));
177
	return gpage + pte_index(vaddr) * sizeof(pte_t);
R
Rusty Russell 已提交
178
}
R
Rusty Russell 已提交
179
#endif
180 181
/*:*/

R
Rusty Russell 已提交
182
/*M:007
R
Rusty Russell 已提交
183 184 185
 * get_pfn is slow: we could probably try to grab batches of pages here as
 * an optimization (ie. pre-faulting).
:*/
R
Rusty Russell 已提交
186

R
Rusty Russell 已提交
187 188
/*H:350
 * This routine takes a page number given by the Guest and converts it to
R
Rusty Russell 已提交
189 190 191 192 193
 * an actual, physical page number.  It can fail for several reasons: the
 * virtual address might not be mapped by the Launcher, the write flag is set
 * and the page is read-only, or the write flag was set and the page was
 * shared so had to be copied, but we ran out of memory.
 *
194
 * This holds a reference to the page, so release_pte() is careful to put that
R
Rusty Russell 已提交
195 196
 * back.
 */
R
Rusty Russell 已提交
197 198 199
static unsigned long get_pfn(unsigned long virtpfn, int write)
{
	struct page *page;
200 201 202 203 204

	/* gup me one page at this address please! */
	if (get_user_pages_fast(virtpfn << PAGE_SHIFT, 1, write, &page) == 1)
		return page_to_pfn(page);

R
Rusty Russell 已提交
205
	/* This value indicates failure. */
206
	return -1UL;
R
Rusty Russell 已提交
207 208
}

R
Rusty Russell 已提交
209 210
/*H:340
 * Converting a Guest page table entry to a shadow (ie. real) page table
R
Rusty Russell 已提交
211 212
 * entry can be a little tricky.  The flags are (almost) the same, but the
 * Guest PTE contains a virtual page number: the CPU needs the real page
R
Rusty Russell 已提交
213 214
 * number.
 */
215
static pte_t gpte_to_spte(struct lg_cpu *cpu, pte_t gpte, int write)
R
Rusty Russell 已提交
216
{
217
	unsigned long pfn, base, flags;
R
Rusty Russell 已提交
218

R
Rusty Russell 已提交
219 220
	/*
	 * The Guest sets the global flag, because it thinks that it is using
R
Rusty Russell 已提交
221 222
	 * PGE.  We only told it to use PGE so it would tell us whether it was
	 * flushing a kernel mapping or a userspace mapping.  We don't actually
R
Rusty Russell 已提交
223 224
	 * use the global bit, so throw it away.
	 */
225
	flags = (pte_flags(gpte) & ~_PAGE_GLOBAL);
R
Rusty Russell 已提交
226

227
	/* The Guest's pages are offset inside the Launcher. */
228
	base = (unsigned long)cpu->lg->mem_base / PAGE_SIZE;
229

R
Rusty Russell 已提交
230 231
	/*
	 * We need a temporary "unsigned long" variable to hold the answer from
R
Rusty Russell 已提交
232 233
	 * get_pfn(), because it returns 0xFFFFFFFF on failure, which wouldn't
	 * fit in spte.pfn.  get_pfn() finds the real physical number of the
R
Rusty Russell 已提交
234 235
	 * page, given the virtual number.
	 */
236
	pfn = get_pfn(base + pte_pfn(gpte), write);
R
Rusty Russell 已提交
237
	if (pfn == -1UL) {
238
		kill_guest(cpu, "failed to get page %lu", pte_pfn(gpte));
R
Rusty Russell 已提交
239 240
		/*
		 * When we destroy the Guest, we'll go through the shadow page
R
Rusty Russell 已提交
241
		 * tables and release_pte() them.  Make sure we don't think
R
Rusty Russell 已提交
242 243
		 * this one is valid!
		 */
244
		flags = 0;
R
Rusty Russell 已提交
245
	}
246 247
	/* Now we assemble our shadow PTE from the page number and flags. */
	return pfn_pte(pfn, __pgprot(flags));
R
Rusty Russell 已提交
248 249
}

R
Rusty Russell 已提交
250
/*H:460 And to complete the chain, release_pte() looks like this: */
251
static void release_pte(pte_t pte)
R
Rusty Russell 已提交
252
{
R
Rusty Russell 已提交
253 254 255 256
	/*
	 * Remember that get_user_pages_fast() took a reference to the page, in
	 * get_pfn()?  We have to put it back now.
	 */
257
	if (pte_flags(pte) & _PAGE_PRESENT)
258
		put_page(pte_page(pte));
R
Rusty Russell 已提交
259
}
R
Rusty Russell 已提交
260
/*:*/
R
Rusty Russell 已提交
261

262
static bool check_gpte(struct lg_cpu *cpu, pte_t gpte)
R
Rusty Russell 已提交
263
{
264
	if ((pte_flags(gpte) & _PAGE_PSE) ||
265
	    pte_pfn(gpte) >= cpu->lg->pfn_limit) {
266
		kill_guest(cpu, "bad page table entry");
267 268 269
		return false;
	}
	return true;
R
Rusty Russell 已提交
270 271
}

272
static bool check_gpgd(struct lg_cpu *cpu, pgd_t gpgd)
R
Rusty Russell 已提交
273
{
M
Matias Zabaljauregui 已提交
274
	if ((pgd_flags(gpgd) & ~CHECK_GPGD_MASK) ||
275
	    (pgd_pfn(gpgd) >= cpu->lg->pfn_limit)) {
276
		kill_guest(cpu, "bad page directory entry");
277 278 279
		return false;
	}
	return true;
R
Rusty Russell 已提交
280 281
}

M
Matias Zabaljauregui 已提交
282
#ifdef CONFIG_X86_PAE
283
static bool check_gpmd(struct lg_cpu *cpu, pmd_t gpmd)
M
Matias Zabaljauregui 已提交
284 285
{
	if ((pmd_flags(gpmd) & ~_PAGE_TABLE) ||
286
	    (pmd_pfn(gpmd) >= cpu->lg->pfn_limit)) {
M
Matias Zabaljauregui 已提交
287
		kill_guest(cpu, "bad page middle directory entry");
288 289 290
		return false;
	}
	return true;
M
Matias Zabaljauregui 已提交
291 292 293
}
#endif

294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375
/*H:331
 * This is the core routine to walk the shadow page tables and find the page
 * table entry for a specific address.
 *
 * If allocate is set, then we allocate any missing levels, setting the flags
 * on the new page directory and mid-level directories using the arguments
 * (which are copied from the Guest's page table entries).
 */
static pte_t *find_spte(struct lg_cpu *cpu, unsigned long vaddr, bool allocate,
			int pgd_flags, int pmd_flags)
{
	pgd_t *spgd;
	/* Mid level for PAE. */
#ifdef CONFIG_X86_PAE
	pmd_t *spmd;
#endif

	/* Get top level entry. */
	spgd = spgd_addr(cpu, cpu->cpu_pgd, vaddr);
	if (!(pgd_flags(*spgd) & _PAGE_PRESENT)) {
		/* No shadow entry: allocate a new shadow PTE page. */
		unsigned long ptepage;

		/* If they didn't want us to allocate anything, stop. */
		if (!allocate)
			return NULL;

		ptepage = get_zeroed_page(GFP_KERNEL);
		/*
		 * This is not really the Guest's fault, but killing it is
		 * simple for this corner case.
		 */
		if (!ptepage) {
			kill_guest(cpu, "out of memory allocating pte page");
			return NULL;
		}
		/*
		 * And we copy the flags to the shadow PGD entry.  The page
		 * number in the shadow PGD is the page we just allocated.
		 */
		set_pgd(spgd, __pgd(__pa(ptepage) | pgd_flags));
	}

	/*
	 * Intel's Physical Address Extension actually uses three levels of
	 * page tables, so we need to look in the mid-level.
	 */
#ifdef CONFIG_X86_PAE
	/* Now look at the mid-level shadow entry. */
	spmd = spmd_addr(cpu, *spgd, vaddr);

	if (!(pmd_flags(*spmd) & _PAGE_PRESENT)) {
		/* No shadow entry: allocate a new shadow PTE page. */
		unsigned long ptepage;

		/* If they didn't want us to allocate anything, stop. */
		if (!allocate)
			return NULL;

		ptepage = get_zeroed_page(GFP_KERNEL);

		/*
		 * This is not really the Guest's fault, but killing it is
		 * simple for this corner case.
		 */
		if (!ptepage) {
			kill_guest(cpu, "out of memory allocating pmd page");
			return NULL;
		}

		/*
		 * And we copy the flags to the shadow PMD entry.  The page
		 * number in the shadow PMD is the page we just allocated.
		 */
		set_pmd(spmd, __pmd(__pa(ptepage) | pmd_flags));
	}
#endif

	/* Get the pointer to the shadow PTE entry we're going to set. */
	return spte_addr(cpu, *spgd, vaddr);
}

R
Rusty Russell 已提交
376
/*H:330
R
Rusty Russell 已提交
377
 * (i) Looking up a page table entry when the Guest faults.
R
Rusty Russell 已提交
378 379 380 381 382 383 384
 *
 * We saw this call in run_guest(): when we see a page fault in the Guest, we
 * come here.  That's because we only set up the shadow page tables lazily as
 * they're needed, so we get page faults all the time and quietly fix them up
 * and return to the Guest without it knowing.
 *
 * If we fixed up the fault (ie. we mapped the address), this routine returns
R
Rusty Russell 已提交
385 386
 * true.  Otherwise, it was a real fault and we need to tell the Guest.
 */
387
bool demand_page(struct lg_cpu *cpu, unsigned long vaddr, int errcode)
R
Rusty Russell 已提交
388 389
{
	unsigned long gpte_ptr;
390 391
	pte_t gpte;
	pte_t *spte;
M
Matias Zabaljauregui 已提交
392
	pmd_t gpmd;
393
	pgd_t gpgd;
M
Matias Zabaljauregui 已提交
394

395 396 397 398
	/* We never demand page the Switcher, so trying is a mistake. */
	if (vaddr >= switcher_addr)
		return false;

R
Rusty Russell 已提交
399
	/* First step: get the top-level Guest page table entry. */
400 401 402 403 404 405 406 407
	if (unlikely(cpu->linear_pages)) {
		/* Faking up a linear mapping. */
		gpgd = __pgd(CHECK_GPGD_MASK);
	} else {
		gpgd = lgread(cpu, gpgd_addr(cpu, vaddr), pgd_t);
		/* Toplevel not present?  We can't map it in. */
		if (!(pgd_flags(gpgd) & _PAGE_PRESENT))
			return false;
R
Rusty Russell 已提交
408

409 410 411
		/* 
		 * This kills the Guest if it has weird flags or tries to
		 * refer to a "physical" address outside the bounds.
R
Rusty Russell 已提交
412
		 */
413 414
		if (!check_gpgd(cpu, gpgd))
			return false;
R
Rusty Russell 已提交
415 416
	}

417 418 419
	/* This "mid-level" entry is only used for non-linear, PAE mode. */
	gpmd = __pmd(_PAGE_TABLE);

M
Matias Zabaljauregui 已提交
420
#ifdef CONFIG_X86_PAE
421
	if (likely(!cpu->linear_pages)) {
422 423 424 425
		gpmd = lgread(cpu, gpmd_addr(gpgd, vaddr), pmd_t);
		/* Middle level not present?  We can't map it in. */
		if (!(pmd_flags(gpmd) & _PAGE_PRESENT))
			return false;
M
Matias Zabaljauregui 已提交
426

427 428 429
		/* 
		 * This kills the Guest if it has weird flags or tries to
		 * refer to a "physical" address outside the bounds.
R
Rusty Russell 已提交
430
		 */
431 432
		if (!check_gpmd(cpu, gpmd))
			return false;
M
Matias Zabaljauregui 已提交
433
	}
R
Rusty Russell 已提交
434

R
Rusty Russell 已提交
435 436 437 438
	/*
	 * OK, now we look at the lower level in the Guest page table: keep its
	 * address, because we might update it later.
	 */
R
Rusty Russell 已提交
439 440
	gpte_ptr = gpte_addr(cpu, gpmd, vaddr);
#else
R
Rusty Russell 已提交
441 442 443 444
	/*
	 * OK, now we look at the lower level in the Guest page table: keep its
	 * address, because we might update it later.
	 */
M
Matias Zabaljauregui 已提交
445
	gpte_ptr = gpte_addr(cpu, gpgd, vaddr);
R
Rusty Russell 已提交
446
#endif
R
Rusty Russell 已提交
447

448 449 450 451 452 453 454
	if (unlikely(cpu->linear_pages)) {
		/* Linear?  Make up a PTE which points to same page. */
		gpte = __pte((vaddr & PAGE_MASK) | _PAGE_RW | _PAGE_PRESENT);
	} else {
		/* Read the actual PTE value. */
		gpte = lgread(cpu, gpte_ptr, pte_t);
	}
R
Rusty Russell 已提交
455

R
Rusty Russell 已提交
456
	/* If this page isn't in the Guest page tables, we can't page it in. */
457
	if (!(pte_flags(gpte) & _PAGE_PRESENT))
458
		return false;
R
Rusty Russell 已提交
459

R
Rusty Russell 已提交
460 461 462 463
	/*
	 * Check they're not trying to write to a page the Guest wants
	 * read-only (bit 2 of errcode == write).
	 */
464
	if ((errcode & 2) && !(pte_flags(gpte) & _PAGE_RW))
465
		return false;
R
Rusty Russell 已提交
466

R
Rusty Russell 已提交
467
	/* User access to a kernel-only page? (bit 3 == user access) */
468
	if ((errcode & 4) && !(pte_flags(gpte) & _PAGE_USER))
469
		return false;
R
Rusty Russell 已提交
470

R
Rusty Russell 已提交
471 472 473 474
	/*
	 * Check that the Guest PTE flags are OK, and the page number is below
	 * the pfn_limit (ie. not mapping the Launcher binary).
	 */
475 476
	if (!check_gpte(cpu, gpte))
		return false;
R
Rusty Russell 已提交
477

R
Rusty Russell 已提交
478
	/* Add the _PAGE_ACCESSED and (for a write) _PAGE_DIRTY flag */
479
	gpte = pte_mkyoung(gpte);
R
Rusty Russell 已提交
480
	if (errcode & 2)
481
		gpte = pte_mkdirty(gpte);
R
Rusty Russell 已提交
482

R
Rusty Russell 已提交
483
	/* Get the pointer to the shadow PTE entry we're going to set. */
484 485 486
	spte = find_spte(cpu, vaddr, true, pgd_flags(gpgd), pmd_flags(gpmd));
	if (!spte)
		return false;
R
Rusty Russell 已提交
487 488 489 490 491

	/*
	 * If there was a valid shadow PTE entry here before, we release it.
	 * This can happen with a write to a previously read-only entry.
	 */
R
Rusty Russell 已提交
492 493
	release_pte(*spte);

R
Rusty Russell 已提交
494 495 496 497
	/*
	 * If this is a write, we insist that the Guest page is writable (the
	 * final arg to gpte_to_spte()).
	 */
498
	if (pte_dirty(gpte))
499
		*spte = gpte_to_spte(cpu, gpte, 1);
500
	else
R
Rusty Russell 已提交
501 502
		/*
		 * If this is a read, don't set the "writable" bit in the page
R
Rusty Russell 已提交
503
		 * table entry, even if the Guest says it's writable.  That way
R
Rusty Russell 已提交
504
		 * we will come back here when a write does actually occur, so
R
Rusty Russell 已提交
505 506
		 * we can update the Guest's _PAGE_DIRTY flag.
		 */
507
		set_pte(spte, gpte_to_spte(cpu, pte_wrprotect(gpte), 0));
R
Rusty Russell 已提交
508

R
Rusty Russell 已提交
509 510 511 512
	/*
	 * Finally, we write the Guest PTE entry back: we've set the
	 * _PAGE_ACCESSED and maybe the _PAGE_DIRTY flags.
	 */
513 514
	if (likely(!cpu->linear_pages))
		lgwrite(cpu, gpte_ptr, pte_t, gpte);
R
Rusty Russell 已提交
515

R
Rusty Russell 已提交
516 517
	/*
	 * The fault is fixed, the page table is populated, the mapping
R
Rusty Russell 已提交
518 519
	 * manipulated, the result returned and the code complete.  A small
	 * delay and a trace of alliteration are the only indications the Guest
R
Rusty Russell 已提交
520 521
	 * has that a page fault occurred at all.
	 */
522
	return true;
R
Rusty Russell 已提交
523 524
}

R
Rusty Russell 已提交
525 526
/*H:360
 * (ii) Making sure the Guest stack is mapped.
R
Rusty Russell 已提交
527
 *
R
Rusty Russell 已提交
528 529 530 531
 * Remember that direct traps into the Guest need a mapped Guest kernel stack.
 * pin_stack_pages() calls us here: we could simply call demand_page(), but as
 * we've seen that logic is quite long, and usually the stack pages are already
 * mapped, so it's overkill.
R
Rusty Russell 已提交
532 533
 *
 * This is a quick version which answers the question: is this virtual address
R
Rusty Russell 已提交
534 535
 * mapped by the shadow page tables, and is it writable?
 */
536
static bool page_writable(struct lg_cpu *cpu, unsigned long vaddr)
R
Rusty Russell 已提交
537
{
538
	pte_t *spte;
R
Rusty Russell 已提交
539
	unsigned long flags;
540 541 542 543 544

	/* You can't put your stack in the Switcher! */
	if (vaddr >= switcher_addr)
		return false;

545 546 547
	/* If there's no shadow PTE, it's not writable. */
	spte = find_spte(cpu, vaddr, false, 0, 0);
	if (!spte)
548
		return false;
R
Rusty Russell 已提交
549

R
Rusty Russell 已提交
550 551 552 553
	/*
	 * Check the flags on the pte entry itself: it must be present and
	 * writable.
	 */
554
	flags = pte_flags(*spte);
R
Rusty Russell 已提交
555 556 557
	return (flags & (_PAGE_PRESENT|_PAGE_RW)) == (_PAGE_PRESENT|_PAGE_RW);
}

R
Rusty Russell 已提交
558 559
/*
 * So, when pin_stack_pages() asks us to pin a page, we check if it's already
R
Rusty Russell 已提交
560
 * in the page tables, and if not, we call demand_page() with error code 2
R
Rusty Russell 已提交
561 562
 * (meaning "write").
 */
563
void pin_page(struct lg_cpu *cpu, unsigned long vaddr)
R
Rusty Russell 已提交
564
{
565
	if (!page_writable(cpu, vaddr) && !demand_page(cpu, vaddr, 2))
566
		kill_guest(cpu, "bad stack page %#lx", vaddr);
R
Rusty Russell 已提交
567
}
R
Rusty Russell 已提交
568
/*:*/
R
Rusty Russell 已提交
569

M
Matias Zabaljauregui 已提交
570 571 572 573 574 575 576 577 578 579 580 581 582
#ifdef CONFIG_X86_PAE
static void release_pmd(pmd_t *spmd)
{
	/* If the entry's not present, there's nothing to release. */
	if (pmd_flags(*spmd) & _PAGE_PRESENT) {
		unsigned int i;
		pte_t *ptepage = __va(pmd_pfn(*spmd) << PAGE_SHIFT);
		/* For each entry in the page, we might need to release it. */
		for (i = 0; i < PTRS_PER_PTE; i++)
			release_pte(ptepage[i]);
		/* Now we can free the page of PTEs */
		free_page((long)ptepage);
		/* And zero out the PMD entry so we never release it twice. */
583
		set_pmd(spmd, __pmd(0));
M
Matias Zabaljauregui 已提交
584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604
	}
}

static void release_pgd(pgd_t *spgd)
{
	/* If the entry's not present, there's nothing to release. */
	if (pgd_flags(*spgd) & _PAGE_PRESENT) {
		unsigned int i;
		pmd_t *pmdpage = __va(pgd_pfn(*spgd) << PAGE_SHIFT);

		for (i = 0; i < PTRS_PER_PMD; i++)
			release_pmd(&pmdpage[i]);

		/* Now we can free the page of PMDs */
		free_page((long)pmdpage);
		/* And zero out the PGD entry so we never release it twice. */
		set_pgd(spgd, __pgd(0));
	}
}

#else /* !CONFIG_X86_PAE */
R
Rusty Russell 已提交
605 606 607 608 609
/*H:450
 * If we chase down the release_pgd() code, the non-PAE version looks like
 * this.  The PAE version is almost identical, but instead of calling
 * release_pte it calls release_pmd(), which looks much like this.
 */
610
static void release_pgd(pgd_t *spgd)
R
Rusty Russell 已提交
611
{
R
Rusty Russell 已提交
612
	/* If the entry's not present, there's nothing to release. */
613
	if (pgd_flags(*spgd) & _PAGE_PRESENT) {
R
Rusty Russell 已提交
614
		unsigned int i;
R
Rusty Russell 已提交
615 616
		/*
		 * Converting the pfn to find the actual PTE page is easy: turn
R
Rusty Russell 已提交
617
		 * the page number into a physical address, then convert to a
R
Rusty Russell 已提交
618 619
		 * virtual address (easy for kernel pages like this one).
		 */
620
		pte_t *ptepage = __va(pgd_pfn(*spgd) << PAGE_SHIFT);
R
Rusty Russell 已提交
621
		/* For each entry in the page, we might need to release it. */
622
		for (i = 0; i < PTRS_PER_PTE; i++)
R
Rusty Russell 已提交
623
			release_pte(ptepage[i]);
R
Rusty Russell 已提交
624
		/* Now we can free the page of PTEs */
R
Rusty Russell 已提交
625
		free_page((long)ptepage);
R
Rusty Russell 已提交
626
		/* And zero out the PGD entry so we never release it twice. */
627
		*spgd = __pgd(0);
R
Rusty Russell 已提交
628 629
	}
}
M
Matias Zabaljauregui 已提交
630
#endif
R
Rusty Russell 已提交
631 632 633

/*H:445
 * We saw flush_user_mappings() twice: once from the flush_user_mappings()
R
Rusty Russell 已提交
634
 * hypercall and once in new_pgdir() when we re-used a top-level pgdir page.
R
Rusty Russell 已提交
635 636
 * It simply releases every PTE page from 0 up to the Guest's kernel address.
 */
R
Rusty Russell 已提交
637 638 639
static void flush_user_mappings(struct lguest *lg, int idx)
{
	unsigned int i;
R
Rusty Russell 已提交
640
	/* Release every pgd entry up to the kernel's address. */
641
	for (i = 0; i < pgd_index(lg->kernel_address); i++)
642
		release_pgd(lg->pgdirs[idx].pgdir + i);
R
Rusty Russell 已提交
643 644
}

R
Rusty Russell 已提交
645 646
/*H:440
 * (v) Flushing (throwing away) page tables,
R
Rusty Russell 已提交
647 648
 *
 * The Guest has a hypercall to throw away the page tables: it's used when a
R
Rusty Russell 已提交
649 650
 * large number of mappings have been changed.
 */
651
void guest_pagetable_flush_user(struct lg_cpu *cpu)
R
Rusty Russell 已提交
652
{
R
Rusty Russell 已提交
653
	/* Drop the userspace part of the current page table. */
654
	flush_user_mappings(cpu->lg, cpu->cpu_pgd);
R
Rusty Russell 已提交
655
}
R
Rusty Russell 已提交
656
/*:*/
R
Rusty Russell 已提交
657

658
/* We walk down the guest page tables to get a guest-physical address */
659
unsigned long guest_pa(struct lg_cpu *cpu, unsigned long vaddr)
660 661 662
{
	pgd_t gpgd;
	pte_t gpte;
M
Matias Zabaljauregui 已提交
663 664 665
#ifdef CONFIG_X86_PAE
	pmd_t gpmd;
#endif
666 667 668 669 670

	/* Still not set up?  Just map 1:1. */
	if (unlikely(cpu->linear_pages))
		return vaddr;

671
	/* First step: get the top-level Guest page table entry. */
672
	gpgd = lgread(cpu, gpgd_addr(cpu, vaddr), pgd_t);
673
	/* Toplevel not present?  We can't map it in. */
674
	if (!(pgd_flags(gpgd) & _PAGE_PRESENT)) {
675
		kill_guest(cpu, "Bad address %#lx", vaddr);
676 677
		return -1UL;
	}
678

M
Matias Zabaljauregui 已提交
679 680 681 682
#ifdef CONFIG_X86_PAE
	gpmd = lgread(cpu, gpmd_addr(gpgd, vaddr), pmd_t);
	if (!(pmd_flags(gpmd) & _PAGE_PRESENT))
		kill_guest(cpu, "Bad address %#lx", vaddr);
R
Rusty Russell 已提交
683 684
	gpte = lgread(cpu, gpte_addr(cpu, gpmd, vaddr), pte_t);
#else
M
Matias Zabaljauregui 已提交
685
	gpte = lgread(cpu, gpte_addr(cpu, gpgd, vaddr), pte_t);
R
Rusty Russell 已提交
686
#endif
687
	if (!(pte_flags(gpte) & _PAGE_PRESENT))
688
		kill_guest(cpu, "Bad address %#lx", vaddr);
689 690 691 692

	return pte_pfn(gpte) * PAGE_SIZE | (vaddr & ~PAGE_MASK);
}

R
Rusty Russell 已提交
693 694
/*
 * We keep several page tables.  This is a simple routine to find the page
R
Rusty Russell 已提交
695
 * table (if any) corresponding to this top-level address the Guest has given
R
Rusty Russell 已提交
696 697
 * us.
 */
R
Rusty Russell 已提交
698 699 700 701
static unsigned int find_pgdir(struct lguest *lg, unsigned long pgtable)
{
	unsigned int i;
	for (i = 0; i < ARRAY_SIZE(lg->pgdirs); i++)
702
		if (lg->pgdirs[i].pgdir && lg->pgdirs[i].gpgdir == pgtable)
R
Rusty Russell 已提交
703 704 705 706
			break;
	return i;
}

R
Rusty Russell 已提交
707 708
/*H:435
 * And this is us, creating the new page directory.  If we really do
R
Rusty Russell 已提交
709
 * allocate a new one (and so the kernel parts are not there), we set
R
Rusty Russell 已提交
710 711
 * blank_pgdir.
 */
712
static unsigned int new_pgdir(struct lg_cpu *cpu,
713
			      unsigned long gpgdir,
R
Rusty Russell 已提交
714 715 716
			      int *blank_pgdir)
{
	unsigned int next;
M
Matias Zabaljauregui 已提交
717 718 719
#ifdef CONFIG_X86_PAE
	pmd_t *pmd_table;
#endif
R
Rusty Russell 已提交
720

R
Rusty Russell 已提交
721 722 723 724
	/*
	 * We pick one entry at random to throw out.  Choosing the Least
	 * Recently Used might be better, but this is easy.
	 */
725
	next = random32() % ARRAY_SIZE(cpu->lg->pgdirs);
R
Rusty Russell 已提交
726
	/* If it's never been allocated at all before, try now. */
727 728 729
	if (!cpu->lg->pgdirs[next].pgdir) {
		cpu->lg->pgdirs[next].pgdir =
					(pgd_t *)get_zeroed_page(GFP_KERNEL);
R
Rusty Russell 已提交
730
		/* If the allocation fails, just keep using the one we have */
731
		if (!cpu->lg->pgdirs[next].pgdir)
732
			next = cpu->cpu_pgd;
M
Matias Zabaljauregui 已提交
733 734
		else {
#ifdef CONFIG_X86_PAE
R
Rusty Russell 已提交
735 736 737 738
			/*
			 * In PAE mode, allocate a pmd page and populate the
			 * last pgd entry.
			 */
M
Matias Zabaljauregui 已提交
739 740 741 742 743 744 745 746 747
			pmd_table = (pmd_t *)get_zeroed_page(GFP_KERNEL);
			if (!pmd_table) {
				free_page((long)cpu->lg->pgdirs[next].pgdir);
				set_pgd(cpu->lg->pgdirs[next].pgdir, __pgd(0));
				next = cpu->cpu_pgd;
			} else {
				set_pgd(cpu->lg->pgdirs[next].pgdir +
					SWITCHER_PGD_INDEX,
					__pgd(__pa(pmd_table) | _PAGE_PRESENT));
R
Rusty Russell 已提交
748 749 750 751
				/*
				 * This is a blank page, so there are no kernel
				 * mappings: caller must map the stack!
				 */
M
Matias Zabaljauregui 已提交
752 753 754
				*blank_pgdir = 1;
			}
#else
R
Rusty Russell 已提交
755
			*blank_pgdir = 1;
M
Matias Zabaljauregui 已提交
756 757
#endif
		}
R
Rusty Russell 已提交
758
	}
R
Rusty Russell 已提交
759
	/* Record which Guest toplevel this shadows. */
760
	cpu->lg->pgdirs[next].gpgdir = gpgdir;
R
Rusty Russell 已提交
761
	/* Release all the non-kernel mappings. */
762
	flush_user_mappings(cpu->lg, next);
R
Rusty Russell 已提交
763 764 765 766

	return next;
}

R
Rusty Russell 已提交
767 768
/*H:470
 * Finally, a routine which throws away everything: all PGD entries in all
R
Rusty Russell 已提交
769
 * the shadow page tables, including the Guest's kernel mappings.  This is used
R
Rusty Russell 已提交
770 771
 * when we destroy the Guest.
 */
R
Rusty Russell 已提交
772 773 774 775
static void release_all_pagetables(struct lguest *lg)
{
	unsigned int i, j;

R
Rusty Russell 已提交
776
	/* Every shadow pagetable this Guest has */
R
Rusty Russell 已提交
777
	for (i = 0; i < ARRAY_SIZE(lg->pgdirs); i++)
M
Matias Zabaljauregui 已提交
778 779 780 781 782 783 784 785 786 787
		if (lg->pgdirs[i].pgdir) {
#ifdef CONFIG_X86_PAE
			pgd_t *spgd;
			pmd_t *pmdpage;
			unsigned int k;

			/* Get the last pmd page. */
			spgd = lg->pgdirs[i].pgdir + SWITCHER_PGD_INDEX;
			pmdpage = __va(pgd_pfn(*spgd) << PAGE_SHIFT);

R
Rusty Russell 已提交
788 789 790 791
			/*
			 * And release the pmd entries of that pmd page,
			 * except for the switcher pmd.
			 */
M
Matias Zabaljauregui 已提交
792 793 794
			for (k = 0; k < SWITCHER_PMD_INDEX; k++)
				release_pmd(&pmdpage[k]);
#endif
R
Rusty Russell 已提交
795
			/* Every PGD entry except the Switcher at the top */
R
Rusty Russell 已提交
796
			for (j = 0; j < SWITCHER_PGD_INDEX; j++)
797
				release_pgd(lg->pgdirs[i].pgdir + j);
M
Matias Zabaljauregui 已提交
798
		}
R
Rusty Russell 已提交
799 800
}

R
Rusty Russell 已提交
801 802
/*
 * We also throw away everything when a Guest tells us it's changed a kernel
R
Rusty Russell 已提交
803
 * mapping.  Since kernel mappings are in every page table, it's easiest to
R
Rusty Russell 已提交
804
 * throw them all away.  This traps the Guest in amber for a while as
R
Rusty Russell 已提交
805 806
 * everything faults back in, but it's rare.
 */
807
void guest_pagetable_clear_all(struct lg_cpu *cpu)
R
Rusty Russell 已提交
808
{
809
	release_all_pagetables(cpu->lg);
R
Rusty Russell 已提交
810
	/* We need the Guest kernel stack mapped again. */
811
	pin_stack_pages(cpu);
R
Rusty Russell 已提交
812
}
813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850

/*H:430
 * (iv) Switching page tables
 *
 * Now we've seen all the page table setting and manipulation, let's see
 * what happens when the Guest changes page tables (ie. changes the top-level
 * pgdir).  This occurs on almost every context switch.
 */
void guest_new_pagetable(struct lg_cpu *cpu, unsigned long pgtable)
{
	int newpgdir, repin = 0;

	/*
	 * The very first time they call this, we're actually running without
	 * any page tables; we've been making it up.  Throw them away now.
	 */
	if (unlikely(cpu->linear_pages)) {
		release_all_pagetables(cpu->lg);
		cpu->linear_pages = false;
		/* Force allocation of a new pgdir. */
		newpgdir = ARRAY_SIZE(cpu->lg->pgdirs);
	} else {
		/* Look to see if we have this one already. */
		newpgdir = find_pgdir(cpu->lg, pgtable);
	}

	/*
	 * If not, we allocate or mug an existing one: if it's a fresh one,
	 * repin gets set to 1.
	 */
	if (newpgdir == ARRAY_SIZE(cpu->lg->pgdirs))
		newpgdir = new_pgdir(cpu, pgtable, &repin);
	/* Change the current pgd index to the new one. */
	cpu->cpu_pgd = newpgdir;
	/* If it was completely blank, we map in the Guest kernel stack */
	if (repin)
		pin_stack_pages(cpu);
}
R
Rusty Russell 已提交
851
/*:*/
R
Rusty Russell 已提交
852 853 854

/*M:009
 * Since we throw away all mappings when a kernel mapping changes, our
R
Rusty Russell 已提交
855 856 857 858 859
 * performance sucks for guests using highmem.  In fact, a guest with
 * PAGE_OFFSET 0xc0000000 (the default) and more than about 700MB of RAM is
 * usually slower than a Guest with less memory.
 *
 * This, of course, cannot be fixed.  It would take some kind of... well, I
R
Rusty Russell 已提交
860 861
 * don't know, but the term "puissant code-fu" comes to mind.
:*/
R
Rusty Russell 已提交
862

R
Rusty Russell 已提交
863 864
/*H:420
 * This is the routine which actually sets the page table entry for then
R
Rusty Russell 已提交
865 866 867 868 869 870 871 872 873 874 875 876 877
 * "idx"'th shadow page table.
 *
 * Normally, we can just throw out the old entry and replace it with 0: if they
 * use it demand_page() will put the new entry in.  We need to do this anyway:
 * The Guest expects _PAGE_ACCESSED to be set on its PTE the first time a page
 * is read from, and _PAGE_DIRTY when it's written to.
 *
 * But Avi Kivity pointed out that most Operating Systems (Linux included) set
 * these bits on PTEs immediately anyway.  This is done to save the CPU from
 * having to update them, but it helps us the same way: if they set
 * _PAGE_ACCESSED then we can put a read-only PTE entry in immediately, and if
 * they set _PAGE_DIRTY then we can put a writable PTE entry in immediately.
 */
878
static void do_set_pte(struct lg_cpu *cpu, int idx,
879
		       unsigned long vaddr, pte_t gpte)
R
Rusty Russell 已提交
880
{
R
Rusty Russell 已提交
881
	/* Look up the matching shadow page directory entry. */
882
	pgd_t *spgd = spgd_addr(cpu, idx, vaddr);
M
Matias Zabaljauregui 已提交
883 884 885
#ifdef CONFIG_X86_PAE
	pmd_t *spmd;
#endif
R
Rusty Russell 已提交
886 887

	/* If the top level isn't present, there's no entry to update. */
888
	if (pgd_flags(*spgd) & _PAGE_PRESENT) {
M
Matias Zabaljauregui 已提交
889 890 891 892
#ifdef CONFIG_X86_PAE
		spmd = spmd_addr(cpu, *spgd, vaddr);
		if (pmd_flags(*spmd) & _PAGE_PRESENT) {
#endif
R
Rusty Russell 已提交
893
			/* Otherwise, start by releasing the existing entry. */
M
Matias Zabaljauregui 已提交
894 895 896
			pte_t *spte = spte_addr(cpu, *spgd, vaddr);
			release_pte(*spte);

R
Rusty Russell 已提交
897 898 899 900 901 902
			/*
			 * If they're setting this entry as dirty or accessed,
			 * we might as well put that entry they've given us in
			 * now.  This shaves 10% off a copy-on-write
			 * micro-benchmark.
			 */
M
Matias Zabaljauregui 已提交
903
			if (pte_flags(gpte) & (_PAGE_DIRTY | _PAGE_ACCESSED)) {
904 905
				if (!check_gpte(cpu, gpte))
					return;
906 907
				set_pte(spte,
					gpte_to_spte(cpu, gpte,
M
Matias Zabaljauregui 已提交
908
						pte_flags(gpte) & _PAGE_DIRTY));
R
Rusty Russell 已提交
909 910 911 912 913
			} else {
				/*
				 * Otherwise kill it and we can demand_page()
				 * it in later.
				 */
914
				set_pte(spte, __pte(0));
R
Rusty Russell 已提交
915
			}
M
Matias Zabaljauregui 已提交
916 917 918
#ifdef CONFIG_X86_PAE
		}
#endif
R
Rusty Russell 已提交
919 920 921
	}
}

R
Rusty Russell 已提交
922 923
/*H:410
 * Updating a PTE entry is a little trickier.
R
Rusty Russell 已提交
924 925 926 927 928 929 930
 *
 * We keep track of several different page tables (the Guest uses one for each
 * process, so it makes sense to cache at least a few).  Each of these have
 * identical kernel parts: ie. every mapping above PAGE_OFFSET is the same for
 * all processes.  So when the page table above that address changes, we update
 * all the page tables, not just the current one.  This is rare.
 *
931
 * The benefit is that when we have to track a new page table, we can keep all
R
Rusty Russell 已提交
932 933
 * the kernel mappings.  This speeds up context switch immensely.
 */
934
void guest_set_pte(struct lg_cpu *cpu,
935
		   unsigned long gpgdir, unsigned long vaddr, pte_t gpte)
R
Rusty Russell 已提交
936
{
937 938 939 940 941 942
	/* We don't let you remap the Switcher; we need it to get back! */
	if (vaddr >= switcher_addr) {
		kill_guest(cpu, "attempt to set pte into Switcher pages");
		return;
	}

R
Rusty Russell 已提交
943 944 945 946
	/*
	 * Kernel mappings must be changed on all top levels.  Slow, but doesn't
	 * happen often.
	 */
947
	if (vaddr >= cpu->lg->kernel_address) {
R
Rusty Russell 已提交
948
		unsigned int i;
949 950 951
		for (i = 0; i < ARRAY_SIZE(cpu->lg->pgdirs); i++)
			if (cpu->lg->pgdirs[i].pgdir)
				do_set_pte(cpu, i, vaddr, gpte);
R
Rusty Russell 已提交
952
	} else {
R
Rusty Russell 已提交
953
		/* Is this page table one we have a shadow for? */
954 955
		int pgdir = find_pgdir(cpu->lg, gpgdir);
		if (pgdir != ARRAY_SIZE(cpu->lg->pgdirs))
R
Rusty Russell 已提交
956
			/* If so, do the update. */
957
			do_set_pte(cpu, pgdir, vaddr, gpte);
R
Rusty Russell 已提交
958 959 960
	}
}

R
Rusty Russell 已提交
961
/*H:400
R
Rusty Russell 已提交
962
 * (iii) Setting up a page table entry when the Guest tells us one has changed.
R
Rusty Russell 已提交
963 964 965 966 967 968 969 970 971 972
 *
 * Just like we did in interrupts_and_traps.c, it makes sense for us to deal
 * with the other side of page tables while we're here: what happens when the
 * Guest asks for a page table to be updated?
 *
 * We already saw that demand_page() will fill in the shadow page tables when
 * needed, so we can simply remove shadow page table entries whenever the Guest
 * tells us they've changed.  When the Guest tries to use the new entry it will
 * fault and demand_page() will fix it up.
 *
973
 * So with that in mind here's our code to update a (top-level) PGD entry:
R
Rusty Russell 已提交
974
 */
975
void guest_set_pgd(struct lguest *lg, unsigned long gpgdir, u32 idx)
R
Rusty Russell 已提交
976 977 978 979 980 981
{
	int pgdir;

	if (idx >= SWITCHER_PGD_INDEX)
		return;

R
Rusty Russell 已提交
982
	/* If they're talking about a page table we have a shadow for... */
983
	pgdir = find_pgdir(lg, gpgdir);
R
Rusty Russell 已提交
984
	if (pgdir < ARRAY_SIZE(lg->pgdirs))
R
Rusty Russell 已提交
985
		/* ... throw it away. */
986
		release_pgd(lg->pgdirs[pgdir].pgdir + idx);
R
Rusty Russell 已提交
987
}
R
Rusty Russell 已提交
988

M
Matias Zabaljauregui 已提交
989
#ifdef CONFIG_X86_PAE
R
Rusty Russell 已提交
990
/* For setting a mid-level, we just throw everything away.  It's easy. */
M
Matias Zabaljauregui 已提交
991 992 993 994 995
void guest_set_pmd(struct lguest *lg, unsigned long pmdp, u32 idx)
{
	guest_pagetable_clear_all(&lg->cpus[0]);
}
#endif
R
Rusty Russell 已提交
996

R
Rusty Russell 已提交
997 998
/*H:500
 * (vii) Setting up the page tables initially.
R
Rusty Russell 已提交
999
 *
1000 1001 1002 1003
 * When a Guest is first created, set initialize a shadow page table which
 * we will populate on future faults.  The Guest doesn't have any actual
 * pagetables yet, so we set linear_pages to tell demand_page() to fake it
 * for the moment.
R
Rusty Russell 已提交
1004
 */
1005
int init_guest_pagetable(struct lguest *lg)
R
Rusty Russell 已提交
1006
{
1007 1008
	struct lg_cpu *cpu = &lg->cpus[0];
	int allocated = 0;
1009

1010 1011 1012
	/* lg (and lg->cpus[]) starts zeroed: this allocates a new pgdir */
	cpu->cpu_pgd = new_pgdir(cpu, 0, &allocated);
	if (!allocated)
R
Rusty Russell 已提交
1013
		return -ENOMEM;
R
Rusty Russell 已提交
1014

1015 1016
	/* We start with a linear mapping until the initialize. */
	cpu->linear_pages = true;
R
Rusty Russell 已提交
1017 1018 1019
	return 0;
}

R
Rusty Russell 已提交
1020
/*H:508 When the Guest calls LHCALL_LGUEST_INIT we do more setup. */
1021
void page_table_guest_data_init(struct lg_cpu *cpu)
1022
{
1023 1024 1025 1026 1027 1028 1029
	/*
	 * We tell the Guest that it can't use the virtual addresses
	 * used by the Switcher.  This trick is equivalent to 4GB -
	 * switcher_addr.
	 */
	u32 top = ~switcher_addr + 1;

1030
	/* We get the kernel address: above this is all kernel memory. */
1031
	if (get_user(cpu->lg->kernel_address,
1032
		     &cpu->lg->lguest_data->kernel_address)
R
Rusty Russell 已提交
1033
		/*
1034 1035
		 * We tell the Guest that it can't use the top virtual
		 * addresses (used by the Switcher).
R
Rusty Russell 已提交
1036
		 */
1037
	    || put_user(top, &cpu->lg->lguest_data->reserve_mem)) {
1038
		kill_guest(cpu, "bad guest page %p", cpu->lg->lguest_data);
1039 1040
		return;
	}
1041

R
Rusty Russell 已提交
1042 1043
	/*
	 * In flush_user_mappings() we loop from 0 to
1044
	 * "pgd_index(lg->kernel_address)".  This assumes it won't hit the
R
Rusty Russell 已提交
1045 1046
	 * Switcher mappings, so check that now.
	 */
1047
	if (cpu->lg->kernel_address >= switcher_addr)
1048 1049
		kill_guest(cpu, "bad kernel address %#lx",
				 cpu->lg->kernel_address);
1050 1051
}

R
Rusty Russell 已提交
1052
/* When a Guest dies, our cleanup is fairly simple. */
R
Rusty Russell 已提交
1053 1054 1055 1056
void free_guest_pagetable(struct lguest *lg)
{
	unsigned int i;

R
Rusty Russell 已提交
1057
	/* Throw away all page table pages. */
R
Rusty Russell 已提交
1058
	release_all_pagetables(lg);
R
Rusty Russell 已提交
1059
	/* Now free the top levels: free_page() can handle 0 just fine. */
R
Rusty Russell 已提交
1060 1061 1062 1063
	for (i = 0; i < ARRAY_SIZE(lg->pgdirs); i++)
		free_page((long)lg->pgdirs[i].pgdir);
}

R
Rusty Russell 已提交
1064 1065
/*H:480
 * (vi) Mapping the Switcher when the Guest is about to run.
R
Rusty Russell 已提交
1066
 *
R
Rusty Russell 已提交
1067
 * The Switcher and the two pages for this CPU need to be visible in the
R
Rusty Russell 已提交
1068
 * Guest (and not the pages for other CPUs).  We have the appropriate PTE pages
R
Rusty Russell 已提交
1069
 * for each CPU already set up, we just need to hook them in now we know which
R
Rusty Russell 已提交
1070 1071
 * Guest is about to run on this CPU.
 */
1072
void map_switcher_in_guest(struct lg_cpu *cpu, struct lguest_pages *pages)
R
Rusty Russell 已提交
1073
{
C
Christoph Lameter 已提交
1074
	pte_t *switcher_pte_page = __this_cpu_read(switcher_pte_pages);
1075
	pte_t regs_pte;
R
Rusty Russell 已提交
1076

M
Matias Zabaljauregui 已提交
1077 1078 1079 1080
#ifdef CONFIG_X86_PAE
	pmd_t switcher_pmd;
	pmd_t *pmd_table;

1081 1082
	switcher_pmd = pfn_pmd(__pa(switcher_pte_page) >> PAGE_SHIFT,
			       PAGE_KERNEL_EXEC);
M
Matias Zabaljauregui 已提交
1083

R
Rusty Russell 已提交
1084 1085
	/* Figure out where the pmd page is, by reading the PGD, and converting
	 * it to a virtual address. */
M
Matias Zabaljauregui 已提交
1086 1087 1088
	pmd_table = __va(pgd_pfn(cpu->lg->
			pgdirs[cpu->cpu_pgd].pgdir[SWITCHER_PGD_INDEX])
								<< PAGE_SHIFT);
R
Rusty Russell 已提交
1089
	/* Now write it into the shadow page table. */
1090
	set_pmd(&pmd_table[SWITCHER_PMD_INDEX], switcher_pmd);
M
Matias Zabaljauregui 已提交
1091 1092 1093
#else
	pgd_t switcher_pgd;

R
Rusty Russell 已提交
1094 1095 1096 1097
	/*
	 * Make the last PGD entry for this Guest point to the Switcher's PTE
	 * page for this CPU (with appropriate flags).
	 */
1098
	switcher_pgd = __pgd(__pa(switcher_pte_page) | __PAGE_KERNEL_EXEC);
1099

1100
	cpu->lg->pgdirs[cpu->cpu_pgd].pgdir[SWITCHER_PGD_INDEX] = switcher_pgd;
R
Rusty Russell 已提交
1101

M
Matias Zabaljauregui 已提交
1102
#endif
R
Rusty Russell 已提交
1103 1104
	/*
	 * We also change the Switcher PTE page.  When we're running the Guest,
R
Rusty Russell 已提交
1105 1106 1107 1108 1109
	 * we want the Guest's "regs" page to appear where the first Switcher
	 * page for this CPU is.  This is an optimization: when the Switcher
	 * saves the Guest registers, it saves them into the first page of this
	 * CPU's "struct lguest_pages": if we make sure the Guest's register
	 * page is already mapped there, we don't have to copy them out
R
Rusty Russell 已提交
1110 1111
	 * again.
	 */
1112 1113
	regs_pte = pfn_pte(__pa(cpu->regs_page) >> PAGE_SHIFT, PAGE_KERNEL);
	set_pte(&switcher_pte_page[pte_index((unsigned long)pages)], regs_pte);
R
Rusty Russell 已提交
1114
}
R
Rusty Russell 已提交
1115
/*:*/
R
Rusty Russell 已提交
1116 1117 1118 1119 1120 1121 1122 1123 1124

static void free_switcher_pte_pages(void)
{
	unsigned int i;

	for_each_possible_cpu(i)
		free_page((long)switcher_pte_page(i));
}

R
Rusty Russell 已提交
1125 1126
/*H:520
 * Setting up the Switcher PTE page for given CPU is fairly easy, given
1127
 * the CPU number and the "struct page"s for the Switcher and per-cpu pages.
R
Rusty Russell 已提交
1128
 */
R
Rusty Russell 已提交
1129
static __init void populate_switcher_pte_page(unsigned int cpu,
1130
					      struct page *switcher_pages[])
R
Rusty Russell 已提交
1131
{
1132
	pte_t *pte = switcher_pte_page(cpu);
1133
	int i;
R
Rusty Russell 已提交
1134

1135 1136
	/* The first entries maps the Switcher code. */
	set_pte(&pte[0], mk_pte(switcher_pages[0],
1137
				__pgprot(_PAGE_PRESENT|_PAGE_ACCESSED)));
R
Rusty Russell 已提交
1138

R
Rusty Russell 已提交
1139
	/* The only other thing we map is this CPU's pair of pages. */
1140
	i = 1 + cpu*2;
R
Rusty Russell 已提交
1141

R
Rusty Russell 已提交
1142
	/* First page (Guest registers) is writable from the Guest */
1143
	set_pte(&pte[i], pfn_pte(page_to_pfn(switcher_pages[i]),
1144
			 __pgprot(_PAGE_PRESENT|_PAGE_ACCESSED|_PAGE_RW)));
1145

R
Rusty Russell 已提交
1146 1147 1148 1149
	/*
	 * The second page contains the "struct lguest_ro_state", and is
	 * read-only.
	 */
1150
	set_pte(&pte[i+1], pfn_pte(page_to_pfn(switcher_pages[i+1]),
1151
			   __pgprot(_PAGE_PRESENT|_PAGE_ACCESSED)));
R
Rusty Russell 已提交
1152 1153
}

R
Rusty Russell 已提交
1154 1155
/*
 * We've made it through the page table code.  Perhaps our tired brains are
R
Rusty Russell 已提交
1156 1157
 * still processing the details, or perhaps we're simply glad it's over.
 *
1158 1159 1160 1161 1162
 * If nothing else, note that all this complexity in juggling shadow page tables
 * in sync with the Guest's page tables is for one reason: for most Guests this
 * page table dance determines how bad performance will be.  This is why Xen
 * uses exotic direct Guest pagetable manipulation, and why both Intel and AMD
 * have implemented shadow page table support directly into hardware.
R
Rusty Russell 已提交
1163
 *
R
Rusty Russell 已提交
1164 1165
 * There is just one file remaining in the Host.
 */
R
Rusty Russell 已提交
1166

R
Rusty Russell 已提交
1167 1168 1169 1170
/*H:510
 * At boot or module load time, init_pagetables() allocates and populates
 * the Switcher PTE page for each CPU.
 */
1171
__init int init_pagetables(struct page **switcher_pages)
R
Rusty Russell 已提交
1172 1173 1174 1175
{
	unsigned int i;

	for_each_possible_cpu(i) {
1176
		switcher_pte_page(i) = (pte_t *)get_zeroed_page(GFP_KERNEL);
R
Rusty Russell 已提交
1177 1178 1179 1180
		if (!switcher_pte_page(i)) {
			free_switcher_pte_pages();
			return -ENOMEM;
		}
1181
		populate_switcher_pte_page(i, switcher_pages);
R
Rusty Russell 已提交
1182 1183 1184
	}
	return 0;
}
R
Rusty Russell 已提交
1185
/*:*/
R
Rusty Russell 已提交
1186

R
Rusty Russell 已提交
1187
/* Cleaning up simply involves freeing the PTE page for each CPU. */
R
Rusty Russell 已提交
1188 1189 1190 1191
void free_pagetables(void)
{
	free_switcher_pte_pages();
}