init_64.c 28.5 KB
Newer Older
L
Linus Torvalds 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
/*
 *  linux/arch/x86_64/mm/init.c
 *
 *  Copyright (C) 1995  Linus Torvalds
 *  Copyright (C) 2000  Pavel Machek <pavel@suse.cz>
 *  Copyright (C) 2002,2003 Andi Kleen <ak@suse.de>
 */

#include <linux/signal.h>
#include <linux/sched.h>
#include <linux/kernel.h>
#include <linux/errno.h>
#include <linux/string.h>
#include <linux/types.h>
#include <linux/ptrace.h>
#include <linux/mman.h>
#include <linux/mm.h>
#include <linux/swap.h>
#include <linux/smp.h>
#include <linux/init.h>
T
Thomas Gleixner 已提交
21
#include <linux/initrd.h>
L
Linus Torvalds 已提交
22 23 24
#include <linux/pagemap.h>
#include <linux/bootmem.h>
#include <linux/proc_fs.h>
25
#include <linux/pci.h>
26
#include <linux/pfn.h>
27
#include <linux/poison.h>
28
#include <linux/dma-mapping.h>
29 30
#include <linux/module.h>
#include <linux/memory_hotplug.h>
31
#include <linux/nmi.h>
L
Linus Torvalds 已提交
32 33 34 35 36 37 38 39 40 41 42 43 44 45

#include <asm/processor.h>
#include <asm/system.h>
#include <asm/uaccess.h>
#include <asm/pgtable.h>
#include <asm/pgalloc.h>
#include <asm/dma.h>
#include <asm/fixmap.h>
#include <asm/e820.h>
#include <asm/apic.h>
#include <asm/tlb.h>
#include <asm/mmu_context.h>
#include <asm/proto.h>
#include <asm/smp.h>
46
#include <asm/sections.h>
47
#include <asm/kdebug.h>
48
#include <asm/numa.h>
49
#include <asm/cacheflush.h>
L
Linus Torvalds 已提交
50

51 52 53 54 55
/*
 * end_pfn only includes RAM, while max_pfn_mapped includes all e820 entries.
 * The direct mapping extends to max_pfn_mapped, so that we can directly access
 * apertures, ACPI and other tables without having to play with fixmaps.
 */
56
unsigned long max_low_pfn_mapped;
57 58
unsigned long max_pfn_mapped;

59 60
static unsigned long dma_reserve __initdata;

L
Linus Torvalds 已提交
61 62
DEFINE_PER_CPU(struct mmu_gather, mmu_gathers);

H
Hugh Dickins 已提交
63
int direct_gbpages
I
Ingo Molnar 已提交
64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82
#ifdef CONFIG_DIRECT_GBPAGES
				= 1
#endif
;

static int __init parse_direct_gbpages_off(char *arg)
{
	direct_gbpages = 0;
	return 0;
}
early_param("nogbpages", parse_direct_gbpages_off);

static int __init parse_direct_gbpages_on(char *arg)
{
	direct_gbpages = 1;
	return 0;
}
early_param("gbpages", parse_direct_gbpages_on);

L
Linus Torvalds 已提交
83 84 85 86 87 88 89 90
/*
 * NOTE: pagetable_init alloc all the fixmap pagetables contiguous on the
 * physical space so we can cache the place of the first one and move
 * around without checking the pgd every time.
 */

int after_bootmem;

91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144
unsigned long __supported_pte_mask __read_mostly = ~0UL;
EXPORT_SYMBOL_GPL(__supported_pte_mask);

static int do_not_nx __cpuinitdata;

/* noexec=on|off
Control non executable mappings for 64bit processes.

on	Enable(default)
off	Disable
*/
static int __init nonx_setup(char *str)
{
	if (!str)
		return -EINVAL;
	if (!strncmp(str, "on", 2)) {
		__supported_pte_mask |= _PAGE_NX;
		do_not_nx = 0;
	} else if (!strncmp(str, "off", 3)) {
		do_not_nx = 1;
		__supported_pte_mask &= ~_PAGE_NX;
	}
	return 0;
}
early_param("noexec", nonx_setup);

void __cpuinit check_efer(void)
{
	unsigned long efer;

	rdmsrl(MSR_EFER, efer);
	if (!(efer & EFER_NX) || do_not_nx)
		__supported_pte_mask &= ~_PAGE_NX;
}

int force_personality32;

/* noexec32=on|off
Control non executable heap for 32bit processes.
To control the stack too use noexec=off

on	PROT_READ does not imply PROT_EXEC for 32bit processes (default)
off	PROT_READ implies PROT_EXEC
*/
static int __init nonx32_setup(char *str)
{
	if (!strcmp(str, "on"))
		force_personality32 &= ~READ_IMPLIES_EXEC;
	else if (!strcmp(str, "off"))
		force_personality32 |= READ_IMPLIES_EXEC;
	return 1;
}
__setup("noexec32=", nonx32_setup);

145 146 147 148 149
/*
 * NOTE: This function is marked __ref because it calls __init function
 * (alloc_bootmem_pages). It's safe to do it ONLY when after_bootmem == 0.
 */
static __ref void *spp_getpage(void)
T
Thomas Gleixner 已提交
150
{
L
Linus Torvalds 已提交
151
	void *ptr;
T
Thomas Gleixner 已提交
152

L
Linus Torvalds 已提交
153
	if (after_bootmem)
T
Thomas Gleixner 已提交
154
		ptr = (void *) get_zeroed_page(GFP_ATOMIC);
L
Linus Torvalds 已提交
155 156
	else
		ptr = alloc_bootmem_pages(PAGE_SIZE);
T
Thomas Gleixner 已提交
157 158 159 160 161

	if (!ptr || ((unsigned long)ptr & ~PAGE_MASK)) {
		panic("set_pte_phys: cannot allocate page data %s\n",
			after_bootmem ? "after bootmem" : "");
	}
L
Linus Torvalds 已提交
162

163
	pr_debug("spp_getpage %p\n", ptr);
T
Thomas Gleixner 已提交
164

L
Linus Torvalds 已提交
165
	return ptr;
T
Thomas Gleixner 已提交
166
}
L
Linus Torvalds 已提交
167

168
void
169
set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte)
L
Linus Torvalds 已提交
170 171 172
{
	pud_t *pud;
	pmd_t *pmd;
173
	pte_t *pte;
L
Linus Torvalds 已提交
174

175
	pud = pud_page + pud_index(vaddr);
L
Linus Torvalds 已提交
176
	if (pud_none(*pud)) {
T
Thomas Gleixner 已提交
177
		pmd = (pmd_t *) spp_getpage();
178
		pud_populate(&init_mm, pud, pmd);
L
Linus Torvalds 已提交
179
		if (pmd != pmd_offset(pud, 0)) {
180
			printk(KERN_ERR "PAGETABLE BUG #01! %p <-> %p\n",
T
Thomas Gleixner 已提交
181
				pmd, pmd_offset(pud, 0));
L
Linus Torvalds 已提交
182 183 184 185 186 187
			return;
		}
	}
	pmd = pmd_offset(pud, vaddr);
	if (pmd_none(*pmd)) {
		pte = (pte_t *) spp_getpage();
188
		pmd_populate_kernel(&init_mm, pmd, pte);
L
Linus Torvalds 已提交
189
		if (pte != pte_offset_kernel(pmd, 0)) {
190
			printk(KERN_ERR "PAGETABLE BUG #02!\n");
L
Linus Torvalds 已提交
191 192 193 194 195
			return;
		}
	}

	pte = pte_offset_kernel(pmd, vaddr);
I
Ingo Molnar 已提交
196
	if (!pte_none(*pte) && pte_val(new_pte) &&
L
Linus Torvalds 已提交
197 198 199 200 201 202 203 204 205 206 207
	    pte_val(*pte) != (pte_val(new_pte) & __supported_pte_mask))
		pte_ERROR(*pte);
	set_pte(pte, new_pte);

	/*
	 * It's enough to flush this one mapping.
	 * (PGE mappings get flushed as well)
	 */
	__flush_tlb_one(vaddr);
}

208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225
void
set_pte_vaddr(unsigned long vaddr, pte_t pteval)
{
	pgd_t *pgd;
	pud_t *pud_page;

	pr_debug("set_pte_vaddr %lx to %lx\n", vaddr, native_pte_val(pteval));

	pgd = pgd_offset_k(vaddr);
	if (pgd_none(*pgd)) {
		printk(KERN_ERR
			"PGD FIXMAP MISSING, it should be setup in head.S!\n");
		return;
	}
	pud_page = (pud_t*)pgd_page_vaddr(*pgd);
	set_pte_vaddr_pud(pud_page, vaddr, pteval);
}

226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265
/*
 * Create large page table mappings for a range of physical addresses.
 */
static void __init __init_extra_mapping(unsigned long phys, unsigned long size,
						pgprot_t prot)
{
	pgd_t *pgd;
	pud_t *pud;
	pmd_t *pmd;

	BUG_ON((phys & ~PMD_MASK) || (size & ~PMD_MASK));
	for (; size; phys += PMD_SIZE, size -= PMD_SIZE) {
		pgd = pgd_offset_k((unsigned long)__va(phys));
		if (pgd_none(*pgd)) {
			pud = (pud_t *) spp_getpage();
			set_pgd(pgd, __pgd(__pa(pud) | _KERNPG_TABLE |
						_PAGE_USER));
		}
		pud = pud_offset(pgd, (unsigned long)__va(phys));
		if (pud_none(*pud)) {
			pmd = (pmd_t *) spp_getpage();
			set_pud(pud, __pud(__pa(pmd) | _KERNPG_TABLE |
						_PAGE_USER));
		}
		pmd = pmd_offset(pud, phys);
		BUG_ON(!pmd_none(*pmd));
		set_pmd(pmd, __pmd(phys | pgprot_val(prot)));
	}
}

void __init init_extra_mapping_wb(unsigned long phys, unsigned long size)
{
	__init_extra_mapping(phys, size, PAGE_KERNEL_LARGE);
}

void __init init_extra_mapping_uc(unsigned long phys, unsigned long size)
{
	__init_extra_mapping(phys, size, PAGE_KERNEL_LARGE_NOCACHE);
}

266
/*
267 268 269
 * The head.S code sets up the kernel high mapping:
 *
 *   from __START_KERNEL_map to __START_KERNEL_map + size (== _end-_text)
270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286
 *
 * phys_addr holds the negative offset to the kernel, which is added
 * to the compile time generated pmds. This results in invalid pmds up
 * to the point where we hit the physaddr 0 mapping.
 *
 * We limit the mappings to the region from _text to _end.  _end is
 * rounded up to the 2MB boundary. This catches the invalid pmds as
 * well, as they are located before _text:
 */
void __init cleanup_highmap(void)
{
	unsigned long vaddr = __START_KERNEL_map;
	unsigned long end = round_up((unsigned long)_end, PMD_SIZE) - 1;
	pmd_t *pmd = level2_kernel_pgt;
	pmd_t *last_pmd = pmd + PTRS_PER_PMD;

	for (; pmd < last_pmd; pmd++, vaddr += PMD_SIZE) {
287
		if (pmd_none(*pmd))
288 289 290 291 292 293
			continue;
		if (vaddr < (unsigned long) _text || vaddr > end)
			set_pmd(pmd, __pmd(0));
	}
}

294 295
static unsigned long __initdata table_start;
static unsigned long __meminitdata table_end;
296
static unsigned long __meminitdata table_top;
L
Linus Torvalds 已提交
297

298
static __ref void *alloc_low_page(unsigned long *phys)
T
Thomas Gleixner 已提交
299
{
300
	unsigned long pfn = table_end++;
L
Linus Torvalds 已提交
301 302
	void *adr;

303 304 305
	if (after_bootmem) {
		adr = (void *)get_zeroed_page(GFP_ATOMIC);
		*phys = __pa(adr);
T
Thomas Gleixner 已提交
306

307 308 309
		return adr;
	}

310
	if (pfn >= table_top)
T
Thomas Gleixner 已提交
311
		panic("alloc_low_page: ran out of memory");
312 313

	adr = early_ioremap(pfn * PAGE_SIZE, PAGE_SIZE);
314
	memset(adr, 0, PAGE_SIZE);
315 316 317
	*phys  = pfn * PAGE_SIZE;
	return adr;
}
L
Linus Torvalds 已提交
318

319
static __ref void unmap_low_page(void *adr)
T
Thomas Gleixner 已提交
320
{
321 322 323
	if (after_bootmem)
		return;

324
	early_iounmap(adr, PAGE_SIZE);
T
Thomas Gleixner 已提交
325
}
L
Linus Torvalds 已提交
326

327
static unsigned long __meminit
328 329 330
phys_pte_init(pte_t *pte_page, unsigned long addr, unsigned long end)
{
	unsigned pages = 0;
331
	unsigned long last_map_addr = end;
332
	int i;
333

334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352
	pte_t *pte = pte_page + pte_index(addr);

	for(i = pte_index(addr); i < PTRS_PER_PTE; i++, addr += PAGE_SIZE, pte++) {

		if (addr >= end) {
			if (!after_bootmem) {
				for(; i < PTRS_PER_PTE; i++, pte++)
					set_pte(pte, __pte(0));
			}
			break;
		}

		if (pte_val(*pte))
			continue;

		if (0)
			printk("   pte=%p addr=%lx pte=%016lx\n",
			       pte, addr, pfn_pte(addr >> PAGE_SHIFT, PAGE_KERNEL).pte);
		set_pte(pte, pfn_pte(addr >> PAGE_SHIFT, PAGE_KERNEL));
353
		last_map_addr = (addr & PAGE_MASK) + PAGE_SIZE;
354 355 356
		pages++;
	}
	update_page_count(PG_LEVEL_4K, pages);
357 358

	return last_map_addr;
359 360
}

361
static unsigned long __meminit
362 363 364 365
phys_pte_update(pmd_t *pmd, unsigned long address, unsigned long end)
{
	pte_t *pte = (pte_t *)pmd_page_vaddr(*pmd);

366
	return phys_pte_init(pte, address, end);
367 368
}

369
static unsigned long __meminit
370 371
phys_pmd_init(pmd_t *pmd_page, unsigned long address, unsigned long end,
			 unsigned long page_size_mask)
372
{
373
	unsigned long pages = 0;
374
	unsigned long last_map_addr = end;
H
Hugh Dickins 已提交
375
	unsigned long start = address;
376

377
	int i = pmd_index(address);
378

379
	for (; i < PTRS_PER_PMD; i++, address += PMD_SIZE) {
380
		unsigned long pte_phys;
381
		pmd_t *pmd = pmd_page + pmd_index(address);
382
		pte_t *pte;
383

384
		if (address >= end) {
T
Thomas Gleixner 已提交
385
			if (!after_bootmem) {
386 387
				for (; i < PTRS_PER_PMD; i++, pmd++)
					set_pmd(pmd, __pmd(0));
T
Thomas Gleixner 已提交
388
			}
389 390
			break;
		}
391

392
		if (pmd_val(*pmd)) {
393 394
			if (!pmd_large(*pmd)) {
				spin_lock(&init_mm.page_table_lock);
395
				last_map_addr = phys_pte_update(pmd, address,
396 397 398
								end);
				spin_unlock(&init_mm.page_table_lock);
			}
H
Hugh Dickins 已提交
399 400 401
			/* Count entries we're using from level2_ident_pgt */
			if (start == 0)
				pages++;
402 403 404
			continue;
		}

405
		if (page_size_mask & (1<<PG_LEVEL_2M)) {
406
			pages++;
407
			spin_lock(&init_mm.page_table_lock);
408 409
			set_pte((pte_t *)pmd,
				pfn_pte(address >> PAGE_SHIFT, PAGE_KERNEL_LARGE));
410
			spin_unlock(&init_mm.page_table_lock);
411
			last_map_addr = (address & PMD_MASK) + PMD_SIZE;
412
			continue;
413
		}
414

415
		pte = alloc_low_page(&pte_phys);
416
		last_map_addr = phys_pte_init(pte, address, end);
417 418
		unmap_low_page(pte);

419
		spin_lock(&init_mm.page_table_lock);
420
		pmd_populate_kernel(&init_mm, pmd, __va(pte_phys));
421
		spin_unlock(&init_mm.page_table_lock);
422
	}
423
	update_page_count(PG_LEVEL_2M, pages);
424
	return last_map_addr;
425 426
}

427
static unsigned long __meminit
428 429
phys_pmd_update(pud_t *pud, unsigned long address, unsigned long end,
			 unsigned long page_size_mask)
430
{
T
Thomas Gleixner 已提交
431
	pmd_t *pmd = pmd_offset(pud, 0);
432 433
	unsigned long last_map_addr;

434
	last_map_addr = phys_pmd_init(pmd, address, end, page_size_mask);
435
	__flush_tlb_all();
436
	return last_map_addr;
437 438
}

439
static unsigned long __meminit
440 441
phys_pud_init(pud_t *pud_page, unsigned long addr, unsigned long end,
			 unsigned long page_size_mask)
T
Thomas Gleixner 已提交
442
{
443
	unsigned long pages = 0;
444
	unsigned long last_map_addr = end;
445
	int i = pud_index(addr);
446

T
Thomas Gleixner 已提交
447
	for (; i < PTRS_PER_PUD; i++, addr = (addr & PUD_MASK) + PUD_SIZE) {
448 449
		unsigned long pmd_phys;
		pud_t *pud = pud_page + pud_index(addr);
L
Linus Torvalds 已提交
450 451
		pmd_t *pmd;

452
		if (addr >= end)
L
Linus Torvalds 已提交
453 454
			break;

T
Thomas Gleixner 已提交
455 456 457
		if (!after_bootmem &&
				!e820_any_mapped(addr, addr+PUD_SIZE, 0)) {
			set_pud(pud, __pud(0));
L
Linus Torvalds 已提交
458
			continue;
T
Thomas Gleixner 已提交
459
		}
L
Linus Torvalds 已提交
460

461
		if (pud_val(*pud)) {
462
			if (!pud_large(*pud))
463 464
				last_map_addr = phys_pmd_update(pud, addr, end,
							 page_size_mask);
465 466 467
			continue;
		}

468
		if (page_size_mask & (1<<PG_LEVEL_1G)) {
469
			pages++;
470
			spin_lock(&init_mm.page_table_lock);
471 472
			set_pte((pte_t *)pud,
				pfn_pte(addr >> PAGE_SHIFT, PAGE_KERNEL_LARGE));
473
			spin_unlock(&init_mm.page_table_lock);
474
			last_map_addr = (addr & PUD_MASK) + PUD_SIZE;
475 476 477
			continue;
		}

478
		pmd = alloc_low_page(&pmd_phys);
479
		last_map_addr = phys_pmd_init(pmd, addr, end, page_size_mask);
480
		unmap_low_page(pmd);
481 482

		spin_lock(&init_mm.page_table_lock);
483
		pud_populate(&init_mm, pud, __va(pmd_phys));
484
		spin_unlock(&init_mm.page_table_lock);
L
Linus Torvalds 已提交
485
	}
A
Andi Kleen 已提交
486
	__flush_tlb_all();
487
	update_page_count(PG_LEVEL_1G, pages);
488

489
	return last_map_addr;
T
Thomas Gleixner 已提交
490
}
L
Linus Torvalds 已提交
491

492
static unsigned long __meminit
493 494
phys_pud_update(pgd_t *pgd, unsigned long addr, unsigned long end,
		 unsigned long page_size_mask)
495 496 497 498 499
{
	pud_t *pud;

	pud = (pud_t *)pgd_page_vaddr(*pgd);

500
	return phys_pud_init(pud, addr, end, page_size_mask);
501 502
}

L
Linus Torvalds 已提交
503 504
static void __init find_early_table_space(unsigned long end)
{
505
	unsigned long puds, pmds, ptes, tables, start;
L
Linus Torvalds 已提交
506 507

	puds = (end + PUD_SIZE - 1) >> PUD_SHIFT;
508
	tables = round_up(puds * sizeof(pud_t), PAGE_SIZE);
509 510 511 512 513 514 515 516 517 518 519 520 521 522 523
	if (direct_gbpages) {
		unsigned long extra;
		extra = end - ((end>>PUD_SHIFT) << PUD_SHIFT);
		pmds = (extra + PMD_SIZE - 1) >> PMD_SHIFT;
	} else
		pmds = (end + PMD_SIZE - 1) >> PMD_SHIFT;
	tables += round_up(pmds * sizeof(pmd_t), PAGE_SIZE);

	if (cpu_has_pse) {
		unsigned long extra;
		extra = end - ((end>>PMD_SHIFT) << PMD_SHIFT);
		ptes = (extra + PAGE_SIZE - 1) >> PAGE_SHIFT;
	} else
		ptes = (end + PAGE_SIZE - 1) >> PAGE_SHIFT;
	tables += round_up(ptes * sizeof(pte_t), PAGE_SIZE);
L
Linus Torvalds 已提交
524

T
Thomas Gleixner 已提交
525 526 527 528 529 530
	/*
	 * RED-PEN putting page tables only on node 0 could
	 * cause a hotspot and fill up ZONE_DMA. The page tables
	 * need roughly 0.5KB per GB.
	 */
	start = 0x8000;
531
	table_start = find_e820_area(start, end, tables, PAGE_SIZE);
L
Linus Torvalds 已提交
532 533 534 535 536
	if (table_start == -1UL)
		panic("Cannot find space for the kernel page tables");

	table_start >>= PAGE_SHIFT;
	table_end = table_start;
537
	table_top = table_start + (tables >> PAGE_SHIFT);
538

539 540
	printk(KERN_DEBUG "kernel direct mapping tables up to %lx @ %lx-%lx\n",
		end, table_start << PAGE_SHIFT, table_top << PAGE_SHIFT);
L
Linus Torvalds 已提交
541 542
}

543 544 545 546 547 548 549 550
static void __init init_gbpages(void)
{
	if (direct_gbpages && cpu_has_gbpages)
		printk(KERN_INFO "Using GB pages for direct mapping\n");
	else
		direct_gbpages = 0;
}

551 552 553
static unsigned long __init kernel_physical_mapping_init(unsigned long start,
						unsigned long end,
						unsigned long page_size_mask)
T
Thomas Gleixner 已提交
554
{
L
Linus Torvalds 已提交
555

556
	unsigned long next, last_map_addr = end;
L
Linus Torvalds 已提交
557 558 559 560 561

	start = (unsigned long)__va(start);
	end = (unsigned long)__va(end);

	for (; start < end; start = next) {
562
		pgd_t *pgd = pgd_offset_k(start);
T
Thomas Gleixner 已提交
563
		unsigned long pud_phys;
564 565
		pud_t *pud;

566
		next = (start + PGDIR_SIZE) & PGDIR_MASK;
567 568 569 570
		if (next > end)
			next = end;

		if (pgd_val(*pgd)) {
571 572
			last_map_addr = phys_pud_update(pgd, __pa(start),
						 __pa(end), page_size_mask);
573 574 575
			continue;
		}

576
		pud = alloc_low_page(&pud_phys);
577 578
		last_map_addr = phys_pud_init(pud, __pa(start), __pa(next),
						 page_size_mask);
579
		unmap_low_page(pud);
580 581 582 583

		spin_lock(&init_mm.page_table_lock);
		pgd_populate(&init_mm, pgd, __va(pud_phys));
		spin_unlock(&init_mm.page_table_lock);
T
Thomas Gleixner 已提交
584
	}
L
Linus Torvalds 已提交
585

586 587
	return last_map_addr;
}
588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613

struct map_range {
	unsigned long start;
	unsigned long end;
	unsigned page_size_mask;
};

#define NR_RANGE_MR 5

static int save_mr(struct map_range *mr, int nr_range,
		   unsigned long start_pfn, unsigned long end_pfn,
		   unsigned long page_size_mask)
{

	if (start_pfn < end_pfn) {
		if (nr_range >= NR_RANGE_MR)
			panic("run out of range for init_memory_mapping\n");
		mr[nr_range].start = start_pfn<<PAGE_SHIFT;
		mr[nr_range].end   = end_pfn<<PAGE_SHIFT;
		mr[nr_range].page_size_mask = page_size_mask;
		nr_range++;
	}

	return nr_range;
}

614 615 616 617 618 619 620 621
/*
 * Setup the direct mapping of the physical memory at PAGE_OFFSET.
 * This runs before bootmem is initialized and gets pages directly from
 * the physical memory. To access them they are temporarily mapped.
 */
unsigned long __init_refok init_memory_mapping(unsigned long start,
					       unsigned long end)
{
622
	unsigned long last_map_addr = 0;
623
	unsigned long page_size_mask = 0;
624
	unsigned long start_pfn, end_pfn;
625

626 627 628
	struct map_range mr[NR_RANGE_MR];
	int nr_range, i;

629 630 631 632 633 634 635 636 637
	printk(KERN_INFO "init_memory_mapping\n");

	/*
	 * Find space for the kernel direct mapping tables.
	 *
	 * Later we should allocate these tables in the local node of the
	 * memory mapped. Unfortunately this is done currently before the
	 * nodes are discovered.
	 */
638
	if (!after_bootmem)
639 640 641 642 643 644 645
		init_gbpages();

	if (direct_gbpages)
		page_size_mask |= 1 << PG_LEVEL_1G;
	if (cpu_has_pse)
		page_size_mask |= 1 << PG_LEVEL_2M;

646 647 648 649
	memset(mr, 0, sizeof(mr));
	nr_range = 0;

	/* head if not big page alignment ?*/
650 651 652
	start_pfn = start >> PAGE_SHIFT;
	end_pfn = ((start + (PMD_SIZE - 1)) >> PMD_SHIFT)
			<< (PMD_SHIFT - PAGE_SHIFT);
653
	nr_range = save_mr(mr, nr_range, start_pfn, end_pfn, 0);
654 655 656 657 658 659 660 661

	/* big page (2M) range*/
	start_pfn = ((start + (PMD_SIZE - 1))>>PMD_SHIFT)
			 << (PMD_SHIFT - PAGE_SHIFT);
	end_pfn = ((start + (PUD_SIZE - 1))>>PUD_SHIFT)
			 << (PUD_SHIFT - PAGE_SHIFT);
	if (end_pfn > ((end>>PUD_SHIFT)<<(PUD_SHIFT - PAGE_SHIFT)))
		end_pfn = ((end>>PUD_SHIFT)<<(PUD_SHIFT - PAGE_SHIFT));
662 663
	nr_range = save_mr(mr, nr_range, start_pfn, end_pfn,
			page_size_mask & (1<<PG_LEVEL_2M));
664 665 666 667

	/* big page (1G) range */
	start_pfn = end_pfn;
	end_pfn = (end>>PUD_SHIFT) << (PUD_SHIFT - PAGE_SHIFT);
668 669 670
	nr_range = save_mr(mr, nr_range, start_pfn, end_pfn,
				page_size_mask &
				 ((1<<PG_LEVEL_2M)|(1<<PG_LEVEL_1G)));
671 672 673 674

	/* tail is not big page (1G) alignment */
	start_pfn = end_pfn;
	end_pfn = (end>>PMD_SHIFT) << (PMD_SHIFT - PAGE_SHIFT);
675 676 677
	nr_range = save_mr(mr, nr_range, start_pfn, end_pfn,
			page_size_mask & (1<<PG_LEVEL_2M));

678 679 680
	/* tail is not big page (2M) alignment */
	start_pfn = end_pfn;
	end_pfn = end>>PAGE_SHIFT;
681 682
	nr_range = save_mr(mr, nr_range, start_pfn, end_pfn, 0);

Y
Yinghai Lu 已提交
683 684 685 686 687 688 689 690 691 692 693 694 695 696
	/* try to merge same page size and continuous */
	for (i = 0; nr_range > 1 && i < nr_range - 1; i++) {
		unsigned long old_start;
		if (mr[i].end != mr[i+1].start ||
		    mr[i].page_size_mask != mr[i+1].page_size_mask)
			continue;
		/* move it */
		old_start = mr[i].start;
		memmove(&mr[i], &mr[i+1],
			 (nr_range - 1 - i) * sizeof (struct map_range));
		mr[i].start = old_start;
		nr_range--;
	}

697 698 699 700 701 702 703 704 705 706
	for (i = 0; i < nr_range; i++)
		printk(KERN_DEBUG " %010lx - %010lx page %s\n",
				mr[i].start, mr[i].end,
			(mr[i].page_size_mask & (1<<PG_LEVEL_1G))?"1G":(
			 (mr[i].page_size_mask & (1<<PG_LEVEL_2M))?"2M":"4k"));

	if (!after_bootmem)
		find_early_table_space(end);

	for (i = 0; i < nr_range; i++)
707
		last_map_addr = kernel_physical_mapping_init(
708 709
					mr[i].start, mr[i].end,
					mr[i].page_size_mask);
710

711
	if (!after_bootmem)
712
		mmu_cr4_features = read_cr4();
L
Linus Torvalds 已提交
713
	__flush_tlb_all();
714

715
	if (!after_bootmem && table_end > table_start)
716 717
		reserve_early(table_start << PAGE_SHIFT,
				 table_end << PAGE_SHIFT, "PGTABLE");
Y
Yinghai Lu 已提交
718

719 720 721
	printk(KERN_INFO "last_map_addr: %lx end: %lx\n",
			 last_map_addr, end);

Y
Yinghai Lu 已提交
722
	if (!after_bootmem)
723
		early_memtest(start, end);
724

725
	return last_map_addr >> PAGE_SHIFT;
L
Linus Torvalds 已提交
726 727
}

728
#ifndef CONFIG_NUMA
729 730 731 732 733 734 735 736 737
void __init initmem_init(unsigned long start_pfn, unsigned long end_pfn)
{
	unsigned long bootmap_size, bootmap;

	bootmap_size = bootmem_bootmap_pages(end_pfn)<<PAGE_SHIFT;
	bootmap = find_e820_area(0, end_pfn<<PAGE_SHIFT, bootmap_size,
				 PAGE_SIZE);
	if (bootmap == -1L)
		panic("Cannot find bootmem map of size %ld\n", bootmap_size);
Y
Yinghai Lu 已提交
738 739 740
	/* don't touch min_low_pfn */
	bootmap_size = init_bootmem_node(NODE_DATA(0), bootmap >> PAGE_SHIFT,
					 0, end_pfn);
741 742 743 744 745 746
	e820_register_active_regions(0, start_pfn, end_pfn);
	free_bootmem_with_active_regions(0, end_pfn);
	early_res_to_bootmem(0, end_pfn<<PAGE_SHIFT);
	reserve_bootmem(bootmap, bootmap_size, BOOTMEM_DEFAULT);
}

L
Linus Torvalds 已提交
747 748
void __init paging_init(void)
{
749
	unsigned long max_zone_pfns[MAX_NR_ZONES];
T
Thomas Gleixner 已提交
750

751 752 753
	memset(max_zone_pfns, 0, sizeof(max_zone_pfns));
	max_zone_pfns[ZONE_DMA] = MAX_DMA_PFN;
	max_zone_pfns[ZONE_DMA32] = MAX_DMA32_PFN;
Y
Yinghai Lu 已提交
754
	max_zone_pfns[ZONE_NORMAL] = max_pfn;
755

Y
Yinghai Lu 已提交
756
	memory_present(0, 0, max_pfn);
757
	sparse_init();
758
	free_area_init_nodes(max_zone_pfns);
L
Linus Torvalds 已提交
759 760 761
}
#endif

762 763 764
/*
 * Memory hotplug specific functions
 */
765
#ifdef CONFIG_MEMORY_HOTPLUG
766 767 768 769
/*
 * Memory is added always to NORMAL zone. This means you will never get
 * additional DMA/DMA32 memory.
 */
770
int arch_add_memory(int nid, u64 start, u64 size)
771
{
772
	struct pglist_data *pgdat = NODE_DATA(nid);
773
	struct zone *zone = pgdat->node_zones + ZONE_NORMAL;
774
	unsigned long last_mapped_pfn, start_pfn = start >> PAGE_SHIFT;
775 776 777
	unsigned long nr_pages = size >> PAGE_SHIFT;
	int ret;

778 779 780
	last_mapped_pfn = init_memory_mapping(start, start + size-1);
	if (last_mapped_pfn > max_pfn_mapped)
		max_pfn_mapped = last_mapped_pfn;
781

782
	ret = __add_pages(zone, start_pfn, nr_pages);
783
	WARN_ON(1);
784 785 786

	return ret;
}
787
EXPORT_SYMBOL_GPL(arch_add_memory);
788

789
#if !defined(CONFIG_ACPI_NUMA) && defined(CONFIG_NUMA)
790 791 792 793
int memory_add_physaddr_to_nid(u64 start)
{
	return 0;
}
794
EXPORT_SYMBOL_GPL(memory_add_physaddr_to_nid);
795 796
#endif

797 798
#endif /* CONFIG_MEMORY_HOTPLUG */

799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818
/*
 * devmem_is_allowed() checks to see if /dev/mem access to a certain address
 * is valid. The argument is a physical page number.
 *
 *
 * On x86, access has to be given to the first megabyte of ram because that area
 * contains bios code and data regions used by X and dosemu and similar apps.
 * Access has to be given to non-kernel-ram areas as well, these contain the PCI
 * mmio resources as well as potential bios/acpi data regions.
 */
int devmem_is_allowed(unsigned long pagenr)
{
	if (pagenr <= 256)
		return 1;
	if (!page_is_ram(pagenr))
		return 1;
	return 0;
}


T
Thomas Gleixner 已提交
819 820
static struct kcore_list kcore_mem, kcore_vmalloc, kcore_kernel,
			 kcore_modules, kcore_vsyscall;
L
Linus Torvalds 已提交
821 822 823

void __init mem_init(void)
{
824
	long codesize, reservedpages, datasize, initsize;
L
Linus Torvalds 已提交
825

826
	pci_iommu_alloc();
L
Linus Torvalds 已提交
827

828
	/* clear_bss() already clear the empty_zero_page */
L
Linus Torvalds 已提交
829 830 831 832

	reservedpages = 0;

	/* this will put all low memory onto the freelists */
833
#ifdef CONFIG_NUMA
834
	totalram_pages = numa_free_all_bootmem();
L
Linus Torvalds 已提交
835
#else
836
	totalram_pages = free_all_bootmem();
L
Linus Torvalds 已提交
837
#endif
Y
Yinghai Lu 已提交
838 839
	reservedpages = max_pfn - totalram_pages -
					absent_pages_in_range(0, max_pfn);
L
Linus Torvalds 已提交
840 841 842 843 844 845 846
	after_bootmem = 1;

	codesize =  (unsigned long) &_etext - (unsigned long) &_text;
	datasize =  (unsigned long) &_edata - (unsigned long) &_etext;
	initsize =  (unsigned long) &__init_end - (unsigned long) &__init_begin;

	/* Register memory areas for /proc/kcore */
T
Thomas Gleixner 已提交
847 848
	kclist_add(&kcore_mem, __va(0), max_low_pfn << PAGE_SHIFT);
	kclist_add(&kcore_vmalloc, (void *)VMALLOC_START,
L
Linus Torvalds 已提交
849 850 851
		   VMALLOC_END-VMALLOC_START);
	kclist_add(&kcore_kernel, &_stext, _end - _stext);
	kclist_add(&kcore_modules, (void *)MODULES_VADDR, MODULES_LEN);
T
Thomas Gleixner 已提交
852
	kclist_add(&kcore_vsyscall, (void *)VSYSCALL_START,
L
Linus Torvalds 已提交
853 854
				 VSYSCALL_END - VSYSCALL_START);

855
	printk(KERN_INFO "Memory: %luk/%luk available (%ldk kernel code, "
T
Thomas Gleixner 已提交
856
				"%ldk reserved, %ldk data, %ldk init)\n",
L
Linus Torvalds 已提交
857
		(unsigned long) nr_free_pages() << (PAGE_SHIFT-10),
Y
Yinghai Lu 已提交
858
		max_pfn << (PAGE_SHIFT-10),
L
Linus Torvalds 已提交
859 860 861 862
		codesize >> 10,
		reservedpages << (PAGE_SHIFT-10),
		datasize >> 10,
		initsize >> 10);
863 864

	cpa_init();
L
Linus Torvalds 已提交
865 866
}

867
void free_init_pages(char *what, unsigned long begin, unsigned long end)
L
Linus Torvalds 已提交
868
{
869
	unsigned long addr = begin;
L
Linus Torvalds 已提交
870

871
	if (addr >= end)
872 873
		return;

I
Ingo Molnar 已提交
874 875 876 877 878 879 880 881 882 883
	/*
	 * If debugging page accesses then do not free this memory but
	 * mark them not present - any buggy init-section access will
	 * create a kernel page fault:
	 */
#ifdef CONFIG_DEBUG_PAGEALLOC
	printk(KERN_INFO "debug: unmapping init memory %08lx..%08lx\n",
		begin, PAGE_ALIGN(end));
	set_memory_np(begin, (end - begin) >> PAGE_SHIFT);
#else
884
	printk(KERN_INFO "Freeing %s: %luk freed\n", what, (end - begin) >> 10);
T
Thomas Gleixner 已提交
885

886
	for (; addr < end; addr += PAGE_SIZE) {
887 888 889 890 891
		ClearPageReserved(virt_to_page(addr));
		init_page_count(virt_to_page(addr));
		memset((void *)(addr & ~(PAGE_SIZE-1)),
			POISON_FREE_INITMEM, PAGE_SIZE);
		free_page(addr);
L
Linus Torvalds 已提交
892 893
		totalram_pages++;
	}
I
Ingo Molnar 已提交
894
#endif
895 896 897 898 899
}

void free_initmem(void)
{
	free_init_pages("unused kernel memory",
900 901
			(unsigned long)(&__init_begin),
			(unsigned long)(&__init_end));
L
Linus Torvalds 已提交
902 903
}

904
#ifdef CONFIG_DEBUG_RODATA
905 906
const int rodata_test_data = 0xC3;
EXPORT_SYMBOL_GPL(rodata_test_data);
907 908 909

void mark_rodata_ro(void)
{
910
	unsigned long start = PFN_ALIGN(_stext), end = PFN_ALIGN(__end_rodata);
911 912 913 914 915 916 917
	unsigned long rodata_start =
		((unsigned long)__start_rodata + PAGE_SIZE - 1) & PAGE_MASK;

#ifdef CONFIG_DYNAMIC_FTRACE
	/* Dynamic tracing modifies the kernel text section */
	start = rodata_start;
#endif
918

919
	printk(KERN_INFO "Write protecting the kernel read-only data: %luk\n",
920
	       (end - start) >> 10);
921 922 923 924 925 926
	set_memory_ro(start, (end - start) >> PAGE_SHIFT);

	/*
	 * The rodata section (but not the kernel text!) should also be
	 * not-executable.
	 */
927
	set_memory_nx(rodata_start, (end - rodata_start) >> PAGE_SHIFT);
928

929 930
	rodata_test();

931
#ifdef CONFIG_CPA_DEBUG
932
	printk(KERN_INFO "Testing CPA: undo %lx-%lx\n", start, end);
933
	set_memory_rw(start, (end-start) >> PAGE_SHIFT);
934

935
	printk(KERN_INFO "Testing CPA: again\n");
936
	set_memory_ro(start, (end-start) >> PAGE_SHIFT);
937
#endif
938
}
939

940 941
#endif

L
Linus Torvalds 已提交
942 943 944
#ifdef CONFIG_BLK_DEV_INITRD
void free_initrd_mem(unsigned long start, unsigned long end)
{
945
	free_init_pages("initrd memory", start, end);
L
Linus Torvalds 已提交
946 947 948
}
#endif

949 950
int __init reserve_bootmem_generic(unsigned long phys, unsigned long len,
				   int flags)
T
Thomas Gleixner 已提交
951
{
952
#ifdef CONFIG_NUMA
953
	int nid, next_nid;
954
	int ret;
955 956
#endif
	unsigned long pfn = phys >> PAGE_SHIFT;
T
Thomas Gleixner 已提交
957

Y
Yinghai Lu 已提交
958
	if (pfn >= max_pfn) {
T
Thomas Gleixner 已提交
959 960 961 962
		/*
		 * This can happen with kdump kernels when accessing
		 * firmware tables:
		 */
963
		if (pfn < max_pfn_mapped)
964
			return -EFAULT;
T
Thomas Gleixner 已提交
965

966
		printk(KERN_ERR "reserve_bootmem: illegal reserve %lx %lu\n",
967
				phys, len);
968
		return -EFAULT;
969 970 971 972
	}

	/* Should check here against the e820 map to avoid double free */
#ifdef CONFIG_NUMA
973 974 975
	nid = phys_to_nid(phys);
	next_nid = phys_to_nid(phys + len - 1);
	if (nid == next_nid)
976
		ret = reserve_bootmem_node(NODE_DATA(nid), phys, len, flags);
977
	else
978 979 980 981 982
		ret = reserve_bootmem(phys, len, flags);

	if (ret != 0)
		return ret;

T
Thomas Gleixner 已提交
983
#else
984
	reserve_bootmem(phys, len, BOOTMEM_DEFAULT);
L
Linus Torvalds 已提交
985
#endif
986

987
	if (phys+len <= MAX_DMA_PFN*PAGE_SIZE) {
988
		dma_reserve += len / PAGE_SIZE;
989 990
		set_dma_reserve(dma_reserve);
	}
991 992

	return 0;
L
Linus Torvalds 已提交
993 994
}

T
Thomas Gleixner 已提交
995 996
int kern_addr_valid(unsigned long addr)
{
L
Linus Torvalds 已提交
997
	unsigned long above = ((long)addr) >> __VIRTUAL_MASK_SHIFT;
T
Thomas Gleixner 已提交
998 999 1000 1001
	pgd_t *pgd;
	pud_t *pud;
	pmd_t *pmd;
	pte_t *pte;
L
Linus Torvalds 已提交
1002 1003

	if (above != 0 && above != -1UL)
T
Thomas Gleixner 已提交
1004 1005
		return 0;

L
Linus Torvalds 已提交
1006 1007 1008 1009 1010 1011
	pgd = pgd_offset_k(addr);
	if (pgd_none(*pgd))
		return 0;

	pud = pud_offset(pgd, addr);
	if (pud_none(*pud))
T
Thomas Gleixner 已提交
1012
		return 0;
L
Linus Torvalds 已提交
1013 1014 1015 1016

	pmd = pmd_offset(pud, addr);
	if (pmd_none(*pmd))
		return 0;
T
Thomas Gleixner 已提交
1017

L
Linus Torvalds 已提交
1018 1019 1020 1021 1022 1023
	if (pmd_large(*pmd))
		return pfn_valid(pmd_pfn(*pmd));

	pte = pte_offset_kernel(pmd, addr);
	if (pte_none(*pte))
		return 0;
T
Thomas Gleixner 已提交
1024

L
Linus Torvalds 已提交
1025 1026 1027
	return pfn_valid(pte_pfn(*pte));
}

T
Thomas Gleixner 已提交
1028 1029 1030 1031 1032
/*
 * A pseudo VMA to allow ptrace access for the vsyscall page.  This only
 * covers the 64bit vsyscall page now. 32bit has a real VMA now and does
 * not need special handling anymore:
 */
L
Linus Torvalds 已提交
1033
static struct vm_area_struct gate_vma = {
T
Thomas Gleixner 已提交
1034 1035 1036 1037
	.vm_start	= VSYSCALL_START,
	.vm_end		= VSYSCALL_START + (VSYSCALL_MAPPED_PAGES * PAGE_SIZE),
	.vm_page_prot	= PAGE_READONLY_EXEC,
	.vm_flags	= VM_READ | VM_EXEC
L
Linus Torvalds 已提交
1038 1039 1040 1041 1042
};

struct vm_area_struct *get_gate_vma(struct task_struct *tsk)
{
#ifdef CONFIG_IA32_EMULATION
1043 1044
	if (test_tsk_thread_flag(tsk, TIF_IA32))
		return NULL;
L
Linus Torvalds 已提交
1045 1046 1047 1048 1049 1050 1051
#endif
	return &gate_vma;
}

int in_gate_area(struct task_struct *task, unsigned long addr)
{
	struct vm_area_struct *vma = get_gate_vma(task);
T
Thomas Gleixner 已提交
1052

1053 1054
	if (!vma)
		return 0;
T
Thomas Gleixner 已提交
1055

L
Linus Torvalds 已提交
1056 1057 1058
	return (addr >= vma->vm_start) && (addr < vma->vm_end);
}

T
Thomas Gleixner 已提交
1059 1060 1061 1062
/*
 * Use this when you have no reliable task/vma, typically from interrupt
 * context. It is less reliable than using the task's vma and may give
 * false positives:
L
Linus Torvalds 已提交
1063 1064 1065
 */
int in_gate_area_no_task(unsigned long addr)
{
1066
	return (addr >= VSYSCALL_START) && (addr < VSYSCALL_END);
L
Linus Torvalds 已提交
1067
}
1068

1069 1070 1071 1072 1073 1074 1075 1076
const char *arch_vma_name(struct vm_area_struct *vma)
{
	if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso)
		return "[vdso]";
	if (vma == &gate_vma)
		return "[vsyscall]";
	return NULL;
}
1077 1078 1079 1080 1081

#ifdef CONFIG_SPARSEMEM_VMEMMAP
/*
 * Initialise the sparsemem vmemmap using huge-pages at the PMD level.
 */
1082 1083 1084 1085
static long __meminitdata addr_start, addr_end;
static void __meminitdata *p_start, *p_end;
static int __meminitdata node_start;

T
Thomas Gleixner 已提交
1086 1087
int __meminit
vmemmap_populate(struct page *start_page, unsigned long size, int node)
1088 1089 1090 1091 1092 1093 1094 1095 1096
{
	unsigned long addr = (unsigned long)start_page;
	unsigned long end = (unsigned long)(start_page + size);
	unsigned long next;
	pgd_t *pgd;
	pud_t *pud;
	pmd_t *pmd;

	for (; addr < end; addr = next) {
1097
		void *p = NULL;
1098 1099 1100 1101

		pgd = vmemmap_pgd_populate(addr, node);
		if (!pgd)
			return -ENOMEM;
T
Thomas Gleixner 已提交
1102

1103 1104 1105 1106
		pud = vmemmap_pud_populate(pgd, addr, node);
		if (!pud)
			return -ENOMEM;

1107 1108 1109 1110 1111 1112 1113 1114
		if (!cpu_has_pse) {
			next = (addr + PAGE_SIZE) & PAGE_MASK;
			pmd = vmemmap_pmd_populate(pud, addr, node);

			if (!pmd)
				return -ENOMEM;

			p = vmemmap_pte_populate(pmd, addr, node);
T
Thomas Gleixner 已提交
1115

1116 1117 1118
			if (!p)
				return -ENOMEM;

1119 1120
			addr_end = addr + PAGE_SIZE;
			p_end = p + PAGE_SIZE;
T
Thomas Gleixner 已提交
1121
		} else {
1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144
			next = pmd_addr_end(addr, end);

			pmd = pmd_offset(pud, addr);
			if (pmd_none(*pmd)) {
				pte_t entry;

				p = vmemmap_alloc_block(PMD_SIZE, node);
				if (!p)
					return -ENOMEM;

				entry = pfn_pte(__pa(p) >> PAGE_SHIFT,
						PAGE_KERNEL_LARGE);
				set_pmd(pmd, __pmd(pte_val(entry)));

				/* check to see if we have contiguous blocks */
				if (p_end != p || node_start != node) {
					if (p_start)
						printk(KERN_DEBUG " [%lx-%lx] PMD -> [%p-%p] on node %d\n",
						       addr_start, addr_end-1, p_start, p_end-1, node_start);
					addr_start = addr;
					node_start = node;
					p_start = p;
				}
Y
Yinghai Lu 已提交
1145 1146 1147

				addr_end = addr + PMD_SIZE;
				p_end = p + PMD_SIZE;
1148 1149
			} else
				vmemmap_verify((pte_t *)pmd, node, addr, next);
T
Thomas Gleixner 已提交
1150
		}
1151

1152 1153 1154
	}
	return 0;
}
1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165

void __meminit vmemmap_populate_print_last(void)
{
	if (p_start) {
		printk(KERN_DEBUG " [%lx-%lx] PMD -> [%p-%p] on node %d\n",
			addr_start, addr_end-1, p_start, p_end-1, node_start);
		p_start = NULL;
		p_end = NULL;
		node_start = 0;
	}
}
1166
#endif